java - map reduce 程序在线程 "main"java.io.IOException : Job failed 中显示错误异常

标签 java apache hadoop mapreduce

我正在尝试运行我的 map reduce 程序。在我尝试运行它之后,输出如下。 (我只显示输出的最后一部分)

File System Counters
    FILE: Number of bytes read=3052
    FILE: Number of bytes written=224295
    FILE: Number of read operations=0
    FILE: Number of large read operations=0
    FILE: Number of write operations=0
    HDFS: Number of bytes read=0
    HDFS: Number of bytes written=0
    HDFS: Number of read operations=5
    HDFS: Number of large read operations=0
    HDFS: Number of write operations=1
Map-Reduce Framework
    Map input records=4
    Map output records=4
    Map output bytes=120
    Map output materialized bytes=0
    Input split bytes=97
    Combine input records=0
    Combine output records=0
    Spilled Records=0
    Failed Shuffles=0
    Merged Map outputs=0
    GC time elapsed (ms)=40
    CPU time spent (ms)=0
    Physical memory (bytes) snapshot=0
    Virtual memory (bytes) snapshot=0
    Total committed heap usage (bytes)=117927936
File Input Format Counters 
    Bytes Read=272
Exception in thread "main" java.io.IOException: Job failed!
at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:836)
at mapreduceprogram.main(mapreduceprog.java:68)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:483)
at org.apache.hadoop.util.RunJar.main(RunJar.java:212)

不知道哪里出错了 有什么帮助吗?

我的主要方法内容:

public static void main(String[] args) throws Exception {
     JobConf conf = new JobConf(mapreduceprog.class);
     conf.setJobName("mapreduceprog");

     conf.setOutputKeyClass(Text.class);
     conf.setOutputValueClass(IntWritable.class);

         conf.setMapOutputKeyClass(Text.class);
         conf.setMapOutputValueClass(Text.class);

     conf.setMapperClass(Map.class);
     conf.setCombinerClass(Reduce.class);
     conf.setReducerClass(Reduce.class);

     conf.setInputFormat(TextInputFormat.class);
     conf.setOutputFormat(TextOutputFormat.class);

     FileInputFormat.setInputPaths(conf, new Path(args[0]));
     FileOutputFormat.setOutputPath(conf, new Path(args[1]));

     JobClient.runJob(conf);

我的第68行是

JobClient.runJob(conf);

最佳答案

您使用的是较旧的 Api。我建议你使用更新的 Api。代码看起来像这样

import java.io.File;
import java.io.IOException;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.lib.input.FileInputFormat;
import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;



public class MyDriver {
    public static void main(String[] args) throws IOException, InterruptedException, ClassNotFoundException {
        if(args.length!=2){
            System.out.println("Error");

            System.exit(-1);
        }
        Job job=new Job();
        job.setJarByClass(MyDriver.class);
        job.setMapperClass(Map.class);
        job.setReducerClass(Reduce.class);
        job.setMapOutputKeyClass(Text.class);
        job.setMapOutputValueClass(Text.class);
        job.setOutputKeyClass(Text.class); /*Reducer Output Key and value class*/
        job.setOutputValueClass(NullWritable.class);
        job.setInputFormatClass(CustomInputFormat.class);
        FileInputFormat.setInputPaths(job, new Path(args[0]));
        FileOutputFormat.setOutputPath(job, new Path(args[2]));
        boolean success=job.waitForCompletion(true);
        System.exit(success?0:-1);
    }

}

关于java - map reduce 程序在线程 "main"java.io.IOException : Job failed 中显示错误异常,我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/26418768/

相关文章:

java - 什么是 JAXB 生成的 package-info.java

java - 在Java中从 "Document"替换文件中的XML

php - Silex微框架: Run from a subfolder

php - 如何确保只有一个 PHP 脚本实例通过 Apache 运行?

java - 如何让 Hive JDBC 驱动程序和 Hive shell 与同一个数据库通信?

hadoop - 无法使用hbase-shaded-client jar,因为其内部依赖于log4j-1.2.17(CVE-2019-1757)

java - 如何在 Map 中控制 JPA 列名

java - 有什么方法可以像谷歌或雅虎那样实时获取股票报价吗?

php - 使用 Docker 的 Nginx 504 网关超时

apache - 带有Nutch的Apache Hadoop 2.7.4尝试由于任何原因失败