java - apache.hadoop.mapred.JobClient.runJob失败

标签 java ubuntu hadoop mapreduce

我在Ubuntu中使用标准的单独mapreduce作业进行了测试。
但是出现以下错误,然后失败。

conf.myconf是从conf.empty副本创建的

ubuntu@ip-172-31-20-2:/etc/hadoop/conf.myconf$ hadoop jar /usr/lib/hadoop-0.20-mapreduce/hadoop-examples-2.0.0-mr1-cdh4.7.0.jar grep ~/hadoop/input ~/hadoop/output '吾輩'

16/10/16 09:56:45 WARN conf.Configuration: session.id is deprecated. Instead, use dfs.metrics.session-id
16/10/16 09:56:45 INFO jvm.JvmMetrics: Initializing JVM Metrics with processName=JobTracker, sessionId=
16/10/16 09:56:45 WARN mapred.JobClient: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same.
16/10/16 09:56:45 INFO mapred.FileInputFormat: Total input paths to process : 2
16/10/16 09:56:46 INFO mapred.JobClient: Running job: job_local1816293773_0001
16/10/16 09:56:46 INFO mapred.LocalJobRunner: OutputCommitter set in config null
16/10/16 09:56:46 INFO mapred.LocalJobRunner: OutputCommitter is org.apache.hadoop.mapred.FileOutputCommitter
16/10/16 09:56:46 ERROR mapred.FileOutputCommitter: Mkdirs failed to create file:/etc/hadoop/conf.myconf/grep-temp-607671240/_temporary
16/10/16 09:56:46 INFO mapred.LocalJobRunner: Waiting for map tasks
16/10/16 09:56:46 INFO mapred.LocalJobRunner: Starting task: attempt_local1816293773_0001_m_000000_0
16/10/16 09:56:46 WARN mapreduce.Counters: Group org.apache.hadoop.mapred.Task$Counter is deprecated. Use org.apache.hadoop.mapreduce.TaskCounter instead
16/10/16 09:56:46 INFO util.ProcessTree: setsid exited with exit code 0
16/10/16 09:56:46 INFO mapred.Task:  Using ResourceCalculatorPlugin : org.apache.hadoop.util.LinuxResourceCalculatorPlugin@1704878c
16/10/16 09:56:46 INFO mapred.MapTask: Processing split: file:/home/ubuntu/hadoop/input/wagahaiwa_nekodearu.utf8.txt:0+1120636
16/10/16 09:56:46 WARN mapreduce.Counters: Counter name MAP_INPUT_BYTES is deprecated. Use FileInputFormatCounters as group name and  BYTES_READ as counter name instead
16/10/16 09:56:46 INFO mapred.MapTask: numReduceTasks: 1
16/10/16 09:56:46 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
16/10/16 09:56:46 INFO mapred.MapTask: io.sort.mb = 100
16/10/16 09:56:46 INFO mapred.MapTask: data buffer = 79691776/99614720
16/10/16 09:56:46 INFO mapred.MapTask: record buffer = 262144/327680
16/10/16 09:56:46 INFO mapred.MapTask: Starting flush of map output
16/10/16 09:56:46 INFO mapred.MapTask: Finished spill 0
16/10/16 09:56:46 INFO mapred.Task: Task:attempt_local1816293773_0001_m_000000_0 is done. And is in the process of commiting
16/10/16 09:56:46 INFO mapred.LocalJobRunner: file:/home/ubuntu/hadoop/input/wagahaiwa_nekodearu.utf8.txt:0+1120636
16/10/16 09:56:46 INFO mapred.Task: Task 'attempt_local1816293773_0001_m_000000_0' done.
16/10/16 09:56:46 INFO mapred.LocalJobRunner: Finishing task: attempt_local1816293773_0001_m_000000_0
16/10/16 09:56:46 INFO mapred.LocalJobRunner: Starting task: attempt_local1816293773_0001_m_000001_0
16/10/16 09:56:46 WARN mapreduce.Counters: Group org.apache.hadoop.mapred.Task$Counter is deprecated. Use org.apache.hadoop.mapreduce.TaskCounter instead
16/10/16 09:56:46 INFO mapred.Task:  Using ResourceCalculatorPlugin : org.apache.hadoop.util.LinuxResourceCalculatorPlugin@76ae327a
16/10/16 09:56:46 INFO mapred.MapTask: Processing split: file:/home/ubuntu/hadoop/input/wagahaiwa_nekodearu.txt:0+748959
16/10/16 09:56:46 WARN mapreduce.Counters: Counter name MAP_INPUT_BYTES is deprecated. Use FileInputFormatCounters as group name and  BYTES_READ as counter name instead
16/10/16 09:56:46 INFO mapred.MapTask: numReduceTasks: 1
16/10/16 09:56:46 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
16/10/16 09:56:46 INFO mapred.MapTask: io.sort.mb = 100
16/10/16 09:56:46 INFO mapred.MapTask: data buffer = 79691776/99614720
16/10/16 09:56:46 INFO mapred.MapTask: record buffer = 262144/327680
16/10/16 09:56:47 INFO mapred.MapTask: Starting flush of map output
16/10/16 09:56:47 INFO mapred.Task: Task:attempt_local1816293773_0001_m_000001_0 is done. And is in the process of commiting
16/10/16 09:56:47 INFO mapred.LocalJobRunner: file:/home/ubuntu/hadoop/input/wagahaiwa_nekodearu.txt:0+748959
16/10/16 09:56:47 INFO mapred.Task: Task 'attempt_local1816293773_0001_m_000001_0' done.
16/10/16 09:56:47 INFO mapred.LocalJobRunner: Finishing task: attempt_local1816293773_0001_m_000001_0
16/10/16 09:56:47 INFO mapred.LocalJobRunner: Map task executor complete.
16/10/16 09:56:47 WARN mapreduce.Counters: Group org.apache.hadoop.mapred.Task$Counter is deprecated. Use org.apache.hadoop.mapreduce.TaskCounter instead
16/10/16 09:56:47 INFO mapred.Task:  Using ResourceCalculatorPlugin : org.apache.hadoop.util.LinuxResourceCalculatorPlugin@778b3dc2
16/10/16 09:56:47 INFO mapred.LocalJobRunner:
16/10/16 09:56:47 INFO mapred.Merger: Merging 2 sorted segments
16/10/16 09:56:47 INFO mapred.Merger: Down to the last merge-pass, with 1 segments left of total size: 19 bytes
16/10/16 09:56:47 INFO mapred.LocalJobRunner:
16/10/16 09:56:47 WARN mapred.LocalJobRunner: job_local1816293773_0001
java.io.IOException: The temporary job-output directory file:/etc/hadoop/conf.myconf/grep-temp-607671240/_temporary doesn't exist!
    at org.apache.hadoop.mapred.FileOutputCommitter.getWorkPath(FileOutputCommitter.java:250)
    at org.apache.hadoop.mapred.FileOutputFormat.getTaskOutputPath(FileOutputFormat.java:240)
    at org.apache.hadoop.mapred.SequenceFileOutputFormat.getRecordWriter(SequenceFileOutputFormat.java:44)
    at org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:476)
    at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:447)
    at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:449)
16/10/16 09:56:47 INFO mapred.JobClient:  map 100% reduce 0%
16/10/16 09:56:47 INFO mapred.JobClient: Job complete: job_local1816293773_0001
16/10/16 09:56:47 INFO mapred.JobClient: Counters: 21
16/10/16 09:56:47 INFO mapred.JobClient:   File System Counters
16/10/16 09:56:47 INFO mapred.JobClient:     FILE: Number of bytes read=3276711
16/10/16 09:56:47 INFO mapred.JobClient:     FILE: Number of bytes written=479842
16/10/16 09:56:47 INFO mapred.JobClient:     FILE: Number of read operations=0
16/10/16 09:56:47 INFO mapred.JobClient:     FILE: Number of large read operations=0
16/10/16 09:56:47 INFO mapred.JobClient:     FILE: Number of write operations=0
16/10/16 09:56:47 INFO mapred.JobClient:   Map-Reduce Framework
16/10/16 09:56:47 INFO mapred.JobClient:     Map input records=4748
16/10/16 09:56:47 INFO mapred.JobClient:     Map output records=483
16/10/16 09:56:47 INFO mapred.JobClient:     Map output bytes=7245
16/10/16 09:56:47 INFO mapred.JobClient:     Input split bytes=219
16/10/16 09:56:47 INFO mapred.JobClient:     Combine input records=483
16/10/16 09:56:47 INFO mapred.JobClient:     Combine output records=1
16/10/16 09:56:47 INFO mapred.JobClient:     Reduce input groups=0
16/10/16 09:56:47 INFO mapred.JobClient:     Reduce shuffle bytes=0
16/10/16 09:56:47 INFO mapred.JobClient:     Reduce input records=0
16/10/16 09:56:47 INFO mapred.JobClient:     Reduce output records=0
16/10/16 09:56:47 INFO mapred.JobClient:     Spilled Records=1
16/10/16 09:56:47 INFO mapred.JobClient:     CPU time spent (ms)=0
16/10/16 09:56:47 INFO mapred.JobClient:     Physical memory (bytes) snapshot=0
16/10/16 09:56:47 INFO mapred.JobClient:     Virtual memory (bytes) snapshot=0
16/10/16 09:56:47 INFO mapred.JobClient:     Total committed heap usage (bytes)=538443776
16/10/16 09:56:47 INFO mapred.JobClient:   org.apache.hadoop.mapreduce.lib.input.FileInputFormatCounter
16/10/16 09:56:47 INFO mapred.JobClient:     BYTES_READ=1869595
16/10/16 09:56:47 INFO mapred.JobClient: Job Failed: NA
java.io.IOException: Job failed!
    at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1416)
    at org.apache.hadoop.examples.Grep.run(Grep.java:69)
    at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
    at org.apache.hadoop.examples.Grep.main(Grep.java:93)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
    at java.lang.reflect.Method.invoke(Method.java:498)
    at org.apache.hadoop.util.ProgramDriver$ProgramDescription.invoke(ProgramDriver.java:72)
    at org.apache.hadoop.util.ProgramDriver.driver(ProgramDriver.java:144)
    at org.apache.hadoop.examples.ExampleDriver.main(ExampleDriver.java:64)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
    at java.lang.reflect.Method.invoke(Method.java:498)
    at org.apache.hadoop.util.RunJar.main(RunJar.java:208)

但是,当我将用户从ubuntu更改为hdfs时,它可以正常工作。
为什么是这样?
hdfs@ip-172-31-20-2:~$ hadoop jar /usr/lib/hadoop-0.20-mapreduce/hadoop-examples-2.0.0-mr1-cdh4.7.0.jar grep /home/ubuntu/hadoop/input /home/ubuntu/hadoop/output '吾輩'

16/10/16 10:15:04 WARN conf.Configuration: session.id is deprecated. Instead, use dfs.metrics.session-id
16/10/16 10:15:04 INFO jvm.JvmMetrics: Initializing JVM Metrics with processName=JobTracker, sessionId=
16/10/16 10:15:04 WARN mapred.JobClient: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same.
16/10/16 10:15:04 INFO mapred.FileInputFormat: Total input paths to process : 2
16/10/16 10:15:04 INFO mapred.JobClient: Running job: job_local783267739_0001
16/10/16 10:15:04 INFO mapred.LocalJobRunner: OutputCommitter set in config null
16/10/16 10:15:04 INFO mapred.LocalJobRunner: OutputCommitter is org.apache.hadoop.mapred.FileOutputCommitter
16/10/16 10:15:05 INFO mapred.LocalJobRunner: Waiting for map tasks
16/10/16 10:15:05 INFO mapred.LocalJobRunner: Starting task: attempt_local783267739_0001_m_000000_0
16/10/16 10:15:05 WARN mapreduce.Counters: Group org.apache.hadoop.mapred.Task$Counter is deprecated. Use org.apache.hadoop.mapreduce.TaskCounter instead
16/10/16 10:15:05 INFO util.ProcessTree: setsid exited with exit code 0
16/10/16 10:15:05 INFO mapred.Task:  Using ResourceCalculatorPlugin : org.apache.hadoop.util.LinuxResourceCalculatorPlugin@56683a8d
16/10/16 10:15:05 INFO mapred.MapTask: Processing split: file:/home/ubuntu/hadoop/input/wagahaiwa_nekodearu.utf8.txt:0+1120636
16/10/16 10:15:05 WARN mapreduce.Counters: Counter name MAP_INPUT_BYTES is deprecated. Use FileInputFormatCounters as group name and  BYTES_READ as counter name instead
16/10/16 10:15:05 INFO mapred.MapTask: numReduceTasks: 1
16/10/16 10:15:05 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
16/10/16 10:15:05 INFO mapred.MapTask: io.sort.mb = 100
16/10/16 10:15:05 INFO mapred.MapTask: data buffer = 79691776/99614720
16/10/16 10:15:05 INFO mapred.MapTask: record buffer = 262144/327680
16/10/16 10:15:05 INFO mapred.MapTask: Starting flush of map output
16/10/16 10:15:05 INFO mapred.MapTask: Finished spill 0
16/10/16 10:15:05 INFO mapred.Task: Task:attempt_local783267739_0001_m_000000_0 is done. And is in the process of commiting
16/10/16 10:15:05 INFO mapred.LocalJobRunner: file:/home/ubuntu/hadoop/input/wagahaiwa_nekodearu.utf8.txt:0+1120636
16/10/16 10:15:05 INFO mapred.Task: Task 'attempt_local783267739_0001_m_000000_0' done.
16/10/16 10:15:05 INFO mapred.LocalJobRunner: Finishing task: attempt_local783267739_0001_m_000000_0
16/10/16 10:15:05 INFO mapred.LocalJobRunner: Starting task: attempt_local783267739_0001_m_000001_0
16/10/16 10:15:05 WARN mapreduce.Counters: Group org.apache.hadoop.mapred.Task$Counter is deprecated. Use org.apache.hadoop.mapreduce.TaskCounter instead
16/10/16 10:15:05 INFO mapred.Task:  Using ResourceCalculatorPlugin : org.apache.hadoop.util.LinuxResourceCalculatorPlugin@2a801853
16/10/16 10:15:05 INFO mapred.MapTask: Processing split: file:/home/ubuntu/hadoop/input/wagahaiwa_nekodearu.txt:0+748959
16/10/16 10:15:05 WARN mapreduce.Counters: Counter name MAP_INPUT_BYTES is deprecated. Use FileInputFormatCounters as group name and  BYTES_READ as counter name instead
16/10/16 10:15:05 INFO mapred.MapTask: numReduceTasks: 1
16/10/16 10:15:05 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
16/10/16 10:15:05 INFO mapred.MapTask: io.sort.mb = 100
16/10/16 10:15:05 INFO mapred.MapTask: data buffer = 79691776/99614720
16/10/16 10:15:05 INFO mapred.MapTask: record buffer = 262144/327680
16/10/16 10:15:05 INFO mapred.MapTask: Starting flush of map output
16/10/16 10:15:05 INFO mapred.Task: Task:attempt_local783267739_0001_m_000001_0 is done. And is in the process of commiting
16/10/16 10:15:05 INFO mapred.LocalJobRunner: file:/home/ubuntu/hadoop/input/wagahaiwa_nekodearu.txt:0+748959
16/10/16 10:15:05 INFO mapred.Task: Task 'attempt_local783267739_0001_m_000001_0' done.
16/10/16 10:15:05 INFO mapred.LocalJobRunner: Finishing task: attempt_local783267739_0001_m_000001_0
16/10/16 10:15:05 INFO mapred.LocalJobRunner: Map task executor complete.
16/10/16 10:15:05 WARN mapreduce.Counters: Group org.apache.hadoop.mapred.Task$Counter is deprecated. Use org.apache.hadoop.mapreduce.TaskCounter instead
16/10/16 10:15:05 INFO mapred.Task:  Using ResourceCalculatorPlugin : org.apache.hadoop.util.LinuxResourceCalculatorPlugin@6b9918ca
16/10/16 10:15:05 INFO mapred.LocalJobRunner:
16/10/16 10:15:05 INFO mapred.Merger: Merging 2 sorted segments
16/10/16 10:15:05 INFO mapred.Merger: Down to the last merge-pass, with 1 segments left of total size: 19 bytes
16/10/16 10:15:05 INFO mapred.LocalJobRunner:
16/10/16 10:15:05 INFO mapred.JobClient:  map 100% reduce 0%
16/10/16 10:15:05 INFO mapred.Task: Task:attempt_local783267739_0001_r_000000_0 is done. And is in the process of commiting
16/10/16 10:15:05 INFO mapred.LocalJobRunner:
16/10/16 10:15:05 INFO mapred.Task: Task attempt_local783267739_0001_r_000000_0 is allowed to commit now
16/10/16 10:15:06 INFO mapred.FileOutputCommitter: Saved output of task 'attempt_local783267739_0001_r_000000_0' to file:/var/lib/hadoop-hdfs/grep-temp-640163485
16/10/16 10:15:06 INFO mapred.LocalJobRunner: reduce > reduce
16/10/16 10:15:06 INFO mapred.Task: Task 'attempt_local783267739_0001_r_000000_0' done.
16/10/16 10:15:06 INFO mapred.JobClient:  map 100% reduce 100%
16/10/16 10:15:06 INFO mapred.JobClient: Job complete: job_local783267739_0001
16/10/16 10:15:07 INFO mapred.JobClient: Counters: 21
16/10/16 10:15:07 INFO mapred.JobClient:   File System Counters
16/10/16 10:15:07 INFO mapred.JobClient:     FILE: Number of bytes read=5289694
16/10/16 10:15:07 INFO mapred.JobClient:     FILE: Number of bytes written=719795
16/10/16 10:15:07 INFO mapred.JobClient:     FILE: Number of read operations=0
16/10/16 10:15:07 INFO mapred.JobClient:     FILE: Number of large read operations=0
16/10/16 10:15:07 INFO mapred.JobClient:     FILE: Number of write operations=0
16/10/16 10:15:07 INFO mapred.JobClient:   Map-Reduce Framework
16/10/16 10:15:07 INFO mapred.JobClient:     Map input records=4748
16/10/16 10:15:07 INFO mapred.JobClient:     Map output records=483
16/10/16 10:15:07 INFO mapred.JobClient:     Map output bytes=7245
16/10/16 10:15:07 INFO mapred.JobClient:     Input split bytes=219
16/10/16 10:15:07 INFO mapred.JobClient:     Combine input records=483
16/10/16 10:15:07 INFO mapred.JobClient:     Combine output records=1
16/10/16 10:15:07 INFO mapred.JobClient:     Reduce input groups=1
16/10/16 10:15:07 INFO mapred.JobClient:     Reduce shuffle bytes=0
16/10/16 10:15:07 INFO mapred.JobClient:     Reduce input records=1
16/10/16 10:15:07 INFO mapred.JobClient:     Reduce output records=1
16/10/16 10:15:07 INFO mapred.JobClient:     Spilled Records=2
16/10/16 10:15:07 INFO mapred.JobClient:     CPU time spent (ms)=0
16/10/16 10:15:07 INFO mapred.JobClient:     Physical memory (bytes) snapshot=0
16/10/16 10:15:07 INFO mapred.JobClient:     Virtual memory (bytes) snapshot=0
16/10/16 10:15:07 INFO mapred.JobClient:     Total committed heap usage (bytes)=853016576
16/10/16 10:15:07 INFO mapred.JobClient:   org.apache.hadoop.mapreduce.lib.input.FileInputFormatCounter
16/10/16 10:15:07 INFO mapred.JobClient:     BYTES_READ=1869595
16/10/16 10:15:07 INFO jvm.JvmMetrics: Cannot initialize JVM Metrics with processName=JobTracker, sessionId= - already initialized
16/10/16 10:15:07 WARN mapred.JobClient: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same.
16/10/16 10:15:07 INFO mapred.FileInputFormat: Total input paths to process : 1
16/10/16 10:15:07 INFO mapred.JobClient: Running job: job_local196240886_0002
16/10/16 10:15:07 INFO mapred.LocalJobRunner: OutputCommitter set in config null
16/10/16 10:15:07 INFO mapred.LocalJobRunner: OutputCommitter is org.apache.hadoop.mapred.FileOutputCommitter
16/10/16 10:15:07 INFO mapred.LocalJobRunner: Waiting for map tasks
16/10/16 10:15:07 INFO mapred.LocalJobRunner: Starting task: attempt_local196240886_0002_m_000000_0
16/10/16 10:15:07 WARN mapreduce.Counters: Group org.apache.hadoop.mapred.Task$Counter is deprecated. Use org.apache.hadoop.mapreduce.TaskCounter instead
16/10/16 10:15:07 INFO mapred.Task:  Using ResourceCalculatorPlugin : org.apache.hadoop.util.LinuxResourceCalculatorPlugin@5ad979dd
16/10/16 10:15:07 INFO mapred.MapTask: Processing split: file:/var/lib/hadoop-hdfs/grep-temp-640163485/part-00000:0+109
16/10/16 10:15:07 WARN mapreduce.Counters: Counter name MAP_INPUT_BYTES is deprecated. Use FileInputFormatCounters as group name and  BYTES_READ as counter name instead
16/10/16 10:15:07 INFO mapred.MapTask: numReduceTasks: 1
16/10/16 10:15:07 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
16/10/16 10:15:07 INFO mapred.MapTask: io.sort.mb = 100
16/10/16 10:15:07 INFO mapred.MapTask: data buffer = 79691776/99614720
16/10/16 10:15:07 INFO mapred.MapTask: record buffer = 262144/327680
16/10/16 10:15:07 INFO mapred.MapTask: Starting flush of map output
16/10/16 10:15:07 INFO mapred.MapTask: Finished spill 0
16/10/16 10:15:07 INFO mapred.Task: Task:attempt_local196240886_0002_m_000000_0 is done. And is in the process of commiting
16/10/16 10:15:07 INFO mapred.LocalJobRunner: file:/var/lib/hadoop-hdfs/grep-temp-640163485/part-00000:0+109
16/10/16 10:15:07 INFO mapred.Task: Task 'attempt_local196240886_0002_m_000000_0' done.
16/10/16 10:15:07 INFO mapred.LocalJobRunner: Finishing task: attempt_local196240886_0002_m_000000_0
16/10/16 10:15:07 INFO mapred.LocalJobRunner: Map task executor complete.
16/10/16 10:15:07 WARN mapreduce.Counters: Group org.apache.hadoop.mapred.Task$Counter is deprecated. Use org.apache.hadoop.mapreduce.TaskCounter instead
16/10/16 10:15:07 INFO mapred.Task:  Using ResourceCalculatorPlugin : org.apache.hadoop.util.LinuxResourceCalculatorPlugin@7ea3e75d
16/10/16 10:15:07 INFO mapred.LocalJobRunner:
16/10/16 10:15:07 INFO mapred.Merger: Merging 1 sorted segments
16/10/16 10:15:07 INFO mapred.Merger: Down to the last merge-pass, with 1 segments left of total size: 19 bytes
16/10/16 10:15:07 INFO mapred.LocalJobRunner:
16/10/16 10:15:07 INFO mapred.Task: Task:attempt_local196240886_0002_r_000000_0 is done. And is in the process of commiting
16/10/16 10:15:07 INFO mapred.LocalJobRunner:
16/10/16 10:15:07 INFO mapred.Task: Task attempt_local196240886_0002_r_000000_0 is allowed to commit now
16/10/16 10:15:07 INFO mapred.FileOutputCommitter: Saved output of task 'attempt_local196240886_0002_r_000000_0' to file:/home/ubuntu/hadoop/output
16/10/16 10:15:07 INFO mapred.LocalJobRunner: reduce > reduce
16/10/16 10:15:07 INFO mapred.Task: Task 'attempt_local196240886_0002_r_000000_0' done.
16/10/16 10:15:08 INFO mapred.JobClient:  map 100% reduce 100%
16/10/16 10:15:08 INFO mapred.JobClient: Job complete: job_local196240886_0002
16/10/16 10:15:08 INFO mapred.JobClient: Counters: 21
16/10/16 10:15:08 INFO mapred.JobClient:   File System Counters
16/10/16 10:15:08 INFO mapred.JobClient:     FILE: Number of bytes read=4312215
16/10/16 10:15:08 INFO mapred.JobClient:     FILE: Number of bytes written=957027
16/10/16 10:15:08 INFO mapred.JobClient:     FILE: Number of read operations=0
16/10/16 10:15:08 INFO mapred.JobClient:     FILE: Number of large read operations=0
16/10/16 10:15:08 INFO mapred.JobClient:     FILE: Number of write operations=0
16/10/16 10:15:08 INFO mapred.JobClient:   Map-Reduce Framework
16/10/16 10:15:08 INFO mapred.JobClient:     Map input records=1
16/10/16 10:15:08 INFO mapred.JobClient:     Map output records=1
16/10/16 10:15:08 INFO mapred.JobClient:     Map output bytes=15
16/10/16 10:15:08 INFO mapred.JobClient:     Input split bytes=109
16/10/16 10:15:08 INFO mapred.JobClient:     Combine input records=0
16/10/16 10:15:08 INFO mapred.JobClient:     Combine output records=0
16/10/16 10:15:08 INFO mapred.JobClient:     Reduce input groups=1
16/10/16 10:15:08 INFO mapred.JobClient:     Reduce shuffle bytes=0
16/10/16 10:15:08 INFO mapred.JobClient:     Reduce input records=1
16/10/16 10:15:08 INFO mapred.JobClient:     Reduce output records=1
16/10/16 10:15:08 INFO mapred.JobClient:     Spilled Records=2
16/10/16 10:15:08 INFO mapred.JobClient:     CPU time spent (ms)=0
16/10/16 10:15:08 INFO mapred.JobClient:     Physical memory (bytes) snapshot=0
16/10/16 10:15:08 INFO mapred.JobClient:     Virtual memory (bytes) snapshot=0
16/10/16 10:15:08 INFO mapred.JobClient:     Total committed heap usage (bytes)=898629632
16/10/16 10:15:08 INFO mapred.JobClient:   org.apache.hadoop.mapreduce.lib.input.FileInputFormatCounter
16/10/16 10:15:08 INFO mapred.JobClient:     BYTES_READ=23

最佳答案

The temporary job-output directory file:/etc/hadoop/conf.myconf/grep-temp-607671240/_temporary doesn't exist!

我会说ubuntu用户在那里没有访问权限。

关于java - apache.hadoop.mapred.JobClient.runJob失败,我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/40069491/

相关文章:

java - Aspectj:在 aop.xml 中声明时切入点不再起作用(LTW)

java - 使用 PDFbox 将 pdf 转换为 tiff 时,Tiff 图像文件为空白

macos - 在 Mac 中编译 Hadoop 示例

java - Hadoop> WordCount按值排序

scala - 在Hive表中加载 Parquet 文件将返回所有NULL

java - Struts:从 View 访问 bean 属性

java - 无法为 java.util.List Retrofit 2.0.0-beta2 创建转换器

postgresql - pg_restore : [archiver] unsupported version (1. 14) 在文件头中

ubuntu - Vala - fatal error : gtk/gtk. h - 使用 Gtk 库引发的错误

ubuntu - pyopencl 错误 : undefined symbol: clCreateSubDevices