java - Hadoop 2.5.0 作业不成功,流命令失败

标签 java hadoop logging jenkins

Hadoop版本:Hadoop 2.5.0-cdh5.3.1

ma​​pper和reducer脚本都是shell文件

部分打印日志:

AttemptID:attempt_1437751786759_1557_m_007335_0 Timed out after 600 secs 2015-08-21 19:46:55,837 INFO [main] mapreduce.Job

(Job.java:monitorAndPrintJob(1372)) - map 76% reduce 0% 2015-08-21 19:46:57,066 INFO [main] mapreduce.Job

(Job.java:monitorAndPrintJob(1372)) - map 100% reduce 100% 2015-08-21 19:47:03,159 INFO [main] mapreduce.Job

(Job.java:monitorAndPrintJob(1372)) - map 97% reduce 100% 2015-08-21 19:47:04,372 INFO [main] mapreduce.Job

(Job.java:monitorAndPrintJob(1372)) - map 100% reduce 100% 2015-08-21 19:47:04,794 INFO [main] mapreduce.Job

(Job.java:monitorAndPrintJob(1385)) - Job job_1437751786759_1557 failed with state FAILED due to: Task failed task_1437751786759_1557_m_001557 Job failed as tasks failed. failedMaps:1 failedReduces:0

2015-08-21 19:47:04,922 INFO [main] mapreduce.Job

(Job.java:monitorAndPrintJob(1390)) - Counters: 34

File System Counters

FILE: Number of bytes read=0

FILE: Number of bytes written=1415074916

FILE: Number of read operations=0

FILE: Number of large read operations=0

FILE: Number of write operations=0

HDFS: Number of bytes read=501146186

HDFS: Number of bytes written=0

HDFS: Number of read operations=22986

HDFS: Number of large read operations=0

HDFS: Number of write operations=0 Job Counters

Failed map tasks=1137

Killed map tasks=1483

Launched map tasks=10282

Other local map tasks=10438

Total time spent by all maps in occupied slots (ms)=10996762530

Total time spent by all reduces in occupied slots (ms)=0

Total time spent by all map tasks (ms)=5498381265

Total vcore-seconds taken by all map tasks=5498381265

Total megabyte-seconds taken by all map tasks=5630342415360

Map-Reduce Framework

Map input records=7662

Map output records=189860

Map output bytes=8829322

Map output materialized bytes=101153057

Input split bytes=988398

Combine input records=0

Spilled Records=189860

Failed Shuffles=0

Merged Map outputs=0

GC time elapsed (ms)=450437

CPU time spent (ms)=129978840

Physical memory (bytes) snapshot=3951235211264

Virtual memory (bytes) snapshot=13755897688064

Total committed heap usage (bytes)=3860902445056

File Input Format Counters

Bytes Read=500157788

2015-08-21 19:47:04,922 ERROR [main] streaming.StreamJob

(StreamJob.java:submitAndMonitorJob(1019)) - Job not successful!

Streaming Command Failed!

此外 在跟踪 url 中,得到这些日志:

++ date +%Y%m%d%H%M%S + /home/disk1/cloudera/parcels/CDH-5.3.1-1.cdh5.3.1.p0.5/lib/hadoop/bin/hadoop >dfs -D speed.limit.kb=9000 -put ./sites_url hdfs://nameservice1/user/rp->product/dma/newsites/url/ccdb/20150821185246..sites_url

DEPRECATED: Use of this script to execute hdfs command is deprecated. Instead use the hdfs command for it.

put: No lease on /user/rp-product/dma/newsites/url/ccdb/20150821185246..sites_url.COPYING (inode 913353): File does not exist. Holder DFSClient_NONMAPREDUCE_39002115_1 does not have any open files.

++ cat sele_url

++ wc -l

+ cn=32

+ (( 32>0 ))

+ cat sele_url

log4j:WARN No appenders could be found for logger

(org.apache.hadoop.metrics2.impl.MetricsSystemImpl).

log4j:WARN Please initialize the log4j system properly.

log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more info.

还有我还找到了hadoop的java源码Here

我在谷歌上搜索了解决方案但失败了,并且没有从日志中得到一些有用的信息来猜测一些可能的原因,所以我需要帮助或任何提示

非常感谢

此致!

最佳答案

也许是晚了,但现在我知道是什么导致了这个错误。在 map 中,从其他服务获得响应需要很长时间。说明map长时间不输出,job被kill掉然后失败

关于java - Hadoop 2.5.0 作业不成功,流命令失败,我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/32169436/

相关文章:

java - RxJava2 中的条件完成

java - E/FirebaseInstanceId : Failed to get FIS auth token

hadoop - 直接下载数据到HDFS?

python - 如何在 Python 日志记录中设置组合字段的宽度

java - 如何在 application-{profile}.properties 文件中包含 log4j-{profile}.properties 文件

java - 优化java 8流操作

java - 无法调试,0x00007c37 处的致命信号 6 (SIGABRT)(代码=-6)

hadoop - 分布式系统中语义网的推理

hadoop - MapReduce作业花费的时间

configuration - Log4Net 过滤器 "OR"