hadoop wordcount运行出错:解决方法

hadoop wordcount运行出错解决方法

第一次运行wordcount没有成功运行记录:

15/11/18 14:54:14 INFO client.RMProxy: Connecting to ResourceManager at master/125.216.231.188:8032
15/11/18 14:54:16 INFO input.FileInputFormat: Total input paths to process : 1
15/11/18 14:54:17 INFO mapreduce.JobSubmitter: number of splits:1
15/11/18 14:54:17 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_1447829571019_0001
15/11/18 14:54:17 INFO impl.YarnClientImpl: Submitted application application_1447829571019_0001
15/11/18 14:54:17 INFO mapreduce.Job: The url to track the job: http://master:8088/proxy/application_1447829571019_0001/
15/11/18 14:54:17 INFO mapreduce.Job: Running job: job_1447829571019_0001
15/11/18 14:54:22 INFO mapreduce.Job: Job job_1447829571019_0001 running in uber mode : false
15/11/18 14:54:22 INFO mapreduce.Job: map 0% reduce 0%
15/11/18 14:54:28 INFO mapreduce.Job: Task Id : attempt_1447829571019_0001_m_000000_0, Status : FAILED
Container [pid=4777,containerID=container_1447829571019_0001_01_000002] is running beyond virtual memory limits. Current usage: 70.7 MB of 1 GB physical memory used; 4.6 GB of 2.1 GB virtual memory used. Killing container.
Dump of the process-tree for container_1447829571019_0001_01_000002 :
|- PID PPID PGRPID SESSID CMD_NAME USER_MODE_TIME(MILLIS) SYSTEM_TIME(MILLIS) VMEM_USAGE(BYTES) RSSMEM_USAGE(PAGES) FULL_CMD_LINE
|- 4777 4602 4777 4777 (bash) 0 1 17047552 713 /bin/bash -c /home/hadoop/my_project/jdk1.7.0_79/bin/java -XX:-UseGCOverheadLimit -Xmx4096m -Djava.io.tmpdir=/home/hadoop/tmp/nm-local-dir/usercache/hadoop/appcache/application_1447829571019_0001/container_1447829571019_0001_01_000002/tmp -Dlog4j.configuration=container-log4j.properties -Dyarn.app.container.log.dir=/home/hadoop/my_project/hadoop-2.5.2/logs/userlogs/application_1447829571019_0001/container_1447829571019_0001_01_000002 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 125.216.231.188 49678 attempt_1447829571019_0001_m_000000_0 2 1>/home/hadoop/my_project/hadoop-2.5.2/logs/userlogs/application_1447829571019_0001/container_1447829571019_0001_01_000002/stdout 2>/home/hadoop/my_project/hadoop-2.5.2/logs/userlogs/application_1447829571019_0001/container_1447829571019_0001_01_000002/stderr
|- 4782 4777 4777 4777 (java) 189 8 4896141312 17389 /home/hadoop/my_project/jdk1.7.0_79/bin/java -XX:-UseGCOverheadLimit -Xmx4096m -Djava.io.tmpdir=/home/hadoop/tmp/nm-local-dir/usercache/hadoop/appcache/application_1447829571019_0001/container_1447829571019_0001_01_000002/tmp -Dlog4j.configuration=container-log4j.properties -Dyarn.app.container.log.dir=/home/hadoop/my_project/hadoop-2.5.2/logs/userlogs/application_1447829571019_0001/container_1447829571019_0001_01_000002 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 125.216.231.188 49678 attempt_1447829571019_0001_m_000000_0 2

Container killed on request. Exit code is 143
Container exited with a non-zero exit code 143

15/11/18 14:54:38 INFO mapreduce.Job: Task Id : attempt_1447829571019_0001_m_000000_1, Status : FAILED
Container [pid=5097,containerID=container_1447829571019_0001_01_000003] is running beyond virtual memory limits. Current usage: 62.1 MB of 1 GB physical memory used; 4.4 GB of 2.1 GB virtual memory used. Killing container.
Dump of the process-tree for container_1447829571019_0001_01_000003 :
|- PID PPID PGRPID SESSID CMD_NAME USER_MODE_TIME(MILLIS) SYSTEM_TIME(MILLIS) VMEM_USAGE(BYTES) RSSMEM_USAGE(PAGES) FULL_CMD_LINE
|- 5097 4932 5097 5097 (bash) 0 0 17047552 675 /bin/bash -c /home/hadoop/my_project/jdk1.7.0_79/bin/java -XX:-UseGCOverheadLimit -Xmx4096m -Djava.io.tmpdir=/home/hadoop/tmp/nm-local-dir/usercache/hadoop/appcache/application_1447829571019_0001/container_1447829571019_0001_01_000003/tmp -Dlog4j.configuration=container-log4j.properties -Dyarn.app.container.log.dir=/home/hadoop/my_project/hadoop-2.5.2/logs/userlogs/application_1447829571019_0001/container_1447829571019_0001_01_000003 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 125.216.231.188 49678 attempt_1447829571019_0001_m_000000_1 3 1>/home/hadoop/my_project/hadoop-2.5.2/logs/userlogs/application_1447829571019_0001/container_1447829571019_0001_01_000003/stdout 2>/home/hadoop/my_project/hadoop-2.5.2/logs/userlogs/application_1447829571019_0001/container_1447829571019_0001_01_000003/stderr
|- 5102 5097 5097 5097 (java) 279 8 4730163200 15229 /home/hadoop/my_project/jdk1.7.0_79/bin/java -XX:-UseGCOverheadLimit -Xmx4096m -Djava.io.tmpdir=/home/hadoop/tmp/nm-local-dir/usercache/hadoop/appcache/application_1447829571019_0001/container_1447829571019_0001_01_000003/tmp -Dlog4j.configuration=container-log4j.properties -Dyarn.app.container.log.dir=/home/hadoop/my_project/hadoop-2.5.2/logs/userlogs/application_1447829571019_0001/container_1447829571019_0001_01_000003 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 125.216.231.188 49678 attempt_1447829571019_0001_m_000000_1 3

Container killed on request. Exit code is 143
Container exited with a non-zero exit code 143

15/11/18 14:54:42 INFO mapreduce.Job: Task Id : attempt_1447829571019_0001_m_000000_2, Status : FAILED
Container [pid=5125,containerID=container_1447829571019_0001_01_000004] is running beyond virtual memory limits. Current usage: 36.5 MB of 1 GB physical memory used; 4.4 GB of 2.1 GB virtual memory used. Killing container.
Dump of the process-tree for container_1447829571019_0001_01_000004 :
|- PID PPID PGRPID SESSID CMD_NAME USER_MODE_TIME(MILLIS) SYSTEM_TIME(MILLIS) VMEM_USAGE(BYTES) RSSMEM_USAGE(PAGES) FULL_CMD_LINE
|- 5130 5125 5125 5125 (java) 26 1 4717449216 8664 /home/hadoop/my_project/jdk1.7.0_79/bin/java -XX:-UseGCOverheadLimit -Xmx4096m -Djava.io.tmpdir=/home/hadoop/tmp/nm-local-dir/usercache/hadoop/appcache/application_1447829571019_0001/container_1447829571019_0001_01_000004/tmp -Dlog4j.configuration=container-log4j.properties -Dyarn.app.container.log.dir=/home/hadoop/my_project/hadoop-2.5.2/logs/userlogs/application_1447829571019_0001/container_1447829571019_0001_01_000004 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 125.216.231.188 49678 attempt_1447829571019_0001_m_000000_2 4
|- 5125 4932 5125 5125 (bash) 0 0 17047552 691 /bin/bash -c /home/hadoop/my_project/jdk1.7.0_79/bin/java -XX:-UseGCOverheadLimit -Xmx4096m -Djava.io.tmpdir=/home/hadoop/tmp/nm-local-dir/usercache/hadoop/appcache/application_1447829571019_0001/container_1447829571019_0001_01_000004/tmp -Dlog4j.configuration=container-log4j.properties -Dyarn.app.container.log.dir=/home/hadoop/my_project/hadoop-2.5.2/logs/userlogs/application_1447829571019_0001/container_1447829571019_0001_01_000004 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 125.216.231.188 49678 attempt_1447829571019_0001_m_000000_2 4 1>/home/hadoop/my_project/hadoop-2.5.2/logs/userlogs/application_1447829571019_0001/container_1447829571019_0001_01_000004/stdout 2>/home/hadoop/my_project/hadoop-2.5.2/logs/userlogs/application_1447829571019_0001/container_1447829571019_0001_01_000004/stderr

Container killed on request. Exit code is 143
Container exited with a non-zero exit code 143

15/11/18 14:54:50 INFO mapreduce.Job: map 100% reduce 100%
15/11/18 14:54:52 INFO mapreduce.Job: Job job_1447829571019_0001 failed with state FAILED due to: Task failed task_1447829571019_0001_m_000000
Job failed as tasks failed. failedMaps:1 failedReduces:0

15/11/18 14:54:52 INFO mapreduce.Job: Counters: 12
Job Counters
Failed map tasks=4
Launched map tasks=4
Other local map tasks=3
Data-local map tasks=1
Total time spent by all maps in occupied slots (ms)=15352
Total time spent by all reduces in occupied slots (ms)=0
Total time spent by all map tasks (ms)=15352
Total vcore-seconds taken by all map tasks=15352
Total megabyte-seconds taken by all map tasks=15720448
Map-Reduce Framework
CPU time spent (ms)=0
Physical memory (bytes) snapshot=0
Virtual memory (bytes) snapshot=0

配置为

<configuration>

<Site specific YARN configuration properties -->
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>
<!--property>
<name>yarn.resourcemanager.address</name>
<value>master:8032</value>
</property>
<property>
<name>yarn.resourcemanager.scheduler.address</name>
<value>master:8030</value>
</property>
<property>
<name>yarn.resourcemanager.resource-tracker.address</name>
<value>master:8031</value>
</property>
<property>
<name>yarn.resourcemanager.admin.address</name>
<value>master:8033</value>
</property>
<property>
<name>yarn.resourcemanager.webapp.address</name>
<value>master:8088</value>
</property>
<property>
<name>yarn.nodemanager.vmem-pmem-ratio</name>
<value>2.9</value>
</property>

</configuration>

 

运行成功的运行记录:

15/11/18 15:11:57 INFO client.RMProxy: Connecting to ResourceManager at /0.0.0.0:8032
15/11/18 15:12:01 INFO input.FileInputFormat: Total input paths to process : 1
15/11/18 15:12:05 INFO mapreduce.JobSubmitter: number of splits:1
15/11/18 15:12:06 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_1447830680112_0001
15/11/18 15:12:06 INFO impl.YarnClientImpl: Submitted application application_1447830680112_0001
15/11/18 15:12:06 INFO mapreduce.Job: The url to track the job: http://master:8088/proxy/application_1447830680112_0001/
15/11/18 15:12:06 INFO mapreduce.Job: Running job: job_1447830680112_0001
15/11/18 15:12:11 INFO mapreduce.Job: Job job_1447830680112_0001 running in uber mode : false
15/11/18 15:12:11 INFO mapreduce.Job: map 0% reduce 0%
15/11/18 15:12:15 INFO mapreduce.Job: map 100% reduce 0%
15/11/18 15:12:18 INFO mapreduce.Job: Task Id : attempt_1447830680112_0001_r_000000_0, Status : FAILED
Container [pid=16292,containerID=container_1447830680112_0001_01_000003] is running beyond virtual memory limits. Current usage: 132.3 MB of 1 GB physical memory used; 4.6 GB of 2.1 GB virtual memory used. Killing container.
Dump of the process-tree for container_1447830680112_0001_01_000003 :
|- PID PPID PGRPID SESSID CMD_NAME USER_MODE_TIME(MILLIS) SYSTEM_TIME(MILLIS) VMEM_USAGE(BYTES) RSSMEM_USAGE(PAGES) FULL_CMD_LINE
|- 16292 15731 16292 16292 (bash) 0 0 17047552 694 /bin/bash -c /home/hadoop/my_project/jdk1.7.0_79/bin/java -Djava.net.preferIPv4Stack=true -Dhadoop.metrics.log.level=WARN -Xmx4096m -Djava.io.tmpdir=/home/hadoop/tmp/nm-local-dir/usercache/hadoop/appcache/application_1447830680112_0001/container_1447830680112_0001_01_000003/tmp -Dlog4j.configuration=container-log4j.properties -Dyarn.app.container.log.dir=/home/hadoop/my_project/hadoop-2.5.2/logs/userlogs/application_1447830680112_0001/container_1447830680112_0001_01_000003 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 125.216.231.188 39696 attempt_1447830680112_0001_r_000000_0 3 1>/home/hadoop/my_project/hadoop-2.5.2/logs/userlogs/application_1447830680112_0001/container_1447830680112_0001_01_000003/stdout 2>/home/hadoop/my_project/hadoop-2.5.2/logs/userlogs/application_1447830680112_0001/container_1447830680112_0001_01_000003/stderr
|- 16297 16292 16292 16292 (java) 275 2 4917329920 33178 /home/hadoop/my_project/jdk1.7.0_79/bin/java -Djava.net.preferIPv4Stack=true -Dhadoop.metrics.log.level=WARN -Xmx4096m -Djava.io.tmpdir=/home/hadoop/tmp/nm-local-dir/usercache/hadoop/appcache/application_1447830680112_0001/container_1447830680112_0001_01_000003/tmp -Dlog4j.configuration=container-log4j.properties -Dyarn.app.container.log.dir=/home/hadoop/my_project/hadoop-2.5.2/logs/userlogs/application_1447830680112_0001/container_1447830680112_0001_01_000003 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild 125.216.231.188 39696 attempt_1447830680112_0001_r_000000_0 3

Container killed on request. Exit code is 143
Container exited with a non-zero exit code 143

15/11/18 15:12:24 INFO mapreduce.Job: map 100% reduce 100%
15/11/18 15:12:27 INFO mapreduce.Job: Job job_1447830680112_0001 completed successfully
15/11/18 15:12:27 INFO mapreduce.Job: Counters: 50
File System Counters
FILE: Number of bytes read=81
FILE: Number of bytes written=195003
FILE: Number of read operations=0
FILE: Number of large read operations=0
FILE: Number of write operations=0
HDFS: Number of bytes read=174
HDFS: Number of bytes written=51
HDFS: Number of read operations=6
HDFS: Number of large read operations=0
HDFS: Number of write operations=2
Job Counters
Failed reduce tasks=1
Launched map tasks=1
Launched reduce tasks=2
Data-local map tasks=1
Total time spent by all maps in occupied slots (ms)=1661
Total time spent by all reduces in occupied slots (ms)=3933
Total time spent by all map tasks (ms)=1661
Total time spent by all reduce tasks (ms)=3933
Total vcore-seconds taken by all map tasks=1661
Total vcore-seconds taken by all reduce tasks=3933
Total megabyte-seconds taken by all map tasks=1700864
Total megabyte-seconds taken by all reduce tasks=4027392
Map-Reduce Framework
Map input records=5
Map output records=10
Map output bytes=106
Map output materialized bytes=81
Input split bytes=108
Combine input records=10
Combine output records=6
Reduce input groups=6
Reduce shuffle bytes=81
Reduce input records=6
Reduce output records=6
Spilled Records=12
Shuffled Maps =1
Failed Shuffles=0
Merged Map outputs=1
GC time elapsed (ms)=36
CPU time spent (ms)=920
Physical memory (bytes) snapshot=408248320
Virtual memory (bytes) snapshot=9894723584
Total committed heap usage (bytes)=421003264
Shuffle Errors
BAD_ID=0
CONNECTION=0
IO_ERROR=0
WRONG_LENGTH=0
WRONG_MAP=0
WRONG_REDUCE=0
File Input Format Counters
Bytes Read=66
File Output Format Counters
Bytes Written=51

配置为:

...
<property> 
<name>yarn.nodemanager.aux-services</name> 
<value>mapreduce_shuffle</value> 
</property> 

 

posted on 2015-11-18 15:32  小米辣  阅读(2611)  评论(0编辑  收藏  举报

导航