您的位置:首页 > 大数据

大数据平台之MapReduce

2019-06-24 15:34 1176 查看

1…在集群节点中/usr/hdp/2.4.3.0-227/hadoop-mapreduce/目录下,存在一个案例JAR 包hadoop-mapreduce-examples.jar。运行JAR包中的PI程序来进行计算圆周率π的近似值,要求运行5次Map任务,每个Map任务的投掷次数为5,运行完成后输出结果为。

[root@master ~]# hadoop jar/usr/hdp/2.4.3.0-227/hadoop-mapreduce/hadoop-mapreduce-   examples-2.7.1.2.4.3.0-227.jarpi 5 5

WARNING: Use "yarn jar" to launch YARNapplications.

Number of Maps = 5

Samples per Map = 5

Wrote input for Map #0

Wrote input for Map #1

Wrote input for Map #2

Wrote input for Map #3

Wrote input for Map #4

Starting Job

17/05/07 03:25:16 INFO impl.TimelineClientImpl:Timeline service address:      http://slaver1:8188/ws/v1/timeline/

17/05/07 03:25:16 INFO client.RMProxy: Connecting toResourceManager at slaver1/10.0.0.15:8050

17/05/07 03:25:17 INFO input.FileInputFormat: Totalinput paths to process : 5

17/05/07 03:25:17 INFO mapreduce.JobSubmitter: numberof splits:5

17/05/07 03:25:18 INFO mapreduce.JobSubmitter:Submitting tokens for job: job_1494125392913_0001

17/05/07 03:25:19 INFO impl.YarnClientImpl: Submittedapplication application_1494125392913_0001

17/05/07 03:25:19 INFO mapreduce.Job: The url to trackthe job: http://slaver1:8088/proxy/application_1494125392913_0001/

17/05/07 03:25:19 INFO mapreduce.Job: Running job:job_1494125392913_0001

17/05/07 03:25:30 INFO mapreduce.Job: Jobjob_1494125392913_0001 running in uber mode : false

17/05/07 03:25:30 INFO mapreduce.Job:  map 0% reduce 0%

17/05/07 03:25:36 INFO mapreduce.Job:  map 40% reduce 0%

17/05/07 03:25:41 INFO mapreduce.Job:  map 60% reduce 0%

17/05/07 03:25:42 INFO mapreduce.Job:  map 80% reduce 0%

17/05/07 03:25:45 INFO mapreduce.Job:  map 100% reduce 0%

17/05/07 03:25:48 INFO mapreduce.Job:  map 100% reduce 100%

17/05/07 03:25:49 INFO mapreduce.Job: Jobjob_1494125392913_0001 completed successfully

17/05/07 03:25:49 INFO mapreduce.Job: Counters: 49

FileSystem Counters

FILE: Number of bytes read=116

FILE: Number of bytes written=819237

FILE: Number of read operations=0

FILE: Number of large read operations=0

FILE: Number of write operations=0

HDFS: Number of bytes read=1300

HDFS: Number of bytes written=215

HDFS: Number of read operations=23

HDFS: Number of large read operations=0

HDFS: Number of write operations=3

JobCounters

Launched map tasks=5

Launched reduce tasks=1

Data-local map tasks=5

Total time spent by all maps in occupied slots (ms)=50808

Total time spent by all reduces in occupied slots (ms)=10839

Total time spent by all map tasks (ms)=16936

Total time spent by all reduce tasks (ms)=3613

Total vcore-seconds taken by all maptasks=16936

Total vcore-seconds taken by all reduce tasks=3613

Total megabyte-seconds taken by all map tasks=26013696

Total megabyte-seconds taken by all reduce tasks=5549568

Map-Reduce Framework

Map input records=5

Map output records=10

Map output bytes=90

Map output materialized bytes=140

Input split bytes=710

Combine input records=0

Combine output records=0

Reduce input groups=2

Reduce shuffle bytes=140

Reduce input records=10

Reduce output records=0

Spilled Records=20

Shuffled Maps =5

Failed Shuffles=0

Merged Map outputs=5

GC time elapsed (ms)=450

CPU time spent (ms)=4330

Physical memory (bytes) snapshot=5840977920

Virtual memory (bytes) snapshot=19436744704

Total committed heap usage (bytes)=5483528192

ShuffleErrors

BAD_ID=0

CONNECTION=0

IO_ERROR=0

WRONG_LENGTH=0

WRONG_MAP=0

WRONG_REDUCE=0

FileInput Format Counters

Bytes Read=590

FileOutput Format Counters

Bytes Written=97
Job Finished in 32.805 seconds
Estimated value of Pi is 3.680000000000000000

2…在集群节点中/usr/hdp/2.4.3.0-227/hadoop-mapreduce/目录下,存在一个案例JAR 包hadoop-mapreduce-examples.jar。运 行JAR包中的wordcount程序来对/1daoyun/file/BigDataSkills.txt文件进行单词计数,将运算结果输出到/1daoyun/output目录中,使用相关命令查询单词计数结果,输出结果为。

[root@master ~]# hadoop jar/usr/hdp/2.4.3.0-227/hadoop-mapreduce/hadoop-mapreduce-examples-2.7.1.2.4.3.0-    227.jarwordcount /1daoyun/file/BigDataSkills.txt /1daoyun/output

WARNING: Use "yarn jar" to launch YARNapplications.

17/05/07 03:28:10 INFO impl.TimelineClientImpl: Timelineservice address: http://slaver1:8188/ws/v1/timeline/

17/05/07 03:28:10 INFO client.RMProxy: Connecting toResourceManager at slaver1/10.0.0.15:8050

17/05/07 03:28:11 INFO input.FileInputFormat: Totalinput paths to process : 1

17/05/07 03:28:12 INFO mapreduce.JobSubmitter: numberof splits:1

17/05/07 03:28:13 INFO mapreduce.JobSubmitter:Submitting tokens for job: job_1494125392913_0003

17/05/07 03:28:14 INFO impl.YarnClientImpl: Submittedapplication application_1494125392913_0003

17/05/07 03:28:14 INFO mapreduce.Job: The url to trackthe job:      http://slaver1:8088/proxy/application_1494125392913_0003/

17/05/07 03:28:14 INFO mapreduce.Job: Running job:job_1494125392913_0003

17/05/07 03:28:24 INFO mapreduce.Job: Jobjob_1494125392913_0003 running in uber mode : false

17/05/07 03:28:24 INFO mapreduce.Job:  map 0% reduce 0%

17/05/07 03:28:30 INFO mapreduce.Job:  map 100% reduce 0%

17/05/07 03:28:40 INFO mapreduce.Job:  map 100% reduce 100%

17/05/07 03:30:51 INFO mapred.ClientServiceDelegate:Application state is completed. FinalApplicationStatus=SUCCEEDED. Redirectingto job history server

17/05/07 03:30:52 INFO mapreduce.Job: Jobjob_1494125392913_0003 completed successfully

17/05/07 03:30:52 INFO mapreduce.Job: Counters: 49

FileSystem Counters

FILE: Number of bytes read=90

FILE: Number of bytes written=272541

FILE: Number of read operations=0

FILE: Number of large read operations=0

FILE: Number of write operations=0

HDFS: Number of bytes read=163

HDFS: Number of bytes written=60

HDFS: Number of read operations=6

HDFS: Number of large read operations=0

HDFS: Number of write operations=2

Job Counters

Launched map tasks=1

Launched reduce tasks=1

Data-local map tasks=1

Total time spent by all maps in occupied slots (ms)=8748

Total time spent by all reduces in occupied slots (ms)=12042

Total time spent by all map tasks (ms)=2916

Total time spent by all reduce tasks (ms)=4014

Total vcore-seconds taken by all map tasks=2916

Total vcore-seconds taken by all reduce tasks=4014

Total megabyte-seconds taken by all map tasks=4478976

Total megabyte-seconds taken by all reduce tasks=6165504

Map-Reduce Framework

Map input records=3

Map output records=6

Map output bytes=72

Map output materialized bytes=90

Input split bytes=114

Combine input records=6

Combine output records=6

Reduce input groups=6

Reduce shuffle bytes=90

Reduce input records=6

Reduce output records=6

Spilled Records=12

Shuffled Maps =1

Failed Shuffles=0

Merged Map outputs=1

GC time elapsed (ms)=109

CPU time spent (ms)=1870

Physical memory (bytes) snapshot=1347346432

Virtual memory (bytes) snapshot=6500098048

Total committed heap usage (bytes)=1229455360

ShuffleErrors

BAD_ID=0

CONNECTION=0

IO_ERROR=0

WRONG_LENGTH=0

WRONG_MAP=0

WRONG_REDUCE=0

FileInput Format Counters

BytesRead=49

FileOutput Format Counters

Bytes Written=60

[root@master ~]# hadoop fs -cat/1daoyun/output/part-r-00000

"duiya  1

hello   1

nisibusisha    1

wosha"  1

zsh     1

3.在集群节点中/usr/hdp/2.4.3.0-227/hadoop-mapreduce/目录下,存在一个案例JAR 包hadoop-mapreduce-examples.jar。运行JAR包中的sudoku程序来计算下表中数独运算题的结果。运行完成后输出结果为。

[root@master ~]# cat puzzle1.dta

8 ? ? ? ? ? ? ? ?

? ? 3 6 ? ? ? ? ?

? 7 ? ? 9 ? 2 ? ?

? 5 ? ? ? 7 ? ? ?

? ? ? ? 4 5 7 ? ?

? ? ? 1 ? ? ? 3 ?

? ? 1 ? ? ? ? 6 8

? ? 8 5 ? ? ? 1 ?

? 9 ? ? ? ? 4 ? ?

[root@master hadoop-mapreduce]# hadoop jarhadoop-mapreduce-examples-2.7.1.2.4.3.0-227.jar sudoku /root/puzzle1.dta

WARNING: Use "yarn jar" to launch YARNapplications.

Solving /root/puzzle1.dta

8 1 2 7 5 3 6 4 9

9 4 3 6 8 2 1 7 5

6 7 5 4 9 1 2 8 3

1 5 4 2 3 7 8 9 6

3 6 9 8 4 5 7 2 1

2 8 7 1 6 9 5 3 4

5 2 1 9 7 4 3 6 8

4 3 8 5 2 6 9 1 7

7 9 6 3 1 8 4 5 2

Found 1 solutions

4.在集群节点中/usr/hdp/2.4.3.0-227/hadoop-mapreduce/目录下,存在一个案例JAR 包hadoop-mapreduce-examples.jar。运行JAR包中的grep程序来统计文件系统中 /1daoyun/file/BigDataSkills.txt文件中“Hadoop”出现的次数,统计完成后,查询统计结果信息,输出结果为?

[root@master hadoop-mapreduce]# hadoop jarhadoop-mapreduce-examples-2.7.1.2.4.3.0-227.jar grep/1daoyun/file/BigDataSkills.txt /output hadoop

WARNING: Use "yarn jar" to launch YARNapplications.

17/05/07 13:37:06 INFO impl.TimelineClientImpl:Timeline service address: http://slaver1:8188/ws/v1/timeline/

17/05/07 13:37:06 INFO client.RMProxy: Connecting toResourceManager at slaver1/10.0.0.15:8050

17/05/07 13:37:07 INFO input.FileInputFormat: Totalinput paths to process : 1

17/05/07 13:37:07 INFO mapreduce.JobSubmitter: numberof splits:1

17/05/07 13:37:07 INFO mapreduce.JobSubmitter:Submitting tokens for job: job_1494163309183_0003

17/05/07 13:37:07 INFO impl.YarnClientImpl: Submittedapplication application_1494163309183_0003

17/05/07 13:37:07 INFO mapreduce.Job: The url to trackthe job: http://slaver1:8088/proxy/application_1494163309183_0003/

17/05/07 13:37:07 INFO mapreduce.Job: Running job:job_1494163309183_0003

17/05/07 13:37:14 INFO mapreduce.Job: Jobjob_1494163309183_0003 running in uber mode : false

17/05/07 13:37:14 INFO mapreduce.Job:  map 0% reduce 0%

17/05/07 13:37:23 INFO mapreduce.Job: Task Id :attempt_1494163309183_0003_m_000000_0, Status : FAILED

Exception from container-launch.

Container id:container_e08_1494163309183_0003_01_000002

Exit code: 1

Stack trace: ExitCodeException exitCode=1:

atorg.apache.hadoop.util.Shell.runCommand(Shell.java:600)

atorg.apache.hadoop.util.Shell.run(Shell.java:511)

atorg.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:783)

at org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:212)

atorg.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:303)

at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:82)

atjava.util.concurrent.FutureTask.run(FutureTask.java:266)

atjava.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)

atjava.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)

atjava.lang.Thread.run(Thread.java:745)

Container exited with a non-zero exit code 1

17/05/07 13:37:29 INFO mapreduce.Job:  map 100% reduce 0%

17/05/07 13:37:36 INFO mapreduce.Job:  map 100% reduce 100%

17/05/07 13:37:37 INFO mapreduce.Job: Jobjob_1494163309183_0003 completed successfully

17/05/07 13:37:37 INFO mapreduce.Job: Counters: 51

FileSystem Counters

FILE: Number of bytes read=23

FILE: Number of bytes written=273125

FILE: Number of read operations=0

FILE: Number of large read operations=0

FILE: N
1b023
umber of write operations=0

HDFS: Number of bytes read=146

HDFS: Number of bytes written=109

HDFS: Number of read operations=6

HDFS: Number of large read operations=0

HDFS: Number of write operations=2

JobCounters

Failed map tasks=1

Launched map tasks=2

Launched reduce tasks=1

Other local map tasks=1

Data-local map tasks=1

Total time spent by all maps in occupied slots (ms)=33174

Total time spent by all reduces in occupied slots (ms)=9663

Total time spent by all map tasks (ms)=11058

Total time spent by all reduce tasks (ms)=3221

Total vcore-seconds taken by all map tasks=11058

Total vcore-seconds taken by all reduce tasks=3221

Total megabyte-seconds taken by all map tasks=16985088

Total megabyte-seconds taken by all reduce tasks=4947456

Map-Reduce Framework

Map input records=5

Map output records=2

Map output bytes=30

Map output materialized bytes=23

Input split bytes=114

Combine input records=2

Combine output records=1

Reduce input groups=1

Reduce shuffle bytes=23

Reduce input records=1

Reduce output records=1

Spilled Records=2

Shuffled Maps =1

Failed Shuffles=0

Merged Map outputs=1

GC time elapsed (ms)=107

CPU time spent (ms)=2070

Physical memory (bytes) snapshot=1351417856

Virtual memory (bytes) snapshot=6499807232

Total committed heap usage (bytes)=1233649664

ShuffleErrors

BAD_ID=0

CONNECTION=0

IO_ERROR=0

WRONG_LENGTH=0

WRONG_MAP=0

WRONG_REDUCE=0

FileInput Format Counters

Bytes Read=32

FileOutput Format Counters

Bytes Written=109

17/05/07 13:37:37 INFO impl.TimelineClientImpl:Timeline service address: http://slaver1:8188/ws/v1/timeline/

17/05/07 13:37:37 INFO client.RMProxy: Connecting toResourceManager at slaver1/10.0.0.15:8050

17/05/07 13:37:37 INFO input.FileInputFormat: Totalinput paths to process : 1

17/05/07 13:37:37 INFO mapreduce.JobSubmitter: numberof splits:1

17/05/07 13:37:38 INFO mapreduce.JobSubmitter:Submitting tokens for job: job_1494163309183_0004

17/05/07 13:37:38 INFO impl.YarnClientImpl: Submittedapplication application_1494163309183_0004

17/05/07 13:37:38 INFO mapreduce.Job: The url to trackthe job: http://slaver1:8088/proxy/application_1494163309183_0004/

17/05/07 13:37:38 INFO mapreduce.Job: Running job:job_1494163309183_0004

17/05/07 13:37:48 INFO mapreduce.Job: Jobjob_1494163309183_0004 running in uber mode : false

17/05/07 13:37:48 INFO mapreduce.Job:  map 0% reduce 0%

17/05/07 13:38:02 INFO mapreduce.Job: Task Id :attempt_1494163309183_0004_m_000000_0, Status : FAILED

Exception from container-launch.

Container id:container_e08_1494163309183_0004_01_000002

Exit code: 1

Stack trace: ExitCodeException exitCode=1:

atorg.apache.hadoop.util.Shell.runCommand(Shell.java:600)

atorg.apache.hadoop.util.Shell.run(Shell.java:511)

atorg.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:783)

atorg.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:212)

atorg.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:303)

atorg.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:82)

atjava.util.concurrent.FutureTask.run(FutureTask.java:266)

atjava.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)

at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)

atjava.lang.Thread.run(Thread.java:745)

Container exited with a non-zero exit code 1

17/05/07 13:39:58 INFO mapreduce.Job: Task Id :attempt_1494163309183_0004_m_000000_1, Status : FAILED

Container exited with a non-zero exit code 154

17/05/07 13:40:48 INFO mapreduce.Job:  map 100% reduce 0%

17/05/07 13:40:55 INFO mapreduce.Job:  map 100% reduce 100%

17/05/07 13:40:56 INFO mapreduce.Job: Jobjob_1494163309183_0004 completed successfully

17/05/07 13:40:56 INFO mapreduce.Job: Counters: 49

FileSystem Counters

FILE: Number of bytes read=23

FILE: Number of bytes written=272059

FILE: Number of read operations=0

FILE: Number of large read operations=0

FILE: Number of write operations=0

HDFS: Number of bytes read=236

HDFS: Number of bytes written=9

HDFS: Number of read operations=7

HDFS: Number of large read operations=0

HDFS: Number of write operations=2

JobCounters

Launched map tasks=1

Launched reduce tasks=1

Data-local map tasks=1

Total time spent by all maps in occupied slots (ms)=8769

Total time spent by all reduces in occupied slots (ms)=12570

Total time spent by all map tasks (ms)=2923

Total time spent by all reduce tasks (ms)=4190

Total vcore-seconds taken by all maptasks=2923

Total vcore-seconds taken by all reduce tasks=4190

Total megabyte-seconds taken by all map tasks=4489728

Total megabyte-seconds taken by all reduce tasks=6435840

Map-Reduce Framework

Map input records=1

Map output records=1

Map output bytes=15

Map output materialized bytes=23

Input split bytes=127

Combine input records=0

Combine output records=0

Reduce input groups=1

Reduce shuffle bytes=23

Reduce input records=1

Reduce output records=1

Spilled Records=2

Shuffled Maps =1

Failed Shuffles=0

Merged Map outputs=1

GC time elapsed (ms)=122

CPU time spent (ms)=1840

Physical memory (bytes) snapshot=1351442432

Virtual memory (bytes) snapshot=6500921344

Total committed heap usage (bytes)=1233125376

ShuffleErrors

BAD_ID=0

CONNECTION=0

IO_ERROR=0

WRONG_LENGTH=0

WRONG_MAP=0

WRONG_REDUCE=0

FileInput Format Counters

Bytes Read=109

FileOutput Format Counters

Bytes Written=9

[root@master hadoop-mapreduce]# hadoop fs -cat/output/part-r-00000

2       hadoop
内容来自用户分享和网络整理,不保证内容的准确性,如有侵权内容,可联系管理员处理 点击这里给我发消息
标签: