java实现map-reduce代码
2017-04-19 15:10
309 查看
本文记录用java实现map-reduce,并在服务器上运行的代码。
package com.demo.hadoop.test;
import java.io.IOException;
import java.util.StringTokenizer;
import java.util.Arrays;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.conf.Configured;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.FileUtil;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.IntWritable;
import org.apache.hadoop.io.LongWritable;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.Mapper;
import org.apache.hadoop.mapreduce.Reducer;
import org.apache.hadoop.mapreduce.Mapper.Context;
import org.apache.hadoop.mapreduce.lib.input.FileInputFormat;
import org.apache.hadoop.mapreduce.lib.input.TextInputFormat;
import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;
import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat;
import org.apache.hadoop.mapreduce.v2.app.job.TaskAttempt;
import org.apache.hadoop.util.GenericOptionsParser;
import org.apache.hadoop.util.Tool;
import org.apache.hadoop.util.ToolRunner;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import com.demo.hadoop.test.AvgScore.MyMap;
import com.demo.hadoop.test.AvgScore.MyReduce;
import com.sun.media.sound.SoftLowFrequencyOscillator;
public class NbRoute implements Tool {
public static final Logger log = LoggerFactory.getLogger(NbRoute.class);
Configuration configuration;
public static class MyMap extends Mapper<LongWritable, Text, Text, Text> {
public void map(LongWritable key, Text value, Context context) throws IOException, InterruptedException {
String[] citation = value.toString().split(",");
context.write(new Text(citation[0]), new Text(citation[1]));
}
}
public static class MyReduce extends Reducer<Text, Text, Text, Text> {
public void reduce(Text key, Iterable<Text> values, Context context) throws IOException, InterruptedException {
String csv = "";
for (Text val : values) {
if (csv.length() > 0) {
csv += ",";
}
if (!csv.contains(val.toString()))
csv += val.toString();
}
String[] tmp = csv.split(",");
Arrays.sort(tmp);
String scenic = "";
if (tmp.length > 2) {
for (int i = 0; i < tmp.length; i++){
scenic += tmp[i];
scenic += ",";
}
String sctmp = scenic.substring(0,scenic.length()-1);
context.write(key, new Text(sctmp));
}
}
}
@Override
public int run(String[] args) throws Exception {
// TODO Auto-generated method stub
Configuration conf = new Configuration();
Job job = new Job(conf, "NbRoute");
job.setJarByClass(NbRoute.class);
job.setJobName("NbRoute");
job.setOutputKeyClass(Text.class);
job.setOutputValueClass(Text.class);
job.setMapperClass(MyMap.class);
job.setCombinerClass(MyReduce.class);
job.setReducerClass(MyReduce.class);
job.setInputFormatClass(TextInputFormat.class);
job.setOutputFormatClass(TextOutputFormat.class);
FileInputFormat.setInputPaths(job, new Path(args[0]));
FileOutputFormat.setOutputPath(job, new Path(args[1]));
boolean success = job.waitForCompletion(true);
System.exit(job.waitForCompletion(true) ? 0 : 1);
return success ? 0 : 1;
}
@Override
public Configuration getConf() {
return configuration;
}
@Override
public void setConf(Configuration conf) {
conf = new Configuration();
configuration = conf;
}
public static void main(String[] args) throws Exception {
int ret = ToolRunner.run(new NbRoute(), args);
System.exit(ret);
}
}
从eclipse中导成jar包放到服务器上运行,代码运行命令:hadoop jar avgscore.jar com.demo.hadoop.test.NbRoute /APPLICATION/YIWUXSP/oidd/test/201703 /APPLICATION/YIWUXSP/oidd/test/out1
package com.demo.hadoop.test;
import java.io.IOException;
import java.util.StringTokenizer;
import java.util.Arrays;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.conf.Configured;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.FileUtil;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.IntWritable;
import org.apache.hadoop.io.LongWritable;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.Mapper;
import org.apache.hadoop.mapreduce.Reducer;
import org.apache.hadoop.mapreduce.Mapper.Context;
import org.apache.hadoop.mapreduce.lib.input.FileInputFormat;
import org.apache.hadoop.mapreduce.lib.input.TextInputFormat;
import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;
import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat;
import org.apache.hadoop.mapreduce.v2.app.job.TaskAttempt;
import org.apache.hadoop.util.GenericOptionsParser;
import org.apache.hadoop.util.Tool;
import org.apache.hadoop.util.ToolRunner;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import com.demo.hadoop.test.AvgScore.MyMap;
import com.demo.hadoop.test.AvgScore.MyReduce;
import com.sun.media.sound.SoftLowFrequencyOscillator;
public class NbRoute implements Tool {
public static final Logger log = LoggerFactory.getLogger(NbRoute.class);
Configuration configuration;
public static class MyMap extends Mapper<LongWritable, Text, Text, Text> {
public void map(LongWritable key, Text value, Context context) throws IOException, InterruptedException {
String[] citation = value.toString().split(",");
context.write(new Text(citation[0]), new Text(citation[1]));
}
}
public static class MyReduce extends Reducer<Text, Text, Text, Text> {
public void reduce(Text key, Iterable<Text> values, Context context) throws IOException, InterruptedException {
String csv = "";
for (Text val : values) {
if (csv.length() > 0) {
csv += ",";
}
if (!csv.contains(val.toString()))
csv += val.toString();
}
String[] tmp = csv.split(",");
Arrays.sort(tmp);
String scenic = "";
if (tmp.length > 2) {
for (int i = 0; i < tmp.length; i++){
scenic += tmp[i];
scenic += ",";
}
String sctmp = scenic.substring(0,scenic.length()-1);
context.write(key, new Text(sctmp));
}
}
}
@Override
public int run(String[] args) throws Exception {
// TODO Auto-generated method stub
Configuration conf = new Configuration();
Job job = new Job(conf, "NbRoute");
job.setJarByClass(NbRoute.class);
job.setJobName("NbRoute");
job.setOutputKeyClass(Text.class);
job.setOutputValueClass(Text.class);
job.setMapperClass(MyMap.class);
job.setCombinerClass(MyReduce.class);
job.setReducerClass(MyReduce.class);
job.setInputFormatClass(TextInputFormat.class);
job.setOutputFormatClass(TextOutputFormat.class);
FileInputFormat.setInputPaths(job, new Path(args[0]));
FileOutputFormat.setOutputPath(job, new Path(args[1]));
boolean success = job.waitForCompletion(true);
System.exit(job.waitForCompletion(true) ? 0 : 1);
return success ? 0 : 1;
}
@Override
public Configuration getConf() {
return configuration;
}
@Override
public void setConf(Configuration conf) {
conf = new Configuration();
configuration = conf;
}
public static void main(String[] args) throws Exception {
int ret = ToolRunner.run(new NbRoute(), args);
System.exit(ret);
}
}
从eclipse中导成jar包放到服务器上运行,代码运行命令:hadoop jar avgscore.jar com.demo.hadoop.test.NbRoute /APPLICATION/YIWUXSP/oidd/test/201703 /APPLICATION/YIWUXSP/oidd/test/out1
相关文章推荐
- Hadoop MapReduce概念学习系列之map并发任务数和reduce并发任务数的原理和代码实现(十八)
- Java:基于Map实现的频率统计代码
- JAVA遍历map的几种实现方法代码
- Hadoop Streaming: 使用Java以外的语言去实现Map/Reduce
- Java实现将Map转换为List的小代码
- Java代码实现Map和Object互转及Map和Json互转
- Java实现将Map转换为List的小代码
- java Map转Object与Object转Map实现代码
- Jsp/Java代码分离.实现页面真正的代码分离 实现框架代码:BasePage,TextBox,BasePageFilter
- 用java实现论坛ubb的代码
- 用java实现简单的网络通信,相当于一个最简单的控制台qq(附代码) 推荐
- Jsp/Java代码分离.实现页面真正的代码分离 设计原理
- 如何编写出高效的数据库连接池(附带完整代码C#和Java实现)
- Java中数据库事务处理的实现-JSP教程,Java技巧及代码
- 3DES加密java的实现代码
- 购物车的 java 代码实现方法
- JAVA断点续传实现方法代码
- newxy技术零java代码实现数据分页显示 作者:胡立新
- 利用newxy(新坐标),不写一行java代码实现多项查询
- 用C语言实现函数语言中的Map和Reduce操作