如何调试Hadoop的MapReduce 过程

1.Hadoop 2.7.4,hbase 1.2.6
2.目的,从hbase 查询数据到hdfs中
3.运行结果如下

Map-Reduce Framework
        Map input records=2
        Map output records=0
        Input split bytes=62
        Spilled Records=0
        Failed Shuffles=0
        Merged Map outputs=0
        GC time elapsed (ms)=38
        CPU time spent (ms)=1450
        Physical memory (bytes) snapshot=213590016
        Virtual memory (bytes) snapshot=2123476992
        Total committed heap usage (bytes)=99090432

Map 一直输出0,不知道怎么看中间过程,hbase user表中两条记录,所以input=2是对的
代码如下:

    public class Hdfs {
    private static Logger logger = Logger.getLogger(Mysql.class);
    public static class HbaseMapper extends TableMapper<Text, Text> {
        @Override
        protected void map(ImmutableBytesWritable key, Result value, Context context) throws IOException, InterruptedException {
            StringBuffer sb = new StringBuffer("");
            context.write(new Text("test"),new Text("value"));
            for(Cell kv : value.listCells()){
                context.write(new Text(key.get()), new Text(new String(kv.getValue()+"sss")));
            }
        }
    }

    public static class HdfsReducer extends Reducer<Text,Text,Text,Text>{
        private Text result = new Text();

        @Override
        protected void reduce(Text key, Iterable<Text> values, Context context) throws IOException, InterruptedException {
            for(Text val:values){
                result.set(val);
            }
            context.write(key, result);
        }
    }

    public static void main(String[] args)throws Exception{

        String output = "hdfs://*.*.*.*:9000/output";
        System.setProperty("hadoop.home.dir","/Users/*/hadoop-2.7.4");
        Configuration conf = HBaseConfiguration.create();
        conf.set("hbase.zookeeper.quorum","master");
        conf.set("fs.default.name","hdfs://*.*.*.*:9000");
        conf.set("mapreduce.app-submission.cross-platform","true");
        conf.set("mapreduce.framework.name","yarn");
        conf.set("mapred.jar","/Users/*/Downloads/WordCount/target/hadoop_m2-1.0-SNAPSHOT.jar");

        FileSystem fs = FileSystem.get(conf);
        Path p = new Path(output);
        if(fs.exists(p)){fs.delete(p,true);}
        Job job = Job.getInstance(conf,"hbase2hdfs");
        job.setJarByClass(Hdfs.class);
        Scan s = new Scan();
        TableMapReduceUtil.initTableMapperJob("user", s,HbaseMapper.class, Text.class, Text.class, job);
        job.setReducerClass(HdfsReducer.class);
        job.setOutputKeyClass(Text.class);
        job.setOutputValueClass(Text.class);
        job.setNumReduceTasks(0);
        FileOutputFormat.setOutputPath(job, p);
        System.exit(job.waitForCompletion(true)?0:1);
    }
}
阅读 3.5k
1 个回答
  1. hadoop Map过程修改需要重新打包(过程中没法输出)
  2. main函数不需要重新打包
  3. 调试的话可以试试MRUnit
撰写回答
你尚未登录,登录后可以
  • 和开发者交流问题的细节
  • 关注并接收问题和回答的更新提醒
  • 参与内容的编辑和改进,让解决方法与时俱进