远程提交Map/Reduce任务
2013-10-09 12:58
375 查看
1. 将开发好MR代码打包成jar。添加到distributed cache中。
Xml代码
bin/hadoop fs -copyFromLocal /root/stat-analysis-mapred-1.0-SNAPSHOT.jar /user/root/lib
2. 在服务器端创建和你客户端一模一样的用户。创建目录 /tmp/hadoop-root/stagging/用户
3. 客户端提交job的代码
Java代码
Configuration conf = HBaseConfiguration.create();
conf.set("hbase.zookeeper.quorum", "node.tracker1");
conf.set("fs.default.name", "hdfs://node.tracker1:9000/hbase");
conf.set("mapred.job.tracker", "node.tracker1:9001");
Job job = new Job(conf, "Hbase_FreqCounter1");
job.setJarByClass(FreqCounter1.class);
Scan scan = new Scan();
String columns = "details"; // comma seperated
scan.addFamily(Bytes.toBytes(columns));
scan.setFilter(new FirstKeyOnlyFilter());
TableMapReduceUtil.initTableMapperJob("access_logs", scan, Mapper1.class, ImmutableBytesWritable.class,
IntWritable.class, job);
TableMapReduceUtil.initTableReducerJob("summary_user", Reducer1.class, job);
/ TableMapReduceUtil.addDependencyJars(job);
DistributedCache.addFileToClassPath(new Path("hdfs://node.tracker1:9000/user/root/lib/stat-analysis-mapred-1.0-SNAPSHOT.jar"),job.getConfiguration());
job.submit();
4.运行java application,登陆node的MR管理页面,可以看到
Xml代码
bin/hadoop fs -copyFromLocal /root/stat-analysis-mapred-1.0-SNAPSHOT.jar /user/root/lib
bin/hadoop fs -copyFromLocal /root/stat-analysis-mapred-1.0-SNAPSHOT.jar /user/root/lib
2. 在服务器端创建和你客户端一模一样的用户。创建目录 /tmp/hadoop-root/stagging/用户
3. 客户端提交job的代码
Java代码
Configuration conf = HBaseConfiguration.create();
conf.set("hbase.zookeeper.quorum", "node.tracker1");
conf.set("fs.default.name", "hdfs://node.tracker1:9000/hbase");
conf.set("mapred.job.tracker", "node.tracker1:9001");
Job job = new Job(conf, "Hbase_FreqCounter1");
job.setJarByClass(FreqCounter1.class);
Scan scan = new Scan();
String columns = "details"; // comma seperated
scan.addFamily(Bytes.toBytes(columns));
scan.setFilter(new FirstKeyOnlyFilter());
TableMapReduceUtil.initTableMapperJob("access_logs", scan, Mapper1.class, ImmutableBytesWritable.class,
IntWritable.class, job);
TableMapReduceUtil.initTableReducerJob("summary_user", Reducer1.class, job);
/ TableMapReduceUtil.addDependencyJars(job);
DistributedCache.addFileToClassPath(new Path("hdfs://node.tracker1:9000/user/root/lib/stat-analysis-mapred-1.0-SNAPSHOT.jar"),job.getConfiguration());
job.submit();
Configuration conf = HBaseConfiguration.create(); conf.set("hbase.zookeeper.quorum", "node.tracker1"); conf.set("fs.default.name", "hdfs://node.tracker1:9000/hbase"); conf.set("mapred.job.tracker", "node.tracker1:9001"); Job job = new Job(conf, "Hbase_FreqCounter1"); job.setJarByClass(FreqCounter1.class); Scan scan = new Scan(); String columns = "details"; // comma seperated scan.addFamily(Bytes.toBytes(columns)); scan.setFilter(new FirstKeyOnlyFilter()); TableMapReduceUtil.initTableMapperJob("access_logs", scan, Mapper1.class, ImmutableBytesWritable.class, IntWritable.class, job); TableMapReduceUtil.initTableReducerJob("summary_user", Reducer1.class, job); // TableMapReduceUtil.addDependencyJars(job); DistributedCache.addFileToClassPath(new Path("hdfs://node.tracker1:9000/user/root/lib/stat-analysis-mapred-1.0-SNAPSHOT.jar"),job.getConfiguration()); job.submit();
4.运行java application,登陆node的MR管理页面,可以看到
相关文章推荐
- Hadoop MapReduce之ReduceTask任务执行(一):远程拷贝map输出
- 客户端用java api 远程操作HDFS以及远程提交MR任务(源码和异常处理)
- 分析JobInProgress中Map/Reduce任务分配
- 从Java代码远程提交YARN MapReduce任务
- windows下eclipse远程提交hadoop 2.5.2 Mapreduce任务
- Spark历险记之编译和远程任务提交
- hive优化之------控制hive任务中的map数和reduce数
- hive优化之 控制hive任务中的map数和reduce数
- MapReduce剖析笔记之七:Child子进程处理Map和Reduce任务的主要流程
- 实现每个Map或Reduce任务只执行一次map或reduce方法
- hadoop如何处理长时间运行不完成的map/reduce 任务?
- map任务和reduce任务个数如何计算
- 【Hadoop代码笔记】Hadoop作业提交之Child启动map任务
- map-reduce任务的执行流程
- windows eclipse远程连接hadoop集群并提交任务运行
- map和reduce任务各参数参考
- 如何在Hadoop中控制Map&Reduce任务的数量
- 小米wifi远程提交下载任务地址
- Spark历险记之编译和远程任务提交
- map-reduce任务的执行流程