hadoop java jar,hadoop jar 提示 Not a valid JAR

. 编译 WordCount.java

2. 打成 WordCount.jar包 上传的hdfs。。

3. hadoop jar WordCount.jar ***.**.MainClass /**/input /***/output

问题一直提示  Not a valid JAR: /usr/***/***/wordcount.jar

问题是:执行jar包的时候应该是用本地jar包,而不是hdfs上的,所以写出hdfs路径是不对的

下面是代码

package com.gilang.hadoop.wordcount;

import java.io.IOException;

import java.util.Iterator;

import java.util.StringTokenizer;

import org.apache.hadoop.fs.Path;

import org.apache.hadoop.io.IntWritable;

import org.apache.hadoop.io.LongWritable;

import org.apache.hadoop.io.Text;

import org.apache.hadoop.mapred.FileInputFormat;

import org.apache.hadoop.mapred.FileOutputFormat;

import org.apache.hadoop.mapred.JobClient;

import org.apache.hadoop.mapred.JobConf;

import org.apache.hadoop.mapred.MapReduceBase;

import org.apache.hadoop.mapred.Mapper;

import org.apache.hadoop.mapred.OutputCollector;

import org.apache.hadoop.mapred.Reducer;

import org.apache.hadoop.mapred.Reporter;

import org.apache.hadoop.mapred.TextInputFormat;

import org.apache.hadoop.mapred.TextOutputFormat;

public class WordCount {

public static class Map extends MapReduceBase implements

Mapper {

private final static IntWritable one = new IntWritable(1);

private Text word = new Text();

public void map(LongWritable key, Text value,

OutputCollector output, Reporter reporter)

throws IOException {

String line = value.toString();

StringTokenizer tokenizer = new StringTokenizer(line);

while (tokenizer.hasMoreTokens()) {

word.set(tokenizer.nextToken());

output.collect(word, one);

}

}

}

public static class Reduce extends MapReduceBase implements

Reducer {

public void reduce(Text key, Iterator values,

OutputCollector output, Reporter reporter)

throws IOException {

int sum = 0;

while (values.hasNext()) {

sum += values.next().get();

}

output.collect(key, new IntWritable(sum));

}

}

public static void main(String[] args) throws Exception {

System.out.println("-------------------------"+args.length);

/*if (args.length != 2) {

System.err

.println("Usage: WordCount ");

System.exit(-1);

}*/

System.out.println("-------------------------");

JobConf conf = new JobConf(WordCount.class);

conf.setJobName("wordcount");

conf.setOutputKeyClass(Text.class);

conf.setOutputValueClass(IntWritable.class);

conf.setMapperClass(Map.class);

conf.setCombinerClass(Reduce.class);

conf.setReducerClass(Reduce.class);

conf.setInputFormat(TextInputFormat.class);

conf.setOutputFormat(TextOutputFormat.class);

System.out.println(args[0]+"|"+args[1]+"|"+args[2]);

FileInputFormat.setInputPaths(conf, new Path(args[1]));

FileOutputFormat.setOutputPath(conf, new Path(args[2]));

JobClient.runJob(conf);

}

}

当时在 后面的两个输入目录参数 和 输出目录的参数上犯错了,,确定好args下标的位置

输出目录也就是我的那个 wordcount 目录下不能有 output目录,MapReduce 会抛文件已存在异常

我执行的命令:   ./hadoop jar /home/long/Desktop/wordcount.jar com.gilang.hadoop.wordcount.WordCount /user/hadoop/wordcount/input /user/hadoop/wordcount/output

  • 1
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值