import java.io.IOException;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.conf.Configured;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hbase.HBaseConfiguration;
import org.apache.hadoop.hbase.HColumnDescriptor;
import org.apache.hadoop.hbase.HTableDescriptor;
import org.apache.hadoop.hbase.MasterNotRunningException;
import org.apache.hadoop.hbase.ZooKeeperConnectionException;
import org.apache.hadoop.hbase.client.HBaseAdmin;
import org.apache.hadoop.hbase.client.Put;
import org.apache.hadoop.hbase.mapreduce.TableOutputFormat;
import org.apache.hadoop.hbase.mapreduce.TableReducer;
import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.io.IntWritable;
import org.apache.hadoop.io.LongWritable;
import org.apache.hadoop.io.NullWritable;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.Mapper;
import org.apache.hadoop.mapreduce.lib.input.FileInputFormat;
import org.apache.hadoop.mapreduce.lib.input.TextInputFormat;
import org.apache.hadoop.util.Tool;
import org.apache.hadoop.util.ToolRunner;
public class HbaseWordCount extends Configured implements Tool {
private static Configuration configuration;
private static String tablename = "wordcount";
public static void main(String[] args) throws Exception {
configuration = new Configuration();
configuration.set(TableOutputFormat.OUTPUT_TABLE, tablename);
System.out.println("===");
createHbaseTable(tablename);
System.out.println("===");
int ret = ToolRunner.run(new HbaseWordCount(), args);
System.out.println(ret);
System.exit(ret);
}
public static class Map extends
Mapper<LongWritable, Text, Text, IntWritable> {
private IntWritable one = new IntWritable(1);
@Override
protected void map(LongWritable key, Text value, Context context)
throws IOException, InterruptedException {
String[] s = value.toString().trim().split(" ");
for (String string : s) {
context.write(new Text(string), one);
}
}
}
public static class Reduce extends
TableReducer<Text, IntWritable, NullWritable> {
@Override
protected void reduce(Text key, Iterable<IntWritable> values,
Context context) throws IOException, InterruptedException {
int sum = 0;
for (IntWritable intWritable : values) {
sum++;
}
System.out.println(key + "==>" + sum);
Put put = new Put(Bytes.toBytes(key.toString()));
put.add(Bytes.toBytes("cnt"), Bytes.toBytes("count"),
Bytes.toBytes(String.valueOf(sum)));
context.write(NullWritable.get(), put);
}
}
@Override
public int run(String[] arg0) throws Exception {
createHbaseTable(tablename);
Job job = new Job(configuration);
job.setJarByClass(HbaseWordCount.class);
job.setNumReduceTasks(3);
job.setMapperClass(Map.class);
job.setReducerClass(Reduce.class);
job.setMapOutputKeyClass(Text.class);
job.setMapOutputValueClass(IntWritable.class);
job.setInputFormatClass(TextInputFormat.class);
job.setOutputFormatClass(TableOutputFormat.class);
FileInputFormat.setInputPaths(job, new Path("./txt/word.txt"));
FileInputFormat.setInputPaths(job, new Path("./txt/word2.txt"));
boolean success = job.waitForCompletion(true);
return success ? 0 : 1;
}
public static void createHbaseTable(String tablename)
throws MasterNotRunningException, ZooKeeperConnectionException,
IOException {
HBaseConfiguration conf = new HBaseConfiguration();
HBaseAdmin admin = new HBaseAdmin(conf);
HTableDescriptor desc = new HTableDescriptor(tablename);
HColumnDescriptor columnDescriptor = new HColumnDescriptor("cnt");
desc.addFamily(columnDescriptor);
if (admin.tableExists(tablename)) {
admin.disableTable(tablename);
admin.deleteTable(tablename);
}
admin.createTable(desc);
}
}
MR之wc数据写入Hbase
最新推荐文章于 2022-04-15 14:24:40 发布