整体描述:将本地文件的数据整理之后导入到hbase中
成都创新互联专注于澄迈企业网站建设,成都响应式网站建设公司,电子商务商城网站建设。澄迈网站建设公司,为澄迈等地区提供建站服务。全流程按需网站建设,专业设计,全程项目跟踪,成都创新互联专业和态度为您提供的服务
在HBase中创建表

数据格式

MapReduce程序
map程序
package com.hadoop.mapreduce.test.map; import java.io.IOException; import org.apache.hadoop.io.Text; import org.apache.hadoop.mapreduce.Mapper; public class WordCountHBaseMapper extends Mapper
Reduce程序
package com.hadoop.mapreduce.test.reduce; import java.io.IOException; import java.util.Iterator; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.mapreduce.TableReducer; import org.apache.hadoop.io.NullWritable; import org.apache.hadoop.io.Text; public class WordCountHBaseReduce extends TableReducer{ @Override protected void reduce(Text key, Iterable value, Context out) throws IOException, InterruptedException { String keyValue = key.toString(); Iterator valueIterator = value.iterator(); while(valueIterator.hasNext()){ Text valueV = valueIterator.next(); String[] valueArray = valueV.toString().split("#"); Put putRow = new Put(keyValue.getBytes()); putRow.add("address".getBytes(), "baseAddress".getBytes(), valueArray[0].getBytes()); putRow.add("sex".getBytes(), "baseSex".getBytes(), valueArray[1].getBytes()); putRow.add("age".getBytes(), "baseAge".getBytes(), valueArray[2].getBytes()); out.write(NullWritable.get(), putRow); } } }
主程序
package com.hadoop.mapreduce.test;
import java.io.IOException;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hbase.HBaseConfiguration;
import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.lib.input.FileInputFormat;
import com.hadoop.mapreduce.test.map.WordCountHBaseMapper;
import com.hadoop.mapreduce.test.reduce.WordCountHBaseReduce;
/**
* 将hdfs上的内容读取到,并插入到hbase的表中,然后读取hbase表中的内容,将统计结果插入到hbase中
*/
public class WordCountHBase {
public static void main(String args[]) throws IOException,
InterruptedException, ClassNotFoundException{
Configuration conf = HBaseConfiguration.create();
conf.set("hbase.zookeeper.quorum", "192.168.192.137");
Job job = Job.getInstance(conf, "MapReduceHbaseJob");
//各种class
job.setJarByClass(WordCountHBase.class);
job.setMapperClass(WordCountHBaseMapper.class);
TableMapReduceUtil.initTableReducerJob("userInfo3",
WordCountHBaseReduce.class, job);
FileInputFormat.addInputPath(job, new Path(args[0]));
job.setMapOutputKeyClass(Text.class);
job.setMapOutputValueClass(Text.class);
System.exit(job.waitForCompletion(true) ? 0 : 1);
}
}结果:

注:如果运行的client没有hbase,需要在hadoop里面的lib中加入hbase的lib