hadoop-hbase-endpoint例子 count下users表里使用google邮箱的用户数服务端:先定义个接口,extends CoprocessorProtocol,接口里定义个方法public long gmailCounter(String userId) throws IOException;再定义个类,extends BaseEndpointCoprocessor,再implements刚才自己定义的接
hadoop-HBase-过滤器 package test.hbase.inaction.example4_8;import org.apache.hadoop.hbase.KeyValue;import org.apache.hadoop.hbase.client.HTableInterface;import org.apache.hadoop.hbase.client.HTablePool;import org.ap
hadoop-HBase-observer的一个例子 observer的一个例子,也只是个例子start() 和stop()这两个方法,在表enable和disable时分别也会对应执行这start()和stop()在连接池里获得连接与释放连接实在是多余observer类似触发器,操作一张表的时候要注意递归调用,拦截到一次put后再put这个表就会递归这个例子就是这样的,实际上没任何意义,就是个例子而已注意锁,HBase是行级的锁,如果想put同一行是不会成功的 先disable表,然后alert一下,那个1001是优先级,别忘记enable
hadoop-HBase-map/reduce调用HBase及Counter package test.hbase.inaction.example3_5;import java.io.IOException;import org.apache.hadoop.conf.Configuration;import org.apache.hadoop.hbase.HBaseConfiguration;import org.apache.hadoop
hadoop-config hdfs- dfs.name.dir /usr/local/hadoop/datalog1,/usr/local/hadoop/datalog2 dfs.data.dir /usr/local/hadoop/data1,/usr/local/hadoop/data2
hadoop-HBase-基本 create 'tab1', 'col1', 'col2', 'col3'put 'tab1', 'row1', 'col1:1', 'value11'put 'tab1', 'row1', 'col1:2', 'value12'put 'tab1', 'row1', 'col2:1', 'value21'put 'tab1', 'row1', 'col2:2', 'val
Ubuntu-几个命令 sudo update-alternatives --install /usr/bin/jps jps /usr/java/jdk1.6.0_43/bin/jps 300sudo update-alternatives --display jpsapt-cache search ncursesapt-get install libncurses5-devdu -ah
hadoop-hive-DISTINCT @@DISTINCT 并覆盖,采用 GROUP BY 代替INSERT OVERWRITE TABLE tab4 select key, value FROM tab4 GROUP BY key, value;@@关联SELECT tab4.*, tab5.value FROM tab4 JOIN tab5 ON (tab4.key=tab5.key);@@居然互斥
hadoop-hive-创建表及LOAD数据相关 @@创建普通表create table tab5 (key int, value string) row format delimited fields terminated by '\t';@@创建分区表create table partition_tab2(col1 string,col2 string)partitioned by (province string) row