【GP分区】
create table catalog_sales_test(
"id" BIGINT,
"t1" smallint,
"t2" integer,
"t3" bigint,
"t4" decimal(6,2),
"t5" numeric(7,3),
"t6" double precision,
"t7" varchar(255),
"t8" char(10),
"t9" text,
"t10" time,
"t11" date,
"t12" TIMESTAMP
)distributed by (id)
partition by range("t11")
(
partition p1 start ('2020-01-01') inclusive end ('2020-01-31') exclusive,
partition p2 start ('2020-04-01') inclusive end ('2020-04-30') exclusive,
default partition default_p
);
【Phoenix分区】
create table if not exists test.catalog_sales_test(
id varchar(255) primary key,
column1 tinyint,
column2 smallint,
column3 integer,
column4 bigint,
column5 float,
column6 double,
column7 DECIMAL,
column8 TIMESTAMP,
column9 DATE,
column10 varchar(255)
) SPLIT ON ('2020-04-24','2020-04-25','2020-04-26');
--------------------------------------------------------------------------
【Hive】压缩比测试: 无压缩8.940s
create table if not exists test_orc_snappy(
column1 INT,
column2 BIGINT,
column3 STRING,
column4 BIGINT,
column5 FLOAT,
column6 STRING,
column7 STRING,
column8 TIMESTAMP,
column9 STRING,
column10 BOOLEAN
)
row format delimited fields terminated by ','
stored AS orc tblproperties ("orc.compress"="SNAPPY");
create table if not exists test_orc_lzo(
column1 INT,
column2 BIGINT,
column3 STRING,
column4 BIGINT,
column5 FLOAT,
column6 STRING,
column7 STRING,
column8 TIMESTAMP,
column9 STRING,
column10 BOOLEAN
)
row format delimited fields terminated by ','
stored AS orc tblproperties ("orc.compress"="LZO");
create table if not exists test_parquet_zlib(
column1 INT,
column2 BIGINT,
column3 STRING,
column4 BIGINT,
column5 FLOAT,
column6 STRING,
column7 STRING,
column8 TIMESTAMP,
column9 STRING,
column10 BOOLEAN
)
row format delimited fields terminated by ','
stored AS PARQUET tblproperties ("parquet.compress"="ZLIB");
create table if not exists test(
column1 INT,
column2 BIGINT,
column3 STRING,
column4 BIGINT,
column5 FLOAT,
column6 STRING,
column7 STRING,
column8 TIMESTAMP,
column9 STRING,
column10 BOOLEAN
)
row format delimited fields terminated by ','
stored AS textfile;
load data inpath '/exportcsv.csv' into table test_txt;
insert into table test_orc_snappy select * from test_txt;
hadoop fs -du -s -h /warehouse/tablespace/managed/hive/test.db/test_orc_snappy;
hdfs dfs -put /root/exportcsv.csv /exportcsv.csv
数据库:tpcds_text_400,tpcds_bin_orc_400;
表名:store_sales
1151988104
1151988104
---------------------------------------------------------------------------------------
【HBase + Phoenix】压缩比测试: 无压缩18.278s
/usr/hdp/3.1.4.0-315/phoenix/bin/sqlline.py host121:2181
java -cp /testCompress/maven_javase-1.0-SNAPSHOT.jar com.test.maven.