hive join explain


数据:


1,join

2,left outer join

select * from tb1 left outer join tb2 on tb1.num=tb2.num;

hive> select * from tb1 left outer join tb2 on tb1.num=tb2.num
    > ;
Total MapReduce jobs = 1
Launching Job 1 out of 1
Number of reduce tasks not specified. Estimated from input data size: 1
In order to change the average load for a reducer (in bytes):
  set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
  set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
  set mapred.reduce.tasks=<number>
Starting Job = job_201307162036_0003, Tracking URL = http://hadoopmaster:50030/jobdetails.jsp?jobid=job_201307162036_0003
Kill Command = /home/hadoop/bigdata/hadoop-1.0.4/libexec/../bin/hadoop job  -kill job_201307162036_0003
Hadoop job information for Stage-1: number of mappers: 2; number of reducers: 1
2013-07-17 19:30:36,174 Stage-1 map = 0%,  reduce = 0%
2013-07-17 19:30:42,199 Stage-1 map = 50%,  reduce = 0%, Cumulative CPU 0.68 sec
2013-07-17 19:30:43,208 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 1.58 sec
2013-07-17 19:30:44,218 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 1.58 sec
2013-07-17 19:30:45,226 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 1.58 sec
2013-07-17 19:30:46,238 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 1.58 sec
2013-07-17 19:30:47,247 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 1.58 sec
2013-07-17 19:30:48,253 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 1.58 sec
2013-07-17 19:30:49,264 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 1.58 sec
2013-07-17 19:30:50,272 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 1.58 sec
2013-07-17 19:30:51,280 Stage-1 map = 100%,  reduce = 33%, Cumulative CPU 1.58 sec
2013-07-17 19:30:52,288 Stage-1 map = 100%,  reduce = 33%, Cumulative CPU 1.58 sec
2013-07-17 19:30:53,300 Stage-1 map = 100%,  reduce = 33%, Cumulative CPU 1.58 sec
2013-07-17 19:30:54,305 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 3.09 sec
2013-07-17 19:30:55,313 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 3.09 sec
2013-07-17 19:30:56,323 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 3.09 sec
2013-07-17 19:30:57,332 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 3.09 sec
2013-07-17 19:30:58,341 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 3.09 sec
2013-07-17 19:30:59,349 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 3.09 sec
2013-07-17 19:31:00,361 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 3.09 sec
MapReduce Total cumulative CPU time: 3 seconds 90 msec
Ended Job = job_201307162036_0003
MapReduce Jobs Launched: 
Job 0: Map: 2  Reduce: 1   Cumulative CPU: 3.09 sec   HDFS Read: 452 HDFS Write: 31 SUCCESS
Total MapReduce CPU Time Spent: 3 seconds 90 msec
OK
111     111
222     222
333     333
444     NULL
Time taken: 32.755 seconds

3,right outer join

hive> select * from tb1 right outer join tb2 on tb1.num=tb2.num
    > ;
Total MapReduce jobs = 1
Launching Job 1 out of 1
Number of reduce tasks not specified. Estimated from input data size: 1
In order to change the average load for a reducer (in bytes):
  set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
  set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
  set mapred.reduce.tasks=<number>
Starting Job = job_201307162036_0004, Tracking URL = http://hadoopmaster:50030/jobdetails.jsp?jobid=job_201307162036_0004
Kill Command = /home/hadoop/bigdata/hadoop-1.0.4/libexec/../bin/hadoop job  -kill job_201307162036_0004
Hadoop job information for Stage-1: number of mappers: 2; number of reducers: 1
2013-07-17 19:31:44,850 Stage-1 map = 0%,  reduce = 0%
2013-07-17 19:31:50,873 Stage-1 map = 50%,  reduce = 0%, Cumulative CPU 0.62 sec
2013-07-17 19:31:51,885 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 1.64 sec
2013-07-17 19:31:52,898 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 1.64 sec
2013-07-17 19:31:53,906 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 1.64 sec
2013-07-17 19:31:54,914 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 1.64 sec
2013-07-17 19:31:55,923 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 1.64 sec
2013-07-17 19:31:56,934 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 1.64 sec
2013-07-17 19:31:57,939 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 1.64 sec
2013-07-17 19:31:58,946 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 1.64 sec
2013-07-17 19:31:59,955 Stage-1 map = 100%,  reduce = 33%, Cumulative CPU 1.64 sec
2013-07-17 19:32:00,963 Stage-1 map = 100%,  reduce = 33%, Cumulative CPU 1.64 sec
2013-07-17 19:32:01,972 Stage-1 map = 100%,  reduce = 33%, Cumulative CPU 1.64 sec
2013-07-17 19:32:02,979 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 3.19 sec
2013-07-17 19:32:03,986 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 3.19 sec
2013-07-17 19:32:04,996 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 3.19 sec
2013-07-17 19:32:06,004 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 3.19 sec
2013-07-17 19:32:07,015 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 3.19 sec
2013-07-17 19:32:08,025 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 3.19 sec
2013-07-17 19:32:09,034 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 3.19 sec
MapReduce Total cumulative CPU time: 3 seconds 190 msec
Ended Job = job_201307162036_0004
MapReduce Jobs Launched: 
Job 0: Map: 2  Reduce: 1   Cumulative CPU: 3.19 sec   HDFS Read: 452 HDFS Write: 31 SUCCESS
Total MapReduce CPU Time Spent: 3 seconds 190 msec
OK
111     111
222     222
333     333
NULL    555
Time taken: 32.695 seconds

4,full outer join

hive> select * from tb1 full  outer join tb2 on tb1.num=tb2.num
    > ;
Total MapReduce jobs = 1
Launching Job 1 out of 1
Number of reduce tasks not specified. Estimated from input data size: 1
In order to change the average load for a reducer (in bytes):
  set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
  set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
  set mapred.reduce.tasks=<number>
Starting Job = job_201307162036_0005, Tracking URL = http://hadoopmaster:50030/jobdetails.jsp?jobid=job_201307162036_0005
Kill Command = /home/hadoop/bigdata/hadoop-1.0.4/libexec/../bin/hadoop job  -kill job_201307162036_0005
Hadoop job information for Stage-1: number of mappers: 2; number of reducers: 1
2013-07-17 19:32:49,596 Stage-1 map = 0%,  reduce = 0%
2013-07-17 19:32:55,617 Stage-1 map = 50%,  reduce = 0%, Cumulative CPU 0.58 sec
2013-07-17 19:32:56,624 Stage-1 map = 50%,  reduce = 0%, Cumulative CPU 0.58 sec
2013-07-17 19:32:57,631 Stage-1 map = 50%,  reduce = 0%, Cumulative CPU 0.58 sec
2013-07-17 19:32:58,641 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 2.08 sec
2013-07-17 19:32:59,648 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 2.08 sec
2013-07-17 19:33:00,654 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 2.08 sec
2013-07-17 19:33:01,662 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 2.08 sec
2013-07-17 19:33:02,670 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 2.08 sec
2013-07-17 19:33:03,679 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 2.08 sec
2013-07-17 19:33:04,688 Stage-1 map = 100%,  reduce = 17%, Cumulative CPU 2.08 sec
2013-07-17 19:33:05,697 Stage-1 map = 100%,  reduce = 17%, Cumulative CPU 2.08 sec
2013-07-17 19:33:06,706 Stage-1 map = 100%,  reduce = 17%, Cumulative CPU 2.08 sec
2013-07-17 19:33:07,711 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 3.8 sec
2013-07-17 19:33:08,716 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 3.8 sec
2013-07-17 19:33:09,725 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 3.8 sec
2013-07-17 19:33:10,735 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 3.8 sec
2013-07-17 19:33:11,745 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 3.8 sec
2013-07-17 19:33:12,754 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 3.8 sec
2013-07-17 19:33:13,763 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 3.8 sec
MapReduce Total cumulative CPU time: 3 seconds 800 msec
Ended Job = job_201307162036_0005
MapReduce Jobs Launched: 
Job 0: Map: 2  Reduce: 1   Cumulative CPU: 3.8 sec   HDFS Read: 452 HDFS Write: 38 SUCCESS
Total MapReduce CPU Time Spent: 3 seconds 800 msec
OK
111     111
222     222
333     333
444     NULL
NULL    555
Time taken: 32.671 seconds

5,left semi join

select * from tb1 left  semi  join tb2 on tb1.num=tb2.num;

hive> select * from tb1 left  semi  join tb2 on tb1.num=tb2.num;
Total MapReduce jobs = 1
Launching Job 1 out of 1
Number of reduce tasks not specified. Estimated from input data size: 1
In order to change the average load for a reducer (in bytes):
  set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
  set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
  set mapred.reduce.tasks=<number>
Starting Job = job_201307162036_0006, Tracking URL = http://hadoopmaster:50030/jobdetails.jsp?jobid=job_201307162036_0006
Kill Command = /home/hadoop/bigdata/hadoop-1.0.4/libexec/../bin/hadoop job  -kill job_201307162036_0006
Hadoop job information for Stage-1: number of mappers: 2; number of reducers: 1
2013-07-17 19:41:56,040 Stage-1 map = 0%,  reduce = 0%
2013-07-17 19:42:02,465 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 4.3 sec
2013-07-17 19:42:03,479 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 4.3 sec
2013-07-17 19:42:04,491 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 4.3 sec
2013-07-17 19:42:05,504 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 4.3 sec
2013-07-17 19:42:06,514 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 4.3 sec
2013-07-17 19:42:07,521 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 4.3 sec
2013-07-17 19:42:08,528 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 4.3 sec
2013-07-17 19:42:09,537 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 4.3 sec
2013-07-17 19:42:10,546 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 4.3 sec
2013-07-17 19:42:11,552 Stage-1 map = 100%,  reduce = 33%, Cumulative CPU 4.3 sec
2013-07-17 19:42:12,560 Stage-1 map = 100%,  reduce = 33%, Cumulative CPU 4.3 sec
2013-07-17 19:42:13,571 Stage-1 map = 100%,  reduce = 33%, Cumulative CPU 4.3 sec
2013-07-17 19:42:14,586 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 5.83 sec
2013-07-17 19:42:15,595 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 5.83 sec
2013-07-17 19:42:16,605 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 5.83 sec
2013-07-17 19:42:17,615 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 5.83 sec
2013-07-17 19:42:18,626 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 5.83 sec
2013-07-17 19:42:19,635 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 5.83 sec
2013-07-17 19:42:20,645 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 5.83 sec
MapReduce Total cumulative CPU time: 5 seconds 830 msec
Ended Job = job_201307162036_0006
MapReduce Jobs Launched: 
Job 0: Map: 2  Reduce: 1   Cumulative CPU: 5.83 sec   HDFS Read: 452 HDFS Write: 12 SUCCESS
Total MapReduce CPU Time Spent: 5 seconds 830 msec
OK
111
222
333
Time taken: 34.525 seconds

-------------------------------------------EXPLAIN-------------------------------------------------------------------------------------------

1,EXPLAIN:

hive> select *from tb1;
OK
111
222
333
444
Time taken: 0.083 seconds
hive> select *from tb1 where num=333;
Total MapReduce jobs = 1
Launching Job 1 out of 1
Number of reduce tasks is set to 0 since there's no reduce operator
Starting Job = job_201307162036_0030, Tracking URL = http://hadoopmaster:50030/jobdetails.jsp?jobid=job_201307162036_0030
Kill Command = /home/hadoop/bigdata/hadoop-1.0.4/libexec/../bin/hadoop job  -kill job_201307162036_0030
Hadoop job information for Stage-1: number of mappers: 1; number of reducers: 0
2013-07-17 21:40:37,868 Stage-1 map = 0%,  reduce = 0%
2013-07-17 21:40:43,898 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 0.68 sec
2013-07-17 21:40:44,904 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 0.68 sec
2013-07-17 21:40:45,911 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 0.68 sec
2013-07-17 21:40:46,915 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 0.68 sec
2013-07-17 21:40:47,921 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 0.68 sec
2013-07-17 21:40:48,929 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 0.68 sec
2013-07-17 21:40:49,936 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 0.68 sec
MapReduce Total cumulative CPU time: 680 msec
Ended Job = job_201307162036_0030
MapReduce Jobs Launched: 
Job 0: Map: 1   Cumulative CPU: 0.68 sec   HDFS Read: 226 HDFS Write: 4 SUCCESS
Total MapReduce CPU Time Spent: 680 msec
OK
333
Time taken: 20.543 seconds
hive> explain select *from tb1 where num=333;
OK
ABSTRACT SYNTAX TREE:
  (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME tb1))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (= (TOK_TABLE_OR_COL num) 333))))

STAGE DEPENDENCIES:
  Stage-1 is a root stage
  Stage-0 is a root stage

STAGE PLANS:
  Stage: Stage-1
    Map Reduce
      Alias -> Map Operator Tree:
        tb1 
          TableScan
            alias: tb1
            Filter Operator
              predicate:
                  expr: (num = 333)
                  type: boolean
              Select Operator
                expressions:
                      expr: num
                      type: int
                outputColumnNames: _col0
                File Output Operator
                  compressed: false
                  GlobalTableId: 0
                  table:
                      input format: org.apache.hadoop.mapred.TextInputFormat
                      output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat

  Stage: Stage-0
    Fetch Operator
      limit: -1


Time taken: 0.072 seconds

-------------------------------------

hive> explain  select * from tb1 left  semi  join tb2 on tb1.num=tb2.num ;        
OK
ABSTRACT SYNTAX TREE:
  (TOK_QUERY (TOK_FROM (TOK_LEFTSEMIJOIN (TOK_TABREF (TOK_TABNAME tb1)) (TOK_TABREF (TOK_TABNAME tb2)) (= (. (TOK_TABLE_OR_COL tb1) num) (. (TOK_TABLE_OR_COL tb2) num)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF))))

STAGE DEPENDENCIES:
  Stage-1 is a root stage
  Stage-0 is a root stage

STAGE PLANS:
  Stage: Stage-1
    Map Reduce
      Alias -> Map Operator Tree:
        tb1 
          TableScan
            alias: tb1
            Reduce Output Operator
              key expressions:
                    expr: num
                    type: int
              sort order: +
              Map-reduce partition columns:
                    expr: num
                    type: int
              tag: 0
              value expressions:
                    expr: num
                    type: int
        tb2 
          TableScan
            alias: tb2
            Select Operator
              expressions:
                    expr: num
                    type: int
              outputColumnNames: num
              Group By Operator
                bucketGroup: false
                keys:
                      expr: num
                      type: int
                mode: hash
                outputColumnNames: _col0
                Reduce Output Operator
                  key expressions:
                        expr: _col0
                        type: int
                  sort order: +
                  Map-reduce partition columns:
                        expr: _col0
                        type: int
                  tag: 1
      Reduce Operator Tree:
        Join Operator
          condition map:
               Left Semi Join 0 to 1
          condition expressions:
            0 {VALUE._col0}
            1 
          handleSkewJoin: false
          outputColumnNames: _col0
          Select Operator
            expressions:
                  expr: _col0
                  type: int
            outputColumnNames: _col0
            File Output Operator
              compressed: false
              GlobalTableId: 0
              table:
                  input format: org.apache.hadoop.mapred.TextInputFormat
                  output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat

  Stage: Stage-0
    Fetch Operator
      limit: -1


Time taken: 2.215 seconds
hive> 


2,EXPLAIN EXTENDED:

hive> explain extended select * from tb1 left  semi  join tb2 on tb1.num=tb2.num limit 2;
OK
ABSTRACT SYNTAX TREE:
  (TOK_QUERY (TOK_FROM (TOK_LEFTSEMIJOIN (TOK_TABREF (TOK_TABNAME tb1)) (TOK_TABREF (TOK_TABNAME tb2)) (= (. (TOK_TABLE_OR_COL tb1) num) (. (TOK_TABLE_OR_COL tb2) num)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_LIMIT 2)))

STAGE DEPENDENCIES:
  Stage-1 is a root stage
  Stage-0 is a root stage

STAGE PLANS:
  Stage: Stage-1
    Map Reduce
      Alias -> Map Operator Tree:
        tb1 
          TableScan
            alias: tb1
            GatherStats: false
            Reduce Output Operator
              key expressions:
                    expr: num
                    type: int
              sort order: +
              Map-reduce partition columns:
                    expr: num
                    type: int
              tag: 0
              value expressions:
                    expr: num
                    type: int
        tb2 
          TableScan
            alias: tb2
            GatherStats: false
            Select Operator
              expressions:
                    expr: num
                    type: int
              outputColumnNames: num
              Group By Operator
                bucketGroup: false
                keys:
                      expr: num
                      type: int
                mode: hash
                outputColumnNames: _col0
                Reduce Output Operator
                  key expressions:
                        expr: _col0
                        type: int
                  sort order: +
                  Map-reduce partition columns:
                        expr: _col0
                        type: int
                  tag: 1
      Needs Tagging: true
      Path -> Alias:
        hdfs://hadoopmaster:9000/user/hive/warehouse/tb1 [tb1]
        hdfs://hadoopmaster:9000/user/hive/warehouse/tb2 [tb2]
      Path -> Partition:
        hdfs://hadoopmaster:9000/user/hive/warehouse/tb1 
          Partition
            base file name: tb1
            input format: org.apache.hadoop.mapred.TextInputFormat
            output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
            properties:
              bucket_count -1
              columns num
              columns.types int
              file.inputformat org.apache.hadoop.mapred.TextInputFormat
              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
              location hdfs://hadoopmaster:9000/user/hive/warehouse/tb1
              name default.tb1
              numFiles 1
              numPartitions 0
              numRows 0
              rawDataSize 0
              serialization.ddl struct tb1 { i32 num}
              serialization.format 1
              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
              totalSize 16
              transient_lastDdlTime 1374060092
            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
          
              input format: org.apache.hadoop.mapred.TextInputFormat
              output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
              properties:
                bucket_count -1
                columns num
                columns.types int
                file.inputformat org.apache.hadoop.mapred.TextInputFormat
                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                location hdfs://hadoopmaster:9000/user/hive/warehouse/tb1
                name default.tb1
                numFiles 1
                numPartitions 0
                numRows 0
                rawDataSize 0
                serialization.ddl struct tb1 { i32 num}
                serialization.format 1
                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                totalSize 16
                transient_lastDdlTime 1374060092
              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
              name: default.tb1
            name: default.tb1
        hdfs://hadoopmaster:9000/user/hive/warehouse/tb2 
          Partition
            base file name: tb2
            input format: org.apache.hadoop.mapred.TextInputFormat
            output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
            properties:
              bucket_count -1
              columns num
              columns.types int
              file.inputformat org.apache.hadoop.mapred.TextInputFormat
              file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
              location hdfs://hadoopmaster:9000/user/hive/warehouse/tb2
              name default.tb2
              numFiles 1
              numPartitions 0
              numRows 0
              rawDataSize 0
              serialization.ddl struct tb2 { i32 num}
              serialization.format 1
              serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
              totalSize 16
              transient_lastDdlTime 1374059714
            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
          
              input format: org.apache.hadoop.mapred.TextInputFormat
              output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
              properties:
                bucket_count -1
                columns num
                columns.types int
                file.inputformat org.apache.hadoop.mapred.TextInputFormat
                file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                location hdfs://hadoopmaster:9000/user/hive/warehouse/tb2
                name default.tb2
                numFiles 1
                numPartitions 0
                numRows 0
                rawDataSize 0
                serialization.ddl struct tb2 { i32 num}
                serialization.format 1
                serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
                totalSize 16
                transient_lastDdlTime 1374059714
              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
              name: default.tb2
            name: default.tb2
      Reduce Operator Tree:
        Join Operator
          condition map:
               Left Semi Join 0 to 1
          condition expressions:
            0 {VALUE._col0}
            1 
          handleSkewJoin: false
          outputColumnNames: _col0
          Select Operator
            expressions:
                  expr: _col0
                  type: int
            outputColumnNames: _col0
            Limit
              File Output Operator
                compressed: false
                GlobalTableId: 0
                directory: hdfs://hadoopmaster:9000/tmp/hive-hadoop/hive_2013-07-17_20-19-28_883_5816218321243775043/-ext-10001
                NumFilesPerFileSink: 1
                Stats Publishing Key Prefix: hdfs://hadoopmaster:9000/tmp/hive-hadoop/hive_2013-07-17_20-19-28_883_5816218321243775043/-ext-10001/
                table:
                    input format: org.apache.hadoop.mapred.TextInputFormat
                    output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
                    properties:
                      columns _col0
                      columns.types int
                      escape.delim \
                      serialization.format 1
                TotalFiles: 1
                GatherStats: false
                MultiFileSpray: false
      Truncated Path -> Alias:
        /tb1 [tb1]
        /tb2 [tb2]

  Stage: Stage-0
    Fetch Operator
      limit: 2


Time taken: 0.138 seconds



  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值