Return-Path: X-Original-To: apmail-hive-commits-archive@www.apache.org Delivered-To: apmail-hive-commits-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id D3D8810F36 for ; Fri, 13 Dec 2013 10:58:28 +0000 (UTC) Received: (qmail 54529 invoked by uid 500); 13 Dec 2013 10:58:28 -0000 Delivered-To: apmail-hive-commits-archive@hive.apache.org Received: (qmail 54504 invoked by uid 500); 13 Dec 2013 10:58:27 -0000 Mailing-List: contact commits-help@hive.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: hive-dev@hive.apache.org Delivered-To: mailing list commits@hive.apache.org Received: (qmail 54496 invoked by uid 99); 13 Dec 2013 10:58:27 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 13 Dec 2013 10:58:27 +0000 X-ASF-Spam-Status: No, hits=-2000.0 required=5.0 tests=ALL_TRUSTED X-Spam-Check-By: apache.org Received: from [140.211.11.4] (HELO eris.apache.org) (140.211.11.4) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 13 Dec 2013 10:58:22 +0000 Received: from eris.apache.org (localhost [127.0.0.1]) by eris.apache.org (Postfix) with ESMTP id 2B4E42388BEF; Fri, 13 Dec 2013 10:57:37 +0000 (UTC) Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit Subject: svn commit: r1550684 [7/24] - in /hive/branches/tez: ./ ant/src/org/apache/hadoop/hive/ant/ beeline/ beeline/src/java/org/apache/hive/beeline/ common/src/java/org/apache/hadoop/hive/conf/ contrib/src/test/results/clientnegative/ contrib/src/test/result... Date: Fri, 13 Dec 2013 10:56:54 -0000 To: commits@hive.apache.org From: gunther@apache.org X-Mailer: svnmailer-1.0.9 Message-Id: <20131213105737.2B4E42388BEF@eris.apache.org> X-Virus-Checked: Checked by ClamAV on apache.org Modified: hive/branches/tez/ql/src/test/results/clientpositive/annotate_stats_select.q.out URL: http://svn.apache.org/viewvc/hive/branches/tez/ql/src/test/results/clientpositive/annotate_stats_select.q.out?rev=1550684&r1=1550683&r2=1550684&view=diff ============================================================================== --- hive/branches/tez/ql/src/test/results/clientpositive/annotate_stats_select.q.out (original) +++ hive/branches/tez/ql/src/test/results/clientpositive/annotate_stats_select.q.out Fri Dec 13 10:56:38 2013 @@ -10,6 +10,7 @@ PREHOOK: query: create table if not exis ts1 timestamp, da1 timestamp, s1 string, + vc1 varchar(5), m1 map, l1 array, st1 struct @@ -29,6 +30,7 @@ POSTHOOK: query: create table if not exi ts1 timestamp, da1 timestamp, s1 string, + vc1 varchar(5), m1 map, l1 array, st1 struct @@ -78,6 +80,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMP POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] PREHOOK: query: -- basicStatState: COMPLETE colStatState: NONE numRows: 2 rawDataSize: 1514 explain extended select * from alltypes_orc PREHOOK: type: QUERY @@ -98,6 +101,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMP POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)))) @@ -112,7 +116,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: NONE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: NONE GatherStats: false Select Operator expressions: @@ -138,25 +142,26 @@ STAGE PLANS: type: timestamp expr: s1 type: string + expr: vc1 + type: varchar(5) expr: m1 type: map expr: l1 type: array expr: st1 type: struct - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14 Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: NONE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: NONE ListSink - PREHOOK: query: -- statistics for complex types are not supported yet -analyze table alltypes_orc compute statistics for columns bo1, ti1, si1, i1, bi1, f1, d1,s1 +analyze table alltypes_orc compute statistics for columns bo1, ti1, si1, i1, bi1, f1, d1, s1, vc1 PREHOOK: type: QUERY PREHOOK: Input: default@alltypes_orc #### A masked pattern was here #### POSTHOOK: query: -- statistics for complex types are not supported yet -analyze table alltypes_orc compute statistics for columns bo1, ti1, si1, i1, bi1, f1, d1,s1 +analyze table alltypes_orc compute statistics for columns bo1, ti1, si1, i1, bi1, f1, d1, s1, vc1 POSTHOOK: type: QUERY POSTHOOK: Input: default@alltypes_orc #### A masked pattern was here #### @@ -174,6 +179,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMP POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] PREHOOK: query: -- numRows: 2 rawDataSize: 1514 explain extended select * from alltypes_orc PREHOOK: type: QUERY @@ -194,6 +200,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMP POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)))) @@ -208,7 +215,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: PARTIAL + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: PARTIAL GatherStats: false Select Operator expressions: @@ -234,18 +241,19 @@ STAGE PLANS: type: timestamp expr: s1 type: string + expr: vc1 + type: varchar(5) expr: m1 type: map expr: l1 type: array expr: st1 type: struct - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14 Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: PARTIAL + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: PARTIAL ListSink - PREHOOK: query: -- numRows: 2 rawDataSize: 8 explain extended select bo1 from alltypes_orc PREHOOK: type: QUERY @@ -266,6 +274,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMP POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL bo1))))) @@ -281,7 +290,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -324,19 +333,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -346,19 +355,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -370,7 +379,6 @@ STAGE PLANS: Fetch Operator limit: -1 - PREHOOK: query: -- col alias renaming -- numRows: 2 rawDataSize: 8 explain extended select i1 as int1 from alltypes_orc @@ -393,6 +401,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMP POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL i1) int1)))) @@ -408,7 +417,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -451,19 +460,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -473,19 +482,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -497,7 +506,6 @@ STAGE PLANS: Fetch Operator limit: -1 - PREHOOK: query: -- numRows: 2 rawDataSize: 174 explain extended select s1 from alltypes_orc PREHOOK: type: QUERY @@ -518,6 +526,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMP POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL s1))))) @@ -533,7 +542,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -576,19 +585,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -598,19 +607,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -622,7 +631,6 @@ STAGE PLANS: Fetch Operator limit: -1 - PREHOOK: query: -- column statistics for complex types unsupported and so statistics will not be updated -- numRows: 2 rawDataSize: 1514 explain extended select m1 from alltypes_orc @@ -645,6 +653,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMP POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL m1))))) @@ -660,7 +669,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: NONE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: NONE GatherStats: false Select Operator expressions: @@ -668,14 +677,14 @@ STAGE PLANS: type: map outputColumnNames: _col0 Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: NONE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: NONE File Output Operator compressed: false GlobalTableId: 0 #### A masked pattern was here #### NumFilesPerFileSink: 1 Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: NONE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: NONE #### A masked pattern was here #### table: input format: org.apache.hadoop.mapred.TextInputFormat @@ -703,19 +712,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -725,19 +734,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -749,7 +758,6 @@ STAGE PLANS: Fetch Operator limit: -1 - PREHOOK: query: -- numRows: 2 rawDataSize: 246 explain extended select bo1, ti1, si1, i1, bi1, f1, d1,s1 from alltypes_orc PREHOOK: type: QUERY @@ -770,6 +778,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMP POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL bo1)) (TOK_SELEXPR (TOK_TABLE_OR_COL ti1)) (TOK_SELEXPR (TOK_TABLE_OR_COL si1)) (TOK_SELEXPR (TOK_TABLE_OR_COL i1)) (TOK_SELEXPR (TOK_TABLE_OR_COL bi1)) (TOK_SELEXPR (TOK_TABLE_OR_COL f1)) (TOK_SELEXPR (TOK_TABLE_OR_COL d1)) (TOK_SELEXPR (TOK_TABLE_OR_COL s1))))) @@ -785,7 +794,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -842,19 +851,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -864,19 +873,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -888,7 +897,6 @@ STAGE PLANS: Fetch Operator limit: -1 - PREHOOK: query: -- numRows: 2 rawDataSize: 0 explain extended select null from alltypes_orc PREHOOK: type: QUERY @@ -909,6 +917,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMP POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_NULL)))) @@ -924,7 +933,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -967,19 +976,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -989,19 +998,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -1013,7 +1022,6 @@ STAGE PLANS: Fetch Operator limit: -1 - PREHOOK: query: -- numRows: 2 rawDataSize: 8 explain extended select 11 from alltypes_orc PREHOOK: type: QUERY @@ -1034,6 +1042,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMP POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR 11)))) @@ -1049,7 +1058,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -1092,19 +1101,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -1114,19 +1123,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -1138,7 +1147,6 @@ STAGE PLANS: Fetch Operator limit: -1 - PREHOOK: query: -- numRows: 2 rawDataSize: 16 explain extended select 11L from alltypes_orc PREHOOK: type: QUERY @@ -1159,6 +1167,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMP POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR 11L)))) @@ -1174,7 +1183,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -1217,19 +1226,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -1239,19 +1248,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -1263,7 +1272,6 @@ STAGE PLANS: Fetch Operator limit: -1 - PREHOOK: query: -- numRows: 2 rawDataSize: 16 explain extended select 11.0 from alltypes_orc PREHOOK: type: QUERY @@ -1284,6 +1292,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMP POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR 11.0)))) @@ -1299,7 +1308,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -1342,19 +1351,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -1364,19 +1373,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -1388,7 +1397,6 @@ STAGE PLANS: Fetch Operator limit: -1 - PREHOOK: query: -- numRows: 2 rawDataSize: 178 explain extended select "hello" from alltypes_orc PREHOOK: type: QUERY @@ -1409,6 +1417,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMP POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR "hello")))) @@ -1424,7 +1433,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -1467,19 +1476,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -1489,19 +1498,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -1513,7 +1522,6 @@ STAGE PLANS: Fetch Operator limit: -1 - PREHOOK: query: explain extended select cast("hello" as char(5)) from alltypes_orc PREHOOK: type: QUERY POSTHOOK: query: explain extended select cast("hello" as char(5)) from alltypes_orc @@ -1532,6 +1540,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMP POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION (TOK_CHAR 5) "hello"))))) @@ -1547,7 +1556,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -1590,19 +1599,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -1612,19 +1621,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -1636,7 +1645,6 @@ STAGE PLANS: Fetch Operator limit: -1 - PREHOOK: query: explain extended select cast("hello" as varchar(5)) from alltypes_orc PREHOOK: type: QUERY POSTHOOK: query: explain extended select cast("hello" as varchar(5)) from alltypes_orc @@ -1655,6 +1663,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMP POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION (TOK_VARCHAR 5) "hello"))))) @@ -1670,7 +1679,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -1713,19 +1722,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -1735,19 +1744,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -1759,7 +1768,6 @@ STAGE PLANS: Fetch Operator limit: -1 - PREHOOK: query: -- numRows: 2 rawDataSize: 96 explain extended select unbase64("0xe23") from alltypes_orc PREHOOK: type: QUERY @@ -1780,6 +1788,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMP POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION unbase64 "0xe23"))))) @@ -1795,7 +1804,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -1838,19 +1847,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -1860,19 +1869,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -1884,7 +1893,6 @@ STAGE PLANS: Fetch Operator limit: -1 - PREHOOK: query: -- numRows: 2 rawDataSize: 16 explain extended select cast("1" as TINYINT), cast("20" as SMALLINT) from alltypes_orc PREHOOK: type: QUERY @@ -1905,6 +1913,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMP POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION TOK_TINYINT "1")) (TOK_SELEXPR (TOK_FUNCTION TOK_SMALLINT "20"))))) @@ -1920,7 +1929,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -1965,19 +1974,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -1987,19 +1996,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -2011,7 +2020,6 @@ STAGE PLANS: Fetch Operator limit: -1 - PREHOOK: query: -- numRows: 2 rawDataSize: 80 explain extended select cast("1970-12-31 15:59:58.174" as TIMESTAMP) from alltypes_orc PREHOOK: type: QUERY @@ -2032,6 +2040,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMP POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION TOK_TIMESTAMP "1970-12-31 15:59:58.174"))))) @@ -2047,7 +2056,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -2090,19 +2099,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -2112,19 +2121,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -2136,7 +2145,6 @@ STAGE PLANS: Fetch Operator limit: -1 - PREHOOK: query: -- numRows: 2 rawDataSize: 112 explain extended select cast("1970-12-31 15:59:58.174" as DATE) from alltypes_orc PREHOOK: type: QUERY @@ -2157,6 +2165,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMP POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION TOK_DATE "1970-12-31 15:59:58.174"))))) @@ -2172,7 +2181,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -2215,19 +2224,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -2237,19 +2246,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -2261,7 +2270,6 @@ STAGE PLANS: Fetch Operator limit: -1 - PREHOOK: query: -- numRows: 2 rawDataSize: 224 explain extended select cast("58.174" as DECIMAL) from alltypes_orc PREHOOK: type: QUERY @@ -2282,6 +2290,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMP POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION TOK_DECIMAL "58.174"))))) @@ -2297,7 +2306,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -2340,19 +2349,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -2362,19 +2371,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | [... 1185 lines stripped ...]