carbondata-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "QiangCai (JIRA)" <j...@apache.org>
Subject [jira] [Assigned] (CARBONDATA-1237) UndeclaredThrowableException thrown when secondary index created on timestamp column for multiple loads(unique data) with enable.unsafe.columnpage = true
Date Tue, 27 Jun 2017 08:14:00 GMT

     [ https://issues.apache.org/jira/browse/CARBONDATA-1237?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]

QiangCai reassigned CARBONDATA-1237:
------------------------------------

    Assignee: QiangCai

> UndeclaredThrowableException thrown when secondary index created on timestamp column
for multiple loads(unique data) with enable.unsafe.columnpage = true
> ---------------------------------------------------------------------------------------------------------------------------------------------------------
>
>                 Key: CARBONDATA-1237
>                 URL: https://issues.apache.org/jira/browse/CARBONDATA-1237
>             Project: CarbonData
>          Issue Type: Bug
>          Components: data-query
>    Affects Versions: 1.1.0
>         Environment: SUSE 11 SP4 3 node cluster.
>            Reporter: Chetan Bhat
>            Assignee: QiangCai
>
> Steps:
> Configure enable.unsafe.columnpage = true.
> Create table having timestamp column.
> Load data -  2 loads using different csv each csv having unique data.
> Try to create secondary index on timestamp column.
> 0: jdbc:hive2://10.19.91.225:22550/default> CREATE TABLE uniqdata (CUST_ID int,CUST_NAME
String,ACTIVE_EMUI_VERSION string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2
bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 decimal(36,10),Double_COLUMN1 double,
Double_COLUMN2 double,INTEGER_COLUMN1 int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES('table_blocksize'='128','include_dictionary'='CUST_NAME,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1,CUST_ID','sort_columns'='CUST_NAME');
> +---------+--+
> | Result  |
> +---------+--+
> +---------+--+
> No rows selected (1.017 seconds)
> 0: jdbc:hive2://10.19.91.225:22550/default> LOAD DATA INPATH 'hdfs://hacluster/chetan/split1.csv'
into table uniqdata OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');
> +---------+--+
> | Result  |
> +---------+--+
> +---------+--+
> No rows selected (10.798 seconds)
> 0: jdbc:hive2://10.19.91.225:22550/default> LOAD DATA INPATH 'hdfs://hacluster/chetan/split2.csv'
into table uniqdata OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1');
> +---------+--+
> | Result  |
> +---------+--+
> +---------+--+
> No rows selected (16.006 seconds)
> 0: jdbc:hive2://10.19.91.225:22550/default> CREATE INDEX indextable2 ON TABLE uniqdata
(DOB) AS 'org.apache.carbondata.format';
> Actual Result(Issue ) - UndeclaredThrowableException thrown when secondary index tried
to be created on timestamp column.
> 0: jdbc:hive2://10.19.91.225:22550/default> CREATE INDEX indextable2 ON TABLE uniqdata
(DOB) AS 'org.apache.carbondata.format';
> Error: java.lang.reflect.UndeclaredThrowableException (state=,code=0)
> Driver Log - 
> 2017-06-27 14:06:34,557 | INFO  | [pool-23-thread-83] | Parsing command: string | org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)
> 2017-06-27 14:06:34,557 | INFO  | [pool-23-thread-83] | Parsing command: timestamp |
org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)
> 2017-06-27 14:06:34,557 | INFO  | [pool-23-thread-83] | Parsing command: timestamp |
org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)
> 2017-06-27 14:06:34,557 | INFO  | [pool-23-thread-83] | Parsing command: bigint | org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)
> 2017-06-27 14:06:34,558 | INFO  | [pool-23-thread-83] | Parsing command: bigint | org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)
> 2017-06-27 14:06:34,558 | INFO  | [pool-23-thread-83] | Parsing command: decimal(30,10)
| org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)
> 2017-06-27 14:06:34,558 | INFO  | [pool-23-thread-83] | Parsing command: decimal(36,10)
| org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)
> 2017-06-27 14:06:34,558 | INFO  | [pool-23-thread-83] | Parsing command: double | org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)
> 2017-06-27 14:06:34,559 | INFO  | [pool-23-thread-83] | Parsing command: double | org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)
> 2017-06-27 14:06:34,559 | INFO  | [pool-23-thread-83] | Parsing command: int | org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)
> 2017-06-27 14:06:34,559 | INFO  | [pool-23-thread-83] | Closed a connection to metastore,
current connections: 0 | org.apache.hadoop.hive.metastore.HiveMetaStoreClient.close(HiveMetaStoreClient.java:622)
> 2017-06-27 14:06:34,560 | INFO  | [pool-23-thread-83] | Trying to connect to metastore
with URI thrift://10.19.91.224:21088 | org.apache.hadoop.hive.metastore.HiveMetaStoreClient.open(HiveMetaStoreClient.java:469)
> 2017-06-27 14:06:34,585 | INFO  | [pool-23-thread-83] | Opened a connection to metastore,
current connections: 1 | org.apache.hadoop.hive.metastore.HiveMetaStoreClient.open(HiveMetaStoreClient.java:514)
> 2017-06-27 14:06:34,585 | INFO  | [pool-23-thread-83] | Connected to metastore. | org.apache.hadoop.hive.metastore.HiveMetaStoreClient.open(HiveMetaStoreClient.java:593)
> 2017-06-27 14:06:34,793 | AUDIT | [pool-23-thread-83] | [linux-225][chetan][Thread-776]Index
creation with Database name [default] and Index name [indextable2] failed | org.apache.carbondata.common.logging.impl.StandardLogService.audit(StandardLogService.java:207)
> 2017-06-27 14:06:34,793 | INFO  | [pool-23-thread-83] | OperationId=75f45d62-5e90-41f8-8e61-186a5e8e58e3
Result=FAIL | org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:320)
> 2017-06-27 14:06:34,793 | ERROR | [pool-23-thread-83] | Error executing query, currentState
RUNNING,  | org.apache.spark.internal.Logging$class.logError(Logging.scala:91)
> java.lang.reflect.UndeclaredThrowableException
> 	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1787)
> 	at org.apache.spark.util.Utils$.proxyOperate(Utils.scala:2688)
> 	at org.apache.spark.util.ACLFileUtils$.changeOwnerRecursivelyAfterOperation(ACLFileUtils.scala:57)
> 	at org.apache.spark.sql.execution.command.CreateIndexTable.run(carbonTableSchema.scala:359)
> 	at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:59)
> 	at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:57)
> 	at org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:75)
> 	at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114)
> 	at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114)
> 	at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:135)
> 	at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
> 	at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:132)
> 	at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:113)
> 	at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:125)
> 	at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:125)
> 	at org.apache.spark.sql.Dataset.<init>(Dataset.scala:185)
> 	at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:64)
> 	at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:592)
> 	at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:699)
> 	at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:251)
> 	at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:183)
> 	at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:180)
> 	at java.security.AccessController.doPrivileged(Native Method)
> 	at javax.security.auth.Subject.doAs(Subject.java:422)
> 	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1769)
> 	at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1.run(SparkExecuteStatementOperation.scala:193)
> 	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> 	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> 	at java.lang.Thread.run(Thread.java:748)
> Caused by: java.util.concurrent.ExecutionException: org.apache.spark.SparkException:
Job aborted due to stage failure: Task 0 in stage 98.0 failed 4 times, most recent failure:
Lost task 0.3 in stage 98.0 (TID 199, linux-224, executor 30): ExecutorLostFailure (executor
30 exited caused by one of the running tasks) Reason: Container marked as failed: container_1497850829760_1804_01_000042
on host: linux-224. Exit status: 134. Diagnostics: Exception from container-launch.
> Container id: container_1497850829760_1804_01_000042
> Exit code: 134
> Stack trace: ExitCodeException exitCode=134: 
> 	at org.apache.hadoop.util.Shell.runCommand(Shell.java:628)
> 	at org.apache.hadoop.util.Shell.run(Shell.java:525)
> 	at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:819)
> 	at org.apache.hadoop.yarn.server.nodemanager.LinuxContainerExecutor.launchContainer(LinuxContainerExecutor.java:388)
> 	at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:313)
> 	at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:88)
> 	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> 	at java.lang.Thread.run(Thread.java:748)
> Shell output: main : command provided 1
> main : run as user is spark2x
> main : requested yarn user is spark2x
> Before: Going to create dir
> Container exited with a non-zero exit code 134 and the last 4096 bytes from the error
logs are :
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in [jar:file:/opt/huawei/Bigdata/FusionInsight_Spark2x_V100R002C70/install/FusionInsight-Spark2x-2.1/spark/jars/slf4j-log4j12-1.7.16.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in [jar:file:/srv/BigData/hadoop/data1/nm/localdir/filecache/11/spark-archive-2x.zip/slf4j-log4j12-1.7.16.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in [jar:file:/opt/huawei/Bigdata/FusionInsight_HD_V100R002C70/install/FusionInsight-Hadoop-2.7.2/hadoop/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
> Driver stacktrace:
> 	at java.util.concurrent.FutureTask.report(FutureTask.java:122)
> 	at java.util.concurrent.FutureTask.get(FutureTask.java:192)
> 	at org.apache.carbondata.spark.rdd.SecondaryIndexCreator$$anonfun$createSecondaryIndex$1$$anonfun$apply$mcV$sp$1.apply(SecondaryIndexCreator.scala:106)
> 	at org.apache.carbondata.spark.rdd.SecondaryIndexCreator$$anonfun$createSecondaryIndex$1$$anonfun$apply$mcV$sp$1.apply(SecondaryIndexCreator.scala:105)
> 	at scala.collection.immutable.List.foreach(List.scala:381)
> 	at org.apache.carbondata.spark.rdd.SecondaryIndexCreator$$anonfun$createSecondaryIndex$1.apply$mcV$sp(SecondaryIndexCreator.scala:105)
> 	at scala.util.control.Breaks.breakable(Breaks.scala:38)
> 	at org.apache.carbondata.spark.rdd.SecondaryIndexCreator$.createSecondaryIndex(SecondaryIndexCreator.scala:103)
> 	at org.apache.carbondata.spark.rdd.CarbonDataRDDFactory$.secondaryIndexCallable(CarbonDataRDDFactory.scala:156)
> 	at org.apache.carbondata.spark.rdd.CarbonDataRDDFactory$.createSecondaryIndex(CarbonDataRDDFactory.scala:103)
> 	at org.apache.spark.sql.execution.command.loadDataForSecondaryIndex.run(carbonTableSchema.scala:491)
> 	at org.apache.spark.sql.execution.command.CreateIndexTable$$anonfun$run$2.apply$mcV$sp(carbonTableSchema.scala:361)
> 	at org.apache.spark.util.ACLFileUtils$$anonfun$changeOwnerRecursivelyAfterOperation$1.apply$mcV$sp(ACLFileUtils.scala:71)
> 	at org.apache.spark.util.Utils$$anon$7.run(Utils.scala:2690)
> 	at org.apache.spark.util.Utils$$anon$7.run(Utils.scala:2688)
> 	at java.security.AccessController.doPrivileged(Native Method)
> 	at javax.security.auth.Subject.doAs(Subject.java:422)
> 	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1769)
> 	... 30 more
> Caused by: org.apache.spark.SparkException: Job aborted due to stage failure: Task 0
in stage 98.0 failed 4 times, most recent failure: Lost task 0.3 in stage 98.0 (TID 199, linux-224,
executor 30): ExecutorLostFailure (executor 30 exited caused by one of the running tasks)
Reason: Container marked as failed: container_1497850829760_1804_01_000042 on host: linux-224.
Exit status: 134. Diagnostics: Exception from container-launch.
> Container id: container_1497850829760_1804_01_000042
> Exit code: 134
> Stack trace: ExitCodeException exitCode=134: 
> 	at org.apache.hadoop.util.Shell.runCommand(Shell.java:628)
> 	at org.apache.hadoop.util.Shell.run(Shell.java:525)
> 	at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:819)
> 	at org.apache.hadoop.yarn.server.nodemanager.LinuxContainerExecutor.launchContainer(LinuxContainerExecutor.java:388)
> 	at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:313)
> 	at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:88)
> 	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> 	at java.lang.Thread.run(Thread.java:748)
> Shell output: main : command provided 1
> main : run as user is spark2x
> main : requested yarn user is spark2x
> Before: Going to create dir
> Container exited with a non-zero exit code 134 and the last 4096 bytes from the error
logs are :
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in [jar:file:/opt/huawei/Bigdata/FusionInsight_Spark2x_V100R002C70/install/FusionInsight-Spark2x-2.1/spark/jars/slf4j-log4j12-1.7.16.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in [jar:file:/srv/BigData/hadoop/data1/nm/localdir/filecache/11/spark-archive-2x.zip/slf4j-log4j12-1.7.16.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in [jar:file:/opt/huawei/Bigdata/FusionInsight_HD_V100R002C70/install/FusionInsight-Hadoop-2.7.2/hadoop/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
> Driver stacktrace:
> 	at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1478)
> 	at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1466)
> 	at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1465)
> 	at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
> 	at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
> 	at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1465)
> 	at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:813)
> 	at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:813)
> 	at scala.Option.foreach(Option.scala:257)
> 	at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:813)
> 	at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1693)
> 	at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1648)
> 	at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1637)
> 	at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
> 	at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:639)
> 	at org.apache.spark.SparkContext.runJob(SparkContext.scala:1946)
> 	at org.apache.spark.SparkContext.runJob(SparkContext.scala:1959)
> 	at org.apache.spark.SparkContext.runJob(SparkContext.scala:1972)
> 	at org.apache.spark.SparkContext.runJob(SparkContext.scala:1986)
> 	at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:937)
> 	at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
> 	at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
> 	at org.apache.spark.rdd.RDD.withScope(RDD.scala:363)
> 	at org.apache.spark.rdd.RDD.collect(RDD.scala:936)
> 	at org.apache.carbondata.spark.rdd.SecondaryIndexCreator$$anonfun$createSecondaryIndex$2$$anon$1.call(SecondaryIndexCreator.scala:92)
> 	at org.apache.carbondata.spark.rdd.SecondaryIndexCreator$$anonfun$createSecondaryIndex$2$$anon$1.call(SecondaryIndexCreator.scala:77)
> 	... 4 more
> 2017-06-27 14:06:34,794 | ERROR | [pool-23-thread-83] | Error running hive query:  |
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:187)
> org.apache.hive.service.cli.HiveSQLException: java.lang.reflect.UndeclaredThrowableException
> 	at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:333)
> 	at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:183)
> 	at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:180)
> 	at java.security.AccessController.doPrivileged(Native Method)
> 	at javax.security.auth.Subject.doAs(Subject.java:422)
> 	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1769)
> 	at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1.run(SparkExecuteStatementOperation.scala:193)
> 	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> 	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> 	at java.lang.Thread.run(Thread.java:748)
> 2017-06-27 14:06:36,996 | INFO  | [main] | the jdbc URL switch open:jdbc:hive2://10.19.91.225:22550/default;principal=spark2x/hadoop.hadoop.com@HADOOP.COM;healthcheck=true;saslQop=auth-conf;auth=KERBEROS;user.principal=spark2x/hadoop.hadoop.com@HADOOP.COM;user.keytab=/opt/huawei/Bigdata/FusionInsight_Spark2x_V100R002C70/install/FusionInsight-Spark2x-2.1/keytab/spark2x/JDBCServer/spark2x.keytab
| org.apache.hive.jdbc.HiveConnection.<init>(HiveConnection.java:145)
> 2017-06-27 14:06:37,003 | INFO  | [main] | Supplied authorities: 10.19.91.225:22550 |
org.apache.hive.jdbc.Utils.parseURL(Utils.java:391)
> 2017-06-27 14:06:37,625 | WARN  | [main] | HiveConf of name hive.thriftProxy.user does
not exist | org.apache.hadoop.hive.conf.HiveConf.initialize(HiveConf.java:2920)
> 2017-06-27 14:06:37,625 | WARN  | [main] | HiveConf of name hive.thriftProxy.address
does not exist | org.apache.hadoop.hive.conf.HiveConf.initialize(HiveConf.java:2920)
> 2017-06-27 14:06:37,629 | INFO  | [main] | Resolved authority: 10.19.91.225:22550 | org.apache.hive.jdbc.Utils.parseURL(Utils.java:490)
> 2017-06-27 14:06:37,785 | WARN  | [main] | Unable to load native-hadoop library for your
platform... using builtin-java classes where applicable | org.apache.hadoop.util.NativeCodeLoader.<clinit>(NativeCodeLoader.java:62)
> 2017-06-27 14:06:38,050 | INFO  | [main] | user login success. | org.apache.hive.jdbc.HiveConnection.login(HiveConnection.java:295)
> 2017-06-27 14:06:38,118 | INFO  | [main] | Will try to open client transport with JDBC
Uri: jdbc:hive2://10.19.91.225:22550/default;principal=spark2x/hadoop.hadoop.com@HADOOP.COM;healthcheck=true;saslQop=auth-conf;auth=KERBEROS;user.principal=spark2x/hadoop.hadoop.com@HADOOP.COM;user.keytab=/opt/huawei/Bigdata/FusionInsight_Spark2x_V100R002C70/install/FusionInsight-Spark2x-2.1/keytab/spark2x/JDBCServer/spark2x.keytab
| org.apache.hive.jdbc.HiveConnection.openTransport(HiveConnection.java:317)
> 2017-06-27 14:06:38,177 | INFO  | [HiveServer2-Handler-Pool: Thread-258] | Client protocol
version: HIVE_CLI_SERVICE_PROTOCOL_V8 | org.apache.hive.service.cli.thrift.ThriftCLIService.OpenSession(ThriftCLIService.java:461)
> 2017-06-27 14:06:38,177 | INFO  | [HiveServer2-Handler-Pool: Thread-258] | Trying to
connect to metastore with URI thrift://10.19.91.224:21088 | org.apache.hadoop.hive.metastore.HiveMetaStoreClient.open(HiveMetaStoreClient.java:469)
> 2017-06-27 14:06:38,180 | INFO  | [HiveServer2-Handler-Pool: Thread-258] | Opened a connection
to metastore, current connections: 38 | org.apache.hadoop.hive.metastore.HiveMetaStoreClient.open(HiveMetaStoreClient.java:514)
> 2017-06-27 14:06:38,181 | INFO  | [HiveServer2-Handler-Pool: Thread-258] | Connected
to metastore. | org.apache.hadoop.hive.metastore.HiveMetaStoreClient.open(HiveMetaStoreClient.java:593)
> 2017-06-27 14:06:38,245 | WARN  | [HiveServer2-Handler-Pool: Thread-258] | load mapred-default.xml,
HIVE_CONF_DIR env not found! | org.apache.hadoop.hive.ql.session.SessionState.loadMapredDefaultXml(SessionState.java:1101)
> 2017-06-27 14:06:38,246 | INFO  | [HiveServer2-Handler-Pool: Thread-258] | Closed a connection
to metastore, current connections: 37 | org.apache.hadoop.hive.metastore.HiveMetaStoreClient.close(HiveMetaStoreClient.java:622)
> 2017-06-27 14:06:38,247 | INFO  | [HiveServer2-Handler-Pool: Thread-258] | Trying to
connect to metastore with URI thrift://10.19.91.224:21088 | org.apache.hadoop.hive.metastore.HiveMetaStoreClient.open(HiveMetaStoreClient.java:469)
> Expected Result : Secondary index creation on timestamp column should be success.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

Mime
View raw message