drill-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "tooptoop4 (JIRA)" <j...@apache.org>
Subject [jira] [Updated] (DRILL-5932) HiveServer2 queries throw error with jdbc connection
Date Sat, 04 Nov 2017 11:44:00 GMT

     [ https://issues.apache.org/jira/browse/DRILL-5932?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

tooptoop4 updated DRILL-5932:
-----------------------------
    Description: 
Basic hive queries all throw error!

{code:sql}

copied https://repo1.maven.org/maven2/org/apache/hive/hive-jdbc/2.3.0/hive-jdbc-2.3.0-standalone.jar to /usr/lib/apache-drill-1.11.0/jars/3rdparty/hive-jdbc-2.3.0-standalone.jar

added this storage plugin:

{
  "type": "jdbc",
  "driver": "org.apache.hive.jdbc.HiveDriver",
  "url": "jdbc:hive2://host:10000/default",
  "username": "hive",
  "password": "hive1234",
  "enabled": true
}

[ec2-user@host ~]$ cd /usr/lib/apache-drill-1.11.0
[ec2-user@host apache-drill-1.11.0]$ ./bin/drill-embedded
OpenJDK 64-Bit Server VM warning: ignoring option MaxPermSize=512M; support was removed in 8.0
Nov 01, 2017 7:53:53 AM org.glassfish.jersey.server.ApplicationHandler initialize
INFO: Initiating Jersey application, version Jersey: 2.8 2014-04-29 01:25:26...
apache drill 1.11.0
"this isn't your grandfather's sql"
0: jdbc:drill:zk=local> SELECT count(*) FROM hive2.`contact`;
Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.

sql SELECT COUNT(*) AS EXPR$0
FROM (SELECT 0 AS $f0
FROM.default.contact) AS t
plugin hive2
Fragment 0:0

[Error Id: 4b293e97-7547-49c5-91da-b9ee2f2184fc on ip-myip.mydomain.orghere.com:31010] (state=,code=0)




0: jdbc:drill:zk=local> ALTER SESSION SET `exec.errors.verbose` = true;
+-------+-------------------------------+
|  ok   |            summary            |
+-------+-------------------------------+
| true  | exec.errors.verbose updated.  |
+-------+-------------------------------+
1 row selected (0.351 seconds)

0: jdbc:drill:zk=local> select * from hive2.contact;
Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.

sql SELECT *
FROM.default.contact
plugin hive2
Fragment 0:0

[Error Id: fe36b026-e8ff-4354-af6c-6073130680c9 on ip-10-160-153-163.syd.non.c1.macquarie.com:31010]

  (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 2:4 cannot recognize input near '.' 'default' '.' in join source
    org.apache.hive.jdbc.Utils.verifySuccess():267
    org.apache.hive.jdbc.Utils.verifySuccessWithInfo():253
    org.apache.hive.jdbc.HiveStatement.runAsyncOnServer():313
    org.apache.hive.jdbc.HiveStatement.execute():253
    org.apache.hive.jdbc.HiveStatement.executeQuery():476
    org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
    org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
    org.apache.drill.exec.store.jdbc.JdbcRecordReader.setup():177
    org.apache.drill.exec.physical.impl.ScanBatch.<init>():104
    org.apache.drill.exec.physical.impl.ScanBatch.<init>():126
    org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():40
    org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():33
    org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():156
    org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
    org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():136
    org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
    org.apache.drill.exec.physical.impl.ImplCreator.getRootExec():109
    org.apache.drill.exec.physical.impl.ImplCreator.getExec():87
    org.apache.drill.exec.work.fragment.FragmentExecutor.run():207
    org.apache.drill.common.SelfCleaningRunnable.run():38
    java.util.concurrent.ThreadPoolExecutor.runWorker():1149
    java.util.concurrent.ThreadPoolExecutor$Worker.run():624
    java.lang.Thread.run():748
  Caused By (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 2:4 cannot recognize input near '.' 'default' '.' in join source
    org.apache.hive.service.cli.operation.Operation.toSQLException():380
    org.apache.hive.service.cli.operation.SQLOperation.prepare():206
    org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
    org.apache.hive.service.cli.operation.Operation.run():320
    org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
    org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
    sun.reflect.GeneratedMethodAccessor86.invoke():-1
    sun.reflect.DelegatingMethodAccessorImpl.invoke():43
    java.lang.reflect.Method.invoke():498
    org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
    org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
    org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
    java.security.AccessController.doPrivileged():-2
    javax.security.auth.Subject.doAs():422
    org.apache.hadoop.security.UserGroupInformation.doAs():1807
    org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
    com.sun.proxy.$Proxy42.executeStatementAsync():-1
    org.apache.hive.service.cli.CLIService.executeStatementAsync():310
    org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
    org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
    org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
    org.apache.thrift.ProcessFunction.process():39
    org.apache.thrift.TBaseProcessor.process():39
    org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
    org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
    java.util.concurrent.ThreadPoolExecutor.runWorker():1149
    java.util.concurrent.ThreadPoolExecutor$Worker.run():624
    java.lang.Thread.run():748
  Caused By (java.lang.RuntimeException) org.apache.hadoop.hive.ql.parse.ParseException:line 2:4 cannot recognize input near '.' 'default' '.' in join source
    org.apache.hadoop.hive.ql.parse.ParseDriver.parse():207
    org.apache.hadoop.hive.ql.parse.ParseUtils.parse():77
    org.apache.hadoop.hive.ql.parse.ParseUtils.parse():70
    org.apache.hadoop.hive.ql.Driver.compile():468
    org.apache.hadoop.hive.ql.Driver.compileInternal():1316
    org.apache.hadoop.hive.ql.Driver.compileAndRespond():1294
    org.apache.hive.service.cli.operation.SQLOperation.prepare():204
    org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
    org.apache.hive.service.cli.operation.Operation.run():320
    org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
    org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
    sun.reflect.GeneratedMethodAccessor86.invoke():-1
    sun.reflect.DelegatingMethodAccessorImpl.invoke():43
    java.lang.reflect.Method.invoke():498
    org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
    org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
    org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
    java.security.AccessController.doPrivileged():-2
    javax.security.auth.Subject.doAs():422
    org.apache.hadoop.security.UserGroupInformation.doAs():1807
    org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
    com.sun.proxy.$Proxy42.executeStatementAsync():-1
    org.apache.hive.service.cli.CLIService.executeStatementAsync():310
    org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
    org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
    org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
    org.apache.thrift.ProcessFunction.process():39
    org.apache.thrift.TBaseProcessor.process():39
    org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
    org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
    java.util.concurrent.ThreadPoolExecutor.runWorker():1149
    java.util.concurrent.ThreadPoolExecutor$Worker.run():624
    java.lang.Thread.run():748 (state=,code=0)
0: jdbc:drill:zk=local> select * from hive2.`default`.contact;
Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.

sql SELECT *
FROM.default.contact
plugin hive2
Fragment 0:0

[Error Id: 7b7aae92-fabd-49dd-ab84-15ca00189882 on ip-10-160-153-163.syd.non.c1.macquarie.com:31010]

  (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 2:4 cannot recognize input near '.' 'default' '.' in join source
    org.apache.hive.jdbc.Utils.verifySuccess():267
    org.apache.hive.jdbc.Utils.verifySuccessWithInfo():253
    org.apache.hive.jdbc.HiveStatement.runAsyncOnServer():313
    org.apache.hive.jdbc.HiveStatement.execute():253
    org.apache.hive.jdbc.HiveStatement.executeQuery():476
    org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
    org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
    org.apache.drill.exec.store.jdbc.JdbcRecordReader.setup():177
    org.apache.drill.exec.physical.impl.ScanBatch.<init>():104
    org.apache.drill.exec.physical.impl.ScanBatch.<init>():126
    org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():40
    org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():33
    org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():156
    org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
    org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():136
    org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
    org.apache.drill.exec.physical.impl.ImplCreator.getRootExec():109
    org.apache.drill.exec.physical.impl.ImplCreator.getExec():87
    org.apache.drill.exec.work.fragment.FragmentExecutor.run():207
    org.apache.drill.common.SelfCleaningRunnable.run():38
    java.util.concurrent.ThreadPoolExecutor.runWorker():1149
    java.util.concurrent.ThreadPoolExecutor$Worker.run():624
    java.lang.Thread.run():748
  Caused By (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 2:4 cannot recognize input near '.' 'default' '.' in join source
    org.apache.hive.service.cli.operation.Operation.toSQLException():380
    org.apache.hive.service.cli.operation.SQLOperation.prepare():206
    org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
    org.apache.hive.service.cli.operation.Operation.run():320
    org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
    org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
    sun.reflect.GeneratedMethodAccessor86.invoke():-1
    sun.reflect.DelegatingMethodAccessorImpl.invoke():43
    java.lang.reflect.Method.invoke():498
    org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
    org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
    org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
    java.security.AccessController.doPrivileged():-2
    javax.security.auth.Subject.doAs():422
    org.apache.hadoop.security.UserGroupInformation.doAs():1807
    org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
    com.sun.proxy.$Proxy42.executeStatementAsync():-1
    org.apache.hive.service.cli.CLIService.executeStatementAsync():310
    org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
    org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
    org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
    org.apache.thrift.ProcessFunction.process():39
    org.apache.thrift.TBaseProcessor.process():39
    org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
    org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
    java.util.concurrent.ThreadPoolExecutor.runWorker():1149
    java.util.concurrent.ThreadPoolExecutor$Worker.run():624
    java.lang.Thread.run():748
  Caused By (java.lang.RuntimeException) org.apache.hadoop.hive.ql.parse.ParseException:line 2:4 cannot recognize input near '.' 'default' '.' in join source
    org.apache.hadoop.hive.ql.parse.ParseDriver.parse():207
    org.apache.hadoop.hive.ql.parse.ParseUtils.parse():77
    org.apache.hadoop.hive.ql.parse.ParseUtils.parse():70
    org.apache.hadoop.hive.ql.Driver.compile():468
    org.apache.hadoop.hive.ql.Driver.compileInternal():1316
    org.apache.hadoop.hive.ql.Driver.compileAndRespond():1294
    org.apache.hive.service.cli.operation.SQLOperation.prepare():204
    org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
    org.apache.hive.service.cli.operation.Operation.run():320
    org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
    org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
    sun.reflect.GeneratedMethodAccessor86.invoke():-1
    sun.reflect.DelegatingMethodAccessorImpl.invoke():43
    java.lang.reflect.Method.invoke():498
    org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
    org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
    org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
    java.security.AccessController.doPrivileged():-2
    javax.security.auth.Subject.doAs():422
    org.apache.hadoop.security.UserGroupInformation.doAs():1807
    org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
    com.sun.proxy.$Proxy42.executeStatementAsync():-1
    org.apache.hive.service.cli.CLIService.executeStatementAsync():310
    org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
    org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
    org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
    org.apache.thrift.ProcessFunction.process():39
    org.apache.thrift.TBaseProcessor.process():39
    org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
    org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
    java.util.concurrent.ThreadPoolExecutor.runWorker():1149
    java.util.concurrent.ThreadPoolExecutor$Worker.run():624
    java.lang.Thread.run():748 (state=,code=0)
0: jdbc:drill:zk=local>


0: jdbc:drill:zk=local> SELECT count(*) FROM hive2.`contact`;
Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.

sql SELECT COUNT(*) AS EXPR$0
FROM (SELECT 0 AS $f0
FROM.default.contact) AS t
plugin hive2
Fragment 0:0

[Error Id: ac5cc8f2-69a4-455e-b1a6-5d22cde99729 on ip-myip.mydomain.orghere.com:31010]

  (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 1:23 missing EOF at '$' near 'EXPR'
    org.apache.hive.jdbc.Utils.verifySuccess():267
    org.apache.hive.jdbc.Utils.verifySuccessWithInfo():253
    org.apache.hive.jdbc.HiveStatement.runAsyncOnServer():313
    org.apache.hive.jdbc.HiveStatement.execute():253
    org.apache.hive.jdbc.HiveStatement.executeQuery():476
    org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
    org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
    org.apache.drill.exec.store.jdbc.JdbcRecordReader.setup():177
    org.apache.drill.exec.physical.impl.ScanBatch.<init>():104
    org.apache.drill.exec.physical.impl.ScanBatch.<init>():126
    org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():40
    org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():33
    org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():156
    org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
    org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():136
    org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
    org.apache.drill.exec.physical.impl.ImplCreator.getRootExec():109
    org.apache.drill.exec.physical.impl.ImplCreator.getExec():87
    org.apache.drill.exec.work.fragment.FragmentExecutor.run():207
    org.apache.drill.common.SelfCleaningRunnable.run():38
    java.util.concurrent.ThreadPoolExecutor.runWorker():1149
    java.util.concurrent.ThreadPoolExecutor$Worker.run():624
    java.lang.Thread.run():748
  Caused By (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 1:23 missing EOF at '$' near 'EXPR'
    org.apache.hive.service.cli.operation.Operation.toSQLException():380
    org.apache.hive.service.cli.operation.SQLOperation.prepare():206
    org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
    org.apache.hive.service.cli.operation.Operation.run():320
    org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
    org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
    sun.reflect.GeneratedMethodAccessor86.invoke():-1
    sun.reflect.DelegatingMethodAccessorImpl.invoke():43
    java.lang.reflect.Method.invoke():498
    org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
    org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
    org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
    java.security.AccessController.doPrivileged():-2
    javax.security.auth.Subject.doAs():422
    org.apache.hadoop.security.UserGroupInformation.doAs():1807
    org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
    com.sun.proxy.$Proxy42.executeStatementAsync():-1
    org.apache.hive.service.cli.CLIService.executeStatementAsync():310
    org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
    org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
    org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
    org.apache.thrift.ProcessFunction.process():39
    org.apache.thrift.TBaseProcessor.process():39
    org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
    org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
    java.util.concurrent.ThreadPoolExecutor.runWorker():1149
    java.util.concurrent.ThreadPoolExecutor$Worker.run():624
    java.lang.Thread.run():748
  Caused By (java.lang.RuntimeException) org.apache.hadoop.hive.ql.parse.ParseException:line 1:23 missing EOF at '$' near 'EXPR'
    org.apache.hadoop.hive.ql.parse.ParseDriver.parse():215
    org.apache.hadoop.hive.ql.parse.ParseUtils.parse():77
    org.apache.hadoop.hive.ql.parse.ParseUtils.parse():70
    org.apache.hadoop.hive.ql.Driver.compile():468
    org.apache.hadoop.hive.ql.Driver.compileInternal():1316
    org.apache.hadoop.hive.ql.Driver.compileAndRespond():1294
    org.apache.hive.service.cli.operation.SQLOperation.prepare():204
    org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
    org.apache.hive.service.cli.operation.Operation.run():320
    org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
    org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
    sun.reflect.GeneratedMethodAccessor86.invoke():-1
    sun.reflect.DelegatingMethodAccessorImpl.invoke():43
    java.lang.reflect.Method.invoke():498
    org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
    org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
    org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
    java.security.AccessController.doPrivileged():-2
    javax.security.auth.Subject.doAs():422
    org.apache.hadoop.security.UserGroupInformation.doAs():1807
    org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
    com.sun.proxy.$Proxy42.executeStatementAsync():-1
    org.apache.hive.service.cli.CLIService.executeStatementAsync():310
    org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
    org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
    org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
    org.apache.thrift.ProcessFunction.process():39
    org.apache.thrift.TBaseProcessor.process():39
    org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
    org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
    java.util.concurrent.ThreadPoolExecutor.runWorker():1149
    java.util.concurrent.ThreadPoolExecutor$Worker.run():624
    java.lang.Thread.run():748 (state=,code=0)
0: jdbc:drill:zk=local>


{
    "id": {
        "part1": 2738948259170077000,
        "part2": -61360705522985576
    },
    "type": 1,
    "start": 1509772561834,
    "end": 1509772562065,
    "query": "SELECT count(*) FROM hive2.`contact`",
    "plan": "00-00    Screen : rowType = RecordType(BIGINT EXPR$0): rowcount = 1.0, cumulative cost = {1.1 rows, 1.1 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 882\n00-01      Project(EXPR$0=[$0]) : rowType = RecordType(BIGINT EXPR$0): rowcount = 1.0, cumulative cost = {1.0 rows, 1.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 881\n00-02        Jdbc(sql=[SELECT COUNT(*) AS EXPR$0\nFROM (SELECT 0 AS $f0\nFROM.default.contact) AS t]) : rowType = RecordType(BIGINT EXPR$0): rowcount = 1.0, cumulative cost = {1.0 rows, 1.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 846\n",
    "foreman": {
        "address": "ip-myip.mydomain.orghere.com",
        "userPort": 31010,
        "controlPort": 31011,
        "dataPort": 31012,
        "version": "1.11.0"
    },
    "state": 4,
    "totalFragments": 1,
    "finishedFragments": 0,
    "fragmentProfile": [
        {
            "majorFragmentId": 0,
            "minorFragmentProfile": [
                {
                    "state": 0,
                    "minorFragmentId": 0,
                    "endpoint": {
                        "address": "ip-myip.mydomain.orghere.com",
                        "userPort": 31010,
                        "controlPort": 31011,
                        "dataPort": 31012,
                        "version": "1.11.0"
                    },
                    "lastUpdate": 1509772562054,
                    "lastProgress": 1509772562054
                }
            ]
        }
    ],
    "user": "anonymous",
    "error": "DATA_READ ERROR: Drill Remote Exception\n\n",
    "verboseError": "DATA_READ ERROR: Drill Remote Exception\n\n\n\n",
    "errorId": "fd38ddaa-53d3-46cb-b8e1-4bc8fef458d4",
    "errorNode": "ip-myip.mydomain.orghere.com:31010",
    "optionsJson": "[ {\n  \"kind\" : \"BOOLEAN\",\n  \"type\" : \"SESSION\",\n  \"name\" : \"exec.errors.verbose\",\n  \"bool_val\" : true\n} ]",
    "planEnd": 1509772562051,
    "queueWaitEnd": 0
}



0: jdbc:drill:zk=local> SELECT gender FROM hive2.`contact` limit 1;
Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.

sql SELECT *
FROM.default.contact
plugin hive2
Fragment 0:0


To prove that it can connect, here is the query plan, showing all columns from the table:


00-00    Screen : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {103.1 rows, 109.1 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 354
00-01      Project(gender=[$0]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {103.0 rows, 109.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 353
00-02        SelectionVectorRemover : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {103.0 rows, 109.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 352
00-03          Limit(fetch=[1]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {102.0 rows, 108.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 351
00-04            Limit(fetch=[1]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {101.0 rows, 104.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 350
00-05              Project(gender=[$9]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 100.0, cumulative cost = {100.0 rows, 100.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 349
00-06                Jdbc(sql=[SELECT *
FROM.default.contact]) : rowType = RecordType(VARCHAR(50) contact_id, VARCHAR(50) first_name, VARCHAR(50) last_name, VARCHAR(100) address_line_1, VARCHAR(100) address_line_2, VARCHAR(50) city, VARCHAR(10) postcode, VARCHAR(100) email, DATE dob, VARCHAR(1) gender, VARCHAR(1) marital_status, VARCHAR(20) tfn): rowcount = 100.0, cumulative cost = {100.0 rows, 100.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 282


Below connection with beeline proves there is data in the table:

Beeline version 2.3.0 by Apache Hive
0: jdbc:hive2://localhost:10000/default> show create table contact;
+----------------------------------------------------+
|                   createtab_stmt                   |
+----------------------------------------------------+
| CREATE TABLE `contact`(                            |
|   `contact_id` varchar(50),                        |
|   `first_name` varchar(50),                        |
|   `last_name` varchar(50),                         |
|   `address_line_1` varchar(100),                   |
|   `address_line_2` varchar(100),                   |
|   `city` varchar(50),                              |
|   `postcode` varchar(10),                          |
|   `email` varchar(100),                            |
|   `dob` date,                                      |
|   `gender` varchar(1),                             |
|   `marital_status` varchar(1),                     |
|   `tfn` varchar(20))                               |
| ROW FORMAT SERDE                                   |
|   'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'  |
| STORED AS INPUTFORMAT                              |
|   'org.apache.hadoop.mapred.TextInputFormat'       |
| OUTPUTFORMAT                                       |
|   'org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' |
| LOCATION                                           |
|   'file:/home/ec2-user/warehouse/contact'          |
| TBLPROPERTIES (                                    |
|   'transient_lastDdlTime'='1509487563')            |
+----------------------------------------------------+
23 rows selected (0.361 seconds)
0: jdbc:hive2://localhost:10000/default> select * from contact;
+---------------------+---------------------+--------------------+-------------------------+-------------------------+---------------+-------------------+--------------------+--------------+-----------------+-------------------------+--------------+
| contact.contact_id  | contact.first_name  | contact.last_name  | contact.address_line_1  | contact.address_line_2  | contact.city  | contact.postcode  |   contact.email    | contact.dob  | contact.gender  | contact.marital_status  | contact.tfn  |
+---------------------+---------------------+--------------------+-------------------------+-------------------------+---------------+-------------------+--------------------+--------------+-----------------+-------------------------+--------------+
| C1                  | Bob                 | B                  | St                      |                         | Ta            | 2100              | bob@b.com          | 2020-03-01   | M               | M                       | 12           |
+---------------------+---------------------+--------------------+-------------------------+-------------------------+---------------+-------------------+--------------------+--------------+-----------------+-------------------------+--------------+
1 rows selected (0.156 seconds)
0: jdbc:hive2://localhost:10000/default>

{code}


  was:
Basic hive queries all throw error!

{code:sql}

copied https://repo1.maven.org/maven2/org/apache/hive/hive-jdbc/2.3.0/hive-jdbc-2.3.0-standalone.jar to /usr/lib/apache-drill-1.11.0/jars/3rdparty/hive-jdbc-2.3.0-standalone.jar

added this storage plugin:

{
  "type": "jdbc",
  "driver": "org.apache.hive.jdbc.HiveDriver",
  "url": "jdbc:hive2://host:10000/default",
  "username": "hive",
  "password": "hive1234",
  "enabled": true
}

[ec2-user@host ~]$ cd /usr/lib/apache-drill-1.11.0
[ec2-user@host apache-drill-1.11.0]$ ./bin/drill-embedded
OpenJDK 64-Bit Server VM warning: ignoring option MaxPermSize=512M; support was removed in 8.0
Nov 01, 2017 7:53:53 AM org.glassfish.jersey.server.ApplicationHandler initialize
INFO: Initiating Jersey application, version Jersey: 2.8 2014-04-29 01:25:26...
apache drill 1.11.0
"this isn't your grandfather's sql"
0: jdbc:drill:zk=local> SELECT count(*) FROM hive2.`contact`;
Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.

sql SELECT COUNT(*) AS EXPR$0
FROM (SELECT 0 AS $f0
FROM.default.contact) AS t
plugin hive2
Fragment 0:0

[Error Id: 4b293e97-7547-49c5-91da-b9ee2f2184fc on ip-myip.mydomain.orghere.com:31010] (state=,code=0)




0: jdbc:drill:zk=local> ALTER SESSION SET `exec.errors.verbose` = true;
+-------+-------------------------------+
|  ok   |            summary            |
+-------+-------------------------------+
| true  | exec.errors.verbose updated.  |
+-------+-------------------------------+
1 row selected (0.351 seconds)
0: jdbc:drill:zk=local> SELECT count(*) FROM hive2.`contact`;
Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.

sql SELECT COUNT(*) AS EXPR$0
FROM (SELECT 0 AS $f0
FROM.default.contact) AS t
plugin hive2
Fragment 0:0

[Error Id: ac5cc8f2-69a4-455e-b1a6-5d22cde99729 on ip-myip.mydomain.orghere.com:31010]

  (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 1:23 missing EOF at '$' near 'EXPR'
    org.apache.hive.jdbc.Utils.verifySuccess():267
    org.apache.hive.jdbc.Utils.verifySuccessWithInfo():253
    org.apache.hive.jdbc.HiveStatement.runAsyncOnServer():313
    org.apache.hive.jdbc.HiveStatement.execute():253
    org.apache.hive.jdbc.HiveStatement.executeQuery():476
    org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
    org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
    org.apache.drill.exec.store.jdbc.JdbcRecordReader.setup():177
    org.apache.drill.exec.physical.impl.ScanBatch.<init>():104
    org.apache.drill.exec.physical.impl.ScanBatch.<init>():126
    org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():40
    org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():33
    org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():156
    org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
    org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():136
    org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
    org.apache.drill.exec.physical.impl.ImplCreator.getRootExec():109
    org.apache.drill.exec.physical.impl.ImplCreator.getExec():87
    org.apache.drill.exec.work.fragment.FragmentExecutor.run():207
    org.apache.drill.common.SelfCleaningRunnable.run():38
    java.util.concurrent.ThreadPoolExecutor.runWorker():1149
    java.util.concurrent.ThreadPoolExecutor$Worker.run():624
    java.lang.Thread.run():748
  Caused By (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 1:23 missing EOF at '$' near 'EXPR'
    org.apache.hive.service.cli.operation.Operation.toSQLException():380
    org.apache.hive.service.cli.operation.SQLOperation.prepare():206
    org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
    org.apache.hive.service.cli.operation.Operation.run():320
    org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
    org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
    sun.reflect.GeneratedMethodAccessor86.invoke():-1
    sun.reflect.DelegatingMethodAccessorImpl.invoke():43
    java.lang.reflect.Method.invoke():498
    org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
    org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
    org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
    java.security.AccessController.doPrivileged():-2
    javax.security.auth.Subject.doAs():422
    org.apache.hadoop.security.UserGroupInformation.doAs():1807
    org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
    com.sun.proxy.$Proxy42.executeStatementAsync():-1
    org.apache.hive.service.cli.CLIService.executeStatementAsync():310
    org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
    org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
    org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
    org.apache.thrift.ProcessFunction.process():39
    org.apache.thrift.TBaseProcessor.process():39
    org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
    org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
    java.util.concurrent.ThreadPoolExecutor.runWorker():1149
    java.util.concurrent.ThreadPoolExecutor$Worker.run():624
    java.lang.Thread.run():748
  Caused By (java.lang.RuntimeException) org.apache.hadoop.hive.ql.parse.ParseException:line 1:23 missing EOF at '$' near 'EXPR'
    org.apache.hadoop.hive.ql.parse.ParseDriver.parse():215
    org.apache.hadoop.hive.ql.parse.ParseUtils.parse():77
    org.apache.hadoop.hive.ql.parse.ParseUtils.parse():70
    org.apache.hadoop.hive.ql.Driver.compile():468
    org.apache.hadoop.hive.ql.Driver.compileInternal():1316
    org.apache.hadoop.hive.ql.Driver.compileAndRespond():1294
    org.apache.hive.service.cli.operation.SQLOperation.prepare():204
    org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
    org.apache.hive.service.cli.operation.Operation.run():320
    org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
    org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
    sun.reflect.GeneratedMethodAccessor86.invoke():-1
    sun.reflect.DelegatingMethodAccessorImpl.invoke():43
    java.lang.reflect.Method.invoke():498
    org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
    org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
    org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
    java.security.AccessController.doPrivileged():-2
    javax.security.auth.Subject.doAs():422
    org.apache.hadoop.security.UserGroupInformation.doAs():1807
    org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
    com.sun.proxy.$Proxy42.executeStatementAsync():-1
    org.apache.hive.service.cli.CLIService.executeStatementAsync():310
    org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
    org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
    org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
    org.apache.thrift.ProcessFunction.process():39
    org.apache.thrift.TBaseProcessor.process():39
    org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
    org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
    java.util.concurrent.ThreadPoolExecutor.runWorker():1149
    java.util.concurrent.ThreadPoolExecutor$Worker.run():624
    java.lang.Thread.run():748 (state=,code=0)
0: jdbc:drill:zk=local>


{
    "id": {
        "part1": 2738948259170077000,
        "part2": -61360705522985576
    },
    "type": 1,
    "start": 1509772561834,
    "end": 1509772562065,
    "query": "SELECT count(*) FROM hive2.`contact`",
    "plan": "00-00    Screen : rowType = RecordType(BIGINT EXPR$0): rowcount = 1.0, cumulative cost = {1.1 rows, 1.1 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 882\n00-01      Project(EXPR$0=[$0]) : rowType = RecordType(BIGINT EXPR$0): rowcount = 1.0, cumulative cost = {1.0 rows, 1.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 881\n00-02        Jdbc(sql=[SELECT COUNT(*) AS EXPR$0\nFROM (SELECT 0 AS $f0\nFROM.default.contact) AS t]) : rowType = RecordType(BIGINT EXPR$0): rowcount = 1.0, cumulative cost = {1.0 rows, 1.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 846\n",
    "foreman": {
        "address": "ip-myip.mydomain.orghere.com",
        "userPort": 31010,
        "controlPort": 31011,
        "dataPort": 31012,
        "version": "1.11.0"
    },
    "state": 4,
    "totalFragments": 1,
    "finishedFragments": 0,
    "fragmentProfile": [
        {
            "majorFragmentId": 0,
            "minorFragmentProfile": [
                {
                    "state": 0,
                    "minorFragmentId": 0,
                    "endpoint": {
                        "address": "ip-myip.mydomain.orghere.com",
                        "userPort": 31010,
                        "controlPort": 31011,
                        "dataPort": 31012,
                        "version": "1.11.0"
                    },
                    "lastUpdate": 1509772562054,
                    "lastProgress": 1509772562054
                }
            ]
        }
    ],
    "user": "anonymous",
    "error": "DATA_READ ERROR: Drill Remote Exception\n\n",
    "verboseError": "DATA_READ ERROR: Drill Remote Exception\n\n\n\n",
    "errorId": "fd38ddaa-53d3-46cb-b8e1-4bc8fef458d4",
    "errorNode": "ip-myip.mydomain.orghere.com:31010",
    "optionsJson": "[ {\n  \"kind\" : \"BOOLEAN\",\n  \"type\" : \"SESSION\",\n  \"name\" : \"exec.errors.verbose\",\n  \"bool_val\" : true\n} ]",
    "planEnd": 1509772562051,
    "queueWaitEnd": 0
}



0: jdbc:drill:zk=local> SELECT gender FROM hive2.`contact` limit 1;
Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.

sql SELECT *
FROM.default.contact
plugin hive2
Fragment 0:0


To prove that it can connect, here is the query plan, showing all columns from the table:


00-00    Screen : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {103.1 rows, 109.1 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 354
00-01      Project(gender=[$0]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {103.0 rows, 109.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 353
00-02        SelectionVectorRemover : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {103.0 rows, 109.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 352
00-03          Limit(fetch=[1]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {102.0 rows, 108.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 351
00-04            Limit(fetch=[1]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {101.0 rows, 104.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 350
00-05              Project(gender=[$9]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 100.0, cumulative cost = {100.0 rows, 100.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 349
00-06                Jdbc(sql=[SELECT *
FROM.default.contact]) : rowType = RecordType(VARCHAR(50) contact_id, VARCHAR(50) first_name, VARCHAR(50) last_name, VARCHAR(100) address_line_1, VARCHAR(100) address_line_2, VARCHAR(50) city, VARCHAR(10) postcode, VARCHAR(100) email, DATE dob, VARCHAR(1) gender, VARCHAR(1) marital_status, VARCHAR(20) tfn): rowcount = 100.0, cumulative cost = {100.0 rows, 100.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 282


Below connection with beeline proves there is data in the table:

Beeline version 2.3.0 by Apache Hive
0: jdbc:hive2://localhost:10000/default> show create table contact;
+----------------------------------------------------+
|                   createtab_stmt                   |
+----------------------------------------------------+
| CREATE TABLE `contact`(                            |
|   `contact_id` varchar(50),                        |
|   `first_name` varchar(50),                        |
|   `last_name` varchar(50),                         |
|   `address_line_1` varchar(100),                   |
|   `address_line_2` varchar(100),                   |
|   `city` varchar(50),                              |
|   `postcode` varchar(10),                          |
|   `email` varchar(100),                            |
|   `dob` date,                                      |
|   `gender` varchar(1),                             |
|   `marital_status` varchar(1),                     |
|   `tfn` varchar(20))                               |
| ROW FORMAT SERDE                                   |
|   'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'  |
| STORED AS INPUTFORMAT                              |
|   'org.apache.hadoop.mapred.TextInputFormat'       |
| OUTPUTFORMAT                                       |
|   'org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' |
| LOCATION                                           |
|   'file:/home/ec2-user/warehouse/contact'          |
| TBLPROPERTIES (                                    |
|   'transient_lastDdlTime'='1509487563')            |
+----------------------------------------------------+
23 rows selected (0.361 seconds)
0: jdbc:hive2://localhost:10000/default> select * from contact;
+---------------------+---------------------+--------------------+-------------------------+-------------------------+---------------+-------------------+--------------------+--------------+-----------------+-------------------------+--------------+
| contact.contact_id  | contact.first_name  | contact.last_name  | contact.address_line_1  | contact.address_line_2  | contact.city  | contact.postcode  |   contact.email    | contact.dob  | contact.gender  | contact.marital_status  | contact.tfn  |
+---------------------+---------------------+--------------------+-------------------------+-------------------------+---------------+-------------------+--------------------+--------------+-----------------+-------------------------+--------------+
| C1                  | Bob                 | B                  | St                      |                         | Ta            | 2100              | bob@b.com          | 2020-03-01   | M               | M                       | 12           |
+---------------------+---------------------+--------------------+-------------------------+-------------------------+---------------+-------------------+--------------------+--------------+-----------------+-------------------------+--------------+
1 rows selected (0.156 seconds)
0: jdbc:hive2://localhost:10000/default>

{code}



> HiveServer2 queries throw error with jdbc connection
> ----------------------------------------------------
>
>                 Key: DRILL-5932
>                 URL: https://issues.apache.org/jira/browse/DRILL-5932
>             Project: Apache Drill
>          Issue Type: Bug
>          Components: Client - JDBC, Storage - Hive
>    Affects Versions: 1.11.0
>         Environment: linux
> 2.3 hive version
>            Reporter: tooptoop4
>            Priority: Blocker
>
> Basic hive queries all throw error!
> {code:sql}
> copied https://repo1.maven.org/maven2/org/apache/hive/hive-jdbc/2.3.0/hive-jdbc-2.3.0-standalone.jar to /usr/lib/apache-drill-1.11.0/jars/3rdparty/hive-jdbc-2.3.0-standalone.jar
> added this storage plugin:
> {
>   "type": "jdbc",
>   "driver": "org.apache.hive.jdbc.HiveDriver",
>   "url": "jdbc:hive2://host:10000/default",
>   "username": "hive",
>   "password": "hive1234",
>   "enabled": true
> }
> [ec2-user@host ~]$ cd /usr/lib/apache-drill-1.11.0
> [ec2-user@host apache-drill-1.11.0]$ ./bin/drill-embedded
> OpenJDK 64-Bit Server VM warning: ignoring option MaxPermSize=512M; support was removed in 8.0
> Nov 01, 2017 7:53:53 AM org.glassfish.jersey.server.ApplicationHandler initialize
> INFO: Initiating Jersey application, version Jersey: 2.8 2014-04-29 01:25:26...
> apache drill 1.11.0
> "this isn't your grandfather's sql"
> 0: jdbc:drill:zk=local> SELECT count(*) FROM hive2.`contact`;
> Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.
> sql SELECT COUNT(*) AS EXPR$0
> FROM (SELECT 0 AS $f0
> FROM.default.contact) AS t
> plugin hive2
> Fragment 0:0
> [Error Id: 4b293e97-7547-49c5-91da-b9ee2f2184fc on ip-myip.mydomain.orghere.com:31010] (state=,code=0)
> 0: jdbc:drill:zk=local> ALTER SESSION SET `exec.errors.verbose` = true;
> +-------+-------------------------------+
> |  ok   |            summary            |
> +-------+-------------------------------+
> | true  | exec.errors.verbose updated.  |
> +-------+-------------------------------+
> 1 row selected (0.351 seconds)
> 0: jdbc:drill:zk=local> select * from hive2.contact;
> Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.
> sql SELECT *
> FROM.default.contact
> plugin hive2
> Fragment 0:0
> [Error Id: fe36b026-e8ff-4354-af6c-6073130680c9 on ip-10-160-153-163.syd.non.c1.macquarie.com:31010]
>   (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 2:4 cannot recognize input near '.' 'default' '.' in join source
>     org.apache.hive.jdbc.Utils.verifySuccess():267
>     org.apache.hive.jdbc.Utils.verifySuccessWithInfo():253
>     org.apache.hive.jdbc.HiveStatement.runAsyncOnServer():313
>     org.apache.hive.jdbc.HiveStatement.execute():253
>     org.apache.hive.jdbc.HiveStatement.executeQuery():476
>     org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
>     org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
>     org.apache.drill.exec.store.jdbc.JdbcRecordReader.setup():177
>     org.apache.drill.exec.physical.impl.ScanBatch.<init>():104
>     org.apache.drill.exec.physical.impl.ScanBatch.<init>():126
>     org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():40
>     org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():33
>     org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():156
>     org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
>     org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():136
>     org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
>     org.apache.drill.exec.physical.impl.ImplCreator.getRootExec():109
>     org.apache.drill.exec.physical.impl.ImplCreator.getExec():87
>     org.apache.drill.exec.work.fragment.FragmentExecutor.run():207
>     org.apache.drill.common.SelfCleaningRunnable.run():38
>     java.util.concurrent.ThreadPoolExecutor.runWorker():1149
>     java.util.concurrent.ThreadPoolExecutor$Worker.run():624
>     java.lang.Thread.run():748
>   Caused By (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 2:4 cannot recognize input near '.' 'default' '.' in join source
>     org.apache.hive.service.cli.operation.Operation.toSQLException():380
>     org.apache.hive.service.cli.operation.SQLOperation.prepare():206
>     org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
>     org.apache.hive.service.cli.operation.Operation.run():320
>     org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
>     org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
>     sun.reflect.GeneratedMethodAccessor86.invoke():-1
>     sun.reflect.DelegatingMethodAccessorImpl.invoke():43
>     java.lang.reflect.Method.invoke():498
>     org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
>     org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
>     org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
>     java.security.AccessController.doPrivileged():-2
>     javax.security.auth.Subject.doAs():422
>     org.apache.hadoop.security.UserGroupInformation.doAs():1807
>     org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
>     com.sun.proxy.$Proxy42.executeStatementAsync():-1
>     org.apache.hive.service.cli.CLIService.executeStatementAsync():310
>     org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
>     org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
>     org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
>     org.apache.thrift.ProcessFunction.process():39
>     org.apache.thrift.TBaseProcessor.process():39
>     org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
>     org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
>     java.util.concurrent.ThreadPoolExecutor.runWorker():1149
>     java.util.concurrent.ThreadPoolExecutor$Worker.run():624
>     java.lang.Thread.run():748
>   Caused By (java.lang.RuntimeException) org.apache.hadoop.hive.ql.parse.ParseException:line 2:4 cannot recognize input near '.' 'default' '.' in join source
>     org.apache.hadoop.hive.ql.parse.ParseDriver.parse():207
>     org.apache.hadoop.hive.ql.parse.ParseUtils.parse():77
>     org.apache.hadoop.hive.ql.parse.ParseUtils.parse():70
>     org.apache.hadoop.hive.ql.Driver.compile():468
>     org.apache.hadoop.hive.ql.Driver.compileInternal():1316
>     org.apache.hadoop.hive.ql.Driver.compileAndRespond():1294
>     org.apache.hive.service.cli.operation.SQLOperation.prepare():204
>     org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
>     org.apache.hive.service.cli.operation.Operation.run():320
>     org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
>     org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
>     sun.reflect.GeneratedMethodAccessor86.invoke():-1
>     sun.reflect.DelegatingMethodAccessorImpl.invoke():43
>     java.lang.reflect.Method.invoke():498
>     org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
>     org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
>     org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
>     java.security.AccessController.doPrivileged():-2
>     javax.security.auth.Subject.doAs():422
>     org.apache.hadoop.security.UserGroupInformation.doAs():1807
>     org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
>     com.sun.proxy.$Proxy42.executeStatementAsync():-1
>     org.apache.hive.service.cli.CLIService.executeStatementAsync():310
>     org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
>     org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
>     org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
>     org.apache.thrift.ProcessFunction.process():39
>     org.apache.thrift.TBaseProcessor.process():39
>     org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
>     org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
>     java.util.concurrent.ThreadPoolExecutor.runWorker():1149
>     java.util.concurrent.ThreadPoolExecutor$Worker.run():624
>     java.lang.Thread.run():748 (state=,code=0)
> 0: jdbc:drill:zk=local> select * from hive2.`default`.contact;
> Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.
> sql SELECT *
> FROM.default.contact
> plugin hive2
> Fragment 0:0
> [Error Id: 7b7aae92-fabd-49dd-ab84-15ca00189882 on ip-10-160-153-163.syd.non.c1.macquarie.com:31010]
>   (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 2:4 cannot recognize input near '.' 'default' '.' in join source
>     org.apache.hive.jdbc.Utils.verifySuccess():267
>     org.apache.hive.jdbc.Utils.verifySuccessWithInfo():253
>     org.apache.hive.jdbc.HiveStatement.runAsyncOnServer():313
>     org.apache.hive.jdbc.HiveStatement.execute():253
>     org.apache.hive.jdbc.HiveStatement.executeQuery():476
>     org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
>     org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
>     org.apache.drill.exec.store.jdbc.JdbcRecordReader.setup():177
>     org.apache.drill.exec.physical.impl.ScanBatch.<init>():104
>     org.apache.drill.exec.physical.impl.ScanBatch.<init>():126
>     org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():40
>     org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():33
>     org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():156
>     org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
>     org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():136
>     org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
>     org.apache.drill.exec.physical.impl.ImplCreator.getRootExec():109
>     org.apache.drill.exec.physical.impl.ImplCreator.getExec():87
>     org.apache.drill.exec.work.fragment.FragmentExecutor.run():207
>     org.apache.drill.common.SelfCleaningRunnable.run():38
>     java.util.concurrent.ThreadPoolExecutor.runWorker():1149
>     java.util.concurrent.ThreadPoolExecutor$Worker.run():624
>     java.lang.Thread.run():748
>   Caused By (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 2:4 cannot recognize input near '.' 'default' '.' in join source
>     org.apache.hive.service.cli.operation.Operation.toSQLException():380
>     org.apache.hive.service.cli.operation.SQLOperation.prepare():206
>     org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
>     org.apache.hive.service.cli.operation.Operation.run():320
>     org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
>     org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
>     sun.reflect.GeneratedMethodAccessor86.invoke():-1
>     sun.reflect.DelegatingMethodAccessorImpl.invoke():43
>     java.lang.reflect.Method.invoke():498
>     org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
>     org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
>     org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
>     java.security.AccessController.doPrivileged():-2
>     javax.security.auth.Subject.doAs():422
>     org.apache.hadoop.security.UserGroupInformation.doAs():1807
>     org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
>     com.sun.proxy.$Proxy42.executeStatementAsync():-1
>     org.apache.hive.service.cli.CLIService.executeStatementAsync():310
>     org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
>     org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
>     org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
>     org.apache.thrift.ProcessFunction.process():39
>     org.apache.thrift.TBaseProcessor.process():39
>     org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
>     org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
>     java.util.concurrent.ThreadPoolExecutor.runWorker():1149
>     java.util.concurrent.ThreadPoolExecutor$Worker.run():624
>     java.lang.Thread.run():748
>   Caused By (java.lang.RuntimeException) org.apache.hadoop.hive.ql.parse.ParseException:line 2:4 cannot recognize input near '.' 'default' '.' in join source
>     org.apache.hadoop.hive.ql.parse.ParseDriver.parse():207
>     org.apache.hadoop.hive.ql.parse.ParseUtils.parse():77
>     org.apache.hadoop.hive.ql.parse.ParseUtils.parse():70
>     org.apache.hadoop.hive.ql.Driver.compile():468
>     org.apache.hadoop.hive.ql.Driver.compileInternal():1316
>     org.apache.hadoop.hive.ql.Driver.compileAndRespond():1294
>     org.apache.hive.service.cli.operation.SQLOperation.prepare():204
>     org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
>     org.apache.hive.service.cli.operation.Operation.run():320
>     org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
>     org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
>     sun.reflect.GeneratedMethodAccessor86.invoke():-1
>     sun.reflect.DelegatingMethodAccessorImpl.invoke():43
>     java.lang.reflect.Method.invoke():498
>     org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
>     org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
>     org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
>     java.security.AccessController.doPrivileged():-2
>     javax.security.auth.Subject.doAs():422
>     org.apache.hadoop.security.UserGroupInformation.doAs():1807
>     org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
>     com.sun.proxy.$Proxy42.executeStatementAsync():-1
>     org.apache.hive.service.cli.CLIService.executeStatementAsync():310
>     org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
>     org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
>     org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
>     org.apache.thrift.ProcessFunction.process():39
>     org.apache.thrift.TBaseProcessor.process():39
>     org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
>     org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
>     java.util.concurrent.ThreadPoolExecutor.runWorker():1149
>     java.util.concurrent.ThreadPoolExecutor$Worker.run():624
>     java.lang.Thread.run():748 (state=,code=0)
> 0: jdbc:drill:zk=local>
> 0: jdbc:drill:zk=local> SELECT count(*) FROM hive2.`contact`;
> Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.
> sql SELECT COUNT(*) AS EXPR$0
> FROM (SELECT 0 AS $f0
> FROM.default.contact) AS t
> plugin hive2
> Fragment 0:0
> [Error Id: ac5cc8f2-69a4-455e-b1a6-5d22cde99729 on ip-myip.mydomain.orghere.com:31010]
>   (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 1:23 missing EOF at '$' near 'EXPR'
>     org.apache.hive.jdbc.Utils.verifySuccess():267
>     org.apache.hive.jdbc.Utils.verifySuccessWithInfo():253
>     org.apache.hive.jdbc.HiveStatement.runAsyncOnServer():313
>     org.apache.hive.jdbc.HiveStatement.execute():253
>     org.apache.hive.jdbc.HiveStatement.executeQuery():476
>     org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
>     org.apache.commons.dbcp.DelegatingStatement.executeQuery():208
>     org.apache.drill.exec.store.jdbc.JdbcRecordReader.setup():177
>     org.apache.drill.exec.physical.impl.ScanBatch.<init>():104
>     org.apache.drill.exec.physical.impl.ScanBatch.<init>():126
>     org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():40
>     org.apache.drill.exec.store.jdbc.JdbcBatchCreator.getBatch():33
>     org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():156
>     org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
>     org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch():136
>     org.apache.drill.exec.physical.impl.ImplCreator.getChildren():179
>     org.apache.drill.exec.physical.impl.ImplCreator.getRootExec():109
>     org.apache.drill.exec.physical.impl.ImplCreator.getExec():87
>     org.apache.drill.exec.work.fragment.FragmentExecutor.run():207
>     org.apache.drill.common.SelfCleaningRunnable.run():38
>     java.util.concurrent.ThreadPoolExecutor.runWorker():1149
>     java.util.concurrent.ThreadPoolExecutor$Worker.run():624
>     java.lang.Thread.run():748
>   Caused By (org.apache.hive.service.cli.HiveSQLException) Error while compiling statement: FAILED: ParseException line 1:23 missing EOF at '$' near 'EXPR'
>     org.apache.hive.service.cli.operation.Operation.toSQLException():380
>     org.apache.hive.service.cli.operation.SQLOperation.prepare():206
>     org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
>     org.apache.hive.service.cli.operation.Operation.run():320
>     org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
>     org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
>     sun.reflect.GeneratedMethodAccessor86.invoke():-1
>     sun.reflect.DelegatingMethodAccessorImpl.invoke():43
>     java.lang.reflect.Method.invoke():498
>     org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
>     org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
>     org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
>     java.security.AccessController.doPrivileged():-2
>     javax.security.auth.Subject.doAs():422
>     org.apache.hadoop.security.UserGroupInformation.doAs():1807
>     org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
>     com.sun.proxy.$Proxy42.executeStatementAsync():-1
>     org.apache.hive.service.cli.CLIService.executeStatementAsync():310
>     org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
>     org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
>     org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
>     org.apache.thrift.ProcessFunction.process():39
>     org.apache.thrift.TBaseProcessor.process():39
>     org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
>     org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
>     java.util.concurrent.ThreadPoolExecutor.runWorker():1149
>     java.util.concurrent.ThreadPoolExecutor$Worker.run():624
>     java.lang.Thread.run():748
>   Caused By (java.lang.RuntimeException) org.apache.hadoop.hive.ql.parse.ParseException:line 1:23 missing EOF at '$' near 'EXPR'
>     org.apache.hadoop.hive.ql.parse.ParseDriver.parse():215
>     org.apache.hadoop.hive.ql.parse.ParseUtils.parse():77
>     org.apache.hadoop.hive.ql.parse.ParseUtils.parse():70
>     org.apache.hadoop.hive.ql.Driver.compile():468
>     org.apache.hadoop.hive.ql.Driver.compileInternal():1316
>     org.apache.hadoop.hive.ql.Driver.compileAndRespond():1294
>     org.apache.hive.service.cli.operation.SQLOperation.prepare():204
>     org.apache.hive.service.cli.operation.SQLOperation.runInternal():290
>     org.apache.hive.service.cli.operation.Operation.run():320
>     org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal():530
>     org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync():517
>     sun.reflect.GeneratedMethodAccessor86.invoke():-1
>     sun.reflect.DelegatingMethodAccessorImpl.invoke():43
>     java.lang.reflect.Method.invoke():498
>     org.apache.hive.service.cli.session.HiveSessionProxy.invoke():78
>     org.apache.hive.service.cli.session.HiveSessionProxy.access$000():36
>     org.apache.hive.service.cli.session.HiveSessionProxy$1.run():63
>     java.security.AccessController.doPrivileged():-2
>     javax.security.auth.Subject.doAs():422
>     org.apache.hadoop.security.UserGroupInformation.doAs():1807
>     org.apache.hive.service.cli.session.HiveSessionProxy.invoke():59
>     com.sun.proxy.$Proxy42.executeStatementAsync():-1
>     org.apache.hive.service.cli.CLIService.executeStatementAsync():310
>     org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement():530
>     org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1437
>     org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult():1422
>     org.apache.thrift.ProcessFunction.process():39
>     org.apache.thrift.TBaseProcessor.process():39
>     org.apache.hive.service.auth.TSetIpAddressProcessor.process():56
>     org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run():286
>     java.util.concurrent.ThreadPoolExecutor.runWorker():1149
>     java.util.concurrent.ThreadPoolExecutor$Worker.run():624
>     java.lang.Thread.run():748 (state=,code=0)
> 0: jdbc:drill:zk=local>
> {
>     "id": {
>         "part1": 2738948259170077000,
>         "part2": -61360705522985576
>     },
>     "type": 1,
>     "start": 1509772561834,
>     "end": 1509772562065,
>     "query": "SELECT count(*) FROM hive2.`contact`",
>     "plan": "00-00    Screen : rowType = RecordType(BIGINT EXPR$0): rowcount = 1.0, cumulative cost = {1.1 rows, 1.1 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 882\n00-01      Project(EXPR$0=[$0]) : rowType = RecordType(BIGINT EXPR$0): rowcount = 1.0, cumulative cost = {1.0 rows, 1.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 881\n00-02        Jdbc(sql=[SELECT COUNT(*) AS EXPR$0\nFROM (SELECT 0 AS $f0\nFROM.default.contact) AS t]) : rowType = RecordType(BIGINT EXPR$0): rowcount = 1.0, cumulative cost = {1.0 rows, 1.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 846\n",
>     "foreman": {
>         "address": "ip-myip.mydomain.orghere.com",
>         "userPort": 31010,
>         "controlPort": 31011,
>         "dataPort": 31012,
>         "version": "1.11.0"
>     },
>     "state": 4,
>     "totalFragments": 1,
>     "finishedFragments": 0,
>     "fragmentProfile": [
>         {
>             "majorFragmentId": 0,
>             "minorFragmentProfile": [
>                 {
>                     "state": 0,
>                     "minorFragmentId": 0,
>                     "endpoint": {
>                         "address": "ip-myip.mydomain.orghere.com",
>                         "userPort": 31010,
>                         "controlPort": 31011,
>                         "dataPort": 31012,
>                         "version": "1.11.0"
>                     },
>                     "lastUpdate": 1509772562054,
>                     "lastProgress": 1509772562054
>                 }
>             ]
>         }
>     ],
>     "user": "anonymous",
>     "error": "DATA_READ ERROR: Drill Remote Exception\n\n",
>     "verboseError": "DATA_READ ERROR: Drill Remote Exception\n\n\n\n",
>     "errorId": "fd38ddaa-53d3-46cb-b8e1-4bc8fef458d4",
>     "errorNode": "ip-myip.mydomain.orghere.com:31010",
>     "optionsJson": "[ {\n  \"kind\" : \"BOOLEAN\",\n  \"type\" : \"SESSION\",\n  \"name\" : \"exec.errors.verbose\",\n  \"bool_val\" : true\n} ]",
>     "planEnd": 1509772562051,
>     "queueWaitEnd": 0
> }
> 0: jdbc:drill:zk=local> SELECT gender FROM hive2.`contact` limit 1;
> Error: DATA_READ ERROR: The JDBC storage plugin failed while trying setup the SQL query.
> sql SELECT *
> FROM.default.contact
> plugin hive2
> Fragment 0:0
> To prove that it can connect, here is the query plan, showing all columns from the table:
> 00-00    Screen : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {103.1 rows, 109.1 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 354
> 00-01      Project(gender=[$0]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {103.0 rows, 109.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 353
> 00-02        SelectionVectorRemover : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {103.0 rows, 109.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 352
> 00-03          Limit(fetch=[1]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {102.0 rows, 108.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 351
> 00-04            Limit(fetch=[1]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 1.0, cumulative cost = {101.0 rows, 104.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 350
> 00-05              Project(gender=[$9]) : rowType = RecordType(VARCHAR(1) gender): rowcount = 100.0, cumulative cost = {100.0 rows, 100.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 349
> 00-06                Jdbc(sql=[SELECT *
> FROM.default.contact]) : rowType = RecordType(VARCHAR(50) contact_id, VARCHAR(50) first_name, VARCHAR(50) last_name, VARCHAR(100) address_line_1, VARCHAR(100) address_line_2, VARCHAR(50) city, VARCHAR(10) postcode, VARCHAR(100) email, DATE dob, VARCHAR(1) gender, VARCHAR(1) marital_status, VARCHAR(20) tfn): rowcount = 100.0, cumulative cost = {100.0 rows, 100.0 cpu, 0.0 io, 0.0 network, 0.0 memory}, id = 282
> Below connection with beeline proves there is data in the table:
> Beeline version 2.3.0 by Apache Hive
> 0: jdbc:hive2://localhost:10000/default> show create table contact;
> +----------------------------------------------------+
> |                   createtab_stmt                   |
> +----------------------------------------------------+
> | CREATE TABLE `contact`(                            |
> |   `contact_id` varchar(50),                        |
> |   `first_name` varchar(50),                        |
> |   `last_name` varchar(50),                         |
> |   `address_line_1` varchar(100),                   |
> |   `address_line_2` varchar(100),                   |
> |   `city` varchar(50),                              |
> |   `postcode` varchar(10),                          |
> |   `email` varchar(100),                            |
> |   `dob` date,                                      |
> |   `gender` varchar(1),                             |
> |   `marital_status` varchar(1),                     |
> |   `tfn` varchar(20))                               |
> | ROW FORMAT SERDE                                   |
> |   'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'  |
> | STORED AS INPUTFORMAT                              |
> |   'org.apache.hadoop.mapred.TextInputFormat'       |
> | OUTPUTFORMAT                                       |
> |   'org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' |
> | LOCATION                                           |
> |   'file:/home/ec2-user/warehouse/contact'          |
> | TBLPROPERTIES (                                    |
> |   'transient_lastDdlTime'='1509487563')            |
> +----------------------------------------------------+
> 23 rows selected (0.361 seconds)
> 0: jdbc:hive2://localhost:10000/default> select * from contact;
> +---------------------+---------------------+--------------------+-------------------------+-------------------------+---------------+-------------------+--------------------+--------------+-----------------+-------------------------+--------------+
> | contact.contact_id  | contact.first_name  | contact.last_name  | contact.address_line_1  | contact.address_line_2  | contact.city  | contact.postcode  |   contact.email    | contact.dob  | contact.gender  | contact.marital_status  | contact.tfn  |
> +---------------------+---------------------+--------------------+-------------------------+-------------------------+---------------+-------------------+--------------------+--------------+-----------------+-------------------------+--------------+
> | C1                  | Bob                 | B                  | St                      |                         | Ta            | 2100              | bob@b.com          | 2020-03-01   | M               | M                       | 12           |
> +---------------------+---------------------+--------------------+-------------------------+-------------------------+---------------+-------------------+--------------------+--------------+-----------------+-------------------------+--------------+
> 1 rows selected (0.156 seconds)
> 0: jdbc:hive2://localhost:10000/default>
> {code}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

Mime
View raw message