hive-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Sushanth Sowmyan (JIRA)" <j...@apache.org>
Subject [jira] [Updated] (HIVE-13017) Child process of HiveServer2 fails to get delegation token from non default FileSystem
Date Wed, 10 Feb 2016 17:58:18 GMT

     [ https://issues.apache.org/jira/browse/HIVE-13017?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]

Sushanth Sowmyan updated HIVE-13017:
------------------------------------
    Attachment: HIVE-13017.patch

Patch attached.

> Child process of HiveServer2 fails to get delegation token from non default FileSystem
> --------------------------------------------------------------------------------------
>
>                 Key: HIVE-13017
>                 URL: https://issues.apache.org/jira/browse/HIVE-13017
>             Project: Hive
>          Issue Type: Bug
>          Components: Authentication
>    Affects Versions: 1.2.1
>         Environment: Secure 
>            Reporter: Takahiko Saito
>            Assignee: Sushanth Sowmyan
>         Attachments: HIVE-13017.patch
>
>
> The following query fails, when Azure Filesystem is used as default file system, and
HDFS is used for intermediate data.
> {noformat}
> >>>  create temporary table s10k stored as orc as select * from studenttab10k;
> >>>  create temporary table v10k as select * from votertab10k;
> >>>  select registration 
> from s10k s join v10k v 
> on (s.name = v.name) join studentparttab30k p 
> on (p.name = v.name) 
> where s.age < 25 and v.age < 25 and p.age < 25;
> ERROR : Execution failed with exit status: 2
> ERROR : Obtaining error information
> ERROR : 
> Task failed!
> Task ID:
>   Stage-5
> Logs:
> ERROR : /var/log/hive/hiveServer2.log
> Error: Error while processing statement: FAILED: Execution Error, return code 2 from
org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask (state=08S01,code=2)
> Aborting command set because "force" is false and command failed: "select registration

> from s10k s join v10k v 
> on (s.name = v.name) join studentparttab30k p 
> on (p.name = v.name) 
> where s.age < 25 and v.age < 25 and p.age < 25;"
> Closing: 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181/;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=hiveserver2;principal=hive/_HOST@HDINSIGHT.NET;transportMode=http;httpPath=cliservice
> hiveServer2.log shows:
> 2016-02-02 18:04:34,182 INFO  [HiveServer2-Background-Pool: Thread-517]: log.PerfLogger
(PerfLogger.java:PerfLogBegin(135)) - <PERFLOG method=Driver.run from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:34,199 INFO  [HiveServer2-Background-Pool: Thread-517]: log.PerfLogger
(PerfLogger.java:PerfLogBegin(135)) - <PERFLOG method=TimeToSubmit from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:34,212 INFO  [HiveServer2-HttpHandler-Pool: Thread-55]: thrift.ThriftHttpServlet
(ThriftHttpServlet.java:doPost(127)) - Could not validate cookie sent, will try to generate
a new cookie
> 2016-02-02 18:04:34,213 INFO  [HiveServer2-Background-Pool: Thread-517]: ql.Driver (Driver.java:checkConcurrency(168))
- Concurrency mode is disabled, not creating a lock manager
> 2016-02-02 18:04:34,219 INFO  [HiveServer2-HttpHandler-Pool: Thread-55]: thrift.ThriftHttpServlet
(ThriftHttpServlet.java:doKerberosAuth(352)) - Failed to authenticate with http/_HOST kerberos
principal, trying with hive/_HOST kerberos principal
> 2016-02-02 18:04:34,219 INFO  [HiveServer2-Background-Pool: Thread-517]: log.PerfLogger
(PerfLogger.java:PerfLogBegin(135)) - <PERFLOG method=Driver.execute from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:34,225 INFO  [HiveServer2-Background-Pool: Thread-517]: ql.Driver (Driver.java:execute(1390))
- Setting caller context to query id hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0
> 2016-02-02 18:04:34,226 INFO  [HiveServer2-Background-Pool: Thread-517]: ql.Driver (Driver.java:execute(1393))
- Starting command(queryId=hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0): select
registration
> from s10k s join v10k v
> on (s.name = v.name) join studentparttab30k p
> on (p.name = v.name)
> where s.age < 25 and v.age < 25 and p.age < 25
> 2016-02-02 18:04:34,228 INFO  [HiveServer2-Background-Pool: Thread-517]: hooks.ATSHook
(ATSHook.java:<init>(90)) - Created ATS Hook
> 2016-02-02 18:04:34,229 INFO  [HiveServer2-Background-Pool: Thread-517]: log.PerfLogger
(PerfLogger.java:PerfLogBegin(135)) - <PERFLOG method=PreHook.org.apache.hadoop.hive.ql.hooks.ATSHook
from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:34,237 INFO  [HiveServer2-HttpHandler-Pool: Thread-55]: thrift.ThriftHttpServlet
(ThriftHttpServlet.java:doPost(169)) - Cookie added for clientUserName hrt_qa
> 2016-02-02 18:04:34,238 INFO  [HiveServer2-Background-Pool: Thread-517]: log.PerfLogger
(PerfLogger.java:PerfLogEnd(162)) - </PERFLOG method=PreHook.org.apache.hadoop.hive.ql.hooks.ATSHook
start=1454436274229 end=1454436274238 duration=9 from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:34,239 INFO  [HiveServer2-Background-Pool: Thread-517]: log.PerfLogger
(PerfLogger.java:PerfLogBegin(135)) - <PERFLOG method=PreHook.org.apache.hadoop.hive.ql.security.authorization.plugin.DisallowTransformHook
from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:34,240 INFO  [HiveServer2-Background-Pool: Thread-517]: log.PerfLogger
(PerfLogger.java:PerfLogEnd(162)) - </PERFLOG method=PreHook.org.apache.hadoop.hive.ql.security.authorization.plugin.DisallowTransformHook
start=1454436274239 end=1454436274240 duration=1 from=org.apache.hadoop.hive.ql.Driver>
> Query ID = hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0
> 2016-02-02 18:04:34,242 INFO  [HiveServer2-Background-Pool: Thread-517]: ql.Driver (SessionState.java:printInfo(923))
- Query ID = hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0
> Total jobs = 1
> 2016-02-02 18:04:34,243 INFO  [HiveServer2-Background-Pool: Thread-517]: ql.Driver (SessionState.java:printInfo(923))
- Total jobs = 1
> 2016-02-02 18:04:34,245 INFO  [HiveServer2-Background-Pool: Thread-517]: log.PerfLogger
(PerfLogger.java:PerfLogEnd(162)) - </PERFLOG method=TimeToSubmit start=1454436274199 end=1454436274245
duration=46 from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:34,246 INFO  [HiveServer2-Background-Pool: Thread-517]: log.PerfLogger
(PerfLogger.java:PerfLogBegin(135)) - <PERFLOG method=runTasks from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:34,247 INFO  [HiveServer2-Background-Pool: Thread-517]: log.PerfLogger
(PerfLogger.java:PerfLogBegin(135)) - <PERFLOG method=task.MAPREDLOCAL.Stage-5 from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:34,258 INFO  [HiveServer2-Background-Pool: Thread-517]: ql.Driver (Driver.java:launchTask(1718))
- Starting task [Stage-5:MAPREDLOCAL] in serial mode
> 2016-02-02 18:04:34,280 INFO  [HiveServer2-Background-Pool: Thread-517]: mr.MapredLocalTask
(MapredLocalTask.java:executeInChildVM(158)) - Generating plan file file:/tmp/hive/916e3dbb-a10d-4888-a063-52fb058ea421/hive_2016-02-02_18-04-29_153_625340340820843828-4/-local-10006/plan.xml
> 2016-02-02 18:04:34,288 INFO  [HiveServer2-Background-Pool: Thread-517]: log.PerfLogger
(PerfLogger.java:PerfLogBegin(135)) - <PERFLOG method=serializePlan from=org.apache.hadoop.hive.ql.exec.Utilities>
> 2016-02-02 18:04:34,289 INFO  [HiveServer2-Background-Pool: Thread-517]: exec.Utilities
(Utilities.java:serializePlan(1028)) - Serializing MapredLocalWork via kryo
> 2016-02-02 18:04:34,290 INFO  [ATS Logger 0]: hooks.ATSHook (ATSHook.java:createPreHookEvent(158))
- Received pre-hook notification for :hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0
> 2016-02-02 18:04:34,358 INFO  [HiveServer2-Background-Pool: Thread-517]: log.PerfLogger
(PerfLogger.java:PerfLogEnd(162)) - </PERFLOG method=serializePlan start=1454436274288
end=1454436274358 duration=70 from=org.apache.hadoop.hive.ql.exec.Utilities>
> 2016-02-02 18:04:34,737 INFO  [HiveServer2-Background-Pool: Thread-517]: mr.MapredLocalTask
(MapredLocalTask.java:executeInChildVM(287)) - Executing: /usr/hdp/2.4.1.0-170/hadoop/bin/hadoop
jar /usr/hdp/2.4.1.0-170/hive/lib/hive-common-1.2.1000.2.4.1.0-170.jar org.apache.hadoop.hive.ql.exec.mr.ExecDriver
-localtask -plan file:/tmp/hive/916e3dbb-a10d-4888-a063-52fb058ea421/hive_2016-02-02_18-04-29_153_625340340820843828-4/-local-10006/plan.xml
  -jobconffile file:/tmp/hive/916e3dbb-a10d-4888-a063-52fb058ea421/hive_2016-02-02_18-04-29_153_625340340820843828-4/-local-10007/jobconf.xml
> WARNING: Use "yarn jar" to launch YARN applications.
> 2016-02-02 18:04:37,450 INFO  [org.apache.ranger.audit.queue.AuditBatchQueue0]: provider.BaseAuditHandler
(BaseAuditHandler.java:logStatus(312)) - Audit Status Log: name=hiveServer2.async.summary.batch.solr,
interval=01:21.012 minutes, events=2, succcessCount=2, totalEvents=4, totalSuccessCount=4
> Execution log at: /tmp/hive/hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0.log
> 2016-02-02 18:04:39,248 INFO  [HiveServer2-HttpHandler-Pool: Thread-55]: thrift.ThriftHttpServlet
(ThriftHttpServlet.java:doPost(127)) - Could not validate cookie sent, will try to generate
a new cookie
> 2016-02-02 18:04:39,254 INFO  [HiveServer2-HttpHandler-Pool: Thread-55]: thrift.ThriftHttpServlet
(ThriftHttpServlet.java:doKerberosAuth(352)) - Failed to authenticate with http/_HOST kerberos
principal, trying with hive/_HOST kerberos principal
> 2016-02-02 18:04:39,261 INFO  [HiveServer2-HttpHandler-Pool: Thread-55]: thrift.ThriftHttpServlet
(ThriftHttpServlet.java:doPost(169)) - Cookie added for clientUserName hrt_qa
> 2016-02-02 18:04:40     Starting to launch local task to process map join;      maximum
memory = 477102080
> Execution failed with exit status: 2
> 2016-02-02 18:04:43,728 ERROR [HiveServer2-Background-Pool: Thread-517]: exec.Task (SessionState.java:printError(932))
- Execution failed with exit status: 2
> Obtaining error information
> 2016-02-02 18:04:43,730 ERROR [HiveServer2-Background-Pool: Thread-517]: exec.Task (SessionState.java:printError(932))
- Obtaining error information
> Task failed!
> Task ID:
>   Stage-5
> Logs:
> 2016-02-02 18:04:43,730 ERROR [HiveServer2-Background-Pool: Thread-517]: exec.Task (SessionState.java:printError(932))
-
> Task failed!
> Task ID:
>   Stage-5
> Logs:
> /var/log/hive/hiveServer2.log
> 2016-02-02 18:04:43,732 ERROR [HiveServer2-Background-Pool: Thread-517]: exec.Task (SessionState.java:printError(932))
- /var/log/hive/hiveServer2.log
> 2016-02-02 18:04:43,732 ERROR [HiveServer2-Background-Pool: Thread-517]: mr.MapredLocalTask
(MapredLocalTask.java:executeInChildVM(307)) - Execution failed with exit status: 2
> 2016-02-02 18:04:43,733 INFO  [HiveServer2-Background-Pool: Thread-517]: hooks.ATSHook
(ATSHook.java:<init>(90)) - Created ATS Hook
> 2016-02-02 18:04:43,734 INFO  [HiveServer2-Background-Pool: Thread-517]: log.PerfLogger
(PerfLogger.java:PerfLogBegin(135)) - <PERFLOG method=FailureHook.org.apache.hadoop.hive.ql.hooks.ATSHook
from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:43,736 INFO  [HiveServer2-Background-Pool: Thread-517]: log.PerfLogger
(PerfLogger.java:PerfLogEnd(162)) - </PERFLOG method=FailureHook.org.apache.hadoop.hive.ql.hooks.ATSHook
start=1454436283734 end=1454436283736 duration=2 from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:43,736 INFO  [ATS Logger 0]: hooks.ATSHook (ATSHook.java:createPostHookEvent(193))
- Received post-hook notification for :hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0
> FAILED: Execution Error, return code 2 from org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask
> 2016-02-02 18:04:43,757 ERROR [HiveServer2-Background-Pool: Thread-517]: ql.Driver (SessionState.java:printError(932))
- FAILED: Execution Error, return code 2 from org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask
> 2016-02-02 18:04:43,758 INFO  [HiveServer2-Background-Pool: Thread-517]: ql.Driver (Driver.java:execute(1621))
- Resetting the caller context to HIVE_SSN_ID:916e3dbb-a10d-4888-a063-52fb058ea421
> 2016-02-02 18:04:43,759 INFO  [HiveServer2-Background-Pool: Thread-517]: log.PerfLogger
(PerfLogger.java:PerfLogEnd(162)) - </PERFLOG method=Driver.execute start=1454436274219
end=1454436283759 duration=9540 from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:43,760 INFO  [HiveServer2-Background-Pool: Thread-517]: log.PerfLogger
(PerfLogger.java:PerfLogBegin(135)) - <PERFLOG method=releaseLocks from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:43,761 INFO  [HiveServer2-Background-Pool: Thread-517]: log.PerfLogger
(PerfLogger.java:PerfLogEnd(162)) - </PERFLOG method=releaseLocks start=1454436283760 end=1454436283761
duration=1 from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:43,766 ERROR [HiveServer2-Background-Pool: Thread-517]: operation.Operation
(SQLOperation.java:run(209)) - Error running hive query:
> org.apache.hive.service.cli.HiveSQLException: Error while processing statement: FAILED:
Execution Error, return code 2 from org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask
>         at org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:315)
>         at org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:156)
>         at org.apache.hive.service.cli.operation.SQLOperation.access$100(SQLOperation.java:71)
>         at org.apache.hive.service.cli.operation.SQLOperation$1$1.run(SQLOperation.java:206)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:415)
>         at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
>         at org.apache.hive.service.cli.operation.SQLOperation$1.run(SQLOperation.java:218)
>         at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
>         at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>         at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>         at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>         at java.lang.Thread.run(Thread.java:745)
> hive Configs can be viewed from http://qelog.hortonworks.com/log/hs21-hbs24-1454382123/artifacts/tmpModifyConfDir_1454394513592/
> Attachments
> Drop files to attach, or browse.
> Add LinkIssue Links
> relates to
> Bug - A problem which impairs or prevents the functions of the product. HIVE-739 webhcat
tests failing in HDInsight secure cluster throwing NullPointerException	 Blocker - Blocks
development and/or testing work, production could not run. RESOLVED
> Activity
> All
> Comments
> Work Log
> History
> Activity
> Ascending order - Click to sort in descending order
> Permalink Edit Delete 
> tsaito Takahiko Saito added a comment - 3 days ago
> The test passes via hive CLI and explain shows:
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> explain select registration
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> from s10k s join v10k v
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> on (s.name = v.name) join studentparttab30k
p
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> on (p.name = v.name)
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> where s.age < 25 and v.age <
25 and p.age < 25;
> +--------------------------------------------------------------------------------------------------------------+--+
> |                                                   Explain                         
                          |
> +--------------------------------------------------------------------------------------------------------------+--+
> | STAGE DEPENDENCIES:                                                               
                          |
> |   Stage-5 is a root stage                                                         
                          |
> |   Stage-4 depends on stages: Stage-5                                              
                          |
> |   Stage-0 depends on stages: Stage-4                                              
                          |
> |                                                                                   
                          |
> | STAGE PLANS:                                                                      
                          |
> |   Stage: Stage-5                                                                  
                          |
> |     Map Reduce Local Work                                                         
                          |
> |       Alias -> Map Local Tables:                                               
                             |
> |         s                                                                         
                          |
> |           Fetch Operator                                                          
                          |
> |             limit: -1                                                             
                          |
> |         v                                                                         
                          |
> |           Fetch Operator                                                          
                          |
> |             limit: -1                                                             
                          |
> |       Alias -> Map Local Operator Tree:                                        
                             |
> |         s                                                                         
                          |
> |           TableScan                                                               
                          |
> |             alias: s                                                              
                          |
> |             filterExpr: (name is not null and (age < 25)) (type: boolean)      
                             |
> |             Statistics: Num rows: 459 Data size: 47777 Basic stats: COMPLETE Column
stats: NONE              |
> |             Filter Operator                                                       
                          |
> |               predicate: (name is not null and (age < 25)) (type: boolean)     
                             |
> |               Statistics: Num rows: 76 Data size: 7910 Basic stats: COMPLETE Column
stats: NONE              |
> |               HashTable Sink Operator                                             
                          |
> |                 keys:                                                             
                          |
> |                   0 name (type: string)                                           
                          |
> |                   1 name (type: string)                                           
                          |
> |                   2 name (type: string)                                           
                          |
> |         v                                                                         
                          |
> |           TableScan                                                               
                          |
> |             alias: v                                                              
                          |
> |             filterExpr: (name is not null and (age < 25)) (type: boolean)      
                             |
> |             Statistics: Num rows: 1653 Data size: 337233 Basic stats: COMPLETE Column
stats: NONE            |
> |             Filter Operator                                                       
                          |
> |               predicate: (name is not null and (age < 25)) (type: boolean)     
                             |
> |               Statistics: Num rows: 275 Data size: 56103 Basic stats: COMPLETE Column
stats: NONE            |
> |               HashTable Sink Operator                                             
                          |
> |                 keys:                                                             
                          |
> |                   0 name (type: string)                                           
                          |
> |                   1 name (type: string)                                           
                          |
> |                   2 name (type: string)                                           
                          |
> |                                                                                   
                          |
> |   Stage: Stage-4                                                                  
                          |
> |     Map Reduce                                                                    
                          |
> |       Map Operator Tree:                                                          
                          |
> |           TableScan                                                               
                          |
> |             alias: p                                                              
                          |
> |             filterExpr: (name is not null and (age < 25)) (type: boolean)      
                             |
> |             Statistics: Num rows: 30000 Data size: 627520 Basic stats: COMPLETE Column
stats: COMPLETE       |
> |             Filter Operator                                                       
                          |
> |               predicate: (name is not null and (age < 25)) (type: boolean)     
                             |
> |               Statistics: Num rows: 10000 Data size: 1010000 Basic stats: COMPLETE
Column stats: COMPLETE    |
> |               Map Join Operator                                                   
                          |
> |                 condition map:                                                    
                          |
> |                      Inner Join 0 to 1                                            
                          |
> |                      Inner Join 1 to 2                                            
                          |
> |                 keys:                                                             
                          |
> |                   0 name (type: string)                                           
                          |
> |                   1 name (type: string)                                           
                          |
> |                   2 name (type: string)                                           
                          |
> |                 outputColumnNames: _col8                                          
                          |
> |                 Statistics: Num rows: 22000 Data size: 2222000 Basic stats: COMPLETE
Column stats: NONE      |
> |                 Select Operator                                                   
                          |
> |                   expressions: _col8 (type: string)                               
                          |
> |                   outputColumnNames: _col0                                        
                          |
> |                   Statistics: Num rows: 22000 Data size: 2222000 Basic stats: COMPLETE
Column stats: NONE    |
> |                   File Output Operator                                            
                          |
> |                     compressed: false                                             
                          |
> |                     Statistics: Num rows: 22000 Data size: 2222000 Basic stats: COMPLETE
Column stats: NONE  |
> |                     table:                                                        
                          |
> |                         input format: org.apache.hadoop.mapred.TextInputFormat    
                          |
> |                         output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
           |
> |                         serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe 
                          |
> |       Local Work:                                                                 
                          |
> |         Map Reduce Local Work                                                     
                          |
> |                                                                                   
                          |
> |   Stage: Stage-0                                                                  
                          |
> |     Fetch Operator                                                                
                          |
> |       limit: -1                                                                   
                          |
> |       Processor Tree:                                                             
                          |
> |         ListSink                                                                  
                          |
> |                                                                                   
                          |
> +--------------------------------------------------------------------------------------------------------------+--+
> 83 rows selected (2.473 seconds)
> Permalink Edit Delete 
> tsaito Takahiko Saito added a comment - 3 days ago
> emptablemisc_8 also fails with the same error:
> >>>  create temporary table temp1 as select * from votertab10k;
> >>>  select * 
> from studenttab10k s 
> where s.name not in 
> (select name from temp1);
> INFO  : Number of reduce tasks determined at compile time: 1
> INFO  : In order to change the average load for a reducer (in bytes):
> INFO  :   set hive.exec.reducers.bytes.per.reducer=<number>
> INFO  : In order to limit the maximum number of reducers:
> INFO  :   set hive.exec.reducers.max=<number>
> INFO  : In order to set a constant number of reducers:
> INFO  :   set mapreduce.job.reduces=<number>
> INFO  : number of splits:1
> INFO  : Submitting tokens for job: job_1454394534358_0164
> INFO  : Kind: HDFS_DELEGATION_TOKEN, Service: 10.0.0.36:8020, Ident: (HDFS_DELEGATION_TOKEN
token 246 for hrt_qa)
> INFO  : The url to track the job: http://hn0-hs21-h.hdinsight.net:8088/proxy/application_1454394534358_0164/
> INFO  : Starting Job = job_1454394534358_0164, Tracking URL = http://hn0-hs21-h.hdinsight.net:8088/proxy/application_1454394534358_0164/
> INFO  : Kill Command = /usr/hdp/2.4.1.0-170/hadoop/bin/hadoop job  -kill job_1454394534358_0164
> INFO  : Hadoop job information for Stage-4: number of mappers: 1; number of reducers:
1
> INFO  : 2016-02-02 10:11:02,367 Stage-4 map = 0%,  reduce = 0%
> INFO  : 2016-02-02 10:11:26,060 Stage-4 map = 100%,  reduce = 0%, Cumulative CPU 5.48
sec
> INFO  : 2016-02-02 10:11:39,024 Stage-4 map = 100%,  reduce = 100%, Cumulative CPU 11.18
sec
> INFO  : MapReduce Total cumulative CPU time: 11 seconds 180 msec
> INFO  : Ended Job = job_1454394534358_0164
> INFO  : Stage-9 is selected by condition resolver.
> INFO  : Stage-1 is filtered out by condition resolver.
> ERROR : Execution failed with exit status: 2
> ERROR : Obtaining error information
> ERROR : 
> Task failed!
> Task ID:
>   Stage-9
> Logs:
> ERROR : /var/log/hive/hiveServer2.log
> INFO  : Number of reduce tasks determined at compile time: 1
> INFO  : In order to change the average load for a reducer (in bytes):
> INFO  :   set hive.exec.reducers.bytes.per.reducer=<number>
> INFO  : In order to limit the maximum number of reducers:
> INFO  :   set hive.exec.reducers.max=<number>
> INFO  : In order to set a constant number of reducers:
> INFO  :   set mapreduce.job.reduces=<number>
> INFO  : number of splits:2
> INFO  : Submitting tokens for job: job_1454394534358_0169
> INFO  : Kind: HDFS_DELEGATION_TOKEN, Service: 10.0.0.36:8020, Ident: (HDFS_DELEGATION_TOKEN
token 252 for hrt_qa)
> INFO  : The url to track the job: http://hn0-hs21-h.hdinsight.net:8088/proxy/application_1454394534358_0169/
> INFO  : Starting Job = job_1454394534358_0169, Tracking URL = http://hn0-hs21-h.hdinsight.net:8088/proxy/application_1454394534358_0169/
> INFO  : Kill Command = /usr/hdp/2.4.1.0-170/hadoop/bin/hadoop job  -kill job_1454394534358_0169
> INFO  : Hadoop job information for Stage-1: number of mappers: 2; number of reducers:
1
> INFO  : 2016-02-02 10:16:38,027 Stage-1 map = 0%,  reduce = 0%
> INFO  : 2016-02-02 10:16:52,498 Stage-1 map = 50%,  reduce = 0%, Cumulative CPU 3.8 sec
> INFO  : 2016-02-02 10:16:53,566 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 9.94
sec
> INFO  : 2016-02-02 10:17:16,202 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 17.04
sec
> INFO  : MapReduce Total cumulative CPU time: 17 seconds 40 msec
> INFO  : Ended Job = job_1454394534358_0169
> ERROR : Execution failed with exit status: 2
> ERROR : Obtaining error information
> ERROR : 
> Task failed!
> Task ID:
>   Stage-8
> Logs:
> ERROR : /var/log/hive/hiveServer2.log
> Error: Error while processing statement: FAILED: Execution Error, return code 2 from
org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask (state=08S01,code=2)
> Aborting command set because "force" is false and command failed: "select * 
> from studenttab10k s 
> where s.name not in 
> (select name from temp1);"
> Closing: 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181/;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=hiveserver2;principal=hive/_HOST@HDINSIGHT.NET;transportMode=http;httpPath=cliservice
> Its app log can be viewed at http://qelog.hortonworks.com/log/hs21-hbs24-1454382123/app-logs/application_1454394534358_0169.log
> Permalink Edit Delete 
> tsaito Takahiko Saito added a comment - 3 days ago
> The query works without 'temporary table':
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> select registration from studenttab10k
s join votertab10k v
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> on (s.name = v.name) join studentparttab30k
p
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> on (p.name = v.name)
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> where s.age < 25 and v.age <
25 and p.age < 25;
> Permalink Edit Delete 
> tsaito Takahiko Saito added a comment - 3 days ago
> More info about temporary table:
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> describe formatted s10k;
> +-------------------------------+---------------------------------------------------------------------------------------------------------------------------------------------+-----------------------+--+
> |           col_name            |                                                   
              data_type                                                                  |
       comment        |
> +-------------------------------+---------------------------------------------------------------------------------------------------------------------------------------------+-----------------------+--+
> | # col_name                    | data_type                                         
                                                                                         |
comment               |
> |                               | NULL                                              
                                                                                         |
NULL                  |
> | name                          | string                                            
                                                                                         |
                      |
> | age                           | int                                               
                                                                                         |
                      |
> | gpa                           | double                                            
                                                                                         |
                      |
> |                               | NULL                                              
                                                                                         |
NULL                  |
> | # Detailed Table Information  | NULL                                              
                                                                                         |
NULL                  |
> | Database:                     | default                                           
                                                                                         |
NULL                  |
> | Owner:                        | hrt_qa                                            
                                                                                         |
NULL                  |
> | CreateTime:                   | Tue Feb 02 23:02:31 UTC 2016                      
                                                                                         |
NULL                  |
> | LastAccessTime:               | UNKNOWN                                           
                                                                                         |
NULL                  |
> | Protect Mode:                 | None                                              
                                                                                         |
NULL                  |
> | Retention:                    | 0                                                 
                                                                                         |
NULL                  |
> | Location:                     | hdfs://hn0-hs21-h.hdinsight.net:8020/tmp/hive/hive/de2667ea-5bc1-4548-a4c4-97e46d6081f8/_tmp_space.db/8e8482c9-a2a4-4e8c-ad0f-2cd6e8fcdb4c
 | NULL                  |
> | Table Type:                   | MANAGED_TABLE                                     
                                                                                         |
NULL                  |
> |                               | NULL                                              
                                                                                         |
NULL                  |
> | # Storage Information         | NULL                                              
                                                                                         |
NULL                  |
> | SerDe Library:                | org.apache.hadoop.hive.ql.io.orc.OrcSerde         
                                                                                         |
NULL                  |
> | InputFormat:                  | org.apache.hadoop.hive.ql.io.orc.OrcInputFormat   
                                                                                         |
NULL                  |
> | OutputFormat:                 | org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat  
                                                                                         |
NULL                  |
> | Compressed:                   | No                                                
                                                                                         |
NULL                  |
> | Num Buckets:                  | -1                                                
                                                                                         |
NULL                  |
> | Bucket Columns:               | []                                                
                                                                                         |
NULL                  |
> | Sort Columns:                 | []                                                
                                                                                         |
NULL                  |
> | Storage Desc Params:          | NULL                                              
                                                                                         |
NULL                  |
> |                               | serialization.format                              
                                                                                         |
1                     |
> +-------------------------------+---------------------------------------------------------------------------------------------------------------------------------------------+-----------------------+--+
> 26 rows selected (0.22 seconds)
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> set hive.server2.enable.doAs;
> +---------------------------------+--+
> |               set               |
> +---------------------------------+--+
> | hive.server2.enable.doAs=false  |
> +---------------------------------+--+
> Permalink Edit Delete 
> tsaito Takahiko Saito added a comment - 3 days ago
> hdfs dir of temporary table is owned by hive as expected since hive.server2.enable.doAs=false:
> hdfs@hn0-hs21-h:~$ hdfs dfs -ls hdfs://hn0-hs21-h.hdinsight.net:8020/tmp/hive/hive/de2667ea-5bc1-4548-a4c4-97e46d6081f8/_tmp_space.db/8e8482c9-a2a4-4e8c-ad0f-2cd6e8fcdb4c
> Found 1 items
> -rwx------   3 hive hdfs      47777 2016-02-02 23:02 hdfs://hn0-hs21-h.hdinsight.net:8020/tmp/hive/hive/de2667ea-5bc1-4548-a4c4-97e46d6081f8/_tmp_space.db/8e8482c9-a2a4-4e8c-ad0f-2cd6e8fcdb4c/000000_0
> Not sure if it's related to the JIRA, but one thing I noticed was that dfs cmd throws
error via beeline:
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> dfs -ls hdfs://hn0-hs21-h.hdinsight.net:8020/tmp/hive/hive/de2667ea-5bc1-4548-a4c4-97e46d6081f8/_tmp_space.db/8e8482c9-a2a4-4e8c-ad0f-2cd6e8fcdb4c;
> Error: Error while processing statement: Permission denied: user [hrt_qa] does not have
privilege for [DFS] command (state=,code=1)
> Permalink Edit 
> thejas Thejas Nair added a comment - 5 hours ago - edited
> Takahiko Saito 
> The above DFS error would be from Ranger of SQL standard authorization. It is not related
to the test failure.
> But you are right that it looks like a permission issue.
> The issue seen in HIVE-739 also exists in this part of Hive. Hive in MR mode launches
a new child process to process the small table and create a hash table. This child process
needs credentials from HDFS. However, in this setup, azure fs is the default file system.
We should also get delegation token from all FS URIs listed under mapreduce.job.hdfs-servers
config.
> There are hueristics around when map-join gets used. That is why you don't see it unless
temp table is used (that uses ORC format and might have stats as well, while the original
table is probaly text format).
> https://github.com/hortonworks/hive/blob/2.4-maint/ql/src/java/org/apache/hadoop/hive/ql/exec/SecureCmdDoAs.java
needs change similar to webhcat change in HIVE-739
> Permalink Edit 
> sush Sushanth Sowmyan added a comment - 7 minutes ago
> Takahiko Saito, we should file an apache jira for this as well. Would you like to do
so, so the bug report has attribution to you?
> Permalink Edit Delete 
> tsaito Takahiko Saito added a comment - 4 minutes ago
> Sushanth Sowmyan I will file one and update with that JIRA.
>  Comment	
> People
> Assignee:	 sush Sushanth Sowmyan
> Assign to me
> Reporter:	 tsaito Takahiko Saito
> QEAssignee:	Takahiko Saito
> Votes:	0
> Watchers:	3 Stop watching this issue 
> Dates
> Created:	3 days ago
> Updated:	4 minutes ago
> Who's Looking?
> Agile
> View on Board
> {noformat}
> hivesever2 log shows:
> {noformat}
> 2016-02-02 18:04:43,766 ERROR [HiveServer2-Background-Pool: Thread-517]: operation.Operation
(SQLOperation.java:run(209)) - Error running hive query:
> org.apache.hive.service.cli.HiveSQLException: Error while processing statement: FAILED:
Execution Error, return code 2 from org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask
>         at org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:315)
>         at org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:156)
>         at org.apache.hive.service.cli.operation.SQLOperation.access$100(SQLOperation.java:71)
>         at org.apache.hive.service.cli.operation.SQLOperation$1$1.run(SQLOperation.java:206)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:415)
>         at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
>         at org.apache.hive.service.cli.operation.SQLOperation$1.run(SQLOperation.java:218)
>         at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
>         at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>         at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>         at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>         at java.lang.Thread.run(Thread.java:745)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Mime
View raw message