ambari-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Sumit Mohanty" <smoha...@hortonworks.com>
Subject Re: Review Request 27600: Pig Service check fails because tez tarball is not copied to HDFS
Date Wed, 05 Nov 2014 01:05:33 GMT

-----------------------------------------------------------
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/27600/#review59901
-----------------------------------------------------------

Ship it!


Ship It!

- Sumit Mohanty


On Nov. 5, 2014, 1:03 a.m., Alejandro Fernandez wrote:
> 
> -----------------------------------------------------------
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/27600/
> -----------------------------------------------------------
> 
> (Updated Nov. 5, 2014, 1:03 a.m.)
> 
> 
> Review request for Ambari, Mahadev Konar, Sumit Mohanty, Srimanth Gunturi, and Sid Wagle.
> 
> 
> Bugs: AMBARI-8152
>     https://issues.apache.org/jira/browse/AMBARI-8152
> 
> 
> Repository: ambari
> 
> 
> Description
> -------
> 
> Pig service check fails with the below error on a secure cluster with 2.2 stack on SUSE11.3.

> 
> ```
> 2014-11-04 18:49:13,411 - Error while executing command 'service_check':
> Traceback (most recent call last):
>   File "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py",
line 122, in execute
>     method(env)
>   File "/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/PIG/package/scripts/service_check.py",
line 87, in service_check
>     user      = params.smokeuser
>   File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", line 148,
in __init__
>     self.env.run()
>   File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line
149, in run
>     self.run_action(resource, action)
>   File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line
115, in run_action
>     provider_action()
>   File "/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py",
line 241, in action_run
>     raise ex
> Fail: Execution of 'pig -x tez /var/lib/ambari-agent/data/tmp/pigSmoke.sh' returned 2.
14/11/04 18:49:03 INFO pig.ExecTypeProvider: Trying ExecType : LOCAL
> 14/11/04 18:49:03 INFO pig.ExecTypeProvider: Trying ExecType : MAPREDUCE
> 14/11/04 18:49:03 INFO pig.ExecTypeProvider: Trying ExecType : TEZ_LOCAL
> 14/11/04 18:49:03 INFO pig.ExecTypeProvider: Trying ExecType : TEZ
> 14/11/04 18:49:03 INFO pig.ExecTypeProvider: Picked TEZ as the ExecType
> 2014-11-04 18:49:04,007 [main] INFO  org.apache.pig.Main - Apache Pig version 0.14.0.2.2.0.0-1756
(rexported) compiled Nov 04 2014, 04:52:01
> 2014-11-04 18:49:04,007 [main] INFO  org.apache.pig.Main - Logging error messages to:
/home/ambari-qa/pig_1415126944005.log
> 2014-11-04 18:49:05,706 [main] INFO  org.apache.pig.impl.util.Utils - Default bootup
file /home/ambari-qa/.pigbootup not found
> 2014-11-04 18:49:06,048 [main] INFO  org.apache.pig.backend.hadoop.executionengine.HExecutionEngine
- Connecting to hadoop file system at: hdfs://nameservice
> 2014-11-04 18:49:08,530 [main] INFO  org.apache.hadoop.hdfs.DFSClient - Created HDFS_DELEGATION_TOKEN
token 20 for ambari-qa on ha-hdfs:nameservice
> 2014-11-04 18:49:08,634 [main] INFO  org.apache.hadoop.mapreduce.security.TokenCache
- Got dt for hdfs://nameservice; Kind: HDFS_DELEGATION_TOKEN, Service: ha-hdfs:nameservice,
Ident: (HDFS_DELEGATION_TOKEN token 20 for ambari-qa)
> 2014-11-04 18:49:08,674 [main] INFO  org.apache.pig.tools.pigstats.ScriptState - Pig
features used in the script: UNKNOWN
> 2014-11-04 18:49:08,731 [main] INFO  org.apache.pig.data.SchemaTupleBackend - Key [pig.schematuple]
was not set... will not generate code.
> 2014-11-04 18:49:08,780 [main] INFO  org.apache.pig.newplan.logical.optimizer.LogicalPlanOptimizer
- {RULES_ENABLED=[AddForEach, ColumnMapKeyPrune, ConstantCalculator, GroupByConstParallelSetter,
LimitOptimizer, LoadTypeCastInserter, MergeFilter, MergeForEach, PartitionFilterOptimizer,
PredicatePushdownOptimizer, PushDownForEachFlatten, PushUpFilter, SplitFilter, StreamTypeCastInserter]}
> 2014-11-04 18:49:09,005 [main] INFO  org.apache.pig.backend.hadoop.executionengine.tez.TezLauncher
- Tez staging directory is /tmp/temp-1770525346
> 2014-11-04 18:49:09,062 [main] INFO  org.apache.pig.backend.hadoop.executionengine.tez.plan.TezCompiler
- File concatenation threshold: 100 optimistic? false
> 2014-11-04 18:49:09,234 [main] INFO  org.apache.tez.mapreduce.hadoop.MRInputHelpers -
Generating mapreduce api input splits
> 2014-11-04 18:49:09,385 [main] INFO  org.apache.hadoop.hdfs.DFSClient - Created HDFS_DELEGATION_TOKEN
token 21 for ambari-qa on ha-hdfs:nameservice
> 2014-11-04 18:49:09,385 [main] INFO  org.apache.hadoop.mapreduce.security.TokenCache
- Got dt for hdfs://nameservice; Kind: HDFS_DELEGATION_TOKEN, Service: ha-hdfs:nameservice,
Ident: (HDFS_DELEGATION_TOKEN token 21 for ambari-qa)
> 2014-11-04 18:49:09,390 [main] INFO  org.apache.hadoop.mapreduce.lib.input.FileInputFormat
- Total input paths to process : 1
> 2014-11-04 18:49:09,390 [main] INFO  org.apache.pig.backend.hadoop.executionengine.util.MapRedUtil
- Total input paths to process : 1
> 2014-11-04 18:49:09,484 [main] INFO  org.apache.pig.backend.hadoop.executionengine.util.MapRedUtil
- Total input paths (combined) to process : 1
> 2014-11-04 18:49:09,521 [main] INFO  org.apache.tez.mapreduce.hadoop.MRInputHelpers -
NumSplits: 1, SerializedSize: 374
> 2014-11-04 18:49:10,394 [main] INFO  org.apache.pig.backend.hadoop.executionengine.tez.TezJobCompiler
- Local resource: guava-11.0.2.jar
> 2014-11-04 18:49:10,394 [main] INFO  org.apache.pig.backend.hadoop.executionengine.tez.TezJobCompiler
- Local resource: antlr-runtime-3.4.jar
> 2014-11-04 18:49:10,394 [main] INFO  org.apache.pig.backend.hadoop.executionengine.tez.TezJobCompiler
- Local resource: joda-time-2.5.jar
> 2014-11-04 18:49:10,394 [main] INFO  org.apache.pig.backend.hadoop.executionengine.tez.TezJobCompiler
- Local resource: pig-0.14.0.2.2.0.0-1756-core-h2.jar
> 2014-11-04 18:49:10,394 [main] INFO  org.apache.pig.backend.hadoop.executionengine.tez.TezJobCompiler
- Local resource: automaton-1.11-8.jar
> 2014-11-04 18:49:10,455 [main] INFO  org.apache.hadoop.hdfs.DFSClient - Created HDFS_DELEGATION_TOKEN
token 22 for ambari-qa on ha-hdfs:nameservice
> 2014-11-04 18:49:10,456 [main] INFO  org.apache.hadoop.mapreduce.security.TokenCache
- Got dt for hdfs://nameservice; Kind: HDFS_DELEGATION_TOKEN, Service: ha-hdfs:nameservice,
Ident: (HDFS_DELEGATION_TOKEN token 22 for ambari-qa)
> 2014-11-04 18:49:10,607 [main] INFO  org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper
- Setting tez.runtime.shuffle.merge.percent to 0.66 from MR setting mapreduce.reduce.shuffle.merge.percent
> 2014-11-04 18:49:10,607 [main] INFO  org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper
- Setting tez.runtime.shuffle.fetch.buffer.percent to 0.7 from MR setting mapreduce.reduce.shuffle.input.buffer.percent
> 2014-11-04 18:49:10,607 [main] INFO  org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper
- Setting tez.runtime.io.sort.mb to 200 from MR setting mapreduce.task.io.sort.mb
> 2014-11-04 18:49:10,607 [main] INFO  org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper
- Setting tez.runtime.shuffle.memory.limit.percent to 0.25 from MR setting mapreduce.reduce.shuffle.memory.limit.percent
> 2014-11-04 18:49:10,608 [main] INFO  org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper
- Setting tez.runtime.io.sort.factor to 100 from MR setting mapreduce.task.io.sort.factor
> 2014-11-04 18:49:10,608 [main] INFO  org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper
- Setting tez.runtime.shuffle.connect.timeout to 180000 from MR setting mapreduce.reduce.shuffle.connect.timeout
> 2014-11-04 18:49:10,608 [main] INFO  org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper
- Setting tez.runtime.internal.sorter.class to org.apache.hadoop.util.QuickSort from MR setting
map.sort.class
> 2014-11-04 18:49:10,608 [main] INFO  org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper
- Setting tez.runtime.merge.progress.records to 10000 from MR setting mapreduce.task.merge.progress.records
> 2014-11-04 18:49:10,608 [main] INFO  org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper
- Setting tez.runtime.compress to false from MR setting mapreduce.map.output.compress
> 2014-11-04 18:49:10,608 [main] INFO  org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper
- Setting tez.runtime.sort.spill.percent to 0.7 from MR setting mapreduce.map.sort.spill.percent
> 2014-11-04 18:49:10,608 [main] INFO  org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper
- Setting tez.runtime.shuffle.ssl.enable to false from MR setting mapreduce.shuffle.ssl.enabled
> 2014-11-04 18:49:10,608 [main] INFO  org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper
- Setting tez.runtime.ifile.readahead to true from MR setting mapreduce.ifile.readahead
> 2014-11-04 18:49:10,609 [main] INFO  org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper
- Setting tez.runtime.shuffle.parallel.copies to 30 from MR setting mapreduce.reduce.shuffle.parallelcopies
> 2014-11-04 18:49:10,609 [main] INFO  org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper
- Setting tez.runtime.ifile.readahead.bytes to 4194304 from MR setting mapreduce.ifile.readahead.bytes
> 2014-11-04 18:49:10,609 [main] INFO  org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper
- Setting tez.runtime.task.input.post-merge.buffer.percent to 0.0 from MR setting mapreduce.reduce.input.buffer.percent
> 2014-11-04 18:49:10,609 [main] INFO  org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper
- Setting tez.runtime.shuffle.read.timeout to 180000 from MR setting mapreduce.reduce.shuffle.read.timeout
> 2014-11-04 18:49:10,609 [main] INFO  org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper
- Setting tez.runtime.compress.codec to org.apache.hadoop.io.compress.DefaultCodec from MR
setting mapreduce.map.output.compress.codec
> 2014-11-04 18:49:10,686 [main] INFO  org.apache.pig.backend.hadoop.executionengine.tez.TezJobCompiler
- For vertex - scope-5: parallelism=1, memory=1024, java opts=-server -XX:NewRatio=8 -Djava.net.preferIPv4Stack=true
-Dhdp.version=2.2.0.0-1756 -Xmx756m -Dlog4j.configuration=tez-container-log4j.properties -Dyarn.app.container.log.dir=<LOG_DIR>
-Dtez.root.logger=INFO,CLA 
> 2014-11-04 18:49:11,909 [pool-1-thread-1] INFO  org.apache.tez.client.TezClient - Tez
Client Version: [ component=tez-api, version=0.5.2.2.2.0.0-1756, revision=8655203fa290004209ec9c1e0779c14450f33bc2,
SCM-URL=scm:git:https://git-wip-us.apache.org/repos/asf/tez.git, buildTIme=20141104-0436 ]
> 2014-11-04 18:49:11,971 [pool-1-thread-1] INFO  org.apache.hadoop.yarn.client.RMProxy
- Connecting to ResourceManager at ip-172-31-34-55.ec2.internal/172.31.34.55:8050
> 2014-11-04 18:49:12,161 [pool-1-thread-1] INFO  org.apache.tez.client.TezClient - Session
mode. Starting session.
> 2014-11-04 18:49:12,162 [pool-1-thread-1] INFO  org.apache.tez.client.TezClientUtils
- Using tez.lib.uris value from configuration: /hdp/apps/2.2.0.0-1756/tez/tez.tar.gz
> 2014-11-04 18:49:12,165 [pool-1-thread-1] ERROR org.apache.pig.backend.hadoop.executionengine.tez.TezJob
- Cannot submit DAG
> java.io.FileNotFoundException: File does not exist: /hdp/apps/2.2.0.0-1756/tez/tez.tar.gz
> 	at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1122)
> 	at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1114)
> 	at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
> 	at org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1114)
> 	at org.apache.hadoop.fs.FileSystem.resolvePath(FileSystem.java:750)
> 	at org.apache.tez.client.TezClientUtils.getLRFileStatus(TezClientUtils.java:126)
> 	at org.apache.tez.client.TezClientUtils.setupTezJarsLocalResources(TezClientUtils.java:177)
> 	at org.apache.tez.client.TezClient.getTezJarResources(TezClient.java:711)
> 	at org.apache.tez.client.TezClient.start(TezClient.java:294)
> 	at org.apache.pig.backend.hadoop.executionengine.tez.TezSessionManager.createSession(TezSessionManager.java:92)
> 	at org.apache.pig.backend.hadoop.executionengine.tez.TezSessionManager.getClient(TezSessionManager.java:192)
> 	at org.apache.pig.backend.hadoop.executionengine.tez.TezJob.run(TezJob.java:153)
> 	at java.lang.Thread.run(Thread.java:745)
> 	at org.apache.pig.backend.hadoop.executionengine.tez.TezLauncher$1.run(TezLauncher.java:154)
> 	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
> 	at java.util.concurrent.FutureTask.run(FutureTask.java:262)
> 	at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:178)
> 	at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:292)
> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> 	at java.lang.Thread.run(Thread.java:745)
> 2014-11-04 18:49:12,906 [main] INFO  org.apache.pig.tools.pigstats.tez.TezStats - Script
Statistics:
> 
>        HadoopVersion: 2.6.0.2.2.0.0-1756                                            
                                     
>           PigVersion: 0.14.0.2.2.0.0-1756                                           
                                     
>           TezVersion: 0.5.2.2.2.0.0-1756                                            
                                     
>               UserId: ambari-qa                                                     
                                     
>             FileName: /var/lib/ambari-agent/data/tmp/pigSmoke.sh                    
                                     
>            StartedAt: 2014-11-04 18:49:09                                           
                                     
>           FinishedAt: 2014-11-04 18:49:12                                           
                                     
>             Features: UNKNOWN                                                       
                                     
> 
> Failed!
> 
> Input(s):
> 
> Output(s):
> 
> 2014-11-04 18:49:12,909 [main] ERROR org.apache.pig.tools.grunt.GruntParser - ERROR 2244:
Job scope-5 failed, hadoop does not return any error message
> Details at logfile: /home/ambari-qa/pig_1415126944005.log
> 2014-11-04 18:49:12,939 [main] INFO  org.apache.pig.Main - Pig script completed in 9
seconds and 133 milliseconds (9133 ms)
> 2014-11-04 18:49:12,943 [Thread-1] ERROR org.apache.pig.impl.io.FileLocalizer - java.io.IOException:
Filesystem closed
> ```
> 
> 
> Diffs
> -----
> 
>   ambari-server/src/main/resources/stacks/HDP/2.0.6/services/PIG/package/scripts/params.py
5f658b4 
>   ambari-server/src/main/resources/stacks/HDP/2.0.6/services/PIG/package/scripts/service_check.py
88c46e2 
> 
> Diff: https://reviews.apache.org/r/27600/diff/
> 
> 
> Testing
> -------
> 
> Ran unit tests in ambari-server
> 
> ----------------------------------------------------------------------
> Total run:684
> Total errors:0
> Total failures:0
> OK
> 
> 
> Verified on a live cluster with HDP 2.2 and Kerberos enabled that Pig Service check passed
without installing HiveServer2 (which also copies the tez tarball to HDFS)
> 
> 
> Thanks,
> 
> Alejandro Fernandez
> 
>


Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message