[ 
https://issues.apache.org/jira/browse/AMBARI-8152?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14197350#comment-14197350
 ] 

Hudson commented on AMBARI-8152:
--------------------------------

SUCCESS: Integrated in Ambari-trunk-Commit #829 (See 
[https://builds.apache.org/job/Ambari-trunk-Commit/829/])
AMBARI-8152. Pig Service check fails because tez tarball is not copied to HDFS 
(alejandro) (afernandez: 
http://git-wip-us.apache.org/repos/asf?p=ambari.git&a=commit&h=d6b88232cf3356e6b707b77841c7903a8ed97a96)
* 
ambari-server/src/main/resources/stacks/HDP/2.0.6/services/PIG/package/scripts/params.py
* 
ambari-server/src/main/resources/stacks/HDP/2.0.6/services/PIG/package/scripts/service_check.py


> Pig Service check fails because tez tarball is not copied to HDFS
> -----------------------------------------------------------------
>
>                 Key: AMBARI-8152
>                 URL: https://issues.apache.org/jira/browse/AMBARI-8152
>             Project: Ambari
>          Issue Type: Bug
>          Components: ambari-server
>    Affects Versions: 1.7.0
>            Reporter: Alejandro Fernandez
>            Assignee: Alejandro Fernandez
>            Priority: Critical
>             Fix For: 1.7.0
>
>         Attachments: AMBARI-8152.patch
>
>
> Pig service check fails with the below error on a secure cluster with 2.2 
> stack on SUSE11.3. 
> {code}
> 2014-11-04 18:49:13,411 - Error while executing command 'service_check':
> Traceback (most recent call last):
>   File 
> "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py",
>  line 122, in execute
>     method(env)
>   File 
> "/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/PIG/package/scripts/service_check.py",
>  line 87, in service_check
>     user      = params.smokeuser
>   File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", 
> line 148, in __init__
>     self.env.run()
>   File 
> "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", 
> line 149, in run
>     self.run_action(resource, action)
>   File 
> "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", 
> line 115, in run_action
>     provider_action()
>   File 
> "/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py",
>  line 241, in action_run
>     raise ex
> Fail: Execution of 'pig -x tez /var/lib/ambari-agent/data/tmp/pigSmoke.sh' 
> returned 2. 14/11/04 18:49:03 INFO pig.ExecTypeProvider: Trying ExecType : 
> LOCAL
> 14/11/04 18:49:03 INFO pig.ExecTypeProvider: Trying ExecType : MAPREDUCE
> 14/11/04 18:49:03 INFO pig.ExecTypeProvider: Trying ExecType : TEZ_LOCAL
> 14/11/04 18:49:03 INFO pig.ExecTypeProvider: Trying ExecType : TEZ
> 14/11/04 18:49:03 INFO pig.ExecTypeProvider: Picked TEZ as the ExecType
> 2014-11-04 18:49:04,007 [main] INFO  org.apache.pig.Main - Apache Pig version 
> 0.14.0.2.2.0.0-1756 (rexported) compiled Nov 04 2014, 04:52:01
> 2014-11-04 18:49:04,007 [main] INFO  org.apache.pig.Main - Logging error 
> messages to: /home/ambari-qa/pig_1415126944005.log
> 2014-11-04 18:49:05,706 [main] INFO  org.apache.pig.impl.util.Utils - Default 
> bootup file /home/ambari-qa/.pigbootup not found
> 2014-11-04 18:49:06,048 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.HExecutionEngine - Connecting 
> to hadoop file system at: hdfs://nameservice
> 2014-11-04 18:49:08,530 [main] INFO  org.apache.hadoop.hdfs.DFSClient - 
> Created HDFS_DELEGATION_TOKEN token 20 for ambari-qa on ha-hdfs:nameservice
> 2014-11-04 18:49:08,634 [main] INFO  
> org.apache.hadoop.mapreduce.security.TokenCache - Got dt for 
> hdfs://nameservice; Kind: HDFS_DELEGATION_TOKEN, Service: 
> ha-hdfs:nameservice, Ident: (HDFS_DELEGATION_TOKEN token 20 for ambari-qa)
> 2014-11-04 18:49:08,674 [main] INFO  
> org.apache.pig.tools.pigstats.ScriptState - Pig features used in the script: 
> UNKNOWN
> 2014-11-04 18:49:08,731 [main] INFO  org.apache.pig.data.SchemaTupleBackend - 
> Key [pig.schematuple] was not set... will not generate code.
> 2014-11-04 18:49:08,780 [main] INFO  
> org.apache.pig.newplan.logical.optimizer.LogicalPlanOptimizer - 
> {RULES_ENABLED=[AddForEach, ColumnMapKeyPrune, ConstantCalculator, 
> GroupByConstParallelSetter, LimitOptimizer, LoadTypeCastInserter, 
> MergeFilter, MergeForEach, PartitionFilterOptimizer, 
> PredicatePushdownOptimizer, PushDownForEachFlatten, PushUpFilter, 
> SplitFilter, StreamTypeCastInserter]}
> 2014-11-04 18:49:09,005 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.tez.TezLauncher - Tez staging 
> directory is /tmp/temp-1770525346
> 2014-11-04 18:49:09,062 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.tez.plan.TezCompiler - File 
> concatenation threshold: 100 optimistic? false
> 2014-11-04 18:49:09,234 [main] INFO  
> org.apache.tez.mapreduce.hadoop.MRInputHelpers - Generating mapreduce api 
> input splits
> 2014-11-04 18:49:09,385 [main] INFO  org.apache.hadoop.hdfs.DFSClient - 
> Created HDFS_DELEGATION_TOKEN token 21 for ambari-qa on ha-hdfs:nameservice
> 2014-11-04 18:49:09,385 [main] INFO  
> org.apache.hadoop.mapreduce.security.TokenCache - Got dt for 
> hdfs://nameservice; Kind: HDFS_DELEGATION_TOKEN, Service: 
> ha-hdfs:nameservice, Ident: (HDFS_DELEGATION_TOKEN token 21 for ambari-qa)
> 2014-11-04 18:49:09,390 [main] INFO  
> org.apache.hadoop.mapreduce.lib.input.FileInputFormat - Total input paths to 
> process : 1
> 2014-11-04 18:49:09,390 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.util.MapRedUtil - Total input 
> paths to process : 1
> 2014-11-04 18:49:09,484 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.util.MapRedUtil - Total input 
> paths (combined) to process : 1
> 2014-11-04 18:49:09,521 [main] INFO  
> org.apache.tez.mapreduce.hadoop.MRInputHelpers - NumSplits: 1, 
> SerializedSize: 374
> 2014-11-04 18:49:10,394 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.tez.TezJobCompiler - Local 
> resource: guava-11.0.2.jar
> 2014-11-04 18:49:10,394 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.tez.TezJobCompiler - Local 
> resource: antlr-runtime-3.4.jar
> 2014-11-04 18:49:10,394 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.tez.TezJobCompiler - Local 
> resource: joda-time-2.5.jar
> 2014-11-04 18:49:10,394 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.tez.TezJobCompiler - Local 
> resource: pig-0.14.0.2.2.0.0-1756-core-h2.jar
> 2014-11-04 18:49:10,394 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.tez.TezJobCompiler - Local 
> resource: automaton-1.11-8.jar
> 2014-11-04 18:49:10,455 [main] INFO  org.apache.hadoop.hdfs.DFSClient - 
> Created HDFS_DELEGATION_TOKEN token 22 for ambari-qa on ha-hdfs:nameservice
> 2014-11-04 18:49:10,456 [main] INFO  
> org.apache.hadoop.mapreduce.security.TokenCache - Got dt for 
> hdfs://nameservice; Kind: HDFS_DELEGATION_TOKEN, Service: 
> ha-hdfs:nameservice, Ident: (HDFS_DELEGATION_TOKEN token 22 for ambari-qa)
> 2014-11-04 18:49:10,607 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper - 
> Setting tez.runtime.shuffle.merge.percent to 0.66 from MR setting 
> mapreduce.reduce.shuffle.merge.percent
> 2014-11-04 18:49:10,607 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper - 
> Setting tez.runtime.shuffle.fetch.buffer.percent to 0.7 from MR setting 
> mapreduce.reduce.shuffle.input.buffer.percent
> 2014-11-04 18:49:10,607 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper - 
> Setting tez.runtime.io.sort.mb to 200 from MR setting 
> mapreduce.task.io.sort.mb
> 2014-11-04 18:49:10,607 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper - 
> Setting tez.runtime.shuffle.memory.limit.percent to 0.25 from MR setting 
> mapreduce.reduce.shuffle.memory.limit.percent
> 2014-11-04 18:49:10,608 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper - 
> Setting tez.runtime.io.sort.factor to 100 from MR setting 
> mapreduce.task.io.sort.factor
> 2014-11-04 18:49:10,608 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper - 
> Setting tez.runtime.shuffle.connect.timeout to 180000 from MR setting 
> mapreduce.reduce.shuffle.connect.timeout
> 2014-11-04 18:49:10,608 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper - 
> Setting tez.runtime.internal.sorter.class to org.apache.hadoop.util.QuickSort 
> from MR setting map.sort.class
> 2014-11-04 18:49:10,608 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper - 
> Setting tez.runtime.merge.progress.records to 10000 from MR setting 
> mapreduce.task.merge.progress.records
> 2014-11-04 18:49:10,608 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper - 
> Setting tez.runtime.compress to false from MR setting 
> mapreduce.map.output.compress
> 2014-11-04 18:49:10,608 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper - 
> Setting tez.runtime.sort.spill.percent to 0.7 from MR setting 
> mapreduce.map.sort.spill.percent
> 2014-11-04 18:49:10,608 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper - 
> Setting tez.runtime.shuffle.ssl.enable to false from MR setting 
> mapreduce.shuffle.ssl.enabled
> 2014-11-04 18:49:10,608 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper - 
> Setting tez.runtime.ifile.readahead to true from MR setting 
> mapreduce.ifile.readahead
> 2014-11-04 18:49:10,609 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper - 
> Setting tez.runtime.shuffle.parallel.copies to 30 from MR setting 
> mapreduce.reduce.shuffle.parallelcopies
> 2014-11-04 18:49:10,609 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper - 
> Setting tez.runtime.ifile.readahead.bytes to 4194304 from MR setting 
> mapreduce.ifile.readahead.bytes
> 2014-11-04 18:49:10,609 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper - 
> Setting tez.runtime.task.input.post-merge.buffer.percent to 0.0 from MR 
> setting mapreduce.reduce.input.buffer.percent
> 2014-11-04 18:49:10,609 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper - 
> Setting tez.runtime.shuffle.read.timeout to 180000 from MR setting 
> mapreduce.reduce.shuffle.read.timeout
> 2014-11-04 18:49:10,609 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.tez.util.MRToTezHelper - 
> Setting tez.runtime.compress.codec to 
> org.apache.hadoop.io.compress.DefaultCodec from MR setting 
> mapreduce.map.output.compress.codec
> 2014-11-04 18:49:10,686 [main] INFO  
> org.apache.pig.backend.hadoop.executionengine.tez.TezJobCompiler - For vertex 
> - scope-5: parallelism=1, memory=1024, java opts=-server -XX:NewRatio=8 
> -Djava.net.preferIPv4Stack=true -Dhdp.version=2.2.0.0-1756 -Xmx756m 
> -Dlog4j.configuration=tez-container-log4j.properties 
> -Dyarn.app.container.log.dir=<LOG_DIR> -Dtez.root.logger=INFO,CLA 
> 2014-11-04 18:49:11,909 [pool-1-thread-1] INFO  
> org.apache.tez.client.TezClient - Tez Client Version: [ component=tez-api, 
> version=0.5.2.2.2.0.0-1756, 
> revision=8655203fa290004209ec9c1e0779c14450f33bc2, 
> SCM-URL=scm:git:https://git-wip-us.apache.org/repos/asf/tez.git, 
> buildTIme=20141104-0436 ]
> 2014-11-04 18:49:11,971 [pool-1-thread-1] INFO  
> org.apache.hadoop.yarn.client.RMProxy - Connecting to ResourceManager at 
> ip-172-31-34-55.ec2.internal/172.31.34.55:8050
> 2014-11-04 18:49:12,161 [pool-1-thread-1] INFO  
> org.apache.tez.client.TezClient - Session mode. Starting session.
> 2014-11-04 18:49:12,162 [pool-1-thread-1] INFO  
> org.apache.tez.client.TezClientUtils - Using tez.lib.uris value from 
> configuration: /hdp/apps/2.2.0.0-1756/tez/tez.tar.gz
> 2014-11-04 18:49:12,165 [pool-1-thread-1] ERROR 
> org.apache.pig.backend.hadoop.executionengine.tez.TezJob - Cannot submit DAG
> java.io.FileNotFoundException: File does not exist: 
> /hdp/apps/2.2.0.0-1756/tez/tez.tar.gz
>       at 
> org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1122)
>       at 
> org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1114)
>       at 
> org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
>       at 
> org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1114)
>       at org.apache.hadoop.fs.FileSystem.resolvePath(FileSystem.java:750)
>       at 
> org.apache.tez.client.TezClientUtils.getLRFileStatus(TezClientUtils.java:126)
>       at 
> org.apache.tez.client.TezClientUtils.setupTezJarsLocalResources(TezClientUtils.java:177)
>       at 
> org.apache.tez.client.TezClient.getTezJarResources(TezClient.java:711)
>       at org.apache.tez.client.TezClient.start(TezClient.java:294)
>       at 
> org.apache.pig.backend.hadoop.executionengine.tez.TezSessionManager.createSession(TezSessionManager.java:92)
>       at 
> org.apache.pig.backend.hadoop.executionengine.tez.TezSessionManager.getClient(TezSessionManager.java:192)
>       at 
> org.apache.pig.backend.hadoop.executionengine.tez.TezJob.run(TezJob.java:153)
>       at java.lang.Thread.run(Thread.java:745)
>       at 
> org.apache.pig.backend.hadoop.executionengine.tez.TezLauncher$1.run(TezLauncher.java:154)
>       at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
>       at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>       at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:178)
>       at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:292)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>       at java.lang.Thread.run(Thread.java:745)
> 2014-11-04 18:49:12,906 [main] INFO  
> org.apache.pig.tools.pigstats.tez.TezStats - Script Statistics:
>        HadoopVersion: 2.6.0.2.2.0.0-1756                                      
>                                             
>           PigVersion: 0.14.0.2.2.0.0-1756                                     
>                                             
>           TezVersion: 0.5.2.2.2.0.0-1756                                      
>                                             
>               UserId: ambari-qa                                               
>                                             
>             FileName: /var/lib/ambari-agent/data/tmp/pigSmoke.sh              
>                                             
>            StartedAt: 2014-11-04 18:49:09                                     
>                                             
>           FinishedAt: 2014-11-04 18:49:12                                     
>                                             
>             Features: UNKNOWN                                                 
>                                             
> Failed!
> Input(s):
> Output(s):
> 2014-11-04 18:49:12,909 [main] ERROR org.apache.pig.tools.grunt.GruntParser - 
> ERROR 2244: Job scope-5 failed, hadoop does not return any error message
> Details at logfile: /home/ambari-qa/pig_1415126944005.log
> 2014-11-04 18:49:12,939 [main] INFO  org.apache.pig.Main - Pig script 
> completed in 9 seconds and 133 milliseconds (9133 ms)
> 2014-11-04 18:49:12,943 [Thread-1] ERROR org.apache.pig.impl.io.FileLocalizer 
> - java.io.IOException: Filesystem closed
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to