[
https://issues.apache.org/jira/browse/HIVE-13017?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15141854#comment-15141854
]
Thejas M Nair commented on HIVE-13017:
--------------------------------------
+1
Can you also please run the new lines through a formatter ? Some nits like
space missing after 'for', and after ',' .
> Child process of HiveServer2 fails to get delegation token from non default
> FileSystem
> --------------------------------------------------------------------------------------
>
> Key: HIVE-13017
> URL: https://issues.apache.org/jira/browse/HIVE-13017
> Project: Hive
> Issue Type: Bug
> Components: Authentication
> Affects Versions: 1.2.1
> Environment: Secure
> Reporter: Takahiko Saito
> Assignee: Sushanth Sowmyan
> Attachments: HIVE-13017.2.patch, HIVE-13017.patch
>
>
> The following query fails, when Azure Filesystem is used as default file
> system, and HDFS is used for intermediate data.
> {noformat}
> >>> create temporary table s10k stored as orc as select * from studenttab10k;
> >>> create temporary table v10k as select * from votertab10k;
> >>> select registration
> from s10k s join v10k v
> on (s.name = v.name) join studentparttab30k p
> on (p.name = v.name)
> where s.age < 25 and v.age < 25 and p.age < 25;
> ERROR : Execution failed with exit status: 2
> ERROR : Obtaining error information
> ERROR :
> Task failed!
> Task ID:
> Stage-5
> Logs:
> ERROR : /var/log/hive/hiveServer2.log
> Error: Error while processing statement: FAILED: Execution Error, return code
> 2 from org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask (state=08S01,code=2)
> Aborting command set because "force" is false and command failed: "select
> registration
> from s10k s join v10k v
> on (s.name = v.name) join studentparttab30k p
> on (p.name = v.name)
> where s.age < 25 and v.age < 25 and p.age < 25;"
> Closing: 0:
> jdbc:hive2://zk2-hs21-h.hdinsight.net:2181/;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=hiveserver2;principal=hive/[email protected];transportMode=http;httpPath=cliservice
> hiveServer2.log shows:
> 2016-02-02 18:04:34,182 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
> method=Driver.run from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:34,199 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
> method=TimeToSubmit from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:34,212 INFO [HiveServer2-HttpHandler-Pool: Thread-55]:
> thrift.ThriftHttpServlet (ThriftHttpServlet.java:doPost(127)) - Could not
> validate cookie sent, will try to generate a new cookie
> 2016-02-02 18:04:34,213 INFO [HiveServer2-Background-Pool: Thread-517]:
> ql.Driver (Driver.java:checkConcurrency(168)) - Concurrency mode is disabled,
> not creating a lock manager
> 2016-02-02 18:04:34,219 INFO [HiveServer2-HttpHandler-Pool: Thread-55]:
> thrift.ThriftHttpServlet (ThriftHttpServlet.java:doKerberosAuth(352)) -
> Failed to authenticate with http/_HOST kerberos principal, trying with
> hive/_HOST kerberos principal
> 2016-02-02 18:04:34,219 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
> method=Driver.execute from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:34,225 INFO [HiveServer2-Background-Pool: Thread-517]:
> ql.Driver (Driver.java:execute(1390)) - Setting caller context to query id
> hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0
> 2016-02-02 18:04:34,226 INFO [HiveServer2-Background-Pool: Thread-517]:
> ql.Driver (Driver.java:execute(1393)) - Starting
> command(queryId=hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0):
> select registration
> from s10k s join v10k v
> on (s.name = v.name) join studentparttab30k p
> on (p.name = v.name)
> where s.age < 25 and v.age < 25 and p.age < 25
> 2016-02-02 18:04:34,228 INFO [HiveServer2-Background-Pool: Thread-517]:
> hooks.ATSHook (ATSHook.java:<init>(90)) - Created ATS Hook
> 2016-02-02 18:04:34,229 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
> method=PreHook.org.apache.hadoop.hive.ql.hooks.ATSHook
> from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:34,237 INFO [HiveServer2-HttpHandler-Pool: Thread-55]:
> thrift.ThriftHttpServlet (ThriftHttpServlet.java:doPost(169)) - Cookie added
> for clientUserName hrt_qa
> 2016-02-02 18:04:34,238 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
> method=PreHook.org.apache.hadoop.hive.ql.hooks.ATSHook start=1454436274229
> end=1454436274238 duration=9 from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:34,239 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
> method=PreHook.org.apache.hadoop.hive.ql.security.authorization.plugin.DisallowTransformHook
> from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:34,240 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
> method=PreHook.org.apache.hadoop.hive.ql.security.authorization.plugin.DisallowTransformHook
> start=1454436274239 end=1454436274240 duration=1
> from=org.apache.hadoop.hive.ql.Driver>
> Query ID = hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0
> 2016-02-02 18:04:34,242 INFO [HiveServer2-Background-Pool: Thread-517]:
> ql.Driver (SessionState.java:printInfo(923)) - Query ID =
> hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0
> Total jobs = 1
> 2016-02-02 18:04:34,243 INFO [HiveServer2-Background-Pool: Thread-517]:
> ql.Driver (SessionState.java:printInfo(923)) - Total jobs = 1
> 2016-02-02 18:04:34,245 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
> method=TimeToSubmit start=1454436274199 end=1454436274245 duration=46
> from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:34,246 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG method=runTasks
> from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:34,247 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
> method=task.MAPREDLOCAL.Stage-5 from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:34,258 INFO [HiveServer2-Background-Pool: Thread-517]:
> ql.Driver (Driver.java:launchTask(1718)) - Starting task
> [Stage-5:MAPREDLOCAL] in serial mode
> 2016-02-02 18:04:34,280 INFO [HiveServer2-Background-Pool: Thread-517]:
> mr.MapredLocalTask (MapredLocalTask.java:executeInChildVM(158)) - Generating
> plan file
> file:/tmp/hive/916e3dbb-a10d-4888-a063-52fb058ea421/hive_2016-02-02_18-04-29_153_625340340820843828-4/-local-10006/plan.xml
> 2016-02-02 18:04:34,288 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
> method=serializePlan from=org.apache.hadoop.hive.ql.exec.Utilities>
> 2016-02-02 18:04:34,289 INFO [HiveServer2-Background-Pool: Thread-517]:
> exec.Utilities (Utilities.java:serializePlan(1028)) - Serializing
> MapredLocalWork via kryo
> 2016-02-02 18:04:34,290 INFO [ATS Logger 0]: hooks.ATSHook
> (ATSHook.java:createPreHookEvent(158)) - Received pre-hook notification for
> :hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0
> 2016-02-02 18:04:34,358 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
> method=serializePlan start=1454436274288 end=1454436274358 duration=70
> from=org.apache.hadoop.hive.ql.exec.Utilities>
> 2016-02-02 18:04:34,737 INFO [HiveServer2-Background-Pool: Thread-517]:
> mr.MapredLocalTask (MapredLocalTask.java:executeInChildVM(287)) - Executing:
> /usr/hdp/2.4.1.0-170/hadoop/bin/hadoop jar
> /usr/hdp/2.4.1.0-170/hive/lib/hive-common-1.2.1000.2.4.1.0-170.jar
> org.apache.hadoop.hive.ql.exec.mr.ExecDriver -localtask -plan
> file:/tmp/hive/916e3dbb-a10d-4888-a063-52fb058ea421/hive_2016-02-02_18-04-29_153_625340340820843828-4/-local-10006/plan.xml
> -jobconffile
> file:/tmp/hive/916e3dbb-a10d-4888-a063-52fb058ea421/hive_2016-02-02_18-04-29_153_625340340820843828-4/-local-10007/jobconf.xml
> WARNING: Use "yarn jar" to launch YARN applications.
> 2016-02-02 18:04:37,450 INFO
> [org.apache.ranger.audit.queue.AuditBatchQueue0]: provider.BaseAuditHandler
> (BaseAuditHandler.java:logStatus(312)) - Audit Status Log:
> name=hiveServer2.async.summary.batch.solr, interval=01:21.012 minutes,
> events=2, succcessCount=2, totalEvents=4, totalSuccessCount=4
> Execution log at:
> /tmp/hive/hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0.log
> 2016-02-02 18:04:39,248 INFO [HiveServer2-HttpHandler-Pool: Thread-55]:
> thrift.ThriftHttpServlet (ThriftHttpServlet.java:doPost(127)) - Could not
> validate cookie sent, will try to generate a new cookie
> 2016-02-02 18:04:39,254 INFO [HiveServer2-HttpHandler-Pool: Thread-55]:
> thrift.ThriftHttpServlet (ThriftHttpServlet.java:doKerberosAuth(352)) -
> Failed to authenticate with http/_HOST kerberos principal, trying with
> hive/_HOST kerberos principal
> 2016-02-02 18:04:39,261 INFO [HiveServer2-HttpHandler-Pool: Thread-55]:
> thrift.ThriftHttpServlet (ThriftHttpServlet.java:doPost(169)) - Cookie added
> for clientUserName hrt_qa
> 2016-02-02 18:04:40 Starting to launch local task to process map join;
> maximum memory = 477102080
> Execution failed with exit status: 2
> 2016-02-02 18:04:43,728 ERROR [HiveServer2-Background-Pool: Thread-517]:
> exec.Task (SessionState.java:printError(932)) - Execution failed with exit
> status: 2
> Obtaining error information
> 2016-02-02 18:04:43,730 ERROR [HiveServer2-Background-Pool: Thread-517]:
> exec.Task (SessionState.java:printError(932)) - Obtaining error information
> Task failed!
> Task ID:
> Stage-5
> Logs:
> 2016-02-02 18:04:43,730 ERROR [HiveServer2-Background-Pool: Thread-517]:
> exec.Task (SessionState.java:printError(932)) -
> Task failed!
> Task ID:
> Stage-5
> Logs:
> /var/log/hive/hiveServer2.log
> 2016-02-02 18:04:43,732 ERROR [HiveServer2-Background-Pool: Thread-517]:
> exec.Task (SessionState.java:printError(932)) - /var/log/hive/hiveServer2.log
> 2016-02-02 18:04:43,732 ERROR [HiveServer2-Background-Pool: Thread-517]:
> mr.MapredLocalTask (MapredLocalTask.java:executeInChildVM(307)) - Execution
> failed with exit status: 2
> 2016-02-02 18:04:43,733 INFO [HiveServer2-Background-Pool: Thread-517]:
> hooks.ATSHook (ATSHook.java:<init>(90)) - Created ATS Hook
> 2016-02-02 18:04:43,734 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
> method=FailureHook.org.apache.hadoop.hive.ql.hooks.ATSHook
> from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:43,736 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
> method=FailureHook.org.apache.hadoop.hive.ql.hooks.ATSHook
> start=1454436283734 end=1454436283736 duration=2
> from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:43,736 INFO [ATS Logger 0]: hooks.ATSHook
> (ATSHook.java:createPostHookEvent(193)) - Received post-hook notification for
> :hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0
> FAILED: Execution Error, return code 2 from
> org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask
> 2016-02-02 18:04:43,757 ERROR [HiveServer2-Background-Pool: Thread-517]:
> ql.Driver (SessionState.java:printError(932)) - FAILED: Execution Error,
> return code 2 from org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask
> 2016-02-02 18:04:43,758 INFO [HiveServer2-Background-Pool: Thread-517]:
> ql.Driver (Driver.java:execute(1621)) - Resetting the caller context to
> HIVE_SSN_ID:916e3dbb-a10d-4888-a063-52fb058ea421
> 2016-02-02 18:04:43,759 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
> method=Driver.execute start=1454436274219 end=1454436283759 duration=9540
> from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:43,760 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
> method=releaseLocks from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:43,761 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
> method=releaseLocks start=1454436283760 end=1454436283761 duration=1
> from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:43,766 ERROR [HiveServer2-Background-Pool: Thread-517]:
> operation.Operation (SQLOperation.java:run(209)) - Error running hive query:
> org.apache.hive.service.cli.HiveSQLException: Error while processing
> statement: FAILED: Execution Error, return code 2 from
> org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask
> at
> org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:315)
> at
> org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:156)
> at
> org.apache.hive.service.cli.operation.SQLOperation.access$100(SQLOperation.java:71)
> at
> org.apache.hive.service.cli.operation.SQLOperation$1$1.run(SQLOperation.java:206)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:415)
> at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
> at
> org.apache.hive.service.cli.operation.SQLOperation$1.run(SQLOperation.java:218)
> at
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
> at java.util.concurrent.FutureTask.run(FutureTask.java:262)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> hive Configs can be viewed from
> http://qelog.hortonworks.com/log/hs21-hbs24-1454382123/artifacts/tmpModifyConfDir_1454394513592/
> Attachments
> Drop files to attach, or browse.
> Add LinkIssue Links
> relates to
> Bug - A problem which impairs or prevents the functions of the product.
> HIVE-739 webhcat tests failing in HDInsight secure cluster throwing
> NullPointerException Blocker - Blocks development and/or testing work,
> production could not run. RESOLVED
> Activity
> All
> Comments
> Work Log
> History
> Activity
> Ascending order - Click to sort in descending order
> Permalink Edit Delete
> tsaito Takahiko Saito added a comment - 3 days ago
> The test passes via hive CLI and explain shows:
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> explain select registration
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> from s10k s join v10k v
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> on (s.name = v.name) join
> studentparttab30k p
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> on (p.name = v.name)
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> where s.age < 25 and v.age <
> 25 and p.age < 25;
> +--------------------------------------------------------------------------------------------------------------+--+
> | Explain
> |
> +--------------------------------------------------------------------------------------------------------------+--+
> | STAGE DEPENDENCIES:
> |
> | Stage-5 is a root stage
> |
> | Stage-4 depends on stages: Stage-5
> |
> | Stage-0 depends on stages: Stage-4
> |
> |
> |
> | STAGE PLANS:
> |
> | Stage: Stage-5
> |
> | Map Reduce Local Work
> |
> | Alias -> Map Local Tables:
> |
> | s
> |
> | Fetch Operator
> |
> | limit: -1
> |
> | v
> |
> | Fetch Operator
> |
> | limit: -1
> |
> | Alias -> Map Local Operator Tree:
> |
> | s
> |
> | TableScan
> |
> | alias: s
> |
> | filterExpr: (name is not null and (age < 25)) (type: boolean)
> |
> | Statistics: Num rows: 459 Data size: 47777 Basic stats:
> COMPLETE Column stats: NONE |
> | Filter Operator
> |
> | predicate: (name is not null and (age < 25)) (type: boolean)
> |
> | Statistics: Num rows: 76 Data size: 7910 Basic stats:
> COMPLETE Column stats: NONE |
> | HashTable Sink Operator
> |
> | keys:
> |
> | 0 name (type: string)
> |
> | 1 name (type: string)
> |
> | 2 name (type: string)
> |
> | v
> |
> | TableScan
> |
> | alias: v
> |
> | filterExpr: (name is not null and (age < 25)) (type: boolean)
> |
> | Statistics: Num rows: 1653 Data size: 337233 Basic stats:
> COMPLETE Column stats: NONE |
> | Filter Operator
> |
> | predicate: (name is not null and (age < 25)) (type: boolean)
> |
> | Statistics: Num rows: 275 Data size: 56103 Basic stats:
> COMPLETE Column stats: NONE |
> | HashTable Sink Operator
> |
> | keys:
> |
> | 0 name (type: string)
> |
> | 1 name (type: string)
> |
> | 2 name (type: string)
> |
> |
> |
> | Stage: Stage-4
> |
> | Map Reduce
> |
> | Map Operator Tree:
> |
> | TableScan
> |
> | alias: p
> |
> | filterExpr: (name is not null and (age < 25)) (type: boolean)
> |
> | Statistics: Num rows: 30000 Data size: 627520 Basic stats:
> COMPLETE Column stats: COMPLETE |
> | Filter Operator
> |
> | predicate: (name is not null and (age < 25)) (type: boolean)
> |
> | Statistics: Num rows: 10000 Data size: 1010000 Basic stats:
> COMPLETE Column stats: COMPLETE |
> | Map Join Operator
> |
> | condition map:
> |
> | Inner Join 0 to 1
> |
> | Inner Join 1 to 2
> |
> | keys:
> |
> | 0 name (type: string)
> |
> | 1 name (type: string)
> |
> | 2 name (type: string)
> |
> | outputColumnNames: _col8
> |
> | Statistics: Num rows: 22000 Data size: 2222000 Basic stats:
> COMPLETE Column stats: NONE |
> | Select Operator
> |
> | expressions: _col8 (type: string)
> |
> | outputColumnNames: _col0
> |
> | Statistics: Num rows: 22000 Data size: 2222000 Basic
> stats: COMPLETE Column stats: NONE |
> | File Output Operator
> |
> | compressed: false
> |
> | Statistics: Num rows: 22000 Data size: 2222000 Basic
> stats: COMPLETE Column stats: NONE |
> | table:
> |
> | input format:
> org.apache.hadoop.mapred.TextInputFormat |
> | output format:
> org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat |
> | serde:
> org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
> |
> | Local Work:
> |
> | Map Reduce Local Work
> |
> |
> |
> | Stage: Stage-0
> |
> | Fetch Operator
> |
> | limit: -1
> |
> | Processor Tree:
> |
> | ListSink
> |
> |
> |
> +--------------------------------------------------------------------------------------------------------------+--+
> 83 rows selected (2.473 seconds)
> Permalink Edit Delete
> tsaito Takahiko Saito added a comment - 3 days ago
> emptablemisc_8 also fails with the same error:
> >>> create temporary table temp1 as select * from votertab10k;
> >>> select *
> from studenttab10k s
> where s.name not in
> (select name from temp1);
> INFO : Number of reduce tasks determined at compile time: 1
> INFO : In order to change the average load for a reducer (in bytes):
> INFO : set hive.exec.reducers.bytes.per.reducer=<number>
> INFO : In order to limit the maximum number of reducers:
> INFO : set hive.exec.reducers.max=<number>
> INFO : In order to set a constant number of reducers:
> INFO : set mapreduce.job.reduces=<number>
> INFO : number of splits:1
> INFO : Submitting tokens for job: job_1454394534358_0164
> INFO : Kind: HDFS_DELEGATION_TOKEN, Service: 10.0.0.36:8020, Ident:
> (HDFS_DELEGATION_TOKEN token 246 for hrt_qa)
> INFO : The url to track the job:
> http://hn0-hs21-h.hdinsight.net:8088/proxy/application_1454394534358_0164/
> INFO : Starting Job = job_1454394534358_0164, Tracking URL =
> http://hn0-hs21-h.hdinsight.net:8088/proxy/application_1454394534358_0164/
> INFO : Kill Command = /usr/hdp/2.4.1.0-170/hadoop/bin/hadoop job -kill
> job_1454394534358_0164
> INFO : Hadoop job information for Stage-4: number of mappers: 1; number of
> reducers: 1
> INFO : 2016-02-02 10:11:02,367 Stage-4 map = 0%, reduce = 0%
> INFO : 2016-02-02 10:11:26,060 Stage-4 map = 100%, reduce = 0%, Cumulative
> CPU 5.48 sec
> INFO : 2016-02-02 10:11:39,024 Stage-4 map = 100%, reduce = 100%,
> Cumulative CPU 11.18 sec
> INFO : MapReduce Total cumulative CPU time: 11 seconds 180 msec
> INFO : Ended Job = job_1454394534358_0164
> INFO : Stage-9 is selected by condition resolver.
> INFO : Stage-1 is filtered out by condition resolver.
> ERROR : Execution failed with exit status: 2
> ERROR : Obtaining error information
> ERROR :
> Task failed!
> Task ID:
> Stage-9
> Logs:
> ERROR : /var/log/hive/hiveServer2.log
> INFO : Number of reduce tasks determined at compile time: 1
> INFO : In order to change the average load for a reducer (in bytes):
> INFO : set hive.exec.reducers.bytes.per.reducer=<number>
> INFO : In order to limit the maximum number of reducers:
> INFO : set hive.exec.reducers.max=<number>
> INFO : In order to set a constant number of reducers:
> INFO : set mapreduce.job.reduces=<number>
> INFO : number of splits:2
> INFO : Submitting tokens for job: job_1454394534358_0169
> INFO : Kind: HDFS_DELEGATION_TOKEN, Service: 10.0.0.36:8020, Ident:
> (HDFS_DELEGATION_TOKEN token 252 for hrt_qa)
> INFO : The url to track the job:
> http://hn0-hs21-h.hdinsight.net:8088/proxy/application_1454394534358_0169/
> INFO : Starting Job = job_1454394534358_0169, Tracking URL =
> http://hn0-hs21-h.hdinsight.net:8088/proxy/application_1454394534358_0169/
> INFO : Kill Command = /usr/hdp/2.4.1.0-170/hadoop/bin/hadoop job -kill
> job_1454394534358_0169
> INFO : Hadoop job information for Stage-1: number of mappers: 2; number of
> reducers: 1
> INFO : 2016-02-02 10:16:38,027 Stage-1 map = 0%, reduce = 0%
> INFO : 2016-02-02 10:16:52,498 Stage-1 map = 50%, reduce = 0%, Cumulative
> CPU 3.8 sec
> INFO : 2016-02-02 10:16:53,566 Stage-1 map = 100%, reduce = 0%, Cumulative
> CPU 9.94 sec
> INFO : 2016-02-02 10:17:16,202 Stage-1 map = 100%, reduce = 100%,
> Cumulative CPU 17.04 sec
> INFO : MapReduce Total cumulative CPU time: 17 seconds 40 msec
> INFO : Ended Job = job_1454394534358_0169
> ERROR : Execution failed with exit status: 2
> ERROR : Obtaining error information
> ERROR :
> Task failed!
> Task ID:
> Stage-8
> Logs:
> ERROR : /var/log/hive/hiveServer2.log
> Error: Error while processing statement: FAILED: Execution Error, return code
> 2 from org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask (state=08S01,code=2)
> Aborting command set because "force" is false and command failed: "select *
> from studenttab10k s
> where s.name not in
> (select name from temp1);"
> Closing: 0:
> jdbc:hive2://zk2-hs21-h.hdinsight.net:2181/;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=hiveserver2;principal=hive/[email protected];transportMode=http;httpPath=cliservice
> Its app log can be viewed at
> http://qelog.hortonworks.com/log/hs21-hbs24-1454382123/app-logs/application_1454394534358_0169.log
> Permalink Edit Delete
> tsaito Takahiko Saito added a comment - 3 days ago
> The query works without 'temporary table':
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> select registration from
> studenttab10k s join votertab10k v
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> on (s.name = v.name) join
> studentparttab30k p
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> on (p.name = v.name)
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> where s.age < 25 and v.age <
> 25 and p.age < 25;
> Permalink Edit Delete
> tsaito Takahiko Saito added a comment - 3 days ago
> More info about temporary table:
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> describe formatted s10k;
> +-------------------------------+---------------------------------------------------------------------------------------------------------------------------------------------+-----------------------+--+
> | col_name |
> data_type
> | comment |
> +-------------------------------+---------------------------------------------------------------------------------------------------------------------------------------------+-----------------------+--+
> | # col_name | data_type
>
> | comment |
> | | NULL
>
> | NULL |
> | name | string
>
> | |
> | age | int
>
> | |
> | gpa | double
>
> | |
> | | NULL
>
> | NULL |
> | # Detailed Table Information | NULL
>
> | NULL |
> | Database: | default
>
> | NULL |
> | Owner: | hrt_qa
>
> | NULL |
> | CreateTime: | Tue Feb 02 23:02:31 UTC 2016
>
> | NULL |
> | LastAccessTime: | UNKNOWN
>
> | NULL |
> | Protect Mode: | None
>
> | NULL |
> | Retention: | 0
>
> | NULL |
> | Location: |
> hdfs://hn0-hs21-h.hdinsight.net:8020/tmp/hive/hive/de2667ea-5bc1-4548-a4c4-97e46d6081f8/_tmp_space.db/8e8482c9-a2a4-4e8c-ad0f-2cd6e8fcdb4c
> | NULL |
> | Table Type: | MANAGED_TABLE
>
> | NULL |
> | | NULL
>
> | NULL |
> | # Storage Information | NULL
>
> | NULL |
> | SerDe Library: | org.apache.hadoop.hive.ql.io.orc.OrcSerde
>
> | NULL |
> | InputFormat: |
> org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
> | NULL
> |
> | OutputFormat: |
> org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
> | NULL
> |
> | Compressed: | No
>
> | NULL |
> | Num Buckets: | -1
>
> | NULL |
> | Bucket Columns: | []
>
> | NULL |
> | Sort Columns: | []
>
> | NULL |
> | Storage Desc Params: | NULL
>
> | NULL |
> | | serialization.format
>
> | 1 |
> +-------------------------------+---------------------------------------------------------------------------------------------------------------------------------------------+-----------------------+--+
> 26 rows selected (0.22 seconds)
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> set hive.server2.enable.doAs;
> +---------------------------------+--+
> | set |
> +---------------------------------+--+
> | hive.server2.enable.doAs=false |
> +---------------------------------+--+
> Permalink Edit Delete
> tsaito Takahiko Saito added a comment - 3 days ago
> hdfs dir of temporary table is owned by hive as expected since
> hive.server2.enable.doAs=false:
> hdfs@hn0-hs21-h:~$ hdfs dfs -ls
> hdfs://hn0-hs21-h.hdinsight.net:8020/tmp/hive/hive/de2667ea-5bc1-4548-a4c4-97e46d6081f8/_tmp_space.db/8e8482c9-a2a4-4e8c-ad0f-2cd6e8fcdb4c
> Found 1 items
> -rwx------ 3 hive hdfs 47777 2016-02-02 23:02
> hdfs://hn0-hs21-h.hdinsight.net:8020/tmp/hive/hive/de2667ea-5bc1-4548-a4c4-97e46d6081f8/_tmp_space.db/8e8482c9-a2a4-4e8c-ad0f-2cd6e8fcdb4c/000000_0
> Not sure if it's related to the JIRA, but one thing I noticed was that dfs
> cmd throws error via beeline:
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> dfs -ls
> hdfs://hn0-hs21-h.hdinsight.net:8020/tmp/hive/hive/de2667ea-5bc1-4548-a4c4-97e46d6081f8/_tmp_space.db/8e8482c9-a2a4-4e8c-ad0f-2cd6e8fcdb4c;
> Error: Error while processing statement: Permission denied: user [hrt_qa]
> does not have privilege for [DFS] command (state=,code=1)
> Permalink Edit
> thejas Thejas Nair added a comment - 5 hours ago - edited
> Takahiko Saito
> The above DFS error would be from Ranger of SQL standard authorization. It is
> not related to the test failure.
> But you are right that it looks like a permission issue.
> The issue seen in HIVE-739 also exists in this part of Hive. Hive in MR mode
> launches a new child process to process the small table and create a hash
> table. This child process needs credentials from HDFS. However, in this
> setup, azure fs is the default file system. We should also get delegation
> token from all FS URIs listed under mapreduce.job.hdfs-servers config.
> There are hueristics around when map-join gets used. That is why you don't
> see it unless temp table is used (that uses ORC format and might have stats
> as well, while the original table is probaly text format).
> https://github.com/hortonworks/hive/blob/2.4-maint/ql/src/java/org/apache/hadoop/hive/ql/exec/SecureCmdDoAs.java
> needs change similar to webhcat change in HIVE-739
> Permalink Edit
> sush Sushanth Sowmyan added a comment - 7 minutes ago
> Takahiko Saito, we should file an apache jira for this as well. Would you
> like to do so, so the bug report has attribution to you?
> Permalink Edit Delete
> tsaito Takahiko Saito added a comment - 4 minutes ago
> Sushanth Sowmyan I will file one and update with that JIRA.
> Comment
> People
> Assignee: sush Sushanth Sowmyan
> Assign to me
> Reporter: tsaito Takahiko Saito
> QEAssignee: Takahiko Saito
> Votes: 0
> Watchers: 3 Stop watching this issue
> Dates
> Created: 3 days ago
> Updated: 4 minutes ago
> Who's Looking?
> Agile
> View on Board
> {noformat}
> hivesever2 log shows:
> {noformat}
> 2016-02-02 18:04:43,766 ERROR [HiveServer2-Background-Pool: Thread-517]:
> operation.Operation (SQLOperation.java:run(209)) - Error running hive query:
> org.apache.hive.service.cli.HiveSQLException: Error while processing
> statement: FAILED: Execution Error, return code 2 from
> org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask
> at
> org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:315)
> at
> org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:156)
> at
> org.apache.hive.service.cli.operation.SQLOperation.access$100(SQLOperation.java:71)
> at
> org.apache.hive.service.cli.operation.SQLOperation$1$1.run(SQLOperation.java:206)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:415)
> at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
> at
> org.apache.hive.service.cli.operation.SQLOperation$1.run(SQLOperation.java:218)
> at
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
> at java.util.concurrent.FutureTask.run(FutureTask.java:262)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> {noformat}
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)