[
https://issues.apache.org/jira/browse/HIVE-13017?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Thejas M Nair updated HIVE-13017:
---------------------------------
Description:
The following query fails, when Azure Filesystem is used as default file
system, and HDFS is used for intermediate data.
{noformat}
>>> create temporary table s10k stored as orc as select * from studenttab10k;
>>> create temporary table v10k as select * from votertab10k;
>>> select registration
from s10k s join v10k v
on (s.name = v.name) join studentparttab30k p
on (p.name = v.name)
where s.age < 25 and v.age < 25 and p.age < 25;
ERROR : Execution failed with exit status: 2
ERROR : Obtaining error information
ERROR :
Task failed!
Task ID:
Stage-5
Logs:
ERROR : /var/log/hive/hiveServer2.log
Error: Error while processing statement: FAILED: Execution Error, return code 2
from org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask (state=08S01,code=2)
Aborting command set because "force" is false and command failed: "select
registration
from s10k s join v10k v
on (s.name = v.name) join studentparttab30k p
on (p.name = v.name)
where s.age < 25 and v.age < 25 and p.age < 25;"
Closing: 0:
jdbc:hive2://zk2-hs21-h.hdinsight.net:2181/;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=hiveserver2;principal=hive/[email protected];transportMode=http;httpPath=cliservice
hiveServer2.log shows:
2016-02-02 18:04:34,182 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG method=Driver.run
from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:34,199 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
method=TimeToSubmit from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:34,212 INFO [HiveServer2-HttpHandler-Pool: Thread-55]:
thrift.ThriftHttpServlet (ThriftHttpServlet.java:doPost(127)) - Could not
validate cookie sent, will try to generate a new cookie
2016-02-02 18:04:34,213 INFO [HiveServer2-Background-Pool: Thread-517]:
ql.Driver (Driver.java:checkConcurrency(168)) - Concurrency mode is disabled,
not creating a lock manager
2016-02-02 18:04:34,219 INFO [HiveServer2-HttpHandler-Pool: Thread-55]:
thrift.ThriftHttpServlet (ThriftHttpServlet.java:doKerberosAuth(352)) - Failed
to authenticate with http/_HOST kerberos principal, trying with hive/_HOST
kerberos principal
2016-02-02 18:04:34,219 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
method=Driver.execute from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:34,225 INFO [HiveServer2-Background-Pool: Thread-517]:
ql.Driver (Driver.java:execute(1390)) - Setting caller context to query id
hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0
2016-02-02 18:04:34,226 INFO [HiveServer2-Background-Pool: Thread-517]:
ql.Driver (Driver.java:execute(1393)) - Starting
command(queryId=hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0):
select registration
from s10k s join v10k v
on (s.name = v.name) join studentparttab30k p
on (p.name = v.name)
where s.age < 25 and v.age < 25 and p.age < 25
2016-02-02 18:04:34,228 INFO [HiveServer2-Background-Pool: Thread-517]:
hooks.ATSHook (ATSHook.java:<init>(90)) - Created ATS Hook
2016-02-02 18:04:34,229 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
method=PreHook.org.apache.hadoop.hive.ql.hooks.ATSHook
from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:34,237 INFO [HiveServer2-HttpHandler-Pool: Thread-55]:
thrift.ThriftHttpServlet (ThriftHttpServlet.java:doPost(169)) - Cookie added
for clientUserName hrt_qa
2016-02-02 18:04:34,238 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
method=PreHook.org.apache.hadoop.hive.ql.hooks.ATSHook start=1454436274229
end=1454436274238 duration=9 from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:34,239 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
method=PreHook.org.apache.hadoop.hive.ql.security.authorization.plugin.DisallowTransformHook
from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:34,240 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
method=PreHook.org.apache.hadoop.hive.ql.security.authorization.plugin.DisallowTransformHook
start=1454436274239 end=1454436274240 duration=1
from=org.apache.hadoop.hive.ql.Driver>
Query ID = hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0
2016-02-02 18:04:34,242 INFO [HiveServer2-Background-Pool: Thread-517]:
ql.Driver (SessionState.java:printInfo(923)) - Query ID =
hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0
Total jobs = 1
2016-02-02 18:04:34,243 INFO [HiveServer2-Background-Pool: Thread-517]:
ql.Driver (SessionState.java:printInfo(923)) - Total jobs = 1
2016-02-02 18:04:34,245 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
method=TimeToSubmit start=1454436274199 end=1454436274245 duration=46
from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:34,246 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG method=runTasks
from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:34,247 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
method=task.MAPREDLOCAL.Stage-5 from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:34,258 INFO [HiveServer2-Background-Pool: Thread-517]:
ql.Driver (Driver.java:launchTask(1718)) - Starting task [Stage-5:MAPREDLOCAL]
in serial mode
2016-02-02 18:04:34,280 INFO [HiveServer2-Background-Pool: Thread-517]:
mr.MapredLocalTask (MapredLocalTask.java:executeInChildVM(158)) - Generating
plan file
file:/tmp/hive/916e3dbb-a10d-4888-a063-52fb058ea421/hive_2016-02-02_18-04-29_153_625340340820843828-4/-local-10006/plan.xml
2016-02-02 18:04:34,288 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
method=serializePlan from=org.apache.hadoop.hive.ql.exec.Utilities>
2016-02-02 18:04:34,289 INFO [HiveServer2-Background-Pool: Thread-517]:
exec.Utilities (Utilities.java:serializePlan(1028)) - Serializing
MapredLocalWork via kryo
2016-02-02 18:04:34,290 INFO [ATS Logger 0]: hooks.ATSHook
(ATSHook.java:createPreHookEvent(158)) - Received pre-hook notification for
:hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0
2016-02-02 18:04:34,358 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
method=serializePlan start=1454436274288 end=1454436274358 duration=70
from=org.apache.hadoop.hive.ql.exec.Utilities>
2016-02-02 18:04:34,737 INFO [HiveServer2-Background-Pool: Thread-517]:
mr.MapredLocalTask (MapredLocalTask.java:executeInChildVM(287)) - Executing:
/usr/hdp/2.4.1.0-170/hadoop/bin/hadoop jar
/usr/hdp/2.4.1.0-170/hive/lib/hive-common-1.2.1000.2.4.1.0-170.jar
org.apache.hadoop.hive.ql.exec.mr.ExecDriver -localtask -plan
file:/tmp/hive/916e3dbb-a10d-4888-a063-52fb058ea421/hive_2016-02-02_18-04-29_153_625340340820843828-4/-local-10006/plan.xml
-jobconffile
file:/tmp/hive/916e3dbb-a10d-4888-a063-52fb058ea421/hive_2016-02-02_18-04-29_153_625340340820843828-4/-local-10007/jobconf.xml
WARNING: Use "yarn jar" to launch YARN applications.
2016-02-02 18:04:37,450 INFO [org.apache.ranger.audit.queue.AuditBatchQueue0]:
provider.BaseAuditHandler (BaseAuditHandler.java:logStatus(312)) - Audit Status
Log: name=hiveServer2.async.summary.batch.solr, interval=01:21.012 minutes,
events=2, succcessCount=2, totalEvents=4, totalSuccessCount=4
Execution log at:
/tmp/hive/hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0.log
2016-02-02 18:04:39,248 INFO [HiveServer2-HttpHandler-Pool: Thread-55]:
thrift.ThriftHttpServlet (ThriftHttpServlet.java:doPost(127)) - Could not
validate cookie sent, will try to generate a new cookie
2016-02-02 18:04:39,254 INFO [HiveServer2-HttpHandler-Pool: Thread-55]:
thrift.ThriftHttpServlet (ThriftHttpServlet.java:doKerberosAuth(352)) - Failed
to authenticate with http/_HOST kerberos principal, trying with hive/_HOST
kerberos principal
2016-02-02 18:04:39,261 INFO [HiveServer2-HttpHandler-Pool: Thread-55]:
thrift.ThriftHttpServlet (ThriftHttpServlet.java:doPost(169)) - Cookie added
for clientUserName hrt_qa
2016-02-02 18:04:40 Starting to launch local task to process map join;
maximum memory = 477102080
Execution failed with exit status: 2
2016-02-02 18:04:43,728 ERROR [HiveServer2-Background-Pool: Thread-517]:
exec.Task (SessionState.java:printError(932)) - Execution failed with exit
status: 2
Obtaining error information
2016-02-02 18:04:43,730 ERROR [HiveServer2-Background-Pool: Thread-517]:
exec.Task (SessionState.java:printError(932)) - Obtaining error information
Task failed!
Task ID:
Stage-5
Logs:
2016-02-02 18:04:43,730 ERROR [HiveServer2-Background-Pool: Thread-517]:
exec.Task (SessionState.java:printError(932)) -
Task failed!
Task ID:
Stage-5
Logs:
/var/log/hive/hiveServer2.log
2016-02-02 18:04:43,732 ERROR [HiveServer2-Background-Pool: Thread-517]:
exec.Task (SessionState.java:printError(932)) - /var/log/hive/hiveServer2.log
2016-02-02 18:04:43,732 ERROR [HiveServer2-Background-Pool: Thread-517]:
mr.MapredLocalTask (MapredLocalTask.java:executeInChildVM(307)) - Execution
failed with exit status: 2
2016-02-02 18:04:43,733 INFO [HiveServer2-Background-Pool: Thread-517]:
hooks.ATSHook (ATSHook.java:<init>(90)) - Created ATS Hook
2016-02-02 18:04:43,734 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
method=FailureHook.org.apache.hadoop.hive.ql.hooks.ATSHook
from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:43,736 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
method=FailureHook.org.apache.hadoop.hive.ql.hooks.ATSHook start=1454436283734
end=1454436283736 duration=2 from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:43,736 INFO [ATS Logger 0]: hooks.ATSHook
(ATSHook.java:createPostHookEvent(193)) - Received post-hook notification for
:hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0
FAILED: Execution Error, return code 2 from
org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask
2016-02-02 18:04:43,757 ERROR [HiveServer2-Background-Pool: Thread-517]:
ql.Driver (SessionState.java:printError(932)) - FAILED: Execution Error, return
code 2 from org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask
2016-02-02 18:04:43,758 INFO [HiveServer2-Background-Pool: Thread-517]:
ql.Driver (Driver.java:execute(1621)) - Resetting the caller context to
HIVE_SSN_ID:916e3dbb-a10d-4888-a063-52fb058ea421
2016-02-02 18:04:43,759 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
method=Driver.execute start=1454436274219 end=1454436283759 duration=9540
from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:43,760 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
method=releaseLocks from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:43,761 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
method=releaseLocks start=1454436283760 end=1454436283761 duration=1
from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:43,766 ERROR [HiveServer2-Background-Pool: Thread-517]:
operation.Operation (SQLOperation.java:run(209)) - Error running hive query:
org.apache.hive.service.cli.HiveSQLException: Error while processing statement:
FAILED: Execution Error, return code 2 from
org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask
at
org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:315)
at
org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:156)
at
org.apache.hive.service.cli.operation.SQLOperation.access$100(SQLOperation.java:71)
at
org.apache.hive.service.cli.operation.SQLOperation$1$1.run(SQLOperation.java:206)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
at
org.apache.hive.service.cli.operation.SQLOperation$1.run(SQLOperation.java:218)
at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
hive Configs can be viewed from
http://qelog.hortonworks.com/log/hs21-hbs24-1454382123/artifacts/tmpModifyConfDir_1454394513592/
Attachments
Drop files to attach, or browse.
Add LinkIssue Links
relates to
Bug - A problem which impairs or prevents the functions of the product.
HIVE-739 webhcat tests failing in HDInsight secure cluster throwing
NullPointerException Blocker - Blocks development and/or testing work,
production could not run. RESOLVED
Activity
All
Comments
Work Log
History
Activity
Ascending order - Click to sort in descending order
Permalink Edit Delete
tsaito Takahiko Saito added a comment - 3 days ago
The test passes via hive CLI and explain shows:
0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> explain select registration
0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> from s10k s join v10k v
0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> on (s.name = v.name) join
studentparttab30k p
0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> on (p.name = v.name)
0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> where s.age < 25 and v.age < 25
and p.age < 25;
+--------------------------------------------------------------------------------------------------------------+--+
| Explain
|
+--------------------------------------------------------------------------------------------------------------+--+
| STAGE DEPENDENCIES:
|
| Stage-5 is a root stage
|
| Stage-4 depends on stages: Stage-5
|
| Stage-0 depends on stages: Stage-4
|
|
|
| STAGE PLANS:
|
| Stage: Stage-5
|
| Map Reduce Local Work
|
| Alias -> Map Local Tables:
|
| s
|
| Fetch Operator
|
| limit: -1
|
| v
|
| Fetch Operator
|
| limit: -1
|
| Alias -> Map Local Operator Tree:
|
| s
|
| TableScan
|
| alias: s
|
| filterExpr: (name is not null and (age < 25)) (type: boolean)
|
| Statistics: Num rows: 459 Data size: 47777 Basic stats: COMPLETE
Column stats: NONE |
| Filter Operator
|
| predicate: (name is not null and (age < 25)) (type: boolean)
|
| Statistics: Num rows: 76 Data size: 7910 Basic stats: COMPLETE
Column stats: NONE |
| HashTable Sink Operator
|
| keys:
|
| 0 name (type: string)
|
| 1 name (type: string)
|
| 2 name (type: string)
|
| v
|
| TableScan
|
| alias: v
|
| filterExpr: (name is not null and (age < 25)) (type: boolean)
|
| Statistics: Num rows: 1653 Data size: 337233 Basic stats:
COMPLETE Column stats: NONE |
| Filter Operator
|
| predicate: (name is not null and (age < 25)) (type: boolean)
|
| Statistics: Num rows: 275 Data size: 56103 Basic stats:
COMPLETE Column stats: NONE |
| HashTable Sink Operator
|
| keys:
|
| 0 name (type: string)
|
| 1 name (type: string)
|
| 2 name (type: string)
|
|
|
| Stage: Stage-4
|
| Map Reduce
|
| Map Operator Tree:
|
| TableScan
|
| alias: p
|
| filterExpr: (name is not null and (age < 25)) (type: boolean)
|
| Statistics: Num rows: 30000 Data size: 627520 Basic stats:
COMPLETE Column stats: COMPLETE |
| Filter Operator
|
| predicate: (name is not null and (age < 25)) (type: boolean)
|
| Statistics: Num rows: 10000 Data size: 1010000 Basic stats:
COMPLETE Column stats: COMPLETE |
| Map Join Operator
|
| condition map:
|
| Inner Join 0 to 1
|
| Inner Join 1 to 2
|
| keys:
|
| 0 name (type: string)
|
| 1 name (type: string)
|
| 2 name (type: string)
|
| outputColumnNames: _col8
|
| Statistics: Num rows: 22000 Data size: 2222000 Basic stats:
COMPLETE Column stats: NONE |
| Select Operator
|
| expressions: _col8 (type: string)
|
| outputColumnNames: _col0
|
| Statistics: Num rows: 22000 Data size: 2222000 Basic stats:
COMPLETE Column stats: NONE |
| File Output Operator
|
| compressed: false
|
| Statistics: Num rows: 22000 Data size: 2222000 Basic
stats: COMPLETE Column stats: NONE |
| table:
|
| input format:
org.apache.hadoop.mapred.TextInputFormat |
| output format:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat |
| serde:
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe |
| Local Work:
|
| Map Reduce Local Work
|
|
|
| Stage: Stage-0
|
| Fetch Operator
|
| limit: -1
|
| Processor Tree:
|
| ListSink
|
|
|
+--------------------------------------------------------------------------------------------------------------+--+
83 rows selected (2.473 seconds)
Permalink Edit Delete
tsaito Takahiko Saito added a comment - 3 days ago
emptablemisc_8 also fails with the same error:
>>> create temporary table temp1 as select * from votertab10k;
>>> select *
from studenttab10k s
where s.name not in
(select name from temp1);
INFO : Number of reduce tasks determined at compile time: 1
INFO : In order to change the average load for a reducer (in bytes):
INFO : set hive.exec.reducers.bytes.per.reducer=<number>
INFO : In order to limit the maximum number of reducers:
INFO : set hive.exec.reducers.max=<number>
INFO : In order to set a constant number of reducers:
INFO : set mapreduce.job.reduces=<number>
INFO : number of splits:1
INFO : Submitting tokens for job: job_1454394534358_0164
INFO : Kind: HDFS_DELEGATION_TOKEN, Service: 10.0.0.36:8020, Ident:
(HDFS_DELEGATION_TOKEN token 246 for hrt_qa)
INFO : The url to track the job:
http://hn0-hs21-h.hdinsight.net:8088/proxy/application_1454394534358_0164/
INFO : Starting Job = job_1454394534358_0164, Tracking URL =
http://hn0-hs21-h.hdinsight.net:8088/proxy/application_1454394534358_0164/
INFO : Kill Command = /usr/hdp/2.4.1.0-170/hadoop/bin/hadoop job -kill
job_1454394534358_0164
INFO : Hadoop job information for Stage-4: number of mappers: 1; number of
reducers: 1
INFO : 2016-02-02 10:11:02,367 Stage-4 map = 0%, reduce = 0%
INFO : 2016-02-02 10:11:26,060 Stage-4 map = 100%, reduce = 0%, Cumulative
CPU 5.48 sec
INFO : 2016-02-02 10:11:39,024 Stage-4 map = 100%, reduce = 100%, Cumulative
CPU 11.18 sec
INFO : MapReduce Total cumulative CPU time: 11 seconds 180 msec
INFO : Ended Job = job_1454394534358_0164
INFO : Stage-9 is selected by condition resolver.
INFO : Stage-1 is filtered out by condition resolver.
ERROR : Execution failed with exit status: 2
ERROR : Obtaining error information
ERROR :
Task failed!
Task ID:
Stage-9
Logs:
ERROR : /var/log/hive/hiveServer2.log
INFO : Number of reduce tasks determined at compile time: 1
INFO : In order to change the average load for a reducer (in bytes):
INFO : set hive.exec.reducers.bytes.per.reducer=<number>
INFO : In order to limit the maximum number of reducers:
INFO : set hive.exec.reducers.max=<number>
INFO : In order to set a constant number of reducers:
INFO : set mapreduce.job.reduces=<number>
INFO : number of splits:2
INFO : Submitting tokens for job: job_1454394534358_0169
INFO : Kind: HDFS_DELEGATION_TOKEN, Service: 10.0.0.36:8020, Ident:
(HDFS_DELEGATION_TOKEN token 252 for hrt_qa)
INFO : The url to track the job:
http://hn0-hs21-h.hdinsight.net:8088/proxy/application_1454394534358_0169/
INFO : Starting Job = job_1454394534358_0169, Tracking URL =
http://hn0-hs21-h.hdinsight.net:8088/proxy/application_1454394534358_0169/
INFO : Kill Command = /usr/hdp/2.4.1.0-170/hadoop/bin/hadoop job -kill
job_1454394534358_0169
INFO : Hadoop job information for Stage-1: number of mappers: 2; number of
reducers: 1
INFO : 2016-02-02 10:16:38,027 Stage-1 map = 0%, reduce = 0%
INFO : 2016-02-02 10:16:52,498 Stage-1 map = 50%, reduce = 0%, Cumulative CPU
3.8 sec
INFO : 2016-02-02 10:16:53,566 Stage-1 map = 100%, reduce = 0%, Cumulative
CPU 9.94 sec
INFO : 2016-02-02 10:17:16,202 Stage-1 map = 100%, reduce = 100%, Cumulative
CPU 17.04 sec
INFO : MapReduce Total cumulative CPU time: 17 seconds 40 msec
INFO : Ended Job = job_1454394534358_0169
ERROR : Execution failed with exit status: 2
ERROR : Obtaining error information
ERROR :
Task failed!
Task ID:
Stage-8
Logs:
ERROR : /var/log/hive/hiveServer2.log
Error: Error while processing statement: FAILED: Execution Error, return code 2
from org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask (state=08S01,code=2)
Aborting command set because "force" is false and command failed: "select *
from studenttab10k s
where s.name not in
(select name from temp1);"
Closing: 0:
jdbc:hive2://zk2-hs21-h.hdinsight.net:2181/;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=hiveserver2;principal=hive/[email protected];transportMode=http;httpPath=cliservice
Its app log can be viewed at
http://qelog.hortonworks.com/log/hs21-hbs24-1454382123/app-logs/application_1454394534358_0169.log
Permalink Edit Delete
tsaito Takahiko Saito added a comment - 3 days ago
The query works without 'temporary table':
0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> select registration from
studenttab10k s join votertab10k v
0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> on (s.name = v.name) join
studentparttab30k p
0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> on (p.name = v.name)
0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> where s.age < 25 and v.age < 25
and p.age < 25;
Permalink Edit Delete
tsaito Takahiko Saito added a comment - 3 days ago
More info about temporary table:
0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> describe formatted s10k;
+-------------------------------+---------------------------------------------------------------------------------------------------------------------------------------------+-----------------------+--+
| col_name |
data_type
| comment |
+-------------------------------+---------------------------------------------------------------------------------------------------------------------------------------------+-----------------------+--+
| # col_name | data_type
| comment |
| | NULL
| NULL |
| name | string
| |
| age | int
| |
| gpa | double
| |
| | NULL
| NULL |
| # Detailed Table Information | NULL
| NULL |
| Database: | default
| NULL |
| Owner: | hrt_qa
| NULL |
| CreateTime: | Tue Feb 02 23:02:31 UTC 2016
| NULL |
| LastAccessTime: | UNKNOWN
| NULL |
| Protect Mode: | None
| NULL |
| Retention: | 0
| NULL |
| Location: |
hdfs://hn0-hs21-h.hdinsight.net:8020/tmp/hive/hive/de2667ea-5bc1-4548-a4c4-97e46d6081f8/_tmp_space.db/8e8482c9-a2a4-4e8c-ad0f-2cd6e8fcdb4c
| NULL |
| Table Type: | MANAGED_TABLE
| NULL |
| | NULL
| NULL |
| # Storage Information | NULL
| NULL |
| SerDe Library: | org.apache.hadoop.hive.ql.io.orc.OrcSerde
| NULL |
| InputFormat: |
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
| NULL
|
| OutputFormat: |
org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
| NULL
|
| Compressed: | No
| NULL |
| Num Buckets: | -1
| NULL |
| Bucket Columns: | []
| NULL |
| Sort Columns: | []
| NULL |
| Storage Desc Params: | NULL
| NULL |
| | serialization.format
| 1 |
+-------------------------------+---------------------------------------------------------------------------------------------------------------------------------------------+-----------------------+--+
26 rows selected (0.22 seconds)
0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> set hive.server2.enable.doAs;
+---------------------------------+--+
| set |
+---------------------------------+--+
| hive.server2.enable.doAs=false |
+---------------------------------+--+
Permalink Edit Delete
tsaito Takahiko Saito added a comment - 3 days ago
hdfs dir of temporary table is owned by hive as expected since
hive.server2.enable.doAs=false:
hdfs@hn0-hs21-h:~$ hdfs dfs -ls
hdfs://hn0-hs21-h.hdinsight.net:8020/tmp/hive/hive/de2667ea-5bc1-4548-a4c4-97e46d6081f8/_tmp_space.db/8e8482c9-a2a4-4e8c-ad0f-2cd6e8fcdb4c
Found 1 items
-rwx------ 3 hive hdfs 47777 2016-02-02 23:02
hdfs://hn0-hs21-h.hdinsight.net:8020/tmp/hive/hive/de2667ea-5bc1-4548-a4c4-97e46d6081f8/_tmp_space.db/8e8482c9-a2a4-4e8c-ad0f-2cd6e8fcdb4c/000000_0
Not sure if it's related to the JIRA, but one thing I noticed was that dfs cmd
throws error via beeline:
0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> dfs -ls
hdfs://hn0-hs21-h.hdinsight.net:8020/tmp/hive/hive/de2667ea-5bc1-4548-a4c4-97e46d6081f8/_tmp_space.db/8e8482c9-a2a4-4e8c-ad0f-2cd6e8fcdb4c;
Error: Error while processing statement: Permission denied: user [hrt_qa] does
not have privilege for [DFS] command (state=,code=1)
Permalink Edit
thejas Thejas Nair added a comment - 5 hours ago - edited
Takahiko Saito
The above DFS error would be from Ranger of SQL standard authorization. It is
not related to the test failure.
But you are right that it looks like a permission issue.
The issue seen in HIVE-739 also exists in this part of Hive. Hive in MR mode
launches a new child process to process the small table and create a hash
table. This child process needs credentials from HDFS. However, in this setup,
azure fs is the default file system. We should also get delegation token from
all FS URIs listed under mapreduce.job.hdfs-servers config.
There are hueristics around when map-join gets used. That is why you don't see
it unless temp table is used (that uses ORC format and might have stats as
well, while the original table is probaly text format).
https://github.com/hortonworks/hive/blob/2.4-maint/ql/src/java/org/apache/hadoop/hive/ql/exec/SecureCmdDoAs.java
needs change similar to webhcat change in HIVE-739
Permalink Edit
sush Sushanth Sowmyan added a comment - 7 minutes ago
Takahiko Saito, we should file an apache jira for this as well. Would you like
to do so, so the bug report has attribution to you?
Permalink Edit Delete
tsaito Takahiko Saito added a comment - 4 minutes ago
Sushanth Sowmyan I will file one and update with that JIRA.
Comment
People
Assignee: sush Sushanth Sowmyan
Assign to me
Reporter: tsaito Takahiko Saito
QEAssignee: Takahiko Saito
Votes: 0
Watchers: 3 Stop watching this issue
Dates
Created: 3 days ago
Updated: 4 minutes ago
Who's Looking?
Agile
View on Board
{noformat}
hivesever2 log shows:
{noformat}
2016-02-02 18:04:43,766 ERROR [HiveServer2-Background-Pool: Thread-517]:
operation.Operation (SQLOperation.java:run(209)) - Error running hive query:
org.apache.hive.service.cli.HiveSQLException: Error while processing statement:
FAILED: Execution Error, return code 2 from
org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask
at
org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:315)
at
org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:156)
at
org.apache.hive.service.cli.operation.SQLOperation.access$100(SQLOperation.java:71)
at
org.apache.hive.service.cli.operation.SQLOperation$1$1.run(SQLOperation.java:206)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
at
org.apache.hive.service.cli.operation.SQLOperation$1.run(SQLOperation.java:218)
at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
{noformat}
was:
The following query fails:
{noformat}
>>> create temporary table s10k stored as orc as select * from studenttab10k;
>>> create temporary table v10k as select * from votertab10k;
>>> select registration
from s10k s join v10k v
on (s.name = v.name) join studentparttab30k p
on (p.name = v.name)
where s.age < 25 and v.age < 25 and p.age < 25;
ERROR : Execution failed with exit status: 2
ERROR : Obtaining error information
ERROR :
Task failed!
Task ID:
Stage-5
Logs:
ERROR : /var/log/hive/hiveServer2.log
Error: Error while processing statement: FAILED: Execution Error, return code 2
from org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask (state=08S01,code=2)
Aborting command set because "force" is false and command failed: "select
registration
from s10k s join v10k v
on (s.name = v.name) join studentparttab30k p
on (p.name = v.name)
where s.age < 25 and v.age < 25 and p.age < 25;"
Closing: 0:
jdbc:hive2://zk2-hs21-h.hdinsight.net:2181/;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=hiveserver2;principal=hive/[email protected];transportMode=http;httpPath=cliservice
hiveServer2.log shows:
2016-02-02 18:04:34,182 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG method=Driver.run
from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:34,199 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
method=TimeToSubmit from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:34,212 INFO [HiveServer2-HttpHandler-Pool: Thread-55]:
thrift.ThriftHttpServlet (ThriftHttpServlet.java:doPost(127)) - Could not
validate cookie sent, will try to generate a new cookie
2016-02-02 18:04:34,213 INFO [HiveServer2-Background-Pool: Thread-517]:
ql.Driver (Driver.java:checkConcurrency(168)) - Concurrency mode is disabled,
not creating a lock manager
2016-02-02 18:04:34,219 INFO [HiveServer2-HttpHandler-Pool: Thread-55]:
thrift.ThriftHttpServlet (ThriftHttpServlet.java:doKerberosAuth(352)) - Failed
to authenticate with http/_HOST kerberos principal, trying with hive/_HOST
kerberos principal
2016-02-02 18:04:34,219 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
method=Driver.execute from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:34,225 INFO [HiveServer2-Background-Pool: Thread-517]:
ql.Driver (Driver.java:execute(1390)) - Setting caller context to query id
hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0
2016-02-02 18:04:34,226 INFO [HiveServer2-Background-Pool: Thread-517]:
ql.Driver (Driver.java:execute(1393)) - Starting
command(queryId=hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0):
select registration
from s10k s join v10k v
on (s.name = v.name) join studentparttab30k p
on (p.name = v.name)
where s.age < 25 and v.age < 25 and p.age < 25
2016-02-02 18:04:34,228 INFO [HiveServer2-Background-Pool: Thread-517]:
hooks.ATSHook (ATSHook.java:<init>(90)) - Created ATS Hook
2016-02-02 18:04:34,229 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
method=PreHook.org.apache.hadoop.hive.ql.hooks.ATSHook
from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:34,237 INFO [HiveServer2-HttpHandler-Pool: Thread-55]:
thrift.ThriftHttpServlet (ThriftHttpServlet.java:doPost(169)) - Cookie added
for clientUserName hrt_qa
2016-02-02 18:04:34,238 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
method=PreHook.org.apache.hadoop.hive.ql.hooks.ATSHook start=1454436274229
end=1454436274238 duration=9 from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:34,239 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
method=PreHook.org.apache.hadoop.hive.ql.security.authorization.plugin.DisallowTransformHook
from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:34,240 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
method=PreHook.org.apache.hadoop.hive.ql.security.authorization.plugin.DisallowTransformHook
start=1454436274239 end=1454436274240 duration=1
from=org.apache.hadoop.hive.ql.Driver>
Query ID = hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0
2016-02-02 18:04:34,242 INFO [HiveServer2-Background-Pool: Thread-517]:
ql.Driver (SessionState.java:printInfo(923)) - Query ID =
hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0
Total jobs = 1
2016-02-02 18:04:34,243 INFO [HiveServer2-Background-Pool: Thread-517]:
ql.Driver (SessionState.java:printInfo(923)) - Total jobs = 1
2016-02-02 18:04:34,245 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
method=TimeToSubmit start=1454436274199 end=1454436274245 duration=46
from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:34,246 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG method=runTasks
from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:34,247 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
method=task.MAPREDLOCAL.Stage-5 from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:34,258 INFO [HiveServer2-Background-Pool: Thread-517]:
ql.Driver (Driver.java:launchTask(1718)) - Starting task [Stage-5:MAPREDLOCAL]
in serial mode
2016-02-02 18:04:34,280 INFO [HiveServer2-Background-Pool: Thread-517]:
mr.MapredLocalTask (MapredLocalTask.java:executeInChildVM(158)) - Generating
plan file
file:/tmp/hive/916e3dbb-a10d-4888-a063-52fb058ea421/hive_2016-02-02_18-04-29_153_625340340820843828-4/-local-10006/plan.xml
2016-02-02 18:04:34,288 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
method=serializePlan from=org.apache.hadoop.hive.ql.exec.Utilities>
2016-02-02 18:04:34,289 INFO [HiveServer2-Background-Pool: Thread-517]:
exec.Utilities (Utilities.java:serializePlan(1028)) - Serializing
MapredLocalWork via kryo
2016-02-02 18:04:34,290 INFO [ATS Logger 0]: hooks.ATSHook
(ATSHook.java:createPreHookEvent(158)) - Received pre-hook notification for
:hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0
2016-02-02 18:04:34,358 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
method=serializePlan start=1454436274288 end=1454436274358 duration=70
from=org.apache.hadoop.hive.ql.exec.Utilities>
2016-02-02 18:04:34,737 INFO [HiveServer2-Background-Pool: Thread-517]:
mr.MapredLocalTask (MapredLocalTask.java:executeInChildVM(287)) - Executing:
/usr/hdp/2.4.1.0-170/hadoop/bin/hadoop jar
/usr/hdp/2.4.1.0-170/hive/lib/hive-common-1.2.1000.2.4.1.0-170.jar
org.apache.hadoop.hive.ql.exec.mr.ExecDriver -localtask -plan
file:/tmp/hive/916e3dbb-a10d-4888-a063-52fb058ea421/hive_2016-02-02_18-04-29_153_625340340820843828-4/-local-10006/plan.xml
-jobconffile
file:/tmp/hive/916e3dbb-a10d-4888-a063-52fb058ea421/hive_2016-02-02_18-04-29_153_625340340820843828-4/-local-10007/jobconf.xml
WARNING: Use "yarn jar" to launch YARN applications.
2016-02-02 18:04:37,450 INFO [org.apache.ranger.audit.queue.AuditBatchQueue0]:
provider.BaseAuditHandler (BaseAuditHandler.java:logStatus(312)) - Audit Status
Log: name=hiveServer2.async.summary.batch.solr, interval=01:21.012 minutes,
events=2, succcessCount=2, totalEvents=4, totalSuccessCount=4
Execution log at:
/tmp/hive/hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0.log
2016-02-02 18:04:39,248 INFO [HiveServer2-HttpHandler-Pool: Thread-55]:
thrift.ThriftHttpServlet (ThriftHttpServlet.java:doPost(127)) - Could not
validate cookie sent, will try to generate a new cookie
2016-02-02 18:04:39,254 INFO [HiveServer2-HttpHandler-Pool: Thread-55]:
thrift.ThriftHttpServlet (ThriftHttpServlet.java:doKerberosAuth(352)) - Failed
to authenticate with http/_HOST kerberos principal, trying with hive/_HOST
kerberos principal
2016-02-02 18:04:39,261 INFO [HiveServer2-HttpHandler-Pool: Thread-55]:
thrift.ThriftHttpServlet (ThriftHttpServlet.java:doPost(169)) - Cookie added
for clientUserName hrt_qa
2016-02-02 18:04:40 Starting to launch local task to process map join;
maximum memory = 477102080
Execution failed with exit status: 2
2016-02-02 18:04:43,728 ERROR [HiveServer2-Background-Pool: Thread-517]:
exec.Task (SessionState.java:printError(932)) - Execution failed with exit
status: 2
Obtaining error information
2016-02-02 18:04:43,730 ERROR [HiveServer2-Background-Pool: Thread-517]:
exec.Task (SessionState.java:printError(932)) - Obtaining error information
Task failed!
Task ID:
Stage-5
Logs:
2016-02-02 18:04:43,730 ERROR [HiveServer2-Background-Pool: Thread-517]:
exec.Task (SessionState.java:printError(932)) -
Task failed!
Task ID:
Stage-5
Logs:
/var/log/hive/hiveServer2.log
2016-02-02 18:04:43,732 ERROR [HiveServer2-Background-Pool: Thread-517]:
exec.Task (SessionState.java:printError(932)) - /var/log/hive/hiveServer2.log
2016-02-02 18:04:43,732 ERROR [HiveServer2-Background-Pool: Thread-517]:
mr.MapredLocalTask (MapredLocalTask.java:executeInChildVM(307)) - Execution
failed with exit status: 2
2016-02-02 18:04:43,733 INFO [HiveServer2-Background-Pool: Thread-517]:
hooks.ATSHook (ATSHook.java:<init>(90)) - Created ATS Hook
2016-02-02 18:04:43,734 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
method=FailureHook.org.apache.hadoop.hive.ql.hooks.ATSHook
from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:43,736 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
method=FailureHook.org.apache.hadoop.hive.ql.hooks.ATSHook start=1454436283734
end=1454436283736 duration=2 from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:43,736 INFO [ATS Logger 0]: hooks.ATSHook
(ATSHook.java:createPostHookEvent(193)) - Received post-hook notification for
:hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0
FAILED: Execution Error, return code 2 from
org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask
2016-02-02 18:04:43,757 ERROR [HiveServer2-Background-Pool: Thread-517]:
ql.Driver (SessionState.java:printError(932)) - FAILED: Execution Error, return
code 2 from org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask
2016-02-02 18:04:43,758 INFO [HiveServer2-Background-Pool: Thread-517]:
ql.Driver (Driver.java:execute(1621)) - Resetting the caller context to
HIVE_SSN_ID:916e3dbb-a10d-4888-a063-52fb058ea421
2016-02-02 18:04:43,759 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
method=Driver.execute start=1454436274219 end=1454436283759 duration=9540
from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:43,760 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
method=releaseLocks from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:43,761 INFO [HiveServer2-Background-Pool: Thread-517]:
log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
method=releaseLocks start=1454436283760 end=1454436283761 duration=1
from=org.apache.hadoop.hive.ql.Driver>
2016-02-02 18:04:43,766 ERROR [HiveServer2-Background-Pool: Thread-517]:
operation.Operation (SQLOperation.java:run(209)) - Error running hive query:
org.apache.hive.service.cli.HiveSQLException: Error while processing statement:
FAILED: Execution Error, return code 2 from
org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask
at
org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:315)
at
org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:156)
at
org.apache.hive.service.cli.operation.SQLOperation.access$100(SQLOperation.java:71)
at
org.apache.hive.service.cli.operation.SQLOperation$1$1.run(SQLOperation.java:206)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
at
org.apache.hive.service.cli.operation.SQLOperation$1.run(SQLOperation.java:218)
at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
hive Configs can be viewed from
http://qelog.hortonworks.com/log/hs21-hbs24-1454382123/artifacts/tmpModifyConfDir_1454394513592/
Attachments
Drop files to attach, or browse.
Add LinkIssue Links
relates to
Bug - A problem which impairs or prevents the functions of the product.
HIVE-739 webhcat tests failing in HDInsight secure cluster throwing
NullPointerException Blocker - Blocks development and/or testing work,
production could not run. RESOLVED
Activity
All
Comments
Work Log
History
Activity
Ascending order - Click to sort in descending order
Permalink Edit Delete
tsaito Takahiko Saito added a comment - 3 days ago
The test passes via hive CLI and explain shows:
0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> explain select registration
0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> from s10k s join v10k v
0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> on (s.name = v.name) join
studentparttab30k p
0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> on (p.name = v.name)
0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> where s.age < 25 and v.age < 25
and p.age < 25;
+--------------------------------------------------------------------------------------------------------------+--+
| Explain
|
+--------------------------------------------------------------------------------------------------------------+--+
| STAGE DEPENDENCIES:
|
| Stage-5 is a root stage
|
| Stage-4 depends on stages: Stage-5
|
| Stage-0 depends on stages: Stage-4
|
|
|
| STAGE PLANS:
|
| Stage: Stage-5
|
| Map Reduce Local Work
|
| Alias -> Map Local Tables:
|
| s
|
| Fetch Operator
|
| limit: -1
|
| v
|
| Fetch Operator
|
| limit: -1
|
| Alias -> Map Local Operator Tree:
|
| s
|
| TableScan
|
| alias: s
|
| filterExpr: (name is not null and (age < 25)) (type: boolean)
|
| Statistics: Num rows: 459 Data size: 47777 Basic stats: COMPLETE
Column stats: NONE |
| Filter Operator
|
| predicate: (name is not null and (age < 25)) (type: boolean)
|
| Statistics: Num rows: 76 Data size: 7910 Basic stats: COMPLETE
Column stats: NONE |
| HashTable Sink Operator
|
| keys:
|
| 0 name (type: string)
|
| 1 name (type: string)
|
| 2 name (type: string)
|
| v
|
| TableScan
|
| alias: v
|
| filterExpr: (name is not null and (age < 25)) (type: boolean)
|
| Statistics: Num rows: 1653 Data size: 337233 Basic stats:
COMPLETE Column stats: NONE |
| Filter Operator
|
| predicate: (name is not null and (age < 25)) (type: boolean)
|
| Statistics: Num rows: 275 Data size: 56103 Basic stats:
COMPLETE Column stats: NONE |
| HashTable Sink Operator
|
| keys:
|
| 0 name (type: string)
|
| 1 name (type: string)
|
| 2 name (type: string)
|
|
|
| Stage: Stage-4
|
| Map Reduce
|
| Map Operator Tree:
|
| TableScan
|
| alias: p
|
| filterExpr: (name is not null and (age < 25)) (type: boolean)
|
| Statistics: Num rows: 30000 Data size: 627520 Basic stats:
COMPLETE Column stats: COMPLETE |
| Filter Operator
|
| predicate: (name is not null and (age < 25)) (type: boolean)
|
| Statistics: Num rows: 10000 Data size: 1010000 Basic stats:
COMPLETE Column stats: COMPLETE |
| Map Join Operator
|
| condition map:
|
| Inner Join 0 to 1
|
| Inner Join 1 to 2
|
| keys:
|
| 0 name (type: string)
|
| 1 name (type: string)
|
| 2 name (type: string)
|
| outputColumnNames: _col8
|
| Statistics: Num rows: 22000 Data size: 2222000 Basic stats:
COMPLETE Column stats: NONE |
| Select Operator
|
| expressions: _col8 (type: string)
|
| outputColumnNames: _col0
|
| Statistics: Num rows: 22000 Data size: 2222000 Basic stats:
COMPLETE Column stats: NONE |
| File Output Operator
|
| compressed: false
|
| Statistics: Num rows: 22000 Data size: 2222000 Basic
stats: COMPLETE Column stats: NONE |
| table:
|
| input format:
org.apache.hadoop.mapred.TextInputFormat |
| output format:
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat |
| serde:
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe |
| Local Work:
|
| Map Reduce Local Work
|
|
|
| Stage: Stage-0
|
| Fetch Operator
|
| limit: -1
|
| Processor Tree:
|
| ListSink
|
|
|
+--------------------------------------------------------------------------------------------------------------+--+
83 rows selected (2.473 seconds)
Permalink Edit Delete
tsaito Takahiko Saito added a comment - 3 days ago
emptablemisc_8 also fails with the same error:
>>> create temporary table temp1 as select * from votertab10k;
>>> select *
from studenttab10k s
where s.name not in
(select name from temp1);
INFO : Number of reduce tasks determined at compile time: 1
INFO : In order to change the average load for a reducer (in bytes):
INFO : set hive.exec.reducers.bytes.per.reducer=<number>
INFO : In order to limit the maximum number of reducers:
INFO : set hive.exec.reducers.max=<number>
INFO : In order to set a constant number of reducers:
INFO : set mapreduce.job.reduces=<number>
INFO : number of splits:1
INFO : Submitting tokens for job: job_1454394534358_0164
INFO : Kind: HDFS_DELEGATION_TOKEN, Service: 10.0.0.36:8020, Ident:
(HDFS_DELEGATION_TOKEN token 246 for hrt_qa)
INFO : The url to track the job:
http://hn0-hs21-h.hdinsight.net:8088/proxy/application_1454394534358_0164/
INFO : Starting Job = job_1454394534358_0164, Tracking URL =
http://hn0-hs21-h.hdinsight.net:8088/proxy/application_1454394534358_0164/
INFO : Kill Command = /usr/hdp/2.4.1.0-170/hadoop/bin/hadoop job -kill
job_1454394534358_0164
INFO : Hadoop job information for Stage-4: number of mappers: 1; number of
reducers: 1
INFO : 2016-02-02 10:11:02,367 Stage-4 map = 0%, reduce = 0%
INFO : 2016-02-02 10:11:26,060 Stage-4 map = 100%, reduce = 0%, Cumulative
CPU 5.48 sec
INFO : 2016-02-02 10:11:39,024 Stage-4 map = 100%, reduce = 100%, Cumulative
CPU 11.18 sec
INFO : MapReduce Total cumulative CPU time: 11 seconds 180 msec
INFO : Ended Job = job_1454394534358_0164
INFO : Stage-9 is selected by condition resolver.
INFO : Stage-1 is filtered out by condition resolver.
ERROR : Execution failed with exit status: 2
ERROR : Obtaining error information
ERROR :
Task failed!
Task ID:
Stage-9
Logs:
ERROR : /var/log/hive/hiveServer2.log
INFO : Number of reduce tasks determined at compile time: 1
INFO : In order to change the average load for a reducer (in bytes):
INFO : set hive.exec.reducers.bytes.per.reducer=<number>
INFO : In order to limit the maximum number of reducers:
INFO : set hive.exec.reducers.max=<number>
INFO : In order to set a constant number of reducers:
INFO : set mapreduce.job.reduces=<number>
INFO : number of splits:2
INFO : Submitting tokens for job: job_1454394534358_0169
INFO : Kind: HDFS_DELEGATION_TOKEN, Service: 10.0.0.36:8020, Ident:
(HDFS_DELEGATION_TOKEN token 252 for hrt_qa)
INFO : The url to track the job:
http://hn0-hs21-h.hdinsight.net:8088/proxy/application_1454394534358_0169/
INFO : Starting Job = job_1454394534358_0169, Tracking URL =
http://hn0-hs21-h.hdinsight.net:8088/proxy/application_1454394534358_0169/
INFO : Kill Command = /usr/hdp/2.4.1.0-170/hadoop/bin/hadoop job -kill
job_1454394534358_0169
INFO : Hadoop job information for Stage-1: number of mappers: 2; number of
reducers: 1
INFO : 2016-02-02 10:16:38,027 Stage-1 map = 0%, reduce = 0%
INFO : 2016-02-02 10:16:52,498 Stage-1 map = 50%, reduce = 0%, Cumulative CPU
3.8 sec
INFO : 2016-02-02 10:16:53,566 Stage-1 map = 100%, reduce = 0%, Cumulative
CPU 9.94 sec
INFO : 2016-02-02 10:17:16,202 Stage-1 map = 100%, reduce = 100%, Cumulative
CPU 17.04 sec
INFO : MapReduce Total cumulative CPU time: 17 seconds 40 msec
INFO : Ended Job = job_1454394534358_0169
ERROR : Execution failed with exit status: 2
ERROR : Obtaining error information
ERROR :
Task failed!
Task ID:
Stage-8
Logs:
ERROR : /var/log/hive/hiveServer2.log
Error: Error while processing statement: FAILED: Execution Error, return code 2
from org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask (state=08S01,code=2)
Aborting command set because "force" is false and command failed: "select *
from studenttab10k s
where s.name not in
(select name from temp1);"
Closing: 0:
jdbc:hive2://zk2-hs21-h.hdinsight.net:2181/;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=hiveserver2;principal=hive/[email protected];transportMode=http;httpPath=cliservice
Its app log can be viewed at
http://qelog.hortonworks.com/log/hs21-hbs24-1454382123/app-logs/application_1454394534358_0169.log
Permalink Edit Delete
tsaito Takahiko Saito added a comment - 3 days ago
The query works without 'temporary table':
0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> select registration from
studenttab10k s join votertab10k v
0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> on (s.name = v.name) join
studentparttab30k p
0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> on (p.name = v.name)
0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> where s.age < 25 and v.age < 25
and p.age < 25;
Permalink Edit Delete
tsaito Takahiko Saito added a comment - 3 days ago
More info about temporary table:
0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> describe formatted s10k;
+-------------------------------+---------------------------------------------------------------------------------------------------------------------------------------------+-----------------------+--+
| col_name |
data_type
| comment |
+-------------------------------+---------------------------------------------------------------------------------------------------------------------------------------------+-----------------------+--+
| # col_name | data_type
| comment |
| | NULL
| NULL |
| name | string
| |
| age | int
| |
| gpa | double
| |
| | NULL
| NULL |
| # Detailed Table Information | NULL
| NULL |
| Database: | default
| NULL |
| Owner: | hrt_qa
| NULL |
| CreateTime: | Tue Feb 02 23:02:31 UTC 2016
| NULL |
| LastAccessTime: | UNKNOWN
| NULL |
| Protect Mode: | None
| NULL |
| Retention: | 0
| NULL |
| Location: |
hdfs://hn0-hs21-h.hdinsight.net:8020/tmp/hive/hive/de2667ea-5bc1-4548-a4c4-97e46d6081f8/_tmp_space.db/8e8482c9-a2a4-4e8c-ad0f-2cd6e8fcdb4c
| NULL |
| Table Type: | MANAGED_TABLE
| NULL |
| | NULL
| NULL |
| # Storage Information | NULL
| NULL |
| SerDe Library: | org.apache.hadoop.hive.ql.io.orc.OrcSerde
| NULL |
| InputFormat: |
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
| NULL
|
| OutputFormat: |
org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
| NULL
|
| Compressed: | No
| NULL |
| Num Buckets: | -1
| NULL |
| Bucket Columns: | []
| NULL |
| Sort Columns: | []
| NULL |
| Storage Desc Params: | NULL
| NULL |
| | serialization.format
| 1 |
+-------------------------------+---------------------------------------------------------------------------------------------------------------------------------------------+-----------------------+--+
26 rows selected (0.22 seconds)
0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> set hive.server2.enable.doAs;
+---------------------------------+--+
| set |
+---------------------------------+--+
| hive.server2.enable.doAs=false |
+---------------------------------+--+
Permalink Edit Delete
tsaito Takahiko Saito added a comment - 3 days ago
hdfs dir of temporary table is owned by hive as expected since
hive.server2.enable.doAs=false:
hdfs@hn0-hs21-h:~$ hdfs dfs -ls
hdfs://hn0-hs21-h.hdinsight.net:8020/tmp/hive/hive/de2667ea-5bc1-4548-a4c4-97e46d6081f8/_tmp_space.db/8e8482c9-a2a4-4e8c-ad0f-2cd6e8fcdb4c
Found 1 items
-rwx------ 3 hive hdfs 47777 2016-02-02 23:02
hdfs://hn0-hs21-h.hdinsight.net:8020/tmp/hive/hive/de2667ea-5bc1-4548-a4c4-97e46d6081f8/_tmp_space.db/8e8482c9-a2a4-4e8c-ad0f-2cd6e8fcdb4c/000000_0
Not sure if it's related to the JIRA, but one thing I noticed was that dfs cmd
throws error via beeline:
0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> dfs -ls
hdfs://hn0-hs21-h.hdinsight.net:8020/tmp/hive/hive/de2667ea-5bc1-4548-a4c4-97e46d6081f8/_tmp_space.db/8e8482c9-a2a4-4e8c-ad0f-2cd6e8fcdb4c;
Error: Error while processing statement: Permission denied: user [hrt_qa] does
not have privilege for [DFS] command (state=,code=1)
Permalink Edit
thejas Thejas Nair added a comment - 5 hours ago - edited
Takahiko Saito
The above DFS error would be from Ranger of SQL standard authorization. It is
not related to the test failure.
But you are right that it looks like a permission issue.
The issue seen in HIVE-739 also exists in this part of Hive. Hive in MR mode
launches a new child process to process the small table and create a hash
table. This child process needs credentials from HDFS. However, in this setup,
azure fs is the default file system. We should also get delegation token from
all FS URIs listed under mapreduce.job.hdfs-servers config.
There are hueristics around when map-join gets used. That is why you don't see
it unless temp table is used (that uses ORC format and might have stats as
well, while the original table is probaly text format).
https://github.com/hortonworks/hive/blob/2.4-maint/ql/src/java/org/apache/hadoop/hive/ql/exec/SecureCmdDoAs.java
needs change similar to webhcat change in HIVE-739
Permalink Edit
sush Sushanth Sowmyan added a comment - 7 minutes ago
Takahiko Saito, we should file an apache jira for this as well. Would you like
to do so, so the bug report has attribution to you?
Permalink Edit Delete
tsaito Takahiko Saito added a comment - 4 minutes ago
Sushanth Sowmyan I will file one and update with that JIRA.
Comment
People
Assignee: sush Sushanth Sowmyan
Assign to me
Reporter: tsaito Takahiko Saito
QEAssignee: Takahiko Saito
Votes: 0
Watchers: 3 Stop watching this issue
Dates
Created: 3 days ago
Updated: 4 minutes ago
Who's Looking?
Agile
View on Board
{noformat}
hivesever2 log shows:
{noformat}
2016-02-02 18:04:43,766 ERROR [HiveServer2-Background-Pool: Thread-517]:
operation.Operation (SQLOperation.java:run(209)) - Error running hive query:
org.apache.hive.service.cli.HiveSQLException: Error while processing statement:
FAILED: Execution Error, return code 2 from
org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask
at
org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:315)
at
org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:156)
at
org.apache.hive.service.cli.operation.SQLOperation.access$100(SQLOperation.java:71)
at
org.apache.hive.service.cli.operation.SQLOperation$1$1.run(SQLOperation.java:206)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
at
org.apache.hive.service.cli.operation.SQLOperation$1.run(SQLOperation.java:218)
at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
{noformat}
> Child process of HiveServer2 fails to get delegation token from non default
> FileSystem
> --------------------------------------------------------------------------------------
>
> Key: HIVE-13017
> URL: https://issues.apache.org/jira/browse/HIVE-13017
> Project: Hive
> Issue Type: Bug
> Components: Authentication
> Affects Versions: 1.2.1
> Environment: Secure
> Reporter: Takahiko Saito
> Assignee: Sushanth Sowmyan
>
> The following query fails, when Azure Filesystem is used as default file
> system, and HDFS is used for intermediate data.
> {noformat}
> >>> create temporary table s10k stored as orc as select * from studenttab10k;
> >>> create temporary table v10k as select * from votertab10k;
> >>> select registration
> from s10k s join v10k v
> on (s.name = v.name) join studentparttab30k p
> on (p.name = v.name)
> where s.age < 25 and v.age < 25 and p.age < 25;
> ERROR : Execution failed with exit status: 2
> ERROR : Obtaining error information
> ERROR :
> Task failed!
> Task ID:
> Stage-5
> Logs:
> ERROR : /var/log/hive/hiveServer2.log
> Error: Error while processing statement: FAILED: Execution Error, return code
> 2 from org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask (state=08S01,code=2)
> Aborting command set because "force" is false and command failed: "select
> registration
> from s10k s join v10k v
> on (s.name = v.name) join studentparttab30k p
> on (p.name = v.name)
> where s.age < 25 and v.age < 25 and p.age < 25;"
> Closing: 0:
> jdbc:hive2://zk2-hs21-h.hdinsight.net:2181/;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=hiveserver2;principal=hive/[email protected];transportMode=http;httpPath=cliservice
> hiveServer2.log shows:
> 2016-02-02 18:04:34,182 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
> method=Driver.run from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:34,199 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
> method=TimeToSubmit from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:34,212 INFO [HiveServer2-HttpHandler-Pool: Thread-55]:
> thrift.ThriftHttpServlet (ThriftHttpServlet.java:doPost(127)) - Could not
> validate cookie sent, will try to generate a new cookie
> 2016-02-02 18:04:34,213 INFO [HiveServer2-Background-Pool: Thread-517]:
> ql.Driver (Driver.java:checkConcurrency(168)) - Concurrency mode is disabled,
> not creating a lock manager
> 2016-02-02 18:04:34,219 INFO [HiveServer2-HttpHandler-Pool: Thread-55]:
> thrift.ThriftHttpServlet (ThriftHttpServlet.java:doKerberosAuth(352)) -
> Failed to authenticate with http/_HOST kerberos principal, trying with
> hive/_HOST kerberos principal
> 2016-02-02 18:04:34,219 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
> method=Driver.execute from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:34,225 INFO [HiveServer2-Background-Pool: Thread-517]:
> ql.Driver (Driver.java:execute(1390)) - Setting caller context to query id
> hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0
> 2016-02-02 18:04:34,226 INFO [HiveServer2-Background-Pool: Thread-517]:
> ql.Driver (Driver.java:execute(1393)) - Starting
> command(queryId=hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0):
> select registration
> from s10k s join v10k v
> on (s.name = v.name) join studentparttab30k p
> on (p.name = v.name)
> where s.age < 25 and v.age < 25 and p.age < 25
> 2016-02-02 18:04:34,228 INFO [HiveServer2-Background-Pool: Thread-517]:
> hooks.ATSHook (ATSHook.java:<init>(90)) - Created ATS Hook
> 2016-02-02 18:04:34,229 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
> method=PreHook.org.apache.hadoop.hive.ql.hooks.ATSHook
> from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:34,237 INFO [HiveServer2-HttpHandler-Pool: Thread-55]:
> thrift.ThriftHttpServlet (ThriftHttpServlet.java:doPost(169)) - Cookie added
> for clientUserName hrt_qa
> 2016-02-02 18:04:34,238 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
> method=PreHook.org.apache.hadoop.hive.ql.hooks.ATSHook start=1454436274229
> end=1454436274238 duration=9 from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:34,239 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
> method=PreHook.org.apache.hadoop.hive.ql.security.authorization.plugin.DisallowTransformHook
> from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:34,240 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
> method=PreHook.org.apache.hadoop.hive.ql.security.authorization.plugin.DisallowTransformHook
> start=1454436274239 end=1454436274240 duration=1
> from=org.apache.hadoop.hive.ql.Driver>
> Query ID = hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0
> 2016-02-02 18:04:34,242 INFO [HiveServer2-Background-Pool: Thread-517]:
> ql.Driver (SessionState.java:printInfo(923)) - Query ID =
> hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0
> Total jobs = 1
> 2016-02-02 18:04:34,243 INFO [HiveServer2-Background-Pool: Thread-517]:
> ql.Driver (SessionState.java:printInfo(923)) - Total jobs = 1
> 2016-02-02 18:04:34,245 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
> method=TimeToSubmit start=1454436274199 end=1454436274245 duration=46
> from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:34,246 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG method=runTasks
> from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:34,247 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
> method=task.MAPREDLOCAL.Stage-5 from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:34,258 INFO [HiveServer2-Background-Pool: Thread-517]:
> ql.Driver (Driver.java:launchTask(1718)) - Starting task
> [Stage-5:MAPREDLOCAL] in serial mode
> 2016-02-02 18:04:34,280 INFO [HiveServer2-Background-Pool: Thread-517]:
> mr.MapredLocalTask (MapredLocalTask.java:executeInChildVM(158)) - Generating
> plan file
> file:/tmp/hive/916e3dbb-a10d-4888-a063-52fb058ea421/hive_2016-02-02_18-04-29_153_625340340820843828-4/-local-10006/plan.xml
> 2016-02-02 18:04:34,288 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
> method=serializePlan from=org.apache.hadoop.hive.ql.exec.Utilities>
> 2016-02-02 18:04:34,289 INFO [HiveServer2-Background-Pool: Thread-517]:
> exec.Utilities (Utilities.java:serializePlan(1028)) - Serializing
> MapredLocalWork via kryo
> 2016-02-02 18:04:34,290 INFO [ATS Logger 0]: hooks.ATSHook
> (ATSHook.java:createPreHookEvent(158)) - Received pre-hook notification for
> :hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0
> 2016-02-02 18:04:34,358 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
> method=serializePlan start=1454436274288 end=1454436274358 duration=70
> from=org.apache.hadoop.hive.ql.exec.Utilities>
> 2016-02-02 18:04:34,737 INFO [HiveServer2-Background-Pool: Thread-517]:
> mr.MapredLocalTask (MapredLocalTask.java:executeInChildVM(287)) - Executing:
> /usr/hdp/2.4.1.0-170/hadoop/bin/hadoop jar
> /usr/hdp/2.4.1.0-170/hive/lib/hive-common-1.2.1000.2.4.1.0-170.jar
> org.apache.hadoop.hive.ql.exec.mr.ExecDriver -localtask -plan
> file:/tmp/hive/916e3dbb-a10d-4888-a063-52fb058ea421/hive_2016-02-02_18-04-29_153_625340340820843828-4/-local-10006/plan.xml
> -jobconffile
> file:/tmp/hive/916e3dbb-a10d-4888-a063-52fb058ea421/hive_2016-02-02_18-04-29_153_625340340820843828-4/-local-10007/jobconf.xml
> WARNING: Use "yarn jar" to launch YARN applications.
> 2016-02-02 18:04:37,450 INFO
> [org.apache.ranger.audit.queue.AuditBatchQueue0]: provider.BaseAuditHandler
> (BaseAuditHandler.java:logStatus(312)) - Audit Status Log:
> name=hiveServer2.async.summary.batch.solr, interval=01:21.012 minutes,
> events=2, succcessCount=2, totalEvents=4, totalSuccessCount=4
> Execution log at:
> /tmp/hive/hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0.log
> 2016-02-02 18:04:39,248 INFO [HiveServer2-HttpHandler-Pool: Thread-55]:
> thrift.ThriftHttpServlet (ThriftHttpServlet.java:doPost(127)) - Could not
> validate cookie sent, will try to generate a new cookie
> 2016-02-02 18:04:39,254 INFO [HiveServer2-HttpHandler-Pool: Thread-55]:
> thrift.ThriftHttpServlet (ThriftHttpServlet.java:doKerberosAuth(352)) -
> Failed to authenticate with http/_HOST kerberos principal, trying with
> hive/_HOST kerberos principal
> 2016-02-02 18:04:39,261 INFO [HiveServer2-HttpHandler-Pool: Thread-55]:
> thrift.ThriftHttpServlet (ThriftHttpServlet.java:doPost(169)) - Cookie added
> for clientUserName hrt_qa
> 2016-02-02 18:04:40 Starting to launch local task to process map join;
> maximum memory = 477102080
> Execution failed with exit status: 2
> 2016-02-02 18:04:43,728 ERROR [HiveServer2-Background-Pool: Thread-517]:
> exec.Task (SessionState.java:printError(932)) - Execution failed with exit
> status: 2
> Obtaining error information
> 2016-02-02 18:04:43,730 ERROR [HiveServer2-Background-Pool: Thread-517]:
> exec.Task (SessionState.java:printError(932)) - Obtaining error information
> Task failed!
> Task ID:
> Stage-5
> Logs:
> 2016-02-02 18:04:43,730 ERROR [HiveServer2-Background-Pool: Thread-517]:
> exec.Task (SessionState.java:printError(932)) -
> Task failed!
> Task ID:
> Stage-5
> Logs:
> /var/log/hive/hiveServer2.log
> 2016-02-02 18:04:43,732 ERROR [HiveServer2-Background-Pool: Thread-517]:
> exec.Task (SessionState.java:printError(932)) - /var/log/hive/hiveServer2.log
> 2016-02-02 18:04:43,732 ERROR [HiveServer2-Background-Pool: Thread-517]:
> mr.MapredLocalTask (MapredLocalTask.java:executeInChildVM(307)) - Execution
> failed with exit status: 2
> 2016-02-02 18:04:43,733 INFO [HiveServer2-Background-Pool: Thread-517]:
> hooks.ATSHook (ATSHook.java:<init>(90)) - Created ATS Hook
> 2016-02-02 18:04:43,734 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
> method=FailureHook.org.apache.hadoop.hive.ql.hooks.ATSHook
> from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:43,736 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
> method=FailureHook.org.apache.hadoop.hive.ql.hooks.ATSHook
> start=1454436283734 end=1454436283736 duration=2
> from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:43,736 INFO [ATS Logger 0]: hooks.ATSHook
> (ATSHook.java:createPostHookEvent(193)) - Received post-hook notification for
> :hive_20160202180429_7ffff6ab-64d6-4c89-88b0-6355cc5acbd0
> FAILED: Execution Error, return code 2 from
> org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask
> 2016-02-02 18:04:43,757 ERROR [HiveServer2-Background-Pool: Thread-517]:
> ql.Driver (SessionState.java:printError(932)) - FAILED: Execution Error,
> return code 2 from org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask
> 2016-02-02 18:04:43,758 INFO [HiveServer2-Background-Pool: Thread-517]:
> ql.Driver (Driver.java:execute(1621)) - Resetting the caller context to
> HIVE_SSN_ID:916e3dbb-a10d-4888-a063-52fb058ea421
> 2016-02-02 18:04:43,759 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
> method=Driver.execute start=1454436274219 end=1454436283759 duration=9540
> from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:43,760 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogBegin(135)) - <PERFLOG
> method=releaseLocks from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:43,761 INFO [HiveServer2-Background-Pool: Thread-517]:
> log.PerfLogger (PerfLogger.java:PerfLogEnd(162)) - </PERFLOG
> method=releaseLocks start=1454436283760 end=1454436283761 duration=1
> from=org.apache.hadoop.hive.ql.Driver>
> 2016-02-02 18:04:43,766 ERROR [HiveServer2-Background-Pool: Thread-517]:
> operation.Operation (SQLOperation.java:run(209)) - Error running hive query:
> org.apache.hive.service.cli.HiveSQLException: Error while processing
> statement: FAILED: Execution Error, return code 2 from
> org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask
> at
> org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:315)
> at
> org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:156)
> at
> org.apache.hive.service.cli.operation.SQLOperation.access$100(SQLOperation.java:71)
> at
> org.apache.hive.service.cli.operation.SQLOperation$1$1.run(SQLOperation.java:206)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:415)
> at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
> at
> org.apache.hive.service.cli.operation.SQLOperation$1.run(SQLOperation.java:218)
> at
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
> at java.util.concurrent.FutureTask.run(FutureTask.java:262)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> hive Configs can be viewed from
> http://qelog.hortonworks.com/log/hs21-hbs24-1454382123/artifacts/tmpModifyConfDir_1454394513592/
> Attachments
> Drop files to attach, or browse.
> Add LinkIssue Links
> relates to
> Bug - A problem which impairs or prevents the functions of the product.
> HIVE-739 webhcat tests failing in HDInsight secure cluster throwing
> NullPointerException Blocker - Blocks development and/or testing work,
> production could not run. RESOLVED
> Activity
> All
> Comments
> Work Log
> History
> Activity
> Ascending order - Click to sort in descending order
> Permalink Edit Delete
> tsaito Takahiko Saito added a comment - 3 days ago
> The test passes via hive CLI and explain shows:
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> explain select registration
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> from s10k s join v10k v
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> on (s.name = v.name) join
> studentparttab30k p
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> on (p.name = v.name)
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> where s.age < 25 and v.age <
> 25 and p.age < 25;
> +--------------------------------------------------------------------------------------------------------------+--+
> | Explain
> |
> +--------------------------------------------------------------------------------------------------------------+--+
> | STAGE DEPENDENCIES:
> |
> | Stage-5 is a root stage
> |
> | Stage-4 depends on stages: Stage-5
> |
> | Stage-0 depends on stages: Stage-4
> |
> |
> |
> | STAGE PLANS:
> |
> | Stage: Stage-5
> |
> | Map Reduce Local Work
> |
> | Alias -> Map Local Tables:
> |
> | s
> |
> | Fetch Operator
> |
> | limit: -1
> |
> | v
> |
> | Fetch Operator
> |
> | limit: -1
> |
> | Alias -> Map Local Operator Tree:
> |
> | s
> |
> | TableScan
> |
> | alias: s
> |
> | filterExpr: (name is not null and (age < 25)) (type: boolean)
> |
> | Statistics: Num rows: 459 Data size: 47777 Basic stats:
> COMPLETE Column stats: NONE |
> | Filter Operator
> |
> | predicate: (name is not null and (age < 25)) (type: boolean)
> |
> | Statistics: Num rows: 76 Data size: 7910 Basic stats:
> COMPLETE Column stats: NONE |
> | HashTable Sink Operator
> |
> | keys:
> |
> | 0 name (type: string)
> |
> | 1 name (type: string)
> |
> | 2 name (type: string)
> |
> | v
> |
> | TableScan
> |
> | alias: v
> |
> | filterExpr: (name is not null and (age < 25)) (type: boolean)
> |
> | Statistics: Num rows: 1653 Data size: 337233 Basic stats:
> COMPLETE Column stats: NONE |
> | Filter Operator
> |
> | predicate: (name is not null and (age < 25)) (type: boolean)
> |
> | Statistics: Num rows: 275 Data size: 56103 Basic stats:
> COMPLETE Column stats: NONE |
> | HashTable Sink Operator
> |
> | keys:
> |
> | 0 name (type: string)
> |
> | 1 name (type: string)
> |
> | 2 name (type: string)
> |
> |
> |
> | Stage: Stage-4
> |
> | Map Reduce
> |
> | Map Operator Tree:
> |
> | TableScan
> |
> | alias: p
> |
> | filterExpr: (name is not null and (age < 25)) (type: boolean)
> |
> | Statistics: Num rows: 30000 Data size: 627520 Basic stats:
> COMPLETE Column stats: COMPLETE |
> | Filter Operator
> |
> | predicate: (name is not null and (age < 25)) (type: boolean)
> |
> | Statistics: Num rows: 10000 Data size: 1010000 Basic stats:
> COMPLETE Column stats: COMPLETE |
> | Map Join Operator
> |
> | condition map:
> |
> | Inner Join 0 to 1
> |
> | Inner Join 1 to 2
> |
> | keys:
> |
> | 0 name (type: string)
> |
> | 1 name (type: string)
> |
> | 2 name (type: string)
> |
> | outputColumnNames: _col8
> |
> | Statistics: Num rows: 22000 Data size: 2222000 Basic stats:
> COMPLETE Column stats: NONE |
> | Select Operator
> |
> | expressions: _col8 (type: string)
> |
> | outputColumnNames: _col0
> |
> | Statistics: Num rows: 22000 Data size: 2222000 Basic
> stats: COMPLETE Column stats: NONE |
> | File Output Operator
> |
> | compressed: false
> |
> | Statistics: Num rows: 22000 Data size: 2222000 Basic
> stats: COMPLETE Column stats: NONE |
> | table:
> |
> | input format:
> org.apache.hadoop.mapred.TextInputFormat |
> | output format:
> org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat |
> | serde:
> org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
> |
> | Local Work:
> |
> | Map Reduce Local Work
> |
> |
> |
> | Stage: Stage-0
> |
> | Fetch Operator
> |
> | limit: -1
> |
> | Processor Tree:
> |
> | ListSink
> |
> |
> |
> +--------------------------------------------------------------------------------------------------------------+--+
> 83 rows selected (2.473 seconds)
> Permalink Edit Delete
> tsaito Takahiko Saito added a comment - 3 days ago
> emptablemisc_8 also fails with the same error:
> >>> create temporary table temp1 as select * from votertab10k;
> >>> select *
> from studenttab10k s
> where s.name not in
> (select name from temp1);
> INFO : Number of reduce tasks determined at compile time: 1
> INFO : In order to change the average load for a reducer (in bytes):
> INFO : set hive.exec.reducers.bytes.per.reducer=<number>
> INFO : In order to limit the maximum number of reducers:
> INFO : set hive.exec.reducers.max=<number>
> INFO : In order to set a constant number of reducers:
> INFO : set mapreduce.job.reduces=<number>
> INFO : number of splits:1
> INFO : Submitting tokens for job: job_1454394534358_0164
> INFO : Kind: HDFS_DELEGATION_TOKEN, Service: 10.0.0.36:8020, Ident:
> (HDFS_DELEGATION_TOKEN token 246 for hrt_qa)
> INFO : The url to track the job:
> http://hn0-hs21-h.hdinsight.net:8088/proxy/application_1454394534358_0164/
> INFO : Starting Job = job_1454394534358_0164, Tracking URL =
> http://hn0-hs21-h.hdinsight.net:8088/proxy/application_1454394534358_0164/
> INFO : Kill Command = /usr/hdp/2.4.1.0-170/hadoop/bin/hadoop job -kill
> job_1454394534358_0164
> INFO : Hadoop job information for Stage-4: number of mappers: 1; number of
> reducers: 1
> INFO : 2016-02-02 10:11:02,367 Stage-4 map = 0%, reduce = 0%
> INFO : 2016-02-02 10:11:26,060 Stage-4 map = 100%, reduce = 0%, Cumulative
> CPU 5.48 sec
> INFO : 2016-02-02 10:11:39,024 Stage-4 map = 100%, reduce = 100%,
> Cumulative CPU 11.18 sec
> INFO : MapReduce Total cumulative CPU time: 11 seconds 180 msec
> INFO : Ended Job = job_1454394534358_0164
> INFO : Stage-9 is selected by condition resolver.
> INFO : Stage-1 is filtered out by condition resolver.
> ERROR : Execution failed with exit status: 2
> ERROR : Obtaining error information
> ERROR :
> Task failed!
> Task ID:
> Stage-9
> Logs:
> ERROR : /var/log/hive/hiveServer2.log
> INFO : Number of reduce tasks determined at compile time: 1
> INFO : In order to change the average load for a reducer (in bytes):
> INFO : set hive.exec.reducers.bytes.per.reducer=<number>
> INFO : In order to limit the maximum number of reducers:
> INFO : set hive.exec.reducers.max=<number>
> INFO : In order to set a constant number of reducers:
> INFO : set mapreduce.job.reduces=<number>
> INFO : number of splits:2
> INFO : Submitting tokens for job: job_1454394534358_0169
> INFO : Kind: HDFS_DELEGATION_TOKEN, Service: 10.0.0.36:8020, Ident:
> (HDFS_DELEGATION_TOKEN token 252 for hrt_qa)
> INFO : The url to track the job:
> http://hn0-hs21-h.hdinsight.net:8088/proxy/application_1454394534358_0169/
> INFO : Starting Job = job_1454394534358_0169, Tracking URL =
> http://hn0-hs21-h.hdinsight.net:8088/proxy/application_1454394534358_0169/
> INFO : Kill Command = /usr/hdp/2.4.1.0-170/hadoop/bin/hadoop job -kill
> job_1454394534358_0169
> INFO : Hadoop job information for Stage-1: number of mappers: 2; number of
> reducers: 1
> INFO : 2016-02-02 10:16:38,027 Stage-1 map = 0%, reduce = 0%
> INFO : 2016-02-02 10:16:52,498 Stage-1 map = 50%, reduce = 0%, Cumulative
> CPU 3.8 sec
> INFO : 2016-02-02 10:16:53,566 Stage-1 map = 100%, reduce = 0%, Cumulative
> CPU 9.94 sec
> INFO : 2016-02-02 10:17:16,202 Stage-1 map = 100%, reduce = 100%,
> Cumulative CPU 17.04 sec
> INFO : MapReduce Total cumulative CPU time: 17 seconds 40 msec
> INFO : Ended Job = job_1454394534358_0169
> ERROR : Execution failed with exit status: 2
> ERROR : Obtaining error information
> ERROR :
> Task failed!
> Task ID:
> Stage-8
> Logs:
> ERROR : /var/log/hive/hiveServer2.log
> Error: Error while processing statement: FAILED: Execution Error, return code
> 2 from org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask (state=08S01,code=2)
> Aborting command set because "force" is false and command failed: "select *
> from studenttab10k s
> where s.name not in
> (select name from temp1);"
> Closing: 0:
> jdbc:hive2://zk2-hs21-h.hdinsight.net:2181/;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=hiveserver2;principal=hive/[email protected];transportMode=http;httpPath=cliservice
> Its app log can be viewed at
> http://qelog.hortonworks.com/log/hs21-hbs24-1454382123/app-logs/application_1454394534358_0169.log
> Permalink Edit Delete
> tsaito Takahiko Saito added a comment - 3 days ago
> The query works without 'temporary table':
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> select registration from
> studenttab10k s join votertab10k v
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> on (s.name = v.name) join
> studentparttab30k p
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> on (p.name = v.name)
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> where s.age < 25 and v.age <
> 25 and p.age < 25;
> Permalink Edit Delete
> tsaito Takahiko Saito added a comment - 3 days ago
> More info about temporary table:
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> describe formatted s10k;
> +-------------------------------+---------------------------------------------------------------------------------------------------------------------------------------------+-----------------------+--+
> | col_name |
> data_type
> | comment |
> +-------------------------------+---------------------------------------------------------------------------------------------------------------------------------------------+-----------------------+--+
> | # col_name | data_type
>
> | comment |
> | | NULL
>
> | NULL |
> | name | string
>
> | |
> | age | int
>
> | |
> | gpa | double
>
> | |
> | | NULL
>
> | NULL |
> | # Detailed Table Information | NULL
>
> | NULL |
> | Database: | default
>
> | NULL |
> | Owner: | hrt_qa
>
> | NULL |
> | CreateTime: | Tue Feb 02 23:02:31 UTC 2016
>
> | NULL |
> | LastAccessTime: | UNKNOWN
>
> | NULL |
> | Protect Mode: | None
>
> | NULL |
> | Retention: | 0
>
> | NULL |
> | Location: |
> hdfs://hn0-hs21-h.hdinsight.net:8020/tmp/hive/hive/de2667ea-5bc1-4548-a4c4-97e46d6081f8/_tmp_space.db/8e8482c9-a2a4-4e8c-ad0f-2cd6e8fcdb4c
> | NULL |
> | Table Type: | MANAGED_TABLE
>
> | NULL |
> | | NULL
>
> | NULL |
> | # Storage Information | NULL
>
> | NULL |
> | SerDe Library: | org.apache.hadoop.hive.ql.io.orc.OrcSerde
>
> | NULL |
> | InputFormat: |
> org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
> | NULL
> |
> | OutputFormat: |
> org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
> | NULL
> |
> | Compressed: | No
>
> | NULL |
> | Num Buckets: | -1
>
> | NULL |
> | Bucket Columns: | []
>
> | NULL |
> | Sort Columns: | []
>
> | NULL |
> | Storage Desc Params: | NULL
>
> | NULL |
> | | serialization.format
>
> | 1 |
> +-------------------------------+---------------------------------------------------------------------------------------------------------------------------------------------+-----------------------+--+
> 26 rows selected (0.22 seconds)
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> set hive.server2.enable.doAs;
> +---------------------------------+--+
> | set |
> +---------------------------------+--+
> | hive.server2.enable.doAs=false |
> +---------------------------------+--+
> Permalink Edit Delete
> tsaito Takahiko Saito added a comment - 3 days ago
> hdfs dir of temporary table is owned by hive as expected since
> hive.server2.enable.doAs=false:
> hdfs@hn0-hs21-h:~$ hdfs dfs -ls
> hdfs://hn0-hs21-h.hdinsight.net:8020/tmp/hive/hive/de2667ea-5bc1-4548-a4c4-97e46d6081f8/_tmp_space.db/8e8482c9-a2a4-4e8c-ad0f-2cd6e8fcdb4c
> Found 1 items
> -rwx------ 3 hive hdfs 47777 2016-02-02 23:02
> hdfs://hn0-hs21-h.hdinsight.net:8020/tmp/hive/hive/de2667ea-5bc1-4548-a4c4-97e46d6081f8/_tmp_space.db/8e8482c9-a2a4-4e8c-ad0f-2cd6e8fcdb4c/000000_0
> Not sure if it's related to the JIRA, but one thing I noticed was that dfs
> cmd throws error via beeline:
> 0: jdbc:hive2://zk2-hs21-h.hdinsight.net:2181> dfs -ls
> hdfs://hn0-hs21-h.hdinsight.net:8020/tmp/hive/hive/de2667ea-5bc1-4548-a4c4-97e46d6081f8/_tmp_space.db/8e8482c9-a2a4-4e8c-ad0f-2cd6e8fcdb4c;
> Error: Error while processing statement: Permission denied: user [hrt_qa]
> does not have privilege for [DFS] command (state=,code=1)
> Permalink Edit
> thejas Thejas Nair added a comment - 5 hours ago - edited
> Takahiko Saito
> The above DFS error would be from Ranger of SQL standard authorization. It is
> not related to the test failure.
> But you are right that it looks like a permission issue.
> The issue seen in HIVE-739 also exists in this part of Hive. Hive in MR mode
> launches a new child process to process the small table and create a hash
> table. This child process needs credentials from HDFS. However, in this
> setup, azure fs is the default file system. We should also get delegation
> token from all FS URIs listed under mapreduce.job.hdfs-servers config.
> There are hueristics around when map-join gets used. That is why you don't
> see it unless temp table is used (that uses ORC format and might have stats
> as well, while the original table is probaly text format).
> https://github.com/hortonworks/hive/blob/2.4-maint/ql/src/java/org/apache/hadoop/hive/ql/exec/SecureCmdDoAs.java
> needs change similar to webhcat change in HIVE-739
> Permalink Edit
> sush Sushanth Sowmyan added a comment - 7 minutes ago
> Takahiko Saito, we should file an apache jira for this as well. Would you
> like to do so, so the bug report has attribution to you?
> Permalink Edit Delete
> tsaito Takahiko Saito added a comment - 4 minutes ago
> Sushanth Sowmyan I will file one and update with that JIRA.
> Comment
> People
> Assignee: sush Sushanth Sowmyan
> Assign to me
> Reporter: tsaito Takahiko Saito
> QEAssignee: Takahiko Saito
> Votes: 0
> Watchers: 3 Stop watching this issue
> Dates
> Created: 3 days ago
> Updated: 4 minutes ago
> Who's Looking?
> Agile
> View on Board
> {noformat}
> hivesever2 log shows:
> {noformat}
> 2016-02-02 18:04:43,766 ERROR [HiveServer2-Background-Pool: Thread-517]:
> operation.Operation (SQLOperation.java:run(209)) - Error running hive query:
> org.apache.hive.service.cli.HiveSQLException: Error while processing
> statement: FAILED: Execution Error, return code 2 from
> org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask
> at
> org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:315)
> at
> org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:156)
> at
> org.apache.hive.service.cli.operation.SQLOperation.access$100(SQLOperation.java:71)
> at
> org.apache.hive.service.cli.operation.SQLOperation$1$1.run(SQLOperation.java:206)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:415)
> at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
> at
> org.apache.hive.service.cli.operation.SQLOperation$1.run(SQLOperation.java:218)
> at
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
> at java.util.concurrent.FutureTask.run(FutureTask.java:262)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> {noformat}
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)