Eric Hanson created HIVE-4665:
---------------------------------

             Summary: error at VectorExecMapper.close in group-by-agg query 
over ORC, vectorized
                 Key: HIVE-4665
                 URL: https://issues.apache.org/jira/browse/HIVE-4665
             Project: Hive
          Issue Type: Sub-task
          Components: Query Processor
    Affects Versions: vectorization-branch
            Reporter: Eric Hanson
            Assignee: Jitendra Nath Pandey


CREATE EXTERNAL TABLE FactSqlEngineAM4712( dAppVersionBuild int, 
dAppVersionBuildUNMAPPED32449 int, dAppVersionMajor int, dAppVersionMinor32447 
int, dAverageCols23083 int, dDatabaseSize23090 int, dDate string, 
dIsInternalMSFT16431 int, dLockEscalationDisabled23323 int, 
dLockEscalationEnabled23324 int, dMachineID int, dNumberTables23008 int, 
dNumCompressionPagePartitions23088 int, dNumCompressionRowPartitions23089 int, 
dNumIndexFragmentation23084 int, dNumPartitionedTables23098 int, 
dNumPartitions23099 int, dNumTablesClusterIndex23010 int, dNumTablesHeap23100 
int, dSessionType5618 int, dSqlEdition8213 int, dTempDbSize23103 int, 
mNumColumnStoreIndexesVar48171 bigint, mOccurrences int, mRowFlag int) ROW 
FORMAT DELIMITED FIELDS TERMINATED BY '\t' LOCATION '/user/ehans/SQM'; 

create table FactSqlEngineAM_vec_ORC ROW FORMAT SERDE 
'org.apache.hadoop.hive.ql.io.orc.OrcSerde' stored as INPUTFORMAT 
'org.apache.hadoop.hive.ql.io.orc.CommonOrcInputFormat' OUTPUTFORMAT 
'org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat' AS select * from 
FactSqlEngineAM4712;



hive> select ddate, max(dnumbertables23008) from factsqlengineam_vec_orc group 
by ddate;

Total MapReduce jobs = 1
Launching Job 1 out of 1
Number of reduce tasks not specified. Estimated from input data size: 3
In order to change the average load for a reducer (in bytes):
  set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
  set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
  set mapred.reduce.tasks=<number>
Validating if vectorized execution is applicable
Going down the vectorization path
java.lang.InstantiationException: 
org.apache.hadoop.hive.serde2.objectinspector.StandardStructObjectInspector
Continuing ...
java.lang.RuntimeException: failed to evaluate: <unbound>=Class.new();
Continuing ...
java.lang.InstantiationException: 
org.apache.hadoop.hive.ql.exec.vector.VectorGroupByOperator
Continuing ...
java.lang.Exception: XMLEncoder: discarding statement 
ArrayList.add(VectorGroupByOperator);
Continuing ...
Starting Job = job_201306041757_0016, Tracking URL = 
http://localhost:50030/jobdetails.jsp?jobid=job_201306041757_0016
Kill Command = c:\Hadoop\hadoop-1.1.0-SNAPSHOT\bin\hadoop.cmd job  -kill 
job_201306041757_0016
Hadoop job information for Stage-1: number of mappers: 8; number of reducers: 3
2013-06-05 10:03:06,022 Stage-1 map = 0%,  reduce = 0%
2013-06-05 10:03:51,142 Stage-1 map = 100%,  reduce = 100%
Ended Job = job_201306041757_0016 with errors
Error during job, obtaining debugging information...
Job Tracking URL: 
http://localhost:50030/jobdetails.jsp?jobid=job_201306041757_0016
Examining task ID: task_201306041757_0016_m_000009 (and more) from job 
job_201306041757_0016

Task with the most failures(4):
-----
Task ID:
  task_201306041757_0016_m_000000

URL:
  
http://localhost:50030/taskdetails.jsp?jobid=job_201306041757_0016&tipid=task_201306041757_0016_m_000000
-----
Diagnostic Messages for this Task:
java.lang.RuntimeException: Hive Runtime Error while closing operators
        at 
org.apache.hadoop.hive.ql.exec.vector.VectorExecMapper.close(VectorExecMapper.java:229)
        at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:57)
        at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:436)
        at org.apache.hadoop.mapred.MapTask.run(MapTask.java:372)
        at org.apache.hadoop.mapred.Child$4.run(Child.java:271)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:396)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1135)
        at org.apache.hadoop.mapred.Child.main(Child.java:265)
Caused by: java.lang.ClassCastException: org.apache.hadoop.io.BytesWritable 
cannot be cast to org.apache.hadoop.io.Text
        at 
org.apache.hadoop.hive.serde2.objectinspector.primitive.WritableStringObjectInspector.getPrimitiveWritableObject(Writable
StringObjectInspector.java:40)
        at 
org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorUtils.hashCode(ObjectInspectorUtils.java:481)
        at 
org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:235)
        at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:502)
        at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:832)
        at 
org.apache.hadoop.hive.ql.exec.vector.VectorGroupByOperator.closeOp(VectorGroupByOperator.java:253)
        at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:588)
        at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:597)
        at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:597)
        at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:597)
        at 
org.apache.hadoop.hive.ql.exec.vector.VectorExecMapper.close(VectorExecMapper.java:196)
        ... 8 more


FAILED: Execution Error, return code 2 from 
org.apache.hadoop.hive.ql.exec.MapRedTask
MapReduce Jobs Launched:
Job 0: Map: 8  Reduce: 3   HDFS Read: 0 HDFS Write: 0 FAIL
Total MapReduce CPU Time Spent: 0 msec
hive>

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

Reply via email to