Here is the Hive log when running the "select count(*) from cars2;":


    application_1402243729361_0009
14/06/08 10:27:19 INFO log.PerfLogger: <PERFLOG method=compile 
from=org.apache.hadoop.hive.ql.Driver>
14/06/08 10:27:19 INFO log.PerfLogger: <PERFLOG method=parse 
from=org.apache.hadoop.hive.ql.Driver>
14/06/08 10:27:19 INFO parse.ParseDriver: Parsing command: select count(*) 
from cars2
14/06/08 10:27:19 INFO parse.ParseDriver: Parse Completed
14/06/08 10:27:19 INFO log.PerfLogger: </PERFLOG method=parse 
start=1402248439935 end=1402248439936 duration=1 
from=org.apache.hadoop.hive.ql.Driver>
14/06/08 10:27:19 INFO log.PerfLogger: <PERFLOG method=semanticAnalyze 
from=org.apache.hadoop.hive.ql.Driver>
14/06/08 10:27:19 INFO parse.SemanticAnalyzer: Starting Semantic Analysis
14/06/08 10:27:19 INFO parse.SemanticAnalyzer: Completed phase 1 of 
Semantic Analysis
14/06/08 10:27:19 INFO parse.SemanticAnalyzer: Get metadata for source 
tables
14/06/08 10:27:19 INFO parse.SemanticAnalyzer: Get metadata for subqueries
14/06/08 10:27:19 INFO parse.SemanticAnalyzer: Get metadata for destination 
tables
14/06/08 10:27:19 INFO ql.Context: New scratch dir is 
hdfs://localhost.localdomain:8020/tmp/hive-hive/hive_2014-06-08_10-27-19_935_2002159029513445063-1
14/06/08 10:27:19 INFO parse.SemanticAnalyzer: Completed getting MetaData 
in Semantic Analysis
14/06/08 10:27:19 INFO ppd.OpProcFactory: Processing for FS(18)
14/06/08 10:27:19 INFO ppd.OpProcFactory: Processing for SEL(17)
14/06/08 10:27:19 INFO ppd.OpProcFactory: Processing for GBY(16)
14/06/08 10:27:19 INFO ppd.OpProcFactory: Processing for RS(15)
14/06/08 10:27:19 INFO ppd.OpProcFactory: Processing for GBY(14)
14/06/08 10:27:19 INFO ppd.OpProcFactory: Processing for SEL(13)
14/06/08 10:27:19 INFO ppd.OpProcFactory: Processing for TS(12)
14/06/08 10:27:19 INFO log.PerfLogger: <PERFLOG method=partition-retrieving 
from=org.apache.hadoop.hive.ql.optimizer.ppr.PartitionPruner>
14/06/08 10:27:19 INFO log.PerfLogger: </PERFLOG 
method=partition-retrieving start=1402248439988 end=1402248439989 
duration=1 from=org.apache.hadoop.hive.ql.optimizer.ppr.PartitionPruner>
14/06/08 10:27:19 INFO physical.MetadataOnlyOptimizer: Looking for table 
scans where optimization is applicable
14/06/08 10:27:19 INFO physical.MetadataOnlyOptimizer: Found 0 metadata 
only table scans
14/06/08 10:27:19 INFO parse.SemanticAnalyzer: Completed plan generation
14/06/08 10:27:19 INFO ql.Driver: Semantic Analysis Completed
14/06/08 10:27:19 INFO log.PerfLogger: </PERFLOG method=semanticAnalyze 
start=1402248439936 end=1402248439990 duration=54 
from=org.apache.hadoop.hive.ql.Driver>
14/06/08 10:27:19 INFO exec.ListSinkOperator: Initializing Self 19 OP
14/06/08 10:27:19 INFO exec.ListSinkOperator: Operator 19 OP initialized
14/06/08 10:27:19 INFO exec.ListSinkOperator: Initialization Done 19 OP
14/06/08 10:27:19 INFO ql.Driver: Returning Hive schema: 
Schema(fieldSchemas:[FieldSchema(name:_c0, type:bigint, comment:null)], 
properties:null)
14/06/08 10:27:19 INFO log.PerfLogger: </PERFLOG method=compile 
start=1402248439935 end=1402248439991 duration=56 
from=org.apache.hadoop.hive.ql.Driver>
14/06/08 10:27:19 INFO Configuration.deprecation: 
mapred.input.dir.recursive is deprecated. Instead, use 
mapreduce.input.fileinputformat.input.dir.recursive
14/06/08 10:27:19 INFO log.PerfLogger: <PERFLOG method=Driver.run 
from=org.apache.hadoop.hive.ql.Driver>
14/06/08 10:27:19 INFO log.PerfLogger: <PERFLOG method=TimeToSubmit 
from=org.apache.hadoop.hive.ql.Driver>
14/06/08 10:27:19 INFO ql.Driver: Creating lock manager of type 
org.apache.hadoop.hive.ql.lockmgr.zookeeper.ZooKeeperHiveLockManager
14/06/08 10:27:19 INFO zookeeper.ZooKeeper: Initiating client connection, 
connectString=localhost.localdomain:2181 sessionTimeout=600000 
watcher=org.apache.hadoop.hive.ql.lockmgr.zookeeper.ZooKeeperHiveLockManager$DummyWatcher@d699a84
14/06/08 10:27:19 INFO log.PerfLogger: <PERFLOG 
method=acquireReadWriteLocks from=org.apache.hadoop.hive.ql.Driver>
14/06/08 10:27:20 INFO log.PerfLogger: </PERFLOG 
method=acquireReadWriteLocks start=1402248439999 end=1402248440012 
duration=13 from=org.apache.hadoop.hive.ql.Driver>
14/06/08 10:27:20 INFO log.PerfLogger: <PERFLOG method=Driver.execute 
from=org.apache.hadoop.hive.ql.Driver>
14/06/08 10:27:20 INFO ql.Driver: Starting command: select count(*) from 
cars2
14/06/08 10:27:20 INFO ql.Driver: Total MapReduce jobs = 1
14/06/08 10:27:20 INFO log.PerfLogger: </PERFLOG method=TimeToSubmit 
start=1402248439992 end=1402248440012 duration=20 
from=org.apache.hadoop.hive.ql.Driver>
14/06/08 10:27:20 INFO log.PerfLogger: <PERFLOG method=runTasks 
from=org.apache.hadoop.hive.ql.Driver>
14/06/08 10:27:20 INFO log.PerfLogger: <PERFLOG method=task.MAPRED.Stage-1 
from=org.apache.hadoop.hive.ql.Driver>
14/06/08 10:27:20 INFO ql.Driver: Launching Job 1 out of 1
14/06/08 10:27:20 INFO exec.Task: Number of reduce tasks determined at 
compile time: 1
14/06/08 10:27:20 INFO exec.Task: In order to change the average load for a 
reducer (in bytes):
14/06/08 10:27:20 INFO exec.Task:   set 
hive.exec.reducers.bytes.per.reducer=<number>
14/06/08 10:27:20 INFO exec.Task: In order to limit the maximum number of 
reducers:
14/06/08 10:27:20 INFO exec.Task:   set hive.exec.reducers.max=<number>
14/06/08 10:27:20 INFO exec.Task: In order to set a constant number of 
reducers:
14/06/08 10:27:20 INFO exec.Task:   set mapred.reduce.tasks=<number>
14/06/08 10:27:20 INFO ql.Context: New scratch dir is 
hdfs://localhost.localdomain:8020/tmp/hive-hive/hive_2014-06-08_10-27-19_935_2002159029513445063-3
14/06/08 10:27:20 INFO Configuration.deprecation: 
mapred.reduce.tasks.speculative.execution is deprecated. Instead, use 
mapreduce.reduce.speculative
14/06/08 10:27:20 INFO mr.ExecDriver: Using 
org.apache.hadoop.hive.ql.io.CombineHiveInputFormat
14/06/08 10:27:20 INFO mr.ExecDriver: adding libjars: 
file:///opt/cloudera/parcels/CDH-5.0.0-1.cdh5.0.0.p0.47/lib/hive/lib/hive-hbase-handler-0.12.0-cdh5.0.0.jar,file:///opt/cloudera/parcels/CDH-5.0.0-1.cdh5.0.0.p0.47/lib/hbase/hbase-protocol.jar,file:///opt/cloudera/parcels/CDH-5.0.0-1.cdh5.0.0.p0.47/lib/hbase/hbase-client.jar,file:///opt/cloudera/parcels/CDH-5.0.0-1.cdh5.0.0.p0.47/lib/hbase/lib/htrace-core.jar,file:///opt/cloudera/parcels/CDH-5.0.0-1.cdh5.0.0.p0.47/lib/hbase/lib/htrace-core-2.01.jar,file:///opt/cloudera/parcels/CDH-5.0.0-1.cdh5.0.0.p0.47/lib/hbase/hbase-hadoop-compat.jar,file:///opt/cloudera/parcels/CDH-5.0.0-1.cdh5.0.0.p0.47/lib/hbase/hbase-hadoop2-compat.jar,file:///opt/cloudera/parcels/CDH-5.0.0-1.cdh5.0.0.p0.47/lib/hbase/hbase-server.jar,file:///opt/cloudera/parcels/CDH-5.0.0-1.cdh5.0.0.p0.47/lib/hbase/hbase-common.jar
14/06/08 10:27:20 INFO exec.Utilities: Processing alias cars2
14/06/08 10:27:20 INFO exec.Utilities: Adding input file 
hdfs://localhost.localdomain:8020/user/hive/warehouse/cars2
14/06/08 10:27:20 INFO exec.Utilities: Content Summary not cached for 
hdfs://localhost.localdomain:8020/user/hive/warehouse/cars2
14/06/08 10:27:20 INFO ql.Context: New scratch dir is 
hdfs://localhost.localdomain:8020/tmp/hive-hive/hive_2014-06-08_10-27-19_935_2002159029513445063-3
14/06/08 10:27:20 INFO log.PerfLogger: <PERFLOG method=serializePlan 
from=org.apache.hadoop.hive.ql.exec.Utilities>
14/06/08 10:27:20 INFO exec.Utilities: Serializing MapWork via kryo
14/06/08 10:27:20 INFO log.PerfLogger: </PERFLOG method=serializePlan 
start=1402248440040 end=1402248440067 duration=27 
from=org.apache.hadoop.hive.ql.exec.Utilities>
14/06/08 10:27:20 INFO log.PerfLogger: <PERFLOG method=serializePlan 
from=org.apache.hadoop.hive.ql.exec.Utilities>
14/06/08 10:27:20 INFO exec.Utilities: Serializing ReduceWork via kryo
14/06/08 10:27:20 INFO log.PerfLogger: </PERFLOG method=serializePlan 
start=1402248440074 end=1402248440096 duration=22 
from=org.apache.hadoop.hive.ql.exec.Utilities>
14/06/08 10:27:20 INFO client.RMProxy: Connecting to ResourceManager at 
localhost.localdomain/127.0.0.1:8032
14/06/08 10:27:20 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:27:20 INFO client.RMProxy: Connecting to ResourceManager at 
localhost.localdomain/127.0.0.1:8032
14/06/08 10:27:20 WARN mapreduce.JobSubmitter: Hadoop command-line option 
parsing not performed. Implement the Tool interface and execute your 
application with ToolRunner to remedy this.
14/06/08 10:27:20 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:27:20 INFO log.PerfLogger: <PERFLOG method=getSplits 
from=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat>
14/06/08 10:27:20 INFO log.PerfLogger: <PERFLOG method=getSplits 
from=org.apache.hadoop.hive.ql.io.HiveInputFormat>
14/06/08 10:27:20 INFO mr.EsInputFormat: Reading from [cars/transactions]
14/06/08 10:27:20 INFO mr.EsInputFormat: Discovered mapping 
{cars=[mappings=[transactions=[color=STRING, make=STRING, price=LONG, 
sold=DATE]]]} for [cars/transactions]
14/06/08 10:27:20 INFO mr.EsInputFormat: Created [5] shard-splits
14/06/08 10:27:20 INFO io.HiveInputFormat: number of splits 5
14/06/08 10:27:20 INFO log.PerfLogger: </PERFLOG method=getSplits 
start=1402248440609 end=1402248440652 duration=43 
from=org.apache.hadoop.hive.ql.io.HiveInputFormat>
14/06/08 10:27:20 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:27:21 INFO mapreduce.JobSubmitter: number of splits:5
14/06/08 10:27:21 INFO mapreduce.JobSubmitter: Submitting tokens for job: 
job_1402243729361_0009
14/06/08 10:27:21 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:27:21 INFO impl.YarnClientImpl: Submitted application 
application_1402243729361_0009
14/06/08 10:27:21 INFO mapreduce.Job: The url to track the job: 
http://localhost.localdomain:8088/proxy/application_1402243729361_0009/
14/06/08 10:27:21 INFO exec.Task: Starting Job = job_1402243729361_0009, 
Tracking URL = 
http://localhost.localdomain:8088/proxy/application_1402243729361_0009/
14/06/08 10:27:21 INFO exec.Task: Kill Command = 
/opt/cloudera/parcels/CDH-5.0.0-1.cdh5.0.0.p0.47/lib/hadoop/bin/hadoop job  
-kill job_1402243729361_0009
14/06/08 10:27:21 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:27:22 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:27:23 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:27:24 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:27:25 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:27:26 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:27:27 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:27:27 INFO exec.Task: Hadoop job information for Stage-1: 
number of mappers: 5; number of reducers: 1
14/06/08 10:27:27 WARN mapreduce.Counters: Group 
org.apache.hadoop.mapred.Task$Counter is deprecated. Use 
org.apache.hadoop.mapreduce.TaskCounter instead
14/06/08 10:27:27 INFO exec.Task: 2014-06-08 10:27:27,858 Stage-1 map = 
0%,  reduce = 0%
14/06/08 10:27:28 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:27:29 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:27:31 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:27:33 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:27:34 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:27:36 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:27:38 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:27:40 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:27:42 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:27:44 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:27:46 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:27:48 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:27:51 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:27:53 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:27:55 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:27:57 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:27:59 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:28:01 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:28:03 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:28:05 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:28:08 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:28:10 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:28:12 INFO exec.Task: 2014-06-08 10:28:12,001 Stage-1 map = 
100%,  reduce = 100%
14/06/08 10:28:12 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()
14/06/08 10:28:13 ERROR exec.Task: Ended Job = job_1402243729361_0009 with 
errors
14/06/08 10:28:13 INFO impl.YarnClientImpl: Killed application 
application_1402243729361_0009
14/06/08 10:28:13 INFO log.PerfLogger: </PERFLOG method=task.MAPRED.Stage-1 
start=1402248440012 end=1402248493129 duration=53117 
from=org.apache.hadoop.hive.ql.Driver>
14/06/08 10:28:13 ERROR ql.Driver: FAILED: Execution Error, return code 2 
from org.apache.hadoop.hive.ql.exec.mr.MapRedTask
14/06/08 10:28:13 INFO log.PerfLogger: </PERFLOG method=Driver.execute 
start=1402248440012 end=1402248493130 duration=53118 
from=org.apache.hadoop.hive.ql.Driver>
14/06/08 10:28:13 INFO ql.Driver: MapReduce Jobs Launched: 
14/06/08 10:28:13 WARN mapreduce.Counters: Group FileSystemCounters is 
deprecated. Use org.apache.hadoop.mapreduce.FileSystemCounter instead
14/06/08 10:28:13 INFO ql.Driver: Job 0: Map: 5  Reduce: 1   HDFS Read: 0 
HDFS Write: 0 FAIL
14/06/08 10:28:13 INFO ql.Driver: Total MapReduce CPU Time Spent: 0 msec
14/06/08 10:28:13 INFO log.PerfLogger: <PERFLOG method=releaseLocks 
from=org.apache.hadoop.hive.ql.Driver>
14/06/08 10:28:13 INFO ZooKeeperHiveLockManager:  about to release lock for 
default/cars2
14/06/08 10:28:13 INFO ZooKeeperHiveLockManager:  about to release lock for 
default
14/06/08 10:28:13 INFO log.PerfLogger: </PERFLOG method=releaseLocks 
start=1402248493131 end=1402248493141 duration=10 
from=org.apache.hadoop.hive.ql.Driver>
14/06/08 10:28:13 ERROR operation.Operation: Error: 
org.apache.hive.service.cli.HiveSQLException: Error while processing 
statement: FAILED: Execution Error, return code 2 from 
org.apache.hadoop.hive.ql.exec.mr.MapRedTask
 at 
org.apache.hive.service.cli.operation.SQLOperation.runInternal(SQLOperation.java:146)
 at 
org.apache.hive.service.cli.operation.SQLOperation.access$100(SQLOperation.java:64)
 at 
org.apache.hive.service.cli.operation.SQLOperation$1.run(SQLOperation.java:177)
 at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
 at java.util.concurrent.FutureTask.run(FutureTask.java:262)
 at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
 at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
 at java.lang.Thread.run(Thread.java:744)
14/06/08 10:28:14 INFO cli.CLIService: OperationHandle 
[opType=EXECUTE_STATEMENT, 
getHandleIdentifier()=ef7105d2-e323-44f4-b1b8-3e2aca12b198]: 
getOperationStatus()


On Sunday, June 8, 2014 1:03:43 PM UTC-4, elitem way wrote:

> I am learning the elasticsearch-hadoop. I have a few issues that I do not 
> understand. I am using ES 1.12 on Windows, elasticsearch-hadoop-2.0.0 and 
> cloudera-quickstart-vm-5.0.0-0-vmware sandbox with Hive.
>
> 1. I loaded only 6 rows to ES index car/transactions. Why did Hive return 
> 14 rows instead? See below.
> 2. "select count(*) from cars2" failed with code 2. "Group by", "sum" also 
> failed. Did I miss anything. The similar query are successful when using 
> sample_07 and sample_08 tables that come with Hive.
> 3.  elasticsearch-hadoop-2.0.0 does seem to work with jetty - the 
> authentication plugin. I got errors when I enable jetty and set 'es.nodes' 
> = 'superuser:[email protected]'
> 4. I could not pipe data from Hive to ElasticSearch either.
>
> *--ISSUE 1*:
> --load data to ES
> ­ POST: http://localhost:9200/cars/transactions/_bulk
> { "index": {}}
> { "price" : 30000, "color" : "green", "make" : "ford", "sold" : 
> "2014-05-18" }
> { "index": {}}
> { "price" : 15000, "color" : "blue", "make" : "toyota", "sold" : 
> "2014-07-02" }
> { "index": {}}
> { "price" : 12000, "color" : "green", "make" : "toyota", "sold" : 
> "2014-08-19" }
> { "index": {}}
> { "price" : 20000, "color" : "red", "make" : "honda", "sold" : 
> "2014-11-05" }
> { "index": {}}
> { "price" : 80000, "color" : "red", "make" : "bmw", "sold" : "2014-01-01" }
> { "index": {}}
> { "price" : 25000, "color" : "blue", "make" : "ford", "sold" : 
> "2014-02-12" }
>
> CREATE EXTERNAL TABLE cars2 (color STRING, make STRING, price BIGINT, sold 
> TIMESTAMP)
> STORED BY 'org.elasticsearch.hadoop.hive.EsStorageHandler'
> TBLPROPERTIES('es.resource' = 'cars/transactions',
> 'es.nodes' = '192.168.128.1', 'es.port'='9200');
>
> HIVE: select * from cars2;
> 14 rows returned.
>
>   color make price sold
> 0 red honda 20000 2014-11-05 00:00:00.0
> 1 red honda 10000 2014-10-28 00:00:00.0
> 2 green ford 30000 2014-05-18 00:00:00.0
> 3 green toyota 12000 2014-08-19 00:00:00.0
> 4 blue ford 25000 2014-02-12 00:00:00.0
> 5 blue toyota 15000 2014-07-02 00:00:00.0
> 6 red bmw 80000 2014-01-01 00:00:00.0
> 7 red honda 10000 2014-10-28 00:00:00.0
> 8 blue toyota 15000 2014-07-02 00:00:00.0
> 9 red honda 20000 2014-11-05 00:00:00.0
> 10 green ford 30000 2014-05-18 00:00:00.0
> 11 green toyota 12000 2014-08-19 00:00:00.0
> 12 red honda 20000 2014-11-05 00:00:00.0
> 13 red honda 20000 2014-11-05 00:00:00.0
> 14 red bmw 80000 2014-01-01 00:00:00.0
>
>
> *ISSUE2:*
>
> HIVE: select count(*) from cars2;
>
> Your query has the following error(s):
> Error while processing statement: FAILED: Execution Error, return code 2 
> from org.apache.hadoop.hive.ql.exec.mr.MapRedTask
>
>
> *--ISSUE 4:*
>
> CREATE EXTERNAL TABLE test1 (
>         description STRING)
> STORED BY 'org.elasticsearch.hadoop.hive.EsStorageHandler'
> TBLPROPERTIES('es.host' = '192.168.128.1', 'es.port'='9200', 'es.resource' 
> = 'test1');
>
> INSERT OVERWRITE TABLE test1 select description from sample_07;
>
> Your query has the following error(s):
>
> Error while processing statement: FAILED: Execution Error, return code 2 
> from org.apache.hadoop.hive.ql.exec.mr.MapRedTask
>

-- 
You received this message because you are subscribed to the Google Groups 
"elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/elasticsearch/e17dd332-a8d5-4636-8b12-438559e018de%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.

Reply via email to