Greetings

Thanx  Sure can.   Below is from HIVE CLI

hive> select count(distinct warctype) from commoncrawl18 where
warctype='warcinfo';
Query ID = jmill383_20160518143715_34041e3e-713b-4e35-ae86-a88498192ab1
Total jobs = 1
Launching Job 1 out of 1
Number of reduce tasks determined at compile time: 1
In order to change the average load for a reducer (in bytes):
  set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
  set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
  set mapreduce.job.reduces=<number>
Starting Job = job_1463594979064_0003, Tracking URL =
http://starchild:8088/proxy/application_1463594979064_0003/
Kill Command = /opt/hadoop/bin/hadoop job  -kill job_1463594979064_0003
Hadoop job information for Stage-1: number of mappers: 0; number of
reducers: 0
2016-05-18 14:37:19,794 Stage-1 map = 0%,  reduce = 0%
Ended Job = job_1463594979064_0003 with errors
Error during job, obtaining debugging information...
Job Tracking URL:
http://starchild:8088/cluster/app/application_1463594979064_0003
FAILED: Execution Error, return code 2 from
org.apache.hadoop.hive.ql.exec.mr.MapRedTask
MapReduce Jobs Launched:
Stage-Stage-1:  HDFS Read: 0 HDFS Write: 0 FAIL
Total MapReduce CPU Time Spent: 0 msec
hive>


This one is from cascading-hive

[jmill383@starchild demo]$ /opt/hadoop/bin/hadoop jar
build/libs/cascading-hive-demo-1.0.jar cascading.hive.HiveDemo
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in
[jar:file:/opt/hadoop/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in
[jar:file:/usr/local/hive/lib/slf4j-log4j12-1.6.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
16/05/18 14:40:20 WARN util.NativeCodeLoader: Unable to load native-hadoop
library for your platform... using builtin-java classes where applicable
16/05/18 14:40:20 INFO metastore.HiveMetaStore: 0: Opening raw store with
implemenation class:org.apache.hadoop.hive.metastore.ObjectStore
16/05/18 14:40:20 INFO metastore.ObjectStore: ObjectStore, initialize called
16/05/18 14:40:21 WARN DataNucleus.General: Plugin (Bundle)
"org.datanucleus.api.jdo" is already registered. Ensure you dont have
multiple JAR versions of the same plugin in the classpath. The URL
"file:/tmp/hadoop-unjar5875988559818865999/lib/datanucleus-api-jdo-3.2.6.jar"
is already registered, and you are trying to register an identical plugin
located at URL "file:/usr/local/hive/lib/datanucleus-api-jdo-3.2.6.jar."
16/05/18 14:40:21 WARN DataNucleus.General: Plugin (Bundle)
"org.datanucleus" is already registered. Ensure you dont have multiple JAR
versions of the same plugin in the classpath. The URL
"file:/tmp/hadoop-unjar5875988559818865999/lib/datanucleus-core-3.2.10.jar"
is already registered, and you are trying to register an identical plugin
located at URL "file:/usr/local/hive/lib/datanucleus-core-3.2.10.jar."
16/05/18 14:40:21 WARN DataNucleus.General: Plugin (Bundle)
"org.datanucleus.store.rdbms" is already registered. Ensure you dont have
multiple JAR versions of the same plugin in the classpath. The URL
"file:/tmp/hadoop-unjar5875988559818865999/lib/datanucleus-rdbms-3.2.9.jar"
is already registered, and you are trying to register an identical plugin
located at URL "file:/usr/local/hive/lib/datanucleus-rdbms-3.2.9.jar."
16/05/18 14:40:21 INFO DataNucleus.Persistence: Property
datanucleus.cache.level2 unknown - will be ignored
16/05/18 14:40:21 INFO DataNucleus.Persistence: Property
hive.metastore.integral.jdo.pushdown unknown - will be ignored
16/05/18 14:40:22 INFO metastore.ObjectStore: Setting MetaStore object pin
classes with
hive.metastore.cache.pinobjtypes="Table,StorageDescriptor,SerDeInfo,Partition,Database,Type,FieldSchema,Order"
16/05/18 14:40:23 INFO DataNucleus.Datastore: The class
"org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as
"embedded-only" so does not have its own datastore table.
16/05/18 14:40:23 INFO DataNucleus.Datastore: The class
"org.apache.hadoop.hive.metastore.model.MOrder" is tagged as
"embedded-only" so does not have its own datastore table.
16/05/18 14:40:23 INFO DataNucleus.Datastore: The class
"org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as
"embedded-only" so does not have its own datastore table.
16/05/18 14:40:23 INFO DataNucleus.Datastore: The class
"org.apache.hadoop.hive.metastore.model.MOrder" is tagged as
"embedded-only" so does not have its own datastore table.
16/05/18 14:40:24 INFO DataNucleus.Query: Reading in results for query
"org.datanucleus.store.rdbms.query.SQLQuery@0" since the connection used is
closing
16/05/18 14:40:24 INFO metastore.MetaStoreDirectSql: Using direct SQL,
underlying DB is DERBY
16/05/18 14:40:24 INFO metastore.ObjectStore: Initialized ObjectStore
16/05/18 14:40:24 INFO metastore.HiveMetaStore: Added admin role in
metastore
16/05/18 14:40:24 INFO metastore.HiveMetaStore: Added public role in
metastore
16/05/18 14:40:24 INFO metastore.HiveMetaStore: No user is added in admin
role, since config is empty
16/05/18 14:40:24 INFO metastore.HiveMetaStore: 0: get_table : db=default
tbl=dual
16/05/18 14:40:24 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=get_table : db=default tbl=dual
16/05/18 14:40:24 INFO metastore.HiveMetaStore: 0: Shutting down the object
store...
16/05/18 14:40:24 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=Shutting down the object store...
16/05/18 14:40:24 INFO metastore.HiveMetaStore: 0: Metastore shutdown
complete.
16/05/18 14:40:24 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=Metastore shutdown complete.
16/05/18 14:40:24 INFO metastore.HiveMetaStore: 0: get_table : db=default
tbl=dual
16/05/18 14:40:24 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=get_table : db=default tbl=dual
16/05/18 14:40:24 INFO metastore.HiveMetaStore: 0: Opening raw store with
implemenation class:org.apache.hadoop.hive.metastore.ObjectStore
16/05/18 14:40:24 INFO metastore.ObjectStore: ObjectStore, initialize called
16/05/18 14:40:24 INFO DataNucleus.Query: Reading in results for query
"org.datanucleus.store.rdbms.query.SQLQuery@0" since the connection used is
closing
16/05/18 14:40:24 INFO metastore.MetaStoreDirectSql: Using direct SQL,
underlying DB is DERBY
16/05/18 14:40:24 INFO metastore.ObjectStore: Initialized ObjectStore
16/05/18 14:40:24 INFO metastore.HiveMetaStore: 0: Shutting down the object
store...
16/05/18 14:40:24 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=Shutting down the object store...
16/05/18 14:40:24 INFO metastore.HiveMetaStore: 0: Metastore shutdown
complete.
16/05/18 14:40:24 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=Metastore shutdown complete.
16/05/18 14:40:24 INFO property.AppProps: using app.id:
954F6CFECF794BC191AB3296A6FAC1F5
16/05/18 14:40:24 INFO metastore.HiveMetaStore: 0: get_table : db=default
tbl=keyvalue
16/05/18 14:40:24 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=get_table : db=default tbl=keyvalue
16/05/18 14:40:24 INFO metastore.HiveMetaStore: 0: Opening raw store with
implemenation class:org.apache.hadoop.hive.metastore.ObjectStore
16/05/18 14:40:24 INFO metastore.ObjectStore: ObjectStore, initialize called
16/05/18 14:40:24 INFO DataNucleus.Query: Reading in results for query
"org.datanucleus.store.rdbms.query.SQLQuery@0" since the connection used is
closing
16/05/18 14:40:24 INFO metastore.MetaStoreDirectSql: Using direct SQL,
underlying DB is DERBY
16/05/18 14:40:24 INFO metastore.ObjectStore: Initialized ObjectStore
16/05/18 14:40:24 INFO metastore.HiveMetaStore: 0: Shutting down the object
store...
16/05/18 14:40:24 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=Shutting down the object store...
16/05/18 14:40:24 INFO metastore.HiveMetaStore: 0: Metastore shutdown
complete.
16/05/18 14:40:24 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=Metastore shutdown complete.
16/05/18 14:40:24 INFO metastore.HiveMetaStore: 0: get_table : db=default
tbl=keyvalue
16/05/18 14:40:24 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=get_table : db=default tbl=keyvalue
16/05/18 14:40:24 INFO metastore.HiveMetaStore: 0: Opening raw store with
implemenation class:org.apache.hadoop.hive.metastore.ObjectStore
16/05/18 14:40:24 INFO metastore.ObjectStore: ObjectStore, initialize called
16/05/18 14:40:24 INFO DataNucleus.Query: Reading in results for query
"org.datanucleus.store.rdbms.query.SQLQuery@0" since the connection used is
closing
16/05/18 14:40:24 INFO metastore.MetaStoreDirectSql: Using direct SQL,
underlying DB is DERBY
16/05/18 14:40:24 INFO metastore.ObjectStore: Initialized ObjectStore
16/05/18 14:40:24 INFO metastore.HiveMetaStore: 0: Shutting down the object
store...
16/05/18 14:40:24 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=Shutting down the object store...
16/05/18 14:40:24 INFO metastore.HiveMetaStore: 0: Metastore shutdown
complete.
16/05/18 14:40:24 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=Metastore shutdown complete.
16/05/18 14:40:24 INFO metastore.HiveMetaStore: 0: get_table : db=default
tbl=keyvalue2
16/05/18 14:40:24 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=get_table : db=default tbl=keyvalue2
16/05/18 14:40:24 INFO metastore.HiveMetaStore: 0: Opening raw store with
implemenation class:org.apache.hadoop.hive.metastore.ObjectStore
16/05/18 14:40:24 INFO metastore.ObjectStore: ObjectStore, initialize called
16/05/18 14:40:24 INFO DataNucleus.Query: Reading in results for query
"org.datanucleus.store.rdbms.query.SQLQuery@0" since the connection used is
closing
16/05/18 14:40:24 INFO metastore.MetaStoreDirectSql: Using direct SQL,
underlying DB is DERBY
16/05/18 14:40:24 INFO metastore.ObjectStore: Initialized ObjectStore
16/05/18 14:40:24 INFO metastore.HiveMetaStore: 0: Shutting down the object
store...
16/05/18 14:40:24 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=Shutting down the object store...
16/05/18 14:40:24 INFO metastore.HiveMetaStore: 0: Metastore shutdown
complete.
16/05/18 14:40:24 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=Metastore shutdown complete.
16/05/18 14:40:24 INFO metastore.HiveMetaStore: 0: get_table : db=default
tbl=keyvalue2
16/05/18 14:40:24 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=get_table : db=default tbl=keyvalue2
16/05/18 14:40:24 INFO metastore.HiveMetaStore: 0: Opening raw store with
implemenation class:org.apache.hadoop.hive.metastore.ObjectStore
16/05/18 14:40:24 INFO metastore.ObjectStore: ObjectStore, initialize called
16/05/18 14:40:24 INFO DataNucleus.Query: Reading in results for query
"org.datanucleus.store.rdbms.query.SQLQuery@0" since the connection used is
closing
16/05/18 14:40:24 INFO metastore.MetaStoreDirectSql: Using direct SQL,
underlying DB is DERBY
16/05/18 14:40:24 INFO metastore.ObjectStore: Initialized ObjectStore
16/05/18 14:40:24 INFO metastore.HiveMetaStore: 0: Shutting down the object
store...
16/05/18 14:40:24 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=Shutting down the object store...
16/05/18 14:40:24 INFO metastore.HiveMetaStore: 0: Metastore shutdown
complete.
16/05/18 14:40:24 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=Metastore shutdown complete.
16/05/18 14:40:24 INFO util.Util: resolving application jar from found main
method on: cascading.hive.HiveDemo
16/05/18 14:40:24 INFO planner.HadoopPlanner: using application jar:
/home/jmill383/cascading-hive/demo/build/libs/cascading-hive-demo-1.0.jar
16/05/18 14:40:25 INFO flow.Flow: [uppercase kv -> kv2 ] executed rule
registry: MapReduceHadoopRuleRegistry, completed as: SUCCESS, in: 00:00.050
16/05/18 14:40:25 INFO flow.Flow: [uppercase kv -> kv2 ] rule registry:
MapReduceHadoopRuleRegistry, supports assembly with steps: 1, nodes: 1
16/05/18 14:40:25 INFO flow.Flow: [uppercase kv -> kv2 ] rule registry:
MapReduceHadoopRuleRegistry, result was selected using: 'default
comparator: selects plan with fewest steps and fewest nodes'
16/05/18 14:40:25 INFO Configuration.deprecation:
mapred.used.genericoptionsparser is deprecated. Instead, use
mapreduce.client.genericoptionsparser.used
16/05/18 14:40:25 INFO Configuration.deprecation: mapred.working.dir is
deprecated. Instead, use mapreduce.job.working.dir
16/05/18 14:40:25 INFO Configuration.deprecation: mapred.input.dir is
deprecated. Instead, use mapreduce.input.fileinputformat.inputdir
16/05/18 14:40:25 INFO Configuration.deprecation: mapred.output.dir is
deprecated. Instead, use mapreduce.output.fileoutputformat.outputdir
16/05/18 14:40:25 INFO Configuration.deprecation: mapred.output.compress is
deprecated. Instead, use mapreduce.output.fileoutputformat.compress
16/05/18 14:40:25 INFO Configuration.deprecation: mapred.output.key.class
is deprecated. Instead, use mapreduce.job.output.key.class
16/05/18 14:40:25 INFO Configuration.deprecation: mapred.output.value.class
is deprecated. Instead, use mapreduce.job.output.value.class
16/05/18 14:40:25 INFO util.Version: Concurrent, Inc - Cascading
3.1.0-wip-57
16/05/18 14:40:25 INFO cascade.Cascade: [uppercase kv -> kv2 +l...] starting
16/05/18 14:40:25 INFO cascade.Cascade: [uppercase kv -> kv2 +l...]
parallel execution of flows is enabled: false
16/05/18 14:40:25 INFO cascade.Cascade: [uppercase kv -> kv2 +l...]
executing total flows: 3
16/05/18 14:40:25 INFO cascade.Cascade: [uppercase kv -> kv2 +l...]
allocating management threads: 1
16/05/18 14:40:25 INFO cascade.Cascade: [uppercase kv -> kv2 +l...]
starting flow: load data into dual
16/05/18 14:40:25 INFO flow.Flow: [load data into dual] at least one sink
is marked for delete
16/05/18 14:40:25 INFO flow.Flow: [load data into dual] sink oldest
modified date: Wed Dec 31 18:59:59 EST 1969
16/05/18 14:40:25 INFO metastore.HiveMetaStore: 1: get_table : db=default
tbl=dual
16/05/18 14:40:25 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=get_table : db=default tbl=dual
16/05/18 14:40:25 INFO metastore.HiveMetaStore: 1: Opening raw store with
implemenation class:org.apache.hadoop.hive.metastore.ObjectStore
16/05/18 14:40:25 INFO metastore.ObjectStore: ObjectStore, initialize called
16/05/18 14:40:25 INFO DataNucleus.Query: Reading in results for query
"org.datanucleus.store.rdbms.query.SQLQuery@0" since the connection used is
closing
16/05/18 14:40:25 INFO metastore.MetaStoreDirectSql: Using direct SQL,
underlying DB is DERBY
16/05/18 14:40:25 INFO metastore.ObjectStore: Initialized ObjectStore
16/05/18 14:40:25 INFO hive.HiveTap: strict mode: comparing existing hive
table with table descriptor
16/05/18 14:40:25 INFO metastore.HiveMetaStore: 1: Shutting down the object
store...
16/05/18 14:40:25 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=Shutting down the object store...
16/05/18 14:40:25 INFO metastore.HiveMetaStore: 1: Metastore shutdown
complete.
16/05/18 14:40:25 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=Metastore shutdown complete.
16/05/18 14:40:25 INFO metastore.HiveMetaStore: 2: get_all_databases
16/05/18 14:40:25 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=get_all_databases
16/05/18 14:40:25 INFO metastore.HiveMetaStore: 2: Opening raw store with
implemenation class:org.apache.hadoop.hive.metastore.ObjectStore
16/05/18 14:40:25 INFO metastore.ObjectStore: ObjectStore, initialize called
16/05/18 14:40:25 INFO DataNucleus.Query: Reading in results for query
"org.datanucleus.store.rdbms.query.SQLQuery@0" since the connection used is
closing
16/05/18 14:40:25 INFO metastore.MetaStoreDirectSql: Using direct SQL,
underlying DB is DERBY
16/05/18 14:40:25 INFO metastore.ObjectStore: Initialized ObjectStore
16/05/18 14:40:25 INFO metastore.HiveMetaStore: 2: get_functions:
db=default pat=*
16/05/18 14:40:25 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=get_functions: db=default pat=*
16/05/18 14:40:25 INFO DataNucleus.Datastore: The class
"org.apache.hadoop.hive.metastore.model.MResourceUri" is tagged as
"embedded-only" so does not have its own datastore table.
16/05/18 14:40:25 INFO session.SessionState: Created local directory:
/tmp/e9c0df1e-0647-47df-ad9b-ddc1dcdb6054_resources
16/05/18 14:40:25 INFO session.SessionState: Created HDFS directory:
/tmp/hive/jmill383/e9c0df1e-0647-47df-ad9b-ddc1dcdb6054
16/05/18 14:40:25 INFO session.SessionState: Created local directory:
/tmp/jmill383/e9c0df1e-0647-47df-ad9b-ddc1dcdb6054
16/05/18 14:40:25 INFO session.SessionState: Created HDFS directory:
/tmp/hive/jmill383/e9c0df1e-0647-47df-ad9b-ddc1dcdb6054/_tmp_space.db
16/05/18 14:40:25 INFO hive.HiveQueryRunner: running hive query: 'load data
local inpath
'file:///home/jmill383/cascading-hive/demo/src/main/resources/data.txt'
overwrite into table dual'
16/05/18 14:40:25 INFO log.PerfLogger: <PERFLOG method=Driver.run
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:25 INFO log.PerfLogger: <PERFLOG method=TimeToSubmit
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:25 INFO log.PerfLogger: <PERFLOG method=compile
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:25 INFO log.PerfLogger: <PERFLOG method=parse
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:25 INFO parse.ParseDriver: Parsing command: load data local
inpath
'file:///home/jmill383/cascading-hive/demo/src/main/resources/data.txt'
overwrite into table dual
16/05/18 14:40:26 INFO parse.ParseDriver: Parse Completed
16/05/18 14:40:26 INFO log.PerfLogger: </PERFLOG method=parse
start=1463596825623 end=1463596826130 duration=507
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:26 INFO log.PerfLogger: <PERFLOG method=semanticAnalyze
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:26 INFO metastore.HiveMetaStore: 2: get_table : db=default
tbl=dual
16/05/18 14:40:26 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=get_table : db=default tbl=dual
16/05/18 14:40:26 INFO ql.Driver: Semantic Analysis Completed
16/05/18 14:40:26 INFO log.PerfLogger: </PERFLOG method=semanticAnalyze
start=1463596826132 end=1463596826291 duration=159
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:26 INFO ql.Driver: Returning Hive schema:
Schema(fieldSchemas:null, properties:null)
16/05/18 14:40:26 INFO log.PerfLogger: </PERFLOG method=compile
start=1463596825602 end=1463596826295 duration=693
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:26 INFO ql.Driver: Concurrency mode is disabled, not
creating a lock manager
16/05/18 14:40:26 INFO log.PerfLogger: <PERFLOG method=Driver.execute
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:26 INFO ql.Driver: Starting
command(queryId=jmill383_20160518144025_dbe97c43-6e94-43a9-bf17-14dd2d88f490):
load data local inpath
'file:///home/jmill383/cascading-hive/demo/src/main/resources/data.txt'
overwrite into table dual
16/05/18 14:40:26 INFO log.PerfLogger: </PERFLOG method=TimeToSubmit
start=1463596825602 end=1463596826298 duration=696
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:26 INFO log.PerfLogger: <PERFLOG method=runTasks
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:26 INFO log.PerfLogger: <PERFLOG method=task.MOVE.Stage-0
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:26 INFO ql.Driver: Starting task [Stage-0:MOVE] in serial
mode
Loading data to table default.dual
16/05/18 14:40:26 INFO exec.Task: Loading data to table default.dual from
file:/home/jmill383/cascading-hive/demo/src/main/resources/data.txt
16/05/18 14:40:26 INFO metastore.HiveMetaStore: 2: get_table : db=default
tbl=dual
16/05/18 14:40:26 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=get_table : db=default tbl=dual
16/05/18 14:40:26 INFO metastore.HiveMetaStore: 2: get_table : db=default
tbl=dual
16/05/18 14:40:26 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=get_table : db=default tbl=dual
16/05/18 14:40:26 INFO common.FileUtils: deleting
hdfs://localhost:8025/user/hive/warehouse/dual/data.txt
16/05/18 14:40:26 INFO fs.TrashPolicyDefault: Namenode trash configuration:
Deletion interval = 0 minutes, Emptier interval = 0 minutes.
16/05/18 14:40:26 INFO metadata.Hive: Replacing
src:file:/home/jmill383/cascading-hive/demo/src/main/resources/data.txt,
dest: hdfs://localhost:8025/user/hive/warehouse/dual/data.txt, Status:true
16/05/18 14:40:26 INFO metastore.HiveMetaStore: 2: alter_table: db=default
tbl=dual newtbl=dual
16/05/18 14:40:26 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=alter_table: db=default tbl=dual newtbl=dual
16/05/18 14:40:26 INFO hive.log: Updating table stats fast for dual
16/05/18 14:40:26 INFO hive.log: Updated size of table dual to 2
16/05/18 14:40:26 INFO log.PerfLogger: <PERFLOG method=task.STATS.Stage-1
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:26 INFO ql.Driver: Starting task [Stage-1:STATS] in serial
mode
16/05/18 14:40:26 INFO exec.StatsTask: Executing stats task
16/05/18 14:40:26 INFO metastore.HiveMetaStore: 2: get_table : db=default
tbl=dual
16/05/18 14:40:26 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=get_table : db=default tbl=dual
16/05/18 14:40:26 INFO metastore.HiveMetaStore: 2: get_table : db=default
tbl=dual
16/05/18 14:40:26 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=get_table : db=default tbl=dual
16/05/18 14:40:26 INFO metastore.HiveMetaStore: 2: alter_table: db=default
tbl=dual newtbl=dual
16/05/18 14:40:26 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=alter_table: db=default tbl=dual newtbl=dual
16/05/18 14:40:26 INFO hive.log: Updating table stats fast for dual
16/05/18 14:40:26 INFO hive.log: Updated size of table dual to 2
Table default.dual stats: [numFiles=1, numRows=0, totalSize=2,
rawDataSize=0]
16/05/18 14:40:26 INFO exec.Task: Table default.dual stats: [numFiles=1,
numRows=0, totalSize=2, rawDataSize=0]
16/05/18 14:40:26 INFO log.PerfLogger: </PERFLOG method=runTasks
start=1463596826298 end=1463596826726 duration=428
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:26 INFO log.PerfLogger: </PERFLOG method=Driver.execute
start=1463596826296 end=1463596826726 duration=430
from=org.apache.hadoop.hive.ql.Driver>
OK
16/05/18 14:40:26 INFO ql.Driver: OK
16/05/18 14:40:26 INFO log.PerfLogger: <PERFLOG method=releaseLocks
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:26 INFO log.PerfLogger: </PERFLOG method=releaseLocks
start=1463596826727 end=1463596826727 duration=0
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:26 INFO log.PerfLogger: </PERFLOG method=Driver.run
start=1463596825602 end=1463596826727 duration=1125
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:26 INFO log.PerfLogger: <PERFLOG method=releaseLocks
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:26 INFO log.PerfLogger: </PERFLOG method=releaseLocks
start=1463596826728 end=1463596826728 duration=0
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:26 INFO cascade.Cascade: [uppercase kv -> kv2 +l...]
completed flow: load data into dual
16/05/18 14:40:26 INFO cascade.Cascade: [uppercase kv -> kv2 +l...]
starting flow: select data from dual into keyvalue
16/05/18 14:40:26 INFO flow.Flow: [select data from dual ...] at least one
sink is marked for delete
16/05/18 14:40:26 INFO flow.Flow: [select data from dual ...] sink oldest
modified date: Wed Dec 31 18:59:59 EST 1969
16/05/18 14:40:26 INFO metastore.HiveMetaStore: 1: get_table : db=default
tbl=keyvalue
16/05/18 14:40:26 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=get_table : db=default tbl=keyvalue
16/05/18 14:40:26 INFO metastore.HiveMetaStore: 1: Opening raw store with
implemenation class:org.apache.hadoop.hive.metastore.ObjectStore
16/05/18 14:40:26 INFO metastore.ObjectStore: ObjectStore, initialize called
16/05/18 14:40:26 INFO DataNucleus.Query: Reading in results for query
"org.datanucleus.store.rdbms.query.SQLQuery@0" since the connection used is
closing
16/05/18 14:40:26 INFO metastore.MetaStoreDirectSql: Using direct SQL,
underlying DB is DERBY
16/05/18 14:40:26 INFO metastore.ObjectStore: Initialized ObjectStore
16/05/18 14:40:26 INFO hive.HiveTap: strict mode: comparing existing hive
table with table descriptor
16/05/18 14:40:26 INFO metastore.HiveMetaStore: 1: Shutting down the object
store...
16/05/18 14:40:26 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=Shutting down the object store...
16/05/18 14:40:26 INFO metastore.HiveMetaStore: 1: Metastore shutdown
complete.
16/05/18 14:40:26 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=Metastore shutdown complete.
16/05/18 14:40:26 INFO session.SessionState: Created local directory:
/tmp/b51c81d8-23b9-49ea-b012-92f81bc1c0ce_resources
16/05/18 14:40:26 INFO session.SessionState: Created HDFS directory:
/tmp/hive/jmill383/b51c81d8-23b9-49ea-b012-92f81bc1c0ce
16/05/18 14:40:26 INFO session.SessionState: Created local directory:
/tmp/jmill383/b51c81d8-23b9-49ea-b012-92f81bc1c0ce
16/05/18 14:40:26 INFO session.SessionState: Created HDFS directory:
/tmp/hive/jmill383/b51c81d8-23b9-49ea-b012-92f81bc1c0ce/_tmp_space.db
16/05/18 14:40:26 INFO hive.HiveQueryRunner: running hive query: 'insert
overwrite table keyvalue select 'Hello' as key, 'hive!' as value from dual'
16/05/18 14:40:26 INFO log.PerfLogger: <PERFLOG method=Driver.run
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:26 INFO log.PerfLogger: <PERFLOG method=TimeToSubmit
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:26 INFO log.PerfLogger: <PERFLOG method=compile
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:26 INFO log.PerfLogger: <PERFLOG method=parse
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:26 INFO parse.ParseDriver: Parsing command: insert overwrite
table keyvalue select 'Hello' as key, 'hive!' as value from dual
16/05/18 14:40:26 INFO parse.ParseDriver: Parse Completed
16/05/18 14:40:26 INFO log.PerfLogger: </PERFLOG method=parse
start=1463596826826 end=1463596826833 duration=7
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:26 INFO log.PerfLogger: <PERFLOG method=semanticAnalyze
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:26 INFO parse.CalcitePlanner: Starting Semantic Analysis
16/05/18 14:40:26 INFO parse.CalcitePlanner: Completed phase 1 of Semantic
Analysis
16/05/18 14:40:26 INFO parse.CalcitePlanner: Get metadata for source tables
16/05/18 14:40:26 INFO metastore.HiveMetaStore: 3: get_table : db=default
tbl=dual
16/05/18 14:40:26 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=get_table : db=default tbl=dual
16/05/18 14:40:26 INFO metastore.HiveMetaStore: 3: Opening raw store with
implemenation class:org.apache.hadoop.hive.metastore.ObjectStore
16/05/18 14:40:26 INFO metastore.ObjectStore: ObjectStore, initialize called
16/05/18 14:40:26 INFO DataNucleus.Query: Reading in results for query
"org.datanucleus.store.rdbms.query.SQLQuery@0" since the connection used is
closing
16/05/18 14:40:26 INFO metastore.MetaStoreDirectSql: Using direct SQL,
underlying DB is DERBY
16/05/18 14:40:26 INFO metastore.ObjectStore: Initialized ObjectStore
16/05/18 14:40:26 INFO parse.CalcitePlanner: Get metadata for subqueries
16/05/18 14:40:26 INFO parse.CalcitePlanner: Get metadata for destination
tables
16/05/18 14:40:26 INFO metastore.HiveMetaStore: 3: get_table : db=default
tbl=keyvalue
16/05/18 14:40:26 INFO HiveMetaStore.audit: ugi=jmill383
ip=unknown-ip-addr    cmd=get_table : db=default tbl=keyvalue
16/05/18 14:40:26 INFO parse.CalcitePlanner: Completed getting MetaData in
Semantic Analysis
16/05/18 14:40:26 INFO parse.BaseSemanticAnalyzer: Not invoking CBO because
the statement has too few joins
16/05/18 14:40:26 INFO common.FileUtils: Creating directory if it doesn't
exist:
hdfs://localhost:8025/user/hive/warehouse/keyvalue/.hive-staging_hive_2016-05-18_14-40-26_826_7796779660082577343-1
16/05/18 14:40:26 INFO parse.CalcitePlanner: Set stats collection dir :
hdfs://localhost:8025/user/hive/warehouse/keyvalue/.hive-staging_hive_2016-05-18_14-40-26_826_7796779660082577343-1/-ext-10001
16/05/18 14:40:27 INFO ppd.OpProcFactory: Processing for FS(2)
16/05/18 14:40:27 INFO ppd.OpProcFactory: Processing for SEL(1)
16/05/18 14:40:27 INFO ppd.OpProcFactory: Processing for TS(0)
16/05/18 14:40:27 INFO log.PerfLogger: <PERFLOG method=partition-retrieving
from=org.apache.hadoop.hive.ql.optimizer.ppr.PartitionPruner>
16/05/18 14:40:27 INFO log.PerfLogger: </PERFLOG
method=partition-retrieving start=1463596827044 end=1463596827044
duration=0 from=org.apache.hadoop.hive.ql.optimizer.ppr.PartitionPruner>
16/05/18 14:40:27 INFO optimizer.GenMRFileSink1: using
CombineHiveInputformat for the merge job
16/05/18 14:40:27 INFO physical.NullScanTaskDispatcher: Looking for table
scans where optimization is applicable
16/05/18 14:40:27 INFO physical.NullScanTaskDispatcher: Found 0 null table
scans
16/05/18 14:40:27 INFO physical.NullScanTaskDispatcher: Looking for table
scans where optimization is applicable
16/05/18 14:40:27 INFO physical.NullScanTaskDispatcher: Found 0 null table
scans
16/05/18 14:40:27 INFO physical.NullScanTaskDispatcher: Looking for table
scans where optimization is applicable
16/05/18 14:40:27 INFO physical.NullScanTaskDispatcher: Found 0 null table
scans
16/05/18 14:40:27 INFO parse.CalcitePlanner: Completed plan generation
16/05/18 14:40:27 INFO ql.Driver: Semantic Analysis Completed
16/05/18 14:40:27 INFO log.PerfLogger: </PERFLOG method=semanticAnalyze
start=1463596826833 end=1463596827060 duration=227
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:27 INFO ql.Driver: Returning Hive schema:
Schema(fieldSchemas:[FieldSchema(name:key, type:string, comment:null),
FieldSchema(name:value, type:string, comment:null)], properties:null)
16/05/18 14:40:27 INFO log.PerfLogger: </PERFLOG method=compile
start=1463596826825 end=1463596827060 duration=235
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:27 INFO ql.Driver: Concurrency mode is disabled, not
creating a lock manager
16/05/18 14:40:27 INFO log.PerfLogger: <PERFLOG method=Driver.execute
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:27 INFO ql.Driver: Starting
command(queryId=jmill383_20160518144026_b7cbea99-d34d-41da-8582-3be591e5d282):
insert overwrite table keyvalue select 'Hello' as key, 'hive!' as value
from dual
Query ID = jmill383_20160518144026_b7cbea99-d34d-41da-8582-3be591e5d282
16/05/18 14:40:27 INFO ql.Driver: Query ID =
jmill383_20160518144026_b7cbea99-d34d-41da-8582-3be591e5d282
Total jobs = 3
16/05/18 14:40:27 INFO ql.Driver: Total jobs = 3
16/05/18 14:40:27 INFO log.PerfLogger: </PERFLOG method=TimeToSubmit
start=1463596826825 end=1463596827061 duration=236
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:27 INFO log.PerfLogger: <PERFLOG method=runTasks
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:27 INFO log.PerfLogger: <PERFLOG method=task.MAPRED.Stage-1
from=org.apache.hadoop.hive.ql.Driver>
Launching Job 1 out of 3
16/05/18 14:40:27 INFO ql.Driver: Launching Job 1 out of 3
16/05/18 14:40:27 INFO ql.Driver: Starting task [Stage-1:MAPRED] in serial
mode
Number of reduce tasks is set to 0 since there's no reduce operator
16/05/18 14:40:27 INFO exec.Task: Number of reduce tasks is set to 0 since
there's no reduce operator
16/05/18 14:40:27 INFO ql.Context: New scratch dir is
hdfs://localhost:8025/tmp/hive/jmill383/b51c81d8-23b9-49ea-b012-92f81bc1c0ce/hive_2016-05-18_14-40-26_826_7796779660082577343-1
16/05/18 14:40:27 INFO mr.ExecDriver: Using
org.apache.hadoop.hive.ql.io.CombineHiveInputFormat
16/05/18 14:40:27 INFO exec.Utilities: Processing alias dual
16/05/18 14:40:27 INFO exec.Utilities: Adding input file
hdfs://localhost:8025/user/hive/warehouse/dual
16/05/18 14:40:27 INFO exec.Utilities: Content Summary not cached for
hdfs://localhost:8025/user/hive/warehouse/dual
16/05/18 14:40:27 INFO ql.Context: New scratch dir is
hdfs://localhost:8025/tmp/hive/jmill383/b51c81d8-23b9-49ea-b012-92f81bc1c0ce/hive_2016-05-18_14-40-26_826_7796779660082577343-1
16/05/18 14:40:27 INFO log.PerfLogger: <PERFLOG method=serializePlan
from=org.apache.hadoop.hive.ql.exec.Utilities>
16/05/18 14:40:27 INFO exec.Utilities: Serializing MapWork via kryo
16/05/18 14:40:27 INFO log.PerfLogger: </PERFLOG method=serializePlan
start=1463596827125 end=1463596827182 duration=57
from=org.apache.hadoop.hive.ql.exec.Utilities>
16/05/18 14:40:27 INFO Configuration.deprecation: mapred.submit.replication
is deprecated. Instead, use mapreduce.client.submit.file.replication
16/05/18 14:40:27 ERROR mr.ExecDriver: yarn
16/05/18 14:40:27 INFO client.RMProxy: Connecting to ResourceManager at /
0.0.0.0:8032
16/05/18 14:40:27 INFO fs.FSStatsPublisher: created :
hdfs://localhost:8025/user/hive/warehouse/keyvalue/.hive-staging_hive_2016-05-18_14-40-26_826_7796779660082577343-1/-ext-10001
16/05/18 14:40:27 INFO client.RMProxy: Connecting to ResourceManager at /
0.0.0.0:8032
16/05/18 14:40:27 INFO exec.Utilities: PLAN PATH =
hdfs://localhost:8025/tmp/hive/jmill383/b51c81d8-23b9-49ea-b012-92f81bc1c0ce/hive_2016-05-18_14-40-26_826_7796779660082577343-1/-mr-10004/2243cc23-ce83-4a00-87ba-d1970db29def/map.xml
16/05/18 14:40:27 INFO exec.Utilities: PLAN PATH =
hdfs://localhost:8025/tmp/hive/jmill383/b51c81d8-23b9-49ea-b012-92f81bc1c0ce/hive_2016-05-18_14-40-26_826_7796779660082577343-1/-mr-10004/2243cc23-ce83-4a00-87ba-d1970db29def/reduce.xml
16/05/18 14:40:27 INFO exec.Utilities: ***************non-local
mode***************
16/05/18 14:40:27 INFO exec.Utilities: local path =
hdfs://localhost:8025/tmp/hive/jmill383/b51c81d8-23b9-49ea-b012-92f81bc1c0ce/hive_2016-05-18_14-40-26_826_7796779660082577343-1/-mr-10004/2243cc23-ce83-4a00-87ba-d1970db29def/reduce.xml
16/05/18 14:40:27 INFO exec.Utilities: Open file to read in plan:
hdfs://localhost:8025/tmp/hive/jmill383/b51c81d8-23b9-49ea-b012-92f81bc1c0ce/hive_2016-05-18_14-40-26_826_7796779660082577343-1/-mr-10004/2243cc23-ce83-4a00-87ba-d1970db29def/reduce.xml
16/05/18 14:40:27 INFO exec.Utilities: File not found: File does not exist:
/tmp/hive/jmill383/b51c81d8-23b9-49ea-b012-92f81bc1c0ce/hive_2016-05-18_14-40-26_826_7796779660082577343-1/-mr-10004/2243cc23-ce83-4a00-87ba-d1970db29def/reduce.xml
    at
org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:66)
    at
org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:56)
    at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocationsUpdateTimes(FSNamesystem.java:1891)
    at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocationsInt(FSNamesystem.java:1832)
    at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1812)
    at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1784)
    at
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getBlockLocations(NameNodeRpcServer.java:542)
    at
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getBlockLocations(ClientNamenodeProtocolServerSideTranslatorPB.java:362)
    at
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
    at
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619)
    at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962)
    at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039)
    at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035)
    at java.security.AccessController.doPrivileged(Native Method)
    at javax.security.auth.Subject.doAs(Subject.java:415)
    at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
    at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033)

16/05/18 14:40:27 INFO exec.Utilities: No plan file found:
hdfs://localhost:8025/tmp/hive/jmill383/b51c81d8-23b9-49ea-b012-92f81bc1c0ce/hive_2016-05-18_14-40-26_826_7796779660082577343-1/-mr-10004/2243cc23-ce83-4a00-87ba-d1970db29def/reduce.xml
16/05/18 14:40:27 WARN mapreduce.JobSubmitter: Hadoop command-line option
parsing not performed. Implement the Tool interface and execute your
application with ToolRunner to remedy this.
16/05/18 14:40:27 INFO log.PerfLogger: <PERFLOG method=getSplits
from=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat>
16/05/18 14:40:27 INFO exec.Utilities: PLAN PATH =
hdfs://localhost:8025/tmp/hive/jmill383/b51c81d8-23b9-49ea-b012-92f81bc1c0ce/hive_2016-05-18_14-40-26_826_7796779660082577343-1/-mr-10004/2243cc23-ce83-4a00-87ba-d1970db29def/map.xml
16/05/18 14:40:27 INFO io.CombineHiveInputFormat: Total number of paths: 1,
launching 1 threads to check non-combinable ones.
16/05/18 14:40:27 INFO io.CombineHiveInputFormat: CombineHiveInputSplit
creating pool for hdfs://localhost:8025/user/hive/warehouse/dual; using
filter path hdfs://localhost:8025/user/hive/warehouse/dual
16/05/18 14:40:27 INFO input.FileInputFormat: Total input paths to process
: 1
16/05/18 14:40:27 INFO input.CombineFileInputFormat: DEBUG: Terminated node
allocation with : CompletedNodes: 1, size left: 0
16/05/18 14:40:27 INFO io.CombineHiveInputFormat: number of splits 1
16/05/18 14:40:27 INFO io.CombineHiveInputFormat: Number of all splits 1
16/05/18 14:40:27 INFO log.PerfLogger: </PERFLOG method=getSplits
start=1463596827579 end=1463596827605 duration=26
from=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat>
16/05/18 14:40:27 INFO mapreduce.JobSubmitter: number of splits:1
16/05/18 14:40:27 INFO mapreduce.JobSubmitter: Submitting tokens for job:
job_1463594979064_0004
16/05/18 14:40:27 INFO impl.YarnClientImpl: Submitted application
application_1463594979064_0004
16/05/18 14:40:27 INFO mapreduce.Job: The url to track the job:
http://starchild:8088/proxy/application_1463594979064_0004/
Starting Job = job_1463594979064_0004, Tracking URL =
http://starchild:8088/proxy/application_1463594979064_0004/
16/05/18 14:40:27 INFO exec.Task: Starting Job = job_1463594979064_0004,
Tracking URL = http://starchild:8088/proxy/application_1463594979064_0004/
Kill Command = /opt/hadoop/bin/hadoop job  -kill job_1463594979064_0004
16/05/18 14:40:27 INFO exec.Task: Kill Command = /opt/hadoop/bin/hadoop
job  -kill job_1463594979064_0004
Hadoop job information for Stage-1: number of mappers: 0; number of
reducers: 0
16/05/18 14:40:31 INFO exec.Task: Hadoop job information for Stage-1:
number of mappers: 0; number of reducers: 0
16/05/18 14:40:31 WARN mapreduce.Counters: Group
org.apache.hadoop.mapred.Task$Counter is deprecated. Use
org.apache.hadoop.mapreduce.TaskCounter instead
2016-05-18 14:40:31,962 Stage-1 map = 0%,  reduce = 0%
16/05/18 14:40:32 INFO exec.Task: 2016-05-18 14:40:31,962 Stage-1 map =
0%,  reduce = 0%
16/05/18 14:40:32 WARN mapreduce.Counters: Group
org.apache.hadoop.mapred.Task$Counter is deprecated. Use
org.apache.hadoop.mapreduce.TaskCounter instead
Ended Job = job_1463594979064_0004 with errors
16/05/18 14:40:32 ERROR exec.Task: Ended Job = job_1463594979064_0004 with
errors
Error during job, obtaining debugging information...
16/05/18 14:40:32 ERROR exec.Task: Error during job, obtaining debugging
information...
16/05/18 14:40:32 INFO Configuration.deprecation: mapred.job.tracker is
deprecated. Instead, use mapreduce.jobtracker.address
Job Tracking URL:
http://starchild:8088/cluster/app/application_1463594979064_0004
16/05/18 14:40:32 ERROR exec.Task: Job Tracking URL:
http://starchild:8088/cluster/app/application_1463594979064_0004
16/05/18 14:40:32 INFO impl.YarnClientImpl: Killed application
application_1463594979064_0004
FAILED: Execution Error, return code 2 from
org.apache.hadoop.hive.ql.exec.mr.MapRedTask
16/05/18 14:40:32 ERROR ql.Driver: FAILED: Execution Error, return code 2
from org.apache.hadoop.hive.ql.exec.mr.MapRedTask
16/05/18 14:40:32 INFO log.PerfLogger: </PERFLOG method=Driver.execute
start=1463596827060 end=1463596832133 duration=5073
from=org.apache.hadoop.hive.ql.Driver>
MapReduce Jobs Launched:
16/05/18 14:40:32 INFO ql.Driver: MapReduce Jobs Launched:
16/05/18 14:40:32 WARN mapreduce.Counters: Group FileSystemCounters is
deprecated. Use org.apache.hadoop.mapreduce.FileSystemCounter instead
Stage-Stage-1:  HDFS Read: 0 HDFS Write: 0 FAIL
16/05/18 14:40:32 INFO ql.Driver: Stage-Stage-1:  HDFS Read: 0 HDFS Write:
0 FAIL
Total MapReduce CPU Time Spent: 0 msec
16/05/18 14:40:32 INFO ql.Driver: Total MapReduce CPU Time Spent: 0 msec
16/05/18 14:40:32 INFO log.PerfLogger: <PERFLOG method=releaseLocks
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:32 INFO log.PerfLogger: </PERFLOG method=releaseLocks
start=1463596832136 end=1463596832136 duration=0
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:32 INFO log.PerfLogger: <PERFLOG method=releaseLocks
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:32 INFO log.PerfLogger: </PERFLOG method=releaseLocks
start=1463596832139 end=1463596832139 duration=0
from=org.apache.hadoop.hive.ql.Driver>
16/05/18 14:40:32 WARN cascade.Cascade: [uppercase kv -> kv2 +l...] flow
failed: select data from dual into keyvalue
cascading.CascadingException: hive error 'FAILED: Execution Error, return
code 2 from org.apache.hadoop.hive.ql.exec.mr.MapRedTask' while running
query insert overwrite table keyvalue select 'Hello' as key, 'hive!' as
value from dual
    at cascading.flow.hive.HiveQueryRunner.run(HiveQueryRunner.java:131)
    at cascading.flow.hive.HiveQueryRunner.call(HiveQueryRunner.java:167)
    at cascading.flow.hive.HiveQueryRunner.call(HiveQueryRunner.java:41)
    at java.util.concurrent.FutureTask.run(FutureTask.java:262)
    at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
    at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
    at java.lang.Thread.run(Thread.java:745)
16/05/18 14:40:32 INFO cascade.Cascade: [uppercase kv -> kv2 +l...]
stopping all flows
16/05/18 14:40:32 INFO cascade.Cascade: [uppercase kv -> kv2 +l...]
stopping flow: uppercase kv -> kv2
16/05/18 14:40:32 INFO flow.Flow: [uppercase kv -> kv2 ] stopping all jobs
16/05/18 14:40:32 INFO flow.Flow: [uppercase kv -> kv2 ] stopping: (1/1)
.../hive/warehouse/keyvalue2
16/05/18 14:40:32 INFO flow.Flow: [uppercase kv -> kv2 ] stopped all jobs
16/05/18 14:40:32 INFO cascade.Cascade: [uppercase kv -> kv2 +l...]
stopping flow: select data from dual into keyvalue
16/05/18 14:40:32 INFO cascade.Cascade: [uppercase kv -> kv2 +l...]
stopping flow: load data into dual
16/05/18 14:40:32 INFO cascade.Cascade: [uppercase kv -> kv2 +l...] stopped
all flows
Exception in thread "main" cascading.cascade.CascadeException: flow failed:
select data from dual into keyvalue
    at cascading.cascade.BaseCascade$CascadeJob.call(BaseCascade.java:963)
    at cascading.cascade.BaseCascade$CascadeJob.call(BaseCascade.java:900)
    at java.util.concurrent.FutureTask.run(FutureTask.java:262)
    at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
    at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
    at java.lang.Thread.run(Thread.java:745)
Caused by: cascading.CascadingException: hive error 'FAILED: Execution
Error, return code 2 from org.apache.hadoop.hive.ql.exec.mr.MapRedTask'
while running query insert overwrite table keyvalue select 'Hello' as key,
'hive!' as value from dual
    at cascading.flow.hive.HiveQueryRunner.run(HiveQueryRunner.java:131)
    at cascading.flow.hive.HiveQueryRunner.call(HiveQueryRunner.java:167)
    at cascading.flow.hive.HiveQueryRunner.call(HiveQueryRunner.java:41)
    ... 4 more
[jmill383@starchild demo]$




On Wed, May 18, 2016 at 1:48 PM, Mich Talebzadeh <mich.talebza...@gmail.com>
wrote:

> Hi John,
>
> can you please a new thread for your problem so we can deal with
> separately.
>
> thanks
>
> Dr Mich Talebzadeh
>
>
>
> LinkedIn * 
> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>*
>
>
>
> http://talebzadehmich.wordpress.com
>
>
>
> On 18 May 2016 at 15:11, JOHN MILLER <jmill...@gmail.com> wrote:
>
>> Greetings Mitch
>>
>> I have an issue with running mapreduce in hive   I am getting a  FAILED:
>> Execution Error, return code 2 from
>> org.apache.hadoop.hive.ql.exec.mr.MapRedTask
>>
>> error while attemtiing to execute SELECT DISTINCT(fieldname) FROM TABLE
>> x  or SELECT  COUNT(*)  FROM TABLE x;;  Trying to run cascading-hive gives
>> me the same problem as well
>>
>> Please advise if u have come across this type of problem or generated
>> some ideas as to resolve this problema
>>
>> On Wed, May 18, 2016 at 9:53 AM, Mich Talebzadeh <
>> mich.talebza...@gmail.com> wrote:
>>
>>> Hi Kuldeep,
>>>
>>> Have you installed hive on any of these nodes.
>>>
>>> Hive is basically an API. You will also need to install sqoop as well if
>>> you are going to import data from other RDBMss like Oracle, Sybase etc.
>>>
>>> Hive has a very small footprint so my suggestion is to install it on all
>>> your boxes and permission granted to Haddop user say hduser.
>>>
>>> Hive will require a metadata in  a database of your choice. default is
>>> derby which I don't use. try to use a reasonable database. ours is on
>>> Oracle
>>>
>>>  Now under directory $HIVE_HOME/conf/hive-site.xml you can set up info
>>> about Hadoop and your metastore etc. You also need to set up environment
>>> variables for both Hadoop and hive in your start up script like .profile
>>> .kshrc etc
>>>
>>> Have a look anyway.
>>>
>>> HTH
>>>
>>> Dr Mich Talebzadeh
>>>
>>>
>>>
>>> LinkedIn * 
>>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
>>> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>*
>>>
>>>
>>>
>>> http://talebzadehmich.wordpress.com
>>>
>>>
>>>
>>> On 18 May 2016 at 13:49, Kuldeep Chitrakar <
>>> kuldeep.chitra...@synechron.com> wrote:
>>>
>>>> I have a very basic question regarding Hadoop & Hive setup.  I have 7
>>>> Machines say M1,M2,M3,M4,M5,M6,M7
>>>>
>>>>
>>>>
>>>> Hadoop Cluster Setup:
>>>>
>>>>
>>>>
>>>> Namenode: M1
>>>>
>>>> Seondary Namenode: M2
>>>>
>>>> Datanodes: M3,M4,M5
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> Now question is:
>>>>
>>>>
>>>>
>>>> Where do I need to install Hive.
>>>>
>>>> 1.       Should I install Hiverserver on M6
>>>>
>>>> a.       if yes does that machine needs core Hadoop JAR’s installed?
>>>>
>>>> b.      How this Hive server knows where Hadoop cluster is. What
>>>> configurations needs to be done?
>>>>
>>>> c.       How can we restrict this machine to be only hive server and
>>>> not datanode of Hadoop cluster?
>>>>
>>>>
>>>>
>>>> 2.       Where do we install Hive CLI
>>>>
>>>> a.       If I want to hive M7 as Hive CLI, then what needs to be
>>>> installed on this machine.
>>>>
>>>> b.      Any required configurations.
>>>>
>>>>
>>>>
>>>> Thanks,
>>>>
>>>> Kuldeep
>>>>
>>>>
>>>>
>>>
>>>
>>
>

Reply via email to