java.io.IOException: OS command error exit with return code: 155, error 
message: SLF4J: Class path contains multiple SLF4J bindings. SLF4J: Found 
binding in 
[jar:file:/opt/cloudera/parcels/CDH-6.2.0-1.cdh6.2.0.p0.967373/jars/log4j-slf4j-impl-2.8.2.jar!/org/slf4j/impl/StaticLoggerBinder.class]
 SLF4J: Found binding in 
[jar:file:/opt/modules/hadoop-2.7.7/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
 SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
explanation. SLF4J: Actual binding is of type 
[org.apache.logging.slf4j.Log4jLoggerFactory] Logging initialized using 
configuration in 
jar:file:/opt/cloudera/parcels/CDH-6.2.0-1.cdh6.2.0.p0.967373/jars/hive-common-2.1.1-cdh6.2.0.jar!/hive-log4j2.properties
 Async: false OK Time taken: 2.513 seconds OK Time taken: 0.385 seconds OK Time 
taken: 0.529 seconds OK Time taken: 0.521 seconds Query ID = 
root_20191208010521_cb1bca53-9910-4ab1-ac6e-a4de6dcdff39 Total jobs = 1 
Execution completed successfully MapredLocal task succeeded Launching Job 1 out 
of 1 Number of reduce tasks is set to 0 since there's no reduce operator 
19/12/08 01:05:42 INFO Configuration.deprecation: mapred.submit.replication is 
deprecated. Instead, use mapreduce.client.submit.file.replication 19/12/08 
01:05:44 INFO client.ConfiguredRMFailoverProxyProvider: Failing over to rm164 
19/12/08 01:05:44 WARN mapreduce.JobResourceUploader: Hadoop command-line 
option parsing not performed. Implement the Tool interface and execute your 
application with ToolRunner to remedy this. 19/12/08 01:05:49 INFO 
input.FileInputFormat: Total input paths to process : 1 19/12/08 01:05:49 INFO 
input.CombineFileInputFormat: DEBUG: Terminated node allocation with : 
CompletedNodes: 3, size left: 0 19/12/08 01:05:49 INFO mapreduce.JobSubmitter: 
number of splits:1 19/12/08 01:05:49 INFO mapreduce.JobSubmitter: Submitting 
tokens for job: job_1575269010295_0006 19/12/08 01:05:50 INFO 
impl.YarnClientImpl: Submitted application application_1575269010295_0006 
19/12/08 01:05:50 INFO mapreduce.Job: The url to track the job: 
http://miao2.com:8088/proxy/application_1575269010295_0006/ Starting Job = 
job_1575269010295_0006, Tracking URL = 
http://miao2.com:8088/proxy/application_1575269010295_0006/ Kill Command = 
/opt/modules/hadoop-2.7.7/bin/hadoop job  -kill job_1575269010295_0006 Hadoop 
job information for Stage-4: number of mappers: 1; number of reducers: 0 
19/12/08 01:06:05 WARN mapreduce.Counters: Group 
org.apache.hadoop.mapred.Task$Counter is deprecated. Use 
org.apache.hadoop.mapreduce.TaskCounter instead 2019-12-08 01:06:05,920 Stage-4 
map = 0%,  reduce = 0% 19/12/08 01:06:19 WARN counters.FileSystemCounterGroup: 
HDFS_BYTES_READ_EC is not a recognized counter. 19/12/08 01:06:19 WARN 
counters.FrameworkCounterGroup: MAP_PHYSICAL_MEMORY_BYTES_MAX is not a 
recognized counter. 19/12/08 01:06:19 WARN counters.FrameworkCounterGroup: 
MAP_VIRTUAL_MEMORY_BYTES_MAX is not a recognized counter. 2019-12-08 
01:06:19,383 Stage-4 map = 100%,  reduce = 0%, Cumulative CPU 2.52 sec 19/12/08 
01:06:20 WARN counters.FileSystemCounterGroup: HDFS_BYTES_READ_EC is not a 
recognized counter. 19/12/08 01:06:20 WARN counters.FrameworkCounterGroup: 
MAP_PHYSICAL_MEMORY_BYTES_MAX is not a recognized counter. 19/12/08 01:06:20 
WARN counters.FrameworkCounterGroup: MAP_VIRTUAL_MEMORY_BYTES_MAX is not a 
recognized counter. 19/12/08 01:06:20 WARN counters.FileSystemCounterGroup: 
HDFS_BYTES_READ_EC is not a recognized counter. 19/12/08 01:06:20 WARN 
counters.FrameworkCounterGroup: MAP_PHYSICAL_MEMORY_BYTES_MAX is not a 
recognized counter. 19/12/08 01:06:20 WARN counters.FrameworkCounterGroup: 
MAP_VIRTUAL_MEMORY_BYTES_MAX is not a recognized counter. MapReduce Total 
cumulative CPU time: 2 seconds 520 msec Ended Job = job_1575269010295_0006 
Loading data to table 
default.kylin_intermediate_emp_cube_da04d657_54f0_d436_1e6f_9b5c1d48d135 
19/12/08 01:06:20 ERROR hdfs.KeyProviderCache: Could not find uri with key 
[dfs.encryption.key.provider.uri] to create a keyProvider !! FAILED: Execution 
Error, return code -101 from org.apache.hadoop.hive.ql.exec.StatsTask. 
org.apache.hadoop.fs.FileStatus.isErasureCoded()Z MapReduce Jobs Launched:  
Stage-Stage-4: Map: 1   Cumulative CPU: 2.52 sec   HDFS Read: 8969 HDFS Write: 
551 SUCCESS Total MapReduce CPU Time Spent: 2 seconds 520 msec The command is:  
hive -e "USE default; DROP TABLE IF EXISTS 
kylin_intermediate_emp_cube_da04d657_54f0_d436_1e6f_9b5c1d48d135; CREATE 
EXTERNAL TABLE IF NOT EXISTS 
kylin_intermediate_emp_cube_da04d657_54f0_d436_1e6f_9b5c1d48d135 ( \`EMP_JOB\` 
string ,\`EMP_HIREDATE\` string ,\`EMP_DEPTNO\` int ,\`DEPT_DNAME\` string 
,\`DEPT_LOC\` int ,\`EMP_SAL\` double ) STORED AS SEQUENCEFILE LOCATION 
'hdfs://nameservice1/kylin/kylin_metadata/kylin-99121fe3-9dce-cdf0-f684-d97cb998c629/kylin_intermediate_emp_cube_da04d657_54f0_d436_1e6f_9b5c1d48d135';
 ALTER TABLE kylin_intermediate_emp_cube_da04d657_54f0_d436_1e6f_9b5c1d48d135 
SET TBLPROPERTIES('auto.purge'='true'); INSERT OVERWRITE TABLE 
\`kylin_intermediate_emp_cube_da04d657_54f0_d436_1e6f_9b5c1d48d135\` SELECT 
\`EMP\`.\`JOB\` as \`EMP_JOB\` ,\`EMP\`.\`HIREDATE\` as \`EMP_HIREDATE\` 
,\`EMP\`.\`DEPTNO\` as \`EMP_DEPTNO\` ,\`DEPT\`.\`DNAME\` as \`DEPT_DNAME\` 
,\`DEPT\`.\`LOC\` as \`DEPT_LOC\` ,\`EMP\`.\`SAL\` as \`EMP_SAL\`  FROM 
\`DEFAULT\`.\`EMP\` as \`EMP\` INNER JOIN \`DEFAULT\`.\`DEPT\` as \`DEPT\` ON 
\`EMP\`.\`DEPTNO\` = \`DEPT\`.\`DEPTNO\` WHERE 1=1; " --hiveconf 
hive.merge.mapredfiles=false --hiveconf hive.auto.convert.join=true --hiveconf 
dfs.replication=2 --hiveconf hive.exec.compress.output=true --hiveconf 
hive.auto.convert.join.noconditionaltask=true --hiveconf 
mapreduce.job.split.metainfo.maxsize=-1 --hiveconf hive.merge.mapfiles=false 
--hiveconf hive.auto.convert.join.noconditionaltask.size=100000000 --hiveconf 
hive.stats.autogather=true     at 
org.apache.kylin.common.util.CliCommandExecutor.execute(CliCommandExecutor.java:96)
  at 
org.apache.kylin.source.hive.CreateFlatHiveTableStep.createFlatHiveTable(CreateFlatHiveTableStep.java:62)
    at 
org.apache.kylin.source.hive.CreateFlatHiveTableStep.doWork(CreateFlatHiveTableStep.java:99)
         at 
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:167)
       at 
org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:71)
     at 
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:167)
       at 
org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:114)
       at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) 
     at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
     at java.lang.Thread.run(Thread.java:748)

Reply via email to