Can you disable kylin.storage.hbase.compression-codec in kylin.properties
and check once.

If this doesn't help you, please check and share the error logs from kylin
log file.

Regards,
Ashsih

On Wed, Aug 8, 2018 at 1:48 PM, Zhixiong Chen <[email protected]>
wrote:

> Hi:
>   非常非常感谢您的回复,我按您的方式,把这些都做了配置,包括环境变量,配置如下:
>
> <property>
>  <name>yarn.app.mapreduce.am.env</name>
>  <value>HADOOP_MAPRED_HOME=/bigdata/tools/hadoop-2.7.3</value>
> </property>
>
> <property>
>  <name>mapreduce.map.env</name>
>  <value>HADOOP_MAPRED_HOME=/bigdata/tools/hadoop-2.7.3</value>
> </property>
>
> <property>
>  <name>mapreduce.reduce.env</name>
>  <value>HADOOP_MAPRED_HOME=/bigdata/tools/hadoop-2.7.3</value>
> </property>
>  <property>
>   <name>yarn.app.mapreduce.am.admin.user.env</name>
>   <value>LD_LIBRARY_PATH=$HADOOP_COMMON_HOME/lib/native:
> $JAVA_LIBRARY_PATH</value>
>  </property>
>
> 另外,还将yarn.resourcemanager.address信息添加到了kylin_job_conf.
> xml这个文件中,原来Cube创建只能执行到第三步就报错,此时可以执行到第十步, 但此问题还是依然存在,只是之前在第三步#3 Step Name:
> Extract Fact Table Distinct Columns时报错,现在第十步又报同样的错误#10 Step Name: Build
> Cube In-Mem Duration: 20.19 mins Waiting: 0 seconds:, 请协助一下,我们万分感激!#10 Step
> Name: Build Cube In-Mem Duration: 20.19 mins Waiting: 0 seconds的错误信息如下:
>
>
> java.net.ConnectException: Call From hsmaster/10.9.0.86 to
> localhost:18032 failed on connection exception: java.net.ConnectException:
> Connection refused; For more details see:  http://wiki.apache.org/hadoop/
> ConnectionRefused
>         at sun.reflect.GeneratedConstructorAccessor76.newInstance(Unknown
> Source)
>         at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(
> DelegatingConstructorAccessorImpl.java:45)
>         at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
>         at org.apache.hadoop.net.NetUtils.wrapWithMessage(
> NetUtils.java:792)
>         at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:732)
>         at org.apache.hadoop.ipc.Client.call(Client.java:1479)
>         at org.apache.hadoop.ipc.Client.call(Client.java:1412)
>         at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.
> invoke(ProtobufRpcEngine.java:229)
>         at com.sun.proxy.$Proxy66.getNewApplication(Unknown Source)
>         at org.apache.hadoop.yarn.api.impl.pb.client.
> ApplicationClientProtocolPBClientImpl.getNewApplication(
> ApplicationClientProtocolPBClientImpl.java:221)
>         at sun.reflect.GeneratedMethodAccessor108.invoke(Unknown Source)
>         at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:498)
>         at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(
> RetryInvocationHandler.java:191)
>         at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(
> RetryInvocationHandler.java:102)
>         at com.sun.proxy.$Proxy67.getNewApplication(Unknown Source)
>         at org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.
> getNewApplication(YarnClientImpl.java:219)
>         at org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.
> createApplication(YarnClientImpl.java:227)
>         at org.apache.hadoop.mapred.ResourceMgrDelegate.getNewJobID(
> ResourceMgrDelegate.java:187)
>         at org.apache.hadoop.mapred.YARNRunner.getNewJobID(
> YARNRunner.java:231)
>         at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(
> JobSubmitter.java:153)
>         at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290)
>         at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:422)
>         at org.apache.hadoop.security.UserGroupInformation.doAs(
> UserGroupInformation.java:1698)
>         at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287)
>         at org.apache.kylin.engine.mr.common.AbstractHadoopJob.
> waitForCompletion(AbstractHadoopJob.java:175)
>         at org.apache.kylin.engine.mr.steps.InMemCuboidJob.run(
> InMemCuboidJob.java:121)
>         at org.apache.kylin.engine.mr.common.MapReduceExecutable.
> doWork(MapReduceExecutable.java:130)
>         at org.apache.kylin.job.execution.AbstractExecutable.
> execute(AbstractExecutable.java:162)
>         at org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(
> DefaultChainedExecutable.java:67)
>         at org.apache.kylin.job.execution.AbstractExecutable.
> execute(AbstractExecutable.java:162)
>         at org.apache.kylin.job.impl.threadpool.DefaultScheduler$
> JobRunner.run(DefaultScheduler.java:300)
>         at java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1142)
>         at java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:617)
>         at java.lang.Thread.run(Thread.java:748)
> Caused by: java.net.ConnectException: Connection refused
>         at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
>         at sun.nio.ch.SocketChannelImpl.finishConnect(
> SocketChannelImpl.java:717)
>         at org.apache.hadoop.net.SocketIOWithTimeout.connect(
> SocketIOWithTimeout.java:206)
>         at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:531)
>         at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:495)
>         at org.apache.hadoop.ipc.Client$Connection.setupConnection(
> Client.java:614)
>         at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(
> Client.java:712)
>         at org.apache.hadoop.ipc.Client$Connection.access$2900(Client.
> java:375)
>         at org.apache.hadoop.ipc.Client.getConnection(Client.java:1528)
>         at org.apache.hadoop.ipc.Client.call(Client.java:1451)
>         ... 31 more
> result code:2
>
>     ________________________________
>     Jason Lu
>   Email:[email protected]
>
>
> On 08/08/2018, 12:22 PM, "Ashish Singhi" <[email protected]> wrote:
>
>     Hi,
>
>     Can you check yarn application logs and check whether AM was started ?
>
>     I think you need to
>     set yarn.app.mapreduce.am.env, mapreduce.map.env, mapreduce.reduce.env
>     and yarn.app.mapreduce.am.admin.user.env in your mapred-site.xml file.
>
>     I too faced this issue a couple of days back, after setting this
>     configurations the issue resolved to me.
>
>     Regards,
>     Ashish
>
>     On Wed, Aug 8, 2018 at 7:54 AM, Zhixiong Chen <
> [email protected]>
>     wrote:
>
>     >
>     > Hi, chen<mailto:[email protected]>:
>     >
>     >
>     > 我们在使用Kylin创建Cube第三步的时候,kylin 返回 localhost:18032
>     > failed错误,我们花了很长时间也没有把这个问题解决,不知道是不是版本BUG还是其它原因,请求chen<mailto:
>     > [email protected]>协助一下,我们万分感激!以下是使用的版本和错误信息,请求chen<mailto:ch
> [email protected]
>     > >协助一下,谢谢!!!
>     > Hadoop(版本apache 2.7.3)集群全部启动正常,Kylin(版本2.3)在创建Cube到#3 Step Name:
> Extract
>     > Fact Table Distinct Columns时报如下错误:
>     >
>     >
>     > java.net.ConnectException: Call From hsmaster/10.9.0.86 to
>     > localhost:18032 failed on connection exception:
> java.net.ConnectException:
>     > Connection refused; For more details see:
> http://wiki.apache.org/hadoop/
>     > ConnectionRefused
>     >
>     >         at sun.reflect.GeneratedConstructorAccessor60
> .newInstance(Unknown
>     > Source)
>     >
>     >         at sun.reflect.DelegatingConstructorAccessorI
> mpl.newInstance(
>     > DelegatingConstructorAccessorImpl.java:45)
>     >
>     >         at java.lang.reflect.Constructor.
> newInstance(Constructor.java:423)
>     >
>     >         at org.apache.hadoop.net.NetUtils.wrapWithMessage(
>     > NetUtils.java:792)
>     >
>     >         at org.apache.hadoop.net.NetUtils.wrapException(
> NetUtils.java:732)
>     >
>     >         at org.apache.hadoop.ipc.Client.call(Client.java:1479)
>     >
>     >         at org.apache.hadoop.ipc.Client.call(Client.java:1412)
>     >
>     >         at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.
>     > invoke(ProtobufRpcEngine.java:229)
>     >
>     >         at com.sun.proxy.$Proxy65.getNewApplication(Unknown Source)
>     >
>     >         at org.apache.hadoop.yarn.api.impl.pb.client.
>     > ApplicationClientProtocolPBClientImpl.getNewApplication(
>     > ApplicationClientProtocolPBClientImpl.java:221)
>     >
>     >         at sun.reflect.GeneratedMethodAccessor87.invoke(Unknown
> Source)
>     >
>     >         at sun.reflect.DelegatingMethodAccessorImpl.invoke(
>     > DelegatingMethodAccessorImpl.java:43)
>     >
>     >         at java.lang.reflect.Method.invoke(Method.java:498)
>     >
>     >         at org.apache.hadoop.io.retry.RetryInvocationHandler.
> invokeMethod(
>     > RetryInvocationHandler.java:191)
>     >
>     >         at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(
>     > RetryInvocationHandler.java:102)
>     >
>     >         at com.sun.proxy.$Proxy66.getNewApplication(Unknown Source)
>     >
>     >         at org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.
>     > getNewApplication(YarnClientImpl.java:219)
>     >
>     >         at org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.
>     > createApplication(YarnClientImpl.java:227)
>     >
>     >         at org.apache.hadoop.mapred.ResourceMgrDelegate.getNewJobID(
>     > ResourceMgrDelegate.java:187)
>     >
>     >         at org.apache.hadoop.mapred.YARNRunner.getNewJobID(
>     > YARNRunner.java:231)
>     >
>     >         at org.apache.hadoop.mapreduce.JobSubmitter.
> submitJobInternal(
>     > JobSubmitter.java:153)
>     >
>     >         at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290)
>     >
>     >         at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287)
>     >
>     >         at java.security.AccessController.doPrivileged(Native
> Method)
>     >
>     >         at javax.security.auth.Subject.doAs(Subject.java:422)
>     >
>     >         at org.apache.hadoop.security.UserGroupInformation.doAs(
>     > UserGroupInformation.java:1698)
>     >
>     >         at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287)
>     >
>     >         at org.apache.kylin.engine.mr.common.AbstractHadoopJob.
>     > waitForCompletion(AbstractHadoopJob.java:175)
>     >
>     >         at org.apache.kylin.engine.mr.steps.FactDistinctColumnsJob.
>     > run(FactDistinctColumnsJob.java:110)
>     >
>     >         at org.apache.kylin.engine.mr.common.MapReduceExecutable.
>     > doWork(MapReduceExecutable.java:130)
>     >
>     >         at org.apache.kylin.job.execution.AbstractExecutable.
>     > execute(AbstractExecutable.java:162)
>     >
>     >         at org.apache.kylin.job.execution.DefaultChainedExecutable.
> doWork(
>     > DefaultChainedExecutable.java:67)
>     >
>     >         at org.apache.kylin.job.execution.AbstractExecutable.
>     > execute(AbstractExecutable.java:162)
>     >
>     >         at org.apache.kylin.job.impl.threadpool.DefaultScheduler$
>     > JobRunner.run(DefaultScheduler.java:300)
>     >
>     >         at java.util.concurrent.ThreadPoolExecutor.runWorker(
>     > ThreadPoolExecutor.java:1142)
>     >
>     >         at java.util.concurrent.ThreadPoolExecutor$Worker.run(
>     > ThreadPoolExecutor.java:617)
>     >
>     >         at java.lang.Thread.run(Thread.java:748)
>     >
>     > Caused by: java.net.ConnectException: Connection refused
>     >
>     >         at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
>     >
>     >         at sun.nio.ch.SocketChannelImpl.finishConnect(
>     > SocketChannelImpl.java:717)
>     >
>     >         at org.apache.hadoop.net.SocketIOWithTimeout.connect(
>     > SocketIOWithTimeout.java:206)
>     >
>     >         at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:531)
>     >
>     >         at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:495)
>     >
>     >         at org.apache.hadoop.ipc.Client$Connection.setupConnection(
>     > Client.java:614)
>     >
>     >         at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(
>     > Client.java:712)
>     >
>     >         at org.apache.hadoop.ipc.Client$
> Connection.access$2900(Client.
>     > java:375)
>     >
>     >         at org.apache.hadoop.ipc.Client.getConnection(Client.java:
> 1528)
>     >
>     >         at org.apache.hadoop.ipc.Client.call(Client.java:1451)
>     >
>     >         ... 31 more
>     >
>     > result code:2
>     >
>     > 在yarn-site.xml 文件中相应的ip和端口配置如下,若把18030、18031、18032、
>     > 19033这几个端口的ip改为localhost,这个问题就可以解决,但Hadoop集群无法正常加载,在htt
>     > p://10.9.0.86:8088/cluster/nodes中只有一个主节点,其它集群节点都无法加载。
>     >
>     >
>     > <property>
>     > <name>yarn.resouremanager.hostname</name>
>     > <value>10.9.0.86</value>
>     > </property>
>     > <property>
>     > <name>yarn.resourcemanager.scheduler.address</name>
>     > <value>10.9.0.86:18030</value>
>     > </property>
>     > <property>
>     > <name>yarn.resourcemanager.resource-tracker.address</name>
>     > <value>10.9.0.86:18031</value>
>     > </property>
>     > <property>
>     > <name>yarn.resourcemanager.address</name>
>     > <value>10.9.0.86:18032</value>
>     > </property>
>     > <property>
>     > <name>yarn.resourcemanager.admin.address</name>
>     > <value>10.9.0.86:18033</value>
>     > </property>
>     > <property>
>     > <name>yarn.resourcemanager.webapp.address</name>
>     > <value>10.9.0.86:8088</value>
>     > </property>
>     >
>     > <property>
>     > <name>yarn.resourcemanager.webapp.https.address</name>
>     > <value>10.9.0.86:8090</value>
>     > </property>
>     >
>     >
>     >
>     >
>     >
>     >
>     > ________________________________
>     > Jason Lu
>     > Email:[email protected]
>     >
>
>
>

Reply via email to