Active cube run set to pending and a new cube request is run without finishing

2021-04-22 Thread xatax
I have obeserved that when the "kylin.job.max-concurrent-jobs" properties is
used for a cluster configuration and set to for e.g. lets say value 1. Then
for a given cube building job cluster, during a cube build if another cube
build request is submitted the currently running job is set to pending and
Kylin starts running the new job on the same cluster. Then again it might
pause the running job and any step and then return to the execution of the
first job. This is usually observed after the first step but can be random.

Is this expected behavior? 

The good thing is ultimately both jobs run to success.

Please share any knowledge on this.

thanks!

--
Sent from: http://apache-kylin.74782.x6.nabble.com/


Can cube be queried while it is refreshing

2021-03-24 Thread xatax
Can a cube be queried while it is refreshing?

If yes, how does it handle the switch over to the refreshed data? Can that
involve downtime?

Thanks!

--
Sent from: http://apache-kylin.74782.x6.nabble.com/


High cardinality column search very slow

2021-03-24 Thread xatax
I have a column with a count distinct of about 12million. This is an integer
field (for e.g. call it account with just numbers in it).
This is account drives a lot of reports where the user enters the account in
a search box on tableau and then tableau sends a query like: 

SELECT account  ,
  SUM(earnings)  
FROM sales 
WHERE (POSITION('538807' IN LOWER(account)) > 0)
GROUP BY 1
;

This search is slow. It takes about 15 to 30 seconds or more. 

The same when implemented on a column with defined as varchar is rather
fast.

Can you please help understand what is going on?

thanks in advance

--
Sent from: http://apache-kylin.74782.x6.nabble.com/


Re: Re:Cube migration across hadoop clusters: What is -updateMappingPath updateTableMapping.json

2021-02-27 Thread xatax
Hi Zhichao,
thank you for your reply. 
Thanks for providing the file format as it is not available anywhere.
I am not sure on what hadoop jars are missing. I am doing the cube migration
on Kylin 3.1.1 on AWS EMR 5.29. 
And I am following the Install Kylin on AWS EMR instructions. 

Here is a new log: Can you please check and may be tell what might be
missing:

Thanks in advance.


[hadoop@ip-10-0-1-141 apache-kylin-3.1.1-bin-hbase1x]$ ./bin/kylin.sh
org.apache.kylin.tool.migration.CubeMigrationCrossClusterCLI -kylinUriSrc
ADMIN:KYLIN@ip-10-0-1-141.ec2.internal:7070 -kylinUriDst
ADMIN:KYLIN@ip-10-0-1-231.ec2.internal:7070 -cube
cube_SAR_EXEC_Cmpny_SAR_Account_3_DNB_Cmpny_Search -execute true -overwrite
true -schemaOnly false
Using cached dependency...
skip flink_dependency
OpenJDK 64-Bit Server VM warning: ignoring option MaxPermSize=512M; support
was removed in 8.0
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in
[jar:file:/usr/local/kylin/apache-kylin-3.1.1-bin-hbase1x/tool/kylin-tool-3.1.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in
[jar:file:/usr/lib/hive/lib/log4j-slf4j-impl-2.6.2.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in
[jar:file:/usr/lib/spark/jars/slf4j-log4j12-1.7.16.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in
[jar:file:/usr/lib/hadoop/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
Running org.apache.kylin.tool.migration.CubeMigrationCrossClusterCLI
-kylinUriSrc ADMIN:KYLIN@ip-10-0-1-141.ec2.internal:7070 -kylinUriDst
ADMIN:KYLIN@ip-10-0-1-231.ec2.internal:7070 -cube
cube_SAR_EXEC_Cmpny_SAR_Account_3_DNB_Cmpny_Search -execute true -overwrite
true -schemaOnly false
2021-02-27 11:29:33,204 INFO  [main] common.KylinConfig:118 : Loading
kylin-defaults.properties from
file:/usr/local/kylin/apache-kylin-3.1.1-bin-hbase1x/tool/kylin-tool-3.1.1.jar!/kylin-defaults.properties
2021-02-27 11:29:33,222 INFO  [main] common.KylinConfig:352 : Use
KYLIN_HOME=/usr/local/kylin/apache-kylin-3.1.1-bin-hbase1x
2021-02-27 11:29:33,224 INFO  [main] common.KylinConfig:153 : Initialized a
new KylinConfig from getInstanceFromEnv : 245765246
2021-02-27 11:29:35,139 INFO  [main] zookeeper.RecoverableZooKeeper:120 :
Process identifier=hconnection-0x2532b351 connecting to ZooKeeper
ensemble=ip-10-0-1-141.ec2.internal:2181
2021-02-27 11:29:35,144 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:zookeeper.version=3.4.14-5f788d5e8f90539ee331702c753fa250727128f4,
built on 12/14/2019 00:35 GMT
2021-02-27 11:29:35,144 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:host.name=ip-10-0-1-141.ec2.internal
2021-02-27 11:29:35,144 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:java.version=1.8.0_272
2021-02-27 11:29:35,144 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:java.vendor=Red Hat, Inc.
2021-02-27 11:29:35,144 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:java.home=/usr/lib/jvm/java-1.8.0-openjdk-1.8.0.272.b10-1.56.amzn1.x86_64/jre
2021-02-27 11:29:35,145 INFO  [main] zookeeper.ZooKeeper:100 : Client

Cube migration across hadoop clusters: What is -updateMappingPath updateTableMapping.json

2021-02-26 Thread xatax
Per documentation: http://kylin.apache.org/docs/tutorial/cube_migration.html

In the example for "II. Migrate across two Hadoop clusters"

./bin/kylin.sh org.apache.kylin.tool.migration.CubeMigrationCrossClusterCLI
-kylinUriSrc ADMIN:ky...@qa.env:17070 -kylinUriDst
ADMIN:ky...@prod.env:1 -cube kylin_sales_cube -updateMappingPath
$KYLIN_HOME/updateTableMapping.json -execute true -schemaOnly false
-overwrite true

What is the updateTableMapping.json file? I have not been able to get any
information on what this file should be. Is this a file to be created by us?
can I have a sample? 

I am encountering error in trying to copy a cube with cube data to another
hadoop cluster. Here is the log:

Command without -updateMappingPath:
./bin/kylin.sh org.apache.kylin.tool.migration.CubeMigrationCrossClusterCLI
-kylinUriSrc ADMIN:KYLIN@ip-172-31-85-18.ec2.internal:7070 -kylinUriDst
ADMIN:KYLIN@ip-172-31-94-63.ec2.internal:7070 -cube kylin_sales_cube
-execute true -overwrite true -schemaOnly false

/*removed previous log*/
2021-02-26 16:48:30,274 INFO  [Cube-kylin_sales_cube-data-migration-pool-0]
migration.CubeMigrationCrossClusterCLI:695 : start to copy hdfs directory
from
hdfs://ip-172-31-85-18.ec2.internal:8020/kylin/kylin_metadata/kylin-8dbe2fbb-7960-c2a5-3a20-ba26958f9691
to
hdfs://ip-172-31-94-63.ec2.internal:8020/kylin/kylin_metadata/kylin-8dbe2fbb-7960-c2a5-3a20-ba26958f9691
2021-02-26 16:48:30,435 INFO  [Cube-kylin_sales_cube-data-migration-pool-1]
client.RMProxy:123 : Connecting to ResourceManager at
ip-172-31-85-18.ec2.internal/172.31.85.18:8032
2021-02-26 16:48:30,435 INFO  [Cube-kylin_sales_cube-data-migration-pool-0]
client.RMProxy:123 : Connecting to ResourceManager at
ip-172-31-85-18.ec2.internal/172.31.85.18:8032
2021-02-26 16:48:30,659 INFO  [Cube-kylin_sales_cube-data-migration-pool-0]
client.RMProxy:123 : Connecting to ResourceManager at
ip-172-31-85-18.ec2.internal/172.31.85.18:8032
2021-02-26 16:48:30,659 INFO  [Cube-kylin_sales_cube-data-migration-pool-1]
client.RMProxy:123 : Connecting to ResourceManager at
ip-172-31-85-18.ec2.internal/172.31.85.18:8032
2021-02-26 16:48:30,877 INFO  [Cube-kylin_sales_cube-data-migration-pool-0]
tools.SimpleCopyListing:589 : Paths (files+dirs) cnt = 71; dirCnt = 15
2021-02-26 16:48:30,877 INFO  [Cube-kylin_sales_cube-data-migration-pool-0]
tools.SimpleCopyListing:384 : Build file listing completed.
2021-02-26 16:48:30,879 INFO  [Cube-kylin_sales_cube-data-migration-pool-0]
Configuration.deprecation:1285 : io.sort.mb is deprecated. Instead, use
mapreduce.task.io.sort.mb
2021-02-26 16:48:30,879 INFO  [Cube-kylin_sales_cube-data-migration-pool-0]
Configuration.deprecation:1285 : io.sort.factor is deprecated. Instead, use
mapreduce.task.io.sort.factor
2021-02-26 16:48:31,370 INFO  [Cube-kylin_sales_cube-data-migration-pool-0]
tools.DistCp:94 : Number of paths in the copy list: 71
2021-02-26 16:48:31,400 INFO  [Cube-kylin_sales_cube-data-migration-pool-0]
tools.DistCp:94 : Number of paths in the copy list: 71
2021-02-26 16:48:31,448 INFO  [Cube-kylin_sales_cube-data-migration-pool-0]
client.RMProxy:123 : Connecting to ResourceManager at
ip-172-31-85-18.ec2.internal/172.31.85.18:8032
2021-02-26 16:48:31,486 WARN  [Cube-kylin_sales_cube-data-migration-pool-0]
mapreduce.JobResourceUploader:64 : Hadoop command-line option parsing not
performed. Implement the Tool interface and execute your application with
ToolRunner to remedy this.
2021-02-26 16:48:32,227 INFO  [Cube-kylin_sales_cube-data-migration-pool-1]
tools.SimpleCopyListing:589 : Paths (files+dirs) cnt = 11; dirCnt = 5
2021-02-26 16:48:32,227 INFO  [Cube-kylin_sales_cube-data-migration-pool-1]
tools.SimpleCopyListing:384 : Build file listing completed.
2021-02-26 16:48:32,243 INFO  [Cube-kylin_sales_cube-data-migration-pool-1]
tools.DistCp:94 : Number of paths in the copy list: 11
2021-02-26 16:48:32,259 INFO  [Cube-kylin_sales_cube-data-migration-pool-1]
tools.DistCp:94 : Number of paths in the copy list: 11
2021-02-26 16:48:32,306 INFO  [Cube-kylin_sales_cube-data-migration-pool-1]
client.RMProxy:123 : Connecting to ResourceManager at
ip-172-31-85-18.ec2.internal/172.31.85.18:8032
2021-02-26 16:48:32,311 WARN  [Cube-kylin_sales_cube-data-migration-pool-1]
mapreduce.JobResourceUploader:64 : Hadoop command-line option parsing not
performed. Implement the Tool interface and execute your application with
ToolRunner to remedy this.
2021-02-26 16:48:32,367 INFO  [Cube-kylin_sales_cube-data-migration-pool-1]
mapreduce.JobSubmitter:200 : number of splits:6
2021-02-26 16:48:32,398 INFO  [Cube-kylin_sales_cube-data-migration-pool-0]
mapreduce.JobSubmitter:200 : number of splits:26
2021-02-26 16:48:32,490 INFO  [Cube-kylin_sales_cube-data-migration-pool-0]
mapreduce.JobSubmitter:289 : Submitting tokens for job:
job_1614247321131_0022
2021-02-26 16:48:32,491 INFO  [Cube-kylin_sales_cube-data-migration-pool-1]
mapreduce.JobSubmitter:289 : Submitting tokens for job:
job_1614247321131_0023
2021-02-26 16:48:32,604 INFO  

Re: Cube migration across hadoop clusters

2021-02-26 Thread xatax
Thanks Zhichao,

I tried that and am now getting this:

[hadoop@ip-172-31-81-255 apache-kylin-3.1.0-bin-hbase1x]$
[hadoop@ip-172-31-81-255 apache-kylin-3.1.0-bin-hbase1x]$ ./bin/kylin.sh
org.apache.kylin.tool.migration.CubeMigrationCrossClusterCLI -kylinUriSrc
ADMIN:KYLIN@ip-172-31-81-255.ec2.internal:7070 -kylinUriDst
ADMIN:KYLIN@ip-172-31-91-198.ec2.internal:7070 -cube kylin_sales_cube
-execute true -overwrite true -schemaOnly false
Using cached dependency...
OpenJDK 64-Bit Server VM warning: ignoring option MaxPermSize=512M; support
was removed in 8.0
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in
[jar:file:/usr/local/kylin/apache-kylin-3.1.0-bin-hbase1x/tool/kylin-tool-3.1.0.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in
[jar:file:/usr/lib/hive/lib/log4j-slf4j-impl-2.6.2.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in
[jar:file:/usr/lib/spark/jars/slf4j-log4j12-1.7.16.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in
[jar:file:/usr/lib/hadoop/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
Running org.apache.kylin.tool.migration.CubeMigrationCrossClusterCLI
-kylinUriSrc ADMIN:KYLIN@ip-172-31-81-255.ec2.internal:7070 -kylinUriDst
ADMIN:KYLIN@ip-172-31-91-198.ec2.internal:7070 -cube kylin_sales_cube
-execute true -overwrite true -schemaOnly false
2021-02-26 10:06:54,000 INFO  [main] common.KylinConfig:118 : Loading
kylin-defaults.properties from
file:/usr/local/kylin/apache-kylin-3.1.0-bin-hbase1x/tool/kylin-tool-3.1.0.jar!/kylin-defaults.properties
2021-02-26 10:06:54,028 DEBUG [main] common.KylinConfig:346 : KYLIN_CONF
property was not set, will seek KYLIN_HOME env variable
2021-02-26 10:06:54,029 INFO  [main] common.KylinConfig:352 : Use
KYLIN_HOME=/usr/local/kylin/apache-kylin-3.1.0-bin-hbase1x
2021-02-26 10:06:54,031 INFO  [main] common.KylinConfig:153 : Initialized a
new KylinConfig from getInstanceFromEnv : 245765246
2021-02-26 10:06:56,326 INFO  [main] zookeeper.RecoverableZooKeeper:120 :
Process identifier=hconnection-0x545607f2 connecting to ZooKeeper
ensemble=ip-172-31-81-255.ec2.internal:2181
2021-02-26 10:06:56,331 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:zookeeper.version=3.4.14-5f788d5e8f90539ee331702c753fa250727128f4,
built on 12/14/2019 00:35 GMT
2021-02-26 10:06:56,331 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:host.name=ip-172-31-81-255.ec2.internal
2021-02-26 10:06:56,332 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:java.version=1.8.0_272
2021-02-26 10:06:56,332 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:java.vendor=Red Hat, Inc.
2021-02-26 10:06:56,332 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:java.home=/usr/lib/jvm/java-1.8.0-openjdk-1.8.0.272.b10-1.56.amzn1.x86_64/jre
2021-02-26 10:06:56,332 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:java.class.path= /*removed*/ 
2021-02-26 10:06:56,333 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:java.library.path=:/usr/lib/hadoop-lzo/lib/native:/usr/lib/hadoop/lib/native
2021-02-26 10:06:56,333 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:java.io.tmpdir=/tmp
2021-02-26 10:06:56,334 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:java.compiler=
2021-02-26 10:06:56,334 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:os.name=Linux
2021-02-26 10:06:56,334 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:os.arch=amd64
2021-02-26 10:06:56,335 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:os.version=4.14.154-99.181.amzn1.x86_64
2021-02-26 10:06:56,335 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:user.name=hadoop
2021-02-26 10:06:56,335 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:user.home=/home/hadoop
2021-02-26 10:06:56,335 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:user.dir=/usr/local/kylin/apache-kylin-3.1.0-bin-hbase1x
2021-02-26 10:06:56,336 INFO  [main] zookeeper.ZooKeeper:442 : Initiating
client connection, connectString=ip-172-31-81-255.ec2.internal:2181
sessionTimeout=9
watcher=org.apache.hadoop.hbase.zookeeper.PendingWatcher@7207cb51
2021-02-26 10:06:56,357 INFO 
[main-SendThread(ip-172-31-81-255.ec2.internal:2181)]
zookeeper.ClientCnxn:1025 : Opening socket connection to server
ip-172-31-81-255.ec2.internal/172.31.81.255:2181. Will not attempt to
authenticate using SASL (unknown error)
2021-02-26 10:06:56,363 INFO 
[main-SendThread(ip-172-31-81-255.ec2.internal:2181)]
zookeeper.ClientCnxn:879 : Socket connection established to
ip-172-31-81-255.ec2.internal/172.31.81.255:2181, initiating session
2021-02-26 10:06:56,369 INFO 
[main-SendThread(ip-172-31-81-255.ec2.internal:2181)]
zookeeper.ClientCnxn:1299 : Session establishment complete on server

Re: Cube migration across hadoop clusters

2021-02-25 Thread xatax
hi Yaqian,
thanks for the response. I am using this method from the documentation: "II.
Migrate across two Hadoop clusters" using Kylin version 3.1.0 on EMR 5.29 on
both source and destination clusters.

I tried in 2 ways:
(1) Migrate from one kylin instance on emr 5.29 to another kylin instance on
emr 5.29 with the Hbase S3 location same on both.
(2) Migrate from one kylin instance on emr 5.29 to another kylin instance on
emr 5.29 with Hbase s3 location different.
In both ways the cube metadata/definition was migrated but not the cube
data.

Here is the log for (1): 

[hadoop@ip-172-31-90-136 apache-kylin-3.1.0-bin-hbase1x]$ ./bin/kylin.sh
org.apache.kylin.tool.migration.CubeMigrationCrossClusterCLI -kylinUriSrc
ADMIN:KYLIN@ip-172-31-90-136.ec2.internal:7070 -kylinUriDst
ADMIN:KYLIN@ip-172-31-81-180.ec2.internal:7070 -cube kylin_sales_cube
-execute true -overwrite true

Using cached dependency...
OpenJDK 64-Bit Server VM warning: ignoring option MaxPermSize=512M; support
was removed in 8.0
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in
[jar:file:/usr/local/kylin/apache-kylin-3.1.0-bin-hbase1x/tool/kylin-tool-3.1.0.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in
[jar:file:/usr/lib/hive/lib/log4j-slf4j-impl-2.6.2.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in
[jar:file:/usr/lib/spark/jars/slf4j-log4j12-1.7.16.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in
[jar:file:/usr/lib/hadoop/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
Running org.apache.kylin.tool.migration.CubeMigrationCrossClusterCLI
-kylinUriSrc ADMIN:KYLIN@ip-172-31-90-136.ec2.internal:7070 -kylinUriDst
ADMIN:KYLIN@ip-172-31-81-180.ec2.internal:7070 -cube kylin_sales_cube
-execute true -overwrite true
2021-02-25 09:36:50,551 INFO  [main] common.KylinConfig:118 : Loading
kylin-defaults.properties from
file:/usr/local/kylin/apache-kylin-3.1.0-bin-hbase1x/tool/kylin-tool-3.1.0.jar!/kylin-defaults.properties
2021-02-25 09:36:50,576 DEBUG [main] common.KylinConfig:346 : KYLIN_CONF
property was not set, will seek KYLIN_HOME env variable
2021-02-25 09:36:50,579 INFO  [main] common.KylinConfig:352 : Use
KYLIN_HOME=/usr/local/kylin/apache-kylin-3.1.0-bin-hbase1x
2021-02-25 09:36:50,582 INFO  [main] common.KylinConfig:153 : Initialized a
new KylinConfig from getInstanceFromEnv : 245765246
2021-02-25 09:36:55,215 INFO  [main] zookeeper.RecoverableZooKeeper:120 :
Process identifier=hconnection-0x5d3b58ca connecting to ZooKeeper
ensemble=ip-172-31-90-136.ec2.internal:2181
2021-02-25 09:36:55,224 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:zookeeper.version=3.4.14-5f788d5e8f90539ee331702c753fa250727128f4,
built on 12/14/2019 00:35 GMT
2021-02-25 09:36:55,224 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:host.name=ip-172-31-90-136.ec2.internal
2021-02-25 09:36:55,225 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:java.version=1.8.0_272
2021-02-25 09:36:55,225 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:java.vendor=Red Hat, Inc.
2021-02-25 09:36:55,225 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:java.home=/usr/lib/jvm/java-1.8.0-openjdk-1.8.0.272.b10-1.56.amzn1.x86_64/jre
2021-02-25 09:36:55,225 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:java.class.path=/*removed*/
2021-02-25 09:36:55,226 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:java.library.path=:/usr/lib/hadoop-lzo/lib/native:/usr/lib/hadoop/lib/native
2021-02-25 09:36:55,227 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:java.io.tmpdir=/tmp
2021-02-25 09:36:55,227 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:java.compiler=
2021-02-25 09:36:55,227 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:os.name=Linux
2021-02-25 09:36:55,228 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:os.arch=amd64
2021-02-25 09:36:55,229 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:os.version=4.14.154-99.181.amzn1.x86_64
2021-02-25 09:36:55,229 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:user.name=hadoop
2021-02-25 09:36:55,229 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:user.home=/home/hadoop
2021-02-25 09:36:55,229 INFO  [main] zookeeper.ZooKeeper:100 : Client
environment:user.dir=/usr/local/kylin/apache-kylin-3.1.0-bin-hbase1x
2021-02-25 09:36:55,230 INFO  [main] zookeeper.ZooKeeper:442 : Initiating
client connection, connectString=ip-172-31-90-136.ec2.internal:2181
sessionTimeout=9
watcher=org.apache.hadoop.hbase.zookeeper.PendingWatcher@7e91ed74
2021-02-25 09:36:55,289 INFO 
[main-SendThread(ip-172-31-90-136.ec2.internal:2181)]
zookeeper.ClientCnxn:1025 : Opening socket connection to server
ip-172-31-90-136.ec2.internal/172.31.90.136:2181. Will not attempt to
authenticate 

Cube migration across hadoop clusters

2021-02-24 Thread xatax
Per: http://kylin.apache.org/docs/tutorial/cube_migration.html

It says: "Support to migrate cube data (segments data on HBase) from QA env
to PROD env."

I tried the migration but it only migrates the cube metadata without the
data. The source kylin instance has the cube buiilt and in ready state.
After I executed the command it said it was successful. But when I checked
the destination kylin instance it only got the cube definition and not the
data.

Can you please clarify if migrating the data is possible or not?

Or is the migration of the data a separate activity which involves the hbase
snapshot import export.

thanks in advance

--
Sent from: http://apache-kylin.74782.x6.nabble.com/


Kylin Cluster mode. SQL works in standalone mode. Errors in Cluster mode. No dictionary found.

2021-02-21 Thread xatax
I installed kylin on EMR in standalone mode. Ran a bunch of SQLs and they run
fine.

Then I add a query cluster. So I have kylin.server.mode in one cluster set
to "all" and one cluster in "query"  mode.

I set kylin.server.cluster-servers=jobnode1:7070,querynode1:7070 
on both query and job cluster.
Restart kylin on both

Then I get this: 
No dictionary found by
/dict/DEFAULT.KYLIN_ACCOUNT/ACCOUNT_COUNTRY/1196ba3d-22e6-0337-cfe8-aaceb02d81dd.dict,
invalid cube state; cube segmentsales_cube1[220100_2012040100],
col BUYER_ACCOUNT:DEFAULT.KYLIN_ACCOUNT.ACCOUNT_COUNTRY while executing SQL: 
"SELECT ACCOUNT_COUNTRY, SUM(PRICE) , COUNT(*) FROM KYLIN_SALES JOIN
KYLIN_ACCOUNT ON KYLIN_SALES.BUYER_ID = KYLIN_ACCOUNT.ACCOUNT_ID GROUP BY
ACCOUNT_COUNTRY LIMIT 200" 

is there some other setting that needs to be done which is not mentioned on
http://kylin.apache.org/docs/install/kylin_cluster.html 

--
Sent from: http://apache-kylin.74782.x6.nabble.com/


Kylin Cluster Mode with EMR. Is a restart of kylin required on all job and query servers when new node is added.

2021-02-21 Thread xatax
Kylin Cluster Mode. I am using Kylin on EMR.
Is a restart of kylin required on all job and query servers when new node is
added. Then there will be downtime. How do we achieve High Availibility.

from the guide:
"
Configure the Kylin node list kylin.server.cluster-servers, including all
nodes (the current node is also included). When the event changes, the node
receiving the change needs to notify all other nodes (the current node is
also included).
"

So if I have 2 job servers and 2 query servers:
then
kylin.server.cluster-servers=jobnode1:7070,jobnode2:7070,querynode1:7070,querynode2:7070

If let's say query node 2 on EMR fails. And I add a new node with new EMR
which will have a new IP. Then for the new node:

kylin.server.cluster-servers=jobnode1:7070,jobnode2:7070,querynode1:7070,querynode3:7070

And then looks like I have to change the string on all the nodes to:
kylin.server.cluster-servers=jobnode1:7070,jobnode2:7070,querynode1:7070,querynode3:7070

And then restart kylin on all nodes.

This will incur downtime. Can you please help me understand how High
Availability can be achieved then.

Thanks in advance!





--
Sent from: http://apache-kylin.74782.x6.nabble.com/


Re: Cluster mode - High availability. Metadata storage options other than MySQL?

2021-02-21 Thread xatax
thanks!
What about something like dynamo db?

--
Sent from: http://apache-kylin.74782.x6.nabble.com/


Cluster mode - High availability. Metadata storage options other than MySQL?

2021-02-14 Thread xatax
What are the recommended High availability options.
>From the apache kylin website it recommends the cluster mode installation
and then there is an option to use MySql for metadata storage. 
MySQL being an RDBMS is not best suited for an HA architecture. 
Any other options that work with Apache Kylin?

Thanks!

--
Sent from: http://apache-kylin.74782.x6.nabble.com/


Cluster mode kylin.metadata.url and kylin.server.cluster-servers examples for AWS EMR

2021-02-03 Thread xatax
Hi 
I am trying to install Kylin in cluster mode on AWS EMRs using the
documentation: http://kylin.apache.org/docs/install/kylin_cluster.html

Can you please help me with example values of kylin.metadata.url and
kylin.server.cluster-servers so I know the format.

thanks!

--
Sent from: http://apache-kylin.74782.x6.nabble.com/


Re: error while using left join

2021-01-14 Thread xatax
We are facing the same issue in tableau when we give a filter on the left
join. It is converting to inner join. 

Can you please let us know if there is any solution for this?

Thanks!

--
Sent from: http://apache-kylin.74782.x6.nabble.com/


jdbc connection not working.

2020-07-12 Thread xatax
I have attempted to connect to Kylin using jdbc driver for the following
Kylin versions 2.6, 3.0.2 and 3.1 and have been unable to make connection. 

The driver files I have used are from the '$KYLIN_HOME/lib' directory:
for version 3.1: kylin-jdbc-3.1.0.jar, jcl-over-slf4j-1.7.21.jar,
slf4j-api-1.7.21.jar
for version 2.6.6: kylin-jdbc-2.6.6.jar
for version 3.0.2: kylin-jdbc-3.0.2.jar

Connection URL I am using:
jdbc:kylin://"EMR cluster hostname":7070/learn_kylin
JDBC Driver class: org.apache.kylin.jdbc.Driver

Will appreciate any insight into what might be going wrong. 

Error logs:

38f9d37464b4% javac ky.java
38f9d37464b4% java ky
Connecting to database...
SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder".
SLF4J: Defaulting to no-operation (NOP) logger implementation
SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further
details.
java.sql.SQLException:
org.apache.kylin.jdbc.shaded.org.apache.http.conn.HttpHostConnectException:
Connection to http://ec2-3-236-65-136.compute-1.amazonaws.com:7070 refused
at org.apache.kylin.jdbc.KylinConnection.(KylinConnection.java:72)
at
org.apache.kylin.jdbc.KylinJdbcFactory.newConnection(KylinJdbcFactory.java:77)
at
org.apache.kylin.jdbc.shaded.org.apache.calcite.avatica.UnregisteredDriver.connect(UnregisteredDriver.java:138)
at java.sql.DriverManager.getConnection(DriverManager.java:664)
at java.sql.DriverManager.getConnection(DriverManager.java:247)
at ky.main(ky.java:22)
Caused by:
org.apache.kylin.jdbc.shaded.org.apache.http.conn.HttpHostConnectException:
Connection to http://ec2-3-236-65-136.compute-1.amazonaws.com:7070 refused
at
org.apache.kylin.jdbc.shaded.org.apache.http.impl.conn.DefaultClientConnectionOperator.openConnection(DefaultClientConnectionOperator.java:190)
at
org.apache.kylin.jdbc.shaded.org.apache.http.impl.conn.ManagedClientConnectionImpl.open(ManagedClientConnectionImpl.java:294)
at
org.apache.kylin.jdbc.shaded.org.apache.http.impl.client.DefaultRequestDirector.tryConnect(DefaultRequestDirector.java:643)
at
org.apache.kylin.jdbc.shaded.org.apache.http.impl.client.DefaultRequestDirector.execute(DefaultRequestDirector.java:479)
at
org.apache.kylin.jdbc.shaded.org.apache.http.impl.client.AbstractHttpClient.execute(AbstractHttpClient.java:906)
at
org.apache.kylin.jdbc.shaded.org.apache.http.impl.client.AbstractHttpClient.execute(AbstractHttpClient.java:805)
at
org.apache.kylin.jdbc.shaded.org.apache.http.impl.client.AbstractHttpClient.execute(AbstractHttpClient.java:784)
at org.apache.kylin.jdbc.KylinClient.connect(KylinClient.java:285)
at org.apache.kylin.jdbc.KylinConnection.(KylinConnection.java:70)
... 5 more
Caused by: java.net.ConnectException: Operation timed out (Connection timed
out)
at java.net.PlainSocketImpl.socketConnect(Native Method)
at
java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:350)
at
java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:206)
at
java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:188)
at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392)
at java.net.Socket.connect(Socket.java:606)
at
org.apache.kylin.jdbc.shaded.org.apache.http.conn.scheme.PlainSocketFactory.connectSocket(PlainSocketFactory.java:127)
at
org.apache.kylin.jdbc.shaded.org.apache.http.impl.conn.DefaultClientConnectionOperator.openConnection(DefaultClientConnectionOperator.java:180)
... 13 more
Goodbye!
38f9d37464b4% javac ky.java
38f9d37464b4% java ky
Connecting to database...
SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder".
SLF4J: Defaulting to no-operation (NOP) logger implementation
SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further
details.
java.sql.SQLException: java.net.ConnectException: Operation timed out
(Connection timed out)
at org.apache.kylin.jdbc.KylinConnection.(KylinConnection.java:72)
at
org.apache.kylin.jdbc.KylinJdbcFactory.newConnection(KylinJdbcFactory.java:77)
at
org.apache.kylin.jdbc.shaded.org.apache.calcite.avatica.UnregisteredDriver.connect(UnregisteredDriver.java:138)
at java.sql.DriverManager.getConnection(DriverManager.java:664)
at java.sql.DriverManager.getConnection(DriverManager.java:247)
at ky.main(ky.java:22)
Caused by: java.net.ConnectException: Operation timed out (Connection timed
out)
at java.net.PlainSocketImpl.socketConnect(Native Method)
at
java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:350)
at
java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:206)
at
java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:188)
at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392)
at 

Build Dimension Dictionary: org.apache.kylin.engine.mr.exception.HadoopShellException: java.lang.ClassCastException: org.apache.hadoop.io.IntWritable cannot be cast to org.apache.hadoop.io.LongWritabl

2020-06-14 Thread xatax
Hi,

I am getting this error during cube build, can anyone please help?
Thanks!!!: 

#4 Step Name: Build Dimension Dictionary
Duration: 0.04 mins Waiting: 0 seconds

org.apache.kylin.engine.mr.exception.HadoopShellException:
java.lang.ClassCastException: org.apache.hadoop.io.IntWritable cannot be
cast to org.apache.hadoop.io.LongWritable
at
org.apache.hadoop.hive.serde2.objectinspector.primitive.WritableLongObjectInspector.getPrimitiveJavaObject(WritableLongObjectInspector.java:46)
at
org.apache.hive.hcatalog.data.HCatRecordSerDe.serializePrimitiveField(HCatRecordSerDe.java:276)
at
org.apache.hive.hcatalog.data.HCatRecordSerDe.serializeField(HCatRecordSerDe.java:197)
at 
org.apache.hive.hcatalog.data.LazyHCatRecord.get(LazyHCatRecord.java:53)
at 
org.apache.hive.hcatalog.data.LazyHCatRecord.get(LazyHCatRecord.java:97)
at
org.apache.hive.hcatalog.mapreduce.HCatRecordReader.nextKeyValue(HCatRecordReader.java:204)
at
org.apache.hive.hcatalog.data.transfer.impl.HCatInputFormatReader$HCatRecordItr.hasNext(HCatInputFormatReader.java:107)
at
org.apache.kylin.source.hive.HiveTableReader.next(HiveTableReader.java:90)
at
org.apache.kylin.dict.lookup.SnapshotTable.takeSnapshot(SnapshotTable.java:98)
at
org.apache.kylin.dict.lookup.SnapshotManager.buildSnapshot(SnapshotManager.java:133)
at
org.apache.kylin.cube.CubeManager$DictionaryAssist.buildSnapshotTable(CubeManager.java:1116)
at
org.apache.kylin.cube.CubeManager.buildSnapshotTable(CubeManager.java:1032)
at
org.apache.kylin.cube.cli.DictionaryGeneratorCLI.processSegment(DictionaryGeneratorCLI.java:90)
at
org.apache.kylin.cube.cli.DictionaryGeneratorCLI.processSegment(DictionaryGeneratorCLI.java:49)
at
org.apache.kylin.engine.mr.steps.CreateDictionaryJob.run(CreateDictionaryJob.java:71)
at org.apache.kylin.engine.mr.MRUtil.runMRJob(MRUtil.java:92)
at
org.apache.kylin.engine.mr.common.HadoopShellExecutable.doWork(HadoopShellExecutable.java:63)
at
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:163)
at
org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:69)
at
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:163)
at
org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:113)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)

result code:2
at
org.apache.kylin.engine.mr.common.HadoopShellExecutable.doWork(HadoopShellExecutable.java:73)
at
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:163)
at
org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:69)
at
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:163)
at
org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:113)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)

--
Sent from: http://apache-kylin.74782.x6.nabble.com/


Re: Cube build failure: org/apache/spark/api/java/function/Function

2020-06-06 Thread xatax
Hi Shaofeng,
Thank you very much for your reply.
Can I please know how I can delete the segments? Are the overlapping
segments causing the error? I get the overlapping segments error when for
example I submit two date ranges as follows:
2020-05-04 to 2020-05-06
and 
2020-05-05 to 2020-05-06.

And if I were to give new date range that I have not used in the past e.g
2020-05-08 to 2020-05-09 then I get "Oops!!!
org/apache/spark/api/java/function/Function"

Please let me know.
thanks

--
Sent from: http://apache-kylin.74782.x6.nabble.com/


Cube build failure: org/apache/spark/api/java/function/Function

2020-06-06 Thread xatax
I am following instructions to build cube with spark from:
http://kylin.apache.org/docs/tutorial/cube_spark.html 

I get the following error when building cube:
Cube build failure: org/apache/spark/api/java/function/Function

Can you please help?

Here is the log:
2020-06-06 14:58:38,289 ERROR [http-bio-7070-exec-2]
controller.CubeController:398 : Segments overlap:
kylin_sales_cube[2012010100_2015010100] and
kylin_sales_cube[2014050100_2020050200]
java.lang.IllegalStateException: Segments overlap:
kylin_sales_cube[2012010100_2015010100] and
kylin_sales_cube[2014050100_2020050200]
at org.apache.kylin.metadata.model.Segments.validate(Segments.java:433)
at
org.apache.kylin.cube.CubeManager.updateCubeWithRetry(CubeManager.java:385)
at org.apache.kylin.cube.CubeManager.updateCube(CubeManager.java:298)
at
org.apache.kylin.cube.CubeManager$SegmentAssist.appendSegment(CubeManager.java:739)
at org.apache.kylin.cube.CubeManager.appendSegment(CubeManager.java:659)
at org.apache.kylin.cube.CubeManager.appendSegment(CubeManager.java:651)
at
org.apache.kylin.rest.service.JobService.submitJobInternal(JobService.java:233)
at 
org.apache.kylin.rest.service.JobService.submitJob(JobService.java:202)
at
org.apache.kylin.rest.controller.CubeController.buildInternal(CubeController.java:395)
at
org.apache.kylin.rest.controller.CubeController.rebuild(CubeController.java:355)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at
org.springframework.web.method.support.InvocableHandlerMethod.doInvoke(InvocableHandlerMethod.java:205)
at
org.springframework.web.method.support.InvocableHandlerMethod.invokeForRequest(InvocableHandlerMethod.java:133)
at
org.springframework.web.servlet.mvc.method.annotation.ServletInvocableHandlerMethod.invokeAndHandle(ServletInvocableHandlerMethod.java:97)
at
org.springframework.web.servlet.mvc.method.annotation.RequestMappingHandlerAdapter.invokeHandlerMethod(RequestMappingHandlerAdapter.java:827)
at
org.springframework.web.servlet.mvc.method.annotation.RequestMappingHandlerAdapter.handleInternal(RequestMappingHandlerAdapter.java:738)
at
org.springframework.web.servlet.mvc.method.AbstractHandlerMethodAdapter.handle(AbstractHandlerMethodAdapter.java:85)
at
org.springframework.web.servlet.DispatcherServlet.doDispatch(DispatcherServlet.java:967)
at
org.springframework.web.servlet.DispatcherServlet.doService(DispatcherServlet.java:901)
at
org.springframework.web.servlet.FrameworkServlet.processRequest(FrameworkServlet.java:970)
at
org.springframework.web.servlet.FrameworkServlet.doPut(FrameworkServlet.java:883)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:653)
at
org.springframework.web.servlet.FrameworkServlet.service(FrameworkServlet.java:846)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:731)
at
org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:303)
at
org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:208)
at 
org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:52)
at
org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:241)
at
org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:208)
at
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:317)
at
org.springframework.security.web.access.intercept.FilterSecurityInterceptor.invoke(FilterSecurityInterceptor.java:127)
at
org.springframework.security.web.access.intercept.FilterSecurityInterceptor.doFilter(FilterSecurityInterceptor.java:91)
at
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
at
org.springframework.security.web.access.ExceptionTranslationFilter.doFilter(ExceptionTranslationFilter.java:114)
at
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
at
org.springframework.security.web.session.SessionManagementFilter.doFilter(SessionManagementFilter.java:137)
at
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
at
org.springframework.security.web.authentication.AnonymousAuthenticationFilter.doFilter(AnonymousAuthenticationFilter.java:111)
at
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
at

Re: No tables on KYLIN WEB UI and NoSuchObjectException(message:default.kylin_intermediate_* table not found) when building cube

2020-04-10 Thread xatax
Thank you very much for your response Yaqian.

So, regarding the question where tables are not showing, I tested this.

I created a hive external table on table data in S3 location (parquet file).
Then I created another table out of it using CTAS so that the new table is
part of the same hive database and has storage in hdfs; however even after
doing that the new tables are not showing in the DEFAULT database which is
the one I installed with what came with the sample.sh file.

So, how do I get table available in the Kylin UI for usage when I am adding
new tables in hive.

Thanks in advance.


--
Sent from: http://apache-kylin.74782.x6.nabble.com/


Re: subscribe kylin

2020-04-05 Thread xatax
I have subscribed but none of my posts are going through. How come your post
of subscribe was accepted whereas none of my posts are getting through.

I am not sure if this reply is going to make it.



--
Sent from: http://apache-kylin.74782.x6.nabble.com/


MetaException(message:Could not connect to meta store using any of the URIs provided. Most recent failure: org.apache.thrift.transport.TTransportException: java.net.ConnectException: Connection refuse

2020-03-14 Thread xatax
I am a beginner. I installed Apache Kylin on EMR per instructions on the
kylin website and am at my first step in the tutorials which is building the
sample kylin_sales cube.

When I try to build I keep getting this error and the second step
"Redistribute Flat Hive Table":

MetaException(message:Could not connect to meta store using any of the URIs
provided. Most recent failure:
org.apache.thrift.transport.TTransportException: java.net.ConnectException:
Connection refused (Connection refused)
at org.apache.thrift.transport.TSocket.open(TSocket.java:226)
at
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.open(HiveMetaStoreClient.java:480)
at
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:247)
at
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:129)
at
org.apache.kylin.source.hive.CLIHiveClient.getMetaStoreClient(CLIHiveClient.java:165)
at
org.apache.kylin.source.hive.CLIHiveClient.getHiveTableRows(CLIHiveClient.java:129)
at
org.apache.kylin.source.hive.RedistributeFlatHiveTableStep.computeRowCount(RedistributeFlatHiveTableStep.java:38)
at
org.apache.kylin.source.hive.RedistributeFlatHiveTableStep.doWork(RedistributeFlatHiveTableStep.java:81)
at
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:179)
at
org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:71)
at
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:179)
at
org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:114)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Caused by: java.net.ConnectException: Connection refused (Connection
refused)
at java.net.PlainSocketImpl.socketConnect(Native Method)
at
java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:350)
at
java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:206)
at
java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:188)
at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392)
at java.net.Socket.connect(Socket.java:607)
at org.apache.thrift.transport.TSocket.open(TSocket.java:221)
... 14 more
)
at
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.open(HiveMetaStoreClient.java:529)
at
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:247)
at
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:129)
at
org.apache.kylin.source.hive.CLIHiveClient.getMetaStoreClient(CLIHiveClient.java:165)
at
org.apache.kylin.source.hive.CLIHiveClient.getHiveTableRows(CLIHiveClient.java:129)
at
org.apache.kylin.source.hive.RedistributeFlatHiveTableStep.computeRowCount(RedistributeFlatHiveTableStep.java:38)
at
org.apache.kylin.source.hive.RedistributeFlatHiveTableStep.doWork(RedistributeFlatHiveTableStep.java:81)
at
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:179)
at
org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:71)
at
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:179)
at
org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:114)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)

--
Sent from: http://apache-kylin.74782.x6.nabble.com/