[
https://issues.apache.org/jira/browse/KYLIN-4625?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Xiaoxiang Yu resolved KYLIN-4625.
---------------------------------
Resolution: Fixed
> Debug the code of Kylin on Parquet without hadoop environment
> -------------------------------------------------------------
>
> Key: KYLIN-4625
> URL: https://issues.apache.org/jira/browse/KYLIN-4625
> Project: Kylin
> Issue Type: Improvement
> Components: Spark Engine
> Reporter: wangrupeng
> Assignee: wangrupeng
> Priority: Major
> Fix For: v4.0.0-alpha
>
> Attachments: image-2020-07-08-17-41-35-954.png,
> image-2020-07-08-17-42-09-603.png, screenshot-1.png
>
>
> Currently, Kylin on Parquet already supports debuging source code with local
> csv files and Not dependent on remote HDP sandbox, but it's a little bit
> complex. The steps are as follows:
> * edit the properties of
> $KYLIN_SOURCE_DIR/examples/test_case_data/sandbox/kylin.properties to local
> {code:java}
> kylin.metadata.url=$LOCAL_META_DIR
> kylin.env.zookeeper-is-local=true
> kylin.env.hdfs-working-dir=file:///path/to/local/dir
> kylin.engine.spark-conf.spark.master=local
> kylin.engine.spark-conf.spark.eventLog.dir=/path/to/local/dir
> kylin.env=UT{code}
> * debug org.apache.kylin.rest.DebugTomcat with IDEA && add VM option
> "-Dspark.local=true"
> !image-2020-07-08-17-41-35-954.png|width=574,height=363!
> * Load csv data source by pressing button "Data Source->Load CSV File as
> Table" on "Model" page, and set the schema for your table. Then press
> "submit" to save.
> !image-2020-07-08-17-42-09-603.png|width=577,height=259!
> Most time we debug just want to build and query cube quickly and focus the
> bug we want to resolve. But current way is complex to load csv tables, create
> model and cube and it's hard to use kylin sample cube. So, I want to add a
> csv source which using the model of kylin sample data directly when debug
> tomcat started.
--
This message was sent by Atlassian Jira
(v8.3.4#803005)