[ 
https://issues.apache.org/jira/browse/KYLIN-4625?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

wangrupeng updated KYLIN-4625:
------------------------------
    Description: 
Currently, Kylin on Parquet already supports debuging source code with local 
csv files, but it's a little bit complex. The steps are as follows:
 * edit the properties of 
$KYLIN_SOURCE_DIR/examples/test_case_data/sandbox/kylin.properties to local

{code:java}
kylin.metadata.url=$LOCAL_META_DIR
 kylin.env.zookeeper-is-local=true
 kylin.env.hdfs-working-dir=file:///path/to/local/dir
 kylin.engine.spark-conf.spark.master=local
 kylin.engine.spark-conf.spark.eventLog.dir=/path/to/local/dir
 kylin.env=UT{code}

 * debug org.apache.kylin.rest.DebugTomcat with IDEA && add VM option 
"-Dspark.local=true"
 !image-2020-07-08-17-41-35-954.png|width=574,height=363!
 * Load csv data source by pressing button "Data Source->Load CSV File as 
Table" on "Model" page, and set the schema for your table. Then press "submit" 
to save.
 !image-2020-07-08-17-42-09-603.png|width=577,height=259!

Most time we debug just want to build and query cube quickly and focus the bug 
we want to resolve. But current way is complex to load csv tables, create model 
and cube and it's hard to use kylin sample cube. So, I want to add a csv source 
which using the model of kylin sample data directly when debug tomcat started.

  was:
Currently, Kylin on Parquet already supports debuging source code with local 
csv files, but it's a little bit complex. The steps are as follows:
 * edit the properties of 
$KYLIN_SOURCE_DIR/examples/test_case_data/sandbox/kylin.properties to local
 ```log
 kylin.metadata.url=$LOCAL_META_DIR
 kylin.env.zookeeper-is-local=true
 kylin.env.hdfs-working-dir=[file:///path/to/local/dir]
 kylin.engine.spark-conf.spark.master=local
 kylin.engine.spark-conf.spark.eventLog.dir=/path/to/local/dir
 kylin.env=UT
 ```
 * debug org.apache.kylin.rest.DebugTomcat with IDEA && add VM option 
"-Dspark.local=true"
 !image-2020-07-08-17-41-35-954.png|width=574,height=363!
 * Load csv data source by pressing button "Data Source->Load CSV File as 
Table" on "Model" page, and set the schema for your table. Then press "submit" 
to save.
 !image-2020-07-08-17-42-09-603.png|width=577,height=259!

Most time we debug just want to build and query cube quickly and focus the bug 
we want to resolve. But current way is complex to load csv tables, create model 
and cube and it's hard to use kylin sample cube. So, I want to add a csv source 
which using the model of kylin sample data directly when debug tomcat started.


> Debug the code of Kylin on Parquet without hadoop environment
> -------------------------------------------------------------
>
>                 Key: KYLIN-4625
>                 URL: https://issues.apache.org/jira/browse/KYLIN-4625
>             Project: Kylin
>          Issue Type: Improvement
>          Components: Spark Engine
>            Reporter: wangrupeng
>            Assignee: wangrupeng
>            Priority: Major
>         Attachments: image-2020-07-08-17-41-35-954.png, 
> image-2020-07-08-17-42-09-603.png, screenshot-1.png
>
>
> Currently, Kylin on Parquet already supports debuging source code with local 
> csv files, but it's a little bit complex. The steps are as follows:
>  * edit the properties of 
> $KYLIN_SOURCE_DIR/examples/test_case_data/sandbox/kylin.properties to local
> {code:java}
> kylin.metadata.url=$LOCAL_META_DIR
>  kylin.env.zookeeper-is-local=true
>  kylin.env.hdfs-working-dir=file:///path/to/local/dir
>  kylin.engine.spark-conf.spark.master=local
>  kylin.engine.spark-conf.spark.eventLog.dir=/path/to/local/dir
>  kylin.env=UT{code}
>  * debug org.apache.kylin.rest.DebugTomcat with IDEA && add VM option 
> "-Dspark.local=true"
>  !image-2020-07-08-17-41-35-954.png|width=574,height=363!
>  * Load csv data source by pressing button "Data Source->Load CSV File as 
> Table" on "Model" page, and set the schema for your table. Then press 
> "submit" to save.
>  !image-2020-07-08-17-42-09-603.png|width=577,height=259!
> Most time we debug just want to build and query cube quickly and focus the 
> bug we want to resolve. But current way is complex to load csv tables, create 
> model and cube and it's hard to use kylin sample cube. So, I want to add a 
> csv source which using the model of kylin sample data directly when debug 
> tomcat started.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to