wangrupeng created KYLIN-4625:
---------------------------------

             Summary: Debug the code of Kylin on Parquet without hadoop 
environment
                 Key: KYLIN-4625
                 URL: https://issues.apache.org/jira/browse/KYLIN-4625
             Project: Kylin
          Issue Type: Improvement
          Components: Spark Engine
            Reporter: wangrupeng
            Assignee: wangrupeng
         Attachments: image-2020-07-08-17-41-35-954.png, 
image-2020-07-08-17-42-09-603.png

Currently, Kylin on Parquet already supports debuging source code with local 
csv files, but it's a little bit complex. The steps are as follows:
* edit the properties of 
$KYLIN_SOURCE_DIR/examples/test_case_data/sandbox/kylin.properties to local
   ```log
   kylin.metadata.url=$LOCAL_META_DIR
   kylin.env.zookeeper-is-local=true
   kylin.env.hdfs-working-dir=file:///path/to/local/dir
   kylin.engine.spark-conf.spark.master=local
   kylin.engine.spark-conf.spark.eventLog.dir=/path/to/local/dir
   ```
* debug org.apache.kylin.rest.DebugTomcat with IDEA && add VM option 
"-Dspark.local=true" 
    !image-2020-07-08-17-41-35-954.png! 
* Load csv data source by pressing button "Data Source->Load CSV File as Table" 
on "Model" page, and set the schema for your table. Then press "submit" to save.
     !image-2020-07-08-17-42-09-603.png! 

Most time we debug just want to build and query cube easy. But current way is 
complex to load csv tables and create model and cube. So, I want to add a csv 
source  which using the model of kylin sample data directly when debug tomcat 
started.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to