Github user chenerlu commented on a diff in the pull request: https://github.com/apache/carbondata/pull/1112#discussion_r124708068 --- Diff: integration/presto/README.md --- @@ -59,28 +55,50 @@ Please follow the below steps to query carbondata in presto ``` * config carbondata-connector for presto - First:compile carbondata-presto integration module + Firstly: Compile carbondata, including carbondata-presto integration module ``` $ git clone https://github.com/apache/carbondata - $ cd carbondata/integration/presto - $ mvn clean package + $ cd carbondata + $ mvn -DskipTests -P{spark-version} -Dspark.version={spark-version-number} -Dhadoop.version={hadoop-version-number} clean package + ``` + Replace the spark and hadoop version with you the version you used in your cluster. + For example, if you use Spark2.1.0 and Hadoop 2.7.3, you would like to compile using: + ``` + mvn -DskipTests -Pspark-2.1 -Dspark.version=2.1.0 -Dhadoop.version=2.7.3 clean package + ``` + + Secondly: Create a folder named 'carbondata' under $PRESTO_HOME$/plugin and + copy all jar from carbondata/integration/presto/target/carbondata-presto-x.x.x-SNAPSHOT + to $PRESTO_HOME$/plugin/carbondata + + Thirdly: Create a carbondata.properties file under $PRESTO_HOME$/etc/catalog/ containing the following contents: ``` - Second:create one folder "carbondata" under ./presto-server-0.166/plugin - Third:copy all jar from ./carbondata/integration/presto/target/carbondata-presto-x.x.x-SNAPSHOT - to ./presto-server-0.166/plugin/carbondata + connector.name=carbondata + carbondata-store={schema-store-path} + ``` + Replace the schema-store-path with the absolute path the directory which is the parent of the schema. + For example, if you have a schema named 'default' stored under hdfs://namenode:9000/test/carbondata/, + Then set carbondata-store=hdfs://namenode:9000/test/carbondata + + If you changed the jar balls or configuration files, make sure you have dispatch the new jar balls + and configuration file to all the presto nodes and restart the nodes in the cluster. A modification of the + carbondata connector will not take an effect automatically. ### Generate CarbonData file -Please refer to quick start : https://github.com/apache/carbondata/blob/master/docs/quick-start-guide.md +Please refer to quick start: https://github.com/apache/carbondata/blob/master/docs/quick-start-guide.md +Load data statement in Spark can be used to create carbondata tables. And you can easily find the creaed --- End diff -- created -> created
--- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---