Hi all: Please help. I directly ran a CarbonData demo program on Eclipse, which copy from carbondata-examples-spark2/src/main/scala/org/apache/carbondata/examples/CarbonSessionExample.scala, but the error occurred, as follows:
*Exception in thread "main" java.io.FileNotFoundException: file:/data/carbon_data/default/carbon_table/Metadata/schema.write * at java.io.FileOutputStream.open0(Native Method) at java.io.FileOutputStream.open(FileOutputStream.java:270) at java.io.FileOutputStream.<init>(FileOutputStream.java:213) at java.io.FileOutputStream.<init>(FileOutputStream.java:101) at org.apache.carbondata.core.datastore.impl.FileFactory.getDataOutputStream(FileFactory.java:188) at org.apache.carbondata.core.fileoperations.AtomicFileOperationsImpl.openForWrite(AtomicFileOperationsImpl.java:61) at org.apache.carbondata.core.writer.ThriftWriter.open(ThriftWriter.java:97) at org.apache.spark.sql.hive.CarbonMetastore.createSchemaThriftFile(CarbonMetastore.scala:412) at org.apache.spark.sql.hive.CarbonMetastore.createTableFromThrift(CarbonMetastore.scala:380) at org.apache.spark.sql.execution.command.CreateTable.run(carbonTableSchema.scala:166) at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:58) at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:56) at org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:74) at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114) at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114) at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:135) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:132) at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:113) at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:87) at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:87) at org.apache.spark.sql.Dataset.<init>(Dataset.scala:185) at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:64) at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:592) at cn.xm.zzc.carbontest.FirstCarbonData$.main(FirstCarbonData.scala:51) at cn.xm.zzc.carbontest.FirstCarbonData.main(FirstCarbonData.scala) but I found a file named 'schema.write' in the path '/data/carbon_data/default/carbon_table/Metadata/', the file size is 0. My program : /val warehouseLocation = Constants.SPARK_WAREHOUSE val storeLocation = Constants.CARBON_FILES //CarbonProperties.getInstance() // .addProperty("carbon.storelocation", storeLocation) CarbonProperties.getInstance() .addProperty(CarbonCommonConstants.CARBON_TIMESTAMP_FORMAT, "yyyy/MM/dd") import org.apache.spark.sql.CarbonSession._ val spark = SparkSession .builder() .appName("FirstCarbonData") .master("local") .config("spark.sql.warehouse.dir", warehouseLocation) //.config("javax.jdo.option.ConnectionURL", // s"jdbc:derby:;databaseName=${Constants.METASTORE_DB};create=true") //.enableHiveSupport() .getOrCreateCarbonSession(storeLocation, Constants.METASTORE_DB) spark.sql("DROP TABLE IF EXISTS carbon_table") // Create table spark.sql( s""" | CREATE TABLE carbon_table( | shortField short, | intField int, | bigintField long, | doubleField double, | stringField string, | timestampField timestamp, | decimalField decimal(18,2), | dateField date, | charField char(5), | floatField float, | complexData array<string> | ) | STORED BY 'carbondata' | TBLPROPERTIES('DICTIONARY_INCLUDE'='dateField, charField') """.stripMargin) val path = "/home/myubuntu/Works/workspace_latest/incubator-carbondata/examples/spark2/src/main/resources/data.csv" // scalastyle:off spark.sql( s""" | LOAD DATA LOCAL INPATH '$path' | INTO TABLE carbon_table | options('FILEHEADER'='shortField,intField,bigintField,doubleField,stringField,timestampField,decimalField,dateField,charField,floatField,complexData','COMPLEX_DELIMITER_LEVEL_1'='#') """.stripMargin) // scalastyle:on spark.sql(""" SELECT * FROM carbon_table where stringfield = 'spark' and decimalField > 40 """).show spark.sql(""" SELECT * FROM carbon_table where length(stringField) = 5 """).show spark.sql(""" SELECT * FROM carbon_table where date_format(dateField, "yyyy-MM-dd") = "2015-07-23" """).show spark.sql(""" select count(stringField) from carbon_table """.stripMargin).show spark.sql(""" SELECT sum(intField), stringField FROM carbon_table GROUP BY stringField """).show spark.sql( """ |select t1.*, t2.* |from carbon_table t1, carbon_table t2 |where t1.stringField = t2.stringField """.stripMargin).show spark.sql( """ |with t1 as ( |select * from carbon_table |union all |select * from carbon_table |) |select t1.*, t2.* |from t1, carbon_table t2 |where t1.stringField = t2.stringField """.stripMargin).show spark.sql(""" SELECT * FROM carbon_table where stringfield = 'spark' and floatField > 2.8 """).show // Drop table // spark.sql("DROP TABLE IF EXISTS carbon_table") spark.stop()/ -- View this message in context: http://apache-carbondata-mailing-list-archive.1130556.n5.nabble.com/java-io-FileNotFoundException-file-data-carbon-data-default-carbon-table-Metadata-schema-write-tp11044.html Sent from the Apache CarbonData Mailing List archive mailing list archive at Nabble.com.