You mean the issue is resolved? Regards Raghunandan
On 06-Feb-2017 1:36 PM, "Sanoj M George" <sanoj.geo...@gmail.com> wrote: Thanks Raghunandan. Checked the thread but it seems this error is due to something else. Below are the parameters that I changed : **** carbon.properties : carbon.storelocation=hdfs://localhost:9000/opt/CarbonStore carbon.ddl.base.hdfs.url=hdfs://localhost:9000/opt/data carbon.kettle.home=/home/cduser/spark/carbonlib/carbonplugins **** spark-defaults.conf : carbon.kettle.home /home/cduser/spark/carbonlib/carbonplugins spark.driver.extraJavaOptions -Dcarbon.properties.filepath=/home/cduser/spark/conf/carbon.properties spark.executor.extraJavaOptions -Dcarbon.properties.filepath=/home/cduser/spark/conf/carbon.properties Although store location is specified in carbon.properties, spark-shell was using "/home/cduser/carbon.store" as store location. Regards On Sun, Feb 5, 2017 at 4:49 PM, Raghunandan S < carbondatacontributi...@gmail.com> wrote: > Dear sanoj, > Pls refer to > http://apache-carbondata-mailing-list-archive.1130556. > n5.nabble.com/Dictionary-file-is-locked-for-updation-td5076.html > > Let me know if this thread didn't address your problem. > > Regards > > > On 05-Feb-2017 5:22 PM, "Sanoj M George" <sanoj.geo...@gmail.com> wrote: > > Hi All, > > I am getting below error while trying out Carbondata with Spark 1.6.2 / > Hadoop 2.6.5 / Carbondata 1. > > ./bin/spark-shell --jars > carbonlib/carbondata_2.10-1.1.0-incubating-SNAPSHOT-shade-hadoop2.2.0.jar > scala> import org.apache.spark.sql.CarbonContext > scala> val cc = new CarbonContext(sc) > scala> cc.sql("CREATE TABLE IF NOT EXISTS t1 (id string, name string, city > string, age Int) STORED BY 'carbondata'") > scala> cc.sql("LOAD DATA INPATH '/home/cduser/spark/sample.csv' INTO TABLE > t1") > INFO 05-02 14:57:22,346 - main Query [LOAD DATA INPATH > '/HOME/CDUSER/SPARK/SAMPLE.CSV' INTO TABLE T1] > INFO 05-02 14:57:37,411 - Table MetaData Unlocked Successfully after data > load > java.lang.RuntimeException: Table is locked for updation. Please try after > some time > at scala.sys.package$.error(package.scala:27) > at > org.apache.spark.sql.execution.command.LoadTable. > run(carbonTableSchema.scala:360) > at > org.apache.spark.sql.execution.ExecutedCommand. > sideEffectResult$lzycompute( > commands.scala:58) > at > org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands. > scala:56) > > > I followed the docs at > https://github.com/apache/incubator-carbondata/blob/ > master/docs/installation-guide.md#installing-and- > configuring-carbondata-on- > standalone-spark-cluster > <http://installation-guide.md#installing-and-configuring- carbondata-on-%0Astandalone-spark-cluster> > and > https://github.com/apache/incubator-carbondata/blob/ > master/docs/quick-start-guide.md > to install carbondata. > > While creating the table, I observed below WARN msg in the log : > > main Query [CREATE TABLE DEFAULT.T1 USING CARBONDATA OPTIONS (TABLENAME > "DEFAULT.T1", TABLEPATH "/HOME/CDUSER/CARBON.STORE/DEFAULT/T1") ] > > WARN 05-02 14:34:30,656 - Couldn't find corresponding Hive SerDe for data > source provider carbondata. Persisting data source relation `default`.`t1` > into Hive metastore in Spark SQL specific format, which is NOT compatible > with Hive. > INFO 05-02 14:34:30,755 - 0: create_table: Table(tableName:t1, > dbName:default, owner:cduser, createTime:1486290870, lastAccessTime:0, > retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:col, > type:array<string>, comment:from deserializer)], location:null, > inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, > outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, > compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, > serializationLib:org.apache.hadoop.hive.serde2. > MetadataTypedColumnsetSerDe, > parameters:{tablePath=/home/cduser/carbon.store/default/t1, > serialization.format=1, tableName=default.t1}), bucketCols:[], sortCols:[], > parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], > skewedColValues:[], > skewedColValueLocationMaps:{})), partitionKeys:[], > parameters:{EXTERNAL=TRUE, spark.sql.sources.provider=carbondata}, > viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE, > privileges:PrincipalPrivilegeSet(userPrivileges:{}, groupPrivileges:null, > rolePrivileges:null)) > > > Appreciate any help in resolving this. > > Thanks, > Sanoj >