Hi

The issue has been solved with Eason's help
@yanwei: Welcome to try Apache CarbonData , look forward to seeing your 
contribution in community :)

Regards
Liang
发件人: [email protected] [mailto:[email protected]] 
发送时间: 2016年6月30日 18:35
收件人: Linyixin (Eason)
抄送: Chenliang (Liang, CarbonData)
主题: 回复: 答复: 按quick start说明在spark-shell执行到create table时报错

Hi Linyixin:
  按您的提示已解决谢谢, 最终解决是把spark安装目录权限改为hdfs,因为我是用hdfs起的spark-shell
4 drwxr-xr-x. 10 hdfs hdfs  4096 Jun 28 17:53 spark
sudo -u hdfs  spark-shell --master local --jars ${carbondata_jar},${mysql_jar}

-----邮件原件-----
发件人: Ravindra Pesala [mailto:[email protected]] 
发送时间: 2016年6月30日 18:35
收件人: [email protected]
主题: Re: 按quick start说明在spark-shell执行到create table时报错

Hi Yangwei,

It seems user does not have permission to create files inside the store path 
(/mnt/resource/opt/cloudera/parcels/CDH-5.6.1-1.cdh5.6.1.
p0.3/lib/spark/carbondata/store) you provided. Please make sure the user has 
read/write permissions to store path.

Regards,
Ravindra.

On 30 June 2016 at 13:07, [email protected] <[email protected]> wrote:

> Hi,
>   我在https://github.com/HuaweiBigData/carbondata/wiki/Quick-Start
>   执行如下:
>  scala> import org.apache.spark.sql.CarbonContext
> import org.apache.spark.sql.CarbonContext
>
> scala> import java.io.File
> import java.io.File
>
> scala> import org.apache.hadoop.hive.conf.HiveConf
> import org.apache.hadoop.hive.conf.HiveConf
>
> scala> val metadata = new File("").getCanonicalPath +
> "/carbondata/metadata"
> metadata: String =
> /mnt/resource/opt/cloudera/parcels/CDH-5.6.1-1.cdh5.6.1.p0.3/lib/spark
> /carbondata/metadata
>
> scala> val cc = new CarbonContext(sc,
> "/mnt/resource/opt/cloudera/parcels/CDH-5.6.1-1.cdh5.6.1.p0.3/lib/spar
> k/carbondata/store")
> cc: org.apache.spark.sql.CarbonContext = 
> org.apache.spark.sql.CarbonContext@2746b25b
>
> scala>
> cc.setConf("carbon.kettle.home","/mnt/resource/opt/cloudera/parcels/CD
> H-5.6.1-1.cdh5.6.1.p0.3/lib/spark/carbondata/carbonplugins")
>
> scala> val metadata = new File("").getCanonicalPath +
> "/carbondata/metadata"
> metadata: String =
> /mnt/resource/opt/cloudera/parcels/CDH-5.6.1-1.cdh5.6.1.p0.3/lib/spark
> /carbondata/metadata
>
> scala> cc.setConf("hive.metastore.warehouse.dir", metadata)
>
> scala> cc.setConf(HiveConf.ConfVars.HIVECHECKFILEFORMAT.varname, 
> scala> "false")
>
> scala> cc.sql("create table if not exists table1 (id string, name 
> scala> string,
> city string, age Int) STORED BY 'org.apache.carbondata.format'") AUDIT 
> 30-06 07:27:48,141 - [BDGroup01]Creating timestamp file
> java.io.IOException: Permission denied at 
> java.io.UnixFileSystem.createFileExclusively(Native Method) at 
> java.io.File.createNewFile(File.java:1006)
> at
> org.carbondata.core.datastorage.store.impl.FileFactory.createNewFile(F
> ileFactory.java:357)
> at
> org.apache.spark.sql.hive.CarbonMetastoreCatalog.updateSchemasUpdatedT
> ime(CarbonMetastoreCatalog.scala:584)
> at
> org.apache.spark.sql.hive.CarbonMetastoreCatalog.loadMetadata(CarbonMe
> tastoreCatalog.scala:225)
> at
> org.apache.spark.sql.hive.CarbonMetastoreCatalog.<init>(CarbonMetastor
> eCatalog.scala:113)
> at
> org.apache.spark.sql.CarbonContext$$anon$1.<init>(CarbonContext.scala:
> 45)
> at
> org.apache.spark.sql.CarbonContext.catalog$lzycompute(CarbonContext.sc
> ala:45) at 
> org.apache.spark.sql.CarbonContext.catalog(CarbonContext.scala:43)
> at
> org.apache.spark.sql.CarbonContext.analyzer$lzycompute(CarbonContext.s
> cala:49) at 
> org.apache.spark.sql.CarbonContext.analyzer(CarbonContext.scala:49)
> at
> org.apache.spark.sql.SQLContext$QueryExecution.assertAnalyzed(SQLConte
> xt.scala:914)
>
>
>
>
> 杨卫
>
> 15116966545
>
>
>


--
Thanks & Regards,
Ravi

Reply via email to