Hi All,
I am trying to access hive from Spark but getting exception
The root scratch dir: /tmp/hive on HDFS should be writable. Current
permissions are: rw-rw-rw-
Code :-
String logFile = "hdfs://hdp23ha/logs"; // Should be some file
on
// your system]
System.setProperty("HADOOP_USER_NAME","hadoop");
SparkConf conf = new SparkConf().setAppName("Simple
Application").setMaster("local[2]");
JavaSparkContext sc = new JavaSparkContext(conf);
sc.hadoopConfiguration().set("fs.defaultFS",
"hdfs://hdp23ha");
sc.hadoopConfiguration().set("hive.metastore.warehouse.dir",
"/apps/hive/warehouse");
sc.hadoopConfiguration().set("hive.exec.local.scratchdir", "D://");
sc.hadoopConfiguration().set("dfs.nameservices",
"hdp23ha");
sc.hadoopConfiguration().set("hive.exec.scratchdir", "/tmp/hive/");
sc.hadoopConfiguration().setInt("hive.exec.scratchdir.permission",
777);
sc.hadoopConfiguration().set("dfs.ha.namenodes.hdp23ha", "nn1,nn2");
sc.hadoopConfiguration().set("dfs.namenode.rpc-address.hdp23ha.nn1",
"ambarimaster:8020");
sc.hadoopConfiguration().set("dfs.namenode.rpc-address.hdp23ha.nn2",
"hdp231:8020");
sc.hadoopConfiguration().set("dfs.client.failover.proxy.provider.hdp23ha",
"org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider");
HiveContext hiveContext = new
org.apache.spark.sql.hive.HiveContext(sc);
JavaRDD<String> logData = sc.textFile(logFile).cache();
hiveContext.sql("CREATE EXTERNAL TABLE IF NOT EXISTS Logs (date
STRING,msg
STRING) STORED AS ORC");
JavaRDD<Logs> logsRDD=logData.map(new Function<String,
com.upwork.sparketl.core.bean.Logs>() {
public Logs call(String arg0) throws Exception {
String array[]=arg0.split(",");
Logs logs=new Logs(array[0],array[1]);
return logs;
}
});
// sc is an existing JavaSparkContext.
SQLContext sqlContext = new org.apache.spark.sql.SQLContext(sc);
DataFrame schemaPeople = sqlContext.createDataFrame(logsRDD,
Logs.class);
schemaPeople.registerTempTable("logs");
DataFrame results = sqlContext.sql("SELECT * FROM
logs");
results.write().format("orc").saveAsTable("Logs");
Any help would of a great help
Thanks
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Unable-To-access-Hive-From-Spark-tp26788.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]