Hi Sandeep can you try this ? results <- sql(hivecontext, "FROM test SELECT id","")
Thanks Deepak On Tue, Jan 5, 2016 at 5:49 PM, Sandeep Khurana <sand...@infoworks.io> wrote: > Thanks Deepak. > > I tried this as well. I created a hivecontext with "hivecontext <<- > sparkRHive.init(sc) " . > > When I tried to read hive table from this , > > results <- sql(hivecontext, "FROM test SELECT id") > > I get below error, > > Error in callJMethod(sqlContext, "sql", sqlQuery) : > Invalid jobj 2. If SparkR was restarted, Spark operations need to be > re-executed. > > > Not sure what is causing this? Any leads or ideas? I am using rstudio. > > > > On Tue, Jan 5, 2016 at 5:35 PM, Deepak Sharma <deepakmc...@gmail.com> > wrote: > >> Hi Sandeep >> I am not sure if ORC can be read directly in R. >> But there can be a workaround .First create hive table on top of ORC >> files and then access hive table in R. >> >> Thanks >> Deepak >> >> On Tue, Jan 5, 2016 at 4:57 PM, Sandeep Khurana <sand...@infoworks.io> >> wrote: >> >>> Hello >>> >>> I need to read an ORC files in hdfs in R using spark. I am not able to >>> find a package to do that. >>> >>> Can anyone help with documentation or example for this purpose? >>> >>> -- >>> Architect >>> Infoworks.io >>> http://Infoworks.io >>> >> >> >> >> -- >> Thanks >> Deepak >> www.bigdatabig.com >> www.keosha.net >> > > > > -- > Architect > Infoworks.io > http://Infoworks.io > -- Thanks Deepak www.bigdatabig.com www.keosha.net