Forgot to add the mailinglist
> On 18. Jan 2018, at 18:55, Jörn Franke <jornfra...@gmail.com> wrote: > > Welll you can use: > https://spark.apache.org/docs/latest/api/java/org/apache/spark/SparkContext.html#hadoopRDD-org.apache.hadoop.mapred.JobConf-java.lang.Class-java.lang.Class-java.lang.Class-int- > > with the following inputformat: > https://hive.apache.org/javadocs/r2.1.1/api/org/apache/hadoop/hive/ql/io/RCFileInputFormat.html > > (note the version of the Javadoc does not matter it is already possible since > a long time). > > Writing is similarly with PairRDD and RCFileOutputFormat > >> On Thu, Jan 18, 2018 at 5:02 PM, Michael Segel <msegel_had...@hotmail.com> >> wrote: >> No idea on how that last line of garbage got in the message. >> >> >> > On Jan 18, 2018, at 9:32 AM, Michael Segel <msegel_had...@hotmail.com> >> > wrote: >> > >> > Hi, >> > >> > I’m trying to find out if there’s a simple way for Spark to be able to >> > read an RCFile. >> > >> > I know I can create a table in Hive, then drop the files in to that >> > directory and use a sql context to read the file from Hive, however I >> > wanted to read the file directly. >> > >> > Not a lot of details to go on… even the Apache site’s links are broken. >> > See : >> > https://cwiki.apache.org/confluence/display/Hive/RCFile >> > >> > Then try to follow the Javadoc link. >> > >> > >> > Any suggestions? >> > >> > Thx >> > >> > -Mike >> > >> > >