join is operation of DataFrame You can call sc.createDataFrame(myRDD) to obtain DataFrame where sc is sqlContext
Cheers On Mon, Jun 8, 2015 at 9:44 PM, amit tewari <amittewar...@gmail.com> wrote: > Hi Dear Spark Users > > I am very new to Spark/Scala. > > Am using Datastax (4.7/Spark 1.2.1) and struggling with following > error/issue. > > Already tried options like import org.apache.spark.SparkContext._ or > explicit import org.apache.spark.SparkContext.rddToPairRDDFunctions. > But error not resolved. > > Help much appreciated. > > Thanks > AT > > scala>val input1 = sc.textFile("/test7").map(line => > line.split(",").map(_.trim)); > scala>val input2 = sc.textFile("/test8").map(line => > line.split(",").map(_.trim)); > scala>val input11 = input1.map(x=>((x(0),x(1)),x(2),x(3))) > scala>val input22 = input2.map(x=>((x(0),x(1)),x(2),x(3))) > > scala> input11.join(input22).take(10) > > <console>:66: error: value join is not a member of > org.apache.spark.rdd.RDD[((String, String), String, String)] > > input11.join(input22).take(10) > > > > > > >