Please take a look at core/src/test/java/org/apache/spark/JavaAPISuite.java in source code.
Cheers On Tue, Jul 7, 2015 at 7:17 AM, 付雅丹 <[email protected]> wrote: > Hi, everyone! > > I've got <key,value> pair in form of <LongWritable, Text>, where I used > the following code: > > SparkConf conf = new SparkConf().setAppName("MapReduceFileInput"); > JavaSparkContext sc = new JavaSparkContext(conf); > Configuration confHadoop = new Configuration(); > > JavaPairRDD<LongWritable,Text> sourceFile=sc.newAPIHadoopFile( > "hdfs://cMaster:9000/wcinput/data.txt", > DataInputFormat.class,LongWritable.class,Text.class,confHadoop); > > Now I want to handle the javapairrdd data from <LongWritable, Text> to > another <LongWritable, Text>, where the Text content is different. After > that, I want to write Text into hdfs in order of LongWritable value. But I > don't know how to write mapreduce function in spark using java language. > Someone can help me? > > > Sincerely, > Missie. >
