Change your import from mapred to mapreduce. like : import org.apache.accumulo.core.client.mapreduce.AccumuloInputFormat;
Thanks Best Regards On Wed, Apr 22, 2015 at 2:42 PM, madhvi <madhvi.gu...@orkash.com> wrote: > Hi, > > I am creating a spark RDD through accumulo writing like: > > JavaPairRDD<Key, Value> accumuloRDD = > sc.newAPIHadoopRDD(accumuloJob.getConfiguration(),AccumuloInputFormat.class,Key.class, > Value.class); > > But I am getting the following error and it is not getting compiled: > > Bound mismatch: The generic method newAPIHadoopRDD(Configuration, > Class<F>, Class<K>, Class<V>) of type JavaSparkContext is not applicable > for the arguments (Configuration, Class<AccumuloInputFormat>, Class<Key>, > Class<Value>). The inferred type AccumuloInputFormat is not a valid > substitute for the bounded parameter <F extends InputFormat<K,V>> > > I am using the following import statements: > > import org.apache.accumulo.core.client.mapred.AccumuloInputFormat; > import org.apache.accumulo.core.data.Key; > import org.apache.accumulo.core.data.Value; > > I am not getting what is the problem in this. > > Thanks > Madhvi > > > --------------------------------------------------------------------- > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > For additional commands, e-mail: user-h...@spark.apache.org > >