How are you creating the object in your Scala shell? Maybe you can write a function that directly returns the RDD, without assigning the object to a temporary variable.
Matei > On Nov 5, 2014, at 2:54 PM, Corey Nolet <cjno...@gmail.com> wrote: > > The closer I look @ the stack trace in the Scala shell, it appears to be the > call to toString() that is causing the construction of the Job object to > fail. Is there a ways to suppress this output since it appears to be > hindering my ability to new up this object? > > On Wed, Nov 5, 2014 at 5:49 PM, Corey Nolet <cjno...@gmail.com > <mailto:cjno...@gmail.com>> wrote: > I'm trying to use a custom input format with SparkContext.newAPIHadoopRDD. > Creating the new RDD works fine but setting up the configuration file via the > static methods on input formats that require a Hadoop Job object is proving > to be difficult. > > Trying to new up my own Job object with the SparkContext.hadoopConfiguration > is throwing the exception on line 283 of this grepcode: > > http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hadoop/hadoop-mapreduce-client-core/2.5.0/org/apache/hadoop/mapreduce/Job.java#Job > > <http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hadoop/hadoop-mapreduce-client-core/2.5.0/org/apache/hadoop/mapreduce/Job.java#Job> > > Looking in the SparkContext code, I'm seeing that it's newing up Job objects > just fine using nothing but the configuraiton. Using SparkContext.textFile() > appears to be working for me. Any ideas? Has anyone else run into this as > well? Is it possible to have a method like SparkContext.getJob() or something > similar? > > Thanks. > >