Yeah, unfortunately that will be up to them to fix, though it wouldn't hurt to 
send them a JIRA mentioning this.

Matei

> On Nov 25, 2014, at 2:58 PM, Corey Nolet <cjno...@gmail.com> wrote:
> 
> I was wiring up my job in the shell while i was learning Spark/Scala. I'm 
> getting more comfortable with them both now so I've been mostly testing 
> through Intellij with mock data as inputs.
> 
> I think the problem lies more on Hadoop than Spark as the Job object seems to 
> check it's state and throw an exception when the toString() method is called 
> before the Job has physically been submitted.
> 
> On Tue, Nov 25, 2014 at 5:31 PM, Matei Zaharia <matei.zaha...@gmail.com 
> <mailto:matei.zaha...@gmail.com>> wrote:
> How are you creating the object in your Scala shell? Maybe you can write a 
> function that directly returns the RDD, without assigning the object to a 
> temporary variable.
> 
> Matei
> 
>> On Nov 5, 2014, at 2:54 PM, Corey Nolet <cjno...@gmail.com 
>> <mailto:cjno...@gmail.com>> wrote:
>> 
>> The closer I look @ the stack trace in the Scala shell, it appears to be the 
>> call to toString() that is causing the construction of the Job object to 
>> fail. Is there a ways to suppress this output since it appears to be 
>> hindering my ability to new up this object?
>> 
>> On Wed, Nov 5, 2014 at 5:49 PM, Corey Nolet <cjno...@gmail.com 
>> <mailto:cjno...@gmail.com>> wrote:
>> I'm trying to use a custom input format with SparkContext.newAPIHadoopRDD. 
>> Creating the new RDD works fine but setting up the configuration file via 
>> the static methods on input formats that require a Hadoop Job object is 
>> proving to be difficult. 
>> 
>> Trying to new up my own Job object with the SparkContext.hadoopConfiguration 
>> is throwing the exception on line 283 of this grepcode:
>> 
>> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hadoop/hadoop-mapreduce-client-core/2.5.0/org/apache/hadoop/mapreduce/Job.java#Job
>>  
>> <http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hadoop/hadoop-mapreduce-client-core/2.5.0/org/apache/hadoop/mapreduce/Job.java#Job>
>> 
>> Looking in the SparkContext code, I'm seeing that it's newing up Job objects 
>> just fine using nothing but the configuraiton. Using SparkContext.textFile() 
>> appears to be working for me. Any ideas? Has anyone else run into this as 
>> well? Is it possible to have a method like SparkContext.getJob() or 
>> something similar?
>> 
>> Thanks.
>> 
>> 
> 
> 

Reply via email to