; >>
>>> >> By setting --hadoop-major-version=2 when using the ec2 scripts,
>>> >> everything worked fine.
>>> >>
>>> >> Darin.
>>> >>
>>> >>
>>> >> - Original Message -
>>>
> >> Just to close the loop in case anyone runs into the same problem I had.
>> >>
>> >> By setting --hadoop-major-version=2 when using the ec2 scripts,
>> >> everything worked fine.
>> >>
>> >> Darin.
>> >>
>>
one runs into the same problem I had.
> >>
> >> By setting --hadoop-major-version=2 when using the ec2 scripts,
> >> everything worked fine.
> >>
> >> Darin.
> >>
> >>
> >> - Original Message -
> >> From: Darin McBeath
&
gt;> >at
>>
>> >org.apache.spark.rdd.PairRDDFunctions.saveAsHadoopFile(PairRDDFunctions.sc
>> >ala:940)
>> >at
>>
>> >org.apache.spark.rdd.PairRDDFunctions.saveAsHadoopFile(PairRDDFunctions.sc
>> >ala:902)
>> >at
>>
>> >org.apache.spark.a
ajor-version=2 when using the ec2 scripts, everything
> worked fine.
>
> Darin.
>
>
> - Original Message -
> From: Darin McBeath
> To: Mingyu Kim ; Aaron Davidson
> Cc: "user@spark.apache.org"
> Sent: Monday, February 23, 2015 3:16 PM
> Subject
nt: Monday, February 23, 2015 3:16 PM
Subject: Re: Which OutputCommitter to use for S3?
Thanks. I think my problem might actually be the other way around.
I'm compiling with hadoop 2, but when I startup Spark, using the ec2 scripts,
I don't specify a
-hadoop-major-version and the
I'll try it and post a response.
- Original Message -
From: Mingyu Kim
To: Darin McBeath ; Aaron Davidson
Cc: "user@spark.apache.org"
Sent: Monday, February 23, 2015 3:06 PM
Subject: Re: Which OutputCommitter to use for S3?
Cool, we will start from there. Thanks Aaron
p.mapred.JobContext.
>
>Is there something obvious that I might be doing wrong (or messed up in
>the translation from Scala to Java) or something I should look into? I'm
>using Spark 1.2 with hadoop 2.4.
>
>
>Thanks.
>
>Darin.
>
>
>______________
From: Aaron Davidson
To: Andrew Ash
Cc: Josh Rosen ; Mingyu Kim ;
"user@spark.apache.org" ; Aaron Davidson
Sent: Saturday, February 21, 2015 7:01 PM
Subject: Re: Which OutputCommitter to use for S3?
Here is the class: https://gist.github.com/aarond
one using
>>> a special OutputCommitter for S3 can comment on this!
>>>
>>> Thanks,
>>> Mingyu
>>>
>>> From: Mingyu Kim
>>> Date: Monday, February 16, 2015 at 1:15 AM
>>> To: "user@spark.apache.org"
>>> Subject: W
nyone using a
>> special OutputCommitter for S3 can comment on this!
>>
>> Thanks,
>> Mingyu
>>
>> From: Mingyu Kim
>> Date: Monday, February 16, 2015 at 1:15 AM
>> To: "user@spark.apache.org"
>> Subject: Which OutputCommitter
any response. It’d be really appreciated if anyone using a
> special OutputCommitter for S3 can comment on this!
>
> Thanks,
> Mingyu
>
> From: Mingyu Kim
> Date: Monday, February 16, 2015 at 1:15 AM
> To: "user@spark.apache.org"
> Subject: Which OutputCommitt
ark.apache.org>"
mailto:user@spark.apache.org>>
Subject: Which OutputCommitter to use for S3?
HI all,
The default OutputCommitter used by RDD, which is FileOutputCommitter, seems to
require moving files at the commit step, which is not a constant operation in
S3, as discussed in
http://mail-
HI all,
The default OutputCommitter used by RDD, which is FileOutputCommitter, seems to
require moving files at the commit step, which is not a constant operation in
S3, as discussed in
http://mail-archives.apache.org/mod_mbox/spark-user/201410.mbox/%3c543e33fa.2000...@entropy.be%3E.
People se
14 matches
Mail list logo