i've eliminated fetch failed with this parameters (don't know which was
the right one for the problem)
to the spark-submit running with 1.2.0
--conf spark.shuffle.compress=false \
--conf spark.file.transferTo=false \
--conf spark.shuffle.manager=hash \
--conf spark.akka.frameSize=50 \
--conf spark.core.connection.ack.wait.timeout=600
..but me too i'm unable to finish a job...now i'm facing OOM's...still
trying...but at
least fetch failed are gone
bye
Il 23/12/2014 21:10, Chen Song ha scritto:
I tried both 1.1.1 and 1.2.0 (built against cdh5.1.0 and hadoop2.3)
but I am still seeing FetchFailedException.
On Mon, Dec 22, 2014 at 8:27 AM, steghe <[email protected]
<mailto:[email protected]>> wrote:
Which version of spark are you running?
It could be related to this
https://issues.apache.org/jira/browse/SPARK-3633
fixed in 1.1.1 and 1.2.0
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Fetch-Failure-tp20787p20811.html
Sent from the Apache Spark User List mailing list archive at
Nabble.com.
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
<mailto:[email protected]>
For additional commands, e-mail: [email protected]
<mailto:[email protected]>
--
Chen Song
--
____________________________________________________________
Stefano Ghezzi ICTeam S.p.A
Project Manager - PMP
tel 035 4232129 fax 035 4522034
email [email protected] url http://www.icteam.com
mobile 335 7308587
____________________________________________________________