I am facing same issue, do you have any solution ?

On Mon, Apr 27, 2015 at 9:43 PM, Deepak Gopalakrishnan <dgk...@gmail.com>
wrote:

> Hello All,
>
> I dug a little deeper and found this error :
>
> 15/04/27 16:05:39 WARN TransportChannelHandler: Exception in connection from 
> /10.1.0.90:40590
> java.io.IOException: Connection reset by peer
>       at sun.nio.ch.FileDispatcherImpl.read0(Native Method)
>       at sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:39)
>       at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:223)
>       at sun.nio.ch.IOUtil.read(IOUtil.java:192)
>       at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:379)
>       at 
> io.netty.buffer.PooledUnsafeDirectByteBuf.setBytes(PooledUnsafeDirectByteBuf.java:311)
>       at io.netty.buffer.AbstractByteBuf.writeBytes(AbstractByteBuf.java:881)
>       at 
> io.netty.channel.socket.nio.NioSocketChannel.doReadBytes(NioSocketChannel.java:225)
>       at 
> io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:119)
>       at 
> io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:511)
>       at 
> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:468)
>       at 
> io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:382)
>       at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:354)
>       at 
> io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:116)
>       at java.lang.Thread.run(Thread.java:745)
> 15/04/27 16:05:39 ERROR TransportRequestHandler: Error sending result 
> ChunkFetchSuccess{streamChunkId=StreamChunkId{streamId=45314884029, 
> chunkIndex=0}, buffer=NioManagedBuffer{buf=java.nio.HeapByteBuffer[pos=0 
> lim=26227673 cap=26227673]}} to /10.1.0.90:40590; closing connection
> java.nio.channels.ClosedChannelException
> 15/04/27 16:05:39 ERROR TransportRequestHandler: Error sending result 
> RpcResponse{requestId=8439869725098873668, response=[B@1bdcdf63} to 
> /10.1.0.90:40590; closing connection
> java.nio.channels.ClosedChannelException
> 15/04/27 16:05:39 ERROR CoarseGrainedExecutorBackend: Driver Disassociated 
> [akka.tcp://sparkexecu...@master.spark.com:60802] -> 
> [akka.tcp://sparkdri...@master.spark.com:37195] disassociated! Shutting down.
> 15/04/27 16:05:39 WARN ReliableDeliverySupervisor: Association with remote 
> system [akka.tcp://sparkdri...@master.spark.com:37195] has failed, address is 
> now gated for [5000] ms. Reason is: [Disassociated].
>
>
> On Mon, Apr 27, 2015 at 8:35 AM, Shixiong Zhu <zsxw...@gmail.com> wrote:
>
>> The configuration key should be "spark.akka.askTimeout" for this timeout.
>> The time unit is seconds.
>>
>> Best Regards,
>> Shixiong(Ryan) Zhu
>>
>> 2015-04-26 15:15 GMT-07:00 Deepak Gopalakrishnan <dgk...@gmail.com>:
>>
>> Hello,
>>>
>>>
>>> Just to add a bit more context :
>>>
>>> I have done that in the code, but I cannot see it change from 30 seconds
>>> in the log.
>>>
>>>                 .set("spark.executor.memory", "10g")
>>>
>>>                 .set("spark.driver.memory", "20g")
>>>
>>>                 .set("spark.akka.timeout","6000")
>>>
>>> PS : I understand that 6000 is quite large, but I'm just trying to see
>>> if it actually changes
>>>
>>>
>>> Here is the command that I'm running
>>>
>>>  sudo MASTER=spark://master.spark.com:7077
>>> /opt/spark/spark-1.3.0-bin-hadoop2.4/bin/spark-submit --class
>>> "<class-name>"   --executor-memory 20G --driver-memory 10G  --deploy-mode
>>> client --conf spark.akka.timeout=6000 --conf spark.akka.askTimeout=6000
>>> <jar file path>
>>>
>>>
>>> and here is how I load the file         JavaPairRDD<String, String>
>>> learningRdd=sc.wholeTextFiles(filePath,10);
>>> Thanks
>>>
>>> On Mon, Apr 27, 2015 at 3:36 AM, Bryan Cutler <cutl...@gmail.com> wrote:
>>>
>>>> I'm not sure what the expected performance should be for this amount of
>>>> data, but you could try to increase the timeout with the property
>>>> "spark.akka.timeout" to see if that helps.
>>>>
>>>> Bryan
>>>>
>>>> On Sun, Apr 26, 2015 at 6:57 AM, Deepak Gopalakrishnan <
>>>> dgk...@gmail.com> wrote:
>>>>
>>>>> Hello All,
>>>>>
>>>>> I'm trying to process a 3.5GB file on standalone mode using spark. I
>>>>> could run my spark job succesfully on a 100MB file and it works as
>>>>> expected. But, when I try to run it on the 3.5GB file, I run into the 
>>>>> below
>>>>> error :
>>>>>
>>>>>
>>>>> 15/04/26 12:45:50 INFO BlockManagerMaster: Updated info of block 
>>>>> taskresult_83
>>>>> 15/04/26 12:46:46 WARN AkkaUtils: Error sending message [message = 
>>>>> Heartbeat(2,[Lscala.Tuple2;@790223d3,BlockManagerId(2, master.spark.com, 
>>>>> 39143))] in 1 attempts
>>>>> java.util.concurrent.TimeoutException: Futures timed out after [30 
>>>>> seconds]
>>>>>   at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>>>>>   at 
>>>>> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
>>>>>   at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
>>>>>   at 
>>>>> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
>>>>>   at scala.concurrent.Await$.result(package.scala:107)
>>>>>   at org.apache.spark.util.AkkaUtils$.askWithReply(AkkaUtils.scala:195)
>>>>>   at org.apache.spark.executor.Executor$$anon$1.run(Executor.scala:427)
>>>>> 15/04/26 12:47:15 INFO MemoryStore: ensureFreeSpace(26227673) called with 
>>>>> curMem=265897, maxMem=5556991426
>>>>> 15/04/26 12:47:15 INFO MemoryStore: Block taskresult_92 stored as bytes 
>>>>> in memory (estimated size 25.0 MB, free 5.2 GB)
>>>>> 15/04/26 12:47:16 INFO MemoryStore: ensureFreeSpace(26272879) called with 
>>>>> curMem=26493570, maxMem=5556991426
>>>>> 15/04/26 12:47:16 INFO MemoryStore: Block taskresult_94 stored as bytes 
>>>>> in memory (estimated size 25.1 MB, free 5.1 GB)
>>>>> 15/04/26 12:47:18 INFO MemoryStore: ensureFreeSpace(26285327) called with 
>>>>> curMem=52766449, maxMem=5556991426
>>>>>
>>>>>
>>>>> and the job fails.
>>>>>
>>>>>
>>>>> I'm on AWS and have opened all ports. Also, since the 100MB file
>>>>> works, it should not be a connection issue.  I've a r3 xlarge and 2 m3
>>>>> large.
>>>>>
>>>>> Can anyone suggest a way to fix this?
>>>>>
>>>>> --
>>>>> Regards,
>>>>> *Deepak Gopalakrishnan*
>>>>> *Mobile*:+918891509774
>>>>> *Skype* : deepakgk87
>>>>> http://myexps.blogspot.com
>>>>>
>>>>>
>>>>
>>>
>>>
>>> --
>>> Regards,
>>> *Deepak Gopalakrishnan*
>>> *Mobile*:+918891509774
>>> *Skype* : deepakgk87
>>> http://myexps.blogspot.com
>>>
>>>
>>
>
>
> --
> Regards,
> *Deepak Gopalakrishnan*
> *Mobile*:+918891509774
> *Skype* : deepakgk87
> http://myexps.blogspot.com
>
>


-- 
Deepak

Reply via email to