Find the log from rm below, before FNFE there is no earlier errors in
driver log,

16/07/04 00:27:56 INFO mapreduce.TableInputFormatBase: Input split length:
0 bytes.
16/07/04 00:27:56 INFO executor.Executor: Executor is trying to kill task
56.0 in stage 2437.0 (TID 328047)
16/07/04 00:27:56 INFO executor.Executor: Executor killed task 266.0 in
stage 2433.0 (TID 328005)
16/07/04 00:27:56 INFO executor.Executor: Executor killed task 206.0 in
stage 2433.0 (TID 327977)
16/07/04 00:27:56 INFO executor.Executor: Executor killed task 318.0 in
stage 2433.0 (TID 328006)
16/07/04 00:27:57 INFO executor.Executor: Executor killed task 56.0 in
stage 2437.0 (TID 328047)
16/07/04 00:27:57 INFO executor.CoarseGrainedExecutorBackend: Driver
commanded a shutdown
16/07/04 00:27:57 INFO storage.MemoryStore: MemoryStore cleared
16/07/04 00:27:57 INFO storage.BlockManager: BlockManager stopped
16/07/04 00:27:57 WARN executor.CoarseGrainedExecutorBackend: An unknown (
driver.domain.com:56055) driver disconnected.
16/07/04 00:27:57 ERROR executor.CoarseGrainedExecutorBackend: Driver
xx:xx:xx:xx:56055 disassociated! Shutting down.
16/07/04 00:27:57 INFO util.ShutdownHookManager: Shutdown hook called
16/07/04 00:27:57 INFO util.ShutdownHookManager: Deleting directory
/opt/mapr/tmp/hadoop-tmp/hadoop-mapr/nm-local-dir/usercache/user/appcache/application_1467474162580_29353/spark-9c0bfccc-74c3-4541-a2fd-19101e47b49a
End of LogType:stderr


On Mon, Jul 4, 2016 at 4:21 PM, Jacek Laskowski <ja...@japila.pl> wrote:

> Can you share some stats from Web UI just before the failure? Any earlier
> errors before FNFE?
>
> Jacek
> On 4 Jul 2016 12:34 p.m., "kishore kumar" <akishore...@gmail.com> wrote:
>
>> @jacek: It is running on yarn-client mode, our code don't support running
>> in yarn-cluster mode and the job is running for around an hour and giving
>> the exception.
>>
>> @karhi: yarn application status is successful, resourcemanager logs did
>> not give any failure info except
>> 16/07/04 00:27:57 INFO executor.CoarseGrainedExecutorBackend: Driver
>> commanded a shutdown
>> 16/07/04 00:27:57 INFO storage.MemoryStore: MemoryStore cleared
>> 16/07/04 00:27:57 INFO storage.BlockManager: BlockManager stopped
>> 16/07/04 00:27:57 WARN executor.CoarseGrainedExecutorBackend: An unknown (
>> slave1.domain.com:56055) driver disconnected.
>> 16/07/04 00:27:57 ERROR executor.CoarseGrainedExecutorBackend: Driver
>> 173.36.88.26:56055 disassociated! Shutting down.
>> 16/07/04 00:27:57 INFO util.ShutdownHookManager: Shutdown hook called
>> 16/07/04 00:27:57 INFO util.ShutdownHookManager: Deleting directory
>> /opt/mapr/tmp/hadoop-tmp/hadoop-mapr/nm-local-dir/usercache/user/appcache/application_1467474162580_29353/spark-9c0bfccc-74c3-4541-a2fd-19101e47b49a
>> End of LogType:stderr
>>
>>
>> On Mon, Jul 4, 2016 at 3:20 PM, Jacek Laskowski <ja...@japila.pl> wrote:
>>
>>> Hi,
>>>
>>> You seem to be using yarn. Is this cluster or client deploy mode? Have
>>> you seen any other exceptions before? How long did the application run
>>> before the exception?
>>>
>>> Pozdrawiam,
>>> Jacek Laskowski
>>> ----
>>> https://medium.com/@jaceklaskowski/
>>> Mastering Apache Spark http://bit.ly/mastering-apache-spark
>>> Follow me at https://twitter.com/jaceklaskowski
>>>
>>>
>>> On Mon, Jul 4, 2016 at 10:57 AM, kishore kumar <akishore...@gmail.com>
>>> wrote:
>>> > We've upgraded spark version from 1.2 to 1.6 still the same problem,
>>> >
>>> > Exception in thread "main" org.apache.spark.SparkException: Job
>>> aborted due
>>> > to stage failure: Task 286 in stage
>>> > 2397.0 failed 4 times, most recent failure: Lost task 286.3 in stage
>>> 2397.0
>>> > (TID 314416, salve-06.domain.com): java.io.FileNotFoundException:
>>> > /opt/mapr/tmp/h
>>> >
>>> adoop-tmp/hadoop-mapr/nm-local-dir/usercache/user1/appcache/application_1467474162580_29353/blockmgr-bd075392-19c2-4cb8-8033-0fe54d683c8f/12/shuffle_530_286_0.inde
>>> > x.c374502a-4cf2-4052-abcf-42977f1623d0 (No such file or directory)
>>> >
>>> > Kindly help me to get rid from this.
>>> >
>>> > On Sun, Jun 5, 2016 at 9:43 AM, kishore kumar <akishore...@gmail.com>
>>> wrote:
>>> >>
>>> >> Hi,
>>> >>
>>> >> Could anyone help me about this error ? why this error comes ?
>>> >>
>>> >> Thanks,
>>> >> KishoreKuamr.
>>> >>
>>> >> On Fri, Jun 3, 2016 at 9:12 PM, kishore kumar <akishore...@gmail.com>
>>> >> wrote:
>>> >>>
>>> >>> Hi Jeff Zhang,
>>> >>>
>>> >>> Thanks for response, could you explain me why this error occurs ?
>>> >>>
>>> >>> On Fri, Jun 3, 2016 at 6:15 PM, Jeff Zhang <zjf...@gmail.com> wrote:
>>> >>>>
>>> >>>> One quick solution is to use spark 1.6.1.
>>> >>>>
>>> >>>> On Fri, Jun 3, 2016 at 8:35 PM, kishore kumar <
>>> akishore...@gmail.com>
>>> >>>> wrote:
>>> >>>>>
>>> >>>>> Could anyone help me on this issue ?
>>> >>>>>
>>> >>>>> On Tue, May 31, 2016 at 8:00 PM, kishore kumar <
>>> akishore...@gmail.com>
>>> >>>>> wrote:
>>> >>>>>>
>>> >>>>>> Hi,
>>> >>>>>>
>>> >>>>>> We installed spark1.2.1 in single node, running a job in
>>> yarn-client
>>> >>>>>> mode on yarn which loads data into hbase and elasticsearch,
>>> >>>>>>
>>> >>>>>> the error which we are encountering is
>>> >>>>>> Exception in thread "main" org.apache.spark.SparkException: Job
>>> >>>>>> aborted due to stage failure: Task 38 in stage 26800.0 failed 4
>>> times, most
>>> >>>>>> recent failure: Lost task 38.3 in stage 26800.0 (TID 4990082,
>>> >>>>>> hdprd-c01-r04-03): java.io.FileNotFoundException:
>>> >>>>>>
>>> /opt/mapr/tmp/hadoop-tmp/hadoop-mapr/nm-local-dir/usercache/sparkuser/appcache/application_1463194314221_211370/spark-3cc37dc7-fa3c-4b98-aa60-0acdfc79c725/28/shuffle_8553_38_0.index
>>> >>>>>> (No such file or directory)
>>> >>>>>>
>>> >>>>>> any idea about this error ?
>>> >>>>>> --
>>> >>>>>> Thanks,
>>> >>>>>> Kishore.
>>> >>>>>
>>> >>>>>
>>> >>>>>
>>> >>>>>
>>> >>>>> --
>>> >>>>> Thanks,
>>> >>>>> Kishore.
>>> >>>>
>>> >>>>
>>> >>>>
>>> >>>>
>>> >>>> --
>>> >>>> Best Regards
>>> >>>>
>>> >>>> Jeff Zhang
>>> >>>
>>> >>>
>>> >>>
>>> >>>
>>> >>> --
>>> >>> Thanks,
>>> >>> Kishore.
>>> >>
>>> >>
>>> >>
>>> >>
>>> >> --
>>> >> Thanks,
>>> >> Kishore.
>>> >
>>> >
>>> >
>>> >
>>> > --
>>> > Thanks,
>>> > Kishore.
>>>
>>
>>
>>
>> --
>> Thanks,
>> Kishore.
>>
>


-- 
Thanks,
Kishore.

Reply via email to