Hi,

We are experiencing a potential bug with how Livy reports the state of a
batch job. We have infrastructure that polls the state of a batch
application and we've noticed that when a fatal exception occurs within the
application and there is a non-zero exit code in spark-submit, we see that
Livy still returns a 'success' state for that batch job.

We confirm the exit code and the exception through the Livy server logs.

So for example... after an exit code of 1, we would see a response from
the API like...

{
    "id": 277,
    "name": null,
    "state": "success",
    "appId": "application_***",
    "appInfo": {
        "driverLogUrl": null,
        "sparkUiUrl": "*********"
    },
    "log": [

"20/05/07 10:25:04 INFO SparkContext: Successfully stopped SparkContext",
        "20/05/07 10:25:04 INFO SparkContext: SparkContext already stopped."
,
   ......

"20/05/07 10:25:04 INFO ShutdownHookManager: Deleting directory
/mnt/tmp/spark-62127d15-3ffd-4ec9-a8c0-c45a26707c3f"
,

"20/05/07 10:25:04 INFO ShutdownHookManager: Deleting directory
/mnt/tmp/spark-67271639-4c3e-4a13-aff3-082e5705cbed"
,
        "\nstderr: ",
        "\nYARN Diagnostics: "
    ]
}

Can you confirm if this is an known issue, expected behaviour or a
misconfiguration? We are running these spark applications on AWS EMR. We
haven't applied any custom configuration to Livy.

Java Spark version 2.4.0
AWS EMR version: emr-5.24.0

Thanks,

Colin

-- 
This message contains proprietary information from Equifax which may be 
confidential. If you are not an intended recipient, please refrain from any 
disclosure, copying, distribution or use of this information and note that 
such actions are prohibited. If you have received this transmission in 
error, please notify by e-mail postmas...@equifax.com 
<mailto:postmas...@equifax.com>. Equifax® is a registered trademark of 
Equifax Inc. All rights reserved.









Reply via email to