Bhagavan created FLINK-17969:
--------------------------------
Summary: Enhance Flink (Task) logging to include job name as
context diagnostic information
Key: FLINK-17969
URL: https://issues.apache.org/jira/browse/FLINK-17969
Project: Flink
Issue Type: Improvement
Components: Runtime / Task
Affects Versions: 1.10.0
Reporter: Bhagavan
Problem statement:
We use a shared session cluster (Standalone/Yarn) to execute jobs. All logs
from the cluster are shipped using log aggregation framework (Logstash/Splunk)
so that application diagnostic is easier.
However, we are missing one vital information in the logline. i.e. Job name so
that we can filter the logs for a single job.
Background
Currently, Flink logging uses SLF4J as API to abstract away from concrete
logging implementation (log4j 1.x, Logback or log4j2) and configuration of
logging pattern and implementation can be configured at deployment, However,
there is no MDC info from framework indicating job context.
Proposed improvement.
Add jobName field to Task class so that we can add it as MDC when task thread
starts executing.
Change is trivial and uses SLF4J MDC API.
With this change, user can customise logging pattern to include MDC (e.g. in
Logback [%X{jobName}])
Change required.
{code:java}
@@ -319,6 +323,7 @@ public class Task implements Runnable, TaskSlotPayload,
TaskActions, PartitionPr
this.jobId = jobInformation.getJobId();
+ this.jobName = jobInformation.getJobName();
this.vertexId = taskInformation.getJobVertexId();
@@ -530,8 +535,10 @@ public class Task implements Runnable, TaskSlotPayload,
TaskActions, PartitionPr
@Override
public void run() {
try {
+ MDC.put("jobName", this.jobName);
doRun();
} finally {
+ MDC.remove("jobName");
terminationFuture.complete(executionState);
}
}
{code}
if we are in agreement for this small change. Will raise PR.
--
This message was sent by Atlassian Jira
(v8.3.4#803005)