XComp commented on a change in pull request #18189:
URL: https://github.com/apache/flink/pull/18189#discussion_r781224232



##########
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/dispatcher/Dispatcher.java
##########
@@ -935,6 +942,19 @@ protected CleanupJobState 
jobReachedTerminalState(ExecutionGraphInfo executionGr
 
         archiveExecutionGraph(executionGraphInfo);
 
+        if (terminalJobStatus.isGloballyTerminalState()) {
+            try {
+                jobResultStore.createDirtyResult(
+                        
JobResult.createFrom(executionGraphInfo.getArchivedExecutionGraph()));
+            } catch (IOException e) {
+                log.error(

Review comment:
       Initially, I thought that we wouldn't need more sophisticated error 
handling here, because we're just falling back to the old behavior if an error 
occurs while accessing the `JobResultStore`. But this reasoning is wrong. Not 
marking a job as dirty in the `JobResultStore` will leave us in an inconsistent 
state. I guess, we should fail fatally (similarly to when a job cannot be 
recovered) in such a case to avoid rerunning the job. I'm gonna utilize the 
`fatalErrorHandler`




-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


Reply via email to