[
https://issues.apache.org/jira/browse/HADOOP-15915?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17082066#comment-17082066
]
Vasii Cosmin Radu commented on HADOOP-15915:
--------------------------------------------
I've encountered the same issue, Flink 1.10, EC2 deployment. Any ideas of how
to fix this? The problem appears on Job Managers as soon as I submit a job.
There is enough disk space and the Job Managers are not actually doing anything
to need such much disk space, they are uploading job graph blobs.
> Report problems w/ local S3A buffer directory meaningfully
> ----------------------------------------------------------
>
> Key: HADOOP-15915
> URL: https://issues.apache.org/jira/browse/HADOOP-15915
> Project: Hadoop Common
> Issue Type: Sub-task
> Components: fs/s3
> Affects Versions: 3.1.1
> Reporter: Steve Loughran
> Priority: Major
>
> When there's a problem working with the temp directory used for block output
> and the staging committers the actual path (and indeed config option) aren't
> printed.
> Improvements: tell the user which directory isn't writeable
--
This message was sent by Atlassian Jira
(v8.3.4#803005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]