Hi to all,
when I run my job within my hadoop cluster (both from command line and from
webapp) the output of my job (HDFS) works fine until I set the write
parallelism to 1 (the output file is created with the user running the
job). If I leave the default parallelism (>1) the job fails because it
The problem is that non-root processes may not be able to read root-owned
files/folders. Therefore, we cannot really check as a non-root users
whether root-owned clusters have been started. It's better not to run Flink
with root permissions.
You're welcome.
Cheers,
Max
On Mon, Oct 26, 2015 at
Hi Flavio,
Are you runing your Flink cluster with root permissions? The directory to
hold the output splits are created by the JobManager. So if you run then
JobManager with root permissions, it will create a folder owned by root. If
the task managers are not run with root permissions, this could