Wrong owner of HDFS output folder

2015-10-26 Thread Flavio Pompermaier
Hi to all, when I run my job within my hadoop cluster (both from command line and from webapp) the output of my job (HDFS) works fine until I set the write parallelism to 1 (the output file is created with the user running the job). If I leave the default parallelism (>1) the job fails because it

Re: Wrong owner of HDFS output folder

2015-10-26 Thread Maximilian Michels
The problem is that non-root processes may not be able to read root-owned files/folders. Therefore, we cannot really check as a non-root users whether root-owned clusters have been started. It's better not to run Flink with root permissions. You're welcome. Cheers, Max On Mon, Oct 26, 2015 at

Re: Wrong owner of HDFS output folder

2015-10-26 Thread Maximilian Michels
Hi Flavio, Are you runing your Flink cluster with root permissions? The directory to hold the output splits are created by the JobManager. So if you run then JobManager with root permissions, it will create a folder owned by root. If the task managers are not run with root permissions, this could