[
https://issues.apache.org/jira/browse/SPARK-10528?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14996667#comment-14996667
]
Michel Lemay commented on SPARK-10528:
--------------------------------------
Looks like the problem is still there with precompiled binaries 1.5.1
I created and chmod using winutils.exe as explained earlier:
from /tmp:
winutils.exe ls hive
drwxrwxrwx
Furthermore, hadoop LocalFileSystem does not seems to be able to change
permissions as shown here:
`import org.apache.hadoop.fs._
val path = new Path("file:/tmp/hive")
val lfs = FileSystem.get(path.toUri(), sc.hadoopConfiguration)
lfs.getFileStatus(path).getPermission()`
Shows: res0: org.apache.hadoop.fs.permission.FsPermission = rw-rw-rw-
`lfs.setPermission(path, new
org.apache.hadoop.fs.permission.FsPermission(0777.toShort))
lfs.getFileStatus(new Path("file:/tmp/hive")).getPermission()`
Still shows rw-rw-rw-
> spark-shell throws java.lang.RuntimeException: The root scratch dir:
> /tmp/hive on HDFS should be writable.
> ----------------------------------------------------------------------------------------------------------
>
> Key: SPARK-10528
> URL: https://issues.apache.org/jira/browse/SPARK-10528
> Project: Spark
> Issue Type: Bug
> Components: Spark Shell
> Affects Versions: 1.5.0
> Environment: Windows 7 x64
> Reporter: Aliaksei Belablotski
> Priority: Minor
>
> Starting spark-shell throws
> java.lang.RuntimeException: java.lang.RuntimeException: The root scratch dir:
> /tmp/hive on HDFS should be writable. Current permissions are: rw-rw-rw-
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]