[ https://issues.apache.org/jira/browse/PIG-555?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Olga Natkovich resolved PIG-555. -------------------------------- Resolution: Duplicate > ORDER by requires write access to /tmp > -------------------------------------- > > Key: PIG-555 > URL: https://issues.apache.org/jira/browse/PIG-555 > Project: Pig > Issue Type: Bug > Components: grunt > Affects Versions: 0.2.0 > Reporter: Ian Holsman > Priority: Minor > > when doing an order by in pig, it relies on you having write access to /tmp/ > in hdfs. (and i think for it to be present in the first place). > this comes down to FileLocalizer creating the default > relativeRoot = pigContext.getDfs().asContainer("/tmp/temp" + r.nextInt()) > i'm not sure if this is due to someone on our side deleting the /tmp dir > accidently, so apolagies for the spam if it is. > 2008-12-03 22:36:41,716 [main] ERROR org.apache.pig.tools.grunt.Grunt - You > don't have permission to perform the operation. Error from the server: Unable > to store for alias: 7 > [org.apache.hadoop.fs.permission.AccessControlException: Permission denied: > user=jholsma, access=WRITE, inode="":hadoop:supergroup:rwxr-xr-x] > java.io.IOException: Unable to store for alias: 7 > [org.apache.hadoop.fs.permission.AccessControlException: Permission denied: > user=jholsma, access=WRITE, inode="":hadoop:supergroup:rwxr-xr-x] > at > org.apache.pig.backend.hadoop.executionengine.HExecutionEngine.execute(HExecutionEngine.java:255) > at > org.apache.pig.PigServer.executeCompiledLogicalPlan(PigServer.java:647) > at org.apache.pig.PigServer.execute(PigServer.java:638) > at org.apache.pig.PigServer.registerQuery(PigServer.java:278) > at > org.apache.pig.tools.grunt.GruntParser.processPig(GruntParser.java:439) > at > org.apache.pig.tools.pigscript.parser.PigScriptParser.parse(PigScriptParser.java:249) > at > org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:84) > at org.apache.pig.tools.grunt.Grunt.exec(Grunt.java:64) > at org.apache.pig.Main.main(Main.java:242) > Caused by: org.apache.pig.backend.executionengine.ExecException: > org.apache.hadoop.fs.permission.AccessControlException: Permission denied: > user=jholsma, access=WRITE, inode="":hadoop:supergroup:rwxr-xr-x > ... 9 more > Caused by: org.apache.pig.impl.plan.VisitorException: > org.apache.hadoop.fs.permission.AccessControlException: Permission denied: > user=jholsma, access=WRITE, inode="":hadoop:supergroup:rwxr-xr-x > at > org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MRCompiler.visitSort(MRCompiler.java:858) > at > org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POSort.visit(POSort.java:304) > at > org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MRCompiler.compile(MRCompiler.java:266) > at > org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MRCompiler.compile(MRCompiler.java:251) > at > org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MRCompiler.compile(MRCompiler.java:193) > at > org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher.compile(MapReduceLauncher.java:134) > at > org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher.launchPig(MapReduceLauncher.java:63) > at > org.apache.pig.backend.hadoop.executionengine.HExecutionEngine.execute(HExecutionEngine.java:245) > ... 8 more > Caused by: org.apache.hadoop.fs.permission.AccessControlException: > org.apache.hadoop.fs.permission.AccessControlException: Permission denied: > user=jholsma, access=WRITE, inode="":hadoop:supergroup:rwxr-xr-x > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native > Method) > at > sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39) > at > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27) > at java.lang.reflect.Constructor.newInstance(Constructor.java:513) > at > org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:90) > at > org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:52) > at org.apache.hadoop.dfs.DFSClient.mkdirs(DFSClient.java:704) > at > org.apache.hadoop.dfs.DistributedFileSystem.mkdirs(DistributedFileSystem.java:236) > at org.apache.hadoop.fs.FileSystem.mkdirs(FileSystem.java:1116) > at > org.apache.pig.backend.hadoop.datastorage.HDirectory.create(HDirectory.java:64) > at > org.apache.pig.backend.hadoop.datastorage.HPath.create(HPath.java:155) > at > org.apache.pig.impl.io.FileLocalizer.getTemporaryPath(FileLocalizer.java:391) > at > org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MRCompiler.getTempFileSpec(MRCompiler.java:479) > at > org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MRCompiler.visitSort(MRCompiler.java:846) > ... 15 more > Caused by: org.apache.hadoop.ipc.RemoteException: > org.apache.hadoop.fs.permission.AccessControlException: Permission denied: > user=jholsma, access=WRITE, inode="":hadoop:supergroup:rwxr-xr-x > at > org.apache.hadoop.dfs.PermissionChecker.check(PermissionChecker.java:175) > at > org.apache.hadoop.dfs.PermissionChecker.check(PermissionChecker.java:156) > at > org.apache.hadoop.dfs.PermissionChecker.checkPermission(PermissionChecker.java:104) > at > org.apache.hadoop.dfs.FSNamesystem.checkPermission(FSNamesystem.java:4228) > at > org.apache.hadoop.dfs.FSNamesystem.checkAncestorAccess(FSNamesystem.java:4198) > at > org.apache.hadoop.dfs.FSNamesystem.mkdirsInternal(FSNamesystem.java:1595) > at org.apache.hadoop.dfs.FSNamesystem.mkdirs(FSNamesystem.java:1564) > at org.apache.hadoop.dfs.NameNode.mkdirs(NameNode.java:450) > at sun.reflect.GeneratedMethodAccessor40.invoke(Unknown Source) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) > at java.lang.reflect.Method.invoke(Method.java:597) > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:452) > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:888) > at org.apache.hadoop.ipc.Client.call(Client.java:715) > at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:216) > at org.apache.hadoop.dfs.$Proxy0.mkdirs(Unknown Source) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) > at java.lang.reflect.Method.invoke(Method.java:597) > at > org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82) > at > org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59) > at org.apache.hadoop.dfs.$Proxy0.mkdirs(Unknown Source) > at org.apache.hadoop.dfs.DFSClient.mkdirs(DFSClient.java:702) > ... 22 more > shell returned 1 -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.