Strange, i've not seen permission issues with the _SUCCESS file! Anyway, 
deleting or not creating the file fixes it. I've committed some fixes for 
readdb -stats and readlinkdb but not for updatedb.

Its related to:
https://issues.apache.org/jira/browse/NUTCH-1029
Can you open an issue?

> Please see following exception. It looks like it is caused by the
> _SUCCESS file created by Hadoop when trying to open map files in a
> permission checked HDFS on CDH3u1. After deleting the _SUCCESS it works.
> What is a better solution for this problem?
> 
> Exception in thread "main"
> org.apache.hadoop.security.AccessControlException:
> org.apache.hadoop.security.AccessControlException: Permission denied:
> user=ferdy, access=EXECUTE,
> inode="/user/ferdy/root10/crawldb/current/_SUCCESS":ferdy:ferdy:-rw-r--r--
>      at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
> Method)
>      at
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAcce
> ssorImpl.java:39) at
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstru
> ctorAccessorImpl.java:27) at
> java.lang.reflect.Constructor.newInstance(Constructor.java:513) at
> org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.
> java:95) at
> org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException
> .java:57) at
> org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:699) at
> org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileS
> ystem.java:542) at
> org.apache.hadoop.fs.FileSystem.getLength(FileSystem.java:776) at
> org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1424)
>      at
> org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1419)
>      at
> org.apache.hadoop.io.MapFile$Reader.createDataFileReader(MapFile.java:302)
>      at org.apache.hadoop.io.MapFile$Reader.open(MapFile.java:284)
>      at org.apache.hadoop.io.MapFile$Reader.<init>(MapFile.java:273)
>      at org.apache.hadoop.io.MapFile$Reader.<init>(MapFile.java:260)
>      at org.apache.hadoop.io.MapFile$Reader.<init>(MapFile.java:253)
>      at
> org.apache.hadoop.mapred.MapFileOutputFormat.getReaders(MapFileOutputFormat
> .java:93) at
> org.apache.nutch.crawl.CrawlDbReader.openReaders(CrawlDbReader.java:81)
>      at org.apache.nutch.crawl.CrawlDbReader.get(CrawlDbReader.java:379)
>      at
> org.apache.nutch.crawl.CrawlDbReader.readUrl(CrawlDbReader.java:386) at
> org.apache.nutch.crawl.CrawlDbReader.main(CrawlDbReader.java:511) at
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:3
> 9) at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImp
> l.java:25) at java.lang.reflect.Method.invoke(Method.java:597)
>      at org.apache.hadoop.util.RunJar.main(RunJar.java:186)
> Caused by: org.apache.hadoop.ipc.RemoteException:
> org.apache.hadoop.security.AccessControlException: Permission denied:
> user=ferdy, access=EXECUTE,
> inode="/user/ferdy/root10/crawldb/current/_SUCCESS":ferdy:ferdy:-rw-r--r--
>      at
> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissi
> onChecker.java:203)

Reply via email to