Apache mailing lists strip attachments. Please host the files somewhere and
provide a link to them.

On Dec 4, 2016 20:54, "Takashi Sasaki" <tsasaki...@gmail.com> wrote:

> Hello,
>
> I'm sorry to take a few wrong infomation at first post.
>
> I asked the project members again about the problem.
> Master server did not throw AccessControlException.
>
> Actually, TabletServer threw AccessControlException.
> And, the stack trace was missing words, wrong path.
>
> Correct full stack trace was line 52 in attached file "tserver.log".
> And I attache "master.log" for your reference.
>
> Unfortunately, I could not still get debug log.
>
> Thank you for your support,
> Takashi
>
>
> 2016-12-04 18:33 GMT+09:00 Takashi Sasaki <tsasaki...@gmail.com>:
> > Hello, Christopher
> >
> >>The stack trace doesn't include anything from Accumulo, so it's not
> clear where in the Accumulo code this occurred. Do you have the full stack
> trace?
> > Yes, I understand the stack trace isn't including from Accumulo.
> > I don't have full stack trace now, but I will try to find it.
> >
> > In additon, I use Accumulo on AWS EMR cluster for Enterprise
> > Production System, so log level isn't debug, becase of disk capacity
> > problem.
> > I will try to reproduce with debug log level.
> >
> > Thank you for your reply,
> > Takashi
> >
> > 2016-12-04 18:00 GMT+09:00 Christopher <ctubb...@apache.org>:
> >> The stack trace doesn't include anything from Accumulo, so it's not
> clear
> >> where in the Accumulo code this occurred. Do you have the full stack
> trace?
> >>
> >> In particular, it's not clear to me that there should be a directory
> called
> >> failed/da at that location, nor is it clear why Accumulo would be
> trying to
> >> check for the execute permission on it, unless it's trying to recurse
> into a
> >> directory. There is one part of the code where, if the directory exists
> when
> >> log recovery begins, it may try to do a recursive delete, but I can't
> see
> >> how this location would have been created by Accumulo. If that is the
> case,
> >> then it should be safe to manually delete this directory and its
> contents.
> >> The failed marker should be a regular file, though, and should not be a
> >> directory with another directory called "da" in it. So, I can't see how
> this
> >> was even created, unless by an older version or another program.
> >>
> >> The only way I can see this occurring is if you recently did an upgrade,
> >> while Accumulo had not yet finished outstanding log recoveries from a
> >> previous shutdown, AND the previous version did something different than
> >> 1.7.2. If that was the case, then perhaps the older version could have
> >> created this problematic directory. It seems unlikely, though... because
> >> directories are usually not created without the execute bit... and the
> error
> >> message looks like a directory missing that bit.
> >>
> >> It's hard to know more without seeing the full stack trace with the
> relevant
> >> accumulo methods included. It might also help to see the master debug
> logs
> >> leading up to the error.
> >>
> >> On Sun, Dec 4, 2016 at 2:35 AM Takashi Sasaki <tsasaki...@gmail.com>
> wrote:
> >>>
> >>> I use Accumulo-1.7.2 with Haddop2.7.2 and ZooKeeper 3.4.8
> >>>
> >>> Master server suddenly throw AccessControlException.
> >>>
> >>> java.io.IOException:
> >>> org.apache.hadoop.security.AccessControlException: Permission denied:
> >>> user=accumulo, access=EXECUTE,
> >>>
> >>> inode="/accumulo/recovery/603194f3-dd41-44ed-8ad6-
> 90d408149952/failed/da":accumulo:accumulo:-rw-r--r--
> >>>  at
> >>> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.
> check(FSPermissionChecker.java:319)
> >>>  at
> >>> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.
> checkTraverse(FSPermissionChecker.java:259)
> >>>  at
> >>> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.
> checkPermission(FSPermissionChecker.java:205)
> >>>  at
> >>> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.
> checkPermission(FSPermissionChecker.java:190)
> >>>  at
> >>> org.apache.hadoop.hdfs.server.namenode.FSDirectory.
> checkPermission(FSDirectory.java:1720)
> >>>  at org.apache.hadoop.hdfs.server.namenode.FSDirSt
> >>>  at AndListingOp.getFileInfo(FSDirSt
> >>>  at AndListingOp.java:108)
> >>>  at
> >>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.
> getFileInfo(FSNamesystem.java:3855)
> >>>  at
> >>> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.
> getFileInfo(NameNodeRpcServer.java:1011)
> >>>  at
> >>> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSi
> deTransl
> >>>  at orPB.getFileInfo(ClientNamenodeProtocolServerSideTransl
> >>>  at orPB.java:843)
> >>>  at
> >>> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$
> ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.
> java)
> >>>  at
> >>> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$
> ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616)
> >>>  at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969)
> >>>  at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049)
> >>>  at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2045)
> >>>  at java.security.AccessController.doPrivileged(N
> >>>  at ive Method)
> >>>  at javax.security.auth.Subject.doAs(Subject.java:422)
> >>>  at org.apache.hadoop.security.UserGroupInform
> >>>  at ion.doAs(UserGroupInform
> >>>  at ion.java:1657)
> >>>  at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2043)
> >>>
> >>>
> >>> How can I solve this Exception?
> >>>
> >>>
> >>> Thank you,
> >>> Takashi.
>

Reply via email to