[ 
https://issues.apache.org/jira/browse/ACCUMULO-2228?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13878768#comment-13878768
 ] 

Bill Havanki commented on ACCUMULO-2228:
----------------------------------------

I'm running on Hadoop 2.0.0-cdh4.5.0. HDFS-4404 was committed after 2.0.0, but 
the CDH release happens to include a backport for it \[1]. I think in my 
particular case that what I'm seeing was actually fixed under HADOOP-9792 - see 
my last comment on ACCUMULO-2227 for details - and that was also fixed after 
2.0.0, but is wasn't backported into what I'm using.

So your hypothesis that I'm using an outdated Hadoop is correct! :) But now I'm 
not sure whether it's worth working around it.

\[1] http://archive.cloudera.com/cdh4/cdh/4/hadoop-2.0.0-cdh4.5.0.CHANGES.txt

> Security randomwalk fails when namenode dies before bulk import
> ---------------------------------------------------------------
>
>                 Key: ACCUMULO-2228
>                 URL: https://issues.apache.org/jira/browse/ACCUMULO-2228
>             Project: Accumulo
>          Issue Type: Bug
>          Components: test
>    Affects Versions: 1.4.4
>            Reporter: Bill Havanki
>              Labels: ha, randomwalk, test
>
> Running Security randomwalk under HDFS HA, if the active namenode is killed:
> {noformat}
> 20 13:55:58,287 [retry.RetryInvocationHandler] WARN : Exception while 
> invoking class org.apache.hado
> op.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.create. Not retrying 
> because the invoked metho
> d is not idempotent, and unable to determine whether it was invoked
> java.io.IOException: Failed on local exception: java.io.IOException: Response 
> is null.; Host Details
>  : local host is: "slave.domain.com/10.20.200.113"; destination host is: 
> "namenode.domain.com":8020;
> ...
>         at 
> org.apache.accumulo.core.file.rfile.RFileOperations.openWriter(RFileOperations.java:124)
>         at 
> org.apache.accumulo.core.file.DispatchingFileFactory.openWriter(FileOperations.java:80)
>         at 
> org.apache.accumulo.server.test.randomwalk.security.TableOp.visit(TableOp.java:177)
>         at 
> org.apache.accumulo.server.test.randomwalk.Module.visit(Module.java:251)
> ...
> Caused by: java.io.IOException: Response is null.
>         at 
> org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.java:952)
>         at org.apache.hadoop.ipc.Client$Connection.run(Client.java:847)
> {noformat}
> This arises from an attempt to write an rfile.The test should be resilient 
> here (and for other file system operations) so that the test can continue 
> once failover has completed.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)

Reply via email to