Is that datanode configured correctly? I wonder why it's excluded.

On Wed, Dec 15, 2021 at 9:45 AM Vincent Russell <vincent.russ...@gmail.com>
wrote:

> Thank you Christopher,
>
> I was able to determine that the ssl settings in core-site.xml are being
> picked up and used.   In fact when accumulo init is run, accumulo is able
> to create the /accumulo directory in HDFS.    What is weird is that when
> the FileSKVWriter that is used when createMetadataFile is called it
> throws an exception when close is called.
>
>
> I get:
>
> org.apache.hadoop.ipc.RemoteException(java.io.IOException): File
> /accumulo/tables/!0/table_info/0_1.rf could only be written to the 0 of the
> 1 minReplication nodes.  There are 1 datanode(s) running and 1 node(s) are
> excluded in this operation.
>         at
>
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:1720)
>         at
>
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3389)
>         at
>
> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:683)
>         at
>
> org.apache.hadoop.hdfs.server.namenode.AuthorizationProviderProxyClientProtocol.addBlock(AuthorizationProviderProxyClientProtocol.java:214)
>         at
>
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:495)
>         at
>
> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
>         at
>
> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:617)
>         at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1073)
>         at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2217)
>         at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2213)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:422)
>         at
>
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1917)
>         at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2211)
>
> I don't get this error when I disable ssl on hadoop.
>
> Any insight would be greatly appreciated.
>
> Thanks,
>
> On Tue, Dec 14, 2021 at 2:23 PM Vincent Russell <vincent.russ...@gmail.com
> >
> wrote:
>
> > Thanks Chris.
> >
> > Yes I do get an error (I can't remember now because it's on a separate
> > computer) during the init and I get a MagicNumber exception on the
> datanode
> > during this process which says something like maybe encryption isn't
> turned
> > on.
> >
> > but let me make sure that the core-default.xml and core-site.xml are on
> > the classpath.  They may not be.
> >
> > Thanks again.
> >
> > On Tue, Dec 14, 2021 at 2:13 PM Christopher <ctubb...@apache.org> wrote:
> >
> >> I have not personally tested HDFS configured for SSL/TLS, but `new
> >> Configuration()` will load the core-default.xml and core-site.xml
> >> files it finds on the class path. So, it looks like it should work.
> >> Have you tried it? Did you get an error?
> >>
> >>
> >> On Tue, Dec 14, 2021 at 1:54 PM Vincent Russell
> >> <vincent.russ...@gmail.com> wrote:
> >> >
> >> > Thank you Mike,
> >> >
> >> > but it appears that accumulo uses those settings to connect accumulo,
> >> but
> >> > not to connect to hdfs.
> >> >
> >> > For instance the VolumeManagementImpl just does this:
> >> >
> >> > VolumeConfiguration.create(new Path(volumeUriOrDir), hadoopConf));
> >> >
> >> > where the hadoopConf is just instantiated in the Initialize class:
> >> >
> >> > Configuration hadoopConfig = new Configuration();
> >> > VolumeManager fs = VolumeManagerImpl.get(siteConfig, hadoopConfig);
> >> >
> >> > Thanks,
> >> > Vincent
> >> >
> >> > On Tue, Dec 14, 2021 at 12:18 PM Mike Miller <mmil...@apache.org>
> >> wrote:
> >> >
> >> > > Checkout the accumulo client properties that start with the "ssl"
> >> prefix.
> >> > >
> https://accumulo.apache.org/docs/2.x/configuration/client-properties
> >> > > This blog post from a few years ago may help:
> >> > >
> >> > >
> >>
> https://accumulo.apache.org/blog/2014/09/02/generating-keystores-for-configuring-accumulo-with-ssl.html
> >> > >
> >> > > On Tue, Dec 14, 2021 at 9:58 AM Vincent Russell <
> >> vincent.russ...@gmail.com
> >> > > >
> >> > > wrote:
> >> > >
> >> > > > Hello,
> >> > > >
> >> > > > I am trying to init a test accumulo instance with an hdfs running
> >> with
> >> > > > SSL.    Is this possible?  I am looking at the code and it doesn't
> >> look
> >> > > > like this is possible.
> >> > > >
> >> > > > The Initialize class just instantiates a Hadoop config and passes
> >> that
> >> > > into
> >> > > > the VolumeManager without sending over any hadoop configs from the
> >> > > core.xml
> >> > > > file.
> >> > > >
> >> > > > Am I missing something?
> >> > > >
> >> > > > Thanks in advance for your help,
> >> > > > Vincent
> >> > > >
> >> > >
> >>
> >
>

Reply via email to