Hi, FYI, I was able to compile and run both the SimpleIngestClient and the SimpleReadClient apps from the minicluster samples.
I think the problem that I asked about was because to support one of the dependencies, I had chosen one of the "fatjar" JARs. I went back and found the class I was looking for, in a different JAR and then I was able to run the apps. Thanks! Jim -------------------------------------------- On Sat, 5/20/17, Josh Elser <els...@apache.org> wrote: Subject: Re: Trying to get SimpleIngestClient example working with Accumulo To: user@accumulo.apache.org Date: Saturday, May 20, 2017, 3:07 PM Have you considered using a tool like Apache Maven to handle dependency management for you? It's pretty uncommon these days to manually compile your code. Another option is using ant+ivy. If you are repackaging JARs, you may have run into a case where the jars are sealed (we've started doing this in Accumulo a while back). The error sounds like it's telling you the checksum for the Manifest file in a JAR is invalid. This means that either the checksum was replaced or you altered the Manifest file itself (and it now has a new checksum). This one is in your camp to figure out.. o haya wrote: > Hi, > > I am just starting to work with Accumulo, and I originally started testing with the minicluster, but now, I have stood up a standalone (single server) Accumulo instance. > > I think that I got that working, but I would like to try the SimpleIngestClient example from the minicluster with my new Accumulo instance. > > So, I've been trying to compile the code for SimpleIngestClient that was in the minicluster. There were a lot of JARs that I had to add to the classpath, but i finally got the application to compile, but when I tried to run it, after adding some more JARs to the classpath, I think I am stuck. > > The error I am getting is shown below. > > > java -classpath /apps/accumulo-1.8.1/lib/accumulo-core.jar:/apps/hadoop-2.7.3/share/hadoop/common/hadoop-common-2.7.3.jar:/apps/hadoop-2.7.3/share/hadoop/common/lib/commons-configuration-1.6.jar:/apps/hadoop-2.7.3/share/hadoop/common/lib/slf4j-api-1.7.10.jar:/apps/hadoop-2.7.3/share/hadoop/common/lib/commons-lang-2.6.jar:/apps/hadoop-2.7.3/share/hadoop/hdfs/lib/commons-logging-1.1.3.jar:/apps/hadoop-2.7.3/share/hadoop/common/lib/commons-collections-3.2.2.jar:/apps/hadoop-2.7.3/share/hadoop/common/lib/guava-11.0.2.jar:/apps/zookeeper-3.4.10/contrib/fatjar/zookeeper-3.4.10-fatjar.jar:/apps/accumulo-1.8.1/lib/slf4j-api.jar:/apps/accumulo-1.8.1/lib/slf4j-log4j12.jar:./ SimpleIngestClient_accumulo1 > > > ++++++++++++++++++++++ In main: About to do authentication... > ++++++++++++++++++++++ In main: About to instantiate new ZooKeeper instance... > SLF4J: Class path contains multiple SLF4J bindings. > SLF4J: Found binding in [jar:file:/apps/zookeeper-3.4.10/contrib/fatjar/zookeeper-3.4.10-fatjar.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: Found binding in [jar:file:/apps/accumulo-1.8.1/lib/slf4j-log4j12.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation. > Failed to instantiate SLF4J LoggerFactory > Reported exception: > java.lang.SecurityException: Invalid signature file digest for Manifest main attributes > at sun.security.util.SignatureFileVerifier.processImpl(SignatureFileVerifier.java:287) > at sun.security.util.SignatureFileVerifier.process(SignatureFileVerifier.java:240) > at java.util.jar.JarVerifier.processEntry(JarVerifier.java:274) > at java.util.jar.JarVerifier.update(JarVerifier.java:228) > at java.util.jar.JarFile.initializeVerifier(JarFile.java:348) > at java.util.jar.JarFile.getInputStream(JarFile.java:415) > at sun.misc.URLClassPath$JarLoader$2.getInputStream(URLClassPath.java:775) > at sun.misc.Resource.cachedInputStream(Resource.java:77) > at sun.misc.Resource.getByteBuffer(Resource.java:160) > at java.net.URLClassLoader.defineClass(URLClassLoader.java:436) > at java.net.URLClassLoader.access$100(URLClassLoader.java:71) > at java.net.URLClassLoader$1.run(URLClassLoader.java:361) > > Does anyone know what might be causing that exception, and also how I can get around it? > > Thanks, > Jim