I don't have any good answer for this. Probably it's rather something to
look for online tutorials to find help with setting up HBase client ...

One thought in my mind though is that HBase client seems very
version-specific. Maybe it's a better route for MetaModel to go through the
REST api than the Java API, although I guess it may come at a severe
performance penalty?


2014-03-19 10:50 GMT+01:00 Tomasz Guziałek <
[email protected]>:

> Hello Kasper
>
> I have also tried to set up HBase (version 0.94.17) on Windows and
> experienced the same issue: some Zookeper classes could not be found on the
> classpath.
>
> But coming back to the Cloudera VM. It seems that it is not a problem with
> the test but the hbase-module itself as I wrote this snippet to read from a
> table that I know it exists in the HBase:
>
> public class HBaseDemo {
>
>         public static void main(String[] args) {
>                 HBaseConfiguration configuration = new
> HBaseConfiguration();
>                 HBaseDataContext dataContext = new
> HBaseDataContext(configuration);
>                 DataSet dataset =
> dataContext.query().from("countrycodes").selectAll().execute();
>
>                 for (Row row : dataset) {
>                         System.out.println(row.toString());
>                 }
>         }
> }
>
> The stacktrace:
>
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in
> [jar:file:/home/cloudera/.m2/repository/org/slf4j/slf4j-nop/1.7.4/slf4j-nop-1.7.4.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in
> [jar:file:/home/cloudera/.m2/repository/org/slf4j/slf4j-log4j12/1.7.4/slf4j-log4j12-1.7.4.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
> explanation.
> SLF4J: Actual binding is of type [org.slf4j.helpers.NOPLoggerFactory]
> Exception in thread "main" org.apache.metamodel.MetaModelException:
> org.apache.hadoop.hbase.exceptions.MasterNotRunningException:
> com.google.protobuf.ServiceException: java.io.IOException: Broken pipe
>         at
> org.apache.metamodel.hbase.HBaseDataContext.getMainSchema(HBaseDataContext.java:131)
>         at
> org.apache.metamodel.QueryPostprocessDataContext.getMainSchemaInternal(QueryPostprocessDataContext.java:498)
>         at
> org.apache.metamodel.QueryPostprocessDataContext.getSchemaByNameInternal(QueryPostprocessDataContext.java:369)
>         at
> org.apache.metamodel.AbstractDataContext.getSchemaByName(AbstractDataContext.java:199)
>         at
> org.apache.metamodel.AbstractDataContext.getDefaultSchema(AbstractDataContext.java:118)
>         at
> org.apache.metamodel.AbstractDataContext.getSchemaByToken(AbstractDataContext.java:503)
>         at
> org.apache.metamodel.AbstractDataContext.getTableByQualifiedLabel(AbstractDataContext.java:389)
>         at
> org.apache.metamodel.query.builder.InitFromBuilderImpl.from(InitFromBuilderImpl.java:78)
>         at info.guzialek.hbase.HBaseDemo.main(HBaseDemo.java:13)
> Caused by: org.apache.hadoop.hbase.exceptions.MasterNotRunningException:
> com.google.protobuf.ServiceException: java.io.IOException: Broken pipe
>         at
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation$StubMaker.makeStub(HConnectionManager.java:1409)
>         at
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation$MasterMonitorServiceStubMaker.makeStub(HConnectionManager.java:1435)
>         at
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.getKeepAliveMasterMonitorService(HConnectionManager.java:1884)
>         at
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.listTables(HConnectionManager.java:2645)
>         at
> org.apache.hadoop.hbase.client.HBaseAdmin.listTables(HBaseAdmin.java:279)
>         at
> org.apache.metamodel.hbase.HBaseDataContext.getMainSchema(HBaseDataContext.java:117)
>         ... 8 more
> Caused by: com.google.protobuf.ServiceException: java.io.IOException:
> Broken pipe
>         at
> org.apache.hadoop.hbase.ipc.RpcClient.callBlockingMethod(RpcClient.java:1556)
>         at
> org.apache.hadoop.hbase.ipc.RpcClient$BlockingRpcChannelImplementation.callBlockingMethod(RpcClient.java:1597)
>         at
> org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos$MasterMonitorService$BlockingStub.isMasterRunning(MasterMonitorProtos.java:3021)
>         at
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation$MasterMonitorServiceStubMaker.isMasterRunning(HConnectionManager.java:1446)
>         at
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation$StubMaker.makeStubNoRetries(HConnectionManager.java:1354)
>         at
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation$StubMaker.makeStub(HConnectionManager.java:1380)
>         ... 13 more
> Caused by: java.io.IOException: Broken pipe
>         at sun.nio.ch.FileDispatcher.write0(Native Method)
>         at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:29)
>         at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:69)
>         at sun.nio.ch.IOUtil.write(IOUtil.java:40)
>         at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:336)
>         at
> org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:55)
>         at
> org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:142)
>         at
> org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:146)
>         at
> org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:107)
>         at
> java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:65)
>         at
> java.io.BufferedOutputStream.flush(BufferedOutputStream.java:123)
>         at java.io.DataOutputStream.flush(DataOutputStream.java:106)
>         at
> org.apache.hadoop.hbase.ipc.RpcClient$Connection.writeConnectionHeader(RpcClient.java:906)
>         at
> org.apache.hadoop.hbase.ipc.RpcClient$Connection.setupIOstreams(RpcClient.java:856)
>         at
> org.apache.hadoop.hbase.ipc.RpcClient.getConnection(RpcClient.java:1422)
>         at org.apache.hadoop.hbase.ipc.RpcClient.call(RpcClient.java:1314)
>         at
> org.apache.hadoop.hbase.ipc.RpcClient.callBlockingMethod(RpcClient.java:1540)
>         ... 18 more
>
> My MapReduce code is able to connect to that instance of HBase from the
> same host (localhost). When I point my browser to http://127.0.0.1:60010I can 
> see the web UI of HBase and I verified that the ZooKeeper default
> port has not been changed (it is 2181 as in the HBaseConfiguration).
>
> Tomasz Guzialek
> Software Engineer
>
> Vesterbrogade 74, 3.
> 1620 København
> Danmark
> Tlf. +45 2613 4307
> [email protected]
>
> Human Inference B.V. - Chamber of Commerce NL - 09106800
>
>
> The information contained in this communication is confidential and is
> intended solely for the person and/or entity addressed above. If you are
> not an intended recipient you are hereby notified that the content of this
> message may not be disclosed, copied or distributed to any other party than
> the addressee. Human Inference does not guarantee that this communication
> is free of viruses or that other parties than the intended recipient will
> not intercept this communication. Human Inference assumes no liability for
> the proper and complete transmission of the information contained in this
> communication or any delay in the receipt of this communication by the
> addressee.
>  Please consider the environment before printing this email.
>
> -----Original Message-----
> From: Kasper Sørensen [mailto:[email protected]]
> Sent: Tuesday, March 18, 2014 19:19
> To: [email protected]
> Subject: Re: [MetaModel hbase-module] HBaseDataContextTest fail in
> Cloudera VM.
>
> Hmm tried some experiments but ultimately I am having trouble even getting
> HBase to work on my Windows box. My "old" hbase installation (version
> 0.95.1-hadoop1) on windows does work, and the unittests pass etc.
> But newer versions I cannot seem to run on Windows (zookeeper issues I
> believe), so I can't reproduce what you're experiencing.
>
>
>
> 2014-03-18 16:09 GMT+01:00 Kasper Sørensen <[email protected]
> >:
>
> > Hi Tomasz
> >
> > It might be that it's just the unittest that's a bit stupid. Or else
> > it must be a HBase configuration thing. I looked at the source code of
> > HBase where the exception is thrown. It is looking for a "location" of
> > the table name. I guess you might want to try to skip the part of the
> > unittest that looks for the table, and instead simply create it (it
> > will not initially exist)...
> >
> > I am doing a reinstall of my HBase here to double verify, but this is
> > my guess.
> >
> > Kasper
> >
> >
> > 2014-03-18 14:52 GMT+01:00 Tomasz Guziałek <
> > [email protected]>:
> >
> >   Hello all
> >>
> >>
> >>
> >> My name is Tomasz and I am working on my Master thesis in
> >> collaboration with Human Inference. Long story short, the thesis
> >> involves creating a Hadoop version of DataCleaner
> >> (http://datacleaner.org), so I am working with HBase too. Within the
> >> project, I expect to make some contributions to the HBase module of
> MetaModel.
> >>
> >>
> >>
> >> However, just at the beginning I encountered problems with the build
> >> of the module.  I use Cloudera VM (
> >> http://www.cloudera.com/content/cloudera-content/cloudera-docs/DemoVM
> >> s/Cloudera-QuickStart-VM/cloudera_quickstart_vm.html)
> >> as a starting point to make sure my HBase cluster is correct (the
> >> HBase version reported by Cloudera Manager: 0.94.6+132). After a
> >> checkout of the MetaModel code, I cannot build it with “maven
> >> install” because of HBaseDataContextTest fail:
> >>
> >>
> >>
> >> -------------------------------------------------------
> >>
> >> T E S T S
> >>
> >> -------------------------------------------------------
> >>
> >> Running org.apache.metamodel.hbase.HBaseDataContextTest
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:zookeeper.version=3.4.5-1392090, built on 09/30/2012
> >> 17:52 GMT
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:host.name =localhost.localdomain
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:java.version=1.6.0_32
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:java.vendor=Sun Microsystems Inc.
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:java.home=/usr/java/jdk1.6.0_32/jre
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >>
> environment:java.class.path=/home/cloudera/git/incubator-metamodel/hbase/target/test-classes:/home/cloudera/git/incubator-metamodel/hbase/target/classes:/home/cloudera/git/incubator-metamodel/core/target/MetaModel-core-4.0.1-incubating-SNAPSHOT.jar:/home/cloudera/.m2/repository/org/slf4j/slf4j-api/1.7.4/slf4j-api-1.7.4.jar:/home/cloudera/.m2/repository/org/apache/hbase/hbase-client/0.95.1-hadoop1/hbase-client-0.95.1-hadoop1.jar:/home/cloudera/.m2/repository/org/apache/hbase/hbase-common/0.95.1-hadoop1/hbase-common-0.95.1-hadoop1.jar:/home/cloudera/.m2/repository/com/google/guava/guava/12.0.1/guava-12.0.1.jar:/home/cloudera/.m2/repository/com/google/code/findbugs/jsr305/1.3.9/jsr305-1.3.9.jar:/home/cloudera/.m2/repository/org/apache/hbase/hbase-protocol/0.95.1-hadoop1/hbase-protocol-0.95.1-hadoop1.jar:/home/cloudera/.m2/repository/com/google/protobuf/protobuf-java/2.4.1/protobuf-java-2.4.1.jar:/home/cloudera/.m2/repository/org/apache/zookeeper/zookeeper/3.4.5/zookeeper-3.4.5.jar:/home/cloudera/.m2/repository/org/cloudera/htrace/htrace/1.50/htrace-1.50.jar:/home/cloudera/.m2/repository/org/mortbay/jetty/jetty-util/6.1.26/jetty-util-6.1.26.jar:/home/cloudera/.m2/repository/io/netty/netty/3.5.9.Final/netty-3.5.9.Final.jar:/home/cloudera/.m2/repository/org/apache/hadoop/hadoop-core/1.1.2/hadoop-core-1.1.2.jar:/home/cloudera/.m2/repository/commons-cli/commons-cli/1.2/commons-cli-1.2.jar:/home/cloudera/.m2/repository/xmlenc/xmlenc/0.52/xmlenc-0.52.jar:/home/cloudera/.m2/repository/com/sun/jersey/jersey-core/1.8/jersey-core-1.8.jar:/home/cloudera/.m2/repository/com/sun/jersey/jersey-json/1.8/jersey-json-1.8.jar:/home/cloudera/.m2/repository/org/codehaus/jettison/jettison/1.1/jettison-1.1.jar:/home/cloudera/.m2/repository/stax/stax-api/1.0.1/stax-api-1.0.1.jar:/home/cloudera/.m2/repository/com/sun/xml/bind/jaxb-impl/2.2.3-1/jaxb-impl-2.2.3-1.jar:/home/cloudera/.m2/repository/javax/xml/bind/jaxb-api/2.2.2/jaxb-api-2.2.2.jar:/home/cloudera/.m2/repository/javax/xml/stream/stax-api/1.0-2/stax-api-1.0-2.jar:/home/cloudera/.m2/repository/javax/activation/activation/1.1/activation-1.1.jar:/home/cloudera/.m2/repository/org/codehaus/jackson/jackson-core-asl/1.9.12/jackson-core-asl-1.9.12.jar:/home/cloudera/.m2/repository/org/codehaus/jackson/jackson-jaxrs/1.7.1/jackson-jaxrs-1.7.1.jar:/home/cloudera/.m2/repository/org/codehaus/jackson/jackson-xc/1.7.1/jackson-xc-1.7.1.jar:/home/cloudera/.m2/repository/com/sun/jersey/jersey-server/1.8/jersey-server-1.8.jar:/home/cloudera/.m2/repository/asm/asm/3.1/asm-3.1.jar:/home/cloudera/.m2/repository/commons-io/commons-io/2.4/commons-io-2.4.jar:/home/cloudera/.m2/repository/commons-httpclient/commons-httpclient/3.0.1/commons-httpclient-3.0.1.jar:/home/cloudera/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar:/home/cloudera/.m2/repository/org/apache/commons/commons-math/2.1/commons-math-2.1.jar:/home/cloudera/.m2/repository/commons-configuration/commons-configuration/1.6/commons-configuration-1.6.jar:/home/cloudera/.m2/repository/commons-collections/commons-collections/3.2.1/commons-collections-3.2.1.jar:/home/cloudera/.m2/repository/commons-lang/commons-lang/2.6/commons-lang-2.6.jar:/home/cloudera/.m2/repository/commons-digester/commons-digester/1.8/commons-digester-1.8.jar:/home/cloudera/.m2/repository/commons-beanutils/commons-beanutils/1.7.0/commons-beanutils-1.7.0.jar:/home/cloudera/.m2/repository/commons-beanutils/commons-beanutils-core/1.8.0/commons-beanutils-core-1.8.0.jar:/home/cloudera/.m2/repository/commons-net/commons-net/1.4.1/commons-net-1.4.1.jar:/home/cloudera/.m2/repository/org/mortbay/jetty/jetty/6.1.26/jetty-6.1.26.jar:/home/cloudera/.m2/repository/org/mortbay/jetty/servlet-api/2.5-20081211/servlet-api-2.5-20081211.jar:/home/cloudera/.m2/repository/tomcat/jasper-runtime/5.5.12/jasper-runtime-5.5.12.jar:/home/cloudera/.m2/repository/tomcat/jasper-compiler/5.5.12/jasper-compiler-5.5.12.jar:/home/cloudera/.m2/repository/org/mortbay/jetty/jsp-api-2.1/6.1.14/jsp-api-2.1-6.1.14.jar:/home/cloudera/.m2/repository/org/mortbay/jetty/servlet-api-2.5/6.1.14/servlet-api-2.5-6.1.14.jar:/home/cloudera/.m2/repository/org/mortbay/jetty/jsp-2.1/6.1.14/jsp-2.1-6.1.14.jar:/home/cloudera/.m2/repository/ant/ant/1.6.5/ant-1.6.5.jar:/home/cloudera/.m2/repository/commons-el/commons-el/1.0/commons-el-1.0.jar:/home/cloudera/.m2/repository/org/codehaus/jackson/jackson-mapper-asl/1.9.12/jackson-mapper-asl-1.9.12.jar:/home/cloudera/.m2/repository/com/github/stephenc/findbugs/findbugs-annotations/1.3.9-1/findbugs-annotations-1.3.9-1.jar:/home/cloudera/.m2/repository/org/slf4j/jcl-over-slf4j/1.7.4/jcl-over-slf4j-1.7.4.jar:/home/cloudera/.m2/repository/org/slf4j/slf4j-log4j12/1.7.4/slf4j-log4j12-1.7.4.jar:/home/cloudera/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar:/home/cloudera/.m2/repository/junit/junit/4.11/junit-4.11.jar:/home/cloudera/.m2/repository/org/hamcrest/hamcrest-core/1.3/hamcrest-core-1.3.jar:
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:java.library.path=/usr/java/jdk1.6.0_32/jre/lib/amd64/ser
> >> ver:/usr/java/jdk1.6.0_32/jre/lib/amd64:/usr/java/jdk1.6.0_32/jre/../
> >> lib/amd64:/usr/java/packages/lib/amd64:/usr/lib64:/lib64:/lib:/usr/li
> >> b
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:java.io.tmpdir=/tmp
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:java.compiler=<NA>
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:os.name =Linux
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:os.arch=amd64
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:os.version=2.6.32-358.18.1.el6.x86_64
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:user.name =cloudera
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:user.home=/home/cloudera
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:user.dir=/home/cloudera/git/incubator-metamodel/hbase
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Initiating client
> >> connection,
> >> connectString=127.0.0.1:2181 sessionTimeout=90000
> >> watcher=hconnection-0x42787d6a
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.RecoverableZooKeeper: Process
> >> identifier=hconnection-0x42787d6a connecting to ZooKeeper ensemble=
> >> 127.0.0.1:2181
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ClientCnxn: Opening socket
> >> connection to server localhost.localdomain/127.0.0.1:2181. Will not
> >> attempt to authenticate using SASL (Unable to locate a login
> >> configuration)
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ClientCnxn: Socket connection
> >> established to localhost.localdomain/127.0.0.1:2181, initiating
> >> session
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ClientCnxn: Session establishment
> >> complete on server localhost.localdomain/127.0.0.1:2181, sessionid =
> >> 0x144d56ddb050007, negotiated timeout = 60000
> >>
> >> 14/03/18 06:47:30 WARN client.ServerCallable: Call exception,
> >> tries=0,
> >> numRetries=20
> >>
> >> java.io.IOException: Failed to find location, tableName=[B@73c6641,
> >> row=table_for_junit,,00000000000000, reload=false
> >>
> >>                 at
> >> org.apache.hadoop.hbase.client.ServerCallable.prepare(ServerCallable.
> >> java:98)
> >>
> >>                 at
> >> org.apache.hadoop.hbase.client.ScannerCallable.prepare(ScannerCallabl
> >> e.java:108)
> >>
> >>                 at
> >> org.apache.hadoop.hbase.client.ServerCallable.withRetries(ServerCalla
> >> ble.java:173)
> >>
> >>                 at
> >> org.apache.hadoop.hbase.client.ClientScanner.nextScanner(ClientScanne
> >> r.java:216)
> >>
> >>                 at
> >> org.apache.hadoop.hbase.client.ClientScanner.<init>(ClientScanner.jav
> >> a:131)
> >>
> >>                 at
> >> org.apache.hadoop.hbase.client.HTable.getScanner(HTable.java:585)
> >>
> >>                 at
> >> org.apache.hadoop.hbase.client.MetaScanner.metaScan(MetaScanner.java:
> >> 164)
> >>
> >>                 at
> >> org.apache.hadoop.hbase.client.MetaScanner.metaScan(MetaScanner.java:
> >> 102)
> >>
> >>                 at
> >> org.apache.hadoop.hbase.client.MetaScanner.metaScan(MetaScanner.java:
> >> 80)
> >>
> >>                 at
> >> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplemen
> >> tation.isTableAvailable(HConnectionManager.java:696)
> >>
> >>                 at
> >> org.apache.hadoop.hbase.client.HBaseAdmin.isTableAvailable(HBaseAdmin
> >> .java:988)
> >>
> >>                 at
> >> org.apache.metamodel.hbase.HBaseDataContextTest.createTableNatively(H
> >> BaseDataContextTest.java:146)
> >>
> >>                 at
> >> org.apache.metamodel.hbase.HBaseDataContextTest.setUp(HBaseDataContex
> >> tTest.java:49)
> >>
> >>                 at
> >> junit.framework.TestCase.runBare(TestCase.java:139)
> >>
> >>                 at
> >> junit.framework.TestResult$1.protect(TestResult.java:122)
> >>
> >>                 at
> >> junit.framework.TestResult.runProtected(TestResult.java:142)
> >>
> >>                 at
> >> junit.framework.TestResult.run(TestResult.java:125)
> >>
> >>                 at junit.framework.TestCase.run(TestCase.java:129)
> >>
> >>                 at
> >> junit.framework.TestSuite.runTest(TestSuite.java:255)
> >>
> >>                 at junit.framework.TestSuite.run(TestSuite.java:250)
> >>
> >>                 at
> >> org.junit.internal.runners.JUnit38ClassRunner.run(JUnit38ClassRunner.
> >> java:84)
> >>
> >>                 at
> >> org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provide
> >> r.java:252)
> >>
> >>                 at
> >> org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4
> >> Provider.java:141)
> >>
> >>                 at
> >> org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider
> >> .java:112)
> >>
> >>                 at
> >> sun.reflect.NativeMethodAccessorImpl.invoke0(Native
> >> Method)
> >>
> >>                 at
> >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.
> >> java:39)
> >>
> >>                 at
> >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces
> >> sorImpl.java:25)
> >>
> >>                 at java.lang.reflect.Method.invoke(Method.java:597)
> >>
> >>                 at
> >> org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(
> >> ReflectionUtils.java:189)
> >>
> >>                 at
> >> org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke
> >> (ProviderFactory.java:165)
> >>
> >>                 at
> >> org.apache.maven.surefire.booter.ProviderFactory.invokeProvider(Provi
> >> derFactory.java:85)
> >>
> >>                 at
> >> org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(Fork
> >> edBooter.java:115)
> >>
> >>                 at
> >> org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:
> >> 75)
> >>
> >>
> >>
> >> Please advise.
> >>
> >> Tomasz Guzialek
> >> Software Engineer
> >>
> >> [image: Description: Description:
> >> http://www.humaninference.com/banners/hilogo25.png]<http://www.humani
> >> nference.com/>
> >>
> >> Vesterbrogade 74, 3.
> >> 1620 København
> >> Danmark
> >>
> >> Tlf. +45 2613 4307
> >> [email protected]
> >>
> >>
> >> Human Inference B.V. - Chamber of Commerce NL - 09106800
> >>
> >> [image: Description: Description:
> >> http://www.humaninference.com/banners/banner%20datavaluetalk.png]<htt
> >> p://www.datavaluetalk.com/>
> >>
> >> The information contained in this communication is confidential and
> >> is intended solely for the person and/or entity addressed above. If
> >> you are not an intended recipient you are hereby notified that the
> >> content of this message may not be disclosed, copied or distributed
> >> to any other party than the addressee. Human Inference does not
> >> guarantee that this communication is free of viruses or that other
> >> parties than the intended recipient will not intercept this
> >> communication. Human Inference assumes no liability for the proper
> >> and complete transmission of the information contained in this
> >> communication or any delay in the receipt of this communication by the
> addressee.
> >> P Please consider the environment before printing this email.
> >>
> >>
> >>
> >
> >
>

Reply via email to