Hello all
The problem with Cloudera VM (CDH 4.5.0) is overcome (at least for me to work
on the code). In the hbase-module/pom.xml the hbase-client dependency is
replaced with Cloudera packages:
<properties>
<hadoop.version>2.0.0-cdh4.5.0</hadoop.version>
<hbase.version>0.94.6-cdh4.5.0</hbase.version>
</properties>
<repositories>
<repository>
<id>cloudera-releases</id>
<url>https://repository.cloudera.com/artifactory/cloudera-repos</url>
<releases>
<enabled>true</enabled>
</releases>
<snapshots>
<enabled>false</enabled>
</snapshots>
</repository>
</repositories>
<dependencies>
<dependency>
<groupId>org.apache.metamodel</groupId>
<artifactId>MetaModel-core</artifactId>
<version>${project.version}</version>
</dependency>
<dependency>
<groupId>org.apache.hadoop</groupId>
<artifactId>hadoop-client</artifactId>
<version>${hadoop.version}</version>
<exclusions>
<exclusion>
<groupId>commons-logging</groupId>
<artifactId>commons-logging</artifactId>
</exclusion>
</exclusions>
</dependency>
<dependency>
<groupId>org.apache.hbase</groupId>
<artifactId>hbase</artifactId>
<version>${hbase.version}</version>
<exclusions>
<exclusion>
<artifactId>log4j</artifactId>
<groupId>log4j</groupId>
</exclusion>
<exclusion>
<artifactId>commons-logging</artifactId>
<groupId>commons-logging</groupId>
</exclusion>
</exclusions>
</dependency>
<dependency>
<groupId>org.slf4j</groupId>
<artifactId>jcl-over-slf4j</artifactId>
</dependency>
<dependency>
<groupId>org.slf4j</groupId>
<artifactId>slf4j-log4j12</artifactId>
<scope>provided</scope>
</dependency>
<!-- Test dependencies -->
<dependency>
<groupId>junit</groupId>
<artifactId>junit</artifactId>
<scope>test</scope>
</dependency>
</dependencies>
Additionally, one line has to be commented in
HBaseDataContext.materializeMainSchemaTable: scan.setMaxResultSize(maxRows);
After doing so, the connection is successful.
Of course, including CDH dependencies is not a way to go for MetaModel, so the
question about REST API is a valid point...
Best regards,
Tomasz
-----Original Message-----
From: Kasper Sørensen [mailto:[email protected]]
Sent: Wednesday, March 19, 2014 21:59
To: [email protected]
Subject: Re: [MetaModel hbase-module] HBaseDataContextTest fail in Cloudera VM.
I don't have any good answer for this. Probably it's rather something to look
for online tutorials to find help with setting up HBase client ...
One thought in my mind though is that HBase client seems very version-specific.
Maybe it's a better route for MetaModel to go through the REST api than the
Java API, although I guess it may come at a severe performance penalty?
2014-03-19 10:50 GMT+01:00 Tomasz Guziałek <
[email protected]>:
> Hello Kasper
>
> I have also tried to set up HBase (version 0.94.17) on Windows and
> experienced the same issue: some Zookeper classes could not be found
> on the classpath.
>
> But coming back to the Cloudera VM. It seems that it is not a problem
> with the test but the hbase-module itself as I wrote this snippet to
> read from a table that I know it exists in the HBase:
>
> public class HBaseDemo {
>
> public static void main(String[] args) {
> HBaseConfiguration configuration = new
> HBaseConfiguration();
> HBaseDataContext dataContext = new
> HBaseDataContext(configuration);
> DataSet dataset =
> dataContext.query().from("countrycodes").selectAll().execute();
>
> for (Row row : dataset) {
> System.out.println(row.toString());
> }
> }
> }
>
> The stacktrace:
>
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in
> [jar:file:/home/cloudera/.m2/repository/org/slf4j/slf4j-nop/1.7.4/slf4
> j-nop-1.7.4.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in
> [jar:file:/home/cloudera/.m2/repository/org/slf4j/slf4j-log4j12/1.7.4/
> slf4j-log4j12-1.7.4.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
> explanation.
> SLF4J: Actual binding is of type [org.slf4j.helpers.NOPLoggerFactory]
> Exception in thread "main" org.apache.metamodel.MetaModelException:
> org.apache.hadoop.hbase.exceptions.MasterNotRunningException:
> com.google.protobuf.ServiceException: java.io.IOException: Broken pipe
> at
> org.apache.metamodel.hbase.HBaseDataContext.getMainSchema(HBaseDataContext.java:131)
> at
> org.apache.metamodel.QueryPostprocessDataContext.getMainSchemaInternal(QueryPostprocessDataContext.java:498)
> at
> org.apache.metamodel.QueryPostprocessDataContext.getSchemaByNameInternal(QueryPostprocessDataContext.java:369)
> at
> org.apache.metamodel.AbstractDataContext.getSchemaByName(AbstractDataContext.java:199)
> at
> org.apache.metamodel.AbstractDataContext.getDefaultSchema(AbstractDataContext.java:118)
> at
> org.apache.metamodel.AbstractDataContext.getSchemaByToken(AbstractDataContext.java:503)
> at
> org.apache.metamodel.AbstractDataContext.getTableByQualifiedLabel(AbstractDataContext.java:389)
> at
> org.apache.metamodel.query.builder.InitFromBuilderImpl.from(InitFromBuilderImpl.java:78)
> at info.guzialek.hbase.HBaseDemo.main(HBaseDemo.java:13)
> Caused by: org.apache.hadoop.hbase.exceptions.MasterNotRunningException:
> com.google.protobuf.ServiceException: java.io.IOException: Broken pipe
> at
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation$StubMaker.makeStub(HConnectionManager.java:1409)
> at
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation$MasterMonitorServiceStubMaker.makeStub(HConnectionManager.java:1435)
> at
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.getKeepAliveMasterMonitorService(HConnectionManager.java:1884)
> at
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.listTables(HConnectionManager.java:2645)
> at
> org.apache.hadoop.hbase.client.HBaseAdmin.listTables(HBaseAdmin.java:279)
> at
> org.apache.metamodel.hbase.HBaseDataContext.getMainSchema(HBaseDataContext.java:117)
> ... 8 more
> Caused by: com.google.protobuf.ServiceException: java.io.IOException:
> Broken pipe
> at
> org.apache.hadoop.hbase.ipc.RpcClient.callBlockingMethod(RpcClient.java:1556)
> at
> org.apache.hadoop.hbase.ipc.RpcClient$BlockingRpcChannelImplementation.callBlockingMethod(RpcClient.java:1597)
> at
> org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos$MasterMonitorService$BlockingStub.isMasterRunning(MasterMonitorProtos.java:3021)
> at
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation$MasterMonitorServiceStubMaker.isMasterRunning(HConnectionManager.java:1446)
> at
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation$StubMaker.makeStubNoRetries(HConnectionManager.java:1354)
> at
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation$StubMaker.makeStub(HConnectionManager.java:1380)
> ... 13 more
> Caused by: java.io.IOException: Broken pipe
> at sun.nio.ch.FileDispatcher.write0(Native Method)
> at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:29)
> at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:69)
> at sun.nio.ch.IOUtil.write(IOUtil.java:40)
> at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:336)
> at
> org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:55)
> at
> org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:142)
> at
> org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:146)
> at
> org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:107)
> at
> java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:65)
> at
> java.io.BufferedOutputStream.flush(BufferedOutputStream.java:123)
> at java.io.DataOutputStream.flush(DataOutputStream.java:106)
> at
> org.apache.hadoop.hbase.ipc.RpcClient$Connection.writeConnectionHeader(RpcClient.java:906)
> at
> org.apache.hadoop.hbase.ipc.RpcClient$Connection.setupIOstreams(RpcClient.java:856)
> at
> org.apache.hadoop.hbase.ipc.RpcClient.getConnection(RpcClient.java:1422)
> at org.apache.hadoop.hbase.ipc.RpcClient.call(RpcClient.java:1314)
> at
> org.apache.hadoop.hbase.ipc.RpcClient.callBlockingMethod(RpcClient.java:1540)
> ... 18 more
>
> My MapReduce code is able to connect to that instance of HBase from
> the same host (localhost). When I point my browser to
> http://127.0.0.1:60010I can see the web UI of HBase and I verified that the
> ZooKeeper default port has not been changed (it is 2181 as in the
> HBaseConfiguration).
>
> Tomasz Guzialek
> Software Engineer
>
> Vesterbrogade 74, 3.
> 1620 København
> Danmark
> Tlf. +45 2613 4307
> [email protected]
>
> Human Inference B.V. - Chamber of Commerce NL - 09106800
>
>
> The information contained in this communication is confidential and is
> intended solely for the person and/or entity addressed above. If you
> are not an intended recipient you are hereby notified that the content
> of this message may not be disclosed, copied or distributed to any
> other party than the addressee. Human Inference does not guarantee
> that this communication is free of viruses or that other parties than
> the intended recipient will not intercept this communication. Human
> Inference assumes no liability for the proper and complete
> transmission of the information contained in this communication or any
> delay in the receipt of this communication by the addressee.
> P Please consider the environment before printing this email.
>
> -----Original Message-----
> From: Kasper Sørensen [mailto:[email protected]]
> Sent: Tuesday, March 18, 2014 19:19
> To: [email protected]
> Subject: Re: [MetaModel hbase-module] HBaseDataContextTest fail in
> Cloudera VM.
>
> Hmm tried some experiments but ultimately I am having trouble even
> getting HBase to work on my Windows box. My "old" hbase installation
> (version
> 0.95.1-hadoop1) on windows does work, and the unittests pass etc.
> But newer versions I cannot seem to run on Windows (zookeeper issues I
> believe), so I can't reproduce what you're experiencing.
>
>
>
> 2014-03-18 16:09 GMT+01:00 Kasper Sørensen
> <[email protected]
> >:
>
> > Hi Tomasz
> >
> > It might be that it's just the unittest that's a bit stupid. Or else
> > it must be a HBase configuration thing. I looked at the source code
> > of HBase where the exception is thrown. It is looking for a
> > "location" of the table name. I guess you might want to try to skip
> > the part of the unittest that looks for the table, and instead
> > simply create it (it will not initially exist)...
> >
> > I am doing a reinstall of my HBase here to double verify, but this
> > is my guess.
> >
> > Kasper
> >
> >
> > 2014-03-18 14:52 GMT+01:00 Tomasz Guziałek <
> > [email protected]>:
> >
> > Hello all
> >>
> >>
> >>
> >> My name is Tomasz and I am working on my Master thesis in
> >> collaboration with Human Inference. Long story short, the thesis
> >> involves creating a Hadoop version of DataCleaner
> >> (http://datacleaner.org), so I am working with HBase too. Within
> >> the project, I expect to make some contributions to the HBase
> >> module of
> MetaModel.
> >>
> >>
> >>
> >> However, just at the beginning I encountered problems with the
> >> build of the module. I use Cloudera VM (
> >> http://www.cloudera.com/content/cloudera-content/cloudera-docs/Demo
> >> VM
> >> s/Cloudera-QuickStart-VM/cloudera_quickstart_vm.html)
> >> as a starting point to make sure my HBase cluster is correct (the
> >> HBase version reported by Cloudera Manager: 0.94.6+132). After a
> >> checkout of the MetaModel code, I cannot build it with “maven
> >> install” because of HBaseDataContextTest fail:
> >>
> >>
> >>
> >> -------------------------------------------------------
> >>
> >> T E S T S
> >>
> >> -------------------------------------------------------
> >>
> >> Running org.apache.metamodel.hbase.HBaseDataContextTest
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:zookeeper.version=3.4.5-1392090, built on 09/30/2012
> >> 17:52 GMT
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:host.name =localhost.localdomain
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:java.version=1.6.0_32
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:java.vendor=Sun Microsystems Inc.
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:java.home=/usr/java/jdk1.6.0_32/jre
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >>
> environment:java.class.path=/home/cloudera/git/incubator-metamodel/hbase/target/test-classes:/home/cloudera/git/incubator-metamodel/hbase/target/classes:/home/cloudera/git/incubator-metamodel/core/target/MetaModel-core-4.0.1-incubating-SNAPSHOT.jar:/home/cloudera/.m2/repository/org/slf4j/slf4j-api/1.7.4/slf4j-api-1.7.4.jar:/home/cloudera/.m2/repository/org/apache/hbase/hbase-client/0.95.1-hadoop1/hbase-client-0.95.1-hadoop1.jar:/home/cloudera/.m2/repository/org/apache/hbase/hbase-common/0.95.1-hadoop1/hbase-common-0.95.1-hadoop1.jar:/home/cloudera/.m2/repository/com/google/guava/guava/12.0.1/guava-12.0.1.jar:/home/cloudera/.m2/repository/com/google/code/findbugs/jsr305/1.3.9/jsr305-1.3.9.jar:/home/cloudera/.m2/repository/org/apache/hbase/hbase-protocol/0.95.1-hadoop1/hbase-protocol-0.95.1-hadoop1.jar:/home/cloudera/.m2/repository/com/google/protobuf/protobuf-java/2.4.1/protobuf-java-2.4.1.jar:/home/cloudera/.m2/repository/org/apache/zookeeper/zookeeper/3.4.5/zookeeper-3.4.5.jar:/home/cloudera/.m2/repository/org/cloudera/htrace/htrace/1.50/htrace-1.50.jar:/home/cloudera/.m2/repository/org/mortbay/jetty/jetty-util/6.1.26/jetty-util-6.1.26.jar:/home/cloudera/.m2/repository/io/netty/netty/3.5.9.Final/netty-3.5.9.Final.jar:/home/cloudera/.m2/repository/org/apache/hadoop/hadoop-core/1.1.2/hadoop-core-1.1.2.jar:/home/cloudera/.m2/repository/commons-cli/commons-cli/1.2/commons-cli-1.2.jar:/home/cloudera/.m2/repository/xmlenc/xmlenc/0.52/xmlenc-0.52.jar:/home/cloudera/.m2/repository/com/sun/jersey/jersey-core/1.8/jersey-core-1.8.jar:/home/cloudera/.m2/repository/com/sun/jersey/jersey-json/1.8/jersey-json-1.8.jar:/home/cloudera/.m2/repository/org/codehaus/jettison/jettison/1.1/jettison-1.1.jar:/home/cloudera/.m2/repository/stax/stax-api/1.0.1/stax-api-1.0.1.jar:/home/cloudera/.m2/repository/com/sun/xml/bind/jaxb-impl/2.2.3-1/jaxb-impl-2.2.3-1.jar:/home/cloudera/.m2/repository/javax/xml/bind/jaxb-api/2.2.2/jaxb-api-2.2.2.jar:/home/cloudera/.m2/repository/javax/xml/stream/stax-api/1.0-2/stax-api-1.0-2.jar:/home/cloudera/.m2/repository/javax/activation/activation/1.1/activation-1.1.jar:/home/cloudera/.m2/repository/org/codehaus/jackson/jackson-core-asl/1.9.12/jackson-core-asl-1.9.12.jar:/home/cloudera/.m2/repository/org/codehaus/jackson/jackson-jaxrs/1.7.1/jackson-jaxrs-1.7.1.jar:/home/cloudera/.m2/repository/org/codehaus/jackson/jackson-xc/1.7.1/jackson-xc-1.7.1.jar:/home/cloudera/.m2/repository/com/sun/jersey/jersey-server/1.8/jersey-server-1.8.jar:/home/cloudera/.m2/repository/asm/asm/3.1/asm-3.1.jar:/home/cloudera/.m2/repository/commons-io/commons-io/2.4/commons-io-2.4.jar:/home/cloudera/.m2/repository/commons-httpclient/commons-httpclient/3.0.1/commons-httpclient-3.0.1.jar:/home/cloudera/.m2/repository/commons-codec/commons-codec/1.7/commons-codec-1.7.jar:/home/cloudera/.m2/repository/org/apache/commons/commons-math/2.1/commons-math-2.1.jar:/home/cloudera/.m2/repository/commons-configuration/commons-configuration/1.6/commons-configuration-1.6.jar:/home/cloudera/.m2/repository/commons-collections/commons-collections/3.2.1/commons-collections-3.2.1.jar:/home/cloudera/.m2/repository/commons-lang/commons-lang/2.6/commons-lang-2.6.jar:/home/cloudera/.m2/repository/commons-digester/commons-digester/1.8/commons-digester-1.8.jar:/home/cloudera/.m2/repository/commons-beanutils/commons-beanutils/1.7.0/commons-beanutils-1.7.0.jar:/home/cloudera/.m2/repository/commons-beanutils/commons-beanutils-core/1.8.0/commons-beanutils-core-1.8.0.jar:/home/cloudera/.m2/repository/commons-net/commons-net/1.4.1/commons-net-1.4.1.jar:/home/cloudera/.m2/repository/org/mortbay/jetty/jetty/6.1.26/jetty-6.1.26.jar:/home/cloudera/.m2/repository/org/mortbay/jetty/servlet-api/2.5-20081211/servlet-api-2.5-20081211.jar:/home/cloudera/.m2/repository/tomcat/jasper-runtime/5.5.12/jasper-runtime-5.5.12.jar:/home/cloudera/.m2/repository/tomcat/jasper-compiler/5.5.12/jasper-compiler-5.5.12.jar:/home/cloudera/.m2/repository/org/mortbay/jetty/jsp-api-2.1/6.1.14/jsp-api-2.1-6.1.14.jar:/home/cloudera/.m2/repository/org/mortbay/jetty/servlet-api-2.5/6.1.14/servlet-api-2.5-6.1.14.jar:/home/cloudera/.m2/repository/org/mortbay/jetty/jsp-2.1/6.1.14/jsp-2.1-6.1.14.jar:/home/cloudera/.m2/repository/ant/ant/1.6.5/ant-1.6.5.jar:/home/cloudera/.m2/repository/commons-el/commons-el/1.0/commons-el-1.0.jar:/home/cloudera/.m2/repository/org/codehaus/jackson/jackson-mapper-asl/1.9.12/jackson-mapper-asl-1.9.12.jar:/home/cloudera/.m2/repository/com/github/stephenc/findbugs/findbugs-annotations/1.3.9-1/findbugs-annotations-1.3.9-1.jar:/home/cloudera/.m2/repository/org/slf4j/jcl-over-slf4j/1.7.4/jcl-over-slf4j-1.7.4.jar:/home/cloudera/.m2/repository/org/slf4j/slf4j-log4j12/1.7.4/slf4j-log4j12-1.7.4.jar:/home/cloudera/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar:/home/cloudera/.m2/repository/junit/junit/4.11/junit-4.11.jar:/home/cloudera/.m2/repository/org/hamcrest/hamcrest-core/1.3/hamcrest-core-1.3.jar:
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:java.library.path=/usr/java/jdk1.6.0_32/jre/lib/amd64/s
> >> er
> >> ver:/usr/java/jdk1.6.0_32/jre/lib/amd64:/usr/java/jdk1.6.0_32/jre/.
> >> ./
> >> lib/amd64:/usr/java/packages/lib/amd64:/usr/lib64:/lib64:/lib:/usr/
> >> li
> >> b
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:java.io.tmpdir=/tmp
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:java.compiler=<NA>
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:os.name =Linux
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:os.arch=amd64
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:os.version=2.6.32-358.18.1.el6.x86_64
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:user.name =cloudera
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:user.home=/home/cloudera
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Client
> >> environment:user.dir=/home/cloudera/git/incubator-metamodel/hbase
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ZooKeeper: Initiating client
> >> connection,
> >> connectString=127.0.0.1:2181 sessionTimeout=90000
> >> watcher=hconnection-0x42787d6a
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.RecoverableZooKeeper: Process
> >> identifier=hconnection-0x42787d6a connecting to ZooKeeper ensemble=
> >> 127.0.0.1:2181
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ClientCnxn: Opening socket
> >> connection to server localhost.localdomain/127.0.0.1:2181. Will not
> >> attempt to authenticate using SASL (Unable to locate a login
> >> configuration)
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ClientCnxn: Socket connection
> >> established to localhost.localdomain/127.0.0.1:2181, initiating
> >> session
> >>
> >> 14/03/18 06:45:30 INFO zookeeper.ClientCnxn: Session establishment
> >> complete on server localhost.localdomain/127.0.0.1:2181, sessionid
> >> = 0x144d56ddb050007, negotiated timeout = 60000
> >>
> >> 14/03/18 06:47:30 WARN client.ServerCallable: Call exception,
> >> tries=0,
> >> numRetries=20
> >>
> >> java.io.IOException: Failed to find location, tableName=[B@73c6641,
> >> row=table_for_junit,,00000000000000, reload=false
> >>
> >> at
> >> org.apache.hadoop.hbase.client.ServerCallable.prepare(ServerCallable.
> >> java:98)
> >>
> >> at
> >> org.apache.hadoop.hbase.client.ScannerCallable.prepare(ScannerCalla
> >> bl
> >> e.java:108)
> >>
> >> at
> >> org.apache.hadoop.hbase.client.ServerCallable.withRetries(ServerCal
> >> la
> >> ble.java:173)
> >>
> >> at
> >> org.apache.hadoop.hbase.client.ClientScanner.nextScanner(ClientScan
> >> ne
> >> r.java:216)
> >>
> >> at
> >> org.apache.hadoop.hbase.client.ClientScanner.<init>(ClientScanner.j
> >> av
> >> a:131)
> >>
> >> at
> >> org.apache.hadoop.hbase.client.HTable.getScanner(HTable.java:585)
> >>
> >> at
> >> org.apache.hadoop.hbase.client.MetaScanner.metaScan(MetaScanner.java:
> >> 164)
> >>
> >> at
> >> org.apache.hadoop.hbase.client.MetaScanner.metaScan(MetaScanner.java:
> >> 102)
> >>
> >> at
> >> org.apache.hadoop.hbase.client.MetaScanner.metaScan(MetaScanner.java:
> >> 80)
> >>
> >> at
> >> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplem
> >> en
> >> tation.isTableAvailable(HConnectionManager.java:696)
> >>
> >> at
> >> org.apache.hadoop.hbase.client.HBaseAdmin.isTableAvailable(HBaseAdm
> >> in
> >> .java:988)
> >>
> >> at
> >> org.apache.metamodel.hbase.HBaseDataContextTest.createTableNatively
> >> (H
> >> BaseDataContextTest.java:146)
> >>
> >> at
> >> org.apache.metamodel.hbase.HBaseDataContextTest.setUp(HBaseDataCont
> >> ex
> >> tTest.java:49)
> >>
> >> at
> >> junit.framework.TestCase.runBare(TestCase.java:139)
> >>
> >> at
> >> junit.framework.TestResult$1.protect(TestResult.java:122)
> >>
> >> at
> >> junit.framework.TestResult.runProtected(TestResult.java:142)
> >>
> >> at
> >> junit.framework.TestResult.run(TestResult.java:125)
> >>
> >> at junit.framework.TestCase.run(TestCase.java:129)
> >>
> >> at
> >> junit.framework.TestSuite.runTest(TestSuite.java:255)
> >>
> >> at
> >> junit.framework.TestSuite.run(TestSuite.java:250)
> >>
> >> at
> >> org.junit.internal.runners.JUnit38ClassRunner.run(JUnit38ClassRunner.
> >> java:84)
> >>
> >> at
> >> org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provi
> >> de
> >> r.java:252)
> >>
> >> at
> >> org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUni
> >> t4
> >> Provider.java:141)
> >>
> >> at
> >> org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provid
> >> er
> >> .java:112)
> >>
> >> at
> >> sun.reflect.NativeMethodAccessorImpl.invoke0(Native
> >> Method)
> >>
> >> at
> >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.
> >> java:39)
> >>
> >> at
> >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcc
> >> es
> >> sorImpl.java:25)
> >>
> >> at java.lang.reflect.Method.invoke(Method.java:597)
> >>
> >> at
> >> org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArra
> >> y(
> >> ReflectionUtils.java:189)
> >>
> >> at
> >> org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invo
> >> ke
> >> (ProviderFactory.java:165)
> >>
> >> at
> >> org.apache.maven.surefire.booter.ProviderFactory.invokeProvider(Pro
> >> vi
> >> derFactory.java:85)
> >>
> >> at
> >> org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(Fo
> >> rk
> >> edBooter.java:115)
> >>
> >> at
> >> org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:
> >> 75)
> >>
> >>
> >>
> >> Please advise.
> >>
> >> Tomasz Guzialek
> >> Software Engineer
> >>
> >> [image: Description: Description:
> >> http://www.humaninference.com/banners/hilogo25.png]<http://www.huma
> >> ni
> >> nference.com/>
> >>
> >> Vesterbrogade 74, 3.
> >> 1620 København
> >> Danmark
> >>
> >> Tlf. +45 2613 4307
> >> [email protected]
> >>
> >>
> >> Human Inference B.V. - Chamber of Commerce NL - 09106800
> >>
> >> [image: Description: Description:
> >> http://www.humaninference.com/banners/banner%20datavaluetalk.png]<h
> >> tt
> >> p://www.datavaluetalk.com/>
> >>
> >> The information contained in this communication is confidential and
> >> is intended solely for the person and/or entity addressed above. If
> >> you are not an intended recipient you are hereby notified that the
> >> content of this message may not be disclosed, copied or distributed
> >> to any other party than the addressee. Human Inference does not
> >> guarantee that this communication is free of viruses or that other
> >> parties than the intended recipient will not intercept this
> >> communication. Human Inference assumes no liability for the proper
> >> and complete transmission of the information contained in this
> >> communication or any delay in the receipt of this communication by
> >> the
> addressee.
> >> P Please consider the environment before printing this email.
> >>
> >>
> >>
> >
> >
>