It's a known issue with Pig 0.10.0 as far as I know, so I usually recommend
the workaround that I suggested to you.

I haven't verified this by myself, but I believe that the recent change to
HBaseStorage (PIG-2821 <https://issues.apache.org/jira/browse/PIG-2821>)
fixes this problem (PIG-2822<https://issues.apache.org/jira/browse/PIG-2822>)
- i.e. hbase-site.xml not being packed into job.xml.

Regarding Pig 0.9.1, I haven't used that version, so I can't answer your
question.

Thanks,
Cheolsoo


On Thu, Sep 13, 2012 at 1:10 AM, lulynn_2008 <[email protected]> wrote:

> Hi Cheolsoo,
>
> Yes, you are right. I am running the ZK quorum on remote machines and you
> way works.
> --  But I have put hbase-site.xml(which include hbase.zookeeper.quorum) in
> pig classpath. But seems pig did not store these hbase configurations to
> job. Is this the design of pig-0.10.0?

--  Seems pig just puts hadoop configurations(hadoop-site.xml,
> core-site.xml...) into job.xml during create job. Is this correct?
> --  I remembered that pig-0.9.1 store hbase configurations into job.xml.
> Did I miss anything?
>
> Thanks.
>
>
>
>
>
>
> At 2012-09-13 02:24:18,"Cheolsoo Park" <[email protected]> wrote:
> >Hi,
> >
> >2012-09-12 00:30:07,198 INFO org.apache.zookeeper.ClientCnxn: Opening
> >> socket connection to server /127.0.0.1:2181
> >
> >
> >This message seems wrong. I assume that you're running the ZK quorum on
> >remote machines, but it is trying to connect to localhost. Can you try to
> >set "hbase.zookeeper.quorum" in "pig.properties" as follows:
> >"hbase.zookeeper.quorum=<your
> >ZK quorum host:port>" ?
> >
> >Thanks,
> >Cheolsoo
> >
> >On Wed, Sep 12, 2012 at 12:39 AM, lulynn_2008 <[email protected]>
> wrote:
> >
> >> Hi Cheolsoo,
> >> Current TestJobSubmission and TestHBaseStorage passed. But if I run
> >> following scripts with hbase-0.94/zookeeper-3.4.3 cluster:
> >>
> >> the same issue happened.
> >>
> >> 2012-09-12 00:30:07,198 INFO org.apache.zookeeper.ClientCnxn: Opening
> >> socket connection to server /127.0.0.1:2181
> >> 2012-09-12 00:30:07,199 INFO
> >> org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: The identifier
> of
> >> this process is [email protected]
> >> 2012-09-12 00:30:07,212 WARN
> >> org.apache.zookeeper.client.ZooKeeperSaslClient: SecurityException:
> >> java.lang.SecurityException: Unable to locate a login configuration
> >> occurred when trying to find JAAS configuration.
> >> 2012-09-12 00:30:07,213 INFO
> >> org.apache.zookeeper.client.ZooKeeperSaslClient: Client will not
> >> SASL-authenticate because the default JAAS configuration section
> 'Client'
> >> could not be found. If you are not using SASL, you may ignore this. On
> the
> >> other hand, if you expected SASL to work, please fix your JAAS
> >> configuration.
> >> 2012-09-12 00:30:07,221 WARN org.apache.zookeeper.ClientCnxn: Session
> 0x0
> >> for server null, unexpected error, closing socket connection and
> attempting
> >> reconnect
> >> java.net.ConnectException: Connection refused
> >>         at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
> >>         at
> >> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:610)
> >>         at
> >>
> org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:286)
> >>         at
> >> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1035)
> >>
> >> Reproduce steps:
> >>
> >> Create hbase table:
> >> ./hbase shell:
> >> create 'employees', 'SN', 'department', 'address'
> >> put 'employees', 'Hong', 'address:country', 'China'
> >>
> >> Run following pig commands:
> >> ./pig
> >> A = load 'hbase://employees' using
> >> org.apache.pig.backend.hadoop.hbase.HBaseStorage( 'address:country',
> >> '-loadKey true') as (SN:bytearray,country:bytearray);
> >> B = filter A by SN == 'Hong';
> >> dump B;
> >>
> >>
> >>
> >>
> >>
> >> At 2012-08-22 05:51:38,"Cheolsoo Park" <[email protected]> wrote:
> >> >OK, I got TestJobSubmission passing. Please apply the following diff to
> >> >your Pig:
> >> >
> >> >
> >> >diff --git test/org/apache/pig/test/TestJobSubmission.java
> >> >test/org/apache/pig/test/TestJobSubmission.java
> >> >index 89ac132..211c4ae 100644
> >> >--- test/org/apache/pig/test/TestJobSubmission.java
> >> >+++ test/org/apache/pig/test/TestJobSubmission.java
> >> >@@ -26,8 +26,10 @@ import java.util.Random;
> >> >
> >> > import org.apache.hadoop.conf.Configuration;
> >> > import org.apache.hadoop.hbase.HBaseTestingUtility;
> >> >*+import org.apache.hadoop.hbase.HConstants;*
> >> > import org.apache.hadoop.hbase.MiniHBaseCluster;
> >> > import org.apache.hadoop.hbase.util.Bytes;
> >> >*+import org.apache.hadoop.hbase.zookeeper.MiniZooKeeperCluster;*
> >> > import org.apache.hadoop.mapred.jobcontrol.Job;
> >> > import org.apache.hadoop.mapred.jobcontrol.JobControl;
> >> > import org.apache.pig.ExecType;
> >> >@@ -541,7 +543,7 @@ public class TestJobSubmission {
> >> >         // use the estimation
> >> >         Configuration conf = cluster.getConfiguration();
> >> >         HBaseTestingUtility util = new HBaseTestingUtility(conf);
> >> >*-        util.startMiniZKCluster();*
> >> >*+        int clientPort = util.startMiniZKCluster().getClientPort();*
> >> >         util.startMiniHBaseCluster(1, 1);
> >> >
> >> >         String query = "a = load '/passwd';" +
> >> >@@ -553,6 +555,7 @@ public class TestJobSubmission {
> >> >
> >> >
> pc.getConf().setProperty("pig.exec.reducers.bytes.per.reducer",
> >> >"100");
> >> >         pc.getConf().setProperty("pig.exec.reducers.max", "10");
> >> >*+        pc.getConf().setProperty(HConstants.ZOOKEEPER_CLIENT_PORT,
> >> >Integer.toString(clientPort));*
> >> >
> ConfigurationValidator.validatePigProperties(pc.getProperties());
> >> >         conf = ConfigurationUtil.toConfiguration(pc.getProperties());
> >> >         JobControlCompiler jcc = new JobControlCompiler(pc, conf);
> >> >
> >> >
> >> >There are several changes to MiniZooKeeperCluster in HBase 0.94
> compared
> >> to
> >> >0.90. In particular, the ZK client port is no longer 21818 by default.
> In
> >> >fact, it is randomized in 0.94. So the client port has to be explicitly
> >> set
> >> >in PigContext at runtime.
> >> >
> >> >Now I am able to run both TestJobSubmission and TestHBaseStorage.
> Please
> >> >let me know if this works for you. I am going to file a jira to capture
> >> >what I found, so we can make use of it when bumping the Hbase version
> in
> >> >the future.
> >> >
> >> >Thanks,
> >> >Cheolsoo
> >> >
> >> >
> >> >On Tue, Aug 21, 2012 at 1:25 AM, Cheolsoo Park <[email protected]
> >> >wrote:
> >> >
> >> >> Hi,
> >> >>
> >> >> TestJobSubmission doesn't pass for me either. But it doesn't throw
> the
> >> >> error that you're seeing. Instead, it hangs at line 583
> >> >> in TestJobSubmission.java. I am not sure yet what's happening though.
> >> >>
> >> >> Thanks,
> >> >> Cheolsoo
> >> >>
> >> >> On Tue, Aug 21, 2012 at 12:24 AM, lulynn_2008 <[email protected]>
> >> wrote:
> >> >>
> >> >>> Hi Cheolsoo,
> >> >>> Please help to check whether TestJobSunbmission passed in your
> >> >>> environment. Thanks
> >> >>>
> >> >>>
> >> >>> Hi Cheolsoo,
> >> >>> TestHBaseStorage passed now with your fix. Thanks.
> >> >>> But TestJobSubmission still failed with the same previous error with
> >> >>> TestHBaseStorage. Please help to check. Thanks
> >> >>>
> >> >>>
> >> >>>
> >> >>>
> >> >>>
> >> >>> At 2012-08-18 14:26:32,"Cheolsoo Park" <[email protected]>
> wrote:
> >> >>> >Hi,
> >> >>> >
> >> >>> >I tried the same version of libraries as you're using and was able
> to
> >> get
> >> >>> >TestHBaseStorage passing in trunk. Here is what I did:
> >> >>> >
> >> >>> >1) Updated the version of hadoop, hbase, and zookeeper in
> >> >>> >ivy/libraries.properties:
> >> >>> >
> >> >>> >hadoop-core.version=1.0.3
> >> >>> >> hadoop-test.version=1.0.3
> >> >>> >> hbase.version=0.94.0
> >> >>> >> zookeeper.version=3.4.3
> >> >>> >
> >> >>> >
> >> >>> >2) Updated TestHBaseStorage.java since it doesn't compile with
> >> >>> hbase-0.94.0:
> >> >>> >
> >> >>> >HTable table = new HTable(tableName); => HTable table = new
> >> HTable(conf,
> >> >>> >> tableName);
> >> >>> >
> >> >>> >
> >> >>> >There are 6 places to fix.
> >> >>> >
> >> >>> >3) Added the following lines to ivy.xml:
> >> >>> >
> >> >>> ><dependency org="com.github.stephenc.high-scale-lib"
> >> >>> name="high-scale-lib"
> >> >>> >> rev="${high-scale-lib.version}" conf="test->default"/>
> >> >>> >> <dependency org="com.google.protobuf" name="protobuf-java"
> >> >>> >> rev="${protobuf-java.version}" conf="test->default"/>
> >> >>> >
> >> >>> >
> >> >>> >I also added the following line to ivy/libraries.properties:
> >> >>> >
> >> >>> >high-scale-lib.version=1.1.1
> >> >>> >
> >> >>> >
> >> >>> >In particular, I was able to reproduce your error without this
> line in
> >> >>> >ivy.xml:
> >> >>> >
> >> >>> ><dependency org="com.google.protobuf" name="protobuf-java"
> >> >>> >> rev="${protobuf-java.version}" conf="test->default"/>
> >> >>> >
> >> >>> >
> >> >>> >With all these changes, TestHBaseStorage passes for me. Please let
> me
> >> >>> know
> >> >>> >if this works for you.
> >> >>> >
> >> >>> >Thanks,
> >> >>> >Cheolsoo
> >> >>> >
> >> >>> >On Thu, Aug 16, 2012 at 11:02 PM, lulynn_2008 <[email protected]
> >
> >> >>> wrote:
> >> >>> >
> >> >>> >> BTW, I am using hadoop-1.0.3, hbase-0.94.0 and zookeeper-3.4.3.
> >> >>> >>
> >> >>> >>
> >> >>> >> Hi Cheolsoo,
> >> >>> >> You are right. TestHBaseStorage also failed with the same error.
> >> >>> >>
> >> >>> >> I got following region server information, and seems hbase region
> >> has
> >> >>> been
> >> >>> >> started successfully. But hbase can not connect to zookeeper
> >> >>> >>
> >> >>> >> 12/08/13 06:34:12 INFO server.PrepRequestProcessor: Got
> user-level
> >> >>> >> KeeperException when processing sessionid:0x1391cf90aed0000
> >> type:delete
> >> >>> >> cxid:0x24 zxid:0x10 txntype:-1 reqpath:n/a Error
> >> >>> >> Path:/hbase/root-region-server Error:KeeperErrorCode = NoNode for
> >> >>> >> /hbase/root-region-server
> >> >>> >>
> >> >>> >> 12/08/13 06:34:16 INFO regionserver.HRegion: creating HRegion
> >> >>> pigtable_1
> >> >>> >> HTD == {NAME => 'pigtable_1', FAMILIES => [{NAME => 'pig',
> >> >>> >> REPLICATION_SCOPE => '0', KEEP_DELETED_CELLS => 'false',
> >> COMPRESSION =>
> >> >>> >> 'NONE', ENCODE_ON_DISK => 'true', BLOCKCACHE => 'true',
> >> MIN_VERSIONS =>
> >> >>> >> '0', DATA_BLOCK_ENCODING => 'NONE', IN_MEMORY => 'false',
> >> BLOOMFILTER
> >> >>> =>
> >> >>> >> 'NONE', TTL => '2147483647', VERSIONS => '3', BLOCKSIZE =>
> >> '65536'}]}
> >> >>> >> RootDir = hdfs://shihc015:54670/user/root/hbase Table name ==
> >> >>> pigtable_1
> >> >>> >>
> >> >>> >> 12/08/13 06:34:43 INFO zookeeper.ClientCnxn: Opening socket
> >> connection
> >> >>> to
> >> >>> >> server shihc015/127.0.0.1:21818
> >> >>> >> 12/08/13 06:34:43 WARN client.ZooKeeperSaslClient:
> >> SecurityException:
> >> >>> >> java.lang.SecurityException: Unable to locate a login
> configuration
> >> >>> >> occurred when trying to find JAAS configuration.
> >> >>> >> 12/08/13 06:34:43 INFO client.ZooKeeperSaslClient: Client will
> not
> >> >>> >> SASL-authenticate because the default JAAS configuration section
> >> >>> 'Client'
> >> >>> >> could not be found. If you are not using SASL, you may ignore
> this.
> >> On
> >> >>> the
> >> >>> >> other hand, if you expected SASL to work, please fix your JAAS
> >> >>> >> configuration.
> >> >>> >> 12/08/13 06:34:43 WARN zookeeper.ClientCnxn: Session 0x0 for
> server
> >> >>> null,
> >> >>> >> unexpected error, closing socket connection and attempting
> reconnect
> >> >>> >> java.net.ConnectException: Connection refused
> >> >>> >>         at sun.nio.ch.SocketChannelImpl.checkConnect(Native
> Method)
> >> >>> >>         at
> >> >>> >>
> >> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:610)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:286)
> >> >>> >>         at
> >> >>> >>
> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1035)
> >> >>> >> 12/08/13 06:34:43 WARN zookeeper.RecoverableZooKeeper: Possibly
> >> >>> transient
> >> >>> >> ZooKeeper exception:
> >> >>> >> org.apache.zookeeper.KeeperException$ConnectionLossException:
> >> >>> >> KeeperErrorCode = ConnectionLoss for /hbase/root-region-server
> >> >>> >>
> >> >>> >>
> >> >>> >> 12/08/16 22:33:05 ERROR zookeeper.ZooKeeperWatcher: hconnection
> >> >>> Received
> >> >>> >> unexpected KeeperException, re-throwing exception
> >> >>> >> org.apache.zookeeper.KeeperException$ConnectionLossException:
> >> >>> >> KeeperErrorCode = ConnectionLoss for /hbase/root-region-server
> >> >>> >>         at
> >> >>> >>
> org.apache.zookeeper.KeeperException.create(KeeperException.java:99)
> >> >>> >>         at
> >> >>> >>
> org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
> >> >>> >>         at
> >> org.apache.zookeeper.ZooKeeper.getData(ZooKeeper.java:1131)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.getData(RecoverableZooKeeper.java:264)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> org.apache.hadoop.hbase.zookeeper.ZKUtil.getDataInternal(ZKUtil.java:522)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> org.apache.hadoop.hbase.zookeeper.ZKUtil.getDataAndWatch(ZKUtil.java:498)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> org.apache.hadoop.hbase.zookeeper.ZooKeeperNodeTracker.getData(ZooKeeperNodeTracker.java:156)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> org.apache.hadoop.hbase.zookeeper.RootRegionTracker.getRootRegionLocation(RootRegionTracker.java:62)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:821)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:801)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegionInMeta(HConnectionManager.java:933)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:832)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:801)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegionInMeta(HConnectionManager.java:933)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:836)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:801)
> >> >>> >>         at
> >> >>> >>
> org.apache.hadoop.hbase.client.HTable.finishSetup(HTable.java:234)
> >> >>> >>         at
> >> >>> org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:174)
> >> >>> >>         at
> >> >>> org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:133)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> org.apache.pig.backend.hadoop.hbase.HBaseStorage.setLocation(HBaseStorage.java:551)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler.getJob(JobControlCompiler.java:385)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler.compile(JobControlCompiler.java:259)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> org.apache.pig.test.TestJobSubmission.testReducerNumEstimation(TestJobSubmission.java:492)
> >> >>> >>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
> >> Method)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:60)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:37)
> >> >>> >>         at java.lang.reflect.Method.invoke(Method.java:611)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:28)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> org.junit.runners.BlockJUnit4ClassRunner.runNotIgnored(BlockJUnit4ClassRunner.java:79)
> >> >>> >>         at
> >> >>> >>
> >> >>>
> >>
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:71)
> >> >>> >>
> >> >>> >>
> >> >>> >>
> >> >>> >>
> >> >>> >>
> >> >>> >>
> >> >>> >> At 2012-08-16 01:30:33,"Cheolsoo Park" <[email protected]>
> >> wrote:
> >> >>> >> >Hi,
> >> >>> >> >
> >> >>> >> >Caused by:
> >> org.apache.hadoop.hbase.client.NoServerForRegionException:
> >> >>> >> >> Unable to find region for passwd,,99999999999999 after 10
> tries.
> >> >>> >> >
> >> >>> >> >
> >> >>> >> >It looks like the region server is not available.
> >> >>> >> >1) Does TestHBaseStorage pass? I guess that it fails with the
> same
> >> >>> error.
> >> >>> >> >2) In your log, do you see any errors related to the HBase
> region
> >> >>> server?
> >> >>> >> >For example, java.lang.IllegalStateException: Could not
> >> instantiate a
> >> >>> >> >region instance.
> >> >>> >> >
> >> >>> >> >If the region server failed to start, you will see this failure.
> >> >>> >> >
> >> >>> >> >Thanks,
> >> >>> >> >Cheolsoo
> >> >>> >> >
> >> >>> >> >
> >> >>> >> >On Wed, Aug 15, 2012 at 12:56 AM, lulynn_2008 <
> [email protected]
> >> >
> >> >>> >> wrote:
> >> >>> >> >
> >> >>> >> >> Hi All,
> >> >>> >> >>
> >> >>> >> >> In TestJobSubmission, testReducerNumEstimation failed with
> >> following
> >> >>> >> error
> >> >>> >> >> information. Please give a glance to check what the problem
> is.
> >> >>> Thanks.
> >> >>> >> >>
> >> >>> >> >> Error information:
> >> >>> >> >> Testcase: testReducerNumEstimation took 1,809.419 sec
> >> >>> >> >>         Caused an ERROR
> >> >>> >> >> Internal error creating job configuration.
> >> >>> >> >>
> >> >>> >>
> >> >>>
> >>
> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobCreationException:
> >> >>> >> >> ERROR 2017: Internal error creating job configuration.
> >> >>> >> >>         at
> >> >>> >> >>
> >> >>> >>
> >> >>>
> >>
> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler.getJob(JobControlCompiler.java:731)
> >> >>> >> >>         at
> >> >>> >> >>
> >> >>> >>
> >> >>>
> >>
> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler.compile(JobControlCompiler.java:259)
> >> >>> >> >>         at
> >> >>> >> >>
> >> >>> >>
> >> >>>
> >>
> org.apache.pig.test.TestJobSubmission.testReducerNumEstimation(TestJobSubmission.java:491)
> >> >>> >> >> Caused by:
> >> >>> org.apache.hadoop.hbase.client.NoServerForRegionException:
> >> >>> >> >> Unable to find region for passwd,,99999999999999 after 10
> tries.
> >> >>> >> >>         at
> >> >>> >> >>
> >> >>> >>
> >> >>>
> >>
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegionInMeta(HConnectionManager.java:926)
> >> >>> >> >>         at
> >> >>> >> >>
> >> >>> >>
> >> >>>
> >>
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:832)
> >> >>> >> >>         at
> >> >>> >> >>
> >> >>> >>
> >> >>>
> >>
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:801)
> >> >>> >> >>         at
> >> >>> >> >>
> >> >>> >>
> >> >>>
> >>
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegionInMeta(HConnectionManager.java:933)
> >> >>> >> >>         at
> >> >>> >> >>
> >> >>> >>
> >> >>>
> >>
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:836)
> >> >>> >> >>         at
> >> >>> >> >>
> >> >>> >>
> >> >>>
> >>
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:801)
> >> >>> >> >>         at
> >> >>> >> >>
> >> org.apache.hadoop.hbase.client.HTable.finishSetup(HTable.java:234)
> >> >>> >> >>         at
> >> >>> org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:174)
> >> >>> >> >>         at
> >> >>> org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:133)
> >> >>> >> >>         at
> >> >>> >> >>
> >> >>> >>
> >> >>>
> >>
> org.apache.pig.backend.hadoop.hbase.HBaseStorage.setLocation(HBaseStorage.java:545)
> >> >>> >> >>         at
> >> >>> >> >>
> >> >>> >>
> >> >>>
> >>
> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler.getJob(JobControlCompiler.java:384)
> >> >>> >> >>
> >> >>> >> >>
> >> >>> >>
> >> >>> >>
> >> >>> >>
> >> >>> >>
> >> >>>
> >> >>>
> >> >>>
> >> >>>
> >> >>
> >>
>

Reply via email to