Hi,
This is also not this, mappers fail even if I run 2, and I have changed that value to 60 and tested it.

--
Wojciech Langiewicz

On 22.01.2011 20:18, Stack wrote:
Does it work if you have<  30 mappers running but if you go>  30, some
start to fail?  Perhaps it this config:
hbase.zookeeper.property.maxClientCnxns.  Its default is 30 max
connections per host.
St.Ack



On Sat, Jan 22, 2011 at 10:42 AM, Wojciech Langiewicz
<[email protected]>  wrote:
I have changed it to 120000, and there's no change.
What is worth noticing is that some mappers finish task without any error,
and during the same job, the same server can fail some tasks, and finishes
other.


On 22.01.2011 18:56, Wojciech Langiewicz wrote:

Hi,
sessionTimeout=60000,
I didn't change it, so it's default.
What value do you recommend?

--
Wojciech Langiewicz

On 22.01.2011 16:49, Ted Yu wrote:

What's the value of 'zookeeper.session.timeout' ?

Maybe you can tune it higher.

On Sat, Jan 22, 2011 at 3:13 AM, Wojciech Langiewicz
<[email protected]>wrote:

Hi,
I have re-run test with 2 or more mappers and looked into logs more
closely: The mapreduce job has finished correctly, some map attempts are
killed, but eventually all of mappers finish. And there's no rule which
mappers on which servers fail (re-run tests multiple times).

So I suspect that this is not classpath problem, maybe there are
settings
that limit number of connections, because before MasterNotRunning I
get this
in logs:

2011-01-22 12:06:43,411 INFO org.apache.zookeeper.ZooKeeper: Initiating
client connection, connectString=hd-master:2181 sessionTimeout=60000
watcher=org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper@3b5e234c
2011-01-22 12:06:43,483 INFO org.apache.zookeeper.ClientCnxn: Opening
socket connection to server hd-master/10.6.75.212:2181
2011-01-22 12:06:43,484 INFO org.apache.zookeeper.ClientCnxn: Socket
connection established to hd-master/10.6.75.212:2181, initiating session
2011-01-22 12:06:43,488 INFO org.apache.zookeeper.ClientCnxn: Unable to
read additional data from server sessionid 0x0, likely server has closed
socket, closing socket connection and attempting reconnect
2011-01-22 12:06:43,605 INFO
org.apache.hadoop.hbase.client.HConnectionManager$TableServers:
getMaster
attempt 0 of 10 failed; retrying after sleep of 1000
java.io.IOException:
org.apache.zookeeper.KeeperException$ConnectionLossException:
KeeperErrorCode = ConnectionLoss for /hbase/master
at

org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper.readAddressOrThrow(ZooKeeperWrapper.java:481)

at

org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper.readMasterAddressOrThrow(ZooKeeperWrapper.java:377)

at

org.apache.hadoop.hbase.client.HConnectionManager$TableServers.getMaster(HConnectionManager.java:381)


at
org.apache.hadoop.hbase.client.HBaseAdmin.<init>(HBaseAdmin.java:78)
at

org.apache.hadoop.hbase.PerformanceEvaluation$Test.testSetup(PerformanceEvaluation.java:745)

at

org.apache.hadoop.hbase.PerformanceEvaluation$Test.test(PerformanceEvaluation.java:764)

at

org.apache.hadoop.hbase.PerformanceEvaluation.runOneClient(PerformanceEvaluation.java:1097)

at

org.apache.hadoop.hbase.PerformanceEvaluation$EvaluationMapTask.map(PerformanceEvaluation.java:446)

at

org.apache.hadoop.hbase.PerformanceEvaluation$EvaluationMapTask.map(PerformanceEvaluation.java:399)

at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144)
at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:639)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:315)
at org.apache.hadoop.mapred.Child$4.run(Child.java:217)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:396)
at

org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1063)

at org.apache.hadoop.mapred.Child.main(Child.java:211)
Caused by: org.apache.zookeeper.KeeperException$ConnectionLossException:
KeeperErrorCode = ConnectionLoss for /hbase/master
at
org.apache.zookeeper.KeeperException.create(KeeperException.java:90)
at
org.apache.zookeeper.KeeperException.create(KeeperException.java:42)
at org.apache.zookeeper.ZooKeeper.getData(ZooKeeper.java:921)
at

org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper.readAddressOrThrow(ZooKeeperWrapper.java:477)

... 16 more


On 22.01.2011 01:36, Stack wrote:

Then its odd that PE fails. Can you figure difference between the two
environments? Perhaps your MR jobs are fat jars that include the conf
and all dependencies whereas PE is dump and expects the dependencies
and conf on CLASSPATH?

St.Ack

On Fri, Jan 21, 2011 at 11:56 AM, Wojciech Langiewicz
<[email protected]>  wrote:

Hi,
I have other mapreduce tasks that are running on this cluster and
using
HBase and they are working correctly. All my servers have the same
configuration.

--
Wojciech Langiewicz

On 21.01.2011 19:20, Stack wrote:


When clients are>  1, then PE tries to run a mapreduce job to host the
loading clients.

Is it possible that the client out in MR task is trying to connect to
wrong location? Perhaps the HBase conf dir is not available to the
running task? Have you seen



http://hbase.apache.org/apidocs/org/apache/hadoop/hbase/mapreduce/package-summary.html#package_description

?
Perhaps this will help?

St.Ack

On Fri, Jan 21, 2011 at 8:12 AM, Wojciech Langiewicz
<[email protected]>  wrote:


Hello
I have problem with running HBase performance tests from
org.apache.hadoop.hbase.PerformanceEvaluation package. I'm using
version
from CDH3,
Tests are ok when argument nclients is 1, but in case of greater
number,
after mappers reaching 100% I get this exception (I didn't test all
tests,
but all of tested by me failed, 'scan' and 'randomRead' fail for
sure):

11/01/21 17:05:19 INFO mapred.JobClient: Task Id :
attempt_201101211442_0005_m_000000_0, Status : FAILED
org.apache.hadoop.hbase.MasterNotRunningException
at



org.apache.hadoop.hbase.client.HConnectionManager$TableServers.getMaster(HConnectionManager.java:416)

at
org.apache.hadoop.hbase.client.HBaseAdmin.<init>(HBaseAdmin.java:78)
at



org.apache.hadoop.hbase.PerformanceEvaluation$Test.testSetup(PerformanceEvaluation.java:745)

at



org.apache.hadoop.hbase.PerformanceEvaluation$Test.test(PerformanceEvaluation.java:764)

at



org.apache.hadoop.hbase.PerformanceEvaluation.runOneClient(PerformanceEvaluation.java:1097)

at



org.apache.hadoop.hbase.PerformanceEvaluation$EvaluationMapTask.map(PerformanceEvaluation.java:446)

at



org.apache.hadoop.hbase.PerformanceEvaluation$EvaluationMapTask.map(PerformanceEvaluation.java:399)

at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144)
at
org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:639)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:315)
at org.apache.hadoop.mapred.Child$4.run(Child.java:217)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:396)
at



org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1063)

at org.apache.hadoop.mapred.Child.main(Child.java:211)

Do you have any ideas how to solve this?

--
Wojciech Langiewicz










Reply via email to