With a single family, ICV finished in 4:21 minutes. So it is a limitation of how many families are there in the mix. Need to re-think the schema ... :-(
[qw...@app48 logs]$ ~/scripts/loadDirect.sh [2009-07-06 20:12:23.542] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV] ... [2009-07-06 20:12:32.895] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [10,000] records [2009-07-06 20:12:42.198] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [20,000] records [2009-07-06 20:12:50.956] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [30,000] records [2009-07-06 20:12:59.087] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [40,000] records [2009-07-06 20:13:08.258] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [50,000] records [2009-07-06 20:13:16.773] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [60,000] records [2009-07-06 20:13:25.128] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [70,000] records [2009-07-06 20:13:34.309] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [80,000] records [2009-07-06 20:13:42.845] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [90,000] records [2009-07-06 20:13:51.363] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [100,000] records [2009-07-06 20:14:00.627] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [110,000] records [2009-07-06 20:14:08.964] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [120,000] records [2009-07-06 20:14:17.896] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [130,000] records [2009-07-06 20:14:27.680] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [140,000] records [2009-07-06 20:14:36.821] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [150,000] records [2009-07-06 20:14:45.966] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [160,000] records [2009-07-06 20:14:54.911] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [170,000] records [2009-07-06 20:15:03.736] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [180,000] records [2009-07-06 20:15:12.037] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [190,000] records [2009-07-06 20:15:20.494] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [200,000] records [2009-07-06 20:15:29.216] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [210,000] records [2009-07-06 20:15:37.809] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [220,000] records [2009-07-06 20:15:46.811] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [230,000] records [2009-07-06 20:15:55.512] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [240,000] records [2009-07-06 20:16:03.961] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [250,000] records [2009-07-06 20:16:12.933] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [260,000] records [2009-07-06 20:16:21.934] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [270,000] records [2009-07-06 20:16:30.435] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [280,000] records [2009-07-06 20:16:39.882] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [290,000] records [2009-07-06 20:16:44.573] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV] completed. # of records processed : [294,786]. Start Time : [2009-07-06 20:12:23], End Time : [2009-07-06 20:16:44] [qw...@app48 logs]$ ----- Original Message ----- From: "Irfan Mohammed" <irfan...@gmail.com> To: hbase-dev@hadoop.apache.org Sent: Monday, July 6, 2009 7:51:00 PM GMT -05:00 US/Canada Eastern Subject: Re: performance help added more instrumentation. it is taking about 2 minutes per 10k records and for 300k records it will take 60 minutes. :-( [qw...@app48 logs]$ ~/scripts/loadDirect.sh [2009-07-06 19:29:20.465] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV] ... [2009-07-06 19:29:21.820] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [100] records [2009-07-06 19:29:23.372] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [200] records [2009-07-06 19:29:24.567] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [300] records [2009-07-06 19:29:25.157] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [400] records [2009-07-06 19:29:26.178] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [500] records [2009-07-06 19:29:27.096] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [600] records [2009-07-06 19:29:28.249] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [700] records [2009-07-06 19:29:28.258] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [800] records [2009-07-06 19:29:28.267] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [900] records [2009-07-06 19:29:28.276] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [1,000] records [2009-07-06 19:29:29.406] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [1,100] records [2009-07-06 19:29:30.094] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [1,200] records [2009-07-06 19:29:30.903] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [1,300] records [2009-07-06 19:29:32.158] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [1,400] records [2009-07-06 19:29:33.483] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [1,500] records [2009-07-06 19:29:34.187] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [1,600] records [2009-07-06 19:29:35.515] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [1,700] records [2009-07-06 19:29:36.610] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [1,800] records [2009-07-06 19:29:37.758] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [1,900] records [2009-07-06 19:29:39.173] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [2,000] records [2009-07-06 19:29:40.443] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [2,100] records [2009-07-06 19:29:41.848] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [2,200] records [2009-07-06 19:29:42.256] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [2,300] records [2009-07-06 19:29:43.520] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [2,400] records [2009-07-06 19:29:44.906] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [2,500] records [2009-07-06 19:29:46.191] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [2,600] records [2009-07-06 19:29:47.502] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [2,700] records [2009-07-06 19:29:48.810] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [2,800] records [2009-07-06 19:29:50.275] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [2,900] records [2009-07-06 19:29:51.579] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [3,000] records [2009-07-06 19:29:52.879] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [3,100] records [2009-07-06 19:29:54.207] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [3,200] records [2009-07-06 19:29:55.619] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [3,300] records [2009-07-06 19:29:56.901] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [3,400] records [2009-07-06 19:29:58.183] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [3,500] records [2009-07-06 19:29:59.555] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [3,600] records [2009-07-06 19:30:00.838] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [3,700] records [2009-07-06 19:30:02.232] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [3,800] records [2009-07-06 19:31:18.371] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [9,900] records [2009-07-06 19:31:19.672] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV]. [10,000] records ----- Original Message ----- From: "Irfan Mohammed" <irfan...@gmail.com> To: hbase-dev@hadoop.apache.org Sent: Monday, July 6, 2009 6:42:10 PM GMT -05:00 US/Canada Eastern Subject: Re: performance help converted the code to just directly use HBase Client API without the M/R framework and the results are interesting ... 1. initially did not use "HTable.incrementColumnValue" and just used "HTable.put" and the process ran in ~5 minutes. 2. after switching to "HTable.incrementColumnValue" it is still running and is about ~30 minutes into the run. I issued couple of "kill -QUIT" to see if the process is moving ahead and looks like it is since the lock object is changing each time. >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> HTable.Put >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> [qw...@app48 transaction_ar20090706_1459.CSV]$ ~/scripts/loadDirect.sh 09/07/06 17:58:43 INFO zookeeper.ZooKeeperWrapper: Quorum servers: app16.qwapi.com:2181,app48.qwapi.com:2181,app122.qwapi.com:2181 09/07/06 17:58:43 INFO zookeeper.ZooKeeper: Client environment:zookeeper.version=3.2.0--1, built on 05/15/2009 06:05 GMT 09/07/06 17:58:43 INFO zookeeper.ZooKeeper: Client environment:host.name=app48 09/07/06 17:58:43 INFO zookeeper.ZooKeeper: Client environment:java.version=1.6.0_13 09/07/06 17:58:43 INFO zookeeper.ZooKeeper: Client environment:java.vendor=Sun Microsystems Inc. 09/07/06 17:58:43 INFO zookeeper.ZooKeeper: Client environment:java.home=/usr/java/jdk1.6.0_13/jre 09/07/06 17:58:43 INFO zookeeper.ZooKeeper: Client environment:java.class.path=/home/qwapi/apps/hbase-latest/lib/zookeeper-r785019-hbase-1329.jar:/home/qwapi/apps/hbase-latest/lib/xmlenc-0.52.jar:/home/qwapi/apps/hbase-latest/lib/servlet-api-2.5-6.1.14.jar:/home/qwapi/apps/hbase-latest/lib/lucene-core-2.2.0.jar:/home/qwapi/apps/hbase-latest/lib/log4j-1.2.15.jar:/home/qwapi/apps/hbase-latest/lib/libthrift-r771587.jar:/home/qwapi/apps/hbase-latest/lib/junit-3.8.1.jar:/home/qwapi/apps/hbase-latest/lib/json.jar:/home/qwapi/apps/hbase-latest/lib/jruby-complete-1.2.0.jar:/home/qwapi/apps/hbase-latest/lib/jetty-util-6.1.14.jar:/home/qwapi/apps/hbase-latest/lib/jetty-6.1.14.jar:/home/qwapi/apps/hbase-latest/lib/jasper-runtime-5.5.12.jar:/home/qwapi/apps/hbase-latest/lib/jasper-compiler-5.5.12.jar:/home/qwapi/apps/hbase-latest/lib/hadoop-0.20.0-test.jar:/home/qwapi/apps/hbase-latest/lib/hadoop-0.20.0-plus4681-core.jar:/home/qwapi/apps/hbase-latest/lib/commons-math-1.1.jar:/home/qwapi/apps/hbase-latest/lib/commons-logging-api-1.0.4.jar:/home/qwapi/apps/hbase-latest/lib/commons-logging-1.0.4.jar:/home/qwapi/apps/hbase-latest/lib/commons-httpclient-3.0.1.jar:/home/qwapi/apps/hbase-latest/lib/commons-el-from-jetty-5.1.4.jar:/home/qwapi/apps/hbase-latest/lib/commons-cli-2.0-SNAPSHOT.jar:/home/qwapi/apps/hbase-latest/lib/AgileJSON-2009-03-30.jar:/home/qwapi/apps/hbase-latest/conf:/home/qwapi/apps/hadoop-latest/hadoop-0.20.0-core.jar:/home/qwapi/apps/hbase-latest/hbase-0.20.0-dev.jar:/home/qwapi/apps/hbase-latest/lib/zookeeper-r785019-hbase-1329.jar:/home/qwapi/txnload/bin/load_direct.jar 09/07/06 17:58:43 INFO zookeeper.ZooKeeper: Client environment:java.library.path=/usr/java/jdk1.6.0_13/jre/lib/i386/server:/usr/java/jdk1.6.0_13/jre/lib/i386:/usr/java/jdk1.6.0_13/jre/../lib/i386:/usr/java/packages/lib/i386:/lib:/usr/lib 09/07/06 17:58:43 INFO zookeeper.ZooKeeper: Client environment:java.io.tmpdir=/tmp 09/07/06 17:58:43 INFO zookeeper.ZooKeeper: Client environment:java.compiler=<NA> 09/07/06 17:58:43 INFO zookeeper.ZooKeeper: Client environment:os.name=Linux 09/07/06 17:58:43 INFO zookeeper.ZooKeeper: Client environment:os.arch=i386 09/07/06 17:58:43 INFO zookeeper.ZooKeeper: Client environment:os.version=2.6.9-67.0.20.ELsmp 09/07/06 17:58:43 INFO zookeeper.ZooKeeper: Client environment:user.name=qwapi 09/07/06 17:58:43 INFO zookeeper.ZooKeeper: Client environment:user.home=/home/qwapi 09/07/06 17:58:43 INFO zookeeper.ZooKeeper: Client environment:user.dir=/home/qwapi/tmp/transaction_ar20090706_1459.CSV 09/07/06 17:58:43 INFO zookeeper.ZooKeeper: Initiating client connection, host=app16.qwapi.com:2181,app48.qwapi.com:2181,app122.qwapi.com:2181 sessionTimeout=10000 watcher=org.apache.hadoop.hbase.zookeeper.watcherwrap...@fbb7cb 09/07/06 17:58:43 INFO zookeeper.ClientCnxn: zookeeper.disableAutoWatchReset is false 09/07/06 17:58:43 INFO zookeeper.ClientCnxn: Attempting connection to server app122.qwapi.com/10.10.0.122:2181 09/07/06 17:58:43 INFO zookeeper.ClientCnxn: Priming connection to java.nio.channels.SocketChannel[connected local=/10.10.0.48:35809 remote=app122.qwapi.com/10.10.0.122:2181] 09/07/06 17:58:43 INFO zookeeper.ClientCnxn: Server connection successful [2009-07-06 17:58:43.425] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV] ... [2009-07-06 18:03:46.104] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV] completed. # of records processed : [294,786] >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> HTable.Put >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> HTable.incrementColumnValue >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> [qw...@app48 transaction_ar20090706_1459.CSV]$ ~/scripts/loadDirect.sh 09/07/06 18:07:12 INFO zookeeper.ZooKeeperWrapper: Quorum servers: app16.qwapi.com:2181,app48.qwapi.com:2181,app122.qwapi.com:2181 09/07/06 18:07:12 INFO zookeeper.ZooKeeper: Client environment:zookeeper.version=3.2.0--1, built on 05/15/2009 06:05 GMT 09/07/06 18:07:12 INFO zookeeper.ZooKeeper: Client environment:host.name=app48 09/07/06 18:07:12 INFO zookeeper.ZooKeeper: Client environment:java.version=1.6.0_13 09/07/06 18:07:12 INFO zookeeper.ZooKeeper: Client environment:java.vendor=Sun Microsystems Inc. 09/07/06 18:07:12 INFO zookeeper.ZooKeeper: Client environment:java.home=/usr/java/jdk1.6.0_13/jre 09/07/06 18:07:12 INFO zookeeper.ZooKeeper: Client environment:java.class.path=/home/qwapi/apps/hbase-latest/lib/zookeeper-r785019-hbase-1329.jar:/home/qwapi/apps/hbase-latest/lib/xmlenc-0.52.jar:/home/qwapi/apps/hbase-latest/lib/servlet-api-2.5-6.1.14.jar:/home/qwapi/apps/hbase-latest/lib/lucene-core-2.2.0.jar:/home/qwapi/apps/hbase-latest/lib/log4j-1.2.15.jar:/home/qwapi/apps/hbase-latest/lib/libthrift-r771587.jar:/home/qwapi/apps/hbase-latest/lib/junit-3.8.1.jar:/home/qwapi/apps/hbase-latest/lib/json.jar:/home/qwapi/apps/hbase-latest/lib/jruby-complete-1.2.0.jar:/home/qwapi/apps/hbase-latest/lib/jetty-util-6.1.14.jar:/home/qwapi/apps/hbase-latest/lib/jetty-6.1.14.jar:/home/qwapi/apps/hbase-latest/lib/jasper-runtime-5.5.12.jar:/home/qwapi/apps/hbase-latest/lib/jasper-compiler-5.5.12.jar:/home/qwapi/apps/hbase-latest/lib/hadoop-0.20.0-test.jar:/home/qwapi/apps/hbase-latest/lib/hadoop-0.20.0-plus4681-core.jar:/home/qwapi/apps/hbase-latest/lib/commons-math-1.1.jar:/home/qwapi/apps/hbase-latest/lib/commons-logging-api-1.0.4.jar:/home/qwapi/apps/hbase-latest/lib/commons-logging-1.0.4.jar:/home/qwapi/apps/hbase-latest/lib/commons-httpclient-3.0.1.jar:/home/qwapi/apps/hbase-latest/lib/commons-el-from-jetty-5.1.4.jar:/home/qwapi/apps/hbase-latest/lib/commons-cli-2.0-SNAPSHOT.jar:/home/qwapi/apps/hbase-latest/lib/AgileJSON-2009-03-30.jar:/home/qwapi/apps/hbase-latest/conf:/home/qwapi/apps/hadoop-latest/hadoop-0.20.0-core.jar:/home/qwapi/apps/hbase-latest/hbase-0.20.0-dev.jar:/home/qwapi/apps/hbase-latest/lib/zookeeper-r785019-hbase-1329.jar:/home/qwapi/txnload/bin/load_direct.jar 09/07/06 18:07:12 INFO zookeeper.ZooKeeper: Client environment:java.library.path=/usr/java/jdk1.6.0_13/jre/lib/i386/server:/usr/java/jdk1.6.0_13/jre/lib/i386:/usr/java/jdk1.6.0_13/jre/../lib/i386:/usr/java/packages/lib/i386:/lib:/usr/lib 09/07/06 18:07:12 INFO zookeeper.ZooKeeper: Client environment:java.io.tmpdir=/tmp 09/07/06 18:07:12 INFO zookeeper.ZooKeeper: Client environment:java.compiler=<NA> 09/07/06 18:07:12 INFO zookeeper.ZooKeeper: Client environment:os.name=Linux 09/07/06 18:07:12 INFO zookeeper.ZooKeeper: Client environment:os.arch=i386 09/07/06 18:07:12 INFO zookeeper.ZooKeeper: Client environment:os.version=2.6.9-67.0.20.ELsmp 09/07/06 18:07:12 INFO zookeeper.ZooKeeper: Client environment:user.name=qwapi 09/07/06 18:07:12 INFO zookeeper.ZooKeeper: Client environment:user.home=/home/qwapi 09/07/06 18:07:12 INFO zookeeper.ZooKeeper: Client environment:user.dir=/home/qwapi/tmp/transaction_ar20090706_1459.CSV 09/07/06 18:07:12 INFO zookeeper.ZooKeeper: Initiating client connection, host=app16.qwapi.com:2181,app48.qwapi.com:2181,app122.qwapi.com:2181 sessionTimeout=10000 watcher=org.apache.hadoop.hbase.zookeeper.watcherwrap...@fbb7cb 09/07/06 18:07:12 INFO zookeeper.ClientCnxn: zookeeper.disableAutoWatchReset is false 09/07/06 18:07:12 INFO zookeeper.ClientCnxn: Attempting connection to server app122.qwapi.com/10.10.0.122:2181 09/07/06 18:07:12 INFO zookeeper.ClientCnxn: Priming connection to java.nio.channels.SocketChannel[connected local=/10.10.0.48:36147 remote=app122.qwapi.com/10.10.0.122:2181] 09/07/06 18:07:12 INFO zookeeper.ClientCnxn: Server connection successful [2009-07-06 18:07:12.735] processing file : [/home/qwapi/tmp/transaction_ar20090706_1459.CSV/transaction_ar20090706_1459.CSV] ... 2009-07-06 18:23:24 Full thread dump Java HotSpot(TM) Server VM (11.3-b02 mixed mode): "IPC Client (47) connection to /10.10.0.163:60020 from an unknown user" daemon prio=10 tid=0xafa1d000 nid=0xd5c runnable [0xaf8ac000..0xaf8ad0b0] java.lang.Thread.State: RUNNABLE at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:215) at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:65) at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:69) - locked <0xb4e9b810> (a sun.nio.ch.Util$1) - locked <0xb4e9b800> (a java.util.Collections$UnmodifiableSet) - locked <0xb4e9b5f8> (a sun.nio.ch.EPollSelectorImpl) at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:80) at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:332) at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:157) at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:155) at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:128) at java.io.FilterInputStream.read(FilterInputStream.java:116) at org.apache.hadoop.hbase.ipc.HBaseClient$Connection$PingInputStream.read(HBaseClient.java:277) at java.io.BufferedInputStream.fill(BufferedInputStream.java:218) at java.io.BufferedInputStream.read(BufferedInputStream.java:237) - locked <0xb4e350c8> (a java.io.BufferedInputStream) at java.io.DataInputStream.readInt(DataInputStream.java:370) at org.apache.hadoop.hbase.ipc.HBaseClient$Connection.receiveResponse(HBaseClient.java:501) at org.apache.hadoop.hbase.ipc.HBaseClient$Connection.run(HBaseClient.java:445) "main-EventThread" daemon prio=10 tid=0x085aec00 nid=0xd59 waiting on condition [0xaf9ad000..0xaf9ade30] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <0xb4e00230> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:158) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1925) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:358) at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:376) "main-SendThread" daemon prio=10 tid=0x08533800 nid=0xd58 runnable [0xaf9fe000..0xaf9feeb0] java.lang.Thread.State: RUNNABLE at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:215) at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:65) at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:69) - locked <0xb4e01130> (a sun.nio.ch.Util$1) - locked <0xb4e01120> (a java.util.Collections$UnmodifiableSet) - locked <0xb4e010e0> (a sun.nio.ch.EPollSelectorImpl) at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:80) at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:873) "Low Memory Detector" daemon prio=10 tid=0x08163800 nid=0xd56 runnable [0x00000000..0x00000000] java.lang.Thread.State: RUNNABLE "CompilerThread1" daemon prio=10 tid=0x08161800 nid=0xd55 waiting on condition [0x00000000..0xafe444e8] java.lang.Thread.State: RUNNABLE "CompilerThread0" daemon prio=10 tid=0x0815d400 nid=0xd54 waiting on condition [0x00000000..0xafec5568] java.lang.Thread.State: RUNNABLE "Signal Dispatcher" daemon prio=10 tid=0x0815b800 nid=0xd53 waiting on condition [0x00000000..0x00000000] java.lang.Thread.State: RUNNABLE "Finalizer" daemon prio=10 tid=0x08148400 nid=0xd52 in Object.wait() [0xb0167000..0xb0167fb0] java.lang.Thread.State: WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <0xb4e030f8> (a java.lang.ref.ReferenceQueue$Lock) at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:116) - locked <0xb4e030f8> (a java.lang.ref.ReferenceQueue$Lock) at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:132) at java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:159) "Reference Handler" daemon prio=10 tid=0x08146c00 nid=0xd51 in Object.wait() [0xb01b8000..0xb01b8e30] java.lang.Thread.State: WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <0xb4e011a8> (a java.lang.ref.Reference$Lock) at java.lang.Object.wait(Object.java:485) at java.lang.ref.Reference$ReferenceHandler.run(Reference.java:116) - locked <0xb4e011a8> (a java.lang.ref.Reference$Lock) "main" prio=10 tid=0x08059c00 nid=0xd47 in Object.wait() [0xf7fc0000..0xf7fc1278] java.lang.Thread.State: WAITING (on object monitor) at java.lang.Object.wait(Native Method) at java.lang.Object.wait(Object.java:485) at org.apache.hadoop.hbase.ipc.HBaseClient.call(HBaseClient.java:712) - locked <0xedf2a8c8> (a org.apache.hadoop.hbase.ipc.HBaseClient$Call) at org.apache.hadoop.hbase.ipc.HBaseRPC$Invoker.invoke(HBaseRPC.java:321) at $Proxy0.incrementColumnValue(Unknown Source) at org.apache.hadoop.hbase.client.HTable$6.call(HTable.java:504) at org.apache.hadoop.hbase.client.HTable$6.call(HTable.java:500) at org.apache.hadoop.hbase.client.HConnectionManager$TableServers.getRegionServerWithRetries(HConnectionManager.java:922) at org.apache.hadoop.hbase.client.HTable.incrementColumnValue(HTable.java:499) at com.qwapi.txnload.LoadDirect.loadRow(LoadDirect.java:157) at com.qwapi.txnload.LoadDirect.loadFile(LoadDirect.java:95) at com.qwapi.txnload.LoadDirect.main(LoadDirect.java:182) "VM Thread" prio=10 tid=0x08143400 nid=0xd50 runnable "GC task thread#0 (ParallelGC)" prio=10 tid=0x08060c00 nid=0xd48 runnable "GC task thread#1 (ParallelGC)" prio=10 tid=0x08062000 nid=0xd49 runnable "GC task thread#2 (ParallelGC)" prio=10 tid=0x08063800 nid=0xd4a runnable "GC task thread#3 (ParallelGC)" prio=10 tid=0x08065000 nid=0xd4b runnable "GC task thread#4 (ParallelGC)" prio=10 tid=0x08066400 nid=0xd4c runnable "GC task thread#5 (ParallelGC)" prio=10 tid=0x08067c00 nid=0xd4d runnable "GC task thread#6 (ParallelGC)" prio=10 tid=0x08069000 nid=0xd4e runnable "GC task thread#7 (ParallelGC)" prio=10 tid=0x0806a800 nid=0xd4f runnable "VM Periodic Task Thread" prio=10 tid=0x08165400 nid=0xd57 waiting on condition JNI global references: 895 Heap PSYoungGen total 14080K, used 3129K [0xedc40000, 0xeea10000, 0xf4e00000) eden space 14016K, 22% used [0xedc40000,0xedf4a4b0,0xee9f0000) from space 64K, 25% used [0xeea00000,0xeea04000,0xeea10000) to space 64K, 0% used [0xee9f0000,0xee9f0000,0xeea00000) PSOldGen total 113472K, used 1795K [0xb4e00000, 0xbbcd0000, 0xedc40000) object space 113472K, 1% used [0xb4e00000,0xb4fc0d00,0xbbcd0000) PSPermGen total 16384K, used 6188K [0xb0e00000, 0xb1e00000, 0xb4e00000) object space 16384K, 37% used [0xb0e00000,0xb140b230,0xb1e00000) 2009-07-06 18:24:59 Full thread dump Java HotSpot(TM) Server VM (11.3-b02 mixed mode): "IPC Client (47) connection to /10.10.0.163:60020 from an unknown user" daemon prio=10 tid=0xafa1d000 nid=0xd5c in Object.wait() [0xaf8ac000..0xaf8ad0b0] java.lang.Thread.State: TIMED_WAITING (on object monitor) at java.lang.Object.wait(Native Method) at org.apache.hadoop.hbase.ipc.HBaseClient$Connection.waitForWork(HBaseClient.java:401) - locked <0xb4e00090> (a org.apache.hadoop.hbase.ipc.HBaseClient$Connection) at org.apache.hadoop.hbase.ipc.HBaseClient$Connection.run(HBaseClient.java:444) "main-EventThread" daemon prio=10 tid=0x085aec00 nid=0xd59 waiting on condition [0xaf9ad000..0xaf9ade30] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <0xb4e00230> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:158) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1925) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:358) at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:376) "main-SendThread" daemon prio=10 tid=0x08533800 nid=0xd58 runnable [0xaf9fe000..0xaf9feeb0] java.lang.Thread.State: RUNNABLE at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:215) at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:65) at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:69) - locked <0xb4e01130> (a sun.nio.ch.Util$1) - locked <0xb4e01120> (a java.util.Collections$UnmodifiableSet) - locked <0xb4e010e0> (a sun.nio.ch.EPollSelectorImpl) at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:80) at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:873) "Low Memory Detector" daemon prio=10 tid=0x08163800 nid=0xd56 runnable [0x00000000..0x00000000] java.lang.Thread.State: RUNNABLE "CompilerThread1" daemon prio=10 tid=0x08161800 nid=0xd55 waiting on condition [0x00000000..0xafe444e8] java.lang.Thread.State: RUNNABLE "CompilerThread0" daemon prio=10 tid=0x0815d400 nid=0xd54 waiting on condition [0x00000000..0xafec5568] java.lang.Thread.State: RUNNABLE "Signal Dispatcher" daemon prio=10 tid=0x0815b800 nid=0xd53 waiting on condition [0x00000000..0x00000000] java.lang.Thread.State: RUNNABLE "Finalizer" daemon prio=10 tid=0x08148400 nid=0xd52 in Object.wait() [0xb0167000..0xb0167fb0] java.lang.Thread.State: WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <0xb4e030f8> (a java.lang.ref.ReferenceQueue$Lock) at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:116) - locked <0xb4e030f8> (a java.lang.ref.ReferenceQueue$Lock) at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:132) at java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:159) "Reference Handler" daemon prio=10 tid=0x08146c00 nid=0xd51 in Object.wait() [0xb01b8000..0xb01b8e30] java.lang.Thread.State: WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <0xb4e011a8> (a java.lang.ref.Reference$Lock) at java.lang.Object.wait(Object.java:485) at java.lang.ref.Reference$ReferenceHandler.run(Reference.java:116) - locked <0xb4e011a8> (a java.lang.ref.Reference$Lock) "main" prio=10 tid=0x08059c00 nid=0xd47 in Object.wait() [0xf7fc0000..0xf7fc1278] java.lang.Thread.State: BLOCKED (on object monitor) at java.lang.Object.wait(Native Method) at java.lang.Object.wait(Object.java:485) at org.apache.hadoop.hbase.ipc.HBaseClient.call(HBaseClient.java:712) - locked <0xee5ecb50> (a org.apache.hadoop.hbase.ipc.HBaseClient$Call) at org.apache.hadoop.hbase.ipc.HBaseRPC$Invoker.invoke(HBaseRPC.java:321) at $Proxy0.incrementColumnValue(Unknown Source) at org.apache.hadoop.hbase.client.HTable$6.call(HTable.java:504) at org.apache.hadoop.hbase.client.HTable$6.call(HTable.java:500) at org.apache.hadoop.hbase.client.HConnectionManager$TableServers.getRegionServerWithRetries(HConnectionManager.java:922) at org.apache.hadoop.hbase.client.HTable.incrementColumnValue(HTable.java:499) at com.qwapi.txnload.LoadDirect.loadRow(LoadDirect.java:157) at com.qwapi.txnload.LoadDirect.loadFile(LoadDirect.java:95) at com.qwapi.txnload.LoadDirect.main(LoadDirect.java:182) "VM Thread" prio=10 tid=0x08143400 nid=0xd50 runnable "GC task thread#0 (ParallelGC)" prio=10 tid=0x08060c00 nid=0xd48 runnable "GC task thread#1 (ParallelGC)" prio=10 tid=0x08062000 nid=0xd49 runnable "GC task thread#2 (ParallelGC)" prio=10 tid=0x08063800 nid=0xd4a runnable "GC task thread#3 (ParallelGC)" prio=10 tid=0x08065000 nid=0xd4b runnable "GC task thread#4 (ParallelGC)" prio=10 tid=0x08066400 nid=0xd4c runnable "GC task thread#5 (ParallelGC)" prio=10 tid=0x08067c00 nid=0xd4d runnable "GC task thread#6 (ParallelGC)" prio=10 tid=0x08069000 nid=0xd4e runnable "GC task thread#7 (ParallelGC)" prio=10 tid=0x0806a800 nid=0xd4f runnable "VM Periodic Task Thread" prio=10 tid=0x08165400 nid=0xd57 waiting on condition JNI global references: 895 Heap PSYoungGen total 14080K, used 10004K [0xedc40000, 0xeea10000, 0xf4e00000) eden space 14016K, 71% used [0xedc40000,0xee601028,0xee9f0000) from space 64K, 25% used [0xeea00000,0xeea04000,0xeea10000) to space 64K, 0% used [0xee9f0000,0xee9f0000,0xeea00000) PSOldGen total 113472K, used 1907K [0xb4e00000, 0xbbcd0000, 0xedc40000) object space 113472K, 1% used [0xb4e00000,0xb4fdcd00,0xbbcd0000) PSPermGen total 16384K, used 6188K [0xb0e00000, 0xb1e00000, 0xb4e00000) object space 16384K, 37% used [0xb0e00000,0xb140b230,0xb1e00000) 2009-07-06 18:30:39 Full thread dump Java HotSpot(TM) Server VM (11.3-b02 mixed mode): "IPC Client (47) connection to /10.10.0.163:60020 from an unknown user" daemon prio=10 tid=0xafa1d000 nid=0xd5c runnable [0xaf8ac000..0xaf8ad0b0] java.lang.Thread.State: RUNNABLE at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:215) at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:65) at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:69) - locked <0xb4e9b810> (a sun.nio.ch.Util$1) - locked <0xb4e9b800> (a java.util.Collections$UnmodifiableSet) - locked <0xb4e9b5f8> (a sun.nio.ch.EPollSelectorImpl) at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:80) at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:332) at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:157) at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:155) at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:128) at java.io.FilterInputStream.read(FilterInputStream.java:116) at org.apache.hadoop.hbase.ipc.HBaseClient$Connection$PingInputStream.read(HBaseClient.java:277) at java.io.BufferedInputStream.fill(BufferedInputStream.java:218) at java.io.BufferedInputStream.read(BufferedInputStream.java:237) - locked <0xb4e350c8> (a java.io.BufferedInputStream) at java.io.DataInputStream.readInt(DataInputStream.java:370) at org.apache.hadoop.hbase.ipc.HBaseClient$Connection.receiveResponse(HBaseClient.java:501) at org.apache.hadoop.hbase.ipc.HBaseClient$Connection.run(HBaseClient.java:445) "main-EventThread" daemon prio=10 tid=0x085aec00 nid=0xd59 waiting on condition [0xaf9ad000..0xaf9ade30] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <0xb4e00230> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:158) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1925) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:358) at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:376) "main-SendThread" daemon prio=10 tid=0x08533800 nid=0xd58 runnable [0xaf9fe000..0xaf9feeb0] java.lang.Thread.State: RUNNABLE at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:215) at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:65) at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:69) - locked <0xb4e01130> (a sun.nio.ch.Util$1) - locked <0xb4e01120> (a java.util.Collections$UnmodifiableSet) - locked <0xb4e010e0> (a sun.nio.ch.EPollSelectorImpl) at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:80) at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:873) "Low Memory Detector" daemon prio=10 tid=0x08163800 nid=0xd56 runnable [0x00000000..0x00000000] java.lang.Thread.State: RUNNABLE "CompilerThread1" daemon prio=10 tid=0x08161800 nid=0xd55 waiting on condition [0x00000000..0xafe444e8] java.lang.Thread.State: RUNNABLE "CompilerThread0" daemon prio=10 tid=0x0815d400 nid=0xd54 waiting on condition [0x00000000..0xafec5568] java.lang.Thread.State: RUNNABLE "Signal Dispatcher" daemon prio=10 tid=0x0815b800 nid=0xd53 waiting on condition [0x00000000..0x00000000] java.lang.Thread.State: RUNNABLE "Finalizer" daemon prio=10 tid=0x08148400 nid=0xd52 in Object.wait() [0xb0167000..0xb0167fb0] java.lang.Thread.State: WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <0xb4e030f8> (a java.lang.ref.ReferenceQueue$Lock) at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:116) - locked <0xb4e030f8> (a java.lang.ref.ReferenceQueue$Lock) at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:132) at java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:159) "Reference Handler" daemon prio=10 tid=0x08146c00 nid=0xd51 in Object.wait() [0xb01b8000..0xb01b8e30] java.lang.Thread.State: WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <0xb4e011a8> (a java.lang.ref.Reference$Lock) at java.lang.Object.wait(Object.java:485) at java.lang.ref.Reference$ReferenceHandler.run(Reference.java:116) - locked <0xb4e011a8> (a java.lang.ref.Reference$Lock) "main" prio=10 tid=0x08059c00 nid=0xd47 in Object.wait() [0xf7fc0000..0xf7fc1278] java.lang.Thread.State: WAITING (on object monitor) at java.lang.Object.wait(Native Method) at java.lang.Object.wait(Object.java:485) at org.apache.hadoop.hbase.ipc.HBaseClient.call(HBaseClient.java:712) - locked <0xee61dfe8> (a org.apache.hadoop.hbase.ipc.HBaseClient$Call) at org.apache.hadoop.hbase.ipc.HBaseRPC$Invoker.invoke(HBaseRPC.java:321) at $Proxy0.incrementColumnValue(Unknown Source) at org.apache.hadoop.hbase.client.HTable$6.call(HTable.java:504) at org.apache.hadoop.hbase.client.HTable$6.call(HTable.java:500) at org.apache.hadoop.hbase.client.HConnectionManager$TableServers.getRegionServerWithRetries(HConnectionManager.java:922) at org.apache.hadoop.hbase.client.HTable.incrementColumnValue(HTable.java:499) at com.qwapi.txnload.LoadDirect.loadRow(LoadDirect.java:157) at com.qwapi.txnload.LoadDirect.loadFile(LoadDirect.java:95) at com.qwapi.txnload.LoadDirect.main(LoadDirect.java:182) "VM Thread" prio=10 tid=0x08143400 nid=0xd50 runnable "GC task thread#0 (ParallelGC)" prio=10 tid=0x08060c00 nid=0xd48 runnable "GC task thread#1 (ParallelGC)" prio=10 tid=0x08062000 nid=0xd49 runnable "GC task thread#2 (ParallelGC)" prio=10 tid=0x08063800 nid=0xd4a runnable "GC task thread#3 (ParallelGC)" prio=10 tid=0x08065000 nid=0xd4b runnable "GC task thread#4 (ParallelGC)" prio=10 tid=0x08066400 nid=0xd4c runnable "GC task thread#5 (ParallelGC)" prio=10 tid=0x08067c00 nid=0xd4d runnable "GC task thread#6 (ParallelGC)" prio=10 tid=0x08069000 nid=0xd4e runnable "GC task thread#7 (ParallelGC)" prio=10 tid=0x0806a800 nid=0xd4f runnable "VM Periodic Task Thread" prio=10 tid=0x08165400 nid=0xd57 waiting on condition JNI global references: 895 Heap PSYoungGen total 14080K, used 10281K [0xedc40000, 0xeea10000, 0xf4e00000) eden space 14016K, 73% used [0xedc40000,0xee6464f0,0xee9f0000) from space 64K, 25% used [0xee9f0000,0xee9f4000,0xeea00000) to space 64K, 0% used [0xeea00000,0xeea00000,0xeea10000) PSOldGen total 113472K, used 2315K [0xb4e00000, 0xbbcd0000, 0xedc40000) object space 113472K, 2% used [0xb4e00000,0xb5042d00,0xbbcd0000) PSPermGen total 16384K, used 6188K [0xb0e00000, 0xb1e00000, 0xb4e00000) object space 16384K, 37% used [0xb0e00000,0xb140b230,0xb1e00000) 2009-07-06 18:31:13 Full thread dump Java HotSpot(TM) Server VM (11.3-b02 mixed mode): "IPC Client (47) connection to /10.10.0.163:60020 from an unknown user" daemon prio=10 tid=0xafa1d000 nid=0xd5c runnable [0xaf8ac000..0xaf8ad0b0] java.lang.Thread.State: RUNNABLE at java.lang.Class.forName0(Native Method) at java.lang.Class.forName(Class.java:247) at org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:761) at org.apache.hadoop.util.ReflectionUtils.setJobConf(ReflectionUtils.java:80) at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:64) at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:117) at org.apache.hadoop.hbase.ipc.HBaseClient$Connection.receiveResponse(HBaseClient.java:513) at org.apache.hadoop.hbase.ipc.HBaseClient$Connection.run(HBaseClient.java:445) "main-EventThread" daemon prio=10 tid=0x085aec00 nid=0xd59 waiting on condition [0xaf9ad000..0xaf9ade30] java.lang.Thread.State: WAITING (parking) at sun.misc.Unsafe.park(Native Method) - parking to wait for <0xb4e00230> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:158) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1925) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:358) at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:376) "main-SendThread" daemon prio=10 tid=0x08533800 nid=0xd58 runnable [0xaf9fe000..0xaf9feeb0] java.lang.Thread.State: RUNNABLE at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:215) at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:65) at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:69) - locked <0xb4e01130> (a sun.nio.ch.Util$1) - locked <0xb4e01120> (a java.util.Collections$UnmodifiableSet) - locked <0xb4e010e0> (a sun.nio.ch.EPollSelectorImpl) at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:80) at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:873) "Low Memory Detector" daemon prio=10 tid=0x08163800 nid=0xd56 runnable [0x00000000..0x00000000] java.lang.Thread.State: RUNNABLE "CompilerThread1" daemon prio=10 tid=0x08161800 nid=0xd55 waiting on condition [0x00000000..0xafe444e8] java.lang.Thread.State: RUNNABLE "CompilerThread0" daemon prio=10 tid=0x0815d400 nid=0xd54 waiting on condition [0x00000000..0xafec5568] java.lang.Thread.State: RUNNABLE "Signal Dispatcher" daemon prio=10 tid=0x0815b800 nid=0xd53 waiting on condition [0x00000000..0x00000000] java.lang.Thread.State: RUNNABLE "Finalizer" daemon prio=10 tid=0x08148400 nid=0xd52 in Object.wait() [0xb0167000..0xb0167fb0] java.lang.Thread.State: WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <0xb4e030f8> (a java.lang.ref.ReferenceQueue$Lock) at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:116) - locked <0xb4e030f8> (a java.lang.ref.ReferenceQueue$Lock) at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:132) at java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:159) "Reference Handler" daemon prio=10 tid=0x08146c00 nid=0xd51 in Object.wait() [0xb01b8000..0xb01b8e30] java.lang.Thread.State: WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <0xb4e011a8> (a java.lang.ref.Reference$Lock) at java.lang.Object.wait(Object.java:485) at java.lang.ref.Reference$ReferenceHandler.run(Reference.java:116) - locked <0xb4e011a8> (a java.lang.ref.Reference$Lock) "main" prio=10 tid=0x08059c00 nid=0xd47 in Object.wait() [0xf7fc0000..0xf7fc1278] java.lang.Thread.State: WAITING (on object monitor) at java.lang.Object.wait(Native Method) at java.lang.Object.wait(Object.java:485) at org.apache.hadoop.hbase.ipc.HBaseClient.call(HBaseClient.java:712) - locked <0xedd8dec0> (a org.apache.hadoop.hbase.ipc.HBaseClient$Call) at org.apache.hadoop.hbase.ipc.HBaseRPC$Invoker.invoke(HBaseRPC.java:321) at $Proxy0.incrementColumnValue(Unknown Source) at org.apache.hadoop.hbase.client.HTable$6.call(HTable.java:504) at org.apache.hadoop.hbase.client.HTable$6.call(HTable.java:500) at org.apache.hadoop.hbase.client.HConnectionManager$TableServers.getRegionServerWithRetries(HConnectionManager.java:922) at org.apache.hadoop.hbase.client.HTable.incrementColumnValue(HTable.java:499) at com.qwapi.txnload.LoadDirect.loadRow(LoadDirect.java:157) at com.qwapi.txnload.LoadDirect.loadFile(LoadDirect.java:95) at com.qwapi.txnload.LoadDirect.main(LoadDirect.java:182) "VM Thread" prio=10 tid=0x08143400 nid=0xd50 runnable "GC task thread#0 (ParallelGC)" prio=10 tid=0x08060c00 nid=0xd48 runnable "GC task thread#1 (ParallelGC)" prio=10 tid=0x08062000 nid=0xd49 runnable "GC task thread#2 (ParallelGC)" prio=10 tid=0x08063800 nid=0xd4a runnable "GC task thread#3 (ParallelGC)" prio=10 tid=0x08065000 nid=0xd4b runnable "GC task thread#4 (ParallelGC)" prio=10 tid=0x08066400 nid=0xd4c runnable "GC task thread#5 (ParallelGC)" prio=10 tid=0x08067c00 nid=0xd4d runnable "GC task thread#6 (ParallelGC)" prio=10 tid=0x08069000 nid=0xd4e runnable "GC task thread#7 (ParallelGC)" prio=10 tid=0x0806a800 nid=0xd4f runnable "VM Periodic Task Thread" prio=10 tid=0x08165400 nid=0xd57 waiting on condition JNI global references: 895 Heap PSYoungGen total 14080K, used 1448K [0xedc40000, 0xeea10000, 0xf4e00000) eden space 14016K, 10% used [0xedc40000,0xedda2018,0xee9f0000) from space 64K, 50% used [0xee9f0000,0xee9f8000,0xeea00000) to space 64K, 0% used [0xeea00000,0xeea00000,0xeea10000) PSOldGen total 113472K, used 2359K [0xb4e00000, 0xbbcd0000, 0xedc40000) object space 113472K, 2% used [0xb4e00000,0xb504dd00,0xbbcd0000) PSPermGen total 16384K, used 6188K [0xb0e00000, 0xb1e00000, 0xb4e00000) object space 16384K, 37% used [0xb0e00000,0xb140b230,0xb1e00000) >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> HTable.incrementColumnValue >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> ----- Original Message ----- From: "Irfan Mohammed" <irfan...@gmail.com> To: hbase-dev@hadoop.apache.org Sent: Monday, July 6, 2009 3:56:57 PM GMT -05:00 US/Canada Eastern Subject: Re: performance help Writing to hdfs directly took just 21 seconds. So I am suspecting that there is something that I am doing incorrectly in my hbase setup or my code. Thanks for the help. [2009-07-06 15:52:47,917] 09/07/06 15:52:22 INFO mapred.FileInputFormat: Total input paths to process : 10 09/07/06 15:52:22 INFO mapred.JobClient: Running job: job_200907052205_0235 09/07/06 15:52:23 INFO mapred.JobClient: map 0% reduce 0% 09/07/06 15:52:37 INFO mapred.JobClient: map 7% reduce 0% 09/07/06 15:52:43 INFO mapred.JobClient: map 100% reduce 0% 09/07/06 15:52:47 INFO mapred.JobClient: Job complete: job_200907052205_0235 09/07/06 15:52:47 INFO mapred.JobClient: Counters: 9 09/07/06 15:52:47 INFO mapred.JobClient: Job Counters 09/07/06 15:52:47 INFO mapred.JobClient: Rack-local map tasks=4 09/07/06 15:52:47 INFO mapred.JobClient: Launched map tasks=10 09/07/06 15:52:47 INFO mapred.JobClient: Data-local map tasks=6 09/07/06 15:52:47 INFO mapred.JobClient: FileSystemCounters 09/07/06 15:52:47 INFO mapred.JobClient: HDFS_BYTES_READ=57966580 09/07/06 15:52:47 INFO mapred.JobClient: HDFS_BYTES_WRITTEN=587539988 09/07/06 15:52:47 INFO mapred.JobClient: Map-Reduce Framework 09/07/06 15:52:47 INFO mapred.JobClient: Map input records=294786 09/07/06 15:52:47 INFO mapred.JobClient: Spilled Records=0 09/07/06 15:52:47 INFO mapred.JobClient: Map input bytes=57966580 09/07/06 15:52:47 INFO mapred.JobClient: Map output records=1160144 ----- Original Message ----- From: "stack" <st...@duboce.net> To: hbase-dev@hadoop.apache.org Sent: Monday, July 6, 2009 2:36:35 PM GMT -05:00 US/Canada Eastern Subject: Re: performance help Sorry, yeah, that'd be 4 tables. So, yeah, it would seem you only have one region in each table. Your cells are small so thats probably about right. So, an hbase client is contacting 4 different servers to do each update. And running with one table made no difference to overall time? St.Ack On Mon, Jul 6, 2009 at 11:24 AM, Irfan Mohammed <irfan...@gmail.com> wrote: > Input is 1 file. > > These are 4 different tables "txn_m1", "txn_m2", "txn_m3", "txn_m4". To me, > it looks like it is always doing 1 region per table and these tables are > always on different regionservers. I never seen the same table on different > regionservers. Does that sound right? > > ----- Original Message ----- > From: "stack" <st...@duboce.net> > To: hbase-dev@hadoop.apache.org > Sent: Monday, July 6, 2009 2:14:43 PM GMT -05:00 US/Canada Eastern > Subject: Re: performance help > > On Mon, Jul 6, 2009 at 11:06 AM, Irfan Mohammed <irfan...@gmail.com> > wrote: > > > I am working on writing to HDFS files. Will update you by end of day > today. > > > > There are always 10 concurrent mappers running. I keep setting the > > setNumMaps(5) and also the following properties in mapred-site.xml to 3 > but > > still end up running 10 concurrent maps. > > > > > Is your input ten files? > > > > > > There are 5 regionservers and the online regions are as follows : > > > > m1 : -ROOT-,,0 > > m2 : txn_m1,,1245462904101 > > m3 : txn_m4,,1245462942282 > > m4 : txn_m2,,1245462890248 > > m5 : .META.,,1 > > txn_m3,,1245460727203 > > > > > So, that looks like 4 regions from table txn? > > So thats about 1 region per regionserver? > > > > I have setAutoFlush(false) and also writeToWal(false) with the same > > behaviour. > > > > If you did above and still takes 10 minutes, then that would seem to rule > out hbase (batching should have big impact on uploads and then setting > writeToWAL to false, should double throughput over whatever you were seeing > previous). > > St.Ack >