Is there a setting to cap row size?
I have a wide table schema for an HBase table, where I model a one-to-many relationship of purchase orders and line items. Each row is a purchase order, and I add columns for each line item. Under normal circumstances I don't expect more than a few thousand columns per row, totalling less than 1mb per row in general. In one of my stress tests, I was inserting many line items into the same row. Eventually, the row's region server shut down. In the log there was an IOException with an error about the wite-ahead log failing to close, with the subsequent regionserver shutdown. Once in this state, the only way I could manage to get my system functional again was to wipe the /hbase directory in HDFS and start from scratch. To avoid having my system susceptible to total data loss due to some bad import data or the like, I'd like to limit the size of the row so it will simply throw an exception if it reaches a certain size (either in bytes or in number of columns). Does such a setting exist?
Re: Is there a setting to cap row size?
Sounds like you are having a HDFS related problem. Check those datanode logs for errors. As for a setting for max row size, this might not be so easy to do, since during the Put time we don't actually know anything about the existing row data. To find that out we'd have to go and read the row first then make a decision. There are some sources for how HBase stores on disk, we are also very similar to the bigtable paper. -ryan On Thu, Apr 7, 2011 at 7:24 AM, Bryan Keller brya...@gmail.com wrote: I have a wide table schema for an HBase table, where I model a one-to-many relationship of purchase orders and line items. Each row is a purchase order, and I add columns for each line item. Under normal circumstances I don't expect more than a few thousand columns per row, totalling less than 1mb per row in general. In one of my stress tests, I was inserting many line items into the same row. Eventually, the row's region server shut down. In the log there was an IOException with an error about the wite-ahead log failing to close, with the subsequent regionserver shutdown. Once in this state, the only way I could manage to get my system functional again was to wipe the /hbase directory in HDFS and start from scratch. To avoid having my system susceptible to total data loss due to some bad import data or the like, I'd like to limit the size of the row so it will simply throw an exception if it reaches a certain size (either in bytes or in number of columns). Does such a setting exist?
timing out for hdfs errors faster
Hello, I get those errors sometimes: 2011-04-07 07:49:41,527 WARN org.apache.hadoop.hdfs.DFSClient: Failed to connect to /10.103.7.5:50010 for file /hbase/media_data/1c95bfcf0dd19800b1f44278627259ae/att/7725092577730365184 for block 802538788372768807:java.net.SocketTimeoutException: 6 millis timeout while waiting for channel to be ready for read. ch : java.nio.channels.SocketChannel[connected local=/10.101.6.8:40801 remote=/10.103.7.5:50010 What would be configuration setting to shorten the timeout say to 5 seconds? What about retries (if any). -Jack
org.apache.hadoop.hbase.ZooKeeperConnectionException: org.apache.hadoop.hbase.ZooKeeperConnectionException: org.apache.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = ConnectionLo
Hi I am trying to read from hbase the following code. http://pastebin.com/wvVVUT3p it reads for first 4-5 times but after that it start throwing this exception SEVERE: null org.apache.hadoop.hbase.ZooKeeperConnectionException: org.apache.hadoop.hbase.ZooKeeperConnectionException: org.apache.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = ConnectionLoss for /hbase at org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.getZooKeeperWatcher(HConnectionManager.java:988) at org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.setupZookeeperTrackers(HConnectionManager.java:301) at org.apache.hadoop.hbase.client.HConnectionManager$HCo The full details of exception are here. http://pastebin.com/Yj4VgvHG I am using CDH3b4 in pseudo enviornment. Any clue??? thnx -- Regards Shuja-ur-Rehman Baig http://pk.linkedin.com/in/shujamughal
file is already being created by NN_Recovery
Hello everybody We've run into this, now popular, error on our cluster 2011-04-07 16:28:00,654 WARN IPC Server handler 0 on 8020 org.apache.hadoop.hdfs.StateChange - DIR* NameSystem.startFile: failed to create file /hbase/.logs/search-hadoop-eu001.v300.gmx.net,60020,1302075782687/search-hadoop-eu001.v300.gmx.net%3A60020.1302075783467 for DFSClient_hb_m_search-namenode-eu002.v300.gmx.net:6_1302186078300 on client 10.1.100.32, because this file is already being created by NN_Recovery on 10.1.100.61 I've read a couple of threads around it, still it seems that nobody pinpointed the cause of it? The only solution here remains to delete the log file and lose data ? I've seen this error on almost any cluster we've installed so far, deleting logs was not concerning since all were test clusters. Now we got this on the production cluster, and strange, this cluster was just installed, there is no table and no data, no activity there. So what logs is master trying to create? We are running the latest CDH3B4 from Cloudera. Thanks for any hints, Daniel
Re: file is already being created by NN_Recovery
If you have socket.dfs.timeout set to 0, consider removing it, most of our issues like that went away after that. This problem occurs when you have datanode crash, and there is a conflict with the lease on the file (which should expire in one hour, this is unconfigurable hard timeout). If you do end up in situation like that, the only way we could resolve it is like this: # stop the master # hadoop fs -cp file new_file # hadoop fs -rm file # hadoop fs -cp new_file file # start master, and watch it replay the log. This appears to break the lease as new .log file does not have this issue. -Jack On Thu, Apr 7, 2011 at 9:35 AM, Daniel Iancu daniel.ia...@1and1.ro wrote: Hello everybody We've run into this, now popular, error on our cluster 2011-04-07 16:28:00,654 WARN IPC Server handler 0 on 8020 org.apache.hadoop.hdfs.StateChange - DIR* NameSystem.startFile: failed to create file /hbase/.logs/search-hadoop-eu001.v300.gmx.net,60020,1302075782687/search-hadoop-eu001.v300.gmx.net%3A60020.1302075783467 for DFSClient_hb_m_search-namenode-eu002.v300.gmx.net:6_1302186078300 on client 10.1.100.32, because this file is already being created by NN_Recovery on 10.1.100.61 I've read a couple of threads around it, still it seems that nobody pinpointed the cause of it? The only solution here remains to delete the log file and lose data ? I've seen this error on almost any cluster we've installed so far, deleting logs was not concerning since all were test clusters. Now we got this on the production cluster, and strange, this cluster was just installed, there is no table and no data, no activity there. So what logs is master trying to create? We are running the latest CDH3B4 from Cloudera. Thanks for any hints, Daniel
Task process exit with nonzero status of 255.
Hi, While executing MR with 472G data, I am running into following error: java.io.IOException: Task process exit with nonzero status of 255. at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:418) 2011-04-07 07:54:15,442 INFO org.apache.zookeeper.ClientCnxn: Opening socket connection to server 10.40.99.22:2181 2011-04-07 07:54:15,442 INFO org.apache.zookeeper.ClientCnxn: Socket connection established to 10.40.99.22:2181, initiating session 2011-04-07 07:54:15,443 WARN org.apache.zookeeper.ClientCnxn: Session 0x0 for server 10.40.99.22:2181, unexpected error, closing socket connection and attempting reconnect java.io.IOException: Connection reset by peer at sun.nio.ch.FileDispatcher.read0(Native Method) at sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:21) at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:237) at sun.nio.ch.IOUtil.read(IOUtil.java:204) at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:236) at org.apache.zookeeper.ClientCnxn$SendThread.doIO(ClientCnxn.java:858) at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1130) 2011-04-07 07:54:15,545 FATAL com.example.dataengine.aggregation.aggregationMR_HDFS: * 2011-04-07 07:54:15,545 FATAL com.example.dataengine.aggregation.aggregationMR_HDFS: org.apache.hadoop.hbase.ZooKeeperConnectionException: org.apache.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = ConnectionLoss for /hbase 2011-04-07 07:54:15,546 FATAL com.example.dataengine.aggregation.aggregationMR_HDFS: org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.getZooKeeperWatcher(HConnectionManager.java:988) org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.setupZookeeperTrackers(HConnectionManager.java:301) org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.init(HConnectionManager.java:292) org.apache.hadoop.hbase.client.HConnectionManager.getConnection(HConnectionManager.java:155) org.apache.hadoop.hbase.client.HTable.init(HTable.java:167) org.apache.hadoop.hbase.client.HTable.init(HTable.java:145) com.example.dataengine.aggregation.aggregationMR_HDFS$aggregationMapper.setup(aggregationMR_HDFS.java:324) org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:142) org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:621) org.apache.hadoop.mapred.MapTask.run(MapTask.java:305) org.apache.hadoop.mapred.Child.main(Child.java:170) 2011-04-07 07:54:15,546 FATAL com.example.dataengine.aggregation.aggregationMR_HDFS: * -- Thanks, Shah
Re: Task process exit with nonzero status of 255.
Check the log of the zookeeper at the adress that's printed, it may be a problem of too many connections (in which case you need to make sure you reuse the configuration objects). J-D On Thu, Apr 7, 2011 at 9:49 AM, Shahnawaz Saifi shahsa...@gmail.com wrote: Hi, While executing MR with 472G data, I am running into following error: java.io.IOException: Task process exit with nonzero status of 255. at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:418) 2011-04-07 07:54:15,442 INFO org.apache.zookeeper.ClientCnxn: Opening socket connection to server 10.40.99.22:2181 2011-04-07 07:54:15,442 INFO org.apache.zookeeper.ClientCnxn: Socket connection established to 10.40.99.22:2181, initiating session 2011-04-07 07:54:15,443 WARN org.apache.zookeeper.ClientCnxn: Session 0x0 for server 10.40.99.22:2181, unexpected error, closing socket connection and attempting reconnect java.io.IOException: Connection reset by peer at sun.nio.ch.FileDispatcher.read0(Native Method) at sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:21) at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:237) at sun.nio.ch.IOUtil.read(IOUtil.java:204) at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:236) at org.apache.zookeeper.ClientCnxn$SendThread.doIO(ClientCnxn.java:858) at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1130) 2011-04-07 07:54:15,545 FATAL com.example.dataengine.aggregation.aggregationMR_HDFS: * 2011-04-07 07:54:15,545 FATAL com.example.dataengine.aggregation.aggregationMR_HDFS: org.apache.hadoop.hbase.ZooKeeperConnectionException: org.apache.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = ConnectionLoss for /hbase 2011-04-07 07:54:15,546 FATAL com.example.dataengine.aggregation.aggregationMR_HDFS: org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.getZooKeeperWatcher(HConnectionManager.java:988) org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.setupZookeeperTrackers(HConnectionManager.java:301) org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.init(HConnectionManager.java:292) org.apache.hadoop.hbase.client.HConnectionManager.getConnection(HConnectionManager.java:155) org.apache.hadoop.hbase.client.HTable.init(HTable.java:167) org.apache.hadoop.hbase.client.HTable.init(HTable.java:145) com.example.dataengine.aggregation.aggregationMR_HDFS$aggregationMapper.setup(aggregationMR_HDFS.java:324) org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:142) org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:621) org.apache.hadoop.mapred.MapTask.run(MapTask.java:305) org.apache.hadoop.mapred.Child.main(Child.java:170) 2011-04-07 07:54:15,546 FATAL com.example.dataengine.aggregation.aggregationMR_HDFS: * -- Thanks, Shah
Re: file is already being created by NN_Recovery
The RegionServer is down for sure? Else it sounds like an issue that was addressed by the addition of a new short-circuit API call added to HDFS on the hadoop-0.20-append branch. The patches that added this new call went into the branch quite a while ago. They are: HDFS-1554. New semantics for recoverLease. (hairong) HDFS-1555. Disallow pipelien recovery if a file is already being lease recovered. (hairong) These patches are not in CDH3b*. They are in the CDH3 release which is due any day now. HBase 0.90.2 makes use of the new API: See https://issues.apache.org/jira/browse/HBASE-3285. Attached to that issue is a patch for CDH3b2, a patch we are running here at SU. Shout if you need a version of this patch for CDH3b3/4. St.Ack On Thu, Apr 7, 2011 at 9:35 AM, Daniel Iancu daniel.ia...@1and1.ro wrote: Hello everybody We've run into this, now popular, error on our cluster 2011-04-07 16:28:00,654 WARN IPC Server handler 0 on 8020 org.apache.hadoop.hdfs.StateChange - DIR* NameSystem.startFile: failed to create file /hbase/.logs/search-hadoop-eu001.v300.gmx.net,60020,1302075782687/search-hadoop-eu001.v300.gmx.net%3A60020.1302075783467 for DFSClient_hb_m_search-namenode-eu002.v300.gmx.net:6_1302186078300 on client 10.1.100.32, because this file is already being created by NN_Recovery on 10.1.100.61 I've read a couple of threads around it, still it seems that nobody pinpointed the cause of it? The only solution here remains to delete the log file and lose data ? I've seen this error on almost any cluster we've installed so far, deleting logs was not concerning since all were test clusters. Now we got this on the production cluster, and strange, this cluster was just installed, there is no table and no data, no activity there. So what logs is master trying to create? We are running the latest CDH3B4 from Cloudera. Thanks for any hints, Daniel
Re: timing out for hdfs errors faster
Another question, why would dfsclient setting for sockettimeout (for data reading) would be set so high by default if HBASE is expected to be real time? Shouldn't it be few seconds (5?). Not all clusters are used for real time applications, also usually users first try to cram as much data as they can and see if it holds, disregard their hardware, if they are swapping, or anything that might make things slow. A lot of configurations are set to high values for those reasons. 2011-04-07 07:49:41,527 WARN org.apache.hadoop.hdfs.DFSClient: Failed to connect to /10.103.7.5:50010 for file /hbase/media_data/1c95bfcf0dd19800b1f44278627259ae/att/7725092577730365184 for block 802538788372768807:java.net.SocketTimeoutException: 6 millis timeout while waiting for channel to be ready for read. ch : java.nio.channels.SocketChannel[connected local=/10.101.6.8:40801 remote=/10.103.7.5:50010 What would be configuration setting to shorten the timeout say to 5 seconds? What about retries (if any). Something is up with that Datanode as the region server isn't even able to establish a channel to it. The retries are done with other replicas (no need to hit the same faulty datanode twice). Looking at the code, the timeout for reads is set with dfs.socket.timeout J-D
Re: org.apache.hadoop.hbase.ZooKeeperConnectionException: org.apache.hadoop.hbase.ZooKeeperConnectionException: org.apache.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = Connecti
You should be seeing more log lines related to ZooKeeper before that. Also make sure your client connects to the zk server. J-D On Thu, Apr 7, 2011 at 9:11 AM, Shuja Rehman shujamug...@gmail.com wrote: Hi I am trying to read from hbase the following code. http://pastebin.com/wvVVUT3p it reads for first 4-5 times but after that it start throwing this exception SEVERE: null org.apache.hadoop.hbase.ZooKeeperConnectionException: org.apache.hadoop.hbase.ZooKeeperConnectionException: org.apache.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = ConnectionLoss for /hbase at org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.getZooKeeperWatcher(HConnectionManager.java:988) at org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.setupZookeeperTrackers(HConnectionManager.java:301) at org.apache.hadoop.hbase.client.HConnectionManager$HCo The full details of exception are here. http://pastebin.com/Yj4VgvHG I am using CDH3b4 in pseudo enviornment. Any clue??? thnx -- Regards Shuja-ur-Rehman Baig http://pk.linkedin.com/in/shujamughal
Re: timing out for hdfs errors faster
On Thu, Apr 7, 2011 at 7:58 AM, Jack Levin magn...@gmail.com wrote: Hello, I get those errors sometimes: 2011-04-07 07:49:41,527 WARN org.apache.hadoop.hdfs.DFSClient: Failed to connect to /10.103.7.5:50010 for file /hbase/media_data/1c95bfcf0dd19800b1f44278627259ae/att/7725092577730365184 for block 802538788372768807:java.net.SocketTimeoutException: 6 millis timeout while waiting for channel to be ready for read. ch : java.nio.channels.SocketChannel[connected local=/10.101.6.8:40801 remote=/10.103.7.5:50010 What would be configuration setting to shorten the timeout say to 5 seconds? What about retries (if any). 0.90.0 added a timeout to the RPC (See HBASE-3154 'HBase RPC should support timeout'). The default is 60 seconds. To change the config., set hbase.rpc.timeout. Retries should be going on in the upper layers. As to why 60 seconds, my guess is that the author and reviewer were being conservative. Previous there was no timeout. St.Ack
Re: HTable.put hangs on bulk loading
There's nothing of use in the pasted logs unfortunately, and the log didn't get attached to your mail (happens often). Consider putting on a web server or pastebin. Also I see you are on an older version, upgrading isn't going to fix your issue (which is probably related to your environment or configuration) but at least it's gonna be easier for us to support you. J-D On Wed, Apr 6, 2011 at 7:10 PM, ajay.gov agovindara...@yahoo.com wrote: I am doing a load test for which I need to load a table with many rows. I have a small java program that has a for loop and calls HTable.put. I am inserting a map of 2 items into a table that has one column family. The limit of the for loop is currently 2. However after 15876 rows the call to Put hangs. I am using autoFlush on the HTable. Any ideas why this may happen? The table configuration: DESCRIPTION ENABLED {NAME = 'TABLE2', FAMILIES = [{NAME = 'TABLE2_CF true 1', BLOOMFILTER = 'NONE', REPLICATION_SCOPE = '0' , COMPRESSION = 'NONE', VERSIONS = '3', TTL = '2 147483647', BLOCKSIZE = '65536', IN_MEMORY = 'fal se', BLOCKCACHE = 'true'}]} The HBase config on the client is the one in the hbase-default.xml. Some values: hbase.client.write.buffer=2097152 hbase.client.pause=1000 hbase.client.retries.number=10 If i use another client I am able to put items to the table. I am also able to scan items from the table using the hbase shell. I have attached the server configuratio I don't see anything in the region server or master logs. I have them here. The master server log: 2011-04-06 19:02:40,149 INFO org.apache.hadoop.hbase.master.BaseScanner: RegionManager.rootScanner scanning meta region {server: 184.106.69.238:60020, regionname: -ROOT-,,0.70236052, startKey: } 2011-04-06 19:02:40,152 INFO org.apache.hadoop.hbase.master.BaseScanner: RegionManager.rootScanner scan of 1 row(s) of meta region {server: 184.106.69.238:60020, regionname: -ROOT-,,0.70236052, startKey: } complete 2011-04-06 19:02:40,157 INFO org.apache.hadoop.hbase.master.ServerManager: 1 region servers, 0 dead, average load 42.0 2011-04-06 19:03:15,252 INFO org.apache.hadoop.hbase.master.BaseScanner: RegionManager.metaScanner scanning meta region {server: 184.106.69.238:60020, regionname: .META.,,1.1028785192, startKey: } 2011-04-06 19:03:15,265 INFO org.apache.hadoop.hbase.master.BaseScanner: RegionManager.metaScanner scan of 40 row(s) of meta region {server: 184.106.69.238:60020, regionname: .META.,,1.1028785192, startKey: } complete 2011-04-06 19:03:15,266 INFO org.apache.hadoop.hbase.master.BaseScanner: All 1 .META. region(s) scanned The region server logs: 2011-04-06 19:02:21,294 DEBUG org.apache.hadoop.hbase.regionserver.HRegion: Creating region TABLE2,,1302141740486.010a5ae704ed53f656cbddb8e489162a. 2011-04-06 19:02:21,295 INFO org.apache.hadoop.hbase.regionserver.HRegion: Onlined TABLE2,,1302141740486.010a5ae704ed53f656cbddb8e489162a.; next sequenceid=1 -- View this message in context: http://old.nabble.com/HTable.put-hangs-on-bulk-loading-tp31338874p31338874.html Sent from the HBase User mailing list archive at Nabble.com.
Re: timing out for hdfs errors faster
Jack: Pardon me. What J-D said. You were asking about DN timeout. Below I write about RS timeout. St.Ack On Thu, Apr 7, 2011 at 10:28 AM, Stack st...@duboce.net wrote: On Thu, Apr 7, 2011 at 7:58 AM, Jack Levin magn...@gmail.com wrote: Hello, I get those errors sometimes: 2011-04-07 07:49:41,527 WARN org.apache.hadoop.hdfs.DFSClient: Failed to connect to /10.103.7.5:50010 for file /hbase/media_data/1c95bfcf0dd19800b1f44278627259ae/att/7725092577730365184 for block 802538788372768807:java.net.SocketTimeoutException: 6 millis timeout while waiting for channel to be ready for read. ch : java.nio.channels.SocketChannel[connected local=/10.101.6.8:40801 remote=/10.103.7.5:50010 What would be configuration setting to shorten the timeout say to 5 seconds? What about retries (if any). 0.90.0 added a timeout to the RPC (See HBASE-3154 'HBase RPC should support timeout'). The default is 60 seconds. To change the config., set hbase.rpc.timeout. Retries should be going on in the upper layers. As to why 60 seconds, my guess is that the author and reviewer were being conservative. Previous there was no timeout. St.Ack
Re: timing out for hdfs errors faster
Thanks, How about setting hbase-site.xml with dfs.datanode.socket.write.timeout dfs.datanode.socket.read.write.timeout If tcp connection is established, but harddrive fails right after that, I do not want to wait 60 seconds to read, I want to quicky timeout and move to next datanode. -Jack On Thu, Apr 7, 2011 at 10:14 AM, Jean-Daniel Cryans jdcry...@apache.org wrote: Another question, why would dfsclient setting for sockettimeout (for data reading) would be set so high by default if HBASE is expected to be real time? Shouldn't it be few seconds (5?). Not all clusters are used for real time applications, also usually users first try to cram as much data as they can and see if it holds, disregard their hardware, if they are swapping, or anything that might make things slow. A lot of configurations are set to high values for those reasons. 2011-04-07 07:49:41,527 WARN org.apache.hadoop.hdfs.DFSClient: Failed to connect to /10.103.7.5:50010 for file /hbase/media_data/1c95bfcf0dd19800b1f44278627259ae/att/7725092577730365184 for block 802538788372768807:java.net.SocketTimeoutException: 6 millis timeout while waiting for channel to be ready for read. ch : java.nio.channels.SocketChannel[connected local=/10.101.6.8:40801 remote=/10.103.7.5:50010 What would be configuration setting to shorten the timeout say to 5 seconds? What about retries (if any). Something is up with that Datanode as the region server isn't even able to establish a channel to it. The retries are done with other replicas (no need to hit the same faulty datanode twice). Looking at the code, the timeout for reads is set with dfs.socket.timeout J-D
Re: timing out for hdfs errors faster
I meant to say dfs.datanode.socket.read.timeout -Jack On Thu, Apr 7, 2011 at 10:54 AM, Jack Levin magn...@gmail.com wrote: Thanks, How about setting hbase-site.xml with dfs.datanode.socket.write.timeout dfs.datanode.socket.read.write.timeout If tcp connection is established, but harddrive fails right after that, I do not want to wait 60 seconds to read, I want to quicky timeout and move to next datanode. -Jack On Thu, Apr 7, 2011 at 10:14 AM, Jean-Daniel Cryans jdcry...@apache.org wrote: Another question, why would dfsclient setting for sockettimeout (for data reading) would be set so high by default if HBASE is expected to be real time? Shouldn't it be few seconds (5?). Not all clusters are used for real time applications, also usually users first try to cram as much data as they can and see if it holds, disregard their hardware, if they are swapping, or anything that might make things slow. A lot of configurations are set to high values for those reasons. 2011-04-07 07:49:41,527 WARN org.apache.hadoop.hdfs.DFSClient: Failed to connect to /10.103.7.5:50010 for file /hbase/media_data/1c95bfcf0dd19800b1f44278627259ae/att/7725092577730365184 for block 802538788372768807:java.net.SocketTimeoutException: 6 millis timeout while waiting for channel to be ready for read. ch : java.nio.channels.SocketChannel[connected local=/10.101.6.8:40801 remote=/10.103.7.5:50010 What would be configuration setting to shorten the timeout say to 5 seconds? What about retries (if any). Something is up with that Datanode as the region server isn't even able to establish a channel to it. The retries are done with other replicas (no need to hit the same faulty datanode twice). Looking at the code, the timeout for reads is set with dfs.socket.timeout J-D
Hadoop Append Github
Is https://github.com/facebook/hadoop-20-append the Github branch for Hadoop Append 0.20?
Re: org.apache.hadoop.hbase.ZooKeeperConnectionException: org.apache.hadoop.hbase.ZooKeeperConnectionException: org.apache.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = Connecti
here is more log. now it is not connecting at all. 11/04/07 23:02:55 WARN hbase.HBaseConfiguration: instantiating HBaseConfiguration() is deprecated. Please use HBaseConfiguration#create() to construct a plain Configuration 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment:zookeeper.version=3.3.2-CDH3B4--1, built on 02/21/2011 22:45 GMT 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment:host.name=DV 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment:java.version=1.6.0_21 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment:java.vendor=Sun Microsystems Inc. 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment:java.home=C:\Program Files\Java\jdk1.6.0_21\jre 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment:java.class.path=D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\build\web\WEB-INF\classes;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\CopyLibs\commons-logging-1.1.1.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\CopyLibs\log4j-1.2.15.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\metro\activation.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\metro\webservices-api.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\metro\webservices-extra-api.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\metro\webservices-extra.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\metro\webservices-rt.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\metro\webservices-tools.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restapi\jsr311-api-1.1.1.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restlib\asm-3.1.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restlib\jackson-core-asl-1.1.1.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restlib\jersey-client-1.1.5.1.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restlib\jersey-core-1.1.5.1.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restlib\jersey-json-1.1.5.1.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restlib\jersey-server-1.1.5.1.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restlib\jersey-spring-1.1.5.1.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restlib\jettison-1.1.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restlib\oauth-client-1.1.5.1.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restlib\oauth-signature-1.1.5.1.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restlib\activation.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restlib\jaxb-api.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restlib\jsr173_api.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\CopyLibs\hadoop-0.20.2-CDH3B4-core.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\CopyLibs\hbase-0.90.1-CDH3B4.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\CopyLibs\mysql-connector-java-5.1.7-bin.jar;D:\FreeLauncing\DataWareHouse\UIDataTransporter\dist\UIDataTransporter.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\CopyLibs\zookeeper-3.3.2-CDH3B4.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\annotations-api.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\catalina-ant.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\catalina-ha.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\catalina-tribes.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\catalina.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\el-api.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\jasper-el.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\jasper.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\jsp-api.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\servlet-api.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\tomcat-coyote.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\tomcat-dbcp.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\tomcat-i18n-es.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\tomcat-i18n-fr.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\tomcat-i18n-ja.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\bin\tomcat-juli.jar 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment:java.library.path=C:\Program Files\Java\jdk1.6.0_21\jre\bin;.;C:\Windows\Sun\Java\bin;C:\Windows\system32;C:\Windows;C:\Program Files\Common Files\Microsoft Shared\Windows Live;C:\Program Files\PC Connectivity
Re: Hadoop Append Github
That's the one published by Facebook, the one maintained by Apache is https://github.com/apache/hadoop-common/tree/branch-0.20-append J-D On Thu, Apr 7, 2011 at 11:04 AM, Jason Rutherglen jason.rutherg...@gmail.com wrote: Is https://github.com/facebook/hadoop-20-append the Github branch for Hadoop Append 0.20?
Re: Hadoop Append Github
That one looks dead Jason. There was a bulk upload in December and nought since. St.Ack On Thu, Apr 7, 2011 at 11:04 AM, Jason Rutherglen jason.rutherg...@gmail.com wrote: Is https://github.com/facebook/hadoop-20-append the Github branch for Hadoop Append 0.20?
Re: Hadoop Append Github
Ah ok, Google turned up the one I posted, I wonder why this one was harder to find? Thanks! On Thu, Apr 7, 2011 at 11:07 AM, Jean-Daniel Cryans jdcry...@apache.org wrote: That's the one published by Facebook, the one maintained by Apache is https://github.com/apache/hadoop-common/tree/branch-0.20-append J-D On Thu, Apr 7, 2011 at 11:04 AM, Jason Rutherglen jason.rutherg...@gmail.com wrote: Is https://github.com/facebook/hadoop-20-append the Github branch for Hadoop Append 0.20?
Re: org.apache.hadoop.hbase.ZooKeeperConnectionException: org.apache.hadoop.hbase.ZooKeeperConnectionException: org.apache.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = Connecti
If you look at 204.13.166.85's zookeeper log, do you see anything that looks bad around the time you ran this? J-D On Thu, Apr 7, 2011 at 11:04 AM, Shuja Rehman shujamug...@gmail.com wrote: here is more log. now it is not connecting at all. 11/04/07 23:02:55 WARN hbase.HBaseConfiguration: instantiating HBaseConfiguration() is deprecated. Please use HBaseConfiguration#create() to construct a plain Configuration 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment:zookeeper.version=3.3.2-CDH3B4--1, built on 02/21/2011 22:45 GMT 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment:host.name=DV 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment:java.version=1.6.0_21 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment:java.vendor=Sun Microsystems Inc. 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment:java.home=C:\Program Files\Java\jdk1.6.0_21\jre 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment:java.class.path=D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\build\web\WEB-INF\classes;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\CopyLibs\commons-logging-1.1.1.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\CopyLibs\log4j-1.2.15.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\metro\activation.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\metro\webservices-api.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\metro\webservices-extra-api.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\metro\webservices-extra.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\metro\webservices-rt.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\metro\webservices-tools.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restapi\jsr311-api-1.1.1.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restlib\asm-3.1.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restlib\jackson-core-asl-1.1.1.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restlib\jersey-client-1.1.5.1.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restlib\jersey-core-1.1.5.1.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restlib\jersey-json-1.1.5.1.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restlib\jersey-server-1.1.5.1.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restlib\jersey-spring-1.1.5.1.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restlib\jettison-1.1.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restlib\oauth-client-1.1.5.1.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restlib\oauth-signature-1.1.5.1.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restlib\activation.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restlib\jaxb-api.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\restlib\jsr173_api.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\CopyLibs\hadoop-0.20.2-CDH3B4-core.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\CopyLibs\hbase-0.90.1-CDH3B4.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\CopyLibs\mysql-connector-java-5.1.7-bin.jar;D:\FreeLauncing\DataWareHouse\UIDataTransporter\dist\UIDataTransporter.jar;D:\FreeLauncing\DataWareHouse\WebServices\WebServiceSvn\lib\CopyLibs\zookeeper-3.3.2-CDH3B4.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\annotations-api.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\catalina-ant.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\catalina-ha.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\catalina-tribes.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\catalina.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\el-api.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\jasper-el.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\jasper.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\jsp-api.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\servlet-api.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\tomcat-coyote.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\tomcat-dbcp.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\tomcat-i18n-es.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\tomcat-i18n-fr.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\lib\tomcat-i18n-ja.jar;C:\Program Files\Apache Software Foundation\Apache Tomcat 6.0.26\bin\tomcat-juli.jar 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment:java.library.path=C:\Program
Re: org.apache.hadoop.hbase.ZooKeeperConnectionException: org.apache.hadoop.hbase.ZooKeeperConnectionException: org.apache.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = Connecti
i got the log file and it says that 2011-04-07 11:17:41,864 - WARN [NIOServerCxn.Factory: 0.0.0.0/0.0.0.0:2181:NIOServerCnxn$Factory@247] - Too many connections from /182.178.254.222 - max is 10 2011-04-07 11:17:45,453 - WARN [NIOServerCxn.Factory: 0.0.0.0/0.0.0.0:2181:NIOServerCnxn$Factory@247] - Too many connections from /182.178.254.222 - max is 10 2011-04-07 11:17:47,470 - WARN [NIOServerCxn.Factory: 0.0.0.0/0.0.0.0:2181:NIOServerCnxn$Factory@247] - Too many connections from /182.178.254.222 - max is 10 2011-04-07 11:17:49,964 - WARN [NIOServerCxn.Factory: 0.0.0.0/0.0.0.0:2181:NIOServerCnxn$Factory@247] - Too many connections from /182.178.254.222 - max is 10 2011-04-07 11:17:52,101 - WARN [NIOServerCxn.Factory: 0.0.0.0/0.0.0.0:2181:NIOServerCnxn$Factory@247] - Too many connections from /182.178.254.222 - max is 10 2011-04-07 11:17:54,211 - WARN [NIOServerCxn.Factory: 0.0.0.0/0.0.0.0:2181:NIOServerCnxn$Factory@247] - Too many connections from /182.178.254.222 - max is 10 2011-04-07 11:17:54,251 - WARN [NIOServerCxn.Factory: 0.0.0.0/0.0.0.0:2181:NIOServerCnxn@639] - Exception causing close of session 0x12f30d63bc90013 due to java.io.IOException: Connection reset by peer 2011-04-07 11:17:54,251 - INFO [NIOServerCxn.Factory: 0.0.0.0/0.0.0.0:2181:NIOServerCnxn@1435] - Closed socket connection for client /182.178.254.222:27546 which had sessionid 0x12f30d63bc90013 2011-04-07 11:18:24,004 - INFO [SessionTracker:ZooKeeperServer@314] - Expiring session 0x12f30d63bc90013, timeout of 4ms exceeded 2011-04-07 11:18:24,005 - INFO [ProcessThread:-1:PrepRequestProcessor@387] - Processed session termination for sessionid: 0x12f30d63bc90013 so how to close the connection? i am also using now htable.close() in the end of funtion. On Thu, Apr 7, 2011 at 11:21 PM, Shuja Rehman shujamug...@gmail.com wrote: where zookeeper keeps its logs? i check the /usr/lib/zookeeper and there is no log file inside. also not in bin folder. On Thu, Apr 7, 2011 at 11:17 PM, Jean-Daniel Cryans jdcry...@apache.orgwrote: If you look at 204.13.166.85's zookeeper log, do you see anything that looks bad around the time you ran this? J-D On Thu, Apr 7, 2011 at 11:04 AM, Shuja Rehman shujamug...@gmail.com wrote: here is more log. now it is not connecting at all. 11/04/07 23:02:55 WARN hbase.HBaseConfiguration: instantiating HBaseConfiguration() is deprecated. Please use HBaseConfiguration#create() to construct a plain Configuration 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment:zookeeper.version=3.3.2-CDH3B4--1, built on 02/21/2011 22:45 GMT 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment: host.name=DV 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment:java.version=1.6.0_21 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment:java.vendor=Sun Microsystems Inc. 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment:java.home=C:\Program Files\Java\jdk1.6.0_21\jre 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client
Re: org.apache.hadoop.hbase.ZooKeeperConnectionException: org.apache.hadoop.hbase.ZooKeeperConnectionException: org.apache.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = Connecti
So regarding finding your logs and other stuff related to that, since you are using CDH you should always check their documentation. In ZooKeeper there's a configurable limit of 30 connections per IP. HTable.close won't close the connection since you can have multiple HTables using the same cluster. Always make sure you reuse the same configuration object and always reuse the HTables, don't create new ones per session or something like that. This is all described in HTable's class javadoc: http://hbase.apache.org/apidocs/org/apache/hadoop/hbase/client/HTable.html J-D On Thu, Apr 7, 2011 at 11:35 AM, Shuja Rehman shujamug...@gmail.com wrote: i got the log file and it says that 2011-04-07 11:17:41,864 - WARN [NIOServerCxn.Factory: 0.0.0.0/0.0.0.0:2181:NIOServerCnxn$Factory@247] - Too many connections from /182.178.254.222 - max is 10 2011-04-07 11:17:45,453 - WARN [NIOServerCxn.Factory: 0.0.0.0/0.0.0.0:2181:NIOServerCnxn$Factory@247] - Too many connections from /182.178.254.222 - max is 10 2011-04-07 11:17:47,470 - WARN [NIOServerCxn.Factory: 0.0.0.0/0.0.0.0:2181:NIOServerCnxn$Factory@247] - Too many connections from /182.178.254.222 - max is 10 2011-04-07 11:17:49,964 - WARN [NIOServerCxn.Factory: 0.0.0.0/0.0.0.0:2181:NIOServerCnxn$Factory@247] - Too many connections from /182.178.254.222 - max is 10 2011-04-07 11:17:52,101 - WARN [NIOServerCxn.Factory: 0.0.0.0/0.0.0.0:2181:NIOServerCnxn$Factory@247] - Too many connections from /182.178.254.222 - max is 10 2011-04-07 11:17:54,211 - WARN [NIOServerCxn.Factory: 0.0.0.0/0.0.0.0:2181:NIOServerCnxn$Factory@247] - Too many connections from /182.178.254.222 - max is 10 2011-04-07 11:17:54,251 - WARN [NIOServerCxn.Factory: 0.0.0.0/0.0.0.0:2181:NIOServerCnxn@639] - Exception causing close of session 0x12f30d63bc90013 due to java.io.IOException: Connection reset by peer 2011-04-07 11:17:54,251 - INFO [NIOServerCxn.Factory: 0.0.0.0/0.0.0.0:2181:NIOServerCnxn@1435] - Closed socket connection for client /182.178.254.222:27546 which had sessionid 0x12f30d63bc90013 2011-04-07 11:18:24,004 - INFO [SessionTracker:ZooKeeperServer@314] - Expiring session 0x12f30d63bc90013, timeout of 4ms exceeded 2011-04-07 11:18:24,005 - INFO [ProcessThread:-1:PrepRequestProcessor@387] - Processed session termination for sessionid: 0x12f30d63bc90013 so how to close the connection? i am also using now htable.close() in the end of funtion. On Thu, Apr 7, 2011 at 11:21 PM, Shuja Rehman shujamug...@gmail.com wrote: where zookeeper keeps its logs? i check the /usr/lib/zookeeper and there is no log file inside. also not in bin folder. On Thu, Apr 7, 2011 at 11:17 PM, Jean-Daniel Cryans jdcry...@apache.orgwrote: If you look at 204.13.166.85's zookeeper log, do you see anything that looks bad around the time you ran this? J-D On Thu, Apr 7, 2011 at 11:04 AM, Shuja Rehman shujamug...@gmail.com wrote: here is more log. now it is not connecting at all. 11/04/07 23:02:55 WARN hbase.HBaseConfiguration: instantiating HBaseConfiguration() is deprecated. Please use HBaseConfiguration#create() to construct a plain Configuration 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment:zookeeper.version=3.3.2-CDH3B4--1, built on 02/21/2011 22:45 GMT 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment: host.name=DV 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment:java.version=1.6.0_21 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment:java.vendor=Sun Microsystems Inc. 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment:java.home=C:\Program Files\Java\jdk1.6.0_21\jre 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client
Re: HTable.put hangs on bulk loading
Sorry, my server config was not attached. Its here: http://pastebin.com/U41QZGiq thanks -ajay ajay.gov wrote: I am doing a load test for which I need to load a table with many rows. I have a small java program that has a for loop and calls HTable.put. I am inserting a map of 2 items into a table that has one column family. The limit of the for loop is currently 2. However after 15876 rows the call to Put hangs. I am using autoFlush on the HTable. Any ideas why this may happen? The table configuration: DESCRIPTION ENABLED {NAME = 'TABLE2', FAMILIES = [{NAME = 'TABLE2_CF true 1', BLOOMFILTER = 'NONE', REPLICATION_SCOPE = '0' , COMPRESSION = 'NONE', VERSIONS = '3', TTL = '2 147483647', BLOCKSIZE = '65536', IN_MEMORY = 'fal se', BLOCKCACHE = 'true'}]} The HBase config on the client is the one in the hbase-default.xml. Some values: hbase.client.write.buffer=2097152 hbase.client.pause=1000 hbase.client.retries.number=10 If i use another client I am able to put items to the table. I am also able to scan items from the table using the hbase shell. I have attached the server configuratio I don't see anything in the region server or master logs. I have them here. The master server log: 2011-04-06 19:02:40,149 INFO org.apache.hadoop.hbase.master.BaseScanner: RegionManager.rootScanner scanning meta region {server: 184.106.69.238:60020, regionname: -ROOT-,,0.70236052, startKey: } 2011-04-06 19:02:40,152 INFO org.apache.hadoop.hbase.master.BaseScanner: RegionManager.rootScanner scan of 1 row(s) of meta region {server: 184.106.69.238:60020, regionname: -ROOT-,,0.70236052, startKey: } complete 2011-04-06 19:02:40,157 INFO org.apache.hadoop.hbase.master.ServerManager: 1 region servers, 0 dead, average load 42.0 2011-04-06 19:03:15,252 INFO org.apache.hadoop.hbase.master.BaseScanner: RegionManager.metaScanner scanning meta region {server: 184.106.69.238:60020, regionname: .META.,,1.1028785192, startKey: } 2011-04-06 19:03:15,265 INFO org.apache.hadoop.hbase.master.BaseScanner: RegionManager.metaScanner scan of 40 row(s) of meta region {server: 184.106.69.238:60020, regionname: .META.,,1.1028785192, startKey: } complete 2011-04-06 19:03:15,266 INFO org.apache.hadoop.hbase.master.BaseScanner: All 1 .META. region(s) scanned The region server logs: 2011-04-06 19:02:21,294 DEBUG org.apache.hadoop.hbase.regionserver.HRegion: Creating region TABLE2,,1302141740486.010a5ae704ed53f656cbddb8e489162a. 2011-04-06 19:02:21,295 INFO org.apache.hadoop.hbase.regionserver.HRegion: Onlined TABLE2,,1302141740486.010a5ae704ed53f656cbddb8e489162a.; next sequenceid=1 -- View this message in context: http://old.nabble.com/HTable.put-hangs-on-bulk-loading-tp31338874p31345580.html Sent from the HBase User mailing list archive at Nabble.com.
Re: org.apache.hadoop.hbase.ZooKeeperConnectionException: org.apache.hadoop.hbase.ZooKeeperConnectionException: org.apache.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = Connecti
To help usability, I created https://issues.apache.org/jira/browse/HBASE-3755 J-D On Thu, Apr 7, 2011 at 11:39 AM, Jean-Daniel Cryans jdcry...@apache.org wrote: So regarding finding your logs and other stuff related to that, since you are using CDH you should always check their documentation. In ZooKeeper there's a configurable limit of 30 connections per IP. HTable.close won't close the connection since you can have multiple HTables using the same cluster. Always make sure you reuse the same configuration object and always reuse the HTables, don't create new ones per session or something like that. This is all described in HTable's class javadoc: http://hbase.apache.org/apidocs/org/apache/hadoop/hbase/client/HTable.html J-D On Thu, Apr 7, 2011 at 11:35 AM, Shuja Rehman shujamug...@gmail.com wrote: i got the log file and it says that 2011-04-07 11:17:41,864 - WARN [NIOServerCxn.Factory: 0.0.0.0/0.0.0.0:2181:NIOServerCnxn$Factory@247] - Too many connections from /182.178.254.222 - max is 10 2011-04-07 11:17:45,453 - WARN [NIOServerCxn.Factory: 0.0.0.0/0.0.0.0:2181:NIOServerCnxn$Factory@247] - Too many connections from /182.178.254.222 - max is 10 2011-04-07 11:17:47,470 - WARN [NIOServerCxn.Factory: 0.0.0.0/0.0.0.0:2181:NIOServerCnxn$Factory@247] - Too many connections from /182.178.254.222 - max is 10 2011-04-07 11:17:49,964 - WARN [NIOServerCxn.Factory: 0.0.0.0/0.0.0.0:2181:NIOServerCnxn$Factory@247] - Too many connections from /182.178.254.222 - max is 10 2011-04-07 11:17:52,101 - WARN [NIOServerCxn.Factory: 0.0.0.0/0.0.0.0:2181:NIOServerCnxn$Factory@247] - Too many connections from /182.178.254.222 - max is 10 2011-04-07 11:17:54,211 - WARN [NIOServerCxn.Factory: 0.0.0.0/0.0.0.0:2181:NIOServerCnxn$Factory@247] - Too many connections from /182.178.254.222 - max is 10 2011-04-07 11:17:54,251 - WARN [NIOServerCxn.Factory: 0.0.0.0/0.0.0.0:2181:NIOServerCnxn@639] - Exception causing close of session 0x12f30d63bc90013 due to java.io.IOException: Connection reset by peer 2011-04-07 11:17:54,251 - INFO [NIOServerCxn.Factory: 0.0.0.0/0.0.0.0:2181:NIOServerCnxn@1435] - Closed socket connection for client /182.178.254.222:27546 which had sessionid 0x12f30d63bc90013 2011-04-07 11:18:24,004 - INFO [SessionTracker:ZooKeeperServer@314] - Expiring session 0x12f30d63bc90013, timeout of 4ms exceeded 2011-04-07 11:18:24,005 - INFO [ProcessThread:-1:PrepRequestProcessor@387] - Processed session termination for sessionid: 0x12f30d63bc90013 so how to close the connection? i am also using now htable.close() in the end of funtion. On Thu, Apr 7, 2011 at 11:21 PM, Shuja Rehman shujamug...@gmail.com wrote: where zookeeper keeps its logs? i check the /usr/lib/zookeeper and there is no log file inside. also not in bin folder. On Thu, Apr 7, 2011 at 11:17 PM, Jean-Daniel Cryans jdcry...@apache.orgwrote: If you look at 204.13.166.85's zookeeper log, do you see anything that looks bad around the time you ran this? J-D On Thu, Apr 7, 2011 at 11:04 AM, Shuja Rehman shujamug...@gmail.com wrote: here is more log. now it is not connecting at all. 11/04/07 23:02:55 WARN hbase.HBaseConfiguration: instantiating HBaseConfiguration() is deprecated. Please use HBaseConfiguration#create() to construct a plain Configuration 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment:zookeeper.version=3.3.2-CDH3B4--1, built on 02/21/2011 22:45 GMT 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment: host.name=DV 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment:java.version=1.6.0_21 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment:java.vendor=Sun Microsystems Inc. 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client environment:java.home=C:\Program Files\Java\jdk1.6.0_21\jre 11/04/07 23:02:55 INFO zookeeper.ZooKeeper: Client
Re: Hadoop Append Github
Is http://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20-append different than the Github one at https://github.com/apache/hadoop-common/tree/branch-0.20-append ? As I can apply the HDFS-347 patch successfully to the SVN version, however the Github one has a number of rejects. Are they different or is there some uniqueness involved with Github, or something simple I'm missing. On Thu, Apr 7, 2011 at 11:10 AM, Stack st...@duboce.net wrote: That one looks dead Jason. There was a bulk upload in December and nought since. St.Ack On Thu, Apr 7, 2011 at 11:04 AM, Jason Rutherglen jason.rutherg...@gmail.com wrote: Is https://github.com/facebook/hadoop-20-append the Github branch for Hadoop Append 0.20?
Re: Hadoop Append Github
As far as I can tell, they are at the same revision. J-D On Thu, Apr 7, 2011 at 1:19 PM, Jason Rutherglen jason.rutherg...@gmail.com wrote: Is http://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20-append different than the Github one at https://github.com/apache/hadoop-common/tree/branch-0.20-append ? As I can apply the HDFS-347 patch successfully to the SVN version, however the Github one has a number of rejects. Are they different or is there some uniqueness involved with Github, or something simple I'm missing. On Thu, Apr 7, 2011 at 11:10 AM, Stack st...@duboce.net wrote: That one looks dead Jason. There was a bulk upload in December and nought since. St.Ack On Thu, Apr 7, 2011 at 11:04 AM, Jason Rutherglen jason.rutherg...@gmail.com wrote: Is https://github.com/facebook/hadoop-20-append the Github branch for Hadoop Append 0.20?
Re: Hadoop Append Github
How did you compare? On Thu, Apr 7, 2011 at 1:37 PM, Jean-Daniel Cryans jdcry...@apache.org wrote: As far as I can tell, they are at the same revision. J-D On Thu, Apr 7, 2011 at 1:19 PM, Jason Rutherglen jason.rutherg...@gmail.com wrote: Is http://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20-append different than the Github one at https://github.com/apache/hadoop-common/tree/branch-0.20-append ? As I can apply the HDFS-347 patch successfully to the SVN version, however the Github one has a number of rejects. Are they different or is there some uniqueness involved with Github, or something simple I'm missing. On Thu, Apr 7, 2011 at 11:10 AM, Stack st...@duboce.net wrote: That one looks dead Jason. There was a bulk upload in December and nought since. St.Ack On Thu, Apr 7, 2011 at 11:04 AM, Jason Rutherglen jason.rutherg...@gmail.com wrote: Is https://github.com/facebook/hadoop-20-append the Github branch for Hadoop Append 0.20?
Re: Hadoop Append Github
It looks like they may [somehow] be different? The latest change to SVN happened 2011-01-10 whereas the Github one was changed Wed Apr 6 19:03:37 2011? Here's from Github: commit 53d6ff79e8c4ee850cf4e592ddd20b8e116a8513 Author: Konstantin Shvachko s...@apache.org Date: Wed Apr 6 19:03:37 2011 + Committing HADOOP-6949 to branch 0.22. git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1089562 13f79535-47bb-0310-9956-ffa450edef68 from http://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20-append Path: . URL: http://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20-append Repository Root: http://svn.apache.org/repos/asf Repository UUID: 13f79535-47bb-0310-9956-ffa450edef68 Revision: 1089967 Node Kind: directory Schedule: normal Last Changed Author: hairong Last Changed Rev: 1057313 Last Changed Date: 2011-01-10 11:01:36 -0800 (Mon, 10 Jan 2011) On Thu, Apr 7, 2011 at 2:05 PM, Jason Rutherglen jason.rutherg...@gmail.com wrote: How did you compare? On Thu, Apr 7, 2011 at 1:37 PM, Jean-Daniel Cryans jdcry...@apache.org wrote: As far as I can tell, they are at the same revision. J-D On Thu, Apr 7, 2011 at 1:19 PM, Jason Rutherglen jason.rutherg...@gmail.com wrote: Is http://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20-append different than the Github one at https://github.com/apache/hadoop-common/tree/branch-0.20-append ? As I can apply the HDFS-347 patch successfully to the SVN version, however the Github one has a number of rejects. Are they different or is there some uniqueness involved with Github, or something simple I'm missing. On Thu, Apr 7, 2011 at 11:10 AM, Stack st...@duboce.net wrote: That one looks dead Jason. There was a bulk upload in December and nought since. St.Ack On Thu, Apr 7, 2011 at 11:04 AM, Jason Rutherglen jason.rutherg...@gmail.com wrote: Is https://github.com/facebook/hadoop-20-append the Github branch for Hadoop Append 0.20?
Re: Hadoop Append Github
That last change on github was for trunk, not the append branch. The last one I see in that branch is: HDFS-1554. New semantics for recoverLease. Contributed by Hairong Kuang. Hairong Kuang (author) January 10, 2011 Same as in SVN. J-D On Thu, Apr 7, 2011 at 2:09 PM, Jason Rutherglen jason.rutherg...@gmail.com wrote: It looks like they may [somehow] be different? The latest change to SVN happened 2011-01-10 whereas the Github one was changed Wed Apr 6 19:03:37 2011? Here's from Github: commit 53d6ff79e8c4ee850cf4e592ddd20b8e116a8513 Author: Konstantin Shvachko s...@apache.org Date: Wed Apr 6 19:03:37 2011 + Committing HADOOP-6949 to branch 0.22. git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1089562 13f79535-47bb-0310-9956-ffa450edef68 from http://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20-append Path: . URL: http://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20-append Repository Root: http://svn.apache.org/repos/asf Repository UUID: 13f79535-47bb-0310-9956-ffa450edef68 Revision: 1089967 Node Kind: directory Schedule: normal Last Changed Author: hairong Last Changed Rev: 1057313 Last Changed Date: 2011-01-10 11:01:36 -0800 (Mon, 10 Jan 2011) On Thu, Apr 7, 2011 at 2:05 PM, Jason Rutherglen jason.rutherg...@gmail.com wrote: How did you compare? On Thu, Apr 7, 2011 at 1:37 PM, Jean-Daniel Cryans jdcry...@apache.org wrote: As far as I can tell, they are at the same revision. J-D On Thu, Apr 7, 2011 at 1:19 PM, Jason Rutherglen jason.rutherg...@gmail.com wrote: Is http://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20-append different than the Github one at https://github.com/apache/hadoop-common/tree/branch-0.20-append ? As I can apply the HDFS-347 patch successfully to the SVN version, however the Github one has a number of rejects. Are they different or is there some uniqueness involved with Github, or something simple I'm missing. On Thu, Apr 7, 2011 at 11:10 AM, Stack st...@duboce.net wrote: That one looks dead Jason. There was a bulk upload in December and nought since. St.Ack On Thu, Apr 7, 2011 at 11:04 AM, Jason Rutherglen jason.rutherg...@gmail.com wrote: Is https://github.com/facebook/hadoop-20-append the Github branch for Hadoop Append 0.20?
Re: Hadoop Append Github
How using Github were you able to see only the log for the given branch/URL? I'm not sure why the patch won't apply. I ran diff, and there are differences, though they're mostly in the scripts and other non-source code files. On Thu, Apr 7, 2011 at 2:20 PM, Jean-Daniel Cryans jdcry...@apache.org wrote: That last change on github was for trunk, not the append branch. The last one I see in that branch is: HDFS-1554. New semantics for recoverLease. Contributed by Hairong Kuang. Hairong Kuang (author) January 10, 2011 Same as in SVN. J-D On Thu, Apr 7, 2011 at 2:09 PM, Jason Rutherglen jason.rutherg...@gmail.com wrote: It looks like they may [somehow] be different? The latest change to SVN happened 2011-01-10 whereas the Github one was changed Wed Apr 6 19:03:37 2011? Here's from Github: commit 53d6ff79e8c4ee850cf4e592ddd20b8e116a8513 Author: Konstantin Shvachko s...@apache.org Date: Wed Apr 6 19:03:37 2011 + Committing HADOOP-6949 to branch 0.22. git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1089562 13f79535-47bb-0310-9956-ffa450edef68 from http://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20-append Path: . URL: http://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20-append Repository Root: http://svn.apache.org/repos/asf Repository UUID: 13f79535-47bb-0310-9956-ffa450edef68 Revision: 1089967 Node Kind: directory Schedule: normal Last Changed Author: hairong Last Changed Rev: 1057313 Last Changed Date: 2011-01-10 11:01:36 -0800 (Mon, 10 Jan 2011) On Thu, Apr 7, 2011 at 2:05 PM, Jason Rutherglen jason.rutherg...@gmail.com wrote: How did you compare? On Thu, Apr 7, 2011 at 1:37 PM, Jean-Daniel Cryans jdcry...@apache.org wrote: As far as I can tell, they are at the same revision. J-D On Thu, Apr 7, 2011 at 1:19 PM, Jason Rutherglen jason.rutherg...@gmail.com wrote: Is http://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20-append different than the Github one at https://github.com/apache/hadoop-common/tree/branch-0.20-append ? As I can apply the HDFS-347 patch successfully to the SVN version, however the Github one has a number of rejects. Are they different or is there some uniqueness involved with Github, or something simple I'm missing. On Thu, Apr 7, 2011 at 11:10 AM, Stack st...@duboce.net wrote: That one looks dead Jason. There was a bulk upload in December and nought since. St.Ack On Thu, Apr 7, 2011 at 11:04 AM, Jason Rutherglen jason.rutherg...@gmail.com wrote: Is https://github.com/facebook/hadoop-20-append the Github branch for Hadoop Append 0.20?
Re: Hadoop Append Github
So from that page: https://github.com/apache/hadoop-common Switch to the append branch then click on the right on the history button. Make sure you switched to the append branch in your cloned git repo too, by default you are on trunk. J-D On Thu, Apr 7, 2011 at 4:08 PM, Jason Rutherglen jason.rutherg...@gmail.com wrote: How using Github were you able to see only the log for the given branch/URL? I'm not sure why the patch won't apply. I ran diff, and there are differences, though they're mostly in the scripts and other non-source code files. On Thu, Apr 7, 2011 at 2:20 PM, Jean-Daniel Cryans jdcry...@apache.org wrote: That last change on github was for trunk, not the append branch. The last one I see in that branch is: HDFS-1554. New semantics for recoverLease. Contributed by Hairong Kuang. Hairong Kuang (author) January 10, 2011 Same as in SVN. J-D On Thu, Apr 7, 2011 at 2:09 PM, Jason Rutherglen jason.rutherg...@gmail.com wrote: It looks like they may [somehow] be different? The latest change to SVN happened 2011-01-10 whereas the Github one was changed Wed Apr 6 19:03:37 2011? Here's from Github: commit 53d6ff79e8c4ee850cf4e592ddd20b8e116a8513 Author: Konstantin Shvachko s...@apache.org Date: Wed Apr 6 19:03:37 2011 + Committing HADOOP-6949 to branch 0.22. git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1089562 13f79535-47bb-0310-9956-ffa450edef68 from http://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20-append Path: . URL: http://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20-append Repository Root: http://svn.apache.org/repos/asf Repository UUID: 13f79535-47bb-0310-9956-ffa450edef68 Revision: 1089967 Node Kind: directory Schedule: normal Last Changed Author: hairong Last Changed Rev: 1057313 Last Changed Date: 2011-01-10 11:01:36 -0800 (Mon, 10 Jan 2011) On Thu, Apr 7, 2011 at 2:05 PM, Jason Rutherglen jason.rutherg...@gmail.com wrote: How did you compare? On Thu, Apr 7, 2011 at 1:37 PM, Jean-Daniel Cryans jdcry...@apache.org wrote: As far as I can tell, they are at the same revision. J-D On Thu, Apr 7, 2011 at 1:19 PM, Jason Rutherglen jason.rutherg...@gmail.com wrote: Is http://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20-append different than the Github one at https://github.com/apache/hadoop-common/tree/branch-0.20-append ? As I can apply the HDFS-347 patch successfully to the SVN version, however the Github one has a number of rejects. Are they different or is there some uniqueness involved with Github, or something simple I'm missing. On Thu, Apr 7, 2011 at 11:10 AM, Stack st...@duboce.net wrote: That one looks dead Jason. There was a bulk upload in December and nought since. St.Ack On Thu, Apr 7, 2011 at 11:04 AM, Jason Rutherglen jason.rutherg...@gmail.com wrote: Is https://github.com/facebook/hadoop-20-append the Github branch for Hadoop Append 0.20?
Re: HTable.put hangs on bulk loading
Thanks for pointing this out. I have uploaded the server config at: http://pastebin.com/U41QZGiq thanks -ajay From: Jean-Daniel Cryans jdcry...@apache.org To: user@hbase.apache.org Sent: Thursday, April 7, 2011 10:29 AM Subject: Re: HTable.put hangs on bulk loading There's nothing of use in the pasted logs unfortunately, and the log didn't get attached to your mail (happens often). Consider putting on a web server or pastebin. Also I see you are on an older version, upgrading isn't going to fix your issue (which is probably related to your environment or configuration) but at least it's gonna be easier for us to support you. J-D On Wed, Apr 6, 2011 at 7:10 PM, ajay.gov agovindara...@yahoo.com wrote: I am doing a load test for which I need to load a table with many rows. I have a small java program that has a for loop and calls HTable.put. I am inserting a map of 2 items into a table that has one column family. The limit of the for loop is currently 2. However after 15876 rows the call to Put hangs. I am using autoFlush on the HTable. Any ideas why this may happen? The table configuration: DESCRIPTION ENABLED {NAME = 'TABLE2', FAMILIES = [{NAME = 'TABLE2_CF true 1', BLOOMFILTER = 'NONE', REPLICATION_SCOPE = '0' , COMPRESSION = 'NONE', VERSIONS = '3', TTL = '2 147483647', BLOCKSIZE = '65536', IN_MEMORY = 'fal se', BLOCKCACHE = 'true'}]} The HBase config on the client is the one in the hbase-default.xml. Some values: hbase.client.write.buffer=2097152 hbase.client.pause=1000 hbase.client.retries.number=10 If i use another client I am able to put items to the table. I am also able to scan items from the table using the hbase shell. I have attached the server configuratio I don't see anything in the region server or master logs. I have them here. The master server log: 2011-04-06 19:02:40,149 INFO org.apache.hadoop.hbase.master.BaseScanner: RegionManager.rootScanner scanning meta region {server: 184.106.69.238:60020, regionname: -ROOT-,,0.70236052, startKey: } 2011-04-06 19:02:40,152 INFO org.apache.hadoop.hbase.master.BaseScanner: RegionManager.rootScanner scan of 1 row(s) of meta region {server: 184.106.69.238:60020, regionname: -ROOT-,,0.70236052, startKey: } complete 2011-04-06 19:02:40,157 INFO org.apache.hadoop.hbase.master.ServerManager: 1 region servers, 0 dead, average load 42.0 2011-04-06 19:03:15,252 INFO org.apache.hadoop.hbase.master.BaseScanner: RegionManager.metaScanner scanning meta region {server: 184.106.69.238:60020, regionname: .META.,,1.1028785192, startKey: } 2011-04-06 19:03:15,265 INFO org.apache.hadoop.hbase.master.BaseScanner: RegionManager.metaScanner scan of 40 row(s) of meta region {server: 184.106.69.238:60020, regionname: .META.,,1.1028785192, startKey: } complete 2011-04-06 19:03:15,266 INFO org.apache.hadoop.hbase.master.BaseScanner: All 1 .META. region(s) scanned The region server logs: 2011-04-06 19:02:21,294 DEBUG org.apache.hadoop.hbase.regionserver.HRegion: Creating region TABLE2,,1302141740486.010a5ae704ed53f656cbddb8e489162a. 2011-04-06 19:02:21,295 INFO org.apache.hadoop.hbase.regionserver.HRegion: Onlined TABLE2,,1302141740486.010a5ae704ed53f656cbddb8e489162a.; next sequenceid=1 -- View this message in context: http://old.nabble.com/HTable.put-hangs-on-bulk-loading-tp31338874p31338874.html Sent from the HBase User mailing list archive at Nabble.com.
zookeper warning with 0.90.1 hbase
I see lot of these warnings..everything seems to be working otherwise..Is this something that can be ignored? 2011-04-07 21:29:15,032 WARN Timer-0-SendThread(..:2181) org.apache.zookeeper.ClientCnxn - Session 0x0 for server :2181, unexpected error, closing socket connection and attempting reconnect java.io.IOException: Connection reset by peer at sun.nio.ch.FileDispatcher.read0(Native Method) at sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:21) at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:233) at sun.nio.ch.IOUtil.read(IOUtil.java:200) at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:236) at org.apache.zookeeper.ClientCnxn$SendThread.doIO(ClientCnxn.java:858) at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1130) 2011-04-07 21:29:15,032 DEBUG Timer-0-SendThread(..:2181) org.apache.zookeeper.ClientCnxn - Ignoring exception during shutdown input java.net.SocketException: Transport endpoint is not connected at sun.nio.ch.SocketChannelImpl.shutdown(Native Method) at sun.nio.ch.SocketChannelImpl.shutdownInput(SocketChannelImpl.java:640) at sun.nio.ch.SocketAdaptor.shutdownInput(SocketAdaptor.java:360) at org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:1205) at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:11
Re: zookeper warning with 0.90.1 hbase
They happen on the end of a map task or on shutdown? If so, yes, ignore (or, if you want to have nice clean shutdown, figure how Session 0x0 was set up -- was it you -- and call appropriate close in time). St.Ack On Thu, Apr 7, 2011 at 6:33 PM, Venkatesh vramanatha...@aol.com wrote: I see lot of these warnings..everything seems to be working otherwise..Is this something that can be ignored? 2011-04-07 21:29:15,032 WARN Timer-0-SendThread(..:2181) org.apache.zookeeper.ClientCnxn - Session 0x0 for server :2181, unexpected error, closing socket connection and attempting reconnect java.io.IOException: Connection reset by peer at sun.nio.ch.FileDispatcher.read0(Native Method) at sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:21) at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:233) at sun.nio.ch.IOUtil.read(IOUtil.java:200) at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:236) at org.apache.zookeeper.ClientCnxn$SendThread.doIO(ClientCnxn.java:858) at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1130) 2011-04-07 21:29:15,032 DEBUG Timer-0-SendThread(..:2181) org.apache.zookeeper.ClientCnxn - Ignoring exception during shutdown input java.net.SocketException: Transport endpoint is not connected at sun.nio.ch.SocketChannelImpl.shutdown(Native Method) at sun.nio.ch.SocketChannelImpl.shutdownInput(SocketChannelImpl.java:640) at sun.nio.ch.SocketAdaptor.shutdownInput(SocketAdaptor.java:360) at org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:1205) at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:11