Look in your datanode logs.  What are these complaining about?

The timeout needs to be seen by dfsclient that hbase is using (add it
to hbase-site.xml or symlink your hadoop-site.xml into
$HBASE_HOME/conf).

How to up file descriptors is also in the FAQ.

St.Ack


On Tue, May 26, 2009 at 10:32 AM, llpind <[email protected]> wrote:
>
> Here are relevent properties:
>
> <property>
>  <name>dfs.replication</name>
>  <value>3</value>
>  <description>Default block replication.
>  The actual number of replications can be specified when the file is
> created.
>  The default is used if replication is not specified in create time.
>  </description>
> </property>
> <property>
>        <name>dfs.datanode.max.xcievers</name>
>        <value>8196</value>
> </property>
> <property>
>  <name>dfs.balance.bandwidthPerSec</name>
>  <value>10485760</value>
>  <description> Specifies the maximum bandwidth that each datanode can
> utilize for the
>   balancing purpose in term of the number of bytes per second. Default is
> 1048576</description>
> </property>
> <property>
>        <name>dfs.datanode.socket.write.timeout</name>
>        <value>0</value>
> </property>
>
> I'm guessing my xceivers and timeout should should be okay?  The one missing
> is the file descriptors you mentioned.  i will try that once this load
> completes (or fails).
>
> It had written out around 1-2 million records at the time of my first tail
> output.   As you mentioned it appears it never splits.
>
> current tail looks like this:
> =================================================
> 2009-05-26 10:20:10,247 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 failed  because recovery
> from primary datanode 192.168.240.175:50010 failed 4 times. Will retry...
> 2009-05-26 10:20:10,250 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 bad datanode[1]
> 192.168.240.180:50010
> 2009-05-26 10:20:10,250 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 in pipeline
> 192.168.240.175:50010, 192.168.240.180:50010: bad datanode
> 192.168.240.180:50010
> 2009-05-26 10:20:10,260 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 failed  because recovery
> from primary datanode 192.168.240.175:50010 failed 5 times. Will retry...
> 2009-05-26 10:20:10,270 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 bad datanode[1]
> 192.168.240.180:50010
> 2009-05-26 10:20:10,270 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 in pipeline
> 192.168.240.175:50010, 192.168.240.180:50010: bad datanode
> 192.168.240.180:50010
> 2009-05-26 10:20:10,278 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 failed  because recovery
> from primary datanode 192.168.240.175:50010 failed 6 times. Marking primary
> datanode as bad.
> 2009-05-26 10:20:10,281 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 bad datanode[0]
> 192.168.240.175:50010
> 2009-05-26 10:20:10,281 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 in pipeline
> 192.168.240.175:50010, 192.168.240.180:50010: bad datanode
> 192.168.240.175:50010
> 2009-05-26 10:20:10,291 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 failed  because recovery
> from primary datanode 192.168.240.180:50010 failed 1 times. Will retry...
> 2009-05-26 10:20:10,294 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 bad datanode[0]
> 192.168.240.175:50010
> 2009-05-26 10:20:10,294 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 in pipeline
> 192.168.240.175:50010, 192.168.240.180:50010: bad datanode
> 192.168.240.175:50010
> 2009-05-26 10:20:10,308 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 failed  because recovery
> from primary datanode 192.168.240.180:50010 failed 2 times. Will retry...
> 2009-05-26 10:20:10,310 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 bad datanode[0]
> 192.168.240.175:50010
> 2009-05-26 10:20:10,310 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 in pipeline
> 192.168.240.175:50010, 192.168.240.180:50010: bad datanode
> 192.168.240.175:50010
> 2009-05-26 10:20:10,325 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 failed  because recovery
> from primary datanode 192.168.240.180:50010 failed 3 times. Will retry...
> 2009-05-26 10:20:10,417 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 bad datanode[0]
> 192.168.240.175:50010
> 2009-05-26 10:20:10,417 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 in pipeline
> 192.168.240.175:50010, 192.168.240.180:50010: bad datanode
> 192.168.240.175:50010
> 2009-05-26 10:20:10,432 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 failed  because recovery
> from primary datanode 192.168.240.180:50010 failed 4 times. Will retry...
> 2009-05-26 10:20:10,435 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 bad datanode[0]
> 192.168.240.175:50010
> 2009-05-26 10:20:10,435 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 in pipeline
> 192.168.240.175:50010, 192.168.240.180:50010: bad datanode
> 192.168.240.175:50010
> 2009-05-26 10:20:11,285 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 failed  because recovery
> from primary datanode 192.168.240.180:50010 failed 5 times. Will retry...
> 2009-05-26 10:20:11,288 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 bad datanode[0]
> 192.168.240.175:50010
> 2009-05-26 10:20:11,288 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 in pipeline
> 192.168.240.175:50010, 192.168.240.180:50010: bad datanode
> 192.168.240.175:50010
> 2009-05-26 10:20:11,297 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 failed  because recovery
> from primary datanode 192.168.240.180:50010 failed 6 times. Marking primary
> datanode as bad.
> 2009-05-26 10:20:11,300 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 bad datanode[1]
> 192.168.240.180:50010
> 2009-05-26 10:20:11,300 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 in pipeline
> 192.168.240.175:50010, 192.168.240.180:50010: bad datanode
> 192.168.240.180:50010
> 2009-05-26 10:20:11,313 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 failed  because recovery
> from primary datanode 192.168.240.175:50010 failed 1 times. Will retry...
> 2009-05-26 10:20:12,581 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 bad datanode[1]
> 192.168.240.180:50010
> 2009-05-26 10:20:12,581 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 in pipeline
> 192.168.240.175:50010, 192.168.240.180:50010: bad datanode
> 192.168.240.180:50010
> 2009-05-26 10:20:12,590 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 failed  because recovery
> from primary datanode 192.168.240.175:50010 failed 2 times. Will retry...
> 2009-05-26 10:20:13,317 INFO org.apache.hadoop.hbase.regionserver.HRegion:
> compaction completed on region tableA,ROW_KEY,1243357190459 in 4sec
> 2009-05-26 10:20:13,594 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 bad datanode[1]
> 192.168.240.180:50010
> 2009-05-26 10:20:13,594 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 in pipeline
> 192.168.240.175:50010, 192.168.240.180:50010: bad datanode
> 192.168.240.180:50010
> 2009-05-26 10:20:13,601 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 failed  because recovery
> from primary datanode 192.168.240.175:50010 failed 3 times. Will retry...
> 2009-05-26 10:20:14,602 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 bad datanode[1]
> 192.168.240.180:50010
> 2009-05-26 10:20:14,936 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 in pipeline
> 192.168.240.175:50010, 192.168.240.180:50010: bad datanode
> 192.168.240.180:50010
> 2009-05-26 10:20:14,945 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 failed  because recovery
> from primary datanode 192.168.240.175:50010 failed 4 times. Will retry...
> 2009-05-26 10:20:15,946 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 bad datanode[1]
> 192.168.240.180:50010
> 2009-05-26 10:20:15,946 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 in pipeline
> 192.168.240.175:50010, 192.168.240.180:50010: bad datanode
> 192.168.240.180:50010
> 2009-05-26 10:20:15,954 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 failed  because recovery
> from primary datanode 192.168.240.175:50010 failed 5 times. Will retry...
> 2009-05-26 10:20:16,958 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 bad datanode[1]
> 192.168.240.180:50010
> 2009-05-26 10:20:16,958 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_6687462549119446006_1241 in pipeline
> 192.168.240.175:50010, 192.168.240.180:50010: bad datanode
> 192.168.240.180:50010
> 2009-05-26 10:20:18,199 INFO org.apache.hadoop.hbase.regionserver.HLog:
> Closed
> hdfs://ats181:54310/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358357536,
> entries=105000. New log writer:
> /hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358418193
> 2009-05-26 10:20:56,012 INFO org.apache.hadoop.hbase.regionserver.HRegion:
> starting  compaction on region tableA,ROW_KEY,1243357190459
> 2009-05-26 10:20:56,059 INFO org.apache.hadoop.hbase.regionserver.HLog:
> Closed
> hdfs://ats181:54310/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358365174,
> entries=105001. New log writer:
> /hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358456051
> 2009-05-26 10:21:03,106 INFO org.apache.hadoop.hbase.regionserver.HRegion:
> compaction completed on region tableA,ROW_KEY,1243357190459 in 7sec
> 2009-05-26 10:21:29,698 INFO org.apache.hadoop.hbase.regionserver.HLog:
> Closed
> hdfs://ats181:54310/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358418193,
> entries=105000. New log writer:
> /hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358489691
> 2009-05-26 10:21:41,243 WARN org.apache.hadoop.hdfs.DFSClient: DataStreamer
> Exception: java.net.SocketTimeoutException: 10000 millis timeout while
> waiting for channel to be ready for write. ch :
> java.nio.channels.SocketChannel[connected local=/192.168.240.175:57592
> remote=/192.168.240.175:50010]
>        at
> org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:162)
>        at
> org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:146)
>        at
> org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:107)
>        at java.io.BufferedOutputStream.write(BufferedOutputStream.java:105)
>        at java.io.DataOutputStream.write(DataOutputStream.java:90)
>        at
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2209)
>
> 2009-05-26 10:21:41,268 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_439016349677707872_1258 bad datanode[0]
> 192.168.240.175:50010
> 2009-05-26 10:21:41,268 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_439016349677707872_1258 in pipeline
> 192.168.240.175:50010, 192.168.240.179:50010: bad datanode
> 192.168.240.175:50010
> 2009-05-26 10:21:41,343 WARN org.apache.hadoop.hdfs.DFSClient: DataStreamer
> Exception: java.net.SocketTimeoutException: 10000 millis timeout while
> waiting for channel to be ready for write. ch :
> java.nio.channels.SocketChannel[connected local=/192.168.240.175:57590
> remote=/192.168.240.175:50010]
>        at
> org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:162)
>        at
> org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:146)
>        at
> org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:107)
>        at java.io.BufferedOutputStream.write(BufferedOutputStream.java:105)
>        at java.io.DataOutputStream.write(DataOutputStream.java:90)
>        at
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2209)
>
> 2009-05-26 10:21:41,355 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_-5277507070339351226_1256 bad datanode[0]
> 192.168.240.175:50010
> 2009-05-26 10:21:41,356 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_-5277507070339351226_1256 in pipeline
> 192.168.240.175:50010, 192.168.240.179:50010: bad datanode
> 192.168.240.175:50010
> 2009-05-26 10:21:42,760 INFO org.apache.hadoop.hbase.regionserver.HRegion:
> starting  compaction on region tableA,ROW_KEY,1243357190459
> 2009-05-26 10:21:42,791 INFO org.apache.hadoop.hbase.regionserver.HRegion:
> compaction completed on region tableA,ROW_KEY,1243357190459 in 0sec
> 2009-05-26 10:22:13,806 INFO org.apache.hadoop.hbase.regionserver.HLog:
> Closed
> hdfs://ats181:54310/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358456051,
> entries=100001. New log writer:
> /hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358533799
> 2009-05-26 10:22:17,209 INFO org.apache.hadoop.hbase.regionserver.HRegion:
> starting  compaction on region tableA,ROW_KEY,1243357190459
> 2009-05-26 10:22:17,250 INFO org.apache.hadoop.hbase.regionserver.HLog:
> Closed
> hdfs://ats181:54310/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358489691,
> entries=125001. New log writer:
> /hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358537229
> 2009-05-26 10:22:21,116 INFO org.apache.hadoop.hbase.regionserver.HRegion:
> compaction completed on region tableA,ROW_KEY,1243357190459 in 3sec
> 2009-05-26 10:22:50,638 INFO org.apache.hadoop.hbase.regionserver.HLog:
> Closed
> hdfs://ats181:54310/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358533799,
> entries=105000. New log writer:
> /hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358570622
> 2009-05-26 10:22:52,213 INFO org.apache.hadoop.hbase.regionserver.HRegion:
> starting  compaction on region tableA,ROW_KEY,1243357190459
> 2009-05-26 10:23:04,025 WARN org.apache.hadoop.hdfs.DFSClient: DataStreamer
> Exception: java.net.SocketTimeoutException: 10000 millis timeout while
> waiting for channel to be ready for write. ch :
> java.nio.channels.SocketChannel[connected local=/192.168.240.175:57657
> remote=/192.168.240.175:50010]
>        at
> org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:162)
>        at
> org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:146)
>        at
> org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:107)
>        at java.io.BufferedOutputStream.write(BufferedOutputStream.java:105)
>        at java.io.DataOutputStream.write(DataOutputStream.java:90)
>        at
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2209)
>
> 2009-05-26 10:23:04,025 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_-5608797308669010722_1275 bad datanode[0]
> 192.168.240.175:50010
> 2009-05-26 10:23:04,025 WARN org.apache.hadoop.hdfs.DFSClient: Error
> Recovery for block blk_-5608797308669010722_1275 in pipeline
> 192.168.240.175:50010, 192.168.240.179:50010: bad datanode
> 192.168.240.175:50010
> 2009-05-26 10:23:23,536 INFO org.apache.hadoop.hbase.regionserver.HRegion:
> compaction completed on region tableA,ROW_KEY,1243357190459 in 31sec
> 2009-05-26 10:23:24,031 INFO org.apache.hadoop.hbase.regionserver.HLog:
> Closed
> hdfs://ats181:54310/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358537229,
> entries=100001. New log writer:
> /hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358604023
> 2009-05-26 10:24:06,109 INFO org.apache.hadoop.hbase.regionserver.HRegion:
> starting  compaction on region tableA,ROW_KEY,1243357190459
> 2009-05-26 10:24:06,180 INFO org.apache.hadoop.hbase.regionserver.HLog:
> Closed
> hdfs://ats181:54310/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358570622,
> entries=155001. New log writer:
> /hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358646150
> 2009-05-26 10:24:06,214 INFO org.apache.hadoop.hbase.regionserver.HRegion:
> compaction completed on region tableA,ROW_KEY,1243357190459 in 0sec
> 2009-05-26 10:24:39,502 INFO org.apache.hadoop.hbase.regionserver.HLog:
> Closed
> hdfs://ats181:54310/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358604023,
> entries=105000. New log writer:
> /hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358679496
> 2009-05-26 10:24:51,089 INFO org.apache.hadoop.hbase.regionserver.HRegion:
> starting  compaction on region tableA,ROW_KEY,1243357190459
> 2009-05-26 10:24:54,372 INFO org.apache.hadoop.hbase.regionserver.HRegion:
> compaction completed on region tableA,ROW_KEY,1243357190459 in 3sec
> 2009-05-26 10:25:22,844 INFO org.apache.hadoop.hbase.regionserver.HLog:
> Closed
> hdfs://ats181:54310/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358646150,
> entries=100001. New log writer:
> /hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358722837
> 2009-05-26 10:25:28,110 INFO org.apache.hadoop.hbase.regionserver.HRegion:
> starting  compaction on region tableA,ROW_KEY,1243357190459
> 2009-05-26 10:25:28,132 INFO org.apache.hadoop.hbase.regionserver.HLog:
> Closed
> hdfs://ats181:54310/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358679496,
> entries=110001. New log writer:
> /hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358728126
> 2009-05-26 10:25:28,158 INFO org.apache.hadoop.hbase.regionserver.HRegion:
> compaction completed on region tableA,ROW_KEY,1243357190459 in 0sec
> 2009-05-26 10:25:59,007 INFO org.apache.hadoop.hbase.regionserver.HLog:
> Closed
> hdfs://ats181:54310/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358722837,
> entries=105000. New log writer:
> /hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358758999
> 2009-05-26 10:26:38,530 INFO org.apache.hadoop.hbase.regionserver.HRegion:
> starting  compaction on region tableA,ROW_KEY,1243357190459
> 2009-05-26 10:26:38,975 INFO org.apache.hadoop.hbase.regionserver.HLog:
> Closed
> hdfs://ats181:54310/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358728126,
> entries=120001. New log writer:
> /hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358798947
>
> =====================================================================
>
>
>
>
> stack-3 wrote:
>>
>> On Tue, May 26, 2009 at 8:41 AM, llpind <[email protected]> wrote:
>>>
>>> Haven't tried the RC yet, but regions do get lost when i do intensive
>>> write
>>> operations (e.g. they are no longer listed under online regions).
>>
>>
>> Have you tried writing less virulently?  Maybe we just can't take your
>> write load on your setup.
>>
>> Above you say all writes are going to one region only.   How many rows
>> have gone in in your estimation?  This is the case when hbase starts
>> up;  all writes go to a single region until sufficient to split.
>> Maybe enough data has not yet gone in?  Perhaps your write rate is
>> such that hbase is unable to split?  Try taking load off.  Try
>> manually splitting regions (See the hbase shell.  Type 'toos' to see
>> list of admin methods).
>>
>> I would suggest you not change default flush and region sizes.
>> We're better able to help if sizes are default.
>>
>>>
>>> Will this RC work with Haoop version 0.19.1?
>>
>>
>> Yes.  An hbase will run on a hadoop of same major and minor version
>> (They can different in the point version number).
>>
>>>
>>> When I mentioned i tried different configurations, I was tweaking
>>> different
>>> properties mentioned in that post.  We are still at a lost for what to
>>> do.
>>>
>>>
>> List is short:
>>
>> + Up your file descriptors.  1024 is not enough.
>> + Up your hadoop xceivers.  256 is too little
>> + Set the timeout on dfsclient to 0.
>>
>> For more detail on how to do above configurations, their exact names
>> and whether client or server-side config, see the FAQ and
>> troubleshooting.
>>
>> Let us know the particular issues you are running into.  We'd like to help
>> out.
>>
>> St.Ack
>>
>>
>>
>>
>>
>>>
>>> stack-3 wrote:
>>>>
>>>> The RC has fixes to help w/ regionserver/master disagreement as to who
>>>> has what regions; i.e. "region loss".  You might give it a go?
>>>> St.Ack
>>>>
>>>> On Sun, May 24, 2009 at 10:33 AM, llpind <[email protected]> wrote:
>>>>>
>>>>> Hey Stack, I'm using 0.19.1.  Also, would like to know if I should
>>>>> check
>>>>> out
>>>>> the latest and try that or try the RC you mentioned above.
>>>>>
>>>>>
>>>>> stack-3 wrote:
>>>>>>
>>>>>> Are you using TRUNK (If you have answered this question already,
>>>>>> please excuse my not remembering)?
>>>>>>
>>>>>> St.Ack
>>>>>>
>>>>>> On Sat, May 23, 2009 at 2:17 PM, llpind <[email protected]>
>>>>>> wrote:
>>>>>>>
>>>>>>> I see similar behavior in my small cluster.  (1 master, 3 datanodes)
>>>>>>>
>>>>>>> I am also planning on trying this RC version.  I've tried various
>>>>>>> configurations, and I continue to lose Regions with intensive writes.
>>>>>>>  I
>>>>>>> really hope something like this will work, because we are starting to
>>>>>>> consider other options now.
>>>>>>>
>>>>>>> --
>>>>>>> View this message in context:
>>>>>>> http://www.nabble.com/HBase-looses-regions.-tp23657983p23688361.html
>>>>>>> Sent from the HBase User mailing list archive at Nabble.com.
>>>>>>>
>>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>> --
>>>>> View this message in context:
>>>>> http://www.nabble.com/HBase-looses-regions.-tp23657983p23696025.html
>>>>> Sent from the HBase User mailing list archive at Nabble.com.
>>>>>
>>>>>
>>>>
>>>>
>>>
>>> --
>>> View this message in context:
>>> http://www.nabble.com/HBase-looses-regions.-tp23657983p23725775.html
>>> Sent from the HBase User mailing list archive at Nabble.com.
>>>
>>>
>>
>>
>
> --
> View this message in context: 
> http://www.nabble.com/HBase-looses-regions.-tp23657983p23727877.html
> Sent from the HBase User mailing list archive at Nabble.com.
>
>

Reply via email to