Here are relevent properties:

<property>
  <name>dfs.replication</name>
  <value>3</value>
  <description>Default block replication.
  The actual number of replications can be specified when the file is
created.
  The default is used if replication is not specified in create time.
  </description>
</property>
<property>
        <name>dfs.datanode.max.xcievers</name>
        <value>8196</value>
</property>
<property> 
  <name>dfs.balance.bandwidthPerSec</name> 
  <value>10485760</value> 
  <description> Specifies the maximum bandwidth that each datanode can
utilize for the 
   balancing purpose in term of the number of bytes per second. Default is
1048576</description> 
</property> 
<property>
        <name>dfs.datanode.socket.write.timeout</name>
        <value>0</value>
</property>
 
I'm guessing my xceivers and timeout should should be okay?  The one missing
is the file descriptors you mentioned.  i will try that once this load
completes (or fails). 

It had written out around 1-2 million records at the time of my first tail
output.   As you mentioned it appears it never splits.

current tail looks like this:
=================================================
2009-05-26 10:20:10,247 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 failed  because recovery
from primary datanode 192.168.240.175:50010 failed 4 times. Will retry...
2009-05-26 10:20:10,250 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 bad datanode[1]
192.168.240.180:50010
2009-05-26 10:20:10,250 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 in pipeline
192.168.240.175:50010, 192.168.240.180:50010: bad datanode
192.168.240.180:50010
2009-05-26 10:20:10,260 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 failed  because recovery
from primary datanode 192.168.240.175:50010 failed 5 times. Will retry...
2009-05-26 10:20:10,270 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 bad datanode[1]
192.168.240.180:50010
2009-05-26 10:20:10,270 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 in pipeline
192.168.240.175:50010, 192.168.240.180:50010: bad datanode
192.168.240.180:50010
2009-05-26 10:20:10,278 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 failed  because recovery
from primary datanode 192.168.240.175:50010 failed 6 times. Marking primary
datanode as bad.
2009-05-26 10:20:10,281 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 bad datanode[0]
192.168.240.175:50010
2009-05-26 10:20:10,281 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 in pipeline
192.168.240.175:50010, 192.168.240.180:50010: bad datanode
192.168.240.175:50010
2009-05-26 10:20:10,291 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 failed  because recovery
from primary datanode 192.168.240.180:50010 failed 1 times. Will retry...
2009-05-26 10:20:10,294 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 bad datanode[0]
192.168.240.175:50010
2009-05-26 10:20:10,294 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 in pipeline
192.168.240.175:50010, 192.168.240.180:50010: bad datanode
192.168.240.175:50010
2009-05-26 10:20:10,308 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 failed  because recovery
from primary datanode 192.168.240.180:50010 failed 2 times. Will retry...
2009-05-26 10:20:10,310 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 bad datanode[0]
192.168.240.175:50010
2009-05-26 10:20:10,310 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 in pipeline
192.168.240.175:50010, 192.168.240.180:50010: bad datanode
192.168.240.175:50010
2009-05-26 10:20:10,325 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 failed  because recovery
from primary datanode 192.168.240.180:50010 failed 3 times. Will retry...
2009-05-26 10:20:10,417 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 bad datanode[0]
192.168.240.175:50010
2009-05-26 10:20:10,417 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 in pipeline
192.168.240.175:50010, 192.168.240.180:50010: bad datanode
192.168.240.175:50010
2009-05-26 10:20:10,432 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 failed  because recovery
from primary datanode 192.168.240.180:50010 failed 4 times. Will retry...
2009-05-26 10:20:10,435 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 bad datanode[0]
192.168.240.175:50010
2009-05-26 10:20:10,435 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 in pipeline
192.168.240.175:50010, 192.168.240.180:50010: bad datanode
192.168.240.175:50010
2009-05-26 10:20:11,285 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 failed  because recovery
from primary datanode 192.168.240.180:50010 failed 5 times. Will retry...
2009-05-26 10:20:11,288 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 bad datanode[0]
192.168.240.175:50010
2009-05-26 10:20:11,288 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 in pipeline
192.168.240.175:50010, 192.168.240.180:50010: bad datanode
192.168.240.175:50010
2009-05-26 10:20:11,297 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 failed  because recovery
from primary datanode 192.168.240.180:50010 failed 6 times. Marking primary
datanode as bad.
2009-05-26 10:20:11,300 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 bad datanode[1]
192.168.240.180:50010
2009-05-26 10:20:11,300 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 in pipeline
192.168.240.175:50010, 192.168.240.180:50010: bad datanode
192.168.240.180:50010
2009-05-26 10:20:11,313 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 failed  because recovery
from primary datanode 192.168.240.175:50010 failed 1 times. Will retry...
2009-05-26 10:20:12,581 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 bad datanode[1]
192.168.240.180:50010
2009-05-26 10:20:12,581 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 in pipeline
192.168.240.175:50010, 192.168.240.180:50010: bad datanode
192.168.240.180:50010
2009-05-26 10:20:12,590 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 failed  because recovery
from primary datanode 192.168.240.175:50010 failed 2 times. Will retry...
2009-05-26 10:20:13,317 INFO org.apache.hadoop.hbase.regionserver.HRegion:
compaction completed on region tableA,ROW_KEY,1243357190459 in 4sec
2009-05-26 10:20:13,594 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 bad datanode[1]
192.168.240.180:50010
2009-05-26 10:20:13,594 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 in pipeline
192.168.240.175:50010, 192.168.240.180:50010: bad datanode
192.168.240.180:50010
2009-05-26 10:20:13,601 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 failed  because recovery
from primary datanode 192.168.240.175:50010 failed 3 times. Will retry...
2009-05-26 10:20:14,602 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 bad datanode[1]
192.168.240.180:50010
2009-05-26 10:20:14,936 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 in pipeline
192.168.240.175:50010, 192.168.240.180:50010: bad datanode
192.168.240.180:50010
2009-05-26 10:20:14,945 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 failed  because recovery
from primary datanode 192.168.240.175:50010 failed 4 times. Will retry...
2009-05-26 10:20:15,946 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 bad datanode[1]
192.168.240.180:50010
2009-05-26 10:20:15,946 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 in pipeline
192.168.240.175:50010, 192.168.240.180:50010: bad datanode
192.168.240.180:50010
2009-05-26 10:20:15,954 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 failed  because recovery
from primary datanode 192.168.240.175:50010 failed 5 times. Will retry...
2009-05-26 10:20:16,958 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 bad datanode[1]
192.168.240.180:50010
2009-05-26 10:20:16,958 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_6687462549119446006_1241 in pipeline
192.168.240.175:50010, 192.168.240.180:50010: bad datanode
192.168.240.180:50010
2009-05-26 10:20:18,199 INFO org.apache.hadoop.hbase.regionserver.HLog:
Closed
hdfs://ats181:54310/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358357536,
entries=105000. New log writer:
/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358418193
2009-05-26 10:20:56,012 INFO org.apache.hadoop.hbase.regionserver.HRegion:
starting  compaction on region tableA,ROW_KEY,1243357190459
2009-05-26 10:20:56,059 INFO org.apache.hadoop.hbase.regionserver.HLog:
Closed
hdfs://ats181:54310/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358365174,
entries=105001. New log writer:
/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358456051
2009-05-26 10:21:03,106 INFO org.apache.hadoop.hbase.regionserver.HRegion:
compaction completed on region tableA,ROW_KEY,1243357190459 in 7sec
2009-05-26 10:21:29,698 INFO org.apache.hadoop.hbase.regionserver.HLog:
Closed
hdfs://ats181:54310/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358418193,
entries=105000. New log writer:
/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358489691
2009-05-26 10:21:41,243 WARN org.apache.hadoop.hdfs.DFSClient: DataStreamer
Exception: java.net.SocketTimeoutException: 10000 millis timeout while
waiting for channel to be ready for write. ch :
java.nio.channels.SocketChannel[connected local=/192.168.240.175:57592
remote=/192.168.240.175:50010]
        at
org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:162)
        at
org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:146)
        at
org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:107)
        at java.io.BufferedOutputStream.write(BufferedOutputStream.java:105)
        at java.io.DataOutputStream.write(DataOutputStream.java:90)
        at
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2209)

2009-05-26 10:21:41,268 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_439016349677707872_1258 bad datanode[0]
192.168.240.175:50010
2009-05-26 10:21:41,268 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_439016349677707872_1258 in pipeline
192.168.240.175:50010, 192.168.240.179:50010: bad datanode
192.168.240.175:50010
2009-05-26 10:21:41,343 WARN org.apache.hadoop.hdfs.DFSClient: DataStreamer
Exception: java.net.SocketTimeoutException: 10000 millis timeout while
waiting for channel to be ready for write. ch :
java.nio.channels.SocketChannel[connected local=/192.168.240.175:57590
remote=/192.168.240.175:50010]
        at
org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:162)
        at
org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:146)
        at
org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:107)
        at java.io.BufferedOutputStream.write(BufferedOutputStream.java:105)
        at java.io.DataOutputStream.write(DataOutputStream.java:90)
        at
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2209)

2009-05-26 10:21:41,355 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_-5277507070339351226_1256 bad datanode[0]
192.168.240.175:50010
2009-05-26 10:21:41,356 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_-5277507070339351226_1256 in pipeline
192.168.240.175:50010, 192.168.240.179:50010: bad datanode
192.168.240.175:50010
2009-05-26 10:21:42,760 INFO org.apache.hadoop.hbase.regionserver.HRegion:
starting  compaction on region tableA,ROW_KEY,1243357190459
2009-05-26 10:21:42,791 INFO org.apache.hadoop.hbase.regionserver.HRegion:
compaction completed on region tableA,ROW_KEY,1243357190459 in 0sec
2009-05-26 10:22:13,806 INFO org.apache.hadoop.hbase.regionserver.HLog:
Closed
hdfs://ats181:54310/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358456051,
entries=100001. New log writer:
/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358533799
2009-05-26 10:22:17,209 INFO org.apache.hadoop.hbase.regionserver.HRegion:
starting  compaction on region tableA,ROW_KEY,1243357190459
2009-05-26 10:22:17,250 INFO org.apache.hadoop.hbase.regionserver.HLog:
Closed
hdfs://ats181:54310/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358489691,
entries=125001. New log writer:
/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358537229
2009-05-26 10:22:21,116 INFO org.apache.hadoop.hbase.regionserver.HRegion:
compaction completed on region tableA,ROW_KEY,1243357190459 in 3sec
2009-05-26 10:22:50,638 INFO org.apache.hadoop.hbase.regionserver.HLog:
Closed
hdfs://ats181:54310/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358533799,
entries=105000. New log writer:
/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358570622
2009-05-26 10:22:52,213 INFO org.apache.hadoop.hbase.regionserver.HRegion:
starting  compaction on region tableA,ROW_KEY,1243357190459
2009-05-26 10:23:04,025 WARN org.apache.hadoop.hdfs.DFSClient: DataStreamer
Exception: java.net.SocketTimeoutException: 10000 millis timeout while
waiting for channel to be ready for write. ch :
java.nio.channels.SocketChannel[connected local=/192.168.240.175:57657
remote=/192.168.240.175:50010]
        at
org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:162)
        at
org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:146)
        at
org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:107)
        at java.io.BufferedOutputStream.write(BufferedOutputStream.java:105)
        at java.io.DataOutputStream.write(DataOutputStream.java:90)
        at
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2209)

2009-05-26 10:23:04,025 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_-5608797308669010722_1275 bad datanode[0]
192.168.240.175:50010
2009-05-26 10:23:04,025 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_-5608797308669010722_1275 in pipeline
192.168.240.175:50010, 192.168.240.179:50010: bad datanode
192.168.240.175:50010
2009-05-26 10:23:23,536 INFO org.apache.hadoop.hbase.regionserver.HRegion:
compaction completed on region tableA,ROW_KEY,1243357190459 in 31sec
2009-05-26 10:23:24,031 INFO org.apache.hadoop.hbase.regionserver.HLog:
Closed
hdfs://ats181:54310/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358537229,
entries=100001. New log writer:
/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358604023
2009-05-26 10:24:06,109 INFO org.apache.hadoop.hbase.regionserver.HRegion:
starting  compaction on region tableA,ROW_KEY,1243357190459
2009-05-26 10:24:06,180 INFO org.apache.hadoop.hbase.regionserver.HLog:
Closed
hdfs://ats181:54310/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358570622,
entries=155001. New log writer:
/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358646150
2009-05-26 10:24:06,214 INFO org.apache.hadoop.hbase.regionserver.HRegion:
compaction completed on region tableA,ROW_KEY,1243357190459 in 0sec
2009-05-26 10:24:39,502 INFO org.apache.hadoop.hbase.regionserver.HLog:
Closed
hdfs://ats181:54310/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358604023,
entries=105000. New log writer:
/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358679496
2009-05-26 10:24:51,089 INFO org.apache.hadoop.hbase.regionserver.HRegion:
starting  compaction on region tableA,ROW_KEY,1243357190459
2009-05-26 10:24:54,372 INFO org.apache.hadoop.hbase.regionserver.HRegion:
compaction completed on region tableA,ROW_KEY,1243357190459 in 3sec
2009-05-26 10:25:22,844 INFO org.apache.hadoop.hbase.regionserver.HLog:
Closed
hdfs://ats181:54310/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358646150,
entries=100001. New log writer:
/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358722837
2009-05-26 10:25:28,110 INFO org.apache.hadoop.hbase.regionserver.HRegion:
starting  compaction on region tableA,ROW_KEY,1243357190459
2009-05-26 10:25:28,132 INFO org.apache.hadoop.hbase.regionserver.HLog:
Closed
hdfs://ats181:54310/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358679496,
entries=110001. New log writer:
/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358728126
2009-05-26 10:25:28,158 INFO org.apache.hadoop.hbase.regionserver.HRegion:
compaction completed on region tableA,ROW_KEY,1243357190459 in 0sec
2009-05-26 10:25:59,007 INFO org.apache.hadoop.hbase.regionserver.HLog:
Closed
hdfs://ats181:54310/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358722837,
entries=105000. New log writer:
/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358758999
2009-05-26 10:26:38,530 INFO org.apache.hadoop.hbase.regionserver.HRegion:
starting  compaction on region tableA,ROW_KEY,1243357190459
2009-05-26 10:26:38,975 INFO org.apache.hadoop.hbase.regionserver.HLog:
Closed
hdfs://ats181:54310/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358728126,
entries=120001. New log writer:
/hbase/log_192.168.240.175_1243356336827_60020/hlog.dat.1243358798947

=====================================================================




stack-3 wrote:
> 
> On Tue, May 26, 2009 at 8:41 AM, llpind <[email protected]> wrote:
>>
>> Haven't tried the RC yet, but regions do get lost when i do intensive
>> write
>> operations (e.g. they are no longer listed under online regions).
> 
> 
> Have you tried writing less virulently?  Maybe we just can't take your
> write load on your setup.
> 
> Above you say all writes are going to one region only.   How many rows
> have gone in in your estimation?  This is the case when hbase starts
> up;  all writes go to a single region until sufficient to split.
> Maybe enough data has not yet gone in?  Perhaps your write rate is
> such that hbase is unable to split?  Try taking load off.  Try
> manually splitting regions (See the hbase shell.  Type 'toos' to see
> list of admin methods).
> 
> I would suggest you not change default flush and region sizes.
> We're better able to help if sizes are default.
> 
>>
>> Will this RC work with Haoop version 0.19.1?
> 
> 
> Yes.  An hbase will run on a hadoop of same major and minor version
> (They can different in the point version number).
> 
>>
>> When I mentioned i tried different configurations, I was tweaking
>> different
>> properties mentioned in that post.  We are still at a lost for what to
>> do.
>>
>>
> List is short:
> 
> + Up your file descriptors.  1024 is not enough.
> + Up your hadoop xceivers.  256 is too little
> + Set the timeout on dfsclient to 0.
> 
> For more detail on how to do above configurations, their exact names
> and whether client or server-side config, see the FAQ and
> troubleshooting.
> 
> Let us know the particular issues you are running into.  We'd like to help
> out.
> 
> St.Ack
> 
> 
> 
> 
> 
>>
>> stack-3 wrote:
>>>
>>> The RC has fixes to help w/ regionserver/master disagreement as to who
>>> has what regions; i.e. "region loss".  You might give it a go?
>>> St.Ack
>>>
>>> On Sun, May 24, 2009 at 10:33 AM, llpind <[email protected]> wrote:
>>>>
>>>> Hey Stack, I'm using 0.19.1.  Also, would like to know if I should
>>>> check
>>>> out
>>>> the latest and try that or try the RC you mentioned above.
>>>>
>>>>
>>>> stack-3 wrote:
>>>>>
>>>>> Are you using TRUNK (If you have answered this question already,
>>>>> please excuse my not remembering)?
>>>>>
>>>>> St.Ack
>>>>>
>>>>> On Sat, May 23, 2009 at 2:17 PM, llpind <[email protected]>
>>>>> wrote:
>>>>>>
>>>>>> I see similar behavior in my small cluster.  (1 master, 3 datanodes)
>>>>>>
>>>>>> I am also planning on trying this RC version.  I've tried various
>>>>>> configurations, and I continue to lose Regions with intensive writes.
>>>>>>  I
>>>>>> really hope something like this will work, because we are starting to
>>>>>> consider other options now.
>>>>>>
>>>>>> --
>>>>>> View this message in context:
>>>>>> http://www.nabble.com/HBase-looses-regions.-tp23657983p23688361.html
>>>>>> Sent from the HBase User mailing list archive at Nabble.com.
>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>
>>>> --
>>>> View this message in context:
>>>> http://www.nabble.com/HBase-looses-regions.-tp23657983p23696025.html
>>>> Sent from the HBase User mailing list archive at Nabble.com.
>>>>
>>>>
>>>
>>>
>>
>> --
>> View this message in context:
>> http://www.nabble.com/HBase-looses-regions.-tp23657983p23725775.html
>> Sent from the HBase User mailing list archive at Nabble.com.
>>
>>
> 
> 

-- 
View this message in context: 
http://www.nabble.com/HBase-looses-regions.-tp23657983p23727877.html
Sent from the HBase User mailing list archive at Nabble.com.

Reply via email to