Were they opening the same region by any chance?

On Mon, Apr 11, 2011 at 6:13 PM, 陈加俊 <[email protected]> wrote:
> There is no big scan,and just norma load. Also strange is when one RS exited
> then another RS exited and others RS like that.
>
> On Tue, Apr 12, 2011 at 8:55 AM, Jean-Daniel Cryans 
> <[email protected]>wrote:
>
>> Ok that looks "fine", did the region server die under heavy load by
>> any chance? Or was it big scans? Or just normal load?
>>
>> J-D
>>
>> On Mon, Apr 11, 2011 at 5:50 PM, 陈加俊 <[email protected]> wrote:
>> > my configuration is follows:
>> >  <property>
>> >                 <name>hbase.client.write.buffer</name>
>> >                 <value>2097152</value>
>> >                 <description>
>> >                         1024*1024*2=2097152
>> >                 </description>
>> >         </property>
>> >         <property>
>> >                 <name>hbase.hstore.blockingStoreFiles</name>
>> >                 <value>14</value>
>> >         </property>
>> >         <property>
>> >                 <name>hbase.hregion.memstore.block.multiplier</name>
>> >                 <value>2</value>
>> >         </property>
>> >         <property>
>> >                 <name>hbase.regionserver.handler.count</name>
>> >                 <value>30</value>
>> >         </property>
>> >  <property>
>> >                 <name>zookeeper.session.timeout</name>
>> >                 <value>90000</value>
>> >         </property>
>> >         <property>
>> >                  <name>hbase.zookeeper.property.tickTime</name>
>> >                  <value>9000</value>
>> >         </property>
>> >         <property>
>> >                  <name>hbase.hregion.max.filesize</name>
>> >                  <value>1073741824</value>
>> >                 <description>
>> >                 Maximum HStoreFile size. If any one of a column families'
>> > HStoreFiles has grown to exceed this value,
>> >                 the hosting HRegion is split in two. Default: 256M.
>> >                 1024*1024*1024=1073741824 (1G)
>> >                 </description>
>> >         </property>
>> >
>> > On Tue, Apr 12, 2011 at 8:43 AM, Jean-Daniel Cryans <[email protected]
>> >
>> > wrote:
>> >>
>> >> And where will they go? The issue isn't the number of regions per say,
>> >> it's the amount of data being served by that region server. Also I
>> >> still don't know if that's really your issue or it's a configuration
>> >> issue (which I have yet to see).
>> >>
>> >> J-D
>> >>
>> >> On Mon, Apr 11, 2011 at 5:41 PM, 陈加俊 <[email protected]> wrote:
>> >> > Can I limit the numbers of regions  on one RegionServer ?
>> >> >
>> >> > On Tue, Apr 12, 2011 at 8:37 AM, Jean-Daniel Cryans
>> >> > <[email protected]>wrote:
>> >> >
>> >> >> It's really a lot yes, but it could also be weird configurations or
>> >> >> too big values.
>> >> >>
>> >> >> J-D
>> >> >>
>> >> >> On Mon, Apr 11, 2011 at 5:35 PM, 陈加俊 <[email protected]> wrote:
>> >> >> > Is it too many regions ? Is the memory enough ?
>> >> >> > HBase-0.20.6
>> >> >> >
>> >> >> > 2011-04-12 00:16:31,844 FATAL
>> >> >> > org.apache.hadoop.hbase.regionserver.HRegionServer:
>> OutOfMemoryError,
>> >> >> > aborting.
>> >> >> > java.lang.OutOfMemoryError: Java heap space
>> >> >> >        at
>> >> >> java.io.BufferedInputStream.<init>(BufferedInputStream.java:178)
>> >> >> >        at
>> >> >> >
>> >> >>
>> >> >>
>> org.apache.hadoop.hdfs.DFSClient$BlockReader.newBlockReader(DFSClient.java:1380)
>> >> >> >        at
>> >> >> >
>> >> >>
>> >> >>
>> org.apache.hadoop.hdfs.DFSClient$DFSInputStream.blockSeekTo(DFSClient.java:1648)
>> >> >> >        at
>> >> >> >
>> >> >> >
>> org.apache.hadoop.hdfs.DFSClient$DFSInputStream.read(DFSClient.java:1767)
>> >> >> >        at
>> java.io.DataInputStream.readFully(DataInputStream.java:178)
>> >> >> >        at
>> java.io.DataInputStream.readFully(DataInputStream.java:152)
>> >> >> >        at
>> >> >> >
>> >> >>
>> >> >>
>> org.apache.hadoop.hbase.io.hfile.HFile$FixedFileTrailer.deserialize(HFile.java:1368)
>> >> >> >        at
>> >> >> >
>> >> >> >
>> org.apache.hadoop.hbase.io.hfile.HFile$Reader.readTrailer(HFile.java:848)
>> >> >> >        at
>> >> >> >
>> >> >>
>> >> >>
>> org.apache.hadoop.hbase.io.hfile.HFile$Reader.loadFileInfo(HFile.java:793)
>> >> >> >        at
>> >> >> >
>> >> >> >
>> org.apache.hadoop.hbase.regionserver.StoreFile.open(StoreFile.java:273)
>> >> >> >        at
>> >> >> >
>> >> >> >
>> org.apache.hadoop.hbase.regionserver.StoreFile.<init>(StoreFile.java:129)
>> >> >> >        at
>> >> >> >
>> >> >> >
>> org.apache.hadoop.hbase.regionserver.Store.loadStoreFiles(Store.java:410)
>> >> >> >        at
>> >> >> org.apache.hadoop.hbase.regionserver.Store.<init>(Store.java:221)
>> >> >> >        at
>> >> >> >
>> >> >>
>> >> >>
>> org.apache.hadoop.hbase.regionserver.HRegion.instantiateHStore(HRegion.java:1636)
>> >> >> >        at
>> >> >> >
>> >> >> >
>> org.apache.hadoop.hbase.regionserver.HRegion.initialize(HRegion.java:321)
>> >> >> >        at
>> >> >> >
>> >> >>
>> >> >>
>> org.apache.hadoop.hbase.regionserver.HRegionServer.instantiateRegion(HRegionServer.java:1566)
>> >> >> >        at
>> >> >> >
>> >> >>
>> >> >>
>> org.apache.hadoop.hbase.regionserver.HRegionServer.openRegion(HRegionServer.java:1533)
>> >> >> >        at
>> >> >> >
>> >> >>
>> >> >>
>> org.apache.hadoop.hbase.regionserver.HRegionServer$Worker.run(HRegionServer.java:1453)
>> >> >> >        at java.lang.Thread.run(Thread.java:619)
>> >> >> > 2011-04-12 00:16:31,847 INFO
>> >> >> > org.apache.hadoop.hbase.regionserver.HRegionServer: Dump of
>> metrics:
>> >> >> > request=0.0, regions=1624, stores=3075, storefiles=2526,
>> >> >> > storefileIndexSize=374, memstoreSize=339, compactionQueueSize=0,
>> >> >> > usedHeap=3990, maxHeap=3991, blockCacheSize=706107016,
>> >> >> > blockCacheFree=131010552, blockCacheCount=10631,
>> >> >> > blockCacheHitRatio=58,
>> >> >> > fsReadLatency=0, fsWriteLatency=0, fsSyncLatency=0
>> >> >> > 2011-04-12 00:16:31,848 INFO
>> >> >> > org.apache.hadoop.hbase.regionserver.HRegionServer: worker thread
>> >> >> > exiting
>> >> >> >
>> >> >> >
>> >> >> > --
>> >> >> > Thanks & Best regards
>> >> >> > jiajun
>> >> >> >
>> >> >>
>> >> >
>> >> >
>> >> >
>> >> > --
>> >> > Thanks & Best regards
>> >> > jiajun
>> >> >
>> >
>> >
>> >
>> > --
>> > Thanks & Best regards
>> > jiajun
>> >
>> >
>>
>
>
>
> --
> Thanks & Best regards
> jiajun
>

Reply via email to