WARN add_table: Missing .regioninfo:.. No server address.. what to do?

2010-08-23 Thread Stuart Smith
Hello, I'm missing several .regioninfo files for several of my regions. Not sure why or when. But after an addtable rebuild of the .META. entries - I get errors about "No server address listed in .META. for region". Which I guess would make sense if add_table failed to update the .META. tabl

Re: WARN add_table: Missing .regioninfo:.. No server address.. what to do?

2010-08-23 Thread Stuart Smith
ut when I came back 4/10 regionservers had died. So a lot of random crashes going on.. Take care, -stu --- On Mon, 8/23/10, Stuart Smith wrote: > From: Stuart Smith > Subject: WARN add_table: Missing .regioninfo:.. No server address.. what to > do? > To: user@hbase.apache.org &

Re: WARN add_table: Missing .regioninfo:.. No server address.. what to do?

2010-08-25 Thread Stuart Smith
-stu --- On Mon, 8/23/10, Stack wrote: > From: Stack > Subject: Re: WARN add_table: Missing .regioninfo:.. No server address.. what > to do? > To: user@hbase.apache.org > Date: Monday, August 23, 2010, 6:08 PM > On Mon, Aug 23, 2010 at 1:35 PM, > Stuart Smith > wrote

Re: WARN add_table: Missing .regioninfo:.. No server address.. what to do?

2010-08-25 Thread Stuart Smith
ioninfo:.. No server address.. what > to do? > To: user@hbase.apache.org > Date: Thursday, August 26, 2010, 2:11 AM > On Wed, Aug 25, 2010 at 11:22 AM, > Stuart Smith > wrote: > > Just curious, though, (if it happens again) - assume > the regions were invalid - I don

Is this indicative of a GC CPU starvation?

2010-11-09 Thread Stuart Smith
Hello, I just wanted to make sure that I'm interpreting a series of common issues correctly. I saw ZK expirations causing regionserver failures, and this in a GC log of one of the regionservers: 16237.033: [GC[YG occupancy: 22353 K (38336 K)]16245.298: [Rescan (parallel) , 0.0264040 secs]1

Re: Is this indicative of a GC CPU starvation?

2010-11-13 Thread Stuart Smith
t; from out of box hbase-env.sh? > > try running this on RS and watch last column, each > increment should be small > sudo -u jstats -gcutil 1000 > > > > On Tue, Nov 9, 2010 at 10:53 AM, Stuart Smith > wrote: > > > Hello, > > > >  I just want

Re: Dealing with single point of failure

2011-10-29 Thread Stuart Smith
I was under the impression you could use HBase with a different distributed filesystem (other than HDFS). That would fix your SPOF. HBase has other issues quite frankly (and I use it, can keep planning on using it). Mainly due to the fact it's under quite heavy development, but I don't think Ca

Re: PENDING_CLOSE for too long

2011-10-29 Thread Stuart Smith
Hello Geoff,   I usually don't show up here, since I use CDH, and good form means I should stay on CDH-users, But!   I've been seeing the same issues for months:  - PENDING_CLOSE too long, master tries to reassign - I see an continuous stream of these.  - WrongRegionExceptions due to overlappin

Re: PENDING_CLOSE for too long

2011-11-14 Thread Stuart Smith
Take care,   -stu From: Geoff Hendrey To: user@hbase.apache.org Cc: user@hbase.apache.org; Stuart Smith Sent: Saturday, October 29, 2011 7:08 PM Subject: Re: PENDING_CLOSE for too long Stuart - Have you disabled splitting? I believe you can work around the issue of PENDInG_CLOSE by prespli

Re: Different value of integer on querying in hbase!

2011-12-14 Thread Stuart Smith
Hello,   How did you query base via a statement object? Are you using Hive? Or is this some new interface I don't know about.. I always had to use Get() or Scan(). And hbase stores everything as bytes, not strings.. unlike C, in java, there is a difference ;) Take care,  -stu _

Re: Questions on timestamps, insights on how timerange/timestamp filter are processed?

2011-12-14 Thread Stuart Smith
Hello Thomas,    Someone here could probably provide more help, but to start you off, the only way I've filtered timestamps is to do a scan, and just filter out rows one by one. This definitely sounds like something coprocessors could help with, but I don't really understand those yet, so someo

Re: Questions on timestamps, insights on how timerange/timestamp filter are processed?

2011-12-14 Thread Stuart Smith
doing this vs doing it client side? Or is it basically the same amount of work - a full scan checking & skipping timestamps.. ? Take care,   -stu From: Carson Hoffacker To: user@hbase.apache.org; Stuart Smith Sent: Wednesday, December 14, 2011 10:29

Re: Different value of integer on querying in hbase!

2011-12-16 Thread Stuart Smith
?), and strings work just fine, so it's all good. Take care,   -stu From: neuron005 To: hbase-u...@hadoop.apache.org Sent: Thursday, December 15, 2011 9:52 PM Subject: Re: Different value of integer on querying in hbase! Stuart Smith-8 wrote:

Ubuntu pulling sun java from partner archives.. Open JDK?

2011-12-19 Thread Stuart Smith
Hello,   Ubuntu is currently pulling the sun java plugin from ubuntu, and will be pulling all sun java packages from partner archives, replacing the current packages with empty ones. They have stated this will break all packages depending on Sun Java: https://lwn.net/Articles/472466/ Thought I

Re: Ubuntu pulling sun java from partner archives.. Open JDK?

2011-12-19 Thread Stuart Smith
I was afraid of that.. thanks! Out of curiousity, though, what version of openjdk did you use? Supposedly version 7 should be much better, as it is now the reference implementation.. but I have no idea how much of that is true, and how much is spin... Take care,    -stu

Trying to write too much to stdout destabilizes cluster across reboots

2010-07-13 Thread Stuart Smith
Hello, I have a M/R task running over hbase tables. I added a one line of output in my reduce tasks running over a set of tens of millions of records. This was a little silly, but I was kind of curious what would happen. Well the reduce ran very slowly, and eventually, I had to kill th

Re: Trying to write too much to stdout destabilizes cluster across reboots

2010-07-13 Thread Stuart Smith
eta. We'd never recover. " They mean never as in _never_ or just as in until the next reboot. Or is my cluster (and my data) dead and gone now? Take care, -stu --- On Tue, 7/13/10, Stuart Smith wrote: > From: Stuart Smith > Subject: Trying to write too much to stdout destabilizes

Re: Trying to write too much to stdout destabilizes cluster across reboots

2010-07-13 Thread Stuart Smith
stdout. Take care, -stu --- On Tue, 7/13/10, Stuart Smith wrote: > From: Stuart Smith > Subject: Re: Trying to write too much to stdout destabilizes cluster across > reboots > To: user@hbase.apache.org > Date: Tuesday, July 13, 2010, 9:10 PM > One follow up. > > I

Cluster destabilizes recovering from crash

2010-07-14 Thread Stuart Smith
start all regionservers" script. Is there anything bad about trying to start regionservers via ./hbase start regionserver every minute or so? They die off pretty quick. Take care, -stu --- On Tue, 7/13/10, Stuart Smith wrote: > From: Stuart Smith > Subject: Re: Trying to write too m

Re: Cluster destabilizes recovering from crash

2010-07-14 Thread Stuart Smith
tion in createBlockOutputStream java.io.EOFException INFO org.apache.hadoop.hdfs.DFSClient: Abandoning block blk_ I already did the file handle thing a while ago. Let's see if this fixes it! Take care, -stu --- On Wed, 7/14/10, Stuart Smith wrote: > From: Stuart Smith > Subject: Cl

Re: Cluster destabilizes recovering from crash

2010-07-14 Thread Stuart Smith
7;t fix HDFS. > > J-D > > On Wed, Jul 14, 2010 at 5:53 PM, Stuart Smith > wrote: > > > > Ok, back up again. > > > > This time, in addition to just watching & > restarting, the error logs led me to this: > > > > http://mail-archives.apach

RE: Run MR job when my data stays in hbase?

2010-07-19 Thread Stuart Smith
Hello, You can ignore this if you're already rock solid on writing M/R jobs, but just in case you're as new to this as I am: Be careful you have all your dependencies lined up in the jar you're creating your M/R job in. If you're using Eclipse this means selecting "Extract required librari

Table goes offline - temporary outage + Retries Exhausted (related?)

2010-07-29 Thread Stuart Smith
Hello, I have two problems that may or may not be related. One is trying to figure out a self-correcting outage I had last evening. I noticed issues starting with clients reporting: RetriesExhaustedException: Trying to contact region server Some server... I didn't see much going on in the re

Re: Table goes offline - temporary outage + Retries Exhausted (related?)

2010-07-29 Thread Stuart Smith
t the Puts() as somewhat unreliable? Take care, -stu --- On Thu, 7/29/10, Stuart Smith wrote: > From: Stuart Smith > Subject: Table goes offline - temporary outage + Retries Exhausted (related?) > To: user@hbase.apache.org > Date: Thursday, July 29, 2010, 2:09 PM > He

Re: Table goes offline - temporary outage + Retries Exhausted (related?)

2010-07-29 Thread Stuart Smith
ay uploads were failing due to the batch issues, and the gui reported the table as disabled, and I connected the two issues incorrectly. Take care, -stu --- On Thu, 7/29/10, Stuart Smith wrote: > From: Stuart Smith > Subject: Re: Table goes offline - temporary outage + Retries Exhausted

Re: Table goes offline - temporary outage + Retries Exhausted (related?)

2010-07-29 Thread Stuart Smith
> > Lastly, I dont think that Put should be unreliable, I have > reliably > imported 10s of billions of rows, so there is something > else going on. > > -ryan > PS: mysql dbas spend tons of time setting up ulimits and > other > esoteric kernel tuning parameters, our r

Re: Table goes offline - temporary outage + Retries Exhausted (related?)

2010-07-29 Thread Stuart Smith
unately it > means lots of log > groveling.  If you want to post logs, try pastebin.com > instead of > trying to attach files. > > Dig some dirt up and lets check it out :-) > > -ryan > > On Thu, Jul 29, 2010 at 4:25 PM, Stuart Smith > wrote: > > Hello Ry

Batch puts interrupted ... Requested row out of range for HRegion filestore ...org.apache.hadoop.hbase.client.RetriesExhaustedException:

2010-08-06 Thread Stuart Smith
Hello, I'm running hbase 0.20.5, and seeing Puts() fail repeatedly when trying to insert a specific item into the database. Client side I see: org.apache.hadoop.hbase.client.RetriesExhaustedException: Trying to contact region server Some server, retryOnlyOne=true, index=0, islastrow=true, tr

Re: Batch puts interrupted ... Requested row out of range for HRegion filestore ...org.apache.hadoop.hbase.client.RetriesExhaustedException:

2010-08-06 Thread Stuart Smith
nd the wrong region and the > regionserver would > reject the request (correctly so). > > That is what is probably happening here.  Check the > table dump in the > master web-ui and see if you can find a 'hole'... where the > end-key > doesnt match up with the star

Re: Batch puts interrupted ... Requested row out of range for HRegion filestore ...org.apache.hadoop.hbase.client.RetriesExhaustedException:

2010-08-06 Thread Stuart Smith
Just to follow up - I ran add_table as I had done when I lost a table before - and it fixed the error. Thanks! Take care, -stu --- On Fri, 8/6/10, Stuart Smith wrote: > From: Stuart Smith > Subject: Re: Batch puts interrupted ... Requested row out of range for > HRegion

Avoiding OutOfMemory Java heap space in region servers

2010-08-10 Thread Stuart Smith
Hello, I'm seeing errors like so: 010-08-10 12:58:38,938 DEBUG org.apache.hadoop.hbase.client.HConnectionManager$ClientZKWatcher: Got ZooKeeper event, state: Disconnected, type: None, path: null 2010-08-10 12:58:38,939 INFO org.apache.hadoop.hbase.regionserver.HRegionServer: Got ZooKeeper e

Re: Avoiding OutOfMemory Java heap space in region servers

2010-08-12 Thread Stuart Smith
andlers could throw you over especially if > your heap > is small. > > What size heap are you running with? > > St.Ack > > > > On Tue, Aug 10, 2010 at 3:26 PM, Stuart Smith > wrote: > > Hello, > > > >   I'm seeing errors like so: > &g

Stream interface to cell Data? Was -> Re: Avoiding OutOfMemory Java heap space in region servers

2010-08-18 Thread Stuart Smith
Hello, I was wondering if there are any plans for a stream interface to Cell data. I saw this: > > or they are using large client write buffers so big > payloads are being > > passed to the server in each RPC request. Our > RPC is not streaming. So I'm guessing there's not one now (and I co