Hello,
I'm missing several .regioninfo files for several of my regions. Not sure why
or when. But after an addtable rebuild of the .META. entries - I get errors
about "No server address listed in .META. for region".
Which I guess would make sense if add_table failed to update the .META. tabl
ut when I
came back 4/10 regionservers had died. So a lot of random crashes going on..
Take care,
-stu
--- On Mon, 8/23/10, Stuart Smith wrote:
> From: Stuart Smith
> Subject: WARN add_table: Missing .regioninfo:.. No server address.. what to
> do?
> To: user@hbase.apache.org
&
-stu
--- On Mon, 8/23/10, Stack wrote:
> From: Stack
> Subject: Re: WARN add_table: Missing .regioninfo:.. No server address.. what
> to do?
> To: user@hbase.apache.org
> Date: Monday, August 23, 2010, 6:08 PM
> On Mon, Aug 23, 2010 at 1:35 PM,
> Stuart Smith
> wrote
ioninfo:.. No server address.. what
> to do?
> To: user@hbase.apache.org
> Date: Thursday, August 26, 2010, 2:11 AM
> On Wed, Aug 25, 2010 at 11:22 AM,
> Stuart Smith
> wrote:
> > Just curious, though, (if it happens again) - assume
> the regions were invalid - I don
Hello,
I just wanted to make sure that I'm interpreting a series of common issues
correctly.
I saw ZK expirations causing regionserver failures, and this in a GC log of one
of the regionservers:
16237.033: [GC[YG occupancy: 22353 K (38336 K)]16245.298: [Rescan (parallel) ,
0.0264040 secs]1
t; from out of box hbase-env.sh?
>
> try running this on RS and watch last column, each
> increment should be small
> sudo -u jstats -gcutil 1000
>
>
>
> On Tue, Nov 9, 2010 at 10:53 AM, Stuart Smith
> wrote:
>
> > Hello,
> >
> > I just want
I was under the impression you could use HBase with a different distributed
filesystem (other than HDFS).
That would fix your SPOF.
HBase has other issues quite frankly (and I use it, can keep planning on using
it).
Mainly due to the fact it's under quite heavy development, but I don't think
Ca
Hello Geoff,
I usually don't show up here, since I use CDH, and good form means I should
stay on CDH-users,
But!
I've been seeing the same issues for months:
- PENDING_CLOSE too long, master tries to reassign - I see an continuous
stream of these.
- WrongRegionExceptions due to overlappin
Take care,
-stu
From: Geoff Hendrey
To: user@hbase.apache.org
Cc: user@hbase.apache.org; Stuart Smith
Sent: Saturday, October 29, 2011 7:08 PM
Subject: Re: PENDING_CLOSE for too long
Stuart -
Have you disabled splitting? I believe you can work around the issue of
PENDInG_CLOSE by prespli
Hello,
How did you query base via a statement object? Are you using Hive?
Or is this some new interface I don't know about.. I always had to use Get() or
Scan().
And hbase stores everything as bytes, not strings.. unlike C, in java, there is
a difference ;)
Take care,
-stu
_
Hello Thomas,
Someone here could probably provide more help, but to start you off, the
only way I've filtered timestamps is to do a scan, and just filter out rows one
by one. This definitely sounds like something coprocessors could help with, but
I don't really understand those yet, so someo
doing this vs doing it client side?
Or is it basically the same amount of work - a full scan checking & skipping
timestamps.. ?
Take care,
-stu
From: Carson Hoffacker
To: user@hbase.apache.org; Stuart Smith
Sent: Wednesday, December 14, 2011 10:29
?), and strings work
just fine, so it's all good.
Take care,
-stu
From: neuron005
To: hbase-u...@hadoop.apache.org
Sent: Thursday, December 15, 2011 9:52 PM
Subject: Re: Different value of integer on querying in hbase!
Stuart Smith-8 wrote:
Hello,
Ubuntu is currently pulling the sun java plugin from ubuntu, and will be
pulling all sun java packages from partner archives, replacing the current
packages with empty ones. They have stated this will break all packages
depending on Sun Java:
https://lwn.net/Articles/472466/
Thought I
I was afraid of that.. thanks!
Out of curiousity, though, what version of openjdk did you use?
Supposedly version 7 should be much better, as it is now the reference
implementation.. but I have no idea how much of that is true, and how much is
spin...
Take care,
-stu
Hello,
I have a M/R task running over hbase tables. I added
a one line of output in my reduce tasks running over a set
of tens of millions of records. This was a little silly, but
I was kind of curious what would happen.
Well the reduce ran very slowly, and eventually, I had to
kill th
eta. We'd never recover. "
They mean never as in _never_ or just as in until the next reboot. Or is my
cluster (and my data) dead and gone now?
Take care,
-stu
--- On Tue, 7/13/10, Stuart Smith wrote:
> From: Stuart Smith
> Subject: Trying to write too much to stdout destabilizes
stdout.
Take care,
-stu
--- On Tue, 7/13/10, Stuart Smith wrote:
> From: Stuart Smith
> Subject: Re: Trying to write too much to stdout destabilizes cluster across
> reboots
> To: user@hbase.apache.org
> Date: Tuesday, July 13, 2010, 9:10 PM
> One follow up.
>
> I
start all regionservers" script. Is there anything bad about trying
to start regionservers via ./hbase start regionserver every minute or so? They
die off pretty quick.
Take care,
-stu
--- On Tue, 7/13/10, Stuart Smith wrote:
> From: Stuart Smith
> Subject: Re: Trying to write too m
tion in
createBlockOutputStream java.io.EOFException
INFO org.apache.hadoop.hdfs.DFSClient: Abandoning
block blk_
I already did the file handle thing a while ago.
Let's see if this fixes it!
Take care,
-stu
--- On Wed, 7/14/10, Stuart Smith wrote:
> From: Stuart Smith
> Subject: Cl
7;t fix HDFS.
>
> J-D
>
> On Wed, Jul 14, 2010 at 5:53 PM, Stuart Smith
> wrote:
> >
> > Ok, back up again.
> >
> > This time, in addition to just watching &
> restarting, the error logs led me to this:
> >
> > http://mail-archives.apach
Hello,
You can ignore this if you're already rock solid on writing M/R jobs, but
just in case you're as new to this as I am:
Be careful you have all your dependencies lined up in the jar you're creating
your M/R job in. If you're using Eclipse this means selecting "Extract
required librari
Hello,
I have two problems that may or may not be related.
One is trying to figure out a self-correcting outage I had last evening.
I noticed issues starting with clients reporting:
RetriesExhaustedException: Trying to contact region server Some server...
I didn't see much going on in the re
t the Puts() as somewhat
unreliable?
Take care,
-stu
--- On Thu, 7/29/10, Stuart Smith wrote:
> From: Stuart Smith
> Subject: Table goes offline - temporary outage + Retries Exhausted (related?)
> To: user@hbase.apache.org
> Date: Thursday, July 29, 2010, 2:09 PM
> He
ay uploads were failing due to the batch issues, and the
gui reported the table as disabled, and I connected the two issues incorrectly.
Take care,
-stu
--- On Thu, 7/29/10, Stuart Smith wrote:
> From: Stuart Smith
> Subject: Re: Table goes offline - temporary outage + Retries Exhausted
>
> Lastly, I dont think that Put should be unreliable, I have
> reliably
> imported 10s of billions of rows, so there is something
> else going on.
>
> -ryan
> PS: mysql dbas spend tons of time setting up ulimits and
> other
> esoteric kernel tuning parameters, our r
unately it
> means lots of log
> groveling. If you want to post logs, try pastebin.com
> instead of
> trying to attach files.
>
> Dig some dirt up and lets check it out :-)
>
> -ryan
>
> On Thu, Jul 29, 2010 at 4:25 PM, Stuart Smith
> wrote:
> > Hello Ry
Hello,
I'm running hbase 0.20.5, and seeing Puts() fail repeatedly when trying to
insert a specific item into the database.
Client side I see:
org.apache.hadoop.hbase.client.RetriesExhaustedException: Trying to contact
region server Some server, retryOnlyOne=true, index=0, islastrow=true, tr
nd the wrong region and the
> regionserver would
> reject the request (correctly so).
>
> That is what is probably happening here. Check the
> table dump in the
> master web-ui and see if you can find a 'hole'... where the
> end-key
> doesnt match up with the star
Just to follow up - I ran add_table as I had done when I lost a table before -
and it fixed the error.
Thanks!
Take care,
-stu
--- On Fri, 8/6/10, Stuart Smith wrote:
> From: Stuart Smith
> Subject: Re: Batch puts interrupted ... Requested row out of range for
> HRegion
Hello,
I'm seeing errors like so:
010-08-10 12:58:38,938 DEBUG
org.apache.hadoop.hbase.client.HConnectionManager$ClientZKWatcher: Got
ZooKeeper event, state: Disconnected, type: None, path: null
2010-08-10 12:58:38,939 INFO
org.apache.hadoop.hbase.regionserver.HRegionServer: Got ZooKeeper e
andlers could throw you over especially if
> your heap
> is small.
>
> What size heap are you running with?
>
> St.Ack
>
>
>
> On Tue, Aug 10, 2010 at 3:26 PM, Stuart Smith
> wrote:
> > Hello,
> >
> > I'm seeing errors like so:
> &g
Hello,
I was wondering if there are any plans for a stream interface to Cell data. I
saw this:
> > or they are using large client write buffers so big
> payloads are being
> > passed to the server in each RPC request. Our
> RPC is not streaming.
So I'm guessing there's not one now (and I co
33 matches
Mail list logo