Re: Hbase tables disappear

2013-01-15 Thread Ibrahim Yakti
HBase Version: 0.92.1-cdh4.1.2, rUnknown, Thu Nov 1 18:03:50 PDT 2012 Sqoop Version: 1.4.1-cdh4.1.2 Stack: yes, it seems that hbase default root directory is /tmp working to fix this now. I am writing the data for the first time into HBase from MySQL, not overwriting it. -- Ibrahim On Mon,

Re: Slow start of HBase operations with YCSB, possibly because of zookeeper ?

2013-01-15 Thread samar kumar
YCSB would be writing all data to one table.. So initially when the table is small or just created all the writes would go to one RS.. As the table grows the Region is split into different RS. The would allow parallel writes, if the keys are random and could possibly make the writes faster. Samar

Re: Hbase tables disappear

2013-01-15 Thread Ibrahim Yakti
The root directory is set to /var/lib/hbase/ The count issue is still exist: *Sqoop: * 13/01/15 08:55:23 INFO mapreduce.ImportJobBase: Retrieved 1754285 records. *MySQL:* +--+ | count(1) | +--+ | 1754303 | +--+ *HBase:* 104210 row(s) in 11.4980 seconds running count

RegionServer level co-processor

2013-01-15 Thread 蔡斌
HiWe try to build an online machine learning system, like Jubatus(http://www-conf.slac.stanford.edu/xldb2012/talks/xldb2012_wed_LT09_HMakino.pdf), based on HBase. Our training data will be stored in HBase, and algorithms will scan these data to build model. We need a region server level

Re: RegionServer level co-processor

2013-01-15 Thread Bing Jiang
In my view, if you want to realize Jubtaus, you can use Endpoint. Which is different from RegionServerObserver. A client will launch a task about a table, and it will execute an Endpoint in each region that belongs to that table. 2013/1/15 蔡斌 caibinb...@hotmail.com HiWe try to build

Re: RegionServer level co-processor

2013-01-15 Thread Jean-Marc Spaggiari
Why not running a MR job? It will run a task per region. Le 15 janv. 2013 06:38, Bing Jiang jiangbinglo...@gmail.com a écrit : In my view, if you want to realize Jubtaus, you can use Endpoint. Which is different from RegionServerObserver. A client will launch a task about a table, and it

Re: How to upgrade HBase from 0.90.5 to 0.94

2013-01-15 Thread Ted
You can upgrade to 0.92.2 first, followed by upgrade to 0.94.4 Please consider upgrading Hadoop to 1.0.4 as well Thanks On Jan 15, 2013, at 4:47 AM, 幻风 huanfeng...@gmail.com wrote: Hi, everybody, I want to upgrade a HBase cluster from 0.90.5 to 0.94. The cluster has about 50 nodes and the

Hbase hbck -repair slow

2013-01-15 Thread Lashing
We recently suffered meta corrupt issue. We tried to use hbck -repair to fix But the performance is not fast (about 1 region for 1 second). Are there any better way to speed up the process? Btw, My Hbase is 0.90.6. Thanks.

Re: Hbase hbck -repair slow

2013-01-15 Thread Kevin O'dell
Lashing, Without knowning the exact errors you are seeing it would be hard to make recommendations. If it is just META that is messed up you can use: hbase hbck -fixMeta -fixAssignments I think(not sure) in that version you can do hbase hbck -repair tablename This will let you target a

Re: Slow start of HBase operations with YCSB, possibly because of zookeeper ?

2013-01-15 Thread Akshay Singh
Thanks Samar. You are right YCSB writes data to a single table 'usertable', but I see very slow operations (in order of 1-2 operations/second) even for read/update workload and not only for inserts. So, the region is already split in to multiple RS before I start my transaction workload. And

Re: Hbase hbck -repair slow

2013-01-15 Thread Lashing
Kevin Thanks for the update. Actually you can think of the whole meta is missing , or say deleted. Can multiple hbck repair table run at the same time ? Kevin O'dell kevin.od...@cloudera.com 於 2013/1/15 下午11:20 寫道: Lashing, Without knowning the exact errors you are seeing it would be

Re: Hbase hbck -repair slow

2013-01-15 Thread Kevin O'dell
If you have only lost your META, I would just use: hbase hbck -fixMeta -fixAssignments This will rebuild your META On Tue, Jan 15, 2013 at 10:53 AM, Lashing lss...@gmail.com wrote: Kevin Thanks for the update. Actually you can think of the whole meta is missing , or say deleted. Can

How to upgrade HBase from 0.90.5 to 0.94

2013-01-15 Thread 幻风
Hi, everybody, I want to upgrade a HBase cluster from 0.90.5 to 0.94. The cluster has about 50 nodes and the hadoop version is 0.20.2. Is there an easy way to do it? Shutdown is acceptable and what I really care is the safety of the data. Hope for your suggestions. Thanks, Mickey

Re: Hbase tables disappear

2013-01-15 Thread Stack
On Tue, Jan 15, 2013 at 1:07 AM, Ibrahim Yakti iya...@souq.com wrote: The root directory is set to /var/lib/hbase/ The count issue is still exist: *Sqoop: * 13/01/15 08:55:23 INFO mapreduce.ImportJobBase: Retrieved 1754285 records. *MySQL:* +--+ | count(1) | +--+ |

Re: Hbase tables disappear

2013-01-15 Thread Ibrahim Yakti
The upload finished and sqoop confirmed it was done successfully. I use sqoop to create the table and schema. I've tried the same command and same data on Amazon EMR instance and it worked without any problem. If it's in schema why the count changes when you run the count command a few minutes

Re: Hbase hbck -repair slow

2013-01-15 Thread Lashing
Thanks Kevin , it looks like this option is only available in 0.90.7 Kevin O'dell kevin.od...@cloudera.com 於 2013/1/15 下午11:56 寫道: If you have only lost your META, I would just use: hbase hbck -fixMeta -fixAssignments This will rebuild your META On Tue, Jan 15, 2013 at 10:53 AM,

Re: Maximizing throughput

2013-01-15 Thread Bryan Keller
I'll follow up on this in case it is useful to anyone. It seems I was network I/O limited. The switch I was using was in managed mode which decreased throughput to 1gbit/sec within the switch, not just on the wire. So with replication set to 2, throughput was about half of the theoretical max

Re: Maximizing throughput

2013-01-15 Thread Andrew Purtell
Thanks Bryan, really appreciate you letting us know the outcome. I'm sure it will be useful to others. On Tue, Jan 15, 2013 at 9:28 AM, Bryan Keller brya...@gmail.com wrote: I'll follow up on this in case it is useful to anyone. It seems I was network I/O limited. The switch I was using was in

Re: Coprocessor / threading model

2013-01-15 Thread Wei Tan
Andrew, could you explain more, why doing cross-table operation is an anti-pattern of using CP? Durability might be an issue, as far as I understand. Thanks, Best Regards, Wei From: Andrew Purtell apurt...@apache.org To: user@hbase.apache.org user@hbase.apache.org, Date: 01/12/2013

Re: Coprocessor / threading model

2013-01-15 Thread Varun Sharma
You should look at the jstack - I think HTablePool is the reason for the large number of threads. Note that HTablePool is a reusable pool HTable(s) and each HTable consists of an ExecutorService containing 1 thread by default. Are you closing the HTable you obtain from HTablePool - if you are not

Re: Coprocessor / threading model

2013-01-15 Thread Andrew Purtell
HTable is a blocking interface. When a client issues a put, for example, we do not want to return until we can confirm the store has been durably persisted. For client convenience many additional details of remote region invocation are hidden, for example META table lookups for relocated regions,

User mailing list

2013-01-15 Thread Renato Marroquín Mogrovejo
Hi all, I am an HBase user but not a developer. The thing is that I keep on getting all the JIRA emails from HBase + the user's emails, isn't that supposed to be only for the dev list? Could anyone with the necessary superpowers take the JIRA emails out of the users list please? Thanks in

Re: User mailing list

2013-01-15 Thread Jean-Daniel Cryans
I checked a few emails and as far as I can see it's going to issues@ not user@ J-D On Tue, Jan 15, 2013 at 11:48 AM, Renato Marroquín Mogrovejo renatoj.marroq...@gmail.com wrote: Hi all, I am an HBase user but not a developer. The thing is that I keep on getting all the JIRA emails from

Re: User mailing list

2013-01-15 Thread Jean-Marc Spaggiari
Renato, can you check the To: from the JIRA your are receiving? Which address is it? What are those JIRAs? Are you watching them? JM 2013/1/15, Jean-Daniel Cryans jdcry...@apache.org: I checked a few emails and as far as I can see it's going to issues@ not user@ J-D On Tue, Jan 15, 2013 at

Re: User mailing list

2013-01-15 Thread Renato Marroquín Mogrovejo
Thanks for the quick replay. I am all confused now. It says: iss...@hbase.apache.org I don't remember subscribing to this list, anyways how do I unsubscribe? the same way as the dev one? Thanks again! Renato M. 2013/1/15 Jean-Marc Spaggiari jean-m...@spaggiari.org: Renato, can you check the

Re: Constructing rowkeys and HBASE-7221

2013-01-15 Thread Doug Meil
Hi there, well, this request for input fell like a thud. :-) But I think perhaps it has to do with the fact that I sent it to the dev-list instead of the user-list, as people that are actively writing HBase itself (devs) need less help with such keybuilding utilities. So one last request for

Re: User mailing list

2013-01-15 Thread Jean-Marc Spaggiari
Yep, the same way ;) https://www.google.com/search?q=unsubscribe+hbase+issues 2013/1/15, Renato Marroquín Mogrovejo renatoj.marroq...@gmail.com: Thanks for the quick replay. I am all confused now. It says: iss...@hbase.apache.org I don't remember subscribing to this list, anyways how do I

Re: Coprocessor / threading model

2013-01-15 Thread Wei Tan
Thanks Andrew for your detailed clarification. Now I understand that in general, the system is subject to CAP theorem. You want good consistency AND latency, then partition tolerance needs to be sacrificed: this is the local index approach, i.e., colocate index and data and avoid RPC.

Bloom filter based scanner/filter

2013-01-15 Thread David G. Boney
I am building a data cube on top of HBase. All access to the data is by map/reduce jobs. I want to build a scanner where its first matching criteria is based on the set intersection of bloom filters, followed by additional matching criteria specified in the current filter architecture. First, I

RE: Coprocessor / threading model

2013-01-15 Thread Anoop Sam John
Thanks Andrew. A detailed and useful reply Nothing more needed to explain the anti pattern.. :) -Anoop- From: Andrew Purtell [apurt...@apache.org] Sent: Wednesday, January 16, 2013 12:50 AM To: user@hbase.apache.org Subject: Re: Coprocessor /