Re: Put while move or split

2012-12-03 Thread Nicolas Liochon
If a server crashes, its regions will be moved to other servers. So, if a query fails because the region server does not respond, the hbase client will retry as well. The principle is to make the failure transparent to the application using the hbase client. Le 3 déc. 2012 03:53, Jean-Marc

Re: Put while move or split

2012-12-03 Thread Michel Segel
He could have meant that the region was unavailable for a longer duration. My take is that you could have limited bandwidth, and a very large region on the move. While I don't imagine that it could take that long to move a region, you never know how things are tuned and you could in theory hit

Re: How to install Snappy?

2012-12-03 Thread Jean-Marc Spaggiari
Thanks all for your replies. So, to reply to all in one. I'm not using CD3. I'm using Hadoop 1.0.3 and HBase 0.94.2 directly from the JARs. Here are all the places where I have put the lib: /home/hadoop/hadoop-1.0.3/lib/native/Linux-amd64-64/libsnappy.so

Re: How to install Snappy?

2012-12-03 Thread Jean-Marc Spaggiari
Ok This: http://code.google.com/p/hadoop-snappy/issues/detail?id=2 helped me and my test program is now working. I'm able to load both libraries. Fine. But the CompressionTest is still not working. What is very strange is that: 12/12/03 08:44:24 WARN snappy.LoadSnappy: Snappy native library

Re: Put while move or split

2012-12-03 Thread Harsh J
Hey Mike, A move is logical in terms of a region and region servers (i.e. only an ID and a command is passed around), but yes the location of blocks could impact its performance a bit (until a further compaction), but would not prevent availability. On Mon, Dec 3, 2012 at 3:53 PM, Michel Segel

Re: How to install Snappy?

2012-12-03 Thread Kevin O'dell
There is a compression test JIRA right now. What are you seeing? On Mon, Dec 3, 2012 at 8:47 AM, Jean-Marc Spaggiari jean-m...@spaggiari.org wrote: Ok This: http://code.google.com/p/hadoop-snappy/issues/detail?id=2 helped me and my test program is now working. I'm able to load both

Re: How to install Snappy?

2012-12-03 Thread Kevin O'dell
Hey JM, Sorry for the quick message earlier. I tracked down the JIRA I was referring to: https://issues.apache.org/jira/browse/HBASE-7080 Does this look like what you are seeing in Compression test? On Mon, Dec 3, 2012 at 9:09 AM, Kevin O'dell kevin.od...@cloudera.comwrote: There is a

Re: How to install Snappy?

2012-12-03 Thread Jean-Marc Spaggiari
Hi Kevin, Thanks for the clarification. No, it's not what I'm seeing. Here is what I'm getting: 12/12/03 09:40:42 WARN snappy.LoadSnappy: Snappy native library is available 12/12/03 09:40:42 WARN snappy.LoadSnappy: Snappy native library not loaded Exception in thread main

Data Locality, HBase? Or Hadoop?

2012-12-03 Thread Jean-Marc Spaggiari
Hi, I'm wondering who is taking care of the data locality. Is it hadoop? Or hbase? Let's say I have disabled the load balancer and I'm manually moving a region to a specific server. Who is going to take care that the data is going to be on the same datanode as the regionserver I moved the region

Re: Data Locality, HBase? Or Hadoop?

2012-12-03 Thread Kevin O'dell
JM, If you have disabled the balancer and are manually moving regions, you will need to run a compaction on those regions. That is the only(logical) way of bringing the data local. HDFS does not have a concept of HBase locality. HBase locality is all managed through major and minor

Re: How to install Snappy?

2012-12-03 Thread Jean-Marc Spaggiari
Ok, I got it I had to copy the hadoop native libs into the hbase native libs directory! Now I get a SUCCESS when I'm doint the CompressionTest... I'm not 100% sure that it's the only think which was missing because I have done so many modifications in the last 3 days... So I will start

Re: Data Locality, HBase? Or Hadoop?

2012-12-03 Thread Jean-Marc Spaggiari
Ok. So each time I move a region manually, I'm better to run a minor compaction to make sure data is moved locally too. Got it. Thanks, JM 2012/12/3, Kevin O'dell kevin.od...@cloudera.com: JM, If you have disabled the balancer and are manually moving regions, you will need to run a

Re: How to install Snappy?

2012-12-03 Thread Kevin O'dell
Never say die! On Mon, Dec 3, 2012 at 10:15 AM, Jean-Marc Spaggiari jean-m...@spaggiari.org wrote: Ok, I got it I had to copy the hadoop native libs into the hbase native libs directory! Now I get a SUCCESS when I'm doint the CompressionTest... I'm not 100% sure that it's the only

Re: Data Locality, HBase? Or Hadoop?

2012-12-03 Thread Kevin O'dell
I would argue to running a major compact on the individual region. Should be quick seeing as how you are on .94. On Mon, Dec 3, 2012 at 10:16 AM, Jean-Marc Spaggiari jean-m...@spaggiari.org wrote: Ok. So each time I move a region manually, I'm better to run a minor compaction to make sure

Re: Data Locality, HBase? Or Hadoop?

2012-12-03 Thread Doug Meil
Hi there- This is also discussed in the Regions section in the RefGuide: http://hbase.apache.org/book.html#regions.arch 9.7.3. Region-RegionServer Locality On 12/3/12 10:08 AM, Kevin O'dell kevin.od...@cloudera.com wrote: JM, If you have disabled the balancer and are manually moving

Re: How to install Snappy?

2012-12-03 Thread Jean-Marc Spaggiari
Ok. I retried and on brand new 0.94.3 installation and the only 2 things required are the libsnappy.so and the libhadoop.so files... And it'S taking 5 minutes to install ;) I will now deploy that all over the cluster and give snappy+0.94.3RC a try... 2012/12/3, Kevin O'dell

Re: Data Locality, HBase? Or Hadoop?

2012-12-03 Thread Jean-Marc Spaggiari
Ok. I will try the major compaction then ;) Doug, thanks for pointing to the doc! I now totally understand why it's moved locally when the compaction occurs! Thanks all! I will give that a try very shortly. JM 2012/12/3, Doug Meil doug.m...@explorysmedical.com: Hi there- This is also

Re: Put while move or split

2012-12-03 Thread Jean-Marc Spaggiari
Thanks all for your comments. I was talking about the 3 minutes period because by default, ZK will have a 3 minute timeout. So if a server crash, it can take up to 3 minutes to Zookeeper to detect that. And in this meantime, the client might not be able to access the impacted regions. So what I

Re: How to install Snappy?

2012-12-03 Thread Mohamed Ibrahim
I got stuck like you before trying to setup snappy on hbase. Here what I recall I did from memory: I think you need to copy the snappy libs inside the hadoop lib/native, then point hbase to it. So in hbase/conf/hbase.env , you should have something similar to this: export

Re: How to install Snappy?

2012-12-03 Thread a...@hsk.hk
Hi, Something more about my workaround last time: I used the following steps to test my workaround: 1) cd $HBASE_HOME ./bin/hbase org.apache.hadoop.hbase.util.CompressionTest file:///tmp/testfile lzo 2) hbase shell create 't1', {NAME = 'cf1', COMPRESSION = 'LZO'} You could modify above

Re: How to install Snappy?

2012-12-03 Thread a...@hsk.hk
Hi JM, I had experienced similar error when I was installing LZO compression to RegionServers: Below is from my record about installing LZO: Issue: java.lang.UnsatisfiedLinkError: no gplcompression in java.library.path ... 12/11/23 19:03:14 ERROR lzo.LzoCodec: Cannot load native-lzo without

Re: How to install Snappy?

2012-12-03 Thread Stack
Any chance of an update to http://hbase.apache.org/book.html#snappy.compression ? If someone writes it up, I'll stitch it in. Thanks, St.Ack On Mon, Dec 3, 2012 at 6:29 AM, a...@hsk.hk a...@hsk.hk wrote: Hi, Something more about my workaround last time: I used the following steps to test

Re: How to install Snappy?

2012-12-03 Thread Jean-Marc Spaggiari
Sure I will. JM 2012/12/3, Stack st...@duboce.net: Any chance of an update to http://hbase.apache.org/book.html#snappy.compression ? If someone writes it up, I'll stitch it in. Thanks, St.Ack On Mon, Dec 3, 2012 at 6:29 AM, a...@hsk.hk a...@hsk.hk wrote: Hi, Something more about my

Re: How to install Snappy?

2012-12-03 Thread Jean-Marc Spaggiari
Done. HBASE-7264 created. I have added few more details about the steps to follow to install Snappy on HBase 0.94.x. I don't know how to build the documentation locally, so I'm not 100% sure about thhe XML structure. I did cutpast so it should be good... JM 2012/12/3, Jean-Marc Spaggiari

Long row + column keys

2012-12-03 Thread Varun Sharma
Hi, I have a schema where the rows are 8 bytes long and the columns are 12 bytes long (roughly 1000 columns per row). The value is 0 bytes. Is this going to be space inefficient in terms of HFile size (large index + blocks) ? The total key size, as far as i know, would be 8 + 12 + 8 (timestamp) =

Re: How to install Snappy?

2012-12-03 Thread Stack
Thank you JM. Let me fold it in St.Ack On Mon, Dec 3, 2012 at 11:48 AM, Jean-Marc Spaggiari jean-m...@spaggiari.org wrote: Done. HBASE-7264 created. I have added few more details about the steps to follow to install Snappy on HBase 0.94.x. I don't know how to build the documentation

Re: Long row + column keys

2012-12-03 Thread Marcos Ortiz
Regards, Varun. I think that you can see the Bernoit Sigoure (@tsuna)´s talk called Lessons learned from OpenTSDB in the last HBaseCon . [1] He explained in great detail how to design your schema to obtain the best performance from HBase. Other recommended talks are: HBase Internals from

Re: Long row + column keys

2012-12-03 Thread Varun Sharma
Hi Marcos, Thanks for the links. We have gone through these and thought about the schema. My question is about whether using PrefixDeltaEncoding makes sense in our situation... Varun On Mon, Dec 3, 2012 at 12:36 PM, Marcos Ortiz mlor...@uci.cu wrote: Regards, Varun. I think that you can see

Multiple regionservers on a single node

2012-12-03 Thread Ishan Chhabra
Hi, Has anybody tried to run multiple RegionServers on a single physical node? Are there deep technical issues or minor impediments that would hinder this? We are trying to do this because we are facing a lot of GC pauses on the large heap sizes (~70G) that we are using, which leads to a lot of

ANN: HBase 0.94.3 is available for download

2012-12-03 Thread lars hofhansl
The HBase Team is pleased to announce the release of HBase 0.94.3. Download it from your favorite Apache mirror [1]. HBase 0.94.3 is a bug fix release and has 51 issues resolved against it, including some performance improvements. 0.94.3 is the current stable release of HBase. All previous

Re: ANN: HBase 0.94.3 is available for download

2012-12-03 Thread Mohammad Tariq
+1. Thank you for the hard work guys. Regards, Mohammad Tariq On Tue, Dec 4, 2012 at 3:12 AM, lars hofhansl lhofha...@yahoo.com wrote: The HBase Team is pleased to announce the release of HBase 0.94.3. Download it from your favorite Apache mirror [1]. HBase 0.94.3 is a bug fix release

Re: Multiple regionservers on a single node

2012-12-03 Thread Marcos Ortiz
Regards, Ishan. On 12/03/2012 04:39 PM, Ishan Chhabra wrote: Hi, Has anybody tried to run multiple RegionServers on a single physical node? Are there deep technical issues or minor impediments that would hinder this? Can you provide more information about your setup? - Network - Disk schema -

Re: Multiple regionservers on a single node

2012-12-03 Thread Doug Meil
Hi there, Not tried multi-RS on a single node, but have you looked at the off-heap cache? It's a part of 0.92.x. From what I understand that feature was designed with this case in mind (I.e., trying to do a lot of caching, but don't want to introduce GC issues in RS).

How to get region location?

2012-12-03 Thread Jean-Marc Spaggiari
Hi, I'm wondering, what's the best way to know which RegionServer a region is hosted on. I looked at the table.jsp but it's using an HMaster from the session and I did not find how it's initialized. From the HBaseAdmin I can get the regions for one table, but I don't have the RegionServer

RE: Data Locality, HBase? Or Hadoop?

2012-12-03 Thread Anoop Sam John
I think all is clear now.. Just to conclude, the data locality is feature provided by HDFS. When DFS client writes some data, hadoop will try to maintain the data locality. HBase region server writes and reads data via the DFS client which is in the same process as that of the RS. When the

Re: How to get region location?

2012-12-03 Thread ramkrishna vasudevan
Hi Actually you can try using HTable.getRegionLocations(). But you need to pass the table name for it. Regards Ram On Tue, Dec 4, 2012 at 6:34 AM, Jean-Marc Spaggiari jean-m...@spaggiari.org wrote: Hi, I'm wondering, what's the best way to know which RegionServer a region is hosted on.

HTablePool dead locks/freezing

2012-12-03 Thread Varun Sharma
Hi, I am using hbase 0.94.0 and am using the HTablePool - Reusable typre with a pool size of 50. I have a lot of threads using the htable pool concurrently (~ 3500) - The client side timeout is 5 seconds and the threads start okay producing good QPS to the hbase cluster, finally the QPS drops

RE: Long row + column keys

2012-12-03 Thread Anoop Sam John
Hi Varun It looks to be very clear that you need to use some sort of encoding scheme. PrefixDeltaEncoding would be fine may be.. You can see the other algos also like the FastDiff... and see how much space it can save in your case. Also suggest you can use the encoding for

Re: HTablePool dead locks/freezing

2012-12-03 Thread Varun Sharma
Okay - this was contention issue - https://issues.apache.org/jira/browse/HBASE-2939 solves the issue - upping the IPC pool size. Thanks ! On Mon, Dec 3, 2012 at 8:04 PM, Varun Sharma va...@pinterest.com wrote: Hi, I am using hbase 0.94.0 and am using the HTablePool - Reusable typre with a

Re: Long row + column keys

2012-12-03 Thread Varun Sharma
Hi Anoop, I agree - I am not so concerned about the savings on disk - rather I am thinking about the savings inside the block cache. I am not sure how stable PrefixDeltaEncoding is and who else uses it. If not, are there people using FastDiff encoding - it seems like any form of encoding scheme

HBase Integration with Active Directory

2012-12-03 Thread anil gupta
Hi All, We would like to integrate HBase with our corporate Active Directory for authentication and authorization of HBase users. I have gone through the following links in HBase: http://blog.cloudera.com/blog/2012/09/understanding-user-authentication-and-authorization-in-apache-hbase/

Re: How to install Snappy?

2012-12-03 Thread surfer
good ! never surrender ! you also could have done like I posted you yesterday which is equivalent: in hbase-env.sh: export HBASE_LIBRARY_PATH=/pathtoyourhadoop/lib/native/Linux-amd64-64 in /pathtoyourhadoop/lib/native/Linux-amd64-64 I have added: libsnappy.a libsnappy.so libsnappy.so.1