Region not assigned

2014-08-14 Thread Kristoffer Sjögren
Hi We are running hbase 0.94.6 cdh 4.4 and have a problem with one table not being assigned to any region. This is the SYSTEM.TABLE in Phoenix so all tables are basically non functional at the moment. When running hbck repair we get the following... ERROR: Region { meta => SYSTEM.TABLE,,13854776

Re: Region not assigned

2014-08-14 Thread Kristoffer Sjögren
org.apache.hadoop.hbase.executor.EventHandler.run(EventHandler.java:175) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:662) On Thu, Aug 14, 2014 a

Re: Region not assigned

2014-08-15 Thread Kristoffer Sjögren
gt; > > > > > > On Thu, Aug 14, 2014 at 8:05 AM, Kristoffer Sjögren > > wrote: > > > >> It seems that the region servers are complaining about wrong phoenix > >> classes for some reason. We are running 2.2.0 which is the version > before > >&

Force remove table

2014-10-14 Thread Kristoffer Sjögren
Hi I accidentally created a few tables with 'lzo' compression without the jar dependencies in hbase 0.94.7. This caused havoc on the master which eventually crashed and now I can't start it again [1]. I tried to remove the tables but I can't - hbase shell is complaining about the master is not ru

Re: Force remove table

2014-10-14 Thread Kristoffer Sjögren
dfs files of those bad tables. > like: bin/hadoop fs -rm /hbase/TABLENAME > > > > > ------ 原始邮件 ------ > 发件人: "Kristoffer Sjögren";; > 发送时间: 2014年10月14日(星期二) 下午3:27 > 收件人: "user"; > > 主题: Force remove table > > > >

Re: Force remove table

2014-10-14 Thread Kristoffer Sjögren
w it goes! > > -Dima > > On Tue, Oct 14, 2014 at 1:11 AM, Kristoffer Sjögren > wrote: > > > I was thinking of doing that but I suspect that zookeeper keeps metadata > of > > tables also. Seems like region servers are fine for now without the > master > > and

Timerange scan

2015-02-28 Thread Kristoffer Sjögren
Hi I want to understand the effectiveness of timerange scans without setting start and stop keys? Will HBase do a full table scan or will the scan be optimized in any way? Cheers, -Kristoffer

Re: Timerange scan

2015-02-28 Thread Kristoffer Sjögren
full table scan, no? > > JM > > 2015-02-28 13:41 GMT-05:00 Kristoffer Sjögren : > > > Hi > > > > I want to understand the effectiveness of timerange scans without setting > > start and stop keys? Will HBase do a full table scan or will the scan be > > optimized in any way? > > > > Cheers, > > -Kristoffer > > >

Re: Timerange scan

2015-02-28 Thread Kristoffer Sjögren
If Scan.setTimeRange is a full table scan then it runs surprisingly fast on tables that host a few hundred million rows :-) On Sat, Feb 28, 2015 at 8:05 PM, Kristoffer Sjögren wrote: > Hi Jean-Marc > > I was thinking of Scan.setTimeRange to only get the x latest rows, but I > w

Re: Timerange scan

2015-03-02 Thread Kristoffer Sjögren
:00 Nick Dimiduk : > > > A Scan without start and end rows will be issued to all regions in the > > table -- a full table scan. Within each region, store files will be > > selected to participate in the scan based on on the min/max timestamps > > from their > > headers.

Re: Timerange scan

2015-03-02 Thread Kristoffer Sjögren
t the end of a key-value's key, but not indexed), so if you > want to model time in your schema, it's best to promote it to an indexed > field -- i.e., make it a component of your row key. > > -n > > On Mon, Mar 2, 2015 at 12:42 AM, Kristoffer Sjögren > wrote: > &g

Re: Timerange scan

2015-03-03 Thread Kristoffer Sjögren
What can I say? Awesome community! :-) On Mon, Mar 2, 2015 at 11:17 PM, Gary Helmling wrote: > Proving it to yourself is sometimes the hardest part! > > On Mon, Mar 2, 2015 at 2:11 PM Nick Dimiduk wrote: > > > Gary to the rescue! Does it still count as being right even if you cannot > > prove i

Stuck closing region / region is flushing

2015-03-14 Thread Kristoffer Sjögren
Hi It seems one of our region servers has been stuck closing a region for almost 22 hours. Puts or gets eventually fail with an exception [1]. Is there any safe way to release the region like restarting the region server? Cheers, -Kristoffer [1] 2015-03-14 21:02:24,316 INFO org.apache.hadoop.

Re: Stuck closing region / region is flushing

2015-03-14 Thread Kristoffer Sjögren
: > Which release of HBase are you using ? > > I wonder if your cluster was hit with HBASE-10499. > > Cheers > > On Sat, Mar 14, 2015 at 1:13 PM, Kristoffer Sjögren > wrote: > > > Hi > > > > It seems one of our region servers has been stuck closing a

Re: Stuck closing region / region is flushing

2015-03-14 Thread Kristoffer Sjögren
estart the server. > > BTW HBASE-10499 didn't go into 0.94 (maybe it should have). Please consider > upgrading. > > Cheers > > On Sat, Mar 14, 2015 at 1:30 PM, Kristoffer Sjögren > wrote: > > > Hi Ted > > > > Sorry I forgot to mention, hbase-0.94.

Re: Stuck closing region / region is flushing

2015-03-14 Thread Kristoffer Sjögren
0-1 > which got stuck, there might be data loss if server is restarted since > there would be some data unable to be flushed. > > Cheers > > On Sat, Mar 14, 2015 at 2:58 PM, Kristoffer Sjögren > wrote: > > > I think I found the thread that is stuck. Is

Re: Stuck closing region / region is flushing

2015-03-15 Thread Kristoffer Sjögren
l edits should be persisted to the > WAL regardless of what Ted said about flushing. > > We are working on the problem, please see HBASE-13238 > > > On Saturday, March 14, 2015, Kristoffer Sjögren > wrote: > > > I think I found the thread that is stuck. Is restar

Rowkey design question

2015-04-07 Thread Kristoffer Sjögren
Hi I have a row with around 100.000 qualifiers with mostly small values around 1-5KB and maybe 5 largers ones around 1-5 MB. A coprocessor do random access of 1-10 qualifiers per row. I would like to understand how HBase loads the data into memory. Will the entire row be loaded or only the qualif

Re: Rowkey design question

2015-04-07 Thread Kristoffer Sjögren
. > > I would suggest rethinking your strategy. > > > On Apr 7, 2015, at 11:13 AM, Kristoffer Sjögren > wrote: > > > > Hi > > > > I have a row with around 100.000 qualifiers with mostly small values > around > > 1-5KB and maybe 5 largers ones aroun

Re: Rowkey design question

2015-04-08 Thread Kristoffer Sjögren
your initial problem statement doesn’t seem to parse … > > > > > > Are you saying that you a single row with approximately 100,000 > elements > > > where each element is roughly 1-5KB in size and in addition there are > ~5 > > > elements which will be betwee

Re: Rowkey design question

2015-04-08 Thread Kristoffer Sjögren
pactions." Is there any chance to run into this problem in the read path for data that is written infrequently and never changed? On Wed, Apr 8, 2015 at 9:30 AM, Kristoffer Sjögren wrote: > A small set of qualifiers will be accessed frequently so keeping them in > block cache would be ve

Re: Rowkey design question

2015-04-08 Thread Kristoffer Sjögren
erms of compaction… if the data is relatively static, you won’t have > compactions because nothing changed. > But if your data is that static… why not put the data in sequence files > and use HBase as the index. Could be faster. > > HTH > > -Mike > > > On Apr 8, 201

Re: Rowkey design question

2015-04-08 Thread Kristoffer Sjögren
to do this is that you’re reading in large amounts of > data and its more efficient to do this from HDFS than through HBase. > > > On Apr 8, 2015, at 8:41 AM, Kristoffer Sjögren wrote: > > > > Yes, I think you're right. Adding one or more dimensions to the rowkey > &

Re: Rowkey design question

2015-04-08 Thread Kristoffer Sjögren
take the > longest, but as long as the row is cached, the results are returned quickly. > > If you’re trying to do a scan with a start/stop row set … your timing then > could vary between sub-second and minutes depending on the query. > > > > On Apr 8, 2015, at 3:10 PM, Kristo

Overwrite a row

2013-04-19 Thread Kristoffer Sjögren
Hi Is it possible to completely overwrite/replace a row in a single _atomic_ action? Already existing columns and qualifiers should be removed if they do not exist in the data inserted into the row. The only way to do this is to first delete the row then insert new data in its place, correct? Or

Re: Overwrite a row

2013-04-19 Thread Kristoffer Sjögren
What would you suggest? I want the operation to be atomic. On Fri, Apr 19, 2013 at 8:32 PM, Ted Yu wrote: > What is the maximum number of versions do you allow for the underlying > table ? > > Thanks > > On Fri, Apr 19, 2013 at 10:53 AM, Kristoffer Sjögren >wrote: &

Re: Overwrite a row

2013-04-20 Thread Kristoffer Sjögren
; > http://hbase.apache.org/apidocs/org/apache/hadoop/hbase/client/HTable.html#mutateRow(org.apache.hadoop.hbase.client.RowMutations) > ). > > >> You can do multiple puts and deletes and they will be performed > > atomically. > > >> So you can remove qualifiers and put ne

Re: Overwrite a row

2013-04-20 Thread Kristoffer Sjögren
. > > Cheers > > On Apr 20, 2013, at 12:17 AM, Kristoffer Sjögren wrote: > > > The schema is known beforehand so this is exactly what I need. Great! > > > > One more question. What guarantees does the batch operation have? Are the > > operations contained w

Re: Overwrite a row

2013-04-22 Thread Kristoffer Sjögren
> For advanced usage, take a look at MultiRowMutationEndpoint: > > > > * This class demonstrates how to implement atomic multi row transactions > > using > > * {@link HRegion#mutateRowsWithLocks(java.util.Collection, > > java.util.Collection)} > > * and Coprocessor endpoi

Wait for master before creating tables

2013-04-22 Thread Kristoffer Sjögren
Hi Im doing an automated install of HBase cluster on EC2. But when the install try to create tables (using hbase shell) the master often have not yet had time start (since this operation is async), so these operations fail. Is there a way to wait until the master is up and know when its safe to

Schema design for filters

2013-06-27 Thread Kristoffer Sjögren
Hi Working with the standard filtering mechanism to scan rows that have columns matching certain criterias. There are columns of numeric (integer and decimal) and string types. These columns are single or multi-valued like "1", "2", "1,2,3", "a", "b" or "a,b,c" - not sure what the separator would

Re: Schema design for filters

2013-06-27 Thread Kristoffer Sjögren
...) > > > On Jun 27, 2013, at 12:59 PM, Kristoffer Sjögren wrote: > > > Hi > > > > Working with the standard filtering mechanism to scan rows that have > > columns matching certain criterias. > > > > There are columns of numeric (integer and decimal)

Re: Schema design for filters

2013-06-27 Thread Kristoffer Sjögren
d be the same, the actual strings could > differ. > > HTH > > -Mike > > On Jun 27, 2013, at 4:41 PM, Kristoffer Sjögren wrote: > > > I realize standard comparators cannot solve this. > > > > However I do know the type of each column so writing custom list

Re: Schema design for filters

2013-06-27 Thread Kristoffer Sjögren
nse. But it should > work ) > > At least it would be a design approach I would talk. YMMV > > Having said that, I expect someone to say its a bad idea and that they > have a better solution. > > HTH > > -Mike > > On Jun 27, 2013, at 5:13 PM, Kristoffer Sjögren wro

Re: Schema design for filters

2013-06-28 Thread Kristoffer Sjögren
t; On Jun 27, 2013, at 4:39 PM, "Kristoffer Sjögren" > wrote: > > > Thanks for your help Mike. Much appreciated. > > > > I dont store rows/columns in JSON format. The schema is exactly that of a > > specific java class, where the rowkey is a unique object ide

Re: Schema design for filters

2013-06-28 Thread Kristoffer Sjögren
slicing and dicing. something to consider... > > Otis > -- > Solr & ElasticSearch Support -- http://sematext.com/ > Performance Monitoring -- http://sematext.com/spm > > > > On Fri, Jun 28, 2013 at 5:24 AM, Kristoffer Sjögren > wrote: > > Interesting. Im actually

Re: Schema design for filters

2013-06-29 Thread Kristoffer Sjögren
ySql? > > > > On Friday, June 28, 2013, Otis Gospodnetic wrote: > > > >> Hi, > >> > >> I see. Btw. isn't HBase for < 1M rows an overkill? > >> Note that Lucene is schemaless and both Solr and Elasticsearch can > >> detect fiel

Filters and versions

2013-07-06 Thread Kristoffer Sjögren
Hi Im writing a custom filter it seems that filterKeyValue(KeyValue kv) give me every version, which is fine, but im only interested in the latest version. I have tried KeyValue.isLatestTimestamp() to filter out older versions but this method always returns false? Also tried setMaxVersions(1) on

Re: Filters and versions

2013-07-06 Thread Kristoffer Sjögren
Ah, the latest version is always returned first. Great, thanks Ted! Cheers, -Kristoffer On Sat, Jul 6, 2013 at 5:25 PM, Ted Yu wrote: > Take a look at SingleColumnValueFilter#filterKeyValue() which respects flag > latestVersionOnly. > > Cheers > > On Sat, Jul 6, 2013 at 8

RowLocks

2013-08-28 Thread Kristoffer Sjögren
Hi About the internals of locking a row in hbase. Does hbase row locks map one-to-one with a locks in zookeeper or are there any optimizations based on the fact that a row only exist on a single machine? Cheers, -Kristoffer

Re: RowLocks

2013-08-28 Thread Kristoffer Sjögren
I want a distributed lock condition for doing certain operations that may or may not be unrelated to hbase. On Thu, Aug 29, 2013 at 12:18 AM, Ted Yu wrote: > RowLock API has been removed in 0.96. > > Can you tell us your use case ? > > > On Wed, Aug 28, 2013 at 3:14 PM,

Re: RowLocks

2013-08-28 Thread Kristoffer Sjögren
rify... > > Do you mean the API is no longer a public API, or do you mean no more RLL > > for atomic writes? > > > > On Aug 28, 2013, at 5:18 PM, Ted Yu wrote: > > > > > RowLock API has been removed in 0.96. > > > > > > Can you tell us your use

Using put for nullifying qualifiers

2012-05-22 Thread Kristoffer Sjögren
Hi I'm trying to use Put operations to replace ("set") already existing rows by nullify certain columns and qualifiers as part of an Put operation. The reason I want to do this is 1) keep the operation atomic/consistent 2) avoid latency from first doing Delete then Put. Is there some way to do t

Re: Using put for nullifying qualifiers

2012-05-22 Thread Kristoffer Sjögren
you can include a delete with a put and keep it atomic. > You could include a null version of the column with your put, though, > for a similar effect. > > --Tom > > On Tue, May 22, 2012 at 10:55 AM, Kristoffer Sjögren > wrote: > > Hi > > > > I'm try

Re: Using put for nullifying qualifiers

2012-05-23 Thread Kristoffer Sjögren
ollection mutations, > Collection rowsToLock) throws IOException { > > It allows you to combine Put's and Delete's for a single region, > atomically. > > On Tue, May 22, 2012 at 1:22 PM, Kristoffer Sjögren >wrote: > > > Thanks, sounds like that should d

Re: Using put for nullifying qualifiers

2012-05-23 Thread Kristoffer Sjögren
Gotcha. Columns are quite dynamic in my case, but since I need to fetch rows first anyways; a KeyOnlyFilter to first find them and then overwrite values will do just fine. Cheers, -Kristoffer

Zookeeper: KeeperErrorCode NoNode for /hbase/backup-masters

2012-08-09 Thread Kristoffer Sjögren
Hi all I have a problem starting hbase in a fully distributed 3 machine setup (2 datanodes/regionservers + 1 master/namenode). For some reason zookeeper on master complains about not finding /hbase/backup-masters in hbase-user-zookeeper-host.out. java.io.IOException: Failed to process transaction

ClosedByInterruptException

2015-09-03 Thread Kristoffer Sjögren
Hi We are running 0.94.6-cdh4.4.0 with phoenix-2.2.3 and recently started seeing connections being forcefully closed, ending with a ClosedByInterruptException [1]. This problem occur on both write and scans. I did some searching and found people that had similar problems. Seems that one answer w

Phantom region server and PENDING_OPEN regions

2015-11-24 Thread Kristoffer Sjögren
Hi I'm trying to install a HBase cluster with 1 master (amb1.service.consul) and 1 region server (amb2.service.consul) using Ambari on docker containers provided by sequenceiq [1] using a custom blueprint [2]. Every component installs correctly except for HBase which get stuck with regions in tra

Re: Phantom region server and PENDING_OPEN regions

2015-11-24 Thread Kristoffer Sjögren
Sorry, I should mention that this is HBase 1.1.2. Zookeeper only report one region server. $ ls /hbase-unsecure/rs [amb2.service.consul,16020,1448353564099] On Tue, Nov 24, 2015 at 9:55 AM, Kristoffer Sjögren wrote: > Hi > > I'm trying to install a HBase cluster with 1

Re: Phantom region server and PENDING_OPEN regions

2015-11-24 Thread Kristoffer Sjögren
om PENDING_OPEN to OPEN state, if hbase:meta table is unavailable master > can not finish initialization. > > Regards > Samir > > On Tue, Nov 24, 2015 at 10:11 AM, Kristoffer Sjögren > wrote: > >> Sorry, I should mention that this is HBase 1.1.2. >> >> Zookeepe

Re: Phantom region server and PENDING_OPEN regions

2015-11-24 Thread Kristoffer Sjögren
d amb2.service.consul,16020,1448357534179 [1] http://pastebin.com/z93p8Mdu On Tue, Nov 24, 2015 at 10:48 AM, Kristoffer Sjögren wrote: > I removed the node.dc1.consul from resolve.conf and restarted the > cluster but it still shows up on the master UI. > > amb2.node.dc1.consul,16020,14

Re: Phantom region server and PENDING_OPEN regions

2015-11-24 Thread Kristoffer Sjögren
roduce wrong > values. Do you have multiple network interfaces on servers? What ping > $HOSTNAME returns? What do you have in /etc/hosts file? Do you have some > local nameserver running on servers ? > > Regards > Samir > On Nov 24, 2015 11:21 AM, "Kristoffer Sjögren"

Re: Phantom region server and PENDING_OPEN regions

2015-11-24 Thread Kristoffer Sjögren
stname to > *.service.consul. > Also try to disable resolution by DNS server, Comment all lines in > /etc/resolve.conf. > > Regards > Samir > > On Tue, Nov 24, 2015 at 12:29 PM, Kristoffer Sjögren > wrote: > >> Only one network interface on all machines. The pin

Re: Hbase + OpenTSDB POC setup ...

2016-08-11 Thread Kristoffer Sjögren
Hi Try to find a docker image. Possibly one that contain both HBase and OpenTSDB. This will get you started in minutes. Cheers, -Kristoffer On Thu, Aug 11, 2016 at 4:02 PM, Kulkarni, Suyog wrote: > Hi, > > I am very new to both HBase and OpenTSDB and I just subscribed to this > use

RegionServer aborting and shutting down

2016-09-22 Thread Kristoffer Sjögren
Hi We are running OpenTSDB 2.2 with HBase 1.1.2 and are having problems with RegionServers that are shutting down sporadically from alleged GC pauses. We run 2 OpenTSDB machines and 30 region servers. 8 GB heaps. The region servers are collocated with data nodes and yarn jobs. Every region server

Guava 15

2013-12-16 Thread Kristoffer Sjögren
Hi At the moment HFileWriterV2.close breaks at startup when using Guava 15. This is not a client problem - it happens because we start a master node to do integration tests. A bit precarious and wonder if there are any plans to support Guava 15, or if there are clever way around this? Cheers, -K

Re: Guava 15

2013-12-16 Thread Kristoffer Sjögren
Thanks! But we cant really upgrade to HBase 0.96 right now, but we need to go to Guava 15 :-( I was thinking of overriding the classes fixed in the patch in our test environment. Could this work maybe? On Mon, Dec 16, 2013 at 11:01 AM, Kristoffer Sjögren wrote: > Hi > > At t

Re: Guava 15

2013-12-16 Thread Kristoffer Sjögren
; > > That means more or less backporting the patch to the 0.94, no? > > It should work imho. > > > > > > > > > > On Mon, Dec 16, 2013 at 3:16 PM, Kristoffer Sjögren > >wrote: > > > > > Thanks! But we cant really upgrade to HBase 0.

Upgrade from HBase 0.94.6-cdh4.4.0 to 0.94.14

2013-12-21 Thread Kristoffer Sjögren
Hi We are running HBase 0.94.6-cdh4.4.0 and wonder what the best way would be to upgrade to 0.94.14 and still have some compliance to cdh? As far as i know, there are no cloudera apt packages for 0.94.14? Cheers, -Kristoffer

Performance tuning

2013-12-21 Thread Kristoffer Sjögren
Hi I have been performance tuning HBase 0.94.6 running Phoenix 2.2.0 the last couple of days and need some help. Background. - 23 machine cluster, 32 cores, 4GB heap per RS. - Table t_24 have 24 online regions (24 salt buckets). - Table t_96 have 96 online regions (96 salt buckets). - 10.5 milli

Re: Performance tuning

2013-12-21 Thread Kristoffer Sjögren
stems administrator to see why pings to these > machines are slow. What are the pings like from a bad RS to another bad RS? > > > On Sat, Dec 21, 2013 at 7:17 PM, Kristoffer Sjögren >wrote: > > > Hi > > > > I have been performance tuning HBase 0.94.6 running Phoe

Re: Performance tuning

2013-12-21 Thread Kristoffer Sjögren
100m (or better 1bn rows). Then we're talking. For anything below > this you wouldn't want to use HBase anyway. > (100k rows I could scan on my phone with a Perl script in less than 1s) > > > With "ping" you mean an actual network ping, or some operation on top

Re: Performance tuning

2013-12-21 Thread Kristoffer Sjögren
Btw, I have tried different number of rows with similar symptom on the bad RS. On Sat, Dec 21, 2013 at 10:28 PM, Kristoffer Sjögren wrote: > @pradeep scanner caching should not be an issue since data transferred to > the client is tiny. > > @lars Yes, the data might be small for thi

Re: Performance tuning

2013-12-21 Thread Kristoffer Sjögren
ight? So > nothing should really be sent across the network. > > When you do the queries, can you check whether there is any network > traffic? > > -- Lars > > > > ____ > From: Kristoffer Sjögren > To: user@hbase.apache.org; lars hofhan

Re: Performance tuning

2013-12-21 Thread Kristoffer Sjögren
Scans on RS 19 and 23, which have 5 regions instead of 4, stands out more than scans on RS 20, 21, 22. But scans on RS 7 and 18, that also have 5 regions are doing fine, not best, but still in the mid-range. On Sat, Dec 21, 2013 at 11:51 PM, Kristoffer Sjögren wrote: > Yeah, im doing a co

Re: Performance tuning

2013-12-21 Thread Kristoffer Sjögren
There are quite a lot of established and time wait connections between the RS on port 50010, but i dont know a good way of monitoring how much data is going through each connection (if that's what you meant)? On Sun, Dec 22, 2013 at 12:00 AM, Kristoffer Sjögren wrote: > Scans on RS 1

Standard vs Asynchbase client reconnect after HBase restart

2014-02-12 Thread Kristoffer Sjögren
Hi I have some tests that check client behaviour during a controlled HBase restart. Everything works as expected and the client is able to recover after a while. However, after doing the same tests with the Asynchbase I noticed that this client recovers almost instantly after HBase comes back up

Re: Standard vs Asynchbase client reconnect after HBase restart

2014-02-12 Thread Kristoffer Sjögren
0.94.x On Wed, Feb 12, 2014 at 4:39 PM, Ted Yu wrote: > Are you using 0.94.x or 0.96.y ? > > Cheers > > > On Wed, Feb 12, 2014 at 12:41 AM, Kristoffer Sjögren >wrote: > > > Hi > > > > I have some tests that check client behaviour during a control

Re: Standard vs Asynchbase client reconnect after HBase restart

2014-02-12 Thread Kristoffer Sjögren
nt has recovered? > > Thx > > -Mike > > > On Feb 12, 2014, at 2:41 AM, Kristoffer Sjögren wrote: > > > Hi > > > > I have some tests that check client behaviour during a controlled HBase > > restart. Everything works as expected and the client is able to recove

Re: Standard vs Asynchbase client reconnect after HBase restart

2014-02-14 Thread Kristoffer Sjögren
oes not > handle well) > > > -- Lars > > > > > From: Kristoffer Sjögren > To: user@hbase.apache.org > Sent: Wednesday, February 12, 2014 11:54 PM > Subject: Re: Standard vs Asynchbase client reconnect after HBase restart > > &

HBase + Spark join

2017-03-13 Thread Kristoffer Sjögren
Hi I want to join a Spark RDD with an HBase table. Im familiar with the different connectors available but couldn't find this functionality. The idea I have is to first sort the RDD according to a byte[] key [1] and rdd.mapPartitions so that I each partition contains a unique and sequentially sor

Parallel range scanners, not getting all rows.

2017-03-16 Thread Kristoffer Sjögren
Hi I'm trying to scan a table using start and stop key ranges based on a single byte. I'm using AsyncHBase where scanners are start key inclusive and stop key exclusive. So for a single byte I generate 256 scanners with key ranges [1]. The "last" key range use start 127 and an empty end key (in

Re: Parallel range scanners, not getting all rows.

2017-03-16 Thread Kristoffer Sjögren
I should mention that I get 126.222.622 of total 126.717.892 so im missing 495.270 rows. 2017-03-16 15:05 GMT+01:00 Kristoffer Sjögren : > Hi > > I'm trying to scan a table using start and stop key ranges based on a > single byte. > > I'm using AsyncHBase where scan

Re: Parallel range scanners, not getting all rows.

2017-03-16 Thread Kristoffer Sjögren
yncHBase, please consider posting on their mailing > list. > > Thanks > > 2017-03-16 7:05 GMT-07:00 Kristoffer Sjögren : > >> Hi >> >> I'm trying to scan a table using start and stop key ranges based on a >> single byte. >> >> I'm

Re: Parallel range scanners, not getting all rows.

2017-03-17 Thread Kristoffer Sjögren
Thanks Ted, that was indeed the problem. 2017-03-17 4:29 GMT+01:00 Ted Yu : > Have you considered the empty start row ? > > 2017-03-16 10:51 GMT-07:00 Kristoffer Sjögren : > >> Thanks Ted, I have posted the question to AsyncHBase. >> >> But, if I may ask. Is my reas

Re: Parallel range scanners, not getting all rows.

2017-03-17 Thread Kristoffer Sjögren
Yes, i already did. Den 17 mars 2017 22:40 skrev "Ted Yu" : > Please also notify AsyncHBase mailing list, if you haven't done so. > > 2017-03-17 9:01 GMT-07:00 Kristoffer Sjögren : > > > Thanks Ted, that was indeed the problem. > > > > 2017-03-17 4:29