*I want my desire output likewise*
ROW CELL+COLUMN
QXM column=stocks_output:average,
timestamp=XX, val
ue=
QTM column=stocks_output:average,
timestamp=XX, val
James,
I have only one family -cp. Yes, that is how I store the Double. No, the
doubles are always positive.
The keys are A14568 Less than a million and I added the alphabets to
randomize them.
I group them based on the C_ suffix and say order them by the Double (to
simplify it).
Is there a
Hi,
I am runing Hive and Hbase on the same Amazon EC2 cluster, where Hbase
is in a pseudo-distributed mode.
After integrating HBase in Hive, I find that it takes a long time when
runing a insert overwrite query from hive in order to load data into a
related HBase table.
In fact, the size
Update:
I messed up some queries, here are the right ones:
CREATE TABLE hbase_table (
material_id int,
new_id_client int,
last_purchase_date int)
STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
WITH SERDEPROPERTIES (hbase.columns.mapping =
:key,cf1:idclt,cf1:dt_last_purchase)
Update:
There are 1 master and 3 slaves in my cluster.
They are all m1.medium instances.
*Instance Family* *Instance Type* *Processor Arch* *vCPU* *ECU*
*Memory (GiB)* *Instance Storage (GB)* *EBS-optimized Available*
*Network Performance*
Hi Manish,
Read your stack trace:
at com.maddy.openaveragestock$map.map(openaveragestock.java:59)
Whatever you are doing on line 59 in openaveragestock.java is causing a
NullPointerException. Make sure you check for nulls or better yet improve
the logic.
Thanks,
Stas
On 19 August 2013 09:13,
Hi guys,
I have the following questions in HBASE 0.90.6
1. Does hbase use only one compaction thread to handle both major and minor
compaction?
2. If hbase uses multiple compaction threads, which configuration parameter
defines the number of compaction threads?
3. After
Hello, all,
I use bucket cache as the main block cache in memory.
The default value of hbase.bucketcache.combinedcache.percentage is 0.9f.
I find that it takes no effect to set it to other values. And more, I can't
find this config string in hbase 0.95.1 source code either.
This config string was
Found it in org.apache.hadoop.hbase.io.hfile.CacheConfig.
The name of this config parameter has been changed to
hbase.bucketcache.percentage.in.combinedcache.
Best Wishes
On Mon, Aug 19, 2013 at 5:58 PM, Xiong LIU liuxiongh...@gmail.com wrote:
Hello, all,
I use
hello,
jean
Did u find it??
On Sun, Aug 18, 2013 at 8:28 AM, manish dunani manishd...@gmail.com wrote:
But i want my output likewise::
ROWCELL+COLUMN
QXM column=stocks_output:average,
timestamp=XX, val
ue=
QTM
Hi all,
I'm getting the following error messages everytime i run the map-reduce job
across multiple hadoop clusters:
java.lang.NullPointerException
at org.apache.hadoop.hbase.util.Bytes.toBytes(Bytes.java:414)
at org.apache.hadoop.hbase.client.HTable.init(HTable.java:170)
at
Also, the same code works perfectly fine when i run it in single node
cluster. I've added the hbase classpath to HADOOP_CLASSPATH and have set
all the other env variables also..
On Mon, Aug 19, 2013 at 6:33 PM, Pavan Sudheendra pavan0...@gmail.comwrote:
Hi all,
I'm getting the following error
Can you please explain or show the flow of the code a bit more? Why are you
create the HTable object again and again in the mapper? Where is
ContentidxTable
(the name of the table, I believe?) defined? What is your actually
requirement?
Also, have you looked into this, the api for wiring HBase
I'm basically trying to do a join across 3 tables in the mapper.. In the
reducer i am doing a group by and writing the output to another table..
Although, i agree that my code is pathetic, what i could actually do is
create a HTable object once and pass it as an extra argument to the map
Hi Manish,
I did not even looked at it ;)
Wha investigation have you done on your side to try to figure what the
issue is?
Have you add some logging to see the values before they are used to figure
what they exactly are?
Have you looked at some code to see what it's doing?
What is the line 59
Hi Monish,
Any reason for you to use 0.90.6? It's a pretty old version. Have you
considered moving to a more recent one?
JM
2013/8/19 Monish r monishs...@gmail.com
Hi guys,
I have the following questions in HBASE 0.90.6
1. Does hbase use only one compaction thread to handle both major and
I could not find it out!!
Can You just look at it !!
*Pastebin:*http://pastebin.com/pFKUhTgJ#
Tell me where i did mistake?
On Mon, Aug 19, 2013 at 7:16 PM, Jean-Marc Spaggiari
jean-m...@spaggiari.org wrote:
Hi Manish,
I did not even looked at it ;)
Wha investigation have you done on
I think you should not try to join the tables this way. It will be against
the recommended design/pattern of HBase (joins in HBase alone go against
the design) and M/R. You should first, maybe through another M/R job or PIg
script, for example, pre-process data and massage it into a uniform or
The line 59 in your example is commented, so I guess it's the line a bit
after.
Really, you should try to investigate a bit on your own. Else you will
always face issues that will take you days to get a reply on. And at some
points, people will stop looking at your questions if they figure you
But there's a lot of processing happening with the table data before sent
over to the reducer.. Theoretically speaking, it should be possible..
Our supervisor strictly wants a mr application to do this..
Do you want to see more code? I'm just baffled as to why it's giving null
pointer when there
Hi Jean,
We are thinking of moving to 0.92 or down the line , but for now we are
using this version.
If anyone can share what the behavior of hbase in 0.90.x for the above
questions will be great.
One more question to add,
1.Major compaction rewrites a region hfile , even if it has only one
Kiru,
Is the column qualifier for the key value storing the double different
for different rows? Not sure I understand what you're grouping over.
Maybe 5 rows worth of sample input and expected output would help.
Thanks,
James
On Aug 19, 2013, at 1:37 AM, Kiru Pakkirisamy
Multiple random seeks?
Sorry, you've lost me.
In simple design, you use an inverted table where the indexed value is the row
key and the columns contain the base table's row key.
One get() and you have all of the rows in the base table that match the key.
The only gotcha… is if your row
I believe hbase keeps info in zk that gives you the count of pending operations
to be replicated. Check into the rz zookeeper node in the hbase replication
documentation.
http://hbase.apache.org/replication.html
On 19/08/2013, at 16:42, Alex Newman posi...@gmail.com wrote:
I have setup
This is related: http://search-hadoop.com/m/SrEIT1jtzPF
Cheers
On Mon, Aug 19, 2013 at 12:50 PM, gordoslocos gordoslo...@gmail.com wrote:
I believe hbase keeps info in zk that gives you the count of pending
operations to be replicated. Check into the rz zookeeper node in the hbase
Just simple canary probe approach
Update cluster1:t1:canary_row with a current time every (say) 1 sec
Read time from cluster2:t1:canary_row every second
Compute the difference
Best regards,
Vladimir Rodionov
Principal Platform Engineer
Carrier IQ, www.carrieriq.com
e-mail:
Vladimir,
I have heard about your approach, kind of a column/timestamper marker, is
implemented by some companies. it is certainly a valid approach, and I am
also looking into this direction.
Just like to put a couple comments on the approach, which I am seeking to
improve:
1) need to manually
Hi Hao,
how do you run HBase in pseudo distributed mode, yet with 3 slaves?
Where is the data written in EC2? EBS or local storage?
Did you do any other tuning at the HBase or HDFS level (server side)?
If your replication level is still set to 3 you're seeing somewhat of a worst
case scenario,
hi all,
I use two threads to write data to three tables (2000 tps), and set
hbase.htable.threads.max as 6 ,
however it generated more than 1000 threads ,and my hbase version:0.94.7 ,
did somebody encounter the same problem? and how to solve it
Could you post the code you used?
If it is too large to fit on a single page, but it it in pastebin.
Thanks.
-- Lars
From: fx_bull javac...@gmail.com
To: user@hbase.apache.org user@hbase.apache.org
Sent: Monday, August 19, 2013 6:46 PM
Subject: problem
30 matches
Mail list logo