Hi Vamsi,
How many number of rows your expecting out of your transformation and what
is the frequency of job?
If there are less number of row (< ~100K and this depends on cluster size
as well), you can go ahead with phoenix-spark plug-in , increase batch
size to accommodate more rows, else use
auto compaction
>
> 4. Select primary key correctly
> 5. Don't use salting . As table will be huge, your phoenix query will fork
> may scanners. Try something like hash on userid.
> 6. Define TTL to purge data periodically
>
>
>
>
>
> Regards,
>
&g
Hi Willem,
Looking at your use case. Phoenix would be a handy client.
Few notes from my experience :
1. Use bulk load rather than psql.py. Load larger files(merge) instead of
small files.
2. Increase HBase block cache
3. Turn off HBase auto compaction
4. Select primary key correctly
5. Don't use
: At my workplace, we have had many challenges in integration of HBase
and Avro with Oozie4.2.
Thanks,
Anil Gupta
On Fri, Aug 14, 2015 at 7:24 AM, Pariksheet Barapatre
pbarapa...@gmail.com wrote:
Hi Ravi/All,
When I use PhoenixPigStorage to load HBase and try to run job through
Oozie, I
2015 at 13:12, Pariksheet Barapatre pbarapa...@gmail.com
wrote:
Hi Russell,
below error I am getting in pig.log
Failed to parse: null
at
org.apache.pig.parser.QueryParserDriver.parse(QueryParserDriver.java:198)
at org.apache.pig.PigServer$Graph.validateQuery(PigServer.java:1648
, Pariksheet Barapatre pbarapa...@gmail.com
wrote:
Hi All,
I am trying to run Pig script on Phoenix table.
I am using same example given in documentation.
CREATE TABLE HIRES( CLIENTID INTEGER NOT NULL, EMPID INTEGER NOT NULL, NAME
VARCHAR CONSTRAINT pk PRIMARY KEY(CLIENTID,EMPID));
raw = LOAD
Hello All,
This is more like a HBase question but as I am planning to use Phoenix as a
access layer, I hope phoenix user will help me.
I would like to create time series data to get on-the-fly analytics.
This use case is for adTech.
Report - what is houly,daily,weekly impression counts at
...@gmail.com wrote:
Hi Pari
For your use-case, having it as part of the rowkey should be a better
design than creating so many columns.
Regards
Skanda
On Tue, Jun 23, 2015 at 6:49 PM, Pariksheet Barapatre
pbarapa...@gmail.com wrote:
Hello All,
This is more like a HBase question but as I am
Hello All,
New year greetings..!!!
My question as follow -
How to create Phoenix Salted table equivalent HFile using MapReduce.
As per my understanding we can create HFile by specifying
HFileOutputFormat.configureIncrementalLoad(job, hTable);
What would be the way to create salt and
* variety. If not, we were
using the API for the Phoenix Data types. I don't know about salting
specifically.
Also, i believe that Phoenix now has an OutputFormat in the latest version
that might help.
Abe
On Sun, Jan 11, 2015 at 12:34 PM, Pariksheet Barapatre
pbarapa...@gmail.com wrote
Hello All,
I am trying to do JOIN of two tables .
Table 1 - 20 columns - 2000 rows
Table 2 - 2 columns - 2 rows
Joins fails for these two tables with ArrayIndexOutOfBoundException.
If I add LIMIT 5 clause , it works. Any idea , why it is happening.
I am running this on CentOS 6 VM - 8GB
Hi All,
I am trying load data into Solr using DIH from jdbc:phoenix.
I have copied phoenix client jar in solr.war file but when I run DIH
update, it just hang..
Anybody with earlier experience with similar issue??
Cheers,
Pari
:
Pariksheet, try editing /etc/hosts in both host os and vm (add zookeeper
quorum host/ip)
On Thu, Aug 21, 2014 at 7:37 AM, Pariksheet Barapatre
pbarapa...@gmail.com wrote:
Hi All,
I have just started exploring both Phoenix and Hbase.
I have configured Hbase0.94.21 and Phoenix 2.2
(PhoenixEmbeddedDriver.java:105)
at java.sql.DriverManager.getConnection(DriverManager.java:571)
at java.sql.DriverManager.getConnection(DriverManager.java:233)
at DataLoader.main(DataLoader.java:10)
Can anybody please help..
Many Thanks
Pari
On 21 August 2014 19:03, Pariksheet Barapatre
14 matches
Mail list logo