Re: Does phoenix CsvBulkLoadTool write to WAL/Memstore

2016-03-19 Thread Pariksheet Barapatre
Hi Vamsi, How many number of rows your expecting out of your transformation and what is the frequency of job? If there are less number of row (< ~100K and this depends on cluster size as well), you can go ahead with phoenix-spark plug-in , increase batch size to accommodate more rows, else use

RE: Telco HBase POC

2016-01-18 Thread Pariksheet Barapatre
auto compaction > > 4. Select primary key correctly > 5. Don't use salting . As table will be huge, your phoenix query will fork > may scanners. Try something like hash on userid. > 6. Define TTL to purge data periodically > > > > > > Regards, > &g

Re: Telco HBase POC

2016-01-15 Thread Pariksheet Barapatre
Hi Willem, Looking at your use case. Phoenix would be a handy client. Few notes from my experience : 1. Use bulk load rather than psql.py. Load larger files(merge) instead of small files. 2. Increase HBase block cache 3. Turn off HBase auto compaction 4. Select primary key correctly 5. Don't use

Re: Phoenix Pig integration

2015-08-14 Thread Pariksheet Barapatre
: At my workplace, we have had many challenges in integration of HBase and Avro with Oozie4.2. Thanks, Anil Gupta On Fri, Aug 14, 2015 at 7:24 AM, Pariksheet Barapatre pbarapa...@gmail.com wrote: Hi Ravi/All, When I use PhoenixPigStorage to load HBase and try to run job through Oozie, I

Re: Phoenix Pig integration

2015-08-10 Thread Pariksheet Barapatre
2015 at 13:12, Pariksheet Barapatre pbarapa...@gmail.com wrote: Hi Russell, below error I am getting in pig.log Failed to parse: null at org.apache.pig.parser.QueryParserDriver.parse(QueryParserDriver.java:198) at org.apache.pig.PigServer$Graph.validateQuery(PigServer.java:1648

Re: Phoenix Pig integration

2015-08-10 Thread Pariksheet Barapatre
, Pariksheet Barapatre pbarapa...@gmail.com wrote: Hi All, I am trying to run Pig script on Phoenix table. I am using same example given in documentation. CREATE TABLE HIRES( CLIENTID INTEGER NOT NULL, EMPID INTEGER NOT NULL, NAME VARCHAR CONSTRAINT pk PRIMARY KEY(CLIENTID,EMPID)); raw = LOAD

Data Model Suggestion

2015-06-23 Thread Pariksheet Barapatre
Hello All, This is more like a HBase question but as I am planning to use Phoenix as a access layer, I hope phoenix user will help me. I would like to create time series data to get on-the-fly analytics. This use case is for adTech. Report - what is houly,daily,weekly impression counts at

Re: Data Model Suggestion

2015-06-23 Thread Pariksheet Barapatre
...@gmail.com wrote: Hi Pari For your use-case, having it as part of the rowkey should be a better design than creating so many columns. Regards Skanda On Tue, Jun 23, 2015 at 6:49 PM, Pariksheet Barapatre pbarapa...@gmail.com wrote: Hello All, This is more like a HBase question but as I am

Pheonix equivalent HFile

2015-01-11 Thread Pariksheet Barapatre
Hello All, New year greetings..!!! My question as follow - How to create Phoenix Salted table equivalent HFile using MapReduce. As per my understanding we can create HFile by specifying HFileOutputFormat.configureIncrementalLoad(job, hTable); What would be the way to create salt and

Re: Pheonix equivalent HFile

2015-01-11 Thread Pariksheet Barapatre
* variety. If not, we were using the API for the Phoenix Data types. I don't know about salting specifically. Also, i believe that Phoenix now has an OutputFormat in the latest version that might help. Abe On Sun, Jan 11, 2015 at 12:34 PM, Pariksheet Barapatre pbarapa...@gmail.com wrote

HASH JOIN - ArrayIndexOutOfBoundException

2014-11-12 Thread Pariksheet Barapatre
Hello All, I am trying to do JOIN of two tables . Table 1 - 20 columns - 2000 rows Table 2 - 2 columns - 2 rows Joins fails for these two tables with ArrayIndexOutOfBoundException. If I add LIMIT 5 clause , it works. Any idea , why it is happening. I am running this on CentOS 6 VM - 8GB

Solr+Phoenix+Hbase

2014-08-21 Thread Pariksheet Barapatre
Hi All, I am trying load data into Solr using DIH from jdbc:phoenix. I have copied phoenix client jar in solr.war file but when I run DIH update, it just hang.. Anybody with earlier experience with similar issue?? Cheers, Pari

Re: jdbc:phoenix connection

2014-08-21 Thread Pariksheet Barapatre
: Pariksheet, try editing /etc/hosts in both host os and vm (add zookeeper quorum host/ip) On Thu, Aug 21, 2014 at 7:37 AM, Pariksheet Barapatre pbarapa...@gmail.com wrote: Hi All, I have just started exploring both Phoenix and Hbase. I have configured Hbase0.94.21 and Phoenix 2.2

Re: jdbc:phoenix connection

2014-08-21 Thread Pariksheet Barapatre
(PhoenixEmbeddedDriver.java:105) at java.sql.DriverManager.getConnection(DriverManager.java:571) at java.sql.DriverManager.getConnection(DriverManager.java:233) at DataLoader.main(DataLoader.java:10) Can anybody please help.. Many Thanks Pari On 21 August 2014 19:03, Pariksheet Barapatre