Hi Willem,
Just to give you my short experience as phoenix user.
I'm using Phoenix4.4 on top of a HBase cluster where I keep 3 billion
entries.
In our use case Phoenix is doing very well and it saved a lot of code
complexity and time. If you guys have already decided that HBase is the way
to go
Hi Willem,
Looking at your use case. Phoenix would be a handy client.
Few notes from my experience :
1. Use bulk load rather than psql.py. Load larger files(merge) instead of
small files.
2. Increase HBase block cache
3. Turn off HBase auto compaction
4. Select primary key correctly
5. Don't use
On the branch: 4.5-HBase-1.0-cdh5, I set cdh version to 5.5.1 in pom and
building the package produces following errors.
Repo: https://github.com/chiastic-security/phoenix-for-cloudera
[ERROR]
Hi James,
Thanks for your reply. My problem was resolved by setting
phoenix.coprocessor.maxServerCacheTimeToLiveMs to 3 minutes and
phoenix.upsert.batch.size to 10. I think, i can increase
phoenix.upsert.batch.size to a higher value but haven't got opportunity to
try that out yet.
Thanks,
Anil
Hi,
I am currently consulting at a client with the following requirements.
They want to make available detailed data usage CDRs for customers to verify
their data usage against the websites that they visited. In short this can be
seen as an itemised bill for data usage. The data is