John,
Thank you so much for responding. Appreciate the link to ppt. Something I
could not find. but read about snowflake
I was looking for guidance on the sequence numbers vs UUID approach.
Could I use sequence numbers ? are the gaps in the sequence numbers ever
back filled?
There is not
Ash,
I built one a while back based on twitter’s snowflake algorithm.
Here is a link to a presentation from twitter on it…
https://www.slideshare.net/davegardnerisme/unique-id-generation-in-distributed-systems
Hello,
Distributed web application. Millions of users connecting to the site.
we are receiving about 150,000 events/ sec through Kinesis Stream.
We need to store these events in a phoenix table identified by an ID the
primary for the table.
what is the best way to accomplish this?
Option 1
I
Developed and test UDF successfully, but was surprised that the same SELECT
with UDF still succeeded after the jar was removed. Same behavior after
restart sqlline.py, even HBase. So is there a permanent cache for UDF jars
on disk in Phoenix?
So far not causing any problem yet, but a little bit
After poking around the Phoenix code (thanks Ankit for the starting point).
Seems there is way to set up a default HBase timestamp, which suites my need
to query existing HBase tables:
Here is the official doc for the UDF config:
https://phoenix.apache.org/udf.html
I was able to get the UDF picked up after following the instruction. Since I
was just testing with local instance, here is my simplified config in
hbase-site.xml under "/bin":
I created a jar with:
jar -cf $HBASE_HOME/lib/phoenix-udj.jar UDF.java
but I get "ClassNotFoundException" when I try to use the UDF in simple
select query
Regards,
Cheyenne O. Forbes
Hello,
We are building an Enterprise Datawarehouse on Phoenix(HBase)
Please refer the diagram attached.
The EDW supports an unified architecture that serves both Streaming and
batch use cases.
I am recommending a staging area that is source compliant (i.e. that mimics
source structure)
In the