See http://search-hadoop.com/m/9UY0h2ZCvra1NlgtC1&subj=Re+Extracting+column+values+from+Phoenix+composite+primary+key
On Fri, Sep 16, 2016 at 10:46 AM, Jonathan Leech <jonat...@gmail.com> wrote: > This would be really useful. The use case I have that is similar is to map > Phoenix data to Hive (but the subset of Hive that Impala understands). I > imagine it could work by reading the System.catalog table, or connection > metadata, and generating Hive create table statements. There would need to > be UDFs to split apart row keys and transform data, e.g. flipping the 1st > byte of numeric types. You could use the same logic in the UDFs to read the > data from a standalone hbase client. > > On Sep 16, 2016, at 11:15 AM, Krishna <research...@gmail.com> wrote: > > Hi, > > Does Phoenix have API for converting a rowkey (made up of multiple > columns) and in ImmutableBytesRow format to split into primary key columns? > I am performing a scan directly from HBase and would like to convert the > rowkey into column values. We used Phoenix standard JDBC API while writing > to the table. > > Thanks > >