The CSV bulk loader in Phoenix actually does this -- it creates HFiles
via MapReduce based on CSV input.

You can take a look at the details of how it works in
CsvBulkLoadTool.java [1] and CsvToKeyValueMapper.java [2]. There isn't
currently a public API for creating Phoenix-compatible HFiles via
MapReduce in Phoenix, but there is a set of utility classes in the
org.apache.phoenix.mapreduce package for writing to Phoenix directly
as the output of a MapReduce program.

- Gabriel


1. 
https://github.com/apache/phoenix/blob/master/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/CsvBulkLoadTool.java
2. 
https://github.com/apache/phoenix/blob/master/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/CsvToKeyValueMapper.java

On Sun, Jan 11, 2015 at 6:34 PM, Pariksheet Barapatre
<pbarapa...@gmail.com> wrote:
> Hello All,
>
> New year greetings..!!!
>
> My question as follow -
>
> How to create Phoenix Salted table  equivalent HFile using MapReduce.
>
> As per my understanding we can create HFile by specifying
>
> HFileOutputFormat.configureIncrementalLoad(job, hTable);
>
> What would be the way to create  salt and generate phoenix equivalent rowkey
> and values.
>
>
> Cheers,
> Pari

Reply via email to