The CSV bulk loader in Phoenix actually does this -- it creates HFiles via MapReduce based on CSV input.
You can take a look at the details of how it works in CsvBulkLoadTool.java [1] and CsvToKeyValueMapper.java [2]. There isn't currently a public API for creating Phoenix-compatible HFiles via MapReduce in Phoenix, but there is a set of utility classes in the org.apache.phoenix.mapreduce package for writing to Phoenix directly as the output of a MapReduce program. - Gabriel 1. https://github.com/apache/phoenix/blob/master/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/CsvBulkLoadTool.java 2. https://github.com/apache/phoenix/blob/master/phoenix-core/src/main/java/org/apache/phoenix/mapreduce/CsvToKeyValueMapper.java On Sun, Jan 11, 2015 at 6:34 PM, Pariksheet Barapatre <pbarapa...@gmail.com> wrote: > Hello All, > > New year greetings..!!! > > My question as follow - > > How to create Phoenix Salted table equivalent HFile using MapReduce. > > As per my understanding we can create HFile by specifying > > HFileOutputFormat.configureIncrementalLoad(job, hTable); > > What would be the way to create salt and generate phoenix equivalent rowkey > and values. > > > Cheers, > Pari