Hi Neelesh, The saveToPhoenix method uses the MapReduce PhoenixOutputFormat under the hood, which is a wrapper over the JDBC driver. It's likely not as efficient as the CSVBulkLoader, although there are performance improvements over a simple JDBC client as the writes are spread across multiple Spark workers (depending on the number of partitions in the RDD/DataFrame).
Regards, Josh On Sun, Apr 10, 2016 at 1:21 AM, Neelesh <neele...@gmail.com> wrote: > Hi , > Does phoenix-spark's saveToPhoenix use the JDBC driver internally, or > does it do something similar to CSVBulkLoader using HFiles? > > Thanks! > >