You can get data type from Phoenix meta, then encode/decode data to
write/read data. I think this way is effective, FYI :)


----------------------------------------
   Yun Zhang
   Best regards!


2018-08-04 21:43 GMT+08:00 Brandon Geise <brandonge...@gmail.com>:

> Good morning,
>
>
>
> I’m looking at using a combination of Hbase, Phoenix and Spark for a
> project and read that using the Spark-Phoenix plugin directly is more
> efficient than JDBC, however it wasn’t entirely clear from examples when
> writing a dataframe if an upsert is performed and how much fine-grained
> options there are for executing the upsert.  Any information someone can
> share would be greatly appreciated!
>
>
>
>
>
> Thanks,
>
> Brandon
>

Reply via email to