i noticed in DataFrame that to get the rdd out of it some conversions are
done:
val converter = CatalystTypeConverters.createToScalaConverter(schema)
rows.map(converter(_).asInstanceOf[Row])
does this mean DataFrame internally does not use the standard scala types?
why not?
In many cases we use more efficient mutable implementations internally
(i.e. mutable undecoded utf8 instead of java.lang.String, or a BigDecimal
implementation that uses a Long when the number is small enough).
On Thu, Jun 25, 2015 at 1:56 PM, Koert Kuipers ko...@tresata.com wrote:
i noticed in