In many cases we use more efficient mutable implementations internally
(i.e. mutable undecoded utf8 instead of java.lang.String, or a BigDecimal
implementation that uses a Long when the number is small enough).
On Thu, Jun 25, 2015 at 1:56 PM, Koert Kuipers wrote:
> i noticed in DataFrame that t
i noticed in DataFrame that to get the rdd out of it some conversions are
done:
val converter = CatalystTypeConverters.createToScalaConverter(schema)
rows.map(converter(_).asInstanceOf[Row])
does this mean DataFrame internally does not use the standard scala types?
why not?