What is the way of testing/building the pyspark part of Spark ? Le jeu. 23 avr. 2015 à 22:06, Olivier Girardot < o.girar...@lateral-thoughts.com> a écrit :
> yep :) I'll open the jira when I've got the time. > Thanks > > Le jeu. 23 avr. 2015 à 19:31, Reynold Xin <r...@databricks.com> a écrit : > >> Ah damn. We need to add it to the Python list. Would you like to give it >> a shot? >> >> >> On Thu, Apr 23, 2015 at 4:31 AM, Olivier Girardot < >> o.girar...@lateral-thoughts.com> wrote: >> >>> Yep no problem, but I can't seem to find the coalesce fonction in >>> pyspark.sql.{*, functions, types or whatever :) } >>> >>> Olivier. >>> >>> Le lun. 20 avr. 2015 à 11:48, Olivier Girardot < >>> o.girar...@lateral-thoughts.com> a écrit : >>> >>> > a UDF might be a good idea no ? >>> > >>> > Le lun. 20 avr. 2015 à 11:17, Olivier Girardot < >>> > o.girar...@lateral-thoughts.com> a écrit : >>> > >>> >> Hi everyone, >>> >> let's assume I'm stuck in 1.3.0, how can I benefit from the *fillna* >>> API >>> >> in PySpark, is there any efficient alternative to mapping the records >>> >> myself ? >>> >> >>> >> Regards, >>> >> >>> >> Olivier. >>> >> >>> > >>> >> >>