Re: Check for null in PySpark DataFrame

2015-07-02 Thread Pedro Rodriguez
>>> "b"), (4, None)])) >>> df.where('_2 is not null').count() >>> >>> However, this won't work >>> df.where(df._2 != None).count() >>> >>> It seems there is no native Python way with DataFrames to do this, but I >

Re: Check for null in PySpark DataFrame

2015-07-01 Thread Michael Armbrust
gt;> It seems there is no native Python way with DataFrames to do this, but I >> find that difficult to believe and more likely that I am missing the >> "right >> way" to do this. >> >> >> >> -- >>

Re: Check for null in PySpark DataFrame

2015-07-01 Thread Olivier Girardot
(df._2 != None).count() > > It seems there is no native Python way with DataFrames to do this, but I > find that difficult to believe and more likely that I am missing the "right > way" to do this. > > > > -- > View this message in context: > http://apache-

Check for null in PySpark DataFrame

2015-06-30 Thread pedro
xt: http://apache-spark-user-list.1001560.n3.nabble.com/Check-for-null-in-PySpark-DataFrame-tp23553.html Sent from the Apache Spark User List mailing list archive at Nabble.com. - To unsubscribe, e-mail: user-unsubscr...@spark.apach