Hi all,
I'm writing here after some intensive usage on pyspark and SparkSQL.
I would like to use a well known function in the R world: coxph() from the
survival package.
>From what I understood, I can't parallelize a function like coxph() because
it isn't provided with the SparkR package. In other words, I should
implement a SparkR compatible algorithm instead of using coxph().
I have no chance to make coxph() parallelizable, right?
More generally, I think this is true for any non-spark function which only
accept data.frame format as the data input. 

Do you plan to implement the coxph() counterpart in Spark? The most useful
version of this model is the Cox Regression Model for Time-Dependent
Covariates, which is missing from ANY ML framework as far as I know.

Thank you
 Pietro



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/Spark-R-guidelines-for-non-spark-functions-and-coxph-Cox-Regression-for-Time-Dependent-Covariates-tp28077.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

---------------------------------------------------------------------
To unsubscribe e-mail: user-unsubscr...@spark.apache.org

Reply via email to