Or RDD.max() and RDD.min() won't work for you?
Yong

Subject: Re: Calculating Min and Max Values using Spark Transformations?
To: as...@wso2.com
CC: user@spark.apache.org
From: jfc...@us.ibm.com
Date: Fri, 28 Aug 2015 09:28:43 -0700


If you already loaded csv data into a dataframe, why not register it as a 
table, and use Spark SQL

to find max/min or any other aggregates? SELECT MAX(column_name) FROM 
dftable_name ... seems natural.












JESSE CHEN

Big Data Performance | IBM Analytics



Office:  408 463 2296

Mobile: 408 828 9068

Email:   jfc...@us.ibm.com








ashensw ---08/28/2015 05:40:07 AM---Hi all, I have a dataset which consist of 
large number of features(columns). It is



From:   ashensw <as...@wso2.com>

To:     user@spark.apache.org

Date:   08/28/2015 05:40 AM

Subject:        Calculating Min and Max Values using Spark Transformations?








Hi all,



I have a dataset which consist of large number of features(columns). It is

in csv format. So I loaded it into a spark dataframe. Then I converted it

into a JavaRDD<Row> Then using a spark transformation I converted that into

JavaRDD<String[]>. Then again converted it into a JavaRDD<double[]>. So now

I have a JavaRDD<double[]>. So is there any method to calculate max and min

values of each columns in this JavaRDD<double[]> ?  



Or Is there any way to access the array if I store max and min values to a

array inside the spark transformation class?



Thanks.







--

View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/Calculating-Min-and-Max-Values-using-Spark-Transformations-tp24491.html

Sent from the Apache Spark User List mailing list archive at Nabble.com.



---------------------------------------------------------------------

To unsubscribe, e-mail: user-unsubscr...@spark.apache.org

For additional commands, e-mail: user-h...@spark.apache.org




                                          

Reply via email to