Hi,

I'm using spark on a cluster of 8 VMs, each with two cores and 3.5GB RAM.

But I need to run a shortest path algorithm on data of 500+GB(textfile, each
line contains a node id and nodes it points to)

I've tested it on the cluster, but the speed seems to be extremely slow, and
haven't got any result yet.

Is it natural to be so slow based on such cluster and data, or there is
something wrong since the problem can be solved much efficiently?(say half
an hour after reading the data?)

Thanks!



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/The-running-time-of-spark-tp12624.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org

Reply via email to