Hi - We have a use case for batch processing for which we are trying to figure out if Apache Spark would be a good fit or not.
We have a universe of identifiers sitting in RDBMS for which we need to go get input data from RDBMS and then pass that input to analytical models that generate some output numbers and store it back to the database. This is one unit of work for us. So basically we are looking where we can do this processing in parallel for the universe of identifiers that we have. All the data is in RDBMS and is not sitting in file system. Can we use spark for this kind of work and would it be a good fit for that? Thanks for your help. -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Evaluating-Spark-just-for-Cluster-Computing-tp6110.html Sent from the Apache Spark User List mailing list archive at Nabble.com.