Hi Dinakar,
Hi Mich,
Posting you my comments,
Right, you seem to have an on-premise Hadoop cluster of 9 physical boxes
and you want to deploy spark on it.
*My comment: Yes.*
What spec do you have for each physical host memory and CPU and disk space?
*My comment: I am not sure, of the exact num
Hi Mich,
Posting you my comments,
Right, you seem to have an on-premise Hadoop cluster of 9 physical boxes
and you want to deploy spark on it.
*My comment: Yes.*
What spec do you have for each physical host memory and CPU and disk space?
*My comment: I am not sure, of the exact numbers. but all
Hi,
Right you seem to have an on-premise hadoop cluster of 9 physical boxes and
you want to deploy spark on it.
What spec do you have for each physical host memory and CPU and disk space?
You can take what is known as data affinity by putting your compute layers
(spark) on the same hadoop nodes.
Hi Dinakar
If you are aim is to run Spark in “distributed mode” then all these cluster
modes (excluding local) runs the cluster in distributed mode anyway.
As I said before the “deployment =cluster” mode is only for Driver application
and executors are running on worker nodes in parallel mode (d