Hello All,
Cordial Greetings,

I am trying to familiarize myself with Apache Hadoop and it's different 
software components and how they can be deployed on physical or virtual 
infrastructure.

I have a few questions:

Q1) Can we use Mapreduce and apache spark in the same cluster
Q2) is it mandatory to use GPUs for apache spark?
Q3) I read that apache spark is in-memory, will it benefit from SSD / Flash for 
caching or persistent storage?
Q4) If we intend to deploy a Hadoop cluster with 6 servers, can we have GPUs in 
only two and restrict apache spark on those servers only?
Q5) Possible to virtualize Spark with GPU Pass thru?
Q6) What GPUs are recommended / compatible with Apache Spark? (NVidia M10 / 
M60)?

Will be grateful for your suggestions and answers - please accept my apologies 
for totally noob questions 😊

Have a good day / evening ahead
Best



Sent from Outlook<http://aka.ms/weboutlook>

Reply via email to