Hello All, Cordial Greetings, I am trying to familiarize myself with Apache Hadoop and it's different software components and how they can be deployed on physical or virtual infrastructure.
I have a few questions: Q1) Can we use Mapreduce and apache spark in the same cluster Q2) is it mandatory to use GPUs for apache spark? Q3) I read that apache spark is in-memory, will it benefit from SSD / Flash for caching or persistent storage? Q4) If we intend to deploy a Hadoop cluster with 6 servers, can we have GPUs in only two and restrict apache spark on those servers only? Q5) Possible to virtualize Spark with GPU Pass thru? Q6) What GPUs are recommended / compatible with Apache Spark? (NVidia M10 / M60)? Will be grateful for your suggestions and answers - please accept my apologies for totally noob questions 😊 Have a good day / evening ahead Best Sent from Outlook<http://aka.ms/weboutlook>