Or, use the SparkOnHBase lab.http://blog.cloudera.com/blog/2014/12/new-in-cloudera-labs-sparkonhbase/ From: Ted Yu <yuzhih...@gmail.com> To: Akhil Das <ak...@sigmoidanalytics.com> Cc: sandeep vura <sandeepv...@gmail.com>; "user@spark.apache.org" <user@spark.apache.org> Sent: Monday, February 23, 2015 8:52 AM Subject: Re: How to integrate HBASE on Spark Installing hbase on hadoop cluster would allow hbase to utilize features provided by hdfs, such as short circuit read (See '90.2. Leveraging local data' under http://hbase.apache.org/book.html#perf.hdfs). Cheers On Sun, Feb 22, 2015 at 11:38 PM, Akhil Das <ak...@sigmoidanalytics.com> wrote:
If you are having both the clusters on the same network, then i'd suggest you installing it on the hadoop cluster. If you install it on the spark cluster itself, then hbase might take up a few cpu cycles and there's a chance for the job to lag. ThanksBest Regards On Mon, Feb 23, 2015 at 12:48 PM, sandeep vura <sandeepv...@gmail.com> wrote: Hi I had installed spark on 3 node cluster. Spark services are up and running.But i want to integrate hbase on spark Do i need to install HBASE on hadoop cluster or spark cluster. Please let me know asap. Regards, Sandeep.v