You can install both the datanode and hbase region server on your cluster machines. This is common practice. Obviously you should use decent hardware, like a multi core processor, more than enough memory and multiple hard disks.
2011/1/5 Friso van Vollenhoven <[email protected]> > Hi Peter, > > Do you mean you want to use the HDFS that HBase relies on for other things > and not just exclusively HBase? That should be just fine. We do it all the > time. > > Are you worried about putting to much load on it? I guess that depends on > the type of work load that you have and what you do with it. But generally I > think it is nice to have all nodes be the same (so all workers are datanode > and region server), such that you don't have to scale out them separately. > > > Friso > > > > On 5 jan 2011, at 14:41, Peter Veentjer wrote: > > > Hi Guys, > > > > I'm currently writing a POC based on hbase and I spend more time on > writing > > a ui than on writing the hbase functionality. So I'm very excited about > > exploring HBase further and doing some serious performance and > scalability > > tests and see if we can use it as core technology instead of the > > time/resource intensive Gigaspaces. > > > > My question: > > > > I'm currently using HBase and I also want to use the HDFS directly to > store > > files. If the HBase server(s) is installed, can I directly access the > HDFS > > of these servers or is it better to set up a seperate Hadoop server for > > running HDFS. > >
