Are the files reopened for each write etc? If locking works glusterfs for 
example could be a nice solution for the replication. Each write would be 
atomically written to all instances, and reads would be local (using AFR with 
preferred servers).

Kind regards,
Fredrik Widlund


-----Original Message-----
From: Suhail Ahmed [mailto:[email protected]]
Sent: den 16 april 2010 10:13
To: [email protected]
Subject: Re: CouchDB and Hadoop

Sure It can be done but for me the whole Java to Erlang layer would be a
mess since they are so different. The better way to go about doing this
would to be implement a distributed file system like Hadoop underneath Couch
for same effect.

On Fri, Apr 16, 2010 at 1:16 AM, Steve-Mustafa Ismail Mustafa <
[email protected]> wrote:

> I swear, I spent over an hour going through the mailing list trying to find
> an answer.
>
> I know that CouchDB is a document oriented DB and I know that Hadoop is a
> File System and that both implement Map/Reduce.  But is it possible to have
> them stacked with Hadoop being the FS in use and CouchDB being the DB? This
> way, wouldn't you get the distributed/clustered FS abilities of Hadoop in
> addition to the powerful retrieval abilities of CouchDB?
>
> If its not possible, and I suspect that it is so, _why_? Don't they operate
> on two seperate levels? Wouldn't CouchDB sort of replace HBase?
>
> Thanks in advance for any and all replies
>

Reply via email to