Edward, You are running with insufficient resources -- too little CPU for your task and too little disk for your data.
If you are running a mapreduce task and DFS runs out of space for the temporary files, then you indeed should expect aberrant job status from the Hadoop job framework, for example such things as completion status running backwards. I do agree that under these circumstances HBase daemons should fail more gracefully, by entering some kind of degraded read only mode, if DFS is not totally dead. I suspect this is already on a to do list somewhere, and I vaguely recall a jira filed on that topic. - Andy > From: Edward J. Yoon <[EMAIL PROTECTED]> > Subject: Re: Bulk import question. > To: [email protected], [EMAIL PROTECTED] > Date: Monday, December 1, 2008, 6:26 PM > It was by 'Datanode DiskOutOfSpaceException'. But, I > think daemons should not dead. > > On Wed, Nov 26, 2008 at 1:08 PM, Edward J. Yoon > <[EMAIL PROTECTED]> wrote: > > Hmm. It often occurs to me. I'll check the logs. > > > > On Fri, Nov 21, 2008 at 9:46 AM, Andrew Purtell > <[EMAIL PROTECTED]> wrote: > > > I think a 2 node cluster is simply too small for > > > the full load of everything. > > >
