Virtual Machines like VMWare give you a lot of control over this.
They're a pain to set up but in production this might be the most
stable solution.


On Tue, Nov 2, 2010 at 6:17 PM, Luke Lu <l...@vicaya.com> wrote:
> nice(1) only changes cpu scheduling priority. It doesn't really help
> if you have tasks (and their child processes) that use too much
> memory, which causes swapping, which is probably the real culprit to
> cause servers to freeze. Decreasing kernel swappiness probably helps.
> Another thing to try is ionice (on linux if you have reasonably recent
> kernel with cfq as io scheduler, default for rhel5) if the freeze is
> caused by io contention (assuming no swapping.)
>
> You can write a simple script to periodically renice(1) and ionice(1)
> these processes to see if they actually work for you.
>
> On Tue, Nov 2, 2010 at 4:51 PM, Jinsong Hu <jinsong...@hotmail.com> wrote:
>> Hi: there:
>>  I have a cluster that is used for both hadoop mapreduce and hbase. What I
>> found is that when I am running map/reduce jobs, the job can be very
>> memory/cpu intensive, and cause hbase or data nodes to freeze. in hbase's
>> case, the region server may shut it self down.
>>  In order to avoid this, I made very conservative configuration of the
>> maximum number of mappers and reducers. However, I am wonder if hadoop
>> allows me to start map/reduce with the command "nice" so that
>> those jobs get lower priority than datanode/tasktracker/hbase regionserver.
>> That way, if there is enough resource, the jobs can fully utilize them. but
>> if not, those jobs will yield to other processes.
>>
>> Jimmy.
>>
>



-- 
Lance Norskog
goks...@gmail.com

Reply via email to