hadoop-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Arun C Murthy <...@hortonworks.com>
Subject Re: Hadoop efficient resource isolation
Date Tue, 26 Feb 2013 14:02:56 GMT
CapacityScheduler has features to allow a user to specify the amount of virtual memory per
map/reduce task and the TaskTracker monitors all tasks and their process-trees to ensure fork-bombs
don't kill the node.

On Feb 25, 2013, at 8:27 PM, Marcin Mejran wrote:

> That won't stop a bad job (say a fork bomb or a massive memory leak in a streaming script)
from taking out a node which is what I believe Dhanasekaran was asking about. He wants to
physically isolate certain lobs to certain "non critical" nodes. I don't believe this is possible
and data would be spread to those nodes, assuming they're data nodes, which would still cause
cluster wide issues (and if data is isolate why not have two separate clusters?),
> I've read references in the docs about some type of memory based contrains in Hadoop
but I don't know of the details. Anyone know how they work?
> Also, I believe there are tools in Linux that can kill processes in case of memory issues
and otherwise restrict what a certain user can do. These seem like a more flexible solution
although they won't cover all potential issues.
> -Marcin
> On Feb 25, 2013, at 7:20 PM, "Arun C Murthy" <acm@hortonworks.com> wrote:
>> CapacityScheduler is what you want...
>> On Feb 21, 2013, at 5:16 AM, Dhanasekaran Anbalagan wrote:
>>> Hi Guys,
>>> It's possible isolation job submission for hadoop cluster, we currently running
48 machine cluster. we  monitor Hadoop is not provides efficient resource isolation. In my
case we ran for tech and research pool, When tech job some memory leak will haven, It's occupy
the hole cluster.  Finally we figure out  issue with tech job. It's  screwed up hole hadoop
cluster. finally 10 data node  are dead.
>>> Any prevention of job submission efficient way resource allocation. When something
wrong in   particular job, effect particular pool, Not effect others job. Any way to archive
>>> Please guide me guys.
>>> My idea is, When tech user submit job means only apply job in for my case submit
24 machine. other machine only for research user.
>>> It's will prevent the memory leak problem. 
>>> -Dhanasekaran.
>>> Did I learn something today? If not, I wasted it.
>> --
>> Arun C. Murthy
>> Hortonworks Inc.
>> http://hortonworks.com/

Arun C. Murthy
Hortonworks Inc.

View raw message