hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Koji Noguchi <knogu...@yahoo-inc.com>
Subject Re: Streaming ignoring stderr output
Date Mon, 26 Oct 2009 15:03:56 GMT
This doesn't solve your stderr/stdout problem, but you can always set the
timeout to be a bigger value if necessary.

-Dmapred.task.timeout=______ (in milliseconds)


On 10/25/09 12:00 PM, "Ryan Rosario" <uclamathguy@gmail.com> wrote:

> I am using a Python script as a mapper for a Hadoop Streaming (hadoop
> 0.20.0) job, with reducer NONE. My jobs keep getting killed with "task
> failed to respond after 600 seconds." I tried sending a heartbeat
> every minute to stderr using sys.stderr.write in my mapper, but
> nothing is being output to stderr either on disk (in
> logs/userlogs/...) or in the web UI. stdout is not even recorded.
> This also means I have no way of knowing what my tasks are doing at
> any given moment except to look at the counts produced in syslog.
> I got it to work once, but have not had any luck since. Any
> suggestions of things to look at as to why I am not able to get any
> output? Help is greatly appreciated.
> - Ryan

View raw message