Return-Path: Delivered-To: apmail-lucene-hadoop-dev-archive@locus.apache.org Received: (qmail 66396 invoked from network); 29 Aug 2006 20:36:31 -0000 Received: from hermes.apache.org (HELO mail.apache.org) (209.237.227.199) by minotaur.apache.org with SMTP; 29 Aug 2006 20:36:31 -0000 Received: (qmail 83242 invoked by uid 500); 29 Aug 2006 20:36:23 -0000 Delivered-To: apmail-lucene-hadoop-dev-archive@lucene.apache.org Received: (qmail 82850 invoked by uid 500); 29 Aug 2006 20:36:21 -0000 Mailing-List: contact hadoop-dev-help@lucene.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: hadoop-dev@lucene.apache.org Delivered-To: mailing list hadoop-dev@lucene.apache.org Received: (qmail 82799 invoked by uid 99); 29 Aug 2006 20:36:18 -0000 Received: from asf.osuosl.org (HELO asf.osuosl.org) (140.211.166.49) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 29 Aug 2006 13:36:18 -0700 X-ASF-Spam-Status: No, hits=1.4 required=10.0 tests=DNS_FROM_RFC_ABUSE,DNS_FROM_RFC_WHOIS X-Spam-Check-By: apache.org Received-SPF: neutral (asf.osuosl.org: local policy) Received: from [207.126.228.150] (HELO rsmtp2.corp.yahoo.com) (207.126.228.150) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 29 Aug 2006 13:36:15 -0700 Received: from [10.72.108.58] (arkady-mac.corp.yahoo.com [10.72.108.58]) by rsmtp2.corp.yahoo.com (8.13.6/8.13.6/y.rout) with ESMTP id k7TKZp6c005783; Tue, 29 Aug 2006 13:35:51 -0700 (PDT) DomainKey-Signature: a=rsa-sha1; s=serpent; d=yahoo-inc.com; c=nofws; q=dns; h=in-reply-to:references:mime-version:content-type:message-id: content-transfer-encoding:from:subject:date:to:x-mailer; b=NHyhKqu+uLl6X57KXwiDGGMnemUG23XP+bmXaJ9E8eVZMDPyr0Pj8eLPDrLH/zgD In-Reply-To: <30846588.1156879178841.JavaMail.jira@brutus> References: <30846588.1156879178841.JavaMail.jira@brutus> Mime-Version: 1.0 (Apple Message framework v624) Content-Type: text/plain; charset=US-ASCII; delsp=yes; format=flowed Message-Id: Content-Transfer-Encoding: 7bit From: Arkady Borkovsky Subject: Re: [jira] Commented: (HADOOP-489) Seperating user logs from system logs in map reduce Date: Tue, 29 Aug 2006 13:35:49 -0700 To: hadoop-dev@lucene.apache.org X-Mailer: Apple Mail (2.624) X-Virus-Checked: Checked by ClamAV on apache.org X-Spam-Rating: minotaur.apache.org 1.6.2 0/1000/N +1 for Michel's points. With these point in mind, solution (3) does not look attractive -- at least as long as the tools for immediate access to such logs are perfect. I like solution (2). Concatenation is separate issue -- the important thing is immediate availability. Once the logs are written, it will be not to hard to provide web UI to access them. Maybe solution (2) be modified so that the messages from all tasks go to the single DFS files -- each line of the logs prefixed with task ID and time stamp? On Aug 29, 2006, at 12:19 PM, Michel Tourn (JIRA) wrote: > [ > http://issues.apache.org/jira/browse/HADOOP-489? > page=comments#action_12431327 ] > > Michel Tourn commented on HADOOP-489: > ------------------------------------- > > I don't know what is the best implementation. > > But the normal requirements for this info are: > > 1. Task log data should be made available in real-time. The main point > is to debug execution problems at startupt. The point is not to gather > after-the-fact throughput metrics on successful jobs. > Example messages we want to see immediately (so we can kill the job): > /usr/custombin/perl : program not found > /usr/bin/perl: myscript.pl: file not found > > 2. Leave any log aggregation to the consumers. Aggregation gets in the > way of timeliness: after the FIRST map task fails as in 1., we want to > access its log. Waiting for failure-everywhere is not an acceptable > alternative. (due to time delays, partial success and 3x reexecution) > > 3. Nice to have: access to a single Task log is as a stream, not a > file or blob. > i.e. we do not need to wait for Task termination to access its log. > However this is more constraining on the implementation, and for the > above use case of failure at launch it is not necessary. > > >> Seperating user logs from system logs in map reduce >> --------------------------------------------------- >> >> Key: HADOOP-489 >> URL: http://issues.apache.org/jira/browse/HADOOP-489 >> Project: Hadoop >> Issue Type: Improvement >> Components: mapred >> Reporter: Mahadev konar >> Assigned To: Mahadev konar >> Priority: Minor >> >> Currently the user logs are a part of system logs in mapreduce. >> Anything logged by the user is logged into the tasktracker log files. >> This create two issues- >> 1) The system log files get cluttered with user output. If the user >> outputs a large amount of logs, the system logs need to be cleaned up >> pretty often. >> 2) For the user, it is difficult to get to each of the machines and >> look for the logs his/her job might have generated. >> I am proposing three solutions to the problem. All of them have >> issues with it - >> Solution 1. >> Output the user logs on the user screen as part of the job submission >> process. >> Merits- >> This will prevent users from printing large amount of logs and the >> user can get runtime feedback on what is wrong with his/her job. >> Issues - >> This proposal will use the framework bandwidth while running jobs for >> the user. The user logs will need to pass from the tasks to the >> tasktrackers, from the tasktrackers to the jobtrackers and then from >> the jobtrackers to the jobclient using a lot of framework bandwidth >> if the user is printing out too much data. >> Solution 2. >> Output the user logs onto a dfs directory and then concatenate these >> files. Each task can create a file for the output in the log >> direcotyr for a given user and jobid. >> Issues - >> This will create a huge amount of small files in DFS which later can >> be concatenated into a single file. Also there is this issue that who >> would concatenate these files into a single file? This could be done >> by the framework (jobtracker) as part of the cleanup for the jobs - >> might stress the jobtracker. >> >> Solution 3. >> Put the user logs into a seperate user log file in the log directory >> on each tasktrackers. We can provide some tools to query these local >> log files. We could have commands like for jobid j and for taskid t >> get me the user log output. These tools could run as a seperate map >> reduce program with each map grepping the user log files and a single >> recude aggregating these logs in to a single dfs file. >> Issues- >> This does sound like more work for the user. Also, the output might >> not be complete since a tasktracker might have went down after it ran >> the job. >> Any thoughts? > > -- > This message is automatically generated by JIRA. > - > If you think it was sent incorrectly contact one of the > administrators: > http://issues.apache.org/jira/secure/Administrators.jspa > - > For more information on JIRA, see: > http://www.atlassian.com/software/jira > >