Return-Path: X-Original-To: apmail-hadoop-hdfs-user-archive@minotaur.apache.org Delivered-To: apmail-hadoop-hdfs-user-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 976D8FC34 for ; Wed, 27 Mar 2013 13:34:51 +0000 (UTC) Received: (qmail 38144 invoked by uid 500); 27 Mar 2013 13:34:46 -0000 Delivered-To: apmail-hadoop-hdfs-user-archive@hadoop.apache.org Received: (qmail 37960 invoked by uid 500); 27 Mar 2013 13:34:45 -0000 Mailing-List: contact user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hadoop.apache.org Delivered-To: mailing list user@hadoop.apache.org Received: (qmail 37947 invoked by uid 99); 27 Mar 2013 13:34:45 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 27 Mar 2013 13:34:45 +0000 X-ASF-Spam-Status: No, hits=1.1 required=5.0 tests=NO_RDNS_DOTCOM_HELO,RCVD_IN_DNSWL_NONE,SPF_NEUTRAL X-Spam-Check-By: apache.org Received-SPF: neutral (athena.apache.org: 98.139.253.105 is neither permitted nor denied by domain of knoguchi@yahoo-inc.com) Received: from [98.139.253.105] (HELO mrout2-b.corp.bf1.yahoo.com) (98.139.253.105) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 27 Mar 2013 13:34:40 +0000 Received: from SP2-EX07CAS02.ds.corp.yahoo.com (sp2-ex07cas02.corp.sp2.yahoo.com [98.137.59.38]) by mrout2-b.corp.bf1.yahoo.com (8.14.4/8.14.4/y.out) with ESMTP id r2RDXkiu050036 (version=TLSv1/SSLv3 cipher=RC4-MD5 bits=128 verify=FAIL) for ; Wed, 27 Mar 2013 06:33:52 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=yahoo-inc.com; s=cobra; t=1364391232; bh=w6UfT9R3QUgzZ5EpVE6BpNhXDaaHRMxT+LsYXGIfBzw=; h=From:To:Date:Subject:Message-ID:References:In-Reply-To: Content-Type:Content-Transfer-Encoding:MIME-Version; b=LGewf/StwghQ3tE3yiIM9dKi5rRLRPoiXXjSR6QkDhEyKmu6QJ+gb5Xx3NkGSB3l8 mD3MmxNZL3JNgNsSyS4R4KToZYhokUi4CGgg5bjQG/3TlRS1PRHrLcr73W9aI0HlHD HfKcChglyJIoGx5IUiUhLJlR6+eaa+u//0YeYszU= Received: from SP2-EX07VS04.ds.corp.yahoo.com ([98.137.59.33]) by SP2-EX07CAS02.ds.corp.yahoo.com ([98.137.59.38]) with mapi; Wed, 27 Mar 2013 06:21:51 -0700 From: Koji Noguchi To: "user@hadoop.apache.org" Date: Wed, 27 Mar 2013 06:21:49 -0700 Subject: Re: Auto clean DistCache? Thread-Topic: Auto clean DistCache? Thread-Index: Ac4q7gngO33oQDUETsCJy3vZfD9LuA== Message-ID: <4265D509-D8B9-48AA-B7B6-DBC19CA807CA@yahoo-inc.com> References: <5BDD5440-304F-44C9-B512-56EDAD21BC39@apache.org> In-Reply-To: Accept-Language: en-US Content-Language: en-US X-MS-Has-Attach: X-MS-TNEF-Correlator: acceptlanguage: en-US Content-Type: text/plain; charset="Windows-1252" Content-Transfer-Encoding: quoted-printable MIME-Version: 1.0 X-Milter-Version: master.31+4-gbc07cd5+ X-CLX-ID: 391232000 X-Virus-Checked: Checked by ClamAV on apache.org > Else, I will go for a customed script to delete all directories (and cont= ent) older than 2 or 3 days=85 > TaskTracker (or NodeManager in 2.*) keeps the list of dist cache entries in= memory. So if external process (like your script) start deleting dist cache files, = there would be inconsistency and you'll start seeing task initialization fa= ilures due to no file found error. Koji On Mar 26, 2013, at 9:00 PM, Jean-Marc Spaggiari wrote: > For the situation I faced I was really a disk space issue, not related > to the number of files. It was writing on a small partition. >=20 > I will try with local.cache.size or > mapreduce.tasktracker.cache.local.size to see if I can keep the final > total size under 5GB... Else, I will go for a customed script to > delete all directories (and content) older than 2 or 3 days... >=20 > Thanks, >=20 > JM >=20 > 2013/3/26 Abdelrahman Shettia : >> Let me clarify , If there are lots of files or directories up to 32K ( >> Depending on the user's # of files sys os config) in those distributed c= ache >> dirs, The OS will not be able to create any more files/dirs, Thus M-R jo= bs >> wont get initiated on those tasktracker machines. Hope this helps. >>=20 >>=20 >> Thanks >>=20 >>=20 >> On Tue, Mar 26, 2013 at 1:44 PM, Vinod Kumar Vavilapalli >> wrote: >>>=20 >>>=20 >>> All the files are not opened at the same time ever, so you shouldn't se= e >>> any "# of open files exceeds error". >>>=20 >>> Thanks, >>> +Vinod Kumar Vavilapalli >>> Hortonworks Inc. >>> http://hortonworks.com/ >>>=20 >>> On Mar 26, 2013, at 12:53 PM, Abdelrhman Shettia wrote: >>>=20 >>> Hi JM , >>>=20 >>> Actually these dirs need to be purged by a script that keeps the last 2 >>> days worth of files, Otherwise you may run into # of open files exceeds >>> error. >>>=20 >>> Thanks >>>=20 >>>=20 >>> On Mar 25, 2013, at 5:16 PM, Jean-Marc Spaggiari >>> wrote: >>>=20 >>> Hi, >>>=20 >>>=20 >>> Each time my MR job is run, a directory is created on the TaskTracker >>>=20 >>> under mapred/local/taskTracker/hadoop/distcache (based on my >>>=20 >>> configuration). >>>=20 >>>=20 >>> I looked at the directory today, and it's hosting thousands of >>>=20 >>> directories and more than 8GB of data there. >>>=20 >>>=20 >>> Is there a way to automatically delete this directory when the job is >>> done? >>>=20 >>>=20 >>> Thanks, >>>=20 >>>=20 >>> JM >>>=20 >>>=20 >>>=20 >>=20