Return-Path: X-Original-To: apmail-spark-user-archive@minotaur.apache.org Delivered-To: apmail-spark-user-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id D12CA18D02 for ; Thu, 8 Oct 2015 10:43:48 +0000 (UTC) Received: (qmail 99811 invoked by uid 500); 8 Oct 2015 10:43:41 -0000 Delivered-To: apmail-spark-user-archive@spark.apache.org Received: (qmail 99715 invoked by uid 500); 8 Oct 2015 10:43:40 -0000 Mailing-List: contact user-help@spark.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Delivered-To: mailing list user@spark.apache.org Received: (qmail 99705 invoked by uid 99); 8 Oct 2015 10:43:40 -0000 Received: from Unknown (HELO spamd1-us-west.apache.org) (209.188.14.142) by apache.org (qpsmtpd/0.29) with ESMTP; Thu, 08 Oct 2015 10:43:40 +0000 Received: from localhost (localhost [127.0.0.1]) by spamd1-us-west.apache.org (ASF Mail Server at spamd1-us-west.apache.org) with ESMTP id 4F206C498E for ; Thu, 8 Oct 2015 10:43:40 +0000 (UTC) X-Virus-Scanned: Debian amavisd-new at spamd1-us-west.apache.org X-Spam-Flag: NO X-Spam-Score: 2.901 X-Spam-Level: ** X-Spam-Status: No, score=2.901 tagged_above=-999 required=6.31 tests=[DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, HTML_MESSAGE=3, URIBL_BLOCKED=0.001] autolearn=disabled Authentication-Results: spamd1-us-west.apache.org (amavisd-new); dkim=pass (1024-bit key) header.d=tink.se Received: from mx1-eu-west.apache.org ([10.40.0.8]) by localhost (spamd1-us-west.apache.org [10.40.0.7]) (amavisd-new, port 10024) with ESMTP id 7ZrYVmg_RfLh for ; Thu, 8 Oct 2015 10:43:27 +0000 (UTC) Received: from mail-yk0-f182.google.com (mail-yk0-f182.google.com [209.85.160.182]) by mx1-eu-west.apache.org (ASF Mail Server at mx1-eu-west.apache.org) with ESMTPS id 8E08F206E8 for ; Thu, 8 Oct 2015 10:43:26 +0000 (UTC) Received: by ykdg206 with SMTP id g206so44439633ykd.1 for ; Thu, 08 Oct 2015 03:43:25 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=tink.se; s=tink; h=mime-version:from:date:message-id:subject:to:content-type; bh=MLcVcNlulTAzUmS2xjWMFVyToBO2T+GmvvIjAW+WgU8=; b=GuPoemmZwrlWMJwGOV1shaEapzkQOQRhSm6ILLY6V+Zl4DJiByKXzZyXIp0LDxFfDp lv7VrWdsOOEU8TqA7kGNqeVGomiuGXTZHAMHPVRwMCmLh2xSlw1hNvQkd/MW+JSh33dq gJSFD4H35y/2I7vClGyv13P27Fulm9IQozPCE= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:mime-version:from:date:message-id:subject:to :content-type; bh=MLcVcNlulTAzUmS2xjWMFVyToBO2T+GmvvIjAW+WgU8=; b=U7AMU5iPDXKuKYhyGIsw8QgMqCciY0qdPeXIfQc48nC4YTAJx50PNm4n279uCBSV5j bvZ4Eyg49BujHuDO8RgOKPonAdPQf4RQVJVLFFECxv2q1dK1b9Op1BUgX55gMZ2iztNP ILsb5u5gCpngCxFvAqgkLUwrYwDS7Z3Ya/WNf0m0uHZspHni1a578WRTB0cvt+AZVXxm RNZUxAhnAfkHcXZuqE/CokUDWC9PED12dsMZOqZ7JOvWVVAHWqnTU53sygzznI7NlmOx tSSeMzWx7bo3ezssrEVC53sLAFMPuDXS7koNuml28xlkkQbjCyPyRiVwmuy9KujOx9Fv Jl6Q== X-Gm-Message-State: ALoCoQnvFGyPDiknI5hfqICHC1XyySWdqIJIl0lVRTlaOdKVqHQRt0R8YYF4MnCZHDaoTrhHjs/d X-Received: by 10.13.205.71 with SMTP id p68mr4542480ywd.164.1444301005440; Thu, 08 Oct 2015 03:43:25 -0700 (PDT) MIME-Version: 1.0 Received: by 10.129.75.195 with HTTP; Thu, 8 Oct 2015 03:43:06 -0700 (PDT) From: Jens Rantil Date: Thu, 8 Oct 2015 12:43:06 +0200 Message-ID: Subject: Best practises to clean up RDDs for old applications To: user@spark.apache.org Content-Type: multipart/alternative; boundary=001a114dab9860f7940521958837 --001a114dab9860f7940521958837 Content-Type: text/plain; charset=UTF-8 Hi, I have a couple of old application RDDs under /var/lib/spark/rdd that haven't been properly cleaned up after themselves. Example: # du -shx /var/lib/spark/rdd/* 44K /var/lib/spark/rdd/liblz4-java1011984124691611873.so 48K /var/lib/spark/rdd/snappy-1.0.5-libsnappyjava.so 2.3G /var/lib/spark/rdd/spark-local-20150903112858-a72d 23M /var/lib/spark/rdd/spark-local-20150929141201-143f The applications (such as "20150903112858") aren't running anymore. What are best practises to clean these up? A cron job? Enabling some kind of cleaner in Spark? I'm currently running Spark 1.1, but will eventually move to 1.2 and then 1.4. Thanks, Jens -- Jens Rantil Backend engineer Tink AB Email: jens.rantil@tink.se Phone: +46 708 84 18 32 Web: www.tink.se Facebook Linkedin Twitter --001a114dab9860f7940521958837 Content-Type: text/html; charset=UTF-8 Content-Transfer-Encoding: quoted-printable
Hi,

I have a couple of old application = RDDs under /var/lib/spark/rdd that haven't been properly cleaned up aft= er themselves. Example:

# du -shx /var/lib/sp= ark/rdd/*
44K /= var/lib/spark/rdd/liblz4-java1011984124691611873.so
48K /var/lib/spark/rdd/snappy-1.0.5-libsnappyjava.so
2.3G /var/lib/spark/rd= d/spark-local-20150903112858-a72d
23M /var/lib/spark/rdd/spark-local-20150929141201-143f

The applications (such as "20150903112858"= ;) aren't running anymore. What are best practises to clean these up? A= cron job? Enabling some kind of cleaner in Spark? I'm currently runnin= g Spark 1.1, but will eventually move to 1.2 and then 1.4.

Thanks,
Jens=C2=A0

--
Jens Rantil
Backend en= gineer
Tink AB

Phone: +46 708 84 18 32
Web:=C2= =A0www.tink.se

--001a114dab9860f7940521958837--