Return-Path: Delivered-To: apmail-hadoop-common-user-archive@www.apache.org Received: (qmail 24236 invoked from network); 15 Jul 2009 05:40:20 -0000 Received: from hermes.apache.org (HELO mail.apache.org) (140.211.11.3) by minotaur.apache.org with SMTP; 15 Jul 2009 05:40:20 -0000 Received: (qmail 96117 invoked by uid 500); 15 Jul 2009 05:40:28 -0000 Delivered-To: apmail-hadoop-common-user-archive@hadoop.apache.org Received: (qmail 96019 invoked by uid 500); 15 Jul 2009 05:40:28 -0000 Mailing-List: contact common-user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: common-user@hadoop.apache.org Delivered-To: mailing list common-user@hadoop.apache.org Received: (qmail 96009 invoked by uid 500); 15 Jul 2009 05:40:28 -0000 Delivered-To: apmail-hadoop-core-user@hadoop.apache.org Received: (qmail 96006 invoked by uid 99); 15 Jul 2009 05:40:28 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 15 Jul 2009 05:40:28 +0000 X-ASF-Spam-Status: No, hits=2.2 required=10.0 tests=HTML_MESSAGE,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: domain of sugandha.n87@gmail.com designates 209.85.222.192 as permitted sender) Received: from [209.85.222.192] (HELO mail-pz0-f192.google.com) (209.85.222.192) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 15 Jul 2009 05:40:18 +0000 Received: by pzk30 with SMTP id 30so4392033pzk.5 for ; Tue, 14 Jul 2009 22:39:57 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=domainkey-signature:mime-version:received:date:message-id:subject :from:to:content-type; bh=59cjDssdjDw2oIc9ciN45n0K8SkCf9KLViDqJyIO+O0=; b=CRZYRAo6vPxxFyv4UbixY8XyZMHtK7phpSaUD/NSLa1akL9t5rtE0sdSXgZoe73WOI BAYHZUYk8ke3p8G3B9rmEmEHOODSEVUQK6eojCKX5O501wbRwlY03+/LELwMjPXiZ7pn NmTNK9IJVVD+YW+cRXelEZcRR3h5I0LOns7rM= DomainKey-Signature: a=rsa-sha1; c=nofws; d=gmail.com; s=gamma; h=mime-version:date:message-id:subject:from:to:content-type; b=loW4iQXarnHvh1sw7wKEiMjdo/KMbOLRTCOfQynePaXaihIipo8ndOfX+UWsdCXhyK 1rbgRfvN5A5mJa4+jNwcxzYIQ9UFQnecnz8RoHYQTWsJn3HxNPMKH88hTjG7kIRlEf6A MN1itrFUroKc7yp5ttEkJm84OPJEaHuFtE4ew= MIME-Version: 1.0 Received: by 10.143.33.16 with SMTP id l16mr1712708wfj.333.1247636397710; Tue, 14 Jul 2009 22:39:57 -0700 (PDT) Date: Wed, 15 Jul 2009 11:09:57 +0530 Message-ID: <6f72e2db0907142239l2b3f687as8820fdc252fe67f6@mail.gmail.com> Subject: Compression issues!! From: Sugandha Naolekar To: core-user@hadoop.apache.org Content-Type: multipart/alternative; boundary=001636e0ad7f4b5eab046eb7fc13 X-Virus-Checked: Checked by ClamAV on apache.org --001636e0ad7f4b5eab046eb7fc13 Content-Type: text/plain; charset=ISO-8859-1 Content-Transfer-Encoding: 7bit Hello! Few days back, I had asked about the compression of data placed in hadoop..I did get apt replies as:: Place the data first in HDFS and then compress it, so that the data would be in sequence files. But, here my query is, I want to compress the data before placing it in HDFS, so that redundancy won't come into picture..! How to do that...!Also, will I have to use external compression algo. or simply api's would solve the purpose? -- Regards! Sugandha --001636e0ad7f4b5eab046eb7fc13--