Return-Path: Delivered-To: apmail-lucene-hadoop-user-archive@locus.apache.org Received: (qmail 42940 invoked from network); 14 Jul 2006 18:10:49 -0000 Received: from hermes.apache.org (HELO mail.apache.org) (209.237.227.199) by minotaur.apache.org with SMTP; 14 Jul 2006 18:10:49 -0000 Received: (qmail 83450 invoked by uid 500); 14 Jul 2006 18:10:49 -0000 Delivered-To: apmail-lucene-hadoop-user-archive@lucene.apache.org Received: (qmail 83258 invoked by uid 500); 14 Jul 2006 18:10:48 -0000 Mailing-List: contact hadoop-user-help@lucene.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: hadoop-user@lucene.apache.org Delivered-To: mailing list hadoop-user@lucene.apache.org Received: (qmail 83249 invoked by uid 99); 14 Jul 2006 18:10:48 -0000 Received: from asf.osuosl.org (HELO asf.osuosl.org) (140.211.166.49) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 14 Jul 2006 11:10:48 -0700 X-ASF-Spam-Status: No, hits=1.4 required=10.0 tests=DNS_FROM_RFC_ABUSE,DNS_FROM_RFC_WHOIS X-Spam-Check-By: apache.org Received-SPF: neutral (asf.osuosl.org: local policy) Received: from [216.145.54.172] (HELO mrout2.yahoo.com) (216.145.54.172) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 14 Jul 2006 11:10:47 -0700 Received: from [10.72.104.183] (enoughhot-lx.corp.yahoo.com [10.72.104.183]) by mrout2.yahoo.com (8.13.6/8.13.6/y.out) with ESMTP id k6EI9xjg042989 for ; Fri, 14 Jul 2006 11:10:01 -0700 (PDT) DomainKey-Signature: a=rsa-sha1; s=serpent; d=yahoo-inc.com; c=nofws; q=dns; h=message-id:date:from:user-agent:x-accept-language: mime-version:to:subject:references:in-reply-to:content-type:content-transfer-encoding; b=kHz12+ZYraP/aQuAx7kefYc35BBPc8b/8W9QwmO6P5MWc6POP8+5m6gPdbfQygzi Message-ID: <44B7DDF4.90709@yahoo-inc.com> Date: Fri, 14 Jul 2006 11:09:56 -0700 From: Konstantin Shvachko User-Agent: Mozilla Thunderbird 1.0.6 (Windows/20050716) X-Accept-Language: en-us, en MIME-Version: 1.0 To: hadoop-user@lucene.apache.org Subject: Re: What about append in hadoop files ? References: <20060714080611.98224.qmail@web34308.mail.mud.yahoo.com> <44B763D6.6000503@apache.org> In-Reply-To: <44B763D6.6000503@apache.org> Content-Type: text/plain; charset=ISO-8859-1; format=flowed Content-Transfer-Encoding: 7bit X-Virus-Checked: Checked by ClamAV on apache.org X-Spam-Rating: minotaur.apache.org 1.6.2 0/1000/N Eric, I remember Doug advised somebody on a related issue to use a directory instead of a file for long lasting appends. You can logically divide your output into smaller files and close them whenever the logical boundary is reached. The directory can be treated as a collection of records. May be this will work for you. IMO the concurrent append feature is a high priority task. --Konstantin Doug Cutting wrote: > drwho wrote: > >> If so, GFS, is also suitable only for large, offline, batch >> computations ? >> I wonder how Google is going to use GFS for writely or their online >> spreadsheet or their BigTable (their gigantic relational DB). > > > Did I say anything about GFS? I don't think so. Also, I said, > "currently" and "primarily", not "forever" and "exclusively". I would > love for DFS to be more suitable for online, incremental stuff, but > we're a ways from that right now. As I said, we're pursuing > reliability, scalability and performance before features like append. > If you'd like to try to implement append w/o disrupting work on > reliability scalability and performance, we'd welcome your > contributions. The project direction is determined by contributors. > > Note that BigTable is a complex layer on top of GFS that caches and > batches i/o. So, while GFS does implement some features that DFS > still does not (like appends), GFS is probably not used directly by, > e.g., writely. Finally, BigTable is not relational. > > Doug > >> Doug Cutting wrote: >> >> DFS is currently primarily used to support large, offline, batch >> computations. For example, a log of critical data with tight >> transactional requirements is probably an inappropriate use of DFS at >> this time. Again, this may change, but that's where we are now. >> >> Doug >> >> >> >> >> Thanks much. >> >> -eric >> > > >