Return-Path: X-Original-To: apmail-hadoop-hdfs-user-archive@minotaur.apache.org Delivered-To: apmail-hadoop-hdfs-user-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 1AF9810A39 for ; Wed, 7 Aug 2013 16:28:42 +0000 (UTC) Received: (qmail 23363 invoked by uid 500); 7 Aug 2013 16:28:36 -0000 Delivered-To: apmail-hadoop-hdfs-user-archive@hadoop.apache.org Received: (qmail 22916 invoked by uid 500); 7 Aug 2013 16:28:34 -0000 Mailing-List: contact user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hadoop.apache.org Delivered-To: mailing list user@hadoop.apache.org Received: (qmail 22908 invoked by uid 99); 7 Aug 2013 16:28:33 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 07 Aug 2013 16:28:33 +0000 X-ASF-Spam-Status: No, hits=1.7 required=5.0 tests=FREEMAIL_ENVFROM_END_DIGIT,HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: domain of vboylin1987@gmail.com designates 209.85.160.50 as permitted sender) Received: from [209.85.160.50] (HELO mail-pb0-f50.google.com) (209.85.160.50) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 07 Aug 2013 16:28:29 +0000 Received: by mail-pb0-f50.google.com with SMTP id uo5so2095673pbc.37 for ; Wed, 07 Aug 2013 09:28:08 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :content-type; bh=FTeHqJgWYvhwYMh3NqqJRVXyBXJGb2Ca8Tn+iwazhu0=; b=A+iBZrpGIaYUTh5jdPLx4PbUZM24sFFrcjx5svlIgIFUz+LcfWKAOcfsZdVBu0HPTu jhQA6z+ANwG37N0oFgjT6S/RaU1fKPy63aydSxexS1RrWaN9x9e0AFqFEUZRfO5f+L8p c8JmGyoQx19cYGoJI9FOu3wyCZJgL7YEOLOH9RaKViWkemj51GeI0WEaMY7TU9pwhe6j SUX+LzYx7NKsUl/PsgTquYBg/kcEF/ujMW3TblAW6coeZ8kU3+q4IJrCKb5K+83uCm7u qumE8N/3BaA78R+eQTHcGNSomgxz1EtN8u3+JTXpCyx6DBS9iU7HK2moYmCoUzArEJE4 C4og== MIME-Version: 1.0 X-Received: by 10.68.96.133 with SMTP id ds5mr1401773pbb.169.1375892888384; Wed, 07 Aug 2013 09:28:08 -0700 (PDT) Received: by 10.70.4.98 with HTTP; Wed, 7 Aug 2013 09:28:08 -0700 (PDT) In-Reply-To: References: Date: Thu, 8 Aug 2013 00:28:08 +0800 Message-ID: Subject: Re: Is there any way to use a hdfs file as a Circular buffer? From: Wukang Lin To: hadoop-user Content-Type: multipart/alternative; boundary=047d7b624c4adc784704e35e06d1 X-Virus-Checked: Checked by ClamAV on apache.org --047d7b624c4adc784704e35e06d1 Content-Type: text/plain; charset=ISO-8859-1 Content-Transfer-Encoding: quoted-printable Hi Niels and Bertrand, Thank you for you great advices. In our scenario, we need to store a steady stream of binary data into a circular storage,throughput and concurrency are the most important indicators.The first way seems work, but as hdfs is not friendly for small files, this approche may be not smooth enough.HBase is good, but not appropriate for us, both for throughput and storage.mongodb is quite good for web applications, but not suitable the scenario we meet all the same. we need a distributed storage system,with Highe throughput, HA,LB and secure. Maybe It act much like hbase, manager a lot of small file(hfile) as a large region. we manager a lot of small file as a large one. Perhaps we should develop it by ourselives. Thank you. Lin Wukang 2013/7/25 Niels Basjes > A circular file on hdfs is not possible. > > Some of the ways around this limitation: > - Create a series of files and delete the oldest file when you have too > much. > - Put the data into an hbase table and do something similar. > - Use completely different technology like mongodb which has built in > support for a circular buffer (capped collection). > > Niels > > Hi all, > Is there any way to use a hdfs file as a Circular buffer? I mean, if I= set a quotas to a directory on hdfs, and writting data to a file in that d= irectory continuously. Once the quotas exceeded, I can redirect the writter= and write the data from the beginning of the file automatically . > > --047d7b624c4adc784704e35e06d1 Content-Type: text/html; charset=ISO-8859-1 Content-Transfer-Encoding: quoted-printable
Hi Niels and Bertrand,
=A0 =A0 Thank you for you great= advices.
=A0 =A0 In our scenario, we need to store=A0a steady st= ream of binary data into a circular=A0storage,throughput and concurrency ar= e the most important indicators.The first way seems work, but as =A0hdfs is= not friendly for small files, this=A0approche may be not smooth enough.HBa= se is good, but =A0not appropriate for us, both for throughput and storage.= mongodb=A0is quite good for web applications, but=A0<= /span>not suitable the scenario we meet al= l the same.
=A0 =A0 we need a distributed st= orage system,with Highe=A0throughput,=A0HA,LB and secure. Maybe It act much like hbase, manager a = lot of small file(hfile) as a large region. we manager a lot of small file = as a large one.=A0Perhaps we should= develop it by ourselives.

Thank you.
Lin Wukang


2013/7/25 Nie= ls Basjes <Niels@basjes.nl>

A circular file on hdfs is not possible.

Some of the ways around this limitation:
- Create a series of files and delete the oldest file when you have too muc= h.
- Put the data into an hbase table and do something similar.
- Use completely different technology like mongodb which has built in suppo= rt for a circular buffer (capped collection).

Niels

Hi all,
   Is there any way to use a hdfs file as a Circular buffer? I mean, if I s=
et a quotas to a directory on hdfs, and writting data to a file in that dir=
ectory continuously. Once the quotas exceeded, I can redirect the writter a=
nd write the data from the beginning of the file automatically .

--047d7b624c4adc784704e35e06d1--