Return-Path: X-Original-To: apmail-accumulo-notifications-archive@minotaur.apache.org Delivered-To: apmail-accumulo-notifications-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 4EEAE176DF for ; Thu, 6 Nov 2014 19:46:34 +0000 (UTC) Received: (qmail 15015 invoked by uid 500); 6 Nov 2014 19:46:34 -0000 Delivered-To: apmail-accumulo-notifications-archive@accumulo.apache.org Received: (qmail 14931 invoked by uid 500); 6 Nov 2014 19:46:34 -0000 Mailing-List: contact notifications-help@accumulo.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: jira@apache.org Delivered-To: mailing list notifications@accumulo.apache.org Received: (qmail 14913 invoked by uid 99); 6 Nov 2014 19:46:34 -0000 Received: from arcas.apache.org (HELO arcas.apache.org) (140.211.11.28) by apache.org (qpsmtpd/0.29) with ESMTP; Thu, 06 Nov 2014 19:46:34 +0000 Date: Thu, 6 Nov 2014 19:46:34 +0000 (UTC) From: "Adam Fuchs (JIRA)" To: notifications@accumulo.apache.org Message-ID: In-Reply-To: References: Subject: [jira] [Commented] (ACCUMULO-3303) funky performance with large WAL MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: 7bit X-JIRA-FingerPrint: 30527f35849b9dde25b450d4833f0394 [ https://issues.apache.org/jira/browse/ACCUMULO-3303?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14200763#comment-14200763 ] Adam Fuchs commented on ACCUMULO-3303: -------------------------------------- bq. Accumulo attempts to choose a block size that is a little bigger than the WAL size, up to MAX_INT. HDFS code takes care of making more blocks, if needed. Can you point to a line of code? I'm looking at DfsLogger:365 in the 1.6 branch, and it looks like we're not limiting block size. Incidentally, my latest performance tests are showing that the WAL bottleneck is almost entirely due to metadata management in most cases, rather than writing to the log itself (as in ACCUMULO-2889). Were you and your intern looking strictly at ingest performance, or also recovery time? > funky performance with large WAL > -------------------------------- > > Key: ACCUMULO-3303 > URL: https://issues.apache.org/jira/browse/ACCUMULO-3303 > Project: Accumulo > Issue Type: Bug > Components: logger, tserver > Affects Versions: 1.6.1 > Reporter: Adam Fuchs > Attachments: 1GB_WAL.png, 2GB_WAL.png, 4GB_WAL.png, 512MB_WAL.png, 8GB_WAL.png, WAL_disabled.png > > > The tserver seems to get into a funky state when writing to a large write-ahead log. I ran some continuous ingest tests varying tserver.walog.max.size in {512M, 1G, 2G, 4G, 8G} and got some results that I have yet to understand. I was expecting to see the effects of walog metadata management as described in ACCUMULO-2889, but I also found an additional behavior of ingest slowing down for long periods when using a large walog size. > The cluster configuration was as follows: > {code} > Accumulo version: 1.6.2-SNAPSHOT (current head of origin/1.6) > Nodes: 4 > Masters: 1 > Slaves: 3 > Cores per node: 24 > Drives per node: 8x1TB data + 2 raided system > Memory per node: 64GB > tserver.memory.maps.max=2G > table.file.compress.type=snappy (for ci table only) > tserver.mutation.queue.max=16M > tserver.wal.sync.method=hflush > Native maps enabled > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)