Return-Path: X-Original-To: apmail-cassandra-commits-archive@www.apache.org Delivered-To: apmail-cassandra-commits-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id AD45B18AA8 for ; Wed, 30 Sep 2015 03:43:05 +0000 (UTC) Received: (qmail 86587 invoked by uid 500); 30 Sep 2015 03:43:05 -0000 Delivered-To: apmail-cassandra-commits-archive@cassandra.apache.org Received: (qmail 86420 invoked by uid 500); 30 Sep 2015 03:43:05 -0000 Mailing-List: contact commits-help@cassandra.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: dev@cassandra.apache.org Delivered-To: mailing list commits@cassandra.apache.org Received: (qmail 86403 invoked by uid 99); 30 Sep 2015 03:43:05 -0000 Received: from arcas.apache.org (HELO arcas.apache.org) (140.211.11.28) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 30 Sep 2015 03:43:05 +0000 Date: Wed, 30 Sep 2015 03:43:05 +0000 (UTC) From: "Jonathan Shook (JIRA)" To: commits@cassandra.apache.org Message-ID: In-Reply-To: References: Subject: [jira] [Updated] (CASSANDRA-10419) Make JBOD compaction and flushing more robust MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: 7bit X-JIRA-FingerPrint: 30527f35849b9dde25b450d4833f0394 [ https://issues.apache.org/jira/browse/CASSANDRA-10419?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jonathan Shook updated CASSANDRA-10419: --------------------------------------- Description: With JBOD and several smaller disks, like SSDs at 1.2 TB or lower, it is possible to run out of space prematurely. With a sufficient ingestion rate, disk selection logic seems to overselect on certain JBOD targets. This causes a premature C* shutdown when there is a significant amount of space left. With DTCS, for example, it should be possible to utilize over 90% of the available space with certain settings. However in the scenario I tested, only about 50% was utilized, before a filesystem full error. (see below). It is likely that this is a scheduling challenge between high rates of ingest and smaller data directories. It would be good to use an anticipatory model if possible to more carefully select compaction targets according to fill rates. As well, if the largest sstable that can be supported is constrained by the largest JBOD extent, we should make that visible to the compaction logic where possible. The attached image shows a test with 12 1.2TB JBOD data directories. At the end, the utilizations are: 59GiB, 83GiB, 83GiB, 97GiB, 330GiB, 589GiB, 604GiB, 630GiB, 697GiB, 1.055TiB, 1.083TB, 1092TiB, was: With JBOD and several smaller disks, like SSDs at 1.2 TB or lower, it is possible to run out of space prematurely. With a sufficient ingestion rate, disk selection logic seems to overselect on certain JBOD targets. This causes a premature C* shutdown when there is a significant amount of space left. With DTCS, for example, it should be possible to utilize over 90% of the available space with certain settings. However in the scenario I tested, only about 50% was utilized, before a filesystem full error. (see below). It is likely that this is a scheduling challenge between high rates of ingest and smaller data directories. It would be good to use an anticipatory model if possible to more carefully select compaction targets according to fill rates. The attached image shows a test with 12 1.2TB JBOD data directories. At the end, the utilizations are: 59GiB, 83GiB, 83GiB, 97GiB, 330GiB, 589GiB, 604GiB, 630GiB, 697GiB, 1.055TiB, 1.083TB, 1092TiB, > Make JBOD compaction and flushing more robust > --------------------------------------------- > > Key: CASSANDRA-10419 > URL: https://issues.apache.org/jira/browse/CASSANDRA-10419 > Project: Cassandra > Issue Type: Improvement > Components: Core > Reporter: Jonathan Shook > Attachments: timeseries-study-overview-jbods.png > > > With JBOD and several smaller disks, like SSDs at 1.2 TB or lower, it is possible to run out of space prematurely. With a sufficient ingestion rate, disk selection logic seems to overselect on certain JBOD targets. This causes a premature C* shutdown when there is a significant amount of space left. With DTCS, for example, it should be possible to utilize over 90% of the available space with certain settings. However in the scenario I tested, only about 50% was utilized, before a filesystem full error. (see below). It is likely that this is a scheduling challenge between high rates of ingest and smaller data directories. It would be good to use an anticipatory model if possible to more carefully select compaction targets according to fill rates. As well, if the largest sstable that can be supported is constrained by the largest JBOD extent, we should make that visible to the compaction logic where possible. > The attached image shows a test with 12 1.2TB JBOD data directories. At the end, the utilizations are: > 59GiB, 83GiB, 83GiB, 97GiB, 330GiB, 589GiB, 604GiB, 630GiB, 697GiB, 1.055TiB, 1.083TB, 1092TiB, -- This message was sent by Atlassian JIRA (v6.3.4#6332)