Return-Path: X-Original-To: apmail-hbase-dev-archive@www.apache.org Delivered-To: apmail-hbase-dev-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id AF3EAEA5 for ; Thu, 21 Jun 2012 19:36:40 +0000 (UTC) Received: (qmail 55907 invoked by uid 500); 21 Jun 2012 19:36:40 -0000 Delivered-To: apmail-hbase-dev-archive@hbase.apache.org Received: (qmail 55853 invoked by uid 500); 21 Jun 2012 19:36:40 -0000 Mailing-List: contact dev-help@hbase.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: dev@hbase.apache.org Delivered-To: mailing list dev@hbase.apache.org Received: (qmail 55841 invoked by uid 99); 21 Jun 2012 19:36:40 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Thu, 21 Jun 2012 19:36:40 +0000 X-ASF-Spam-Status: No, hits=1.5 required=5.0 tests=HTML_MESSAGE,RCVD_IN_DNSWL_LOW X-Spam-Check-By: apache.org Received-SPF: unknown ~alla (nike.apache.org: encountered unrecognized mechanism during SPF processing of domain of mcorgan@hotpads.com) Received: from [209.85.213.169] (HELO mail-yx0-f169.google.com) (209.85.213.169) by apache.org (qpsmtpd/0.29) with ESMTP; Thu, 21 Jun 2012 19:36:34 +0000 Received: by yenr5 with SMTP id r5so1096357yen.14 for ; Thu, 21 Jun 2012 12:36:13 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20120113; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :content-type:x-gm-message-state; bh=f0srp4jctBlRRY8GvHPlqfhnZEGieF2mr68zCbEiNj0=; b=k+9OjH/dPalCaygXQOrCls5BpSag/9rYAO8Ps0PacctP9i/5HXj7Q7jzZ9HyrQckuB bL2BfNelRC7Kb8YKHE2kCYfgYKHU+4AiNL5OQUsVQzELcIFQ3KvdhX5tAJ7HroMvF34I dJLxf4eFOH1tuTUpPfin2PcvbRLwSpbx3blS69HgYnXw9Vs0JleLL99rstRwvkr3QmXR Qhp7RmnktifU+NvG2PCh6noQTxdyWGPak7hHOCXmCcSv5nHTyY2l+MNrlEQbp8Ye8tF0 3ykNVZFDqO7EoUKYGvtuf/Hnwi+CSCU8mAF9GIo2Fw6HtyZdaxsDo+H7f0Z8Dn1RNGn4 ZwHw== MIME-Version: 1.0 Received: by 10.60.26.134 with SMTP id l6mr7192125oeg.40.1340307373097; Thu, 21 Jun 2012 12:36:13 -0700 (PDT) Received: by 10.182.143.36 with HTTP; Thu, 21 Jun 2012 12:36:13 -0700 (PDT) In-Reply-To: References: Date: Thu, 21 Jun 2012 12:36:13 -0700 Message-ID: Subject: Re: Performance Testing From: Matt Corgan To: dev@hbase.apache.org Content-Type: multipart/alternative; boundary=e89a8ff1c63cdd028f04c300a0e5 X-Gm-Message-State: ALoCoQlLGkQxasjvhU0ZROg39Ak/ePpYms6fisD6KLLuIcSA5RD4nna81DnDNVvwbiIgKXXS9DA3 --e89a8ff1c63cdd028f04c300a0e5 Content-Type: text/plain; charset=UTF-8 These are geared more towards development than regression testing, but here are a few ideas that I would find useful: * Ability to run the performance tests (or at least a subset of them) on a development machine would help people avoid committing regressions and would speed development in general * Ability to test a single region without heavier weight servers and clusters * Letting the test run with multiple combinations of input parameters (block size, compression, blooms, encoding, flush size, etc, etc). Possibly many combinations that could take a while to run * Output results to a CSV file that's importable to a spreadsheet for sorting/filtering/charting. * Email the CSV file to the user notifying them the tests have finished. * Getting fancier: ability to specify a list of branches or tags from git or subversion as inputs, which would allow the developer to tag many different performance changes and later figure out which combination is the best (all before submitting a patch) On Thu, Jun 21, 2012 at 12:13 PM, Elliott Clark wrote: > I actually think that more measurements are needed than just per release. > The best I could hope for would be a four node+ cluster(One master and > three slaves) that for every check in on trunk run multiple different perf > tests. > > > - All Reads (Scans) > - Large Writes (Should test compactions/flushes) > - Read Dominated with 10% writes > > Then every checkin can be evaluated and large regressions can be treated as > bugs. And with that we can see the difference between the different > versions as well. http://arewefastyet.com/ is kind of the model that I > would love to see. And I'm more than willing to help where ever needed. > > However in reality every night will probably be more feasible. And Four > nodes is probably not going to happen either. > > On Thu, Jun 21, 2012 at 11:38 AM, Andrew Purtell >wrote: > > > On Wed, Jun 20, 2012 at 10:37 PM, Ryan Ausanka-Crues > > wrote: > > > I think it makes sense to start by defining the goals for the > > performance testing project and then deciding what we'd like to > accomplish. > > As such, I start by soliciting ideas from everyone on what they would > like > > to see from the project. We can then collate those thoughts and > prioritize > > the different features. Does that sound like a reasonable approach? > > > > In terms of defining a goal, the fundamental need I see for us as a > > project is to quantify performance from one release to the next, thus > > be able to avoid regressions by noting adverse changes in release > > candidates. > > > > In terms of defining what "performance" means... well, that's an > > involved and separate discussion I think. > > > > Best regards, > > > > - Andy > > > > Problems worthy of attack prove their worth by hitting back. - Piet > > Hein (via Tom White) > > > --e89a8ff1c63cdd028f04c300a0e5--