Return-Path: X-Original-To: archive-asf-public-internal@cust-asf2.ponee.io Delivered-To: archive-asf-public-internal@cust-asf2.ponee.io Received: from cust-asf.ponee.io (cust-asf.ponee.io [163.172.22.183]) by cust-asf2.ponee.io (Postfix) with ESMTP id 08B31200BA8 for ; Mon, 24 Oct 2016 19:06:33 +0200 (CEST) Received: by cust-asf.ponee.io (Postfix) id 074C8160AEB; Mon, 24 Oct 2016 17:06:33 +0000 (UTC) Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by cust-asf.ponee.io (Postfix) with SMTP id 4EB3B160AD7 for ; Mon, 24 Oct 2016 19:06:32 +0200 (CEST) Received: (qmail 65854 invoked by uid 500); 24 Oct 2016 17:06:26 -0000 Mailing-List: contact user-help@flink.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@flink.apache.org Delivered-To: mailing list user@flink.apache.org Received: (qmail 65844 invoked by uid 99); 24 Oct 2016 17:06:26 -0000 Received: from pnap-us-west-generic-nat.apache.org (HELO spamd3-us-west.apache.org) (209.188.14.142) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 24 Oct 2016 17:06:26 +0000 Received: from localhost (localhost [127.0.0.1]) by spamd3-us-west.apache.org (ASF Mail Server at spamd3-us-west.apache.org) with ESMTP id EFED0180BD7 for ; Mon, 24 Oct 2016 17:06:25 +0000 (UTC) X-Virus-Scanned: Debian amavisd-new at spamd3-us-west.apache.org X-Spam-Flag: NO X-Spam-Score: 2.629 X-Spam-Level: ** X-Spam-Status: No, score=2.629 tagged_above=-999 required=6.31 tests=[DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, FREEMAIL_ENVFROM_END_DIGIT=0.25, HTML_MESSAGE=2, RCVD_IN_DNSWL_NONE=-0.0001, RCVD_IN_MSPIKE_H3=-0.01, RCVD_IN_MSPIKE_WL=-0.01, RCVD_IN_SORBS_SPAM=0.5, SPF_PASS=-0.001] autolearn=disabled Authentication-Results: spamd3-us-west.apache.org (amavisd-new); dkim=pass (2048-bit key) header.d=gmail.com Received: from mx1-lw-eu.apache.org ([10.40.0.8]) by localhost (spamd3-us-west.apache.org [10.40.0.10]) (amavisd-new, port 10024) with ESMTP id 0qabN4_F35I6 for ; Mon, 24 Oct 2016 17:06:24 +0000 (UTC) Received: from mail-wm0-f47.google.com (mail-wm0-f47.google.com [74.125.82.47]) by mx1-lw-eu.apache.org (ASF Mail Server at mx1-lw-eu.apache.org) with ESMTPS id AAFB55FC3D for ; Mon, 24 Oct 2016 17:06:23 +0000 (UTC) Received: by mail-wm0-f47.google.com with SMTP id c78so114846211wme.0 for ; Mon, 24 Oct 2016 10:06:23 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=mime-version:from:date:message-id:subject:to; bh=mp7RwboN9iYHaFkuotdz2IpxFg8huR1rluddyzayXqQ=; b=zo8/lE0TPtIIrLga4vkVBeFeOUhIRdEMCJo425E35ajgx5elMNE1BqW/h68I/L/L9B tnk0G4OZrYrUBA6DesY32mJdIxS/XwHg/IdEV70nf/ir+rG02Q8rC6Iq8LBJuc9i8gwM ZN7HePU3SB59aNc6swyM/2t3klcbxvh+pbKZVJn7eF0yozl2U455isZWLdR7oPWbjoDT x/WY8KjTYD/7AGwH5Q5H43KZHdNgNISoIfENU0vKmPVmMnTF+O0YYCpu2fFED6Zlot6V 7YrRUGSt4nPrEnazh5Z0LuyM0ZOLsRpfwZFsb4LU7ACoay93xL9M0LifOyitDFh7d8+W ziZQ== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:mime-version:from:date:message-id:subject:to; bh=mp7RwboN9iYHaFkuotdz2IpxFg8huR1rluddyzayXqQ=; b=Iypb8oAJ/eF1rNDvVtTHDgYMYZ2kCQNvdoob01SFYXZ0wthUUWKFHpOYbGzbQReDJM NbzLT70n32n+s99RC/I17VaJqsNlc/FMcgDnlZh+INEozmKSbZVKBlgxeg8Icrf2qvwm 2g49hLDCPlBLCTQXD6D3dGmbV6fZiGFsZT46YfyB5sbCeqbGy0a/9QMvLR5cVsOn8EcR VLE2LPcITTLwPeKAlYTDEh5ZV1LpNJYTw1Jw+eoxBcL66JFo6k7y9TTc+gwiwMdRGPqi rR40ZOkKWIZI1vV8/TQb0k9tomieu/YQrykU5V+PJxt4/PJWO5l2giJ9jipl/V4E8KnS iejg== X-Gm-Message-State: AA6/9RlRrMbQfydVwo2Fqhr7Pu1owEmnPmQUFXLO7QQvISOFKnEOw7ILw2SFWExqec5DE62c5S0tbtKswT5xlQ== X-Received: by 10.28.193.65 with SMTP id r62mr22628274wmf.12.1477328782825; Mon, 24 Oct 2016 10:06:22 -0700 (PDT) MIME-Version: 1.0 Received: by 10.80.135.80 with HTTP; Mon, 24 Oct 2016 10:06:22 -0700 (PDT) From: Josh Date: Mon, 24 Oct 2016 18:06:22 +0100 Message-ID: Subject: Checkpointing large RocksDB state to S3 - tips? To: user@flink.apache.org Content-Type: multipart/alternative; boundary=94eb2c0d451a511f37053f9f6905 archived-at: Mon, 24 Oct 2016 17:06:33 -0000 --94eb2c0d451a511f37053f9f6905 Content-Type: text/plain; charset=UTF-8 Hi all, I'm running Flink on EMR/YARN with 2x m3.xlarge instances and am checkpointing a fairly large RocksDB state to S3. I've found that when the state size hits 10GB, the checkpoint takes around 6 minutes, according to the Flink dashboard. Originally my checkpoint interval was 5 minutes for the job, but I've found that the YARN container crashes (I guess because the checkpoint time is greater than the checkpoint interval), so have now decreased the checkpoint frequency to every 10 minutes. I was just wondering if anyone has any tips about how to reduce the checkpoint time. Taking 6 minutes to checkpoint ~10GB state means it's uploading at ~30MB/sec. I believe the m3.xlarge instances should have around 125MB/sec network bandwidth each, so I think the bottleneck is S3. Since there are 2 instances, I'm not sure if that means each instance is uploading at 15MB/sec - do the state uploads get shared equally among the instances, assuming the state is split equally between the task managers? If the state upload is split between the instances, perhaps the only way to speed up the checkpoints is to add more instances and task managers, and split the state equally among the task managers? Also just wondering - is there any chance the incremental checkpoints work will be complete any time soon? Thanks, Josh --94eb2c0d451a511f37053f9f6905 Content-Type: text/html; charset=UTF-8 Content-Transfer-Encoding: quoted-printable
Hi all,

I'm running Flink on EMR/YA= RN with 2x m3.xlarge instances and am checkpointing a fairly large RocksDB = state to S3.

I've found that when the state si= ze hits 10GB, the checkpoint takes around 6 minutes, according to the Flink= dashboard. Originally my checkpoint interval was 5 minutes for the job, bu= t I've found that the YARN container crashes (I guess because the check= point time is greater than the checkpoint interval), so have now decreased = the checkpoint frequency to every 10 minutes.

I wa= s just wondering if anyone has any tips about how to reduce the checkpoint = time. Taking 6 minutes to checkpoint ~10GB state means it's uploading a= t ~30MB/sec. I believe the m3.xlarge instances should have around 125MB/sec= network bandwidth each, so I think the bottleneck is S3.=C2=A0
S= ince there are 2 instances, I'm not sure if that means each instance is= uploading at 15MB/sec - do the state uploads get shared equally among the = instances, assuming the state is split equally between the task managers?

If the state upload is split between the instan= ces, perhaps the only way to speed up the checkpoints is to add more instan= ces and task managers, and split the state equally among the task managers?=

Also just wondering - is there any chance the= incremental checkpoints work will be complete any time soon?

Thanks,
Josh
--94eb2c0d451a511f37053f9f6905--