Return-Path: X-Original-To: archive-asf-public-internal@cust-asf2.ponee.io Delivered-To: archive-asf-public-internal@cust-asf2.ponee.io Received: from cust-asf.ponee.io (cust-asf.ponee.io [163.172.22.183]) by cust-asf2.ponee.io (Postfix) with ESMTP id 41998200BC8 for ; Wed, 23 Nov 2016 18:41:00 +0100 (CET) Received: by cust-asf.ponee.io (Postfix) id 40139160AFD; Wed, 23 Nov 2016 17:41:00 +0000 (UTC) Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by cust-asf.ponee.io (Postfix) with SMTP id 64359160AEC for ; Wed, 23 Nov 2016 18:40:59 +0100 (CET) Received: (qmail 42200 invoked by uid 500); 23 Nov 2016 17:40:58 -0000 Mailing-List: contact user-help@flink.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@flink.apache.org Delivered-To: mailing list user@flink.apache.org Received: (qmail 42190 invoked by uid 99); 23 Nov 2016 17:40:58 -0000 Received: from pnap-us-west-generic-nat.apache.org (HELO spamd4-us-west.apache.org) (209.188.14.142) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 23 Nov 2016 17:40:58 +0000 Received: from localhost (localhost [127.0.0.1]) by spamd4-us-west.apache.org (ASF Mail Server at spamd4-us-west.apache.org) with ESMTP id 03E1BC04C2 for ; Wed, 23 Nov 2016 17:40:58 +0000 (UTC) X-Virus-Scanned: Debian amavisd-new at spamd4-us-west.apache.org X-Spam-Flag: NO X-Spam-Score: 2.38 X-Spam-Level: ** X-Spam-Status: No, score=2.38 tagged_above=-999 required=6.31 tests=[DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, HTML_MESSAGE=2, RCVD_IN_DNSWL_NONE=-0.0001, RCVD_IN_MSPIKE_H3=-0.01, RCVD_IN_MSPIKE_WL=-0.01, RCVD_IN_SORBS_SPAM=0.5, SPF_PASS=-0.001, URIBL_BLOCKED=0.001] autolearn=disabled Authentication-Results: spamd4-us-west.apache.org (amavisd-new); dkim=pass (2048-bit key) header.d=gmail.com Received: from mx1-lw-eu.apache.org ([10.40.0.8]) by localhost (spamd4-us-west.apache.org [10.40.0.11]) (amavisd-new, port 10024) with ESMTP id wL03zDsy_xun for ; Wed, 23 Nov 2016 17:40:56 +0000 (UTC) Received: from mail-yw0-f172.google.com (mail-yw0-f172.google.com [209.85.161.172]) by mx1-lw-eu.apache.org (ASF Mail Server at mx1-lw-eu.apache.org) with ESMTPS id 153FE5FAD8 for ; Wed, 23 Nov 2016 17:40:56 +0000 (UTC) Received: by mail-yw0-f172.google.com with SMTP id r204so17888763ywb.0 for ; Wed, 23 Nov 2016 09:40:56 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=mime-version:in-reply-to:references:from:date:message-id:subject:to; bh=dhCa4cO+DdbCM35X4JLuAAHYw0yCSiCOYyE6XpaOLTk=; b=T7+l1FO9XXHbU1zkoIQgafgWWXdx/Co0jxEu9thiGu0bwkpaPBN5K4tYq9y/LHZGio yBH5seP4MzRl9MeX4wAyBDxidpVUoWMxKgMF0sEQ+96ZH5YCc8eTwVrd5nw+svRgHjwo 8YQBbmxBMEvk5GZSNFo84dycFz7f0tCewC7UzkZnzf7O8ETzCTkpjR1BlQ2rG7++2dnf yuHd/6AL5OuScZLlAARXa5GM5TJl+35/Bnydc5CSapFt3Qu1u12WO/Pz6X/qPtjKxUGM +XpA6xmnlJH0jdvlur1Onb9yLAAv956nRn8lhRkuioCgdvvZKWuUQazGDTYpTz+6JIbo rJcQ== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:mime-version:in-reply-to:references:from:date :message-id:subject:to; bh=dhCa4cO+DdbCM35X4JLuAAHYw0yCSiCOYyE6XpaOLTk=; b=iwNpRwIQKiTGJfCjnB7B5SqC8fSJJMkkP19FtuBRFupVRrg0W0WMwjWHUYRHbREkhj lR19xO1IwbdCx/T5jjOTX4MnRncWPY4Psqz/pYThcMPzmn5k2KY4J+gaZJ86DDsh0Y6K fc6CtFRP37gQtlocb0eKLINUhPDcq/nZgQoBxU7eai94S0D8KO/I6na2EPu3M7Uw/fqq u3FalgE3qTyTuvX4aWZBNDMdWe75jjjY2O6lvkgAoUPCo5qYo7xqqR7ATVjUlsXgTHrV EAOqjWsYypKbPgD9iwNuXg9IQ0ENUJ9fpqUIEq4bx1rUOBPaI95qAtFkEEHokh1k2p50 a16Q== X-Gm-Message-State: AKaTC00uvqDNkbGn4FzsAmP5mn0WrWvsqfYu6BgpN/bsTJnv01dIL8HpqeWfVlLF+WDEWz2uQdMXEs2Zmm1DRQ== X-Received: by 10.129.98.193 with SMTP id w184mr4828925ywb.195.1479922853810; Wed, 23 Nov 2016 09:40:53 -0800 (PST) MIME-Version: 1.0 Received: by 10.37.37.69 with HTTP; Wed, 23 Nov 2016 09:40:53 -0800 (PST) In-Reply-To: References: From: Jonathan Share Date: Wed, 23 Nov 2016 18:40:53 +0100 Message-ID: Subject: Re: S3 checkpointing in AWS in Frankfurt To: user@flink.apache.org Content-Type: multipart/alternative; boundary=001a1146ff48ff39540541fb63c2 archived-at: Wed, 23 Nov 2016 17:41:00 -0000 --001a1146ff48ff39540541fb63c2 Content-Type: text/plain; charset=UTF-8 Hi Greg, Standard storage class, everything is on defaults, we've not done anything special with the bucket. Cloud Watch only appears to give me total billing for S3 in general, I don't see a breakdown unless that's something I can configure somewhere. Regards, Jonathan On 23 November 2016 at 16:29, Greg Hogan wrote: > Hi Jonathan, > > Which S3 storage class are you using? Do you have a breakdown of the S3 > costs as storage / API calls / early deletes / data transfer? > > Greg > > On Wed, Nov 23, 2016 at 2:52 AM, Jonathan Share > wrote: > >> Hi, >> >> I'm interested in hearing if anyone else has experience with using Amazon >> S3 as a state backend in the Frankfurt region. For political reasons we've >> been asked to keep all European data in Amazon's Frankfurt region. This >> causes a problem as the S3 endpoint in Frankfurt requires the use of AWS >> Signature Version 4 "This new Region supports only Signature Version 4" >> [1] and this doesn't appear to work with the Hadoop version that Flink is >> built against [2]. >> >> After some hacking we have managed to create a docker image with a build >> of Flink 1.2 master, copying over jar files from the hadoop >> 3.0.0-alpha1 package and this appears to work, for the most part but we >> still suffer from some classpath problems (conflicts between AWS API used >> in hadoop and those we want to use in out streams for interacting with >> Kinesis) and the whole thing feels a little fragile. Has anyone else tried >> this? Is there a simpler solution? >> >> As a follow-up question, we saw that with checkpointing on three >> relatively simple streams set to 1 second, our S3 costs were higher than >> the EC2 costs for our entire infrastructure. This seems slightly >> disproportionate. For now we have reduced checkpointing interval to 10 >> seconds and that has greatly improved the cost projections graphed via >> Amazon Cloud Watch, but I'm interested in hearing other peoples experience >> with this. Is that the kind of billing level we can expect or is this a >> symptom of a mis-configuration? Is this a setup others are using? As we are >> using Kinesis as the source for all streams I don't see a huge risk with >> larger checkpoint intervals and our Sinks are designed to mostly tolerate >> duplicates (some improvements can be made). >> >> Thanks in advance >> Jonathan >> >> >> [1] https://aws.amazon.com/blogs/aws/aws-region-germany/ >> [2] https://issues.apache.org/jira/browse/HADOOP-13324 >> > > --001a1146ff48ff39540541fb63c2 Content-Type: text/html; charset=UTF-8 Content-Transfer-Encoding: quoted-printable
Hi Greg,

Standard storage class, everyt= hing is on defaults, we've not done anything special with the bucket.

Cloud Watch only appears to give me total billing f= or S3 in general, I don't see a breakdown unless that's something I= can configure somewhere.

Regards,
Jonat= han


On 23 November 2016 at 16:29, Greg Hogan <= ;code@greghogan.com= > wrote:
<= div>Hi Jonathan,

Which S3 storage class are you using? Do= you have a breakdown of the S3 costs as storage / API calls / early delete= s / data transfer?

Greg

On Wed, Nov 23, 2016 at= 2:52 AM, Jonathan Share <jon.share@gmail.com> wrote:
<= blockquote class=3D"gmail_quote" style=3D"margin:0 0 0 .8ex;border-left:1px= #ccc solid;padding-left:1ex">
Hi,

I'= ;m interested in hearing if anyone else has experience with using Amazon S3= as a state backend in the Frankfurt region. For political reasons we'v= e been asked to keep all European data in Amazon's Frankfurt region. Th= is causes a problem as the S3 endpoint in Frankfurt requires the use of AWS= Signature Version 4 "This new = Region supports only Signature Version 4" [1] and this doesn= 9;t appear to work with the Hadoop version that Flink is built against [2].=

After some hacking we have managed to create a do= cker image with a build of Flink 1.2 master, copying over jar files from th= e hadoop 3.0.0-alpha1=C2=A0package and this appears to work, for the most p= art but we still suffer from some classpath problems (conflicts between AWS= API used in hadoop and those we want to use in out streams for interacting= with Kinesis) and the whole thing feels a little fragile. Has anyone else = tried this? Is there a simpler solution?

As a foll= ow-up question, we saw that with checkpointing on three relatively simple s= treams set to 1 second, our S3 costs were higher than the EC2 costs for our= entire infrastructure. This seems slightly disproportionate. For now we ha= ve reduced checkpointing interval to 10 seconds and that has greatly improv= ed the cost projections graphed via Amazon Cloud Watch, but I'm interes= ted in hearing other peoples experience with this. Is that the kind of bill= ing level we can expect or is this a symptom of a mis-configuration? Is thi= s a setup others are using? As we are using Kinesis as the source for all s= treams I don't see a huge risk with larger checkpoint intervals and our= Sinks are designed to mostly tolerate duplicates (some improvements can be= made).=C2=A0

Thanks in advance
Jonathan=




--001a1146ff48ff39540541fb63c2--