Return-Path: X-Original-To: archive-asf-public-internal@cust-asf2.ponee.io Delivered-To: archive-asf-public-internal@cust-asf2.ponee.io Received: from cust-asf.ponee.io (cust-asf.ponee.io [163.172.22.183]) by cust-asf2.ponee.io (Postfix) with ESMTP id 0589A200D11 for ; Mon, 2 Oct 2017 22:13:08 +0200 (CEST) Received: by cust-asf.ponee.io (Postfix) id 03DAB1609EF; Mon, 2 Oct 2017 20:13:08 +0000 (UTC) Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by cust-asf.ponee.io (Postfix) with SMTP id 22E0C1609C0 for ; Mon, 2 Oct 2017 22:13:06 +0200 (CEST) Received: (qmail 50366 invoked by uid 500); 2 Oct 2017 20:13:06 -0000 Mailing-List: contact dev-help@flink.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: dev@flink.apache.org Delivered-To: mailing list dev@flink.apache.org Received: (qmail 50351 invoked by uid 99); 2 Oct 2017 20:13:05 -0000 Received: from pnap-us-west-generic-nat.apache.org (HELO spamd3-us-west.apache.org) (209.188.14.142) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 02 Oct 2017 20:13:05 +0000 Received: from localhost (localhost [127.0.0.1]) by spamd3-us-west.apache.org (ASF Mail Server at spamd3-us-west.apache.org) with ESMTP id 12BDC199F10 for ; Mon, 2 Oct 2017 20:13:05 +0000 (UTC) X-Virus-Scanned: Debian amavisd-new at spamd3-us-west.apache.org X-Spam-Flag: NO X-Spam-Score: 2.379 X-Spam-Level: ** X-Spam-Status: No, score=2.379 tagged_above=-999 required=6.31 tests=[DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, HTML_MESSAGE=2, RCVD_IN_DNSWL_NONE=-0.0001, RCVD_IN_MSPIKE_H3=-0.01, RCVD_IN_MSPIKE_WL=-0.01, RCVD_IN_SORBS_SPAM=0.5, SPF_PASS=-0.001] autolearn=disabled Authentication-Results: spamd3-us-west.apache.org (amavisd-new); dkim=pass (2048-bit key) header.d=gmail.com Received: from mx1-lw-eu.apache.org ([10.40.0.8]) by localhost (spamd3-us-west.apache.org [10.40.0.10]) (amavisd-new, port 10024) with ESMTP id 4-OcQ8lsPcWt for ; Mon, 2 Oct 2017 20:13:02 +0000 (UTC) Received: from mail-vk0-f45.google.com (mail-vk0-f45.google.com [209.85.213.45]) by mx1-lw-eu.apache.org (ASF Mail Server at mx1-lw-eu.apache.org) with ESMTPS id 10A565FB40 for ; Mon, 2 Oct 2017 20:13:02 +0000 (UTC) Received: by mail-vk0-f45.google.com with SMTP id w23so3511128vkw.2 for ; Mon, 02 Oct 2017 13:13:02 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=mime-version:in-reply-to:references:from:date:message-id:subject:to; bh=sqHbnTyR8GmjYiuN09yn2S1xNoNYxoeqFKQF/VaGXoQ=; b=kTS7raxKoYkdDZoj1eSs+4LKmg29MAX7llC5k1AgE0PX9K+pdgo9rgdBkeiMGeM2hS e/xQfnmNYpAF8p7WRfVPYSatgVbbxi5xXya2B7VPgJh7LozAgsvT1GY/xBp+3R5RPnNw c4C1bTgF+b8U+hfM58jsiYDvuJ0chrOLVt63a4Qk36C8mSdLEJk7ryxwpuImShLHAbmc ehqlfgkXuJHsZLo4ixZQH/qHWVt6904xnajrYziEGTq4icfJ0Ek14mHOVMeP8ALSsicR E/CoRfpvv6zzxT3RmtEFV0OQXezJBBpoZ8gYsiuXeR/EC7KKUA6E7Pg8v9SI3uRo5IHm T71g== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:in-reply-to:references:from:date :message-id:subject:to; bh=sqHbnTyR8GmjYiuN09yn2S1xNoNYxoeqFKQF/VaGXoQ=; b=PuJZpMD9X9omYwlrA/DWoaZYO1oKcLj/GZ+kgmmVp6DE1A/jaZpAjFfnDPISYiJCOB gatsX0xCwWfHJAnIWYhI/7FRVTlHlEmMMzrtmFlT1B/sk3QjDRW/1FpV35ySKCfQjn6L 7fjF34/ZDdR2aFWrvDlh/0vxl0d/vHta3tvIB7sXBB4ufWyfb+2E8c3hfeBz28Wf70m9 SolqwiRJcHm/hX40aiwjWxYyUMAZXkN0rc2gUlKIYKdGua1FboUwOvysKLwk/sfBITLQ nQB74N3ETn2XVsJY79z8VMGkZ7x3NaO4SIezz+BbS1AphmZa9I50/QPxnkbmflIiY05L 1AlA== X-Gm-Message-State: AMCzsaV2+JiKSg7D1TGKzYMnbvrgXq1emo61ndxOGU1X/oxKqIN5fFEi u1rSSYe+MeO9D7NFBvfbHqa4jlUbLPHge8sufZG+YZGd X-Google-Smtp-Source: AOwi7QDDNwgEi4I4m0obPwutLiIslfJo0c4vi+gPXl2FsGxSLf3uSa0zaR0sYUmgzj0ShhBRpQ6bShjkV1Qb373ZZiQ= X-Received: by 10.31.159.65 with SMTP id i62mr2038294vke.157.1506975180741; Mon, 02 Oct 2017 13:13:00 -0700 (PDT) MIME-Version: 1.0 Received: by 10.103.12.69 with HTTP; Mon, 2 Oct 2017 13:12:20 -0700 (PDT) In-Reply-To: References: From: Fabian Hueske Date: Mon, 2 Oct 2017 22:12:20 +0200 Message-ID: Subject: Re: Unable to write snapshots to S3 on EMR To: "dev@flink.apache.org" Content-Type: multipart/alternative; boundary="001a1141b33055649d055a9600e0" archived-at: Mon, 02 Oct 2017 20:13:08 -0000 --001a1141b33055649d055a9600e0 Content-Type: text/plain; charset="UTF-8" Content-Transfer-Encoding: quoted-printable Hi Andy, I'm not an AWS expert, so I'll just check on some common issues. I guess you already had a look at the Flink docs for AWS/EMR but I'll post the link just be to sure [1]. Since you are using Flink 1.3.2 (EMR 5.8.0 comes with Flink 1.3.1) did you built Flink yourself or did you download the binaries? Does the Hadoop version of the Flink build match the Hadoop version of EMR 5.8.0, i.e., Hadoop 2.7.x? Best, Fabian [1] https://ci.apache.org/projects/flink/flink-docs-release-1.3/setup/aws.html 2017-10-02 21:51 GMT+02:00 Andy M. : > Hi Fabian, > > Sorry, I just realized I forgot to include that part. The error returned > is: > > java.lang.NoSuchMethodError: > org.apache.hadoop.conf.Configuration.addResource(Lorg/apache/hadoop/conf/ > Configuration;)V > at com.amazon.ws.emr.hadoop.fs.EmrFileSystem.initialize( > EmrFileSystem.java:93) > at org.apache.flink.runtime.fs.hdfs.HadoopFileSystem. > initialize(HadoopFileSystem.java:328) > at org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem( > FileSystem.java:350) > at org.apache.flink.core.fs.FileSystem.get(FileSystem.java:389) > at org.apache.flink.core.fs.Path.getFileSystem(Path.java:293) > at org.apache.flink.runtime.state.filesystem. > FsCheckpointStreamFactory.(FsCheckpointStreamFactory.java:99) > at org.apache.flink.runtime.state.filesystem.FsStateBackend. > createStreamFactory(FsStateBackend.java:282) > at org.apache.flink.contrib.streaming.state.RocksDBStateBackend. > createStreamFactory(RocksDBStateBackend.java:273 > > I believe it has something to do with the classpath, but I am unsure why = or > how to fix it. The classpath being used during the execution is: > /home/hadoop/flink-1.3.2/lib/flink-python_2.11-1.3.2.jar:/ > ho=E2=80=8C=E2=80=8Bme/hadoop/flink-1.3.=E2=80=8C=E2=80=8B2/lib/flink-sha= ded-h=E2=80=8C=E2=80=8Badoop2- > uber-1.3.2.ja=E2=80=8C=E2=80=8Br:/home/hadoop/flink=E2=80=8C=E2=80=8B-1.3= .2/lib/log4j-1.2=E2=80=8C=E2=80=8B. > 17.jar:/home/hadoop=E2=80=8C=E2=80=8B/flink-1.3.2/lib/slf=E2=80=8C=E2=80= =8B4j-log4j12-1.7.7. > jar=E2=80=8C=E2=80=8B:/home/hadoop/flink-=E2=80=8C=E2=80=8B1.3.2/lib/flin= k-dist=E2=80=8C=E2=80=8B_2.11-1.3. > 2.jar::/et=E2=80=8C=E2=80=8Bc/hadoop/conf: > > I decompiled flink-shaded-h=E2=80=8C=E2=80=8Badoop2-uber-1.3.2.ja=E2=80= =8C=E2=80=8Br and it seems the > addResource function does seem to be there. > > Thank you > > On Mon, Oct 2, 2017 at 3:43 PM, Fabian Hueske wrote: > > > Hi Andy, > > > > can you describe in more detail what exactly isn't working? > > Do you see error messages in the log files or on the console? > > > > Thanks, Fabian > > > > 2017-10-02 15:52 GMT+02:00 Andy M. : > > > > > Hello, > > > > > > I am about to deploy my first Flink projects to production, but I am > > > running into a very big hurdle. I am unable to launch my project so = it > > can > > > write to an S3 bucket. My project is running on an EMR cluster, wher= e > I > > > have installed Flink 1.3.2. I am using Yarn to launch the applicatio= n, > > and > > > it seems to run fine unless I am trying to enable check pointing(with= a > > S3 > > > target). I am looking to use RocksDB as my check-pointing backend. = I > > have > > > asked a few places, and I am still unable to find a solution to this > > > problem. Here are my steps for creating a cluster, and launching my > > > application, perhaps I am missing a step. I'd be happy to provide an= y > > > additional information if needed. > > > > > > AWS Portal: > > > > > > 1) EMR -> Create Cluster > > > 2) Advanced Options > > > 3) Release =3D emr-5.8.0 > > > 4) Only select Hadoop 2.7.3 > > > 5) Next -> Next -> Next -> Create Cluster ( I do fill out > > > names/keys/etc) > > > > > > Once the cluster is up I ssh into the Master and do the following: > > > > > > 1 wget > > > http://apache.claz.org/flink/flink-1.3.2/flink-1.3.2-bin- > > > hadoop27-scala_2.11.tgz > > > 2 tar -xzf flink-1.3.2-bin-hadoop27-scala_2.11.tgz > > > 3 cd flink-1.3.2 > > > 4 ./bin/yarn-session.sh -n 2 -tm 5120 -s 4 -d > > > 5 Change conf/flink-conf.yaml > > > 6 ./bin/flink run -m yarn-cluster -yn 1 ~/flink-consumer.jar > > > > > > My conf/flink-conf.yaml I add the following fields: > > > > > > state.backend: rocksdb > > > state.backend.fs.checkpointdir: s3:/bucket/location > > > state.checkpoints.dir: s3:/bucket/location > > > > > > My program's checkpointing setup: > > > > > > > > > env.enableCheckpointing(getCheckpointRate,CheckpointingMode.EXACTLY_ > > ONCE) > > > > > > env.getCheckpointConfig.enableExternalizedCheckpoints( > > > ExternalizedCheckpointCleanup.RETAIN_ON_CANCELLATION) > > > > > > env.getCheckpointConfig.setMinPauseBetweenCheckpoints( > > > getCheckpointMinPause) > > > env.getCheckpointConfig.setCheckpointTimeout(getCheckpointTimeout= ) > > > env.getCheckpointConfig.setMaxConcurrentCheckpoints(1) > > > env.setStateBackend(new RocksDBStateBackend("s3:// > bucket/location", > > > true)) > > > > > > --001a1141b33055649d055a9600e0--