Return-Path: X-Original-To: archive-asf-public-internal@cust-asf2.ponee.io Delivered-To: archive-asf-public-internal@cust-asf2.ponee.io Received: from cust-asf.ponee.io (cust-asf.ponee.io [163.172.22.183]) by cust-asf2.ponee.io (Postfix) with ESMTP id 891A4200B58 for ; Wed, 27 Jul 2016 16:34:14 +0200 (CEST) Received: by cust-asf.ponee.io (Postfix) id 87FB7160A90; Wed, 27 Jul 2016 14:34:14 +0000 (UTC) Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by cust-asf.ponee.io (Postfix) with SMTP id B2077160A6F for ; Wed, 27 Jul 2016 16:34:13 +0200 (CEST) Received: (qmail 2890 invoked by uid 500); 27 Jul 2016 14:34:12 -0000 Mailing-List: contact issues-help@flink.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: dev@flink.apache.org Delivered-To: mailing list issues@flink.apache.org Received: (qmail 2881 invoked by uid 99); 27 Jul 2016 14:34:12 -0000 Received: from pnap-us-west-generic-nat.apache.org (HELO spamd2-us-west.apache.org) (209.188.14.142) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 27 Jul 2016 14:34:12 +0000 Received: from localhost (localhost [127.0.0.1]) by spamd2-us-west.apache.org (ASF Mail Server at spamd2-us-west.apache.org) with ESMTP id 810F71A53F6 for ; Wed, 27 Jul 2016 14:34:12 +0000 (UTC) X-Virus-Scanned: Debian amavisd-new at spamd2-us-west.apache.org X-Spam-Flag: NO X-Spam-Score: -4.646 X-Spam-Level: X-Spam-Status: No, score=-4.646 tagged_above=-999 required=6.31 tests=[KAM_ASCII_DIVIDERS=0.8, KAM_LAZY_DOMAIN_SECURITY=1, RCVD_IN_DNSWL_HI=-5, RCVD_IN_MSPIKE_H3=-0.01, RCVD_IN_MSPIKE_WL=-0.01, RP_MATCHES_RCVD=-1.426] autolearn=disabled Received: from mx1-lw-eu.apache.org ([10.40.0.8]) by localhost (spamd2-us-west.apache.org [10.40.0.9]) (amavisd-new, port 10024) with ESMTP id JPfrc4e8Nm0Y for ; Wed, 27 Jul 2016 14:34:08 +0000 (UTC) Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by mx1-lw-eu.apache.org (ASF Mail Server at mx1-lw-eu.apache.org) with SMTP id 83ECC5F2C4 for ; Wed, 27 Jul 2016 14:34:07 +0000 (UTC) Received: (qmail 2863 invoked by uid 99); 27 Jul 2016 14:34:06 -0000 Received: from git1-us-west.apache.org (HELO git1-us-west.apache.org) (140.211.11.23) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 27 Jul 2016 14:34:06 +0000 Received: by git1-us-west.apache.org (ASF Mail Server at git1-us-west.apache.org, from userid 33) id 9E614E02A2; Wed, 27 Jul 2016 14:34:06 +0000 (UTC) From: mxm To: issues@flink.incubator.apache.org Reply-To: issues@flink.incubator.apache.org References: In-Reply-To: Subject: [GitHub] flink pull request #2275: FLINK-3929 Support for Kerberos Authentication wit... Content-Type: text/plain Message-Id: <20160727143406.9E614E02A2@git1-us-west.apache.org> Date: Wed, 27 Jul 2016 14:34:06 +0000 (UTC) archived-at: Wed, 27 Jul 2016 14:34:14 -0000 Github user mxm commented on a diff in the pull request: https://github.com/apache/flink/pull/2275#discussion_r72450591 --- Diff: flink-test-utils-parent/flink-test-utils/src/main/java/org/apache/flink/test/util/SecureTestEnvironment.java --- @@ -0,0 +1,207 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.test.util; + +import org.apache.flink.configuration.ConfigConstants; +import org.apache.flink.runtime.security.SecurityContext; +import org.apache.hadoop.fs.FileUtil; +import org.apache.hadoop.minikdc.MiniKdc; +import org.junit.rules.TemporaryFolder; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; + +import static org.junit.Assert.assertTrue; + +import javax.annotation.Nullable; +import java.io.File; +import java.util.HashMap; +import java.util.Map; +import java.util.Properties; + +/** + * Helper {@link SecureTestEnvironment} to handle MiniKDC lifecycle. + * This class can be used to start/stop MiniKDC and create secure configurations for MiniDFSCluster + * and MiniYarn + */ + +public class SecureTestEnvironment { + + protected static final Logger LOG = LoggerFactory.getLogger(SecureTestEnvironment.class); + + private static MiniKdc kdc; + + private static String testKeytab = null; + + private static String testPrincipal = null; + + private static String testZkServerPrincipal = null; + + private static String hadoopServicePrincipal = null; + + private static File baseDirForSecureRun = null; + + public static void prepare(TemporaryFolder tempFolder) { + + try { + baseDirForSecureRun = tempFolder.newFolder(); + + LOG.info("Base Directory for Secure Environment: {}", baseDirForSecureRun); + + FileUtil.fullyDelete(baseDirForSecureRun); + + assertTrue(baseDirForSecureRun.mkdirs()); + + Properties kdcConf = MiniKdc.createConf(); + + kdc = new MiniKdc(kdcConf, baseDirForSecureRun); + + kdc.start(); + + LOG.info("Started Mini KDC"); + + File keytabFile = new File(baseDirForSecureRun, "test-users.keytab"); + + testKeytab = keytabFile.getAbsolutePath(); + + testZkServerPrincipal = "zookeeper/127.0.0.1"; + + hadoopServicePrincipal = "hadoop/localhost"; + + testPrincipal = "client/localhost"; + + kdc.createPrincipal(keytabFile, testPrincipal, testZkServerPrincipal, hadoopServicePrincipal); + + + testPrincipal = testPrincipal + "@" + kdc.getRealm(); + testZkServerPrincipal = testZkServerPrincipal + "@" + kdc.getRealm(); + hadoopServicePrincipal = hadoopServicePrincipal + "@" + kdc.getRealm(); + + LOG.info("-------------------------------------------------------------------"); + LOG.info("Test Principal: {}", testPrincipal); + LOG.info("Test ZK Server Principal: {}", testZkServerPrincipal); + LOG.info("Test Hadoop Service Principal: {}", hadoopServicePrincipal); + LOG.info("Test Keytab: {}", testKeytab); + LOG.info("-------------------------------------------------------------------"); + + //Security Context is established to allow non hadoop applications that requires JAAS + //based SASL/Kerberos authentication to work. However, for Hadoop specific applications + //the context can be reinitialized with Hadoop configuration by calling + //ctx.setHadoopConfiguration() for the UGI implementation to work properly. + //See Yarn test case module for reference + SecurityContext.SecurityConfiguration ctx = new SecurityContext.SecurityConfiguration(); + ctx.setCredentials(testKeytab, testPrincipal); + TestingSecurityContext.install(ctx, getClientSecurityConfigurationMap()); + + populateSystemEnvVariables(); + + } catch(Exception e) { + LOG.error("Exception occured while preparing secure environment. Reason: {}", e); + throw new RuntimeException(e); + } + + } + + public static void cleanup() { + + LOG.info("Cleaning up Secure Environment"); + + if( kdc != null) { + kdc.stop(); + LOG.info("Stopped KDC server"); + } + + if(baseDirForSecureRun != null) { + try { + FileUtil.fullyDelete(baseDirForSecureRun); --- End diff -- Not necessary to cleanup `TemporaryFolder`. --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastructure@apache.org or file a JIRA ticket with INFRA. ---