ambari-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From ababiic...@apache.org
Subject [04/63] [abbrv] ambari git commit: AMBARI-21268 Remove Upgrade Catalogs For Every Version Before 2.5 (dgrinenko)
Date Fri, 30 Jun 2017 13:37:54 GMT
http://git-wip-us.apache.org/repos/asf/ambari/blob/af1bf85c/ambari-server/src/test/java/org/apache/ambari/server/upgrade/UpgradeCatalog240Test.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/upgrade/UpgradeCatalog240Test.java b/ambari-server/src/test/java/org/apache/ambari/server/upgrade/UpgradeCatalog240Test.java
deleted file mode 100644
index f106658..0000000
--- a/ambari-server/src/test/java/org/apache/ambari/server/upgrade/UpgradeCatalog240Test.java
+++ /dev/null
@@ -1,2688 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *     http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-package org.apache.ambari.server.upgrade;
-
-
-import static org.easymock.EasyMock.anyObject;
-import static org.easymock.EasyMock.anyString;
-import static org.easymock.EasyMock.capture;
-import static org.easymock.EasyMock.createMock;
-import static org.easymock.EasyMock.createMockBuilder;
-import static org.easymock.EasyMock.createNiceMock;
-import static org.easymock.EasyMock.createStrictMock;
-import static org.easymock.EasyMock.eq;
-import static org.easymock.EasyMock.expect;
-import static org.easymock.EasyMock.expectLastCall;
-import static org.easymock.EasyMock.newCapture;
-import static org.easymock.EasyMock.replay;
-import static org.easymock.EasyMock.reset;
-import static org.easymock.EasyMock.verify;
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertFalse;
-import static org.junit.Assert.assertNull;
-import static org.junit.Assert.assertTrue;
-
-import java.io.File;
-import java.lang.reflect.Field;
-import java.lang.reflect.InvocationTargetException;
-import java.lang.reflect.Method;
-import java.sql.Connection;
-import java.sql.ResultSet;
-import java.sql.SQLException;
-import java.sql.Statement;
-import java.util.ArrayList;
-import java.util.Arrays;
-import java.util.Collections;
-import java.util.HashMap;
-import java.util.HashSet;
-import java.util.List;
-import java.util.Map;
-
-import javax.persistence.EntityManager;
-
-import org.apache.ambari.server.AmbariException;
-import org.apache.ambari.server.H2DatabaseCleaner;
-import org.apache.ambari.server.actionmanager.ActionManager;
-import org.apache.ambari.server.api.services.AmbariMetaInfo;
-import org.apache.ambari.server.configuration.Configuration;
-import org.apache.ambari.server.controller.AmbariManagementController;
-import org.apache.ambari.server.controller.AmbariManagementControllerImpl;
-import org.apache.ambari.server.controller.KerberosHelper;
-import org.apache.ambari.server.controller.MaintenanceStateHelper;
-import org.apache.ambari.server.hooks.HookContextFactory;
-import org.apache.ambari.server.hooks.HookService;
-import org.apache.ambari.server.orm.DBAccessor;
-import org.apache.ambari.server.orm.GuiceJpaInitializer;
-import org.apache.ambari.server.orm.InMemoryDefaultTestModule;
-import org.apache.ambari.server.orm.dao.AlertDefinitionDAO;
-import org.apache.ambari.server.orm.dao.ClusterDAO;
-import org.apache.ambari.server.orm.dao.PrivilegeDAO;
-import org.apache.ambari.server.orm.dao.RemoteAmbariClusterDAO;
-import org.apache.ambari.server.orm.dao.RequestScheduleDAO;
-import org.apache.ambari.server.orm.dao.StackDAO;
-import org.apache.ambari.server.orm.dao.UserDAO;
-import org.apache.ambari.server.orm.dao.ViewInstanceDAO;
-import org.apache.ambari.server.orm.dao.WidgetDAO;
-import org.apache.ambari.server.orm.entities.AlertDefinitionEntity;
-import org.apache.ambari.server.orm.entities.ClusterEntity;
-import org.apache.ambari.server.orm.entities.PermissionEntity;
-import org.apache.ambari.server.orm.entities.PrincipalEntity;
-import org.apache.ambari.server.orm.entities.PrivilegeEntity;
-import org.apache.ambari.server.orm.entities.RemoteAmbariClusterEntity;
-import org.apache.ambari.server.orm.entities.RequestScheduleEntity;
-import org.apache.ambari.server.orm.entities.ResourceEntity;
-import org.apache.ambari.server.orm.entities.ResourceTypeEntity;
-import org.apache.ambari.server.orm.entities.UserEntity;
-import org.apache.ambari.server.orm.entities.ViewInstanceEntity;
-import org.apache.ambari.server.orm.entities.WidgetEntity;
-import org.apache.ambari.server.security.authorization.ResourceType;
-import org.apache.ambari.server.security.authorization.User;
-import org.apache.ambari.server.security.authorization.UserName;
-import org.apache.ambari.server.security.authorization.Users;
-import org.apache.ambari.server.stack.StackManagerFactory;
-import org.apache.ambari.server.state.AlertFirmness;
-import org.apache.ambari.server.state.Cluster;
-import org.apache.ambari.server.state.Clusters;
-import org.apache.ambari.server.state.Config;
-import org.apache.ambari.server.state.ConfigHelper;
-import org.apache.ambari.server.state.PropertyInfo;
-import org.apache.ambari.server.state.SecurityType;
-import org.apache.ambari.server.state.Service;
-import org.apache.ambari.server.state.ServiceInfo;
-import org.apache.ambari.server.state.StackId;
-import org.apache.ambari.server.state.StackInfo;
-import org.apache.ambari.server.state.alert.SourceType;
-import org.apache.ambari.server.state.kerberos.KerberosDescriptor;
-import org.apache.ambari.server.state.kerberos.KerberosIdentityDescriptor;
-import org.apache.ambari.server.state.kerberos.KerberosKeytabDescriptor;
-import org.apache.ambari.server.state.kerberos.KerberosPrincipalDescriptor;
-import org.apache.ambari.server.state.stack.OsFamily;
-import org.apache.ambari.server.view.DefaultMasker;
-import org.apache.ambari.view.ClusterType;
-import org.apache.ambari.view.MaskException;
-import org.apache.commons.io.FileUtils;
-import org.easymock.Capture;
-import org.easymock.CaptureType;
-import org.easymock.EasyMock;
-import org.easymock.EasyMockSupport;
-import org.junit.AfterClass;
-import org.junit.Before;
-import org.junit.BeforeClass;
-import org.junit.Test;
-import org.junit.rules.TemporaryFolder;
-import org.springframework.security.crypto.password.PasswordEncoder;
-
-import com.google.common.collect.ImmutableMap;
-import com.google.common.collect.Maps;
-import com.google.gson.Gson;
-import com.google.inject.AbstractModule;
-import com.google.inject.Binder;
-import com.google.inject.Guice;
-import com.google.inject.Injector;
-import com.google.inject.Module;
-import com.google.inject.Provider;
-
-import junit.framework.Assert;
-
-public class UpgradeCatalog240Test {
-  private static final String CAPACITY_SCHEDULER_CONFIG_TYPE = "capacity-scheduler";
-  private static final String WEBHCAT_SITE_CONFIG_TYPE = "webhcat-site";
-  private static final String TEZ_SITE_CONFIG_TYPE = "tez-site";
-  private static final String MAPRED_SITE_CONFIG_TYPE = "mapred-site";
-  private static final String YARN_ENV_CONFIG_TYPE = "yarn-env";
-  private static Injector injector;
-  private Provider<EntityManager> entityManagerProvider = createStrictMock(Provider.class);
-  private EntityManager entityManager = createNiceMock(EntityManager.class);
-
-
-  @BeforeClass
-  public static void classSetUp() {
-    injector = Guice.createInjector(new InMemoryDefaultTestModule());
-    injector.getInstance(GuiceJpaInitializer.class);
-  }
-
-  @Before
-  public void init() {
-    reset(entityManagerProvider);
-    expect(entityManagerProvider.get()).andReturn(entityManager).anyTimes();
-    replay(entityManagerProvider);
-
-    injector.getInstance(UpgradeCatalogHelper.class);
-    // inject AmbariMetaInfo to ensure that stacks get populated in the DB
-    injector.getInstance(AmbariMetaInfo.class);
-    // load the stack entity
-    StackDAO stackDAO = injector.getInstance(StackDAO.class);
-    stackDAO.find("HDP", "2.2.0");
-  }
-
-  @AfterClass
-  public static void tearDown() throws AmbariException, SQLException {
-    H2DatabaseCleaner.clearDatabaseAndStopPersistenceService(injector);
-  }
-
-  @Test
-  public void testExecuteDDLUpdates() throws SQLException, AmbariException, ClassNotFoundException {
-    Capture<DBAccessor.DBColumnInfo> capturedSortOrderColumnInfo = newCapture();
-    Capture<DBAccessor.DBColumnInfo> capturedPermissionIDColumnInfo = newCapture();
-    Capture<DBAccessor.DBColumnInfo> capturedScColumnInfo = newCapture();
-    Capture<DBAccessor.DBColumnInfo> capturedScDesiredVersionColumnInfo = newCapture();
-
-    final DBAccessor dbAccessor = createStrictMock(DBAccessor.class);
-    Configuration configuration = createNiceMock(Configuration.class);
-    Connection connection = createNiceMock(Connection.class);
-    Statement statement = createNiceMock(Statement.class);
-    ResultSet resultSet = createNiceMock(ResultSet.class);
-
-    Capture<List<DBAccessor.DBColumnInfo>> capturedExtensionColumns = EasyMock.newCapture();
-    Capture<List<DBAccessor.DBColumnInfo>> capturedExtensionLinkColumns = EasyMock.newCapture();
-    Capture<List<DBAccessor.DBColumnInfo>> capturedSettingColumns = EasyMock.newCapture();
-
-    dbAccessor.addColumn(eq("adminpermission"), capture(capturedSortOrderColumnInfo));
-    dbAccessor.addColumn(eq("adminpermission"), capture(capturedPermissionIDColumnInfo));
-    dbAccessor.addColumn(eq(UpgradeCatalog240.SERVICE_COMPONENT_DESIRED_STATE_TABLE), capture(capturedScColumnInfo));
-    dbAccessor.addColumn(eq(UpgradeCatalog240.SERVICE_COMPONENT_DESIRED_STATE_TABLE),
-        capture(capturedScDesiredVersionColumnInfo));
-
-    dbAccessor.createTable(eq("extension"), capture(capturedExtensionColumns), eq("extension_id"));
-    dbAccessor.addUniqueConstraint("extension", "UQ_extension", "extension_name", "extension_version");
-
-    expect(dbAccessor.getConnection()).andReturn(connection);
-    dbAccessor.createTable(eq("extensionlink"), capture(capturedExtensionLinkColumns), eq("link_id"));
-    dbAccessor.addUniqueConstraint("extensionlink", "UQ_extension_link", "stack_id", "extension_id");
-    dbAccessor.addFKConstraint("extensionlink", "FK_extensionlink_extension_id", "extension_id", "extension",
-                               "extension_id", false);
-    dbAccessor.addFKConstraint("extensionlink", "FK_extensionlink_stack_id", "stack_id", "stack",
-                               "stack_id", false);
-
-    expect(dbAccessor.getConnection()).andReturn(connection);
-    dbAccessor.createTable(eq("setting"), capture(capturedSettingColumns), eq("id"));
-
-    expect(configuration.getDatabaseUrl()).andReturn(Configuration.JDBC_IN_MEMORY_URL).anyTimes();
-    expect(dbAccessor.getConnection()).andReturn(connection);
-    expect(connection.createStatement()).andReturn(statement);
-    expect(statement.executeQuery(anyObject(String.class))).andReturn(resultSet);
-
-    Capture<DBAccessor.DBColumnInfo> repoVersionRepoTypeColumnCapture = newCapture();
-    Capture<DBAccessor.DBColumnInfo> repoVersionUrlColumnCapture = newCapture();
-    Capture<DBAccessor.DBColumnInfo> repoVersionXmlColumnCapture = newCapture();
-    Capture<DBAccessor.DBColumnInfo> repoVersionXsdColumnCapture = newCapture();
-    Capture<DBAccessor.DBColumnInfo> repoVersionParentIdColumnCapture = newCapture();
-
-    dbAccessor.addColumn(eq("repo_version"), capture(repoVersionRepoTypeColumnCapture));
-    dbAccessor.addColumn(eq("repo_version"), capture(repoVersionUrlColumnCapture));
-    dbAccessor.addColumn(eq("repo_version"), capture(repoVersionXmlColumnCapture));
-    dbAccessor.addColumn(eq("repo_version"), capture(repoVersionXsdColumnCapture));
-    dbAccessor.addColumn(eq("repo_version"), capture(repoVersionParentIdColumnCapture));
-
-    // skip all of the drama of the servicecomponentdesiredstate table for now
-    expect(dbAccessor.tableHasPrimaryKey("servicecomponentdesiredstate", "id")).andReturn(true);
-
-    Capture<List<DBAccessor.DBColumnInfo>> capturedHistoryColumns = EasyMock.newCapture();
-    dbAccessor.createTable(eq("servicecomponent_history"), capture(capturedHistoryColumns),
-            eq((String[]) null));
-
-    dbAccessor.addPKConstraint("servicecomponent_history", "PK_sc_history", "id");
-    dbAccessor.addFKConstraint("servicecomponent_history", "FK_sc_history_component_id",
-        "component_id", "servicecomponentdesiredstate", "id", false);
-
-    dbAccessor.addFKConstraint("servicecomponent_history", "FK_sc_history_upgrade_id", "upgrade_id",
-        "upgrade", "upgrade_id", false);
-
-    dbAccessor.addFKConstraint("servicecomponent_history", "FK_sc_history_from_stack_id",
-            "from_stack_id", "stack", "stack_id", false);
-
-    dbAccessor.addFKConstraint("servicecomponent_history", "FK_sc_history_to_stack_id",
-            "to_stack_id", "stack", "stack_id", false);
-
-
-    expect(dbAccessor.getConnection()).andReturn(connection);
-    expect(connection.createStatement()).andReturn(statement);
-    expect(statement.executeQuery(anyObject(String.class))).andReturn(resultSet);
-
-    Capture<DBAccessor.DBColumnInfo> capturedClusterUpgradeColumnInfo = newCapture();
-    dbAccessor.addColumn(eq(UpgradeCatalog240.CLUSTER_TABLE), capture(capturedClusterUpgradeColumnInfo));
-    dbAccessor.addFKConstraint(UpgradeCatalog240.CLUSTER_TABLE, "FK_clusters_upgrade_id",
-            UpgradeCatalog240.CLUSTER_UPGRADE_ID_COLUMN, UpgradeCatalog240.UPGRADE_TABLE, "upgrade_id", false);
-
-    Capture<DBAccessor.DBColumnInfo> capturedHelpURLColumnInfo = newCapture();
-    Capture<DBAccessor.DBColumnInfo> capturedRepeatToleranceColumnInfo = newCapture();
-    Capture<DBAccessor.DBColumnInfo> capturedRepeatToleranceEnabledColumnInfo = newCapture();
-    Capture<DBAccessor.DBColumnInfo> capturedOccurrencesColumnInfo = newCapture();
-    Capture<DBAccessor.DBColumnInfo> capturedFirmnessColumnInfo = newCapture();
-    Capture<DBAccessor.DBColumnInfo> capturedTargetEnabledColumnInfo = newCapture();
-
-    dbAccessor.addColumn(eq(UpgradeCatalog240.ALERT_DEFINITION_TABLE), capture(capturedHelpURLColumnInfo));
-    dbAccessor.addColumn(eq(UpgradeCatalog240.ALERT_DEFINITION_TABLE), capture(capturedRepeatToleranceColumnInfo));
-    dbAccessor.addColumn(eq(UpgradeCatalog240.ALERT_DEFINITION_TABLE), capture(capturedRepeatToleranceEnabledColumnInfo));
-    dbAccessor.addColumn(eq(UpgradeCatalog240.ALERT_CURRENT_TABLE), capture(capturedOccurrencesColumnInfo));
-    dbAccessor.addColumn(eq(UpgradeCatalog240.ALERT_CURRENT_TABLE), capture(capturedFirmnessColumnInfo));
-    dbAccessor.addColumn(eq(UpgradeCatalog240.ALERT_TARGET_TABLE), capture(capturedTargetEnabledColumnInfo));
-
-    // Test creation of blueprint_setting table
-    Capture<List<DBAccessor.DBColumnInfo>> capturedBlueprintSettingColumns = EasyMock.newCapture();
-    dbAccessor.createTable(eq(UpgradeCatalog240.BLUEPRINT_SETTING_TABLE), capture(capturedBlueprintSettingColumns));
-    dbAccessor.addPKConstraint(UpgradeCatalog240.BLUEPRINT_SETTING_TABLE, "PK_blueprint_setting", UpgradeCatalog240.ID);
-    dbAccessor.addUniqueConstraint(UpgradeCatalog240.BLUEPRINT_SETTING_TABLE, "UQ_blueprint_setting_name",
-            UpgradeCatalog240.BLUEPRINT_NAME_COL, UpgradeCatalog240.SETTING_NAME_COL);
-    dbAccessor.addFKConstraint(UpgradeCatalog240.BLUEPRINT_SETTING_TABLE, "FK_blueprint_setting_name",
-            UpgradeCatalog240.BLUEPRINT_NAME_COL, UpgradeCatalog240.BLUEPRINT_TABLE,
-            UpgradeCatalog240.BLUEPRINT_NAME_COL, false);
-    expect(dbAccessor.getConnection()).andReturn(connection);
-    expect(connection.createStatement()).andReturn(statement);
-    expect(statement.executeQuery(anyObject(String.class))).andReturn(resultSet);
-
-    // Test host_role_command adds a column called original_start_time
-    Capture<DBAccessor.DBColumnInfo> hostRoleCommandOriginalStartTimeColumnInfo = newCapture();
-    dbAccessor.addColumn(eq(UpgradeCatalog240.HOST_ROLE_COMMAND_TABLE), capture(hostRoleCommandOriginalStartTimeColumnInfo));
-
-    Capture<List<DBAccessor.DBColumnInfo>> capturedViewUrlColums = EasyMock.newCapture();
-    dbAccessor.createTable(eq(UpgradeCatalog240.VIEWURL_TABLE), capture(capturedViewUrlColums),eq("url_id"));
-    expect(dbAccessor.getConnection()).andReturn(connection);
-    expect(connection.createStatement()).andReturn(statement);
-
-    Capture<DBAccessor.DBColumnInfo> viewInstanceShortUrlInfo = newCapture();
-    dbAccessor.addColumn(eq(UpgradeCatalog240.VIEWINSTANCE_TABLE), capture(viewInstanceShortUrlInfo));
-
-    dbAccessor.addFKConstraint(UpgradeCatalog240.VIEWINSTANCE_TABLE, "FK_instance_url_id",
-            UpgradeCatalog240.SHORT_URL_COLUMN, UpgradeCatalog240.VIEWURL_TABLE, "url_id", false);
-
-    Capture<DBAccessor.DBColumnInfo> viewInstanceClusterType = newCapture();
-    dbAccessor.addColumn(eq(UpgradeCatalog240.VIEWINSTANCE_TABLE), capture(viewInstanceClusterType));
-
-    // Test remote Cluster Tables
-    Capture<List<DBAccessor.DBColumnInfo>> capturedRemoteAmbariClusterColumns = EasyMock.newCapture();
-    dbAccessor.createTable(eq(UpgradeCatalog240.REMOTE_AMBARI_CLUSTER_TABLE), capture(capturedRemoteAmbariClusterColumns),anyString());
-    dbAccessor.addUniqueConstraint(UpgradeCatalog240.REMOTE_AMBARI_CLUSTER_TABLE , "UQ_remote_ambari_cluster" , UpgradeCatalog240.CLUSTER_NAME);
-    expect(dbAccessor.getConnection()).andReturn(connection);
-    expect(connection.createStatement()).andReturn(statement);
-
-    Capture<List<DBAccessor.DBColumnInfo>> capturedRemoteClusterServiceColumns = EasyMock.newCapture();
-    dbAccessor.createTable(eq(UpgradeCatalog240.REMOTE_AMBARI_CLUSTER_SERVICE_TABLE), capture(capturedRemoteClusterServiceColumns),anyString());
-    dbAccessor.addFKConstraint(UpgradeCatalog240.REMOTE_AMBARI_CLUSTER_SERVICE_TABLE, "FK_remote_ambari_cluster_id",
-      UpgradeCatalog240.CLUSTER_ID, UpgradeCatalog240.REMOTE_AMBARI_CLUSTER_TABLE, UpgradeCatalog240.CLUSTER_ID, false);
-    expect(dbAccessor.getConnection()).andReturn(connection);
-    expect(connection.createStatement()).andReturn(statement);
-
-    // Test viewInstance update
-    expect(dbAccessor.getColumnClass(UpgradeCatalog240.VIEWINSTANCE_TABLE, UpgradeCatalog240.CLUSTER_HANDLE_COLUMN)).andReturn(String.class);
-    dbAccessor.addColumn(eq(UpgradeCatalog240.VIEWINSTANCE_TABLE), anyObject(DBAccessor.DBColumnInfo.class));
-
-    expect(dbAccessor.getConnection()).andReturn(connection);
-    expect(connection.createStatement()).andReturn(statement);
-
-    dbAccessor.dropColumn(UpgradeCatalog240.VIEWINSTANCE_TABLE, UpgradeCatalog240.CLUSTER_HANDLE_COLUMN);
-
-    Capture<DBAccessor.DBColumnInfo> capturedClusterHandleColumn = EasyMock.newCapture();
-    dbAccessor.renameColumn(eq(UpgradeCatalog240.VIEWINSTANCE_TABLE), anyString(), capture(capturedClusterHandleColumn));
-    Capture<DBAccessor.DBColumnInfo> requestScheduleUserIdInfo = newCapture();
-    dbAccessor.addColumn(eq(UpgradeCatalog240.REQUESTSCHEDULE_TABLE), capture(requestScheduleUserIdInfo));
-
-    Capture<DBAccessor.DBColumnInfo> provisionActionColumnInfo = newCapture();
-    dbAccessor.addColumn(eq(UpgradeCatalog240.TOPOLOGY_REQUEST_TABLE), capture(provisionActionColumnInfo));
-
-    replay(dbAccessor, configuration, connection, statement, resultSet);
-
-    Module module = new Module() {
-      @Override
-      public void configure(Binder binder) {
-        binder.bind(DBAccessor.class).toInstance(dbAccessor);
-        binder.bind(OsFamily.class).toInstance(createNiceMock(OsFamily.class));
-        binder.bind(EntityManager.class).toInstance(entityManager);
-        binder.bind(PasswordEncoder.class).toInstance(createNiceMock(PasswordEncoder.class));
-        binder.bind(HookContextFactory.class).toInstance(createMock(HookContextFactory.class));
-        binder.bind(HookService.class).toInstance(createMock(HookService.class));
-      }
-      };
-
-    Injector injector = Guice.createInjector(module);
-    UpgradeCatalog240 upgradeCatalog240 = injector.getInstance(UpgradeCatalog240.class);
-    upgradeCatalog240.executeDDLUpdates();
-
-    DBAccessor.DBColumnInfo columnSortOrderInfo = capturedSortOrderColumnInfo.getValue();
-    Assert.assertNotNull(columnSortOrderInfo);
-    Assert.assertEquals(UpgradeCatalog240.SORT_ORDER_COL, columnSortOrderInfo.getName());
-    Assert.assertEquals(null, columnSortOrderInfo.getLength());
-    Assert.assertEquals(Short.class, columnSortOrderInfo.getType());
-    Assert.assertEquals(1, columnSortOrderInfo.getDefaultValue());
-    Assert.assertEquals(false, columnSortOrderInfo.isNullable());
-
-    DBAccessor.DBColumnInfo columnPrincipalIDInfo = capturedPermissionIDColumnInfo.getValue();
-    Assert.assertNotNull(columnPrincipalIDInfo);
-    Assert.assertEquals(UpgradeCatalog240.PRINCIPAL_ID_COL, columnPrincipalIDInfo.getName());
-    Assert.assertEquals(null, columnPrincipalIDInfo.getLength());
-    Assert.assertEquals(Long.class, columnPrincipalIDInfo.getType());
-    Assert.assertEquals(null, columnPrincipalIDInfo.getDefaultValue());
-    Assert.assertEquals(true, columnPrincipalIDInfo.isNullable());
-
-    // Verify if recovery_enabled column was added to servicecomponentdesiredstate table
-    DBAccessor.DBColumnInfo columnScInfo = capturedScColumnInfo.getValue();
-    Assert.assertNotNull(columnScInfo);
-    Assert.assertEquals(UpgradeCatalog240.RECOVERY_ENABLED_COL, columnScInfo.getName());
-    Assert.assertEquals(null, columnScInfo.getLength());
-    Assert.assertEquals(Short.class, columnScInfo.getType());
-    Assert.assertEquals(0, columnScInfo.getDefaultValue());
-    Assert.assertEquals(false, columnScInfo.isNullable());
-
-    DBAccessor.DBColumnInfo columnScDesiredVersionInfo = capturedScDesiredVersionColumnInfo.getValue();
-    Assert.assertNotNull(columnScDesiredVersionInfo);
-    Assert.assertEquals(UpgradeCatalog240.DESIRED_VERSION_COLUMN_NAME, columnScDesiredVersionInfo.getName());
-    Assert.assertEquals(Integer.valueOf(255), columnScDesiredVersionInfo.getLength());
-    Assert.assertEquals(String.class, columnScDesiredVersionInfo.getType());
-    Assert.assertEquals("UNKNOWN", columnScDesiredVersionInfo.getDefaultValue());
-    Assert.assertEquals(false, columnScDesiredVersionInfo.isNullable());
-
-    // Verify if upgrade_id column was added to clusters table
-    DBAccessor.DBColumnInfo clusterUpgradeColumnInfo = capturedClusterUpgradeColumnInfo.getValue();
-    Assert.assertNotNull(clusterUpgradeColumnInfo);
-    Assert.assertEquals(UpgradeCatalog240.CLUSTER_UPGRADE_ID_COLUMN, clusterUpgradeColumnInfo.getName());
-    Assert.assertEquals(null, clusterUpgradeColumnInfo.getLength());
-    Assert.assertEquals(Long.class, clusterUpgradeColumnInfo.getType());
-    Assert.assertEquals(null, clusterUpgradeColumnInfo.getDefaultValue());
-    Assert.assertEquals(true, clusterUpgradeColumnInfo.isNullable());
-
-    Map<String, Class> expectedCaptures = new HashMap<>();
-    expectedCaptures.put("id", Long.class);
-    expectedCaptures.put("name", String.class);
-    expectedCaptures.put("setting_type", String.class);
-    expectedCaptures.put("content", String.class);
-    expectedCaptures.put("updated_by", String.class);
-    expectedCaptures.put("update_timestamp", Long.class);
-
-    Map<String, Class> actualCaptures = new HashMap<>();
-    for(DBAccessor.DBColumnInfo settingColumnInfo : capturedSettingColumns.getValue()) {
-      actualCaptures.put(settingColumnInfo.getName(), settingColumnInfo.getType());
-    }
-
-    assertEquals(expectedCaptures, actualCaptures);
-
-    expectedCaptures = new HashMap<>();
-    expectedCaptures.put("extension_id", Long.class);
-    expectedCaptures.put("extension_name", String.class);
-    expectedCaptures.put("extension_version", String.class);
-
-    actualCaptures = new HashMap<>();
-    for(DBAccessor.DBColumnInfo settingColumnInfo : capturedExtensionColumns.getValue()) {
-      actualCaptures.put(settingColumnInfo.getName(), settingColumnInfo.getType());
-    }
-
-    assertEquals(expectedCaptures, actualCaptures);
-
-
-    expectedCaptures = new HashMap<>();
-    expectedCaptures.put("link_id", Long.class);
-    expectedCaptures.put("stack_id", Long.class);
-    expectedCaptures.put("extension_id", Long.class);
-
-    actualCaptures = new HashMap<>();
-    for(DBAccessor.DBColumnInfo settingColumnInfo : capturedExtensionLinkColumns.getValue()) {
-      actualCaptures.put(settingColumnInfo.getName(), settingColumnInfo.getType());
-    }
-
-    assertEquals(expectedCaptures, actualCaptures);
-
-    expectedCaptures = new HashMap<>();
-    expectedCaptures.put("id", Long.class);
-    expectedCaptures.put("component_id", Long.class);
-    expectedCaptures.put("upgrade_id", Long.class);
-    expectedCaptures.put("from_stack_id", Long.class);
-    expectedCaptures.put("to_stack_id", Long.class);
-
-    actualCaptures = new HashMap<>();
-    for (DBAccessor.DBColumnInfo historyColumnInfo : capturedHistoryColumns.getValue()) {
-      actualCaptures.put(historyColumnInfo.getName(), historyColumnInfo.getType());
-    }
-
-    DBAccessor.DBColumnInfo columnHelpURLInfo = capturedHelpURLColumnInfo.getValue();
-    Assert.assertNotNull(columnHelpURLInfo);
-    Assert.assertEquals(UpgradeCatalog240.HELP_URL_COLUMN, columnHelpURLInfo.getName());
-    Assert.assertEquals(Integer.valueOf(512), columnHelpURLInfo.getLength());
-    Assert.assertEquals(String.class, columnHelpURLInfo.getType());
-    Assert.assertEquals(null, columnHelpURLInfo.getDefaultValue());
-    Assert.assertEquals(true, columnHelpURLInfo.isNullable());
-
-    DBAccessor.DBColumnInfo columnRepeatToleranceInfo = capturedRepeatToleranceColumnInfo.getValue();
-    Assert.assertNotNull(columnRepeatToleranceInfo);
-    Assert.assertEquals(UpgradeCatalog240.REPEAT_TOLERANCE_COLUMN, columnRepeatToleranceInfo.getName());
-    Assert.assertEquals(Integer.class, columnRepeatToleranceInfo.getType());
-    Assert.assertEquals(1, columnRepeatToleranceInfo.getDefaultValue());
-    Assert.assertEquals(false, columnRepeatToleranceInfo.isNullable());
-
-    DBAccessor.DBColumnInfo columnRepeatToleranceEnabledInfo = capturedRepeatToleranceEnabledColumnInfo.getValue();
-    Assert.assertNotNull(columnRepeatToleranceEnabledInfo);
-    Assert.assertEquals(UpgradeCatalog240.REPEAT_TOLERANCE_ENABLED_COLUMN, columnRepeatToleranceEnabledInfo.getName());
-    Assert.assertEquals(Short.class, columnRepeatToleranceEnabledInfo.getType());
-    Assert.assertEquals(0, columnRepeatToleranceEnabledInfo.getDefaultValue());
-    Assert.assertEquals(false, columnRepeatToleranceEnabledInfo.isNullable());
-
-    DBAccessor.DBColumnInfo columnOccurrencesInfo = capturedOccurrencesColumnInfo.getValue();
-    Assert.assertNotNull(columnOccurrencesInfo);
-    Assert.assertEquals(UpgradeCatalog240.ALERT_CURRENT_OCCURRENCES_COLUMN, columnOccurrencesInfo.getName());
-    Assert.assertEquals(Long.class, columnOccurrencesInfo.getType());
-    Assert.assertEquals(1, columnOccurrencesInfo.getDefaultValue());
-    Assert.assertEquals(false, columnOccurrencesInfo.isNullable());
-
-    DBAccessor.DBColumnInfo columnFirmnessInfo = capturedFirmnessColumnInfo.getValue();
-    Assert.assertNotNull(columnFirmnessInfo);
-    Assert.assertEquals(UpgradeCatalog240.ALERT_CURRENT_FIRMNESS_COLUMN, columnFirmnessInfo.getName());
-    Assert.assertEquals(String.class, columnFirmnessInfo.getType());
-    Assert.assertEquals(AlertFirmness.HARD.name(), columnFirmnessInfo.getDefaultValue());
-    Assert.assertEquals(false, columnFirmnessInfo.isNullable());
-
-    DBAccessor.DBColumnInfo targetEnabledColumnInfo = capturedTargetEnabledColumnInfo.getValue();
-    Assert.assertNotNull(targetEnabledColumnInfo);
-    Assert.assertEquals(UpgradeCatalog240.ALERT_TARGET_ENABLED_COLUMN, targetEnabledColumnInfo.getName());
-    Assert.assertEquals(Short.class, targetEnabledColumnInfo.getType());
-    Assert.assertEquals(1, targetEnabledColumnInfo.getDefaultValue());
-    Assert.assertEquals(false, targetEnabledColumnInfo.isNullable());
-
-    assertEquals(expectedCaptures, actualCaptures);
-
-    // Verify blueprint_setting columns
-    expectedCaptures = new HashMap<>();
-    expectedCaptures.put(UpgradeCatalog240.ID, Long.class);
-    expectedCaptures.put(UpgradeCatalog240.BLUEPRINT_NAME_COL, String.class);
-    expectedCaptures.put(UpgradeCatalog240.SETTING_NAME_COL, String.class);
-    expectedCaptures.put(UpgradeCatalog240.SETTING_DATA_COL, char[].class);
-
-    actualCaptures = new HashMap<>();
-    for(DBAccessor.DBColumnInfo blueprintSettingsColumnInfo : capturedBlueprintSettingColumns.getValue()) {
-      actualCaptures.put(blueprintSettingsColumnInfo.getName(), blueprintSettingsColumnInfo.getType());
-    }
-
-    assertEquals(expectedCaptures, actualCaptures);
-
-    // Verify host_role_command column
-    DBAccessor.DBColumnInfo originalStartTimeInfo = hostRoleCommandOriginalStartTimeColumnInfo.getValue();
-    Assert.assertNotNull(originalStartTimeInfo);
-    Assert.assertEquals("original_start_time", originalStartTimeInfo.getName());
-    Assert.assertEquals(Long.class, originalStartTimeInfo.getType());
-    Assert.assertEquals(-1L, originalStartTimeInfo.getDefaultValue());
-
-    DBAccessor.DBColumnInfo viewInstanceEntityUrlColInfoValue = viewInstanceShortUrlInfo.getValue();
-    Assert.assertNotNull(viewInstanceEntityUrlColInfoValue);
-    Assert.assertEquals("short_url", viewInstanceEntityUrlColInfoValue.getName());
-    Assert.assertEquals(Long.class, viewInstanceEntityUrlColInfoValue.getType());
-
-    List<DBAccessor.DBColumnInfo> capturedViewUrlColumsValue = capturedViewUrlColums.getValue();
-    Assert.assertNotNull(capturedViewUrlColumsValue);
-    Assert.assertEquals(3, capturedViewUrlColumsValue.size());
-
-    // Verify cluster_type column
-    DBAccessor.DBColumnInfo viewInstanceEntityClusterTypeValue = viewInstanceClusterType.getValue();
-    Assert.assertNotNull(viewInstanceClusterType);
-    Assert.assertEquals("cluster_type", viewInstanceEntityClusterTypeValue.getName());
-    Assert.assertEquals(String.class, viewInstanceEntityClusterTypeValue.getType());
-
-    List<DBAccessor.DBColumnInfo> capturedRemoteAmbariClusterColumnsValue = capturedRemoteAmbariClusterColumns.getValue();
-    Assert.assertNotNull(capturedRemoteAmbariClusterColumnsValue);
-    Assert.assertEquals(5, capturedRemoteAmbariClusterColumnsValue.size());
-
-    List<DBAccessor.DBColumnInfo> capturedRemoteClusterServiceColumnsValue = capturedRemoteClusterServiceColumns.getValue();
-    Assert.assertNotNull(capturedRemoteClusterServiceColumnsValue);
-    Assert.assertEquals(3, capturedRemoteClusterServiceColumnsValue.size());
-
-    DBAccessor.DBColumnInfo clusterHandleColumn = capturedClusterHandleColumn.getValue();
-    Assert.assertEquals(UpgradeCatalog240.CLUSTER_HANDLE_COLUMN, clusterHandleColumn.getName());
-    Assert.assertEquals(Long.class, clusterHandleColumn.getType());
-
-    // Verify authenticated_user_id column
-    DBAccessor.DBColumnInfo requestScheduleUserIdInfoValue = requestScheduleUserIdInfo.getValue();
-    Assert.assertNotNull(requestScheduleUserIdInfoValue);
-    Assert.assertEquals("authenticated_user_id", requestScheduleUserIdInfoValue.getName());
-    Assert.assertEquals(Integer.class, requestScheduleUserIdInfoValue.getType());
-    Assert.assertEquals(null, requestScheduleUserIdInfoValue.getDefaultValue());
-
-    DBAccessor.DBColumnInfo provisionActionColumnInfoValue = provisionActionColumnInfo.getValue();
-    Assert.assertNotNull(provisionActionColumnInfoValue);
-    Assert.assertEquals(UpgradeCatalog240.PROVISION_ACTION_COL, provisionActionColumnInfoValue.getName());
-    Assert.assertEquals(String.class, provisionActionColumnInfoValue.getType());
-    Assert.assertEquals(true, provisionActionColumnInfoValue.isNullable());
-
-    verify(dbAccessor);
-  }
-
-  @Test
-  public void testExecuteDMLUpdates() throws Exception {
-    Method addNewConfigurationsFromXml = AbstractUpgradeCatalog.class.getDeclaredMethod("addNewConfigurationsFromXml");
-    Method updateAlerts = UpgradeCatalog240.class.getDeclaredMethod("updateAlerts");
-    Method addManageUserPersistedDataPermission = UpgradeCatalog240.class.getDeclaredMethod("addManageUserPersistedDataPermission");
-    Method allowClusterOperatorToManageCredentials = UpgradeCatalog240.class.getDeclaredMethod("allowClusterOperatorToManageCredentials");
-    Method addSettingPermission = UpgradeCatalog240.class.getDeclaredMethod("addSettingPermission");
-    Method addViewOperationalLogsPermission = UpgradeCatalog240.class.getDeclaredMethod("addViewOperationalLogsPermission");
-    Method updateHDFSConfigs = UpgradeCatalog240.class.getDeclaredMethod("updateHDFSConfigs");
-    Method updateHIVEConfigs = UpgradeCatalog240.class.getDeclaredMethod("updateHIVEConfigs");
-    Method updateAmsConfigs = UpgradeCatalog240.class.getDeclaredMethod("updateAMSConfigs");
-    Method updateClusterEnv = UpgradeCatalog240.class.getDeclaredMethod("updateClusterEnv");
-    Method updateSequenceForView = UpgradeCatalog240.class.getDeclaredMethod("updateSequenceForView");
-    Method adjustHiveJobTimestamps = UpgradeCatalog240.class.getDeclaredMethod("adjustHiveJobTimestamps");
-    Method updateHostRoleCommandTableDML = UpgradeCatalog240.class.getDeclaredMethod("updateHostRoleCommandTableDML");
-    Method updateKerberosEnv = UpgradeCatalog240.class.getDeclaredMethod("updateKerberosConfigs");
-    Method updateYarnEnv = UpgradeCatalog240.class.getDeclaredMethod("updateYarnEnv");
-    Method removeHiveOozieDBConnectionConfigs = UpgradeCatalog240.class.getDeclaredMethod("removeHiveOozieDBConnectionConfigs");
-    Method updateClustersAndHostsVersionStateTableDML = UpgradeCatalog240.class.getDeclaredMethod("updateClustersAndHostsVersionStateTableDML");
-    Method removeStandardDeviationAlerts = UpgradeCatalog240.class.getDeclaredMethod("removeStandardDeviationAlerts");
-    Method consolidateUserRoles = UpgradeCatalog240.class.getDeclaredMethod("consolidateUserRoles");
-    Method updateClusterInheritedPermissionsConfig = UpgradeCatalog240.class.getDeclaredMethod("updateClusterInheritedPermissionsConfig");
-    Method createRolePrincipals = UpgradeCatalog240.class.getDeclaredMethod("createRolePrincipals");
-    Method updateHDFSWidget = UpgradeCatalog240.class.getDeclaredMethod("updateHDFSWidgetDefinition");
-    Method upgradeCapSchedulerView = UpgradeCatalog240.class.getDeclaredMethod("upgradeCapSchedulerView");
-    Method updatePhoenixConfigs = UpgradeCatalog240.class.getDeclaredMethod("updatePhoenixConfigs");
-    Method updateSparkConfigs = UpgradeCatalog240.class.getDeclaredMethod("updateSparkConfigs");
-    Method updateHBaseConfigs = UpgradeCatalog240.class.getDeclaredMethod("updateHBaseConfigs");
-    Method updateKerberosDescriptorArtifacts = AbstractUpgradeCatalog.class.getDeclaredMethod("updateKerberosDescriptorArtifacts");
-    Method updateFalconConfigs = UpgradeCatalog240.class.getDeclaredMethod("updateFalconConfigs");
-    Method fixAuthorizationDescriptions = UpgradeCatalog240.class.getDeclaredMethod("fixAuthorizationDescriptions");
-    Method removeAuthorizations = UpgradeCatalog240.class.getDeclaredMethod("removeAuthorizations");
-    Method addConnectionTimeoutParamForWebAndMetricAlerts = AbstractUpgradeCatalog.class.getDeclaredMethod("addConnectionTimeoutParamForWebAndMetricAlerts");
-    Method addSliderClientConfig = UpgradeCatalog240.class.getDeclaredMethod("addSliderClientConfig");
-    Method updateRequestScheduleEntityUserIds = UpgradeCatalog240.class.getDeclaredMethod("updateRequestScheduleEntityUserIds");
-    Method updateRecoveryConfigurationDML = UpgradeCatalog240.class.getDeclaredMethod("updateRecoveryConfigurationDML");
-    Method removeAtlasMetaserverAlert = UpgradeCatalog240.class.getDeclaredMethod("removeAtlasMetaserverAlert");
-    Method updateRangerHbasePluginProperties = UpgradeCatalog240.class.getDeclaredMethod("updateRangerHbasePluginProperties");
-    Method updateKAFKAConfigs = UpgradeCatalog240.class.getDeclaredMethod("updateKAFKAConfigs");
-
-    Capture<String> capturedStatements = newCapture(CaptureType.ALL);
-
-    DBAccessor dbAccessor = createStrictMock(DBAccessor.class);
-    expect(dbAccessor.executeUpdate(capture(capturedStatements))).andReturn(1).times(7);
-
-    Capture<String> capturedTezViewUpdate = newCapture();
-    expect(dbAccessor.executeUpdate(capture(capturedTezViewUpdate))).andReturn(1).once();
-
-    Capture<String> capturedPigSmokeTestEntityUpdate = newCapture();
-    expect(dbAccessor.executeUpdate(capture(capturedPigSmokeTestEntityUpdate))).andReturn(1).once();
-
-    UpgradeCatalog240 upgradeCatalog240 = createMockBuilder(UpgradeCatalog240.class)
-            .addMockedMethod(addNewConfigurationsFromXml)
-            .addMockedMethod(updateAlerts)
-            .addMockedMethod(addSettingPermission)
-            .addMockedMethod(addViewOperationalLogsPermission)
-            .addMockedMethod(addManageUserPersistedDataPermission)
-            .addMockedMethod(allowClusterOperatorToManageCredentials)
-            .addMockedMethod(updateHDFSConfigs)
-            .addMockedMethod(updateHIVEConfigs)
-            .addMockedMethod(updateAmsConfigs)
-            .addMockedMethod(updateClusterEnv)
-            .addMockedMethod(updateSequenceForView)
-            .addMockedMethod(adjustHiveJobTimestamps)
-            .addMockedMethod(updateHostRoleCommandTableDML)
-            .addMockedMethod(updateKerberosEnv)
-            .addMockedMethod(updateYarnEnv)
-            .addMockedMethod(removeHiveOozieDBConnectionConfigs)
-            .addMockedMethod(updateClustersAndHostsVersionStateTableDML)
-            .addMockedMethod(removeStandardDeviationAlerts)
-            .addMockedMethod(consolidateUserRoles)
-            .addMockedMethod(updateClusterInheritedPermissionsConfig)
-            .addMockedMethod(createRolePrincipals)
-            .addMockedMethod(updateHDFSWidget)
-            .addMockedMethod(upgradeCapSchedulerView)
-            .addMockedMethod(updatePhoenixConfigs)
-            .addMockedMethod(updateSparkConfigs)
-            .addMockedMethod(updateKerberosDescriptorArtifacts)
-            .addMockedMethod(updateFalconConfigs)
-            .addMockedMethod(fixAuthorizationDescriptions)
-            .addMockedMethod(removeAuthorizations)
-            .addMockedMethod(addConnectionTimeoutParamForWebAndMetricAlerts)
-            .addMockedMethod(updateHBaseConfigs)
-            .addMockedMethod(addSliderClientConfig)
-            .addMockedMethod(updateRequestScheduleEntityUserIds)
-            .addMockedMethod(updateRecoveryConfigurationDML)
-            .addMockedMethod(removeAtlasMetaserverAlert)
-            .addMockedMethod(updateRangerHbasePluginProperties)
-            .addMockedMethod(updateKAFKAConfigs)
-            .createMock();
-
-    Field field = AbstractUpgradeCatalog.class.getDeclaredField("dbAccessor");
-    field.set(upgradeCatalog240, dbAccessor);
-
-    upgradeCatalog240.addNewConfigurationsFromXml();
-    upgradeCatalog240.updateAlerts();
-    upgradeCatalog240.addSettingPermission();
-    upgradeCatalog240.addViewOperationalLogsPermission();
-    upgradeCatalog240.addManageUserPersistedDataPermission();
-    upgradeCatalog240.allowClusterOperatorToManageCredentials();
-    upgradeCatalog240.updateHDFSConfigs();
-    upgradeCatalog240.updateHIVEConfigs();
-    upgradeCatalog240.updateAMSConfigs();
-    upgradeCatalog240.updateClusterEnv();
-    upgradeCatalog240.updateSequenceForView();
-    upgradeCatalog240.updateHostRoleCommandTableDML();
-    upgradeCatalog240.updateKerberosConfigs();
-    upgradeCatalog240.updateYarnEnv();
-    upgradeCatalog240.removeHiveOozieDBConnectionConfigs();
-    upgradeCatalog240.updateClustersAndHostsVersionStateTableDML();
-    upgradeCatalog240.removeStandardDeviationAlerts();
-    upgradeCatalog240.consolidateUserRoles();
-    upgradeCatalog240.createRolePrincipals();
-    upgradeCatalog240.updateClusterInheritedPermissionsConfig();
-    upgradeCatalog240.updateHDFSWidgetDefinition();
-    upgradeCatalog240.upgradeCapSchedulerView();
-    upgradeCatalog240.updatePhoenixConfigs();
-    upgradeCatalog240.updateSparkConfigs();
-    upgradeCatalog240.updateKerberosDescriptorArtifacts();
-    upgradeCatalog240.updateFalconConfigs();
-    upgradeCatalog240.fixAuthorizationDescriptions();
-    upgradeCatalog240.removeAuthorizations();
-    upgradeCatalog240.addConnectionTimeoutParamForWebAndMetricAlerts();
-    upgradeCatalog240.updateHBaseConfigs();
-    upgradeCatalog240.addSliderClientConfig();
-    upgradeCatalog240.updateRequestScheduleEntityUserIds();
-    upgradeCatalog240.updateRecoveryConfigurationDML();
-    upgradeCatalog240.removeAtlasMetaserverAlert();
-    upgradeCatalog240.updateRangerHbasePluginProperties();
-    upgradeCatalog240.adjustHiveJobTimestamps();
-    upgradeCatalog240.updateKAFKAConfigs();
-
-    replay(upgradeCatalog240, dbAccessor);
-
-    upgradeCatalog240.executeDMLUpdates();
-
-    verify(upgradeCatalog240, dbAccessor);
-
-    List<String> statements = capturedStatements.getValues();
-    Assert.assertNotNull(statements);
-    Assert.assertEquals(7, statements.size());
-    Assert.assertTrue(statements.contains("UPDATE adminpermission SET sort_order=1 WHERE permission_name='AMBARI.ADMINISTRATOR'"));
-    Assert.assertTrue(statements.contains("UPDATE adminpermission SET sort_order=2 WHERE permission_name='CLUSTER.ADMINISTRATOR'"));
-    Assert.assertTrue(statements.contains("UPDATE adminpermission SET sort_order=3 WHERE permission_name='CLUSTER.OPERATOR'"));
-    Assert.assertTrue(statements.contains("UPDATE adminpermission SET sort_order=4 WHERE permission_name='SERVICE.ADMINISTRATOR'"));
-    Assert.assertTrue(statements.contains("UPDATE adminpermission SET sort_order=5 WHERE permission_name='SERVICE.OPERATOR'"));
-    Assert.assertTrue(statements.contains("UPDATE adminpermission SET sort_order=6 WHERE permission_name='CLUSTER.USER'"));
-    Assert.assertTrue(statements.contains("UPDATE adminpermission SET sort_order=7 WHERE permission_name='VIEW.USER'"));
-
-    Assert.assertNotNull(capturedTezViewUpdate.getValue());
-    Assert.assertEquals("UPDATE viewinstanceproperty SET name = 'yarn.ats.url' where name = 'yarn.timeline-server.url'",
-      capturedTezViewUpdate.getValue());
-
-    Assert.assertNotNull(capturedPigSmokeTestEntityUpdate.getValue());
-    Assert.assertEquals("UPDATE viewentity " +
-        "SET class_name = 'org.apache.ambari.view.pig.persistence.SmokeTestEntity' " +
-        "WHERE class_name = 'org.apache.ambari.view.pig.persistence.DataStoreStorage$SmokeTestEntity'",
-      capturedPigSmokeTestEntityUpdate.getValue());
-  }
-
-  @Test
-  public void testRemoveHiveOozieDBConnectionConfigs() throws Exception{
-    EasyMockSupport easyMockSupport = new EasyMockSupport();
-    final AmbariManagementController mockAmbariManagementController = easyMockSupport.createNiceMock(
-            AmbariManagementController.class);
-    final ConfigHelper mockConfigHelper = easyMockSupport.createMock(ConfigHelper.class);
-
-    final Clusters mockClusters = easyMockSupport.createStrictMock(Clusters.class);
-    final Cluster mockClusterExpected = easyMockSupport.createNiceMock(Cluster.class);
-
-    final Config mockOozieEnv = easyMockSupport.createNiceMock(Config.class);
-    final Config mockHiveEnv = easyMockSupport.createNiceMock(Config.class);
-
-    final Map<String, String> propertiesExpectedOozieEnv = new HashMap<>();
-    propertiesExpectedOozieEnv.put("oozie_derby_database", "Derby");
-    propertiesExpectedOozieEnv.put("property", "value");
-    // Imitate missing property
-    // propertiesExpectedOozieEnv.put("oozie_hostname", "hostname");
-    final Map<String, String> propertiesExpectedHiveEnv = new HashMap<>();
-    propertiesExpectedHiveEnv.put("hive_hostname", "hostname");
-
-    final Injector mockInjector = Guice.createInjector(new Module() {
-      @Override
-      public void configure(Binder binder) {
-        binder.bind(AmbariManagementController.class).toInstance(mockAmbariManagementController);
-        binder.bind(ConfigHelper.class).toInstance(mockConfigHelper);
-        binder.bind(Clusters.class).toInstance(mockClusters);
-        binder.bind(EntityManager.class).toInstance(entityManager);
-        binder.bind(DBAccessor.class).toInstance(createNiceMock(DBAccessor.class));
-        binder.bind(OsFamily.class).toInstance(createNiceMock(OsFamily.class));
-        binder.bind(PasswordEncoder.class).toInstance(createNiceMock(PasswordEncoder.class));
-        binder.bind(HookContextFactory.class).toInstance(createMock(HookContextFactory.class));
-        binder.bind(HookService.class).toInstance(createMock(HookService.class));
-      }
-    });
-
-    expect(mockAmbariManagementController.getClusters()).andReturn(mockClusters).once();
-    expect(mockClusters.getClusters()).andReturn(new HashMap<String, Cluster>() {{
-      put("normal", mockClusterExpected);
-    }}).once();
-
-    expect(mockClusterExpected.getDesiredConfigByType("oozie-env")).andReturn(mockOozieEnv).atLeastOnce();
-    expect(mockClusterExpected.getDesiredConfigByType("hive-env")).andReturn(mockHiveEnv).atLeastOnce();
-    expect(mockOozieEnv.getProperties()).andReturn(propertiesExpectedOozieEnv).anyTimes();
-    expect(mockHiveEnv.getProperties()).andReturn(propertiesExpectedHiveEnv).anyTimes();
-
-    Capture<Map<String, String>> oozieCapture =  newCapture();
-    Capture<Map<String, String>> hiveCapture =  newCapture();
-    expect(mockAmbariManagementController.createConfig(eq(mockClusterExpected), anyObject(StackId.class), eq("oozie-env"),
-        capture(oozieCapture), anyString(), EasyMock.<Map<String, Map<String, String>>>anyObject())).andReturn(null).once();
-    expect(mockAmbariManagementController.createConfig(eq(mockClusterExpected), anyObject(StackId.class), eq("hive-env"),
-            capture(hiveCapture), anyString(), EasyMock.<Map<String, Map<String, String>>>anyObject())).andReturn(null).once();
-
-    easyMockSupport.replayAll();
-    mockInjector.getInstance(UpgradeCatalog240.class).removeHiveOozieDBConnectionConfigs();
-    easyMockSupport.verifyAll();
-
-    assertEquals("value", oozieCapture.getValue().get("property"));
-    assertNull(oozieCapture.getValue().get("oozie_derby_database"));
-    assertNull(oozieCapture.getValue().get("oozie_hostname"));
-    assertNull(hiveCapture.getValue().get("hive_hostname"));
-  }
-
-  @Test
-  public void testUpdateFalconConfigs() throws Exception{
-    EasyMockSupport easyMockSupport = new EasyMockSupport();
-    final AmbariManagementController mockAmbariManagementController = easyMockSupport.createNiceMock(
-            AmbariManagementController.class);
-    final ConfigHelper mockConfigHelper = easyMockSupport.createMock(ConfigHelper.class);
-
-    final Clusters mockClusters = easyMockSupport.createStrictMock(Clusters.class);
-    final Cluster mockClusterExpected = easyMockSupport.createNiceMock(Cluster.class);
-
-    final Config mockFalconEnv = easyMockSupport.createNiceMock(Config.class);
-
-    final Map<String, String> propertiesExpectedFalconEnv = new HashMap<>();
-    propertiesExpectedFalconEnv.put("falcon_store_uri", "file:///hadoop/falcon/store");
-    propertiesExpectedFalconEnv.put("content", "# content");
-    propertiesExpectedFalconEnv.put("property", "value");
-
-    final String applicationServicesOldPropertyValue =
-        "org.apache.falcon.security.AuthenticationInitializationService,\\\n" +
-        "      org.apache.falcon.workflow.WorkflowJobEndNotificationService, \\\n" +
-        "      org.apache.falcon.service.ProcessSubscriberService,\\\n" +
-        "      org.apache.falcon.entity.store.ConfigurationStore,\\\n" +
-        "      org.apache.falcon.rerun.service.RetryService,\\\n" +
-        "      org.apache.falcon.rerun.service.LateRunService,\\\n" +
-        "      org.apache.falcon.service.LogCleanupService,\\\n" +
-        "      org.apache.falcon.metadata.MetadataMappingService";
-
-    final String applicationServicesExpectedPropertyValue =
-        "org.apache.falcon.security.AuthenticationInitializationService,\\\n" +
-        "      org.apache.falcon.workflow.WorkflowJobEndNotificationService, \\\n" +
-        "      org.apache.falcon.service.ProcessSubscriberService,\\\n" +
-        "      org.apache.falcon.entity.store.ConfigurationStore,\\\n" +
-        "      org.apache.falcon.rerun.service.RetryService,\\\n" +
-        "      org.apache.falcon.rerun.service.LateRunService,\\\n" +
-        "      org.apache.falcon.service.LogCleanupService,\\\n" +
-        "      org.apache.falcon.metadata.MetadataMappingService{{atlas_application_class_addition}}";
-
-    final Config falconStartupConfig = easyMockSupport.createNiceMock(Config.class);
-
-    final Map<String, String> falconStartupConfigProperties= new HashMap<>();
-    falconStartupConfigProperties.put("*.application.services", applicationServicesOldPropertyValue);
-    falconStartupConfigProperties.put("property", "value");
-    final Injector mockInjector = Guice.createInjector(new Module() {
-      @Override
-      public void configure(Binder binder) {
-        binder.bind(AmbariManagementController.class).toInstance(mockAmbariManagementController);
-        binder.bind(ConfigHelper.class).toInstance(mockConfigHelper);
-        binder.bind(Clusters.class).toInstance(mockClusters);
-        binder.bind(EntityManager.class).toInstance(entityManager);
-        binder.bind(DBAccessor.class).toInstance(createNiceMock(DBAccessor.class));
-        binder.bind(OsFamily.class).toInstance(createNiceMock(OsFamily.class));
-        binder.bind(PasswordEncoder.class).toInstance(createNiceMock(PasswordEncoder.class));
-        binder.bind(HookContextFactory.class).toInstance(createMock(HookContextFactory.class));
-        binder.bind(HookService.class).toInstance(createMock(HookService.class));
-      }
-    });
-
-    expect(mockAmbariManagementController.getClusters()).andReturn(mockClusters).once();
-    expect(mockClusters.getClusters()).andReturn(new HashMap<String, Cluster>() {{
-      put("normal", mockClusterExpected);
-    }}).once();
-
-    expect(mockClusterExpected.getDesiredConfigByType("falcon-env")).andReturn(mockFalconEnv).atLeastOnce();
-    expect(mockFalconEnv.getProperties()).andReturn(propertiesExpectedFalconEnv).anyTimes();
-
-    expect(mockClusterExpected.getDesiredConfigByType("falcon-startup.properties")).andReturn(falconStartupConfig).atLeastOnce();
-    expect(falconStartupConfig.getProperties()).andReturn(falconStartupConfigProperties).anyTimes();
-
-    Capture<Map<String, String>> falconCapture =  newCapture();
-    expect(mockAmbariManagementController.createConfig(eq(mockClusterExpected), anyObject(StackId.class),  eq("falcon-env"),
-        capture(falconCapture), anyString(), EasyMock.<Map<String, Map<String, String>>>anyObject())).andReturn(null).once();
-
-    Capture<Map<String, String>> falconCapture2 =  newCapture();
-    expect(mockAmbariManagementController.createConfig(eq(mockClusterExpected), anyObject(StackId.class), eq("falcon-env"),
-        capture(falconCapture2), anyString(), EasyMock.<Map<String, Map<String, String>>>anyObject())).andReturn(null).once();
-
-    Capture<Map<String, String>> falconStartupCapture =  newCapture();
-    expect(mockAmbariManagementController.createConfig(eq(mockClusterExpected), anyObject(StackId.class), eq("falcon-startup.properties"),
-        capture(falconStartupCapture), anyString(), EasyMock.<Map<String, Map<String, String>>>anyObject())).andReturn(null).once();
-
-    easyMockSupport.replayAll();
-    mockInjector.getInstance(UpgradeCatalog240.class).updateFalconConfigs();
-    easyMockSupport.verifyAll();
-
-    final String expectredEnvContent = "# content\n" +
-                                       "\n" +
-                                       "{% if falcon_atlas_support %}\n" +
-                                       "# Add the Atlas Falcon hook to the Falcon classpath\n" +
-                                       "export FALCON_EXTRA_CLASS_PATH={{atlas_hook_cp}}${FALCON_EXTRA_CLASS_PATH}\n" +
-                                       "{% endif %}";
-
-    assertEquals("value", falconCapture.getValue().get("property"));
-    assertEquals("# content", falconCapture.getValue().get("content"));
-    assertNull(falconCapture.getValue().get("falcon_store_uri"));
-
-    assertEquals(expectredEnvContent, falconCapture2.getValue().get("content"));
-
-    assertEquals("value", falconStartupCapture.getValue().get("property"));
-    assertEquals(applicationServicesExpectedPropertyValue, falconStartupCapture.getValue().get("*.application.services"));
-  }
-
-  @Test
-  public void testUpdateHbaseConfigs() throws Exception{
-    EasyMockSupport easyMockSupport = new EasyMockSupport();
-    final AmbariManagementController mockAmbariManagementController = easyMockSupport.createNiceMock(
-        AmbariManagementController.class);
-    final ConfigHelper mockConfigHelper = easyMockSupport.createMock(ConfigHelper.class);
-
-    final Clusters mockClusters = easyMockSupport.createStrictMock(Clusters.class);
-    final Cluster mockCluster = easyMockSupport.createNiceMock(Cluster.class);
-
-    final Injector mockInjector = Guice.createInjector(new Module() {
-      @Override
-      public void configure(Binder binder) {
-        binder.bind(AmbariManagementController.class).toInstance(mockAmbariManagementController);
-        binder.bind(ConfigHelper.class).toInstance(mockConfigHelper);
-        binder.bind(Clusters.class).toInstance(mockClusters);
-        binder.bind(EntityManager.class).toInstance(entityManager);
-        binder.bind(DBAccessor.class).toInstance(createNiceMock(DBAccessor.class));
-        binder.bind(OsFamily.class).toInstance(createNiceMock(OsFamily.class));
-        binder.bind(PasswordEncoder.class).toInstance(createNiceMock(PasswordEncoder.class));
-        binder.bind(HookContextFactory.class).toInstance(createMock(HookContextFactory.class));
-        binder.bind(HookService.class).toInstance(createMock(HookService.class));
-
-      }
-    });
-
-    expect(mockAmbariManagementController.getClusters()).andReturn(mockClusters).anyTimes();
-    expect(mockClusters.getClusters()).andReturn(new HashMap<String, Cluster>() {{
-      put("normal", mockCluster);
-    }}).anyTimes();
-
-    final Service hbaseService = createNiceMock(Service.class);
-    expect(hbaseService.getDesiredStackId()).andReturn(new StackId("HDP-2.4"));
-
-    expect(mockCluster.getServices()).andReturn(new HashMap<String, Service>(){{put("HBASE",hbaseService);}}).anyTimes();
-    expect(mockCluster.getSecurityType()).andReturn(SecurityType.KERBEROS).anyTimes();
-
-    final Config mockHbaseSiteConfigs = easyMockSupport.createNiceMock(Config.class);
-    expect(mockCluster.getDesiredConfigByType("hbase-site")).andReturn(mockHbaseSiteConfigs).atLeastOnce();
-    expect(mockHbaseSiteConfigs.getProperties()).andReturn(new HashMap<String, String>(){{
-      put("hbase.coprocessor.regionserver.classes","{{hbase_coprocessor_regionserver_classes}}");
-      put("hbase.coprocessor.region.classes","{{hbase_coprocessor_region_classes}}");
-    }}).anyTimes();
-
-    final Config mockRangerHbaseConfigs = easyMockSupport.createNiceMock(Config.class);
-    expect(mockCluster.getDesiredConfigByType("ranger-hbase-plugin-properties")).andReturn(mockRangerHbaseConfigs).atLeastOnce();
-    expect(mockRangerHbaseConfigs.getProperties()).andReturn(new HashMap<String, String>(){{
-      put("ranger-hbase-plugin-enabled", "yes");
-    }}).anyTimes();
-
-    final Config mockRangerEnvConfigs = easyMockSupport.createNiceMock(Config.class);
-    expect(mockCluster.getDesiredConfigByType("ranger-env")).andReturn(mockRangerEnvConfigs).atLeastOnce();
-    expect(mockRangerEnvConfigs.getProperties()).andReturn(new HashMap<String, String>(){{
-      put("xml_configurations_supported", "true");
-    }}).anyTimes();
-
-
-    Capture<Map<String, String>> hbaseCapture =  newCapture();
-    expect(mockAmbariManagementController.createConfig(eq(mockCluster), anyObject(StackId.class), eq("hbase-site"),
-        capture(hbaseCapture), anyString(), EasyMock.<Map<String, Map<String, String>>>anyObject())).andReturn(null).once();
-
-    easyMockSupport.replayAll();
-    mockInjector.getInstance(UpgradeCatalog240.class).updateHBaseConfigs();
-    easyMockSupport.verifyAll();
-
-    assertEquals("org.apache.ranger.authorization.hbase.RangerAuthorizationCoprocessor", hbaseCapture.getValue().get("hbase.coprocessor.regionserver.classes"));
-    assertEquals("org.apache.hadoop.hbase.security.token.TokenProvider,org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint,org.apache.ranger.authorization.hbase.RangerAuthorizationCoprocessor", hbaseCapture.getValue().get("hbase.coprocessor.region.classes"));
-  }
-
-  @Test
-  public void test_addParam_ParamsNotAvailable() {
-
-    UpgradeCatalog240 upgradeCatalog240 = new UpgradeCatalog240(injector);
-    String inputSource = "{ \"path\" : \"test_path\", \"type\" : \"SCRIPT\"}";
-    List<String> params = Arrays.asList("connection.timeout", "checkpoint.time.warning.threshold", "checkpoint.time.critical.threshold");
-    String expectedSource = "{\"path\":\"test_path\",\"type\":\"SCRIPT\",\"parameters\":[{\"name\":\"connection.timeout\",\"display_name\":\"Connection Timeout\",\"value\":5.0,\"type\":\"NUMERIC\",\"description\":\"The maximum time before this alert is considered to be CRITICAL\",\"units\":\"seconds\",\"threshold\":\"CRITICAL\"},{\"name\":\"checkpoint.time.warning.threshold\",\"display_name\":\"Checkpoint Warning\",\"value\":2.0,\"type\":\"PERCENT\",\"description\":\"The percentage of the last checkpoint time greater than the interval in order to trigger a warning alert.\",\"units\":\"%\",\"threshold\":\"WARNING\"},{\"name\":\"checkpoint.time.critical.threshold\",\"display_name\":\"Checkpoint Critical\",\"value\":4.0,\"type\":\"PERCENT\",\"description\":\"The percentage of the last checkpoint time greater than the interval in order to trigger a critical alert.\",\"units\":\"%\",\"threshold\":\"CRITICAL\"}]}";
-
-    String result = upgradeCatalog240.addParam(inputSource, params);
-    Assert.assertEquals(result, expectedSource);
-  }
-
-  @Test
-  public void test_addParam_ParamsAvailableWithOneOFNeededItem() {
-
-    UpgradeCatalog240 upgradeCatalog240 = new UpgradeCatalog240(injector);
-    String inputSource = "{\"path\":\"test_path\",\"type\":\"SCRIPT\",\"parameters\":[{\"name\":\"connection.timeout\",\"display_name\":\"Connection Timeout\",\"value\":5.0,\"type\":\"NUMERIC\",\"description\":\"The maximum time before this alert is considered to be CRITICAL\",\"units\":\"seconds\",\"threshold\":\"CRITICAL\"}]}";
-    List<String> params = new ArrayList<>(Arrays.asList("connection.timeout", "checkpoint.time.warning.threshold", "checkpoint.time.critical.threshold"));
-    String expectedSource = "{\"path\":\"test_path\",\"type\":\"SCRIPT\",\"parameters\":[{\"name\":\"connection.timeout\",\"display_name\":\"Connection Timeout\",\"value\":5.0,\"type\":\"NUMERIC\",\"description\":\"The maximum time before this alert is considered to be CRITICAL\",\"units\":\"seconds\",\"threshold\":\"CRITICAL\"},{\"name\":\"checkpoint.time.warning.threshold\",\"display_name\":\"Checkpoint Warning\",\"value\":2.0,\"type\":\"PERCENT\",\"description\":\"The percentage of the last checkpoint time greater than the interval in order to trigger a warning alert.\",\"units\":\"%\",\"threshold\":\"WARNING\"},{\"name\":\"checkpoint.time.critical.threshold\",\"display_name\":\"Checkpoint Critical\",\"value\":4.0,\"type\":\"PERCENT\",\"description\":\"The percentage of the last checkpoint time greater than the interval in order to trigger a critical alert.\",\"units\":\"%\",\"threshold\":\"CRITICAL\"}]}";
-
-    String result = upgradeCatalog240.addParam(inputSource, params);
-    Assert.assertEquals(result, expectedSource);
-  }
-
-  /**
-   * Test that dfs.internal.nameservices is not affected
-   * Also, dfs.client.retry.policy.enabled is reset
-   * @throws Exception
-   */
-  @Test
-  public void testHdfsSiteUpdateConfigs2() throws Exception{
-    Map<String, String> oldPropertiesHdfsSite = new HashMap<String, String>() {
-      {
-        put("dfs.client.retry.policy.enabled", "true");
-      }
-    };
-    Map<String, String> newPropertiesHdfsSite = new HashMap<String, String>() {
-      {
-        put("dfs.client.retry.policy.enabled", "false");
-      }
-    };
-    EasyMockSupport easyMockSupport = new EasyMockSupport();
-
-    Clusters clusters = easyMockSupport.createNiceMock(Clusters.class);
-    final Cluster cluster = easyMockSupport.createNiceMock(Cluster.class);
-    final Service service = createStrictMock(Service.class);
-    final Map<String, Service> services = Collections.singletonMap("HDFS", service);
-    Config mockHdfsSite = easyMockSupport.createNiceMock(Config.class);
-
-    expect(clusters.getClusters()).andReturn(new HashMap<String, Cluster>() {{
-      put("normal", cluster);
-    }}).anyTimes();
-    expect(cluster.getDesiredConfigByType("hdfs-site")).andReturn(mockHdfsSite).atLeastOnce();
-    expect(mockHdfsSite.getProperties()).andReturn(oldPropertiesHdfsSite).anyTimes();
-    expect(cluster.getServices()).andReturn(services).once();
-
-    Injector injector = easyMockSupport.createNiceMock(Injector.class);
-    expect(injector.getInstance(Gson.class)).andReturn(null).anyTimes();
-    expect(injector.getInstance(MaintenanceStateHelper.class)).andReturn(null).anyTimes();
-    expect(injector.getInstance(KerberosHelper.class)).andReturn(createNiceMock(KerberosHelper.class)).anyTimes();
-
-    replay(injector, clusters, mockHdfsSite, cluster);
-
-    AmbariManagementControllerImpl controller = createMockBuilder(AmbariManagementControllerImpl.class)
-        .addMockedMethod("createConfiguration")
-        .addMockedMethod("getClusters", new Class[] { })
-        .addMockedMethod("createConfig")
-        .withConstructor(createNiceMock(ActionManager.class), clusters, injector)
-        .createNiceMock();
-
-    Injector injector2 = easyMockSupport.createNiceMock(Injector.class);
-    Capture<Map<String, String>> propertiesCapture = EasyMock.newCapture();
-
-    expect(injector2.getInstance(AmbariManagementController.class)).andReturn(controller).anyTimes();
-    expect(controller.getClusters()).andReturn(clusters).anyTimes();
-    expect(controller.createConfig(anyObject(Cluster.class), anyObject(StackId.class), anyString(), capture(propertiesCapture), anyString(),
-                                   EasyMock.<Map<String, Map<String, String>>>anyObject())).andReturn(createNiceMock(Config.class)).once();
-
-    replay(controller, injector2);
-    new UpgradeCatalog240(injector2).updateHDFSConfigs();
-    easyMockSupport.verifyAll();
-
-    Map<String, String> updatedProperties = propertiesCapture.getValue();
-    assertTrue(Maps.difference(newPropertiesHdfsSite, updatedProperties).areEqual());
-  }
-
-  /**
-   * Test that dfs.nameservices is copied over to dfs.internal.nameservices
-   * @throws Exception
-   */
-  @Test
-  public void testHdfsSiteUpdateConfigs() throws Exception{
-    Map<String, String> oldPropertiesHdfsSite = new HashMap<String, String>() {
-      {
-        put("dfs.nameservices", "nnha");
-      }
-    };
-    Map<String, String> newPropertiesHdfsSite = new HashMap<String, String>() {
-      {
-        put("dfs.nameservices", "nnha");
-        put("dfs.internal.nameservices", "nnha");
-      }
-    };
-    Map<String, String> oldPropertiesHadoopEnv = new HashMap<String, String>() {
-      {
-        put("keyserver_port", " ");
-      }
-    };
-    Map<String, String> newPropertiesHadoopEnv = new HashMap<String, String>() {
-      {
-        put("keyserver_port", "");
-      }
-    };
-    EasyMockSupport easyMockSupport = new EasyMockSupport();
-
-    Clusters clusters = easyMockSupport.createNiceMock(Clusters.class);
-    final Cluster cluster = easyMockSupport.createNiceMock(Cluster.class);
-    final Service service = createStrictMock(Service.class);
-    final Map<String, Service> services = Collections.singletonMap("HDFS", service);
-    Config mockHdfsSite = easyMockSupport.createNiceMock(Config.class);
-    Config mockHadoopEnv = easyMockSupport.createNiceMock(Config.class);
-
-    expect(clusters.getClusters()).andReturn(new HashMap<String, Cluster>() {{
-      put("normal", cluster);
-    }}).anyTimes();
-    expect(cluster.getDesiredConfigByType("hdfs-site")).andReturn(mockHdfsSite).atLeastOnce();
-    expect(cluster.getDesiredConfigByType("hadoop-env")).andReturn(mockHadoopEnv).atLeastOnce();
-    expect(mockHdfsSite.getProperties()).andReturn(oldPropertiesHdfsSite).anyTimes();
-    expect(mockHadoopEnv.getProperties()).andReturn(oldPropertiesHadoopEnv).anyTimes();
-    expect(cluster.getServices()).andReturn(services).once();
-
-    Injector injector = easyMockSupport.createNiceMock(Injector.class);
-    expect(injector.getInstance(Gson.class)).andReturn(null).anyTimes();
-    expect(injector.getInstance(MaintenanceStateHelper.class)).andReturn(null).anyTimes();
-    expect(injector.getInstance(KerberosHelper.class)).andReturn(createNiceMock(KerberosHelper.class)).anyTimes();
-
-    replay(injector, clusters, mockHdfsSite, mockHadoopEnv, cluster);
-
-    AmbariManagementControllerImpl controller = createMockBuilder(AmbariManagementControllerImpl.class)
-        .addMockedMethod("createConfiguration")
-        .addMockedMethod("getClusters", new Class[] { })
-        .addMockedMethod("createConfig")
-        .withConstructor(createNiceMock(ActionManager.class), clusters, injector)
-        .createNiceMock();
-
-    Injector injector2 = easyMockSupport.createNiceMock(Injector.class);
-    Capture<Map<String, String>> propertiesCaptureHdfsSite = EasyMock.newCapture();
-    Capture<Map<String, String>> propertiesCaptureHadoopEnv = EasyMock.newCapture();
-
-    expect(injector2.getInstance(AmbariManagementController.class)).andReturn(controller).anyTimes();
-    expect(controller.getClusters()).andReturn(clusters).anyTimes();
-    expect(controller.createConfig(anyObject(Cluster.class), anyObject(StackId.class), eq("hdfs-site"), capture(propertiesCaptureHdfsSite), anyString(),
-                                   EasyMock.<Map<String, Map<String, String>>>anyObject())).andReturn(createNiceMock(Config.class)).once();
-    expect(controller.createConfig(anyObject(Cluster.class), anyObject(StackId.class), eq("hadoop-env"), capture(propertiesCaptureHadoopEnv), anyString(),
-        EasyMock.<Map<String, Map<String, String>>>anyObject())).andReturn(createNiceMock(Config.class)).once();
-
-    replay(controller, injector2);
-    new UpgradeCatalog240(injector2).updateHDFSConfigs();
-    easyMockSupport.verifyAll();
-
-    Map<String, String> updatedPropertiesHdfsSite = propertiesCaptureHdfsSite.getValue();
-    assertTrue(Maps.difference(newPropertiesHdfsSite, updatedPropertiesHdfsSite).areEqual());
-
-    Map<String, String> updatedPropertiesHadoopEnv = propertiesCaptureHadoopEnv.getValue();
-    assertTrue(Maps.difference(newPropertiesHadoopEnv, updatedPropertiesHadoopEnv).areEqual());
-  }
-
-  @Test
-  public void testYarnEnvUpdateConfigs() throws Exception{
-
-    Map<String, String> oldPropertiesYarnEnv = new HashMap<String, String>() {
-      {
-        put("content", "export YARN_HISTORYSERVER_HEAPSIZE={{apptimelineserver_heapsize}}");
-      }
-    };
-    Map<String, String> newPropertiesYarnEnv = new HashMap<String, String>() {
-      {
-        put("content", "# export YARN_HISTORYSERVER_HEAPSIZE={{apptimelineserver_heapsize}}" +
-                "\n\n      # Specify the max Heapsize for the timeline server using a numerical value\n" +
-                "      # in the scale of MB. For example, to specify an jvm option of -Xmx1000m, set\n" +
-                "      # the value to 1024.\n" +
-                "      # This value will be overridden by an Xmx setting specified in either YARN_OPTS\n" +
-                "      # and/or YARN_TIMELINESERVER_OPTS.\n" +
-                "      # If not specified, the default value will be picked from either YARN_HEAPMAX\n" +
-                "      # or JAVA_HEAP_MAX with YARN_HEAPMAX as the preferred option of the two.\n" +
-                "      export YARN_TIMELINESERVER_HEAPSIZE={{apptimelineserver_heapsize}}");
-      }
-    };
-    EasyMockSupport easyMockSupport = new EasyMockSupport();
-
-    Clusters clusters = easyMockSupport.createNiceMock(Clusters.class);
-    final Cluster cluster = easyMockSupport.createNiceMock(Cluster.class);
-    Config mockYarnEnv = easyMockSupport.createNiceMock(Config.class);
-
-    expect(clusters.getClusters()).andReturn(new HashMap<String, Cluster>() {{
-      put("normal", cluster);
-    }}).once();
-    expect(cluster.getDesiredConfigByType("yarn-env")).andReturn(mockYarnEnv).atLeastOnce();
-    expect(mockYarnEnv.getProperties()).andReturn(oldPropertiesYarnEnv).anyTimes();
-
-    Injector injector = easyMockSupport.createNiceMock(Injector.class);
-    expect(injector.getInstance(Gson.class)).andReturn(null).anyTimes();
-    expect(injector.getInstance(MaintenanceStateHelper.class)).andReturn(null).anyTimes();
-    expect(injector.getInstance(KerberosHelper.class)).andReturn(createNiceMock(KerberosHelper.class)).anyTimes();
-
-    replay(injector, clusters, mockYarnEnv, cluster);
-
-    AmbariManagementControllerImpl controller = createMockBuilder(AmbariManagementControllerImpl.class)
-            .addMockedMethod("createConfiguration")
-            .addMockedMethod("getClusters", new Class[] { })
-            .addMockedMethod("createConfig")
-            .withConstructor(createNiceMock(ActionManager.class), clusters, injector)
-            .createNiceMock();
-
-    Injector injector2 = easyMockSupport.createNiceMock(Injector.class);
-    Capture<Map<String, String>> propertiesCapture = EasyMock.newCapture();
-
-    expect(injector2.getInstance(AmbariManagementController.class)).andReturn(controller).anyTimes();
-    expect(controller.getClusters()).andReturn(clusters).anyTimes();
-    expect(controller.createConfig(anyObject(Cluster.class), anyObject(StackId.class), anyString(), capture(propertiesCapture), anyString(),
-            EasyMock.<Map<String, Map<String, String>>>anyObject())).andReturn(createNiceMock(Config.class)).once();
-
-    replay(controller, injector2);
-    new UpgradeCatalog240(injector2).updateYarnEnv();
-    easyMockSupport.verifyAll();
-
-    Map<String, String> updatedProperties = propertiesCapture.getValue();
-    assertTrue(Maps.difference(newPropertiesYarnEnv, updatedProperties).areEqual());
-  }
-
-  @Test
-  public void testUpdateKAFKAConfigs() throws Exception{
-    EasyMockSupport easyMockSupport = new EasyMockSupport();
-    final AmbariManagementController mockAmbariManagementController = easyMockSupport.createNiceMock(AmbariManagementController.class);
-    final Clusters mockClusters = easyMockSupport.createStrictMock(Clusters.class);
-    final Cluster mockClusterExpected = easyMockSupport.createNiceMock(Cluster.class);
-
-    final Config kafkaBroker = easyMockSupport.createNiceMock(Config.class);
-    expect(kafkaBroker.getProperties()).andReturn(new HashMap<String, String>(){{
-      put("listeners", "PLAINTEXT://localhost:6667,SSL://localhost:6666");
-    }}
-    ).anyTimes();
-
-    final Injector mockInjector = Guice.createInjector(new AbstractModule() {
-      @Override
-      protected void configure() {
-        bind(AmbariManagementController.class).toInstance(mockAmbariManagementController);
-        bind(Clusters.class).toInstance(mockClusters);
-        bind(EntityManager.class).toInstance(entityManager);
-        bind(DBAccessor.class).toInstance(createNiceMock(DBAccessor.class));
-        bind(OsFamily.class).toInstance(createNiceMock(OsFamily.class));
-        bind(PasswordEncoder.class).toInstance(createNiceMock(PasswordEncoder.class));
-        bind(HookContextFactory.class).toInstance(createMock(HookContextFactory.class));
-        bind(HookService.class).toInstance(createMock(HookService.class));
-      }
-    });
-
-    expect(mockAmbariManagementController.getClusters()).andReturn(mockClusters).once();
-    expect(mockClusters.getClusters()).andReturn(new HashMap<String, Cluster>() {{
-      put("normal", mockClusterExpected);
-    }}).atLeastOnce();
-    expect(mockClusterExpected.getDesiredConfigByType("kafka-broker")).andReturn(kafkaBroker).atLeastOnce();
-    expect(mockClusterExpected.getSecurityType()).andReturn(SecurityType.KERBEROS);
-    expect(mockClusterExpected.getServices()).andReturn(new HashMap<String, Service>() {
-      {
-        put("KAFKA", null);
-      }
-    }).atLeastOnce();
-
-    UpgradeCatalog240 upgradeCatalog240 = createMockBuilder(UpgradeCatalog240.class)
-            .withConstructor(Injector.class)
-            .withArgs(mockInjector)
-            .addMockedMethod("updateConfigurationProperties", String.class,
-                    Map.class, boolean.class, boolean.class)
-            .createMock();
-
-    Map<String, String> expectedUpdates = new HashMap<>();
-    expectedUpdates.put("listeners", "PLAINTEXTSASL://localhost:6667,SSL://localhost:6666");
-
-    upgradeCatalog240.updateConfigurationProperties("kafka-broker", expectedUpdates,
-            true, false);
-    expectLastCall().once();
-
-    easyMockSupport.replayAll();
-    replay(upgradeCatalog240);
-    upgradeCatalog240.updateKAFKAConfigs();
-    easyMockSupport.verifyAll();
-  }
-
-
-  @Test
-  public void testSparkConfigUpdate() throws Exception{
-
-    Map<String, String> oldPropertiesSparkDefaults = new HashMap<String, String>() {
-      {
-        put("spark.driver.extraJavaOptions", "-Dhdp.version={{hdp_full_version}}");
-        put("spark.yarn.am.extraJavaOptions", "-Dhdp.version={{hdp_full_version}}");
-      }
-    };
-    Map<String, String> newPropertiesSparkDefaults = new HashMap<String, String>() {
-      {
-        put("spark.driver.extraJavaOptions", "-Dhdp.version={{full_stack_version}}");
-        put("spark.yarn.am.extraJavaOptions", "-Dhdp.version={{full_stack_version}}");
-      }
-    };
-
-    Map<String, String> oldPropertiesSparkJavaOpts = new HashMap<String, String>() {
-      {
-        put("content", "-Dhdp.version={{hdp_full_version}}");
-      }
-    };
-    Map<String, String> newPropertiesSparkJavaOpts = new HashMap<String, String>() {
-      {
-        put("content", "-Dhdp.version={{full_stack_version}}");
-      }
-    };
-    EasyMockSupport easyMockSupport = new EasyMockSupport();
-
-    Clusters clusters = easyMockSupport.createNiceMock(Clusters.class);
-    final Cluster cluster = easyMockSupport.createNiceMock(Cluster.class);
-    Config mockSparkDefaults = easyMockSupport.createNiceMock(Config.class);
-    Config mockSparkJavaOpts = easyMockSupport.createNiceMock(Config.class);
-
-    expect(clusters.getClusters()).andReturn(new HashMap<String, Cluster>() {{
-      put("normal", cluster);
-    }}).once();
-    expect(cluster.getDesiredConfigByType("spark-defaults")).andReturn(mockSparkDefaults).atLeastOnce();
-    expect(mockSparkDefaults.getProperties()).andReturn(oldPropertiesSparkDefaults).anyTimes();
-    expect(cluster.getDesiredConfigByType("spark-javaopts-properties")).andReturn(mockSparkJavaOpts).atLeastOnce();
-    expect(mockSparkJavaOpts.getProperties()).andReturn(oldPropertiesSparkJavaOpts).anyTimes();
-
-    Injector injector = easyMockSupport.createNiceMock(Injector.class);
-    expect(injector.getInstance(Gson.class)).andReturn(null).anyTimes();
-    expect(injector.getInstance(MaintenanceStateHelper.class)).andReturn(null).anyTimes();
-    expect(injector.getInstance(KerberosHelper.class)).andReturn(createNiceMock(KerberosHelper.class)).anyTimes();
-
-    replay(injector, clusters, mockSparkDefaults, mockSparkJavaOpts, cluster);
-
-    AmbariManagementControllerImpl controller = createMockBuilder(AmbariManagementControllerImpl.class)
-        .addMockedMethod("createConfiguration")
-        .addMockedMethod("getClusters", new Class[] { })
-        .addMockedMethod("createConfig")
-        .withConstructor(createNiceMock(ActionManager.class), clusters, injector)
-        .createNiceMock();
-
-    Injector injector2 = easyMockSupport.createNiceMock(Injector.class);
-    Capture<Map<String, String>> propertiesSparkDefaultsCapture = EasyMock.newCapture();
-    Capture<Map<String, String>> propertiesSparkJavaOptsCapture = EasyMock.newCapture();
-
-    expect(injector2.getInstance(AmbariManagementController.class)).andReturn(controller).anyTimes();
-    expect(controller.getClusters()).andReturn(clusters).anyTimes();
-    expect(controller.createConfig(anyObject(Cluster.class), anyObject(StackId.class), eq("spark-defaults"), capture(propertiesSparkDefaultsCapture), anyString(),
-        EasyMock.<Map<String, Map<String, String>>>anyObject())).andReturn(createNiceMock(Config.class)).once();
-    expect(controller.createConfig(anyObject(Cluster.class), anyObject(StackId.class), eq("spark-javaopts-properties"), capture(propertiesSparkJavaOptsCapture), anyString(),
-        EasyMock.<Map<String, Map<String, String>>>anyObject())).andReturn(createNiceMock(Config.class)).once();
-
-    replay(controller, injector2);
-    new UpgradeCatalog240(injector2).updateSparkConfigs();
-    easyMockSupport.verifyAll();
-
-    Map<String, String> updatedSparkDefaultsProperties = propertiesSparkDefaultsCapture.getValue();
-    assertTrue(Maps.difference(newPropertiesSparkDefaults, updatedSparkDefaultsProperties).areEqual());
-
-    Map<String, String> updatedSparkJavaOptsProperties = propertiesSparkJavaOptsCapture.getValue();
-    assertTrue(Maps.difference(newPropertiesSparkJavaOpts, updatedSparkJavaOptsProperties).areEqual());
-  }
-
-  @Test
-  public void testAmsHbaseEnvUpdateConfigs() throws Exception{
-
-    Map<String, String> oldPropertiesAmsHbaseEnv = new HashMap<String, String>() {
-      {
-        put("content", "some_content");
-      }
-    };
-    Map<String, String> newPropertiesAmsHbaseEnv = new HashMap<String, String>() {
-      {
-        put("content", "some_content"+ "\n # Explicitly Setting HBASE_HOME for AMS HBase so that there is no conflict\n" +
-          "export HBASE_HOME={{ams_hbase_home_dir}}\n");
-      }
-    };
-    EasyMockSupport easyMockSupport = new EasyMockSupport();
-
-    Clusters clusters = easyMockSupport.createNiceMock(Clusters.class);
-    final Cluster cluster = easyMockSupport.createNiceMock(Cluster.class);
-    Config mockAmsHbaseEnv = easyMockSupport.createNiceMock(Config.class);
-
-    expect(clusters.getClusters()).andReturn(new HashMap<String, Cluster>() {{
-      put("normal", cluster);
-    }}).once();
-    expect(cluster.getDesiredConfigByType("ams-hbase-env")).andReturn(mockAmsHbaseEnv).atLeastOnce();
-    expect(mockAmsHbaseEnv.getProperties()).andReturn(oldPropertiesAmsHbaseEnv).anyTimes();
-
-    Injector injector = easyMockSupport.createNiceMock(Injector.class);
-    expect(injector.getInstance(Gson.class)).andReturn(null).anyTimes();
-    expect(injector.getInstance(MaintenanceStateHelper.class)).andReturn(null).anyTimes();
-    expect(injector.getInstance(KerberosHelper.class)).andReturn(createNiceMock(KerberosHelper.class)).anyTimes();
-
-    replay(injector, clusters, mockAmsHbaseEnv, cluster);
-
-    AmbariManagementControllerImpl controller = createMockBuilder(AmbariManagementControllerImpl.class)
-      .addMockedMethod("createConfiguration")
-      .addMockedMethod("getClusters", new Class[] { })
-      .addMockedMethod("createConfig")
-      .withConstructor(createNiceMock(ActionManager.class), clusters, injector)
-      .createNiceMock();
-
-    Injector injector2 = easyMockSupport.createNiceMock(Injector.class);
-    Capture<Map<String, String>> propertiesCapture = EasyMock.newCapture();
-
-    expect(injector2.getInstance(AmbariManagementController.class)).andReturn(controller).anyTimes();
-    expect(controller.getClusters()).andReturn(clusters).anyTimes();
-    expect(controller.createConfig(anyObject(Cluster.class), anyObject(StackId.class), anyString(), capture(propertiesCapture), anyString(),
-      EasyMock.<Map<String, Map<String, String>>>anyObject())).andReturn(createNiceMock(Config.class)).once();
-
-    replay(controller, injector2);
-    new UpgradeCatalog240(injector2).updateAMSConfigs();
-    easyMockSupport.verifyAll();
-
-    Map<String, String> updatedProperties = propertiesCapture.getValue();
-    assertTrue(Maps.difference(newPropertiesAmsHbaseEnv, updatedProperties).areEqual());
-  }
-
-  @Test
-  public void testAmsSiteUpdateConfigs() throws Exception{
-
-    Map<String, String> oldPropertiesAmsSite = new HashMap<String, String>() {
-      {
-        put("timeline.metrics.sink.collection.period", "60");
-        put("timeline.metrics.host.aggregator.ttl", "86400");
-        put("timeline.metrics.cluster.aggregator.second.ttl", "604800");
-      }
-    };
-    Map<String, String> newPropertiesAmsSite = new HashMap<String, String>() {
-      {
-        put("timeline.metrics.sink.collection.period", "10");
-        put("timeline.metrics.host.aggregator.ttl", "86400");
-        put("timeline.metrics.cluster.aggregator.second.ttl", "259200");
-      }
-    };
-    EasyMockSupport easyMockSupport = new EasyMockSupport();
-
-    Clusters clusters = easyMockSupport.createNiceMock(Clusters.class);
-    final Cluster cluster = easyMockSupport.createNiceMock(Cluster.class);
-    Config mockAmsSite = easyMockSupport.createNiceMock(Config.class);
-
-    expect(clusters.getClusters()).andReturn(new HashMap<String, Cluster>() {{
-      put("normal", cluster);
-    }}).once();
-    expect(cluster.getDesiredConfigByType("ams-site")).andReturn(mockAmsSite).atLeastOnce();
-    expect(mockAmsSite.getProperties()).andReturn(oldPropertiesAmsSite).anyTimes();
-
-    Injector injector = easyMockSupport.createNiceMock(Injector.class);
-    expect(injector.getInstance(Gson.class)).andReturn(null).anyTimes();
-    expect(injector.getInstance(MaintenanceStateHelper.class)).andReturn(null).anyTimes();
-    expect(injector.getInstance(KerberosHelper.class)).andReturn(createNiceMock(KerberosHelper.class)).anyTimes();
-
-    replay(injector, clusters, mockAmsSite, cluster);
-
-    AmbariManagementControllerImpl controller = createMockBuilder(AmbariManagementControllerImpl.class)
-      .addMockedMethod("createConfiguration")
-      .addMockedMethod("getClusters", new Class[] { })
-      .addMockedMethod("createConfig")
-      .withConstructor(createNiceMock(ActionManager.class), clusters, injector)
-      .createNiceMock();
-
-    Injector injector2 = easyMockSupport.createNiceMock(Injector.class);
-    Capture<Map<String, String>> propertiesCapture = EasyMock.newCapture();
-
-    expect(injector2.getInstance(AmbariManagementController.class)).andReturn(controller).anyTimes();
-    expect(controller.getClusters()).andReturn(clusters).anyTimes();
-    expect(controller.createConfig(anyObject(Cluster.class), anyObject(StackId.class), anyString(), capture(propertiesCapture), anyString(),
-      EasyMock.<Map<String, Map<String, String>>>anyObject())).andReturn(createNiceMock(Config.class)).once();
-
-    replay(controller, injector2);
-    new UpgradeCatalog240(injector2).updateAMSConfigs();
-    easyMockSupport.verifyAll();
-
-    Map<String, String> updatedProperties = propertiesCapture.getValue();
-    assertTrue(Maps.difference(newPropertiesAmsSite, updatedProperties).areEqual());
-  }
-
-  @Test
-  public void testAmsHbaseSiteUpdateConfigs() throws Exception{
-
-    Map<String, String> oldPropertiesAmsHbaseSite = new HashMap<String, String>() {
-      {
-        put("hbase.rpc.timeout", "30000");
-        put("hbase.normalizer.enabled", String.valueOf(true));
-      }
-    };
-    Map<String, String> newPropertiesAmsHbaseSite = new HashMap<String, String>() {
-      {
-        put("hbase.rpc.timeout", "300000");
-        put("hbase.normalizer.enabled", String.valueOf(false));
-      }
-    };
-    EasyMockSupport easyMockSupport = new EasyMockSupport();
-
-    Clusters clusters = easyMockSupport.createNiceMock(Clusters.class);
-    final Cluster cluster = easyMockSupport.createNiceMock(Cluster.class);
-    Config mockAmsHbaseSite = easyMockSupport.createNiceMock(Config.class);
-
-    expect(clusters.getClusters()).andReturn(new HashMap<String, Cluster>() {{
-      put("normal", cluster);
-    }}).once();
-    expect(cluster.getDesiredConfigByType("ams-hbase-site")).andReturn(mockAmsHbaseSite).atLeastOnce();
-    expect(mockAmsHbaseSite.getProperties()).andReturn(oldPropertiesAmsHbaseSite).anyTimes();
-
-    Injector injector = easyMockSupport.createNiceMock(Injector.class);
-    expect(injector.getInstance(Gson.class)).andReturn(null).anyTimes();
-    expect(injector.getInstance(MaintenanceStateHelper.class)).andReturn(null).anyTimes();
-    expect(injector.getInstance(KerberosHelper.class)).andReturn(createNiceMock(KerberosHelper.class)).anyTimes();
-
-    replay(injector, clusters, mockAmsHbaseSite, cluster);
-
-    AmbariManagementControllerImpl controller = createMockBuilder(AmbariManagementControllerImpl.class)
-      .addMockedMethod("createConfiguration")
-      .addMockedMethod("getClusters", new Class[] { })
-      .addMockedMethod("createConfig")
-      .withConstructor(createNiceMock(ActionManager.class), clusters, injector)
-      .createNiceMock();
-
-    Injector injector2 = easyMockSupport.createNiceMock(Injector.class);
-    Capture<Map<String, String>> propertiesCapture = EasyMock.newCapture();
-
-    expect(injector2.getInstance(AmbariManagementController.class)).andReturn(controller).anyTimes();
-    expect(controller.getClusters()).andReturn(clusters).anyTimes();
-    expect(controller.createConfig(anyObject(Cluster.class), anyObject(StackId.class), anyString(), capture(propertiesCapture), anyString(),
-      EasyMock.<Map<String, Map<String, String>>>anyObject())).andReturn(createNiceMock(Config.class)).once();
-
-    replay(controller, injector2);
-    new UpgradeCatalog240(injector2).updateAMSConfigs();
-    easyMockSupport.verifyAll();
-
-    Map<String, String> updatedProperties = propertiesCapture.getValue();
-    assertTrue(Maps.difference(newPropertiesAmsHbaseSite, updatedProperties).areEqual());
-  }
-
-  @Test
-  public void testUpdateKerberosConfiguration() throws Exception {
-    final AmbariManagementController controller = createMock(AmbariManagementController.class);
-    final AmbariMetaInfo metaInfo = createMock(AmbariMetaInfo.class);
-    final StackInfo stackInfo = createMock(StackInfo.class);
-    final ServiceInfo serviceInfo = createMock(ServiceInfo.class);
-    final DBAccessor dbAccessor = createNiceMock(DBAccessor.class);
-    final OsFamily osFamily = createNiceMock(OsFamily.class);
-
-    final Map<String, String> propertiesKerberosEnv = new HashMap<String, String>() {
-      {
-        put("realm", "EXAMPLE.COM");
-        put("encryption_types", "aes des3-cbc-sha1 rc4 des-cbc-md5");
-        put("kdc_host", "c6407.ambari.apache.org");
-        put("admin_server_host", "c6407.ambari.apache.org");
-        put("kdc_type", "mit-kdc");
-      }
-    };
-
-    final Map<String, String> propertiesKrb5Conf = new HashMap<String, String>() {
-      {
-        put("content", "\n" +
-            "[libdefaults]\n" +
-            "  renew_lifetime = 7d\n" +
-            "  forwardable = true\n" +
-            "  default_realm = {{realm}}\n" +
-            "  ticket_lifetime = 24h\n" +
-            "  dns_lookup_realm = false\n" +
-            "  dns_lookup_kdc = false\n" +
-            "  #default_tgs_enctypes = {{encryption_types}}\n" +
-            "  #default_tkt_enctypes = {{encryption_types}}\n" +
-            "\n" +
-            "{% if domains %}\n" +
-            "[domain_realm]\n" +
-            "{% for domain in domains.split(',') %}\n" +
-            "  {{domain}} = {{realm}}\n" +
-            "{% endfor %}\n" +
-            "{% endif %}\n" +
-            "\n" +
-            "[logging]\n" +
-            "  default = FILE:/var/log/krb5kdc.log\n" +
-            "  admin_server = FILE:/var/log/kadmind.log\n" +
-            "  kdc = FILE:/var/log/krb5kdc.log\n" +
-            "\n" +
-            "[realms]\n" +
-            "  {{realm}} = {\n" +
-            "    admin_server = {{admin_server_host|default(kdc_host, True)}}\n" +
-            "    kdc = {{kdc_host}}\n" +
-            "  }\n" +
-            "\n" +
-            "{# Append additional realm declarations below #}");
-      }
-    };
-
-    final Config configKerberosEnv = createNiceMock(Config.class);
-    expect(configKerberosEnv.getProperties()).andReturn(propertiesKerberosEnv).anyTimes();
-    expect(configKerberosEnv.getTag()).andReturn("tag1").anyTimes();
-
-    final Config configKrb5Conf = createNiceMock(Config.class);
-    expect(configKrb5Conf.getProperties()).andReturn(propertiesKrb

<TRUNCATED>

Mime
View raw message