ambari-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From jlun...@apache.org
Subject [09/13] ambari git commit: Revert "AMBARI-16250: Create STORM service versions in common-services (jluniya)"
Date Fri, 06 May 2016 06:04:49 GMT
http://git-wip-us.apache.org/repos/asf/ambari/blob/d216c128/ambari-server/src/main/resources/common-services/STORM/0.9.1/kerberos.json
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/STORM/0.9.1/kerberos.json b/ambari-server/src/main/resources/common-services/STORM/0.9.1/kerberos.json
deleted file mode 100644
index fcfe524..0000000
--- a/ambari-server/src/main/resources/common-services/STORM/0.9.1/kerberos.json
+++ /dev/null
@@ -1,120 +0,0 @@
-{
-  "services": [
-    {
-      "name": "STORM",
-      "identities": [
-        {
-          "name": "/spnego"
-        },
-        {
-          "name": "/smokeuser"
-        },
-        {
-          "name": "storm_components",
-          "principal": {
-            "value": "${storm-env/storm_user}-${cluster_name|toLower()}@${realm}",
-            "type": "user",
-            "configuration": "storm-env/storm_principal_name"
-          },
-          "keytab": {
-            "file": "${keytab_dir}/storm.headless.keytab",
-            "owner": {
-              "name": "${storm-env/storm_user}",
-              "access": "r"
-            },
-            "group": {
-              "name": "${cluster-env/user_group}",
-              "access": ""
-            },
-            "configuration": "storm-env/storm_keytab"
-          }
-        }
-      ],
-      "configurations": [
-        {
-          "storm-site": {
-            "nimbus.authorizer": "backtype.storm.security.auth.authorizer.SimpleACLAuthorizer",
-            "drpc.authorizer": "backtype.storm.security.auth.authorizer.DRPCSimpleACLAuthorizer",
-            "ui.filter": "org.apache.hadoop.security.authentication.server.AuthenticationFilter",
-            "storm.principal.tolocal": "backtype.storm.security.auth.KerberosPrincipalToLocal",
-            "supervisor.enable": "true",
-            "storm.zookeeper.superACL": "sasl:{{storm_bare_jaas_principal}}",
-            "java.security.auth.login.config": "{{conf_dir}}/storm_jaas.conf",
-            "nimbus.admins": "['{{storm_bare_jaas_principal}}']",
-            "nimbus.supervisor.users": "['{{storm_bare_jaas_principal}}']",
-            "ui.filter.params": "{'type': 'kerberos', 'kerberos.principal': '{{storm_ui_jaas_principal}}', 'kerberos.keytab': '{{storm_ui_keytab_path}}', 'kerberos.name.rules': 'DEFAULT'}"
-          }
-        }
-      ],
-      "components": [
-        {
-          "name": "STORM_UI_SERVER",
-          "identities": [
-            {
-              "name": "/spnego",
-              "principal": {
-                "configuration": "storm-env/storm_ui_principal_name"
-              },
-              "keytab": {
-                "configuration": "storm-env/storm_ui_keytab"
-              }
-            }
-          ]
-        },
-        {
-          "name": "NIMBUS",
-          "identities": [
-            {
-              "name": "nimbus_server",
-              "principal": {
-                "value": "nimbus/_HOST@${realm}",
-                "type": "service",
-                "configuration": "storm-env/nimbus_principal_name"
-              },
-              "keytab": {
-                "file": "${keytab_dir}/nimbus.service.keytab",
-                "owner": {
-                  "name": "${storm-env/storm_user}",
-                  "access": "r"
-                },
-                "group": {
-                  "name": "${cluster-env/user_group}",
-                  "access": ""
-                },
-                "configuration": "storm-env/nimbus_keytab"
-              }
-            }
-          ]
-        },
-        {
-          "name": "DRPC_SERVER",
-          "identities": [
-            {
-              "name": "nimbus_server",
-              "principal": {
-                "value": "nimbus/_HOST@${realm}",
-                "type": "service",
-                "configuration": "storm-env/nimbus_principal_name"
-              },
-              "keytab": {
-                "file": "${keytab_dir}/nimbus.service.keytab",
-                "owner": {
-                  "name": "${storm-env/storm_user}",
-                  "access": "r"
-                },
-                "group": {
-                  "name": "${cluster-env/user_group}",
-                  "access": ""
-                },
-                "configuration": "storm-env/nimbus_keytab"
-              }
-            }
-          ]
-        },
-        {
-          "name" : "SUPERVISOR"
-        }
-      ]
-    }
-  ]
-}

http://git-wip-us.apache.org/repos/asf/ambari/blob/d216c128/ambari-server/src/main/resources/common-services/STORM/0.9.1/metainfo.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/STORM/0.9.1/metainfo.xml b/ambari-server/src/main/resources/common-services/STORM/0.9.1/metainfo.xml
deleted file mode 100644
index 0f4d520..0000000
--- a/ambari-server/src/main/resources/common-services/STORM/0.9.1/metainfo.xml
+++ /dev/null
@@ -1,181 +0,0 @@
-<?xml version="1.0"?>
-<!--
-   Licensed to the Apache Software Foundation (ASF) under one or more
-   contributor license agreements.  See the NOTICE file distributed with
-   this work for additional information regarding copyright ownership.
-   The ASF licenses this file to You under the Apache License, Version 2.0
-   (the "License"); you may not use this file except in compliance with
-   the License.  You may obtain a copy of the License at
-
-       http://www.apache.org/licenses/LICENSE-2.0
-
-   Unless required by applicable law or agreed to in writing, software
-   distributed under the License is distributed on an "AS IS" BASIS,
-   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-   See the License for the specific language governing permissions and
-   limitations under the License.
--->
-
-<metainfo>
-  <schemaVersion>2.0</schemaVersion>
-  <services>
-    <service>
-      <name>STORM</name>
-      <displayName>Storm</displayName>
-      <comment>Apache Hadoop Stream processing framework</comment>
-      <version>0.9.1</version>
-      <components>
-
-        <component>
-          <name>NIMBUS</name>
-          <displayName>Nimbus</displayName>
-          <category>MASTER</category>
-          <cardinality>1</cardinality>
-          <versionAdvertised>true</versionAdvertised>
-          <dependencies>
-            <dependency>
-              <name>ZOOKEEPER/ZOOKEEPER_SERVER</name>
-              <scope>cluster</scope>
-              <auto-deploy>
-                <enabled>true</enabled>
-              </auto-deploy>
-            </dependency>
-          </dependencies>
-          <commandScript>
-            <script>scripts/nimbus.py</script>
-            <scriptType>PYTHON</scriptType>
-            <timeout>1200</timeout>
-          </commandScript>
-          <logs>
-            <log>
-              <logId>storm_nimbus</logId>
-              <primary>true</primary>
-            </log>
-          </logs>
-        </component>
-
-        <component>
-          <name>STORM_REST_API</name>
-          <displayName>Storm REST API Server</displayName>
-          <category>MASTER</category>
-          <cardinality>1</cardinality>
-          <versionAdvertised>true</versionAdvertised>
-          <commandScript>
-            <script>scripts/rest_api.py</script>
-            <scriptType>PYTHON</scriptType>
-            <timeout>1200</timeout>
-          </commandScript>
-        </component>
-
-        <component>
-          <name>SUPERVISOR</name>
-          <displayName>Supervisor</displayName>
-          <category>SLAVE</category>
-          <cardinality>1+</cardinality>
-          <versionAdvertised>true</versionAdvertised>
-          <commandScript>
-            <script>scripts/supervisor.py</script>
-            <scriptType>PYTHON</scriptType>
-            <timeout>1200</timeout>
-          </commandScript>
-          <bulkCommands>
-            <displayName>Supervisors</displayName>
-            <masterComponent>SUPERVISOR</masterComponent>
-          </bulkCommands>
-          <logs>
-            <log>
-              <logId>storm_supervisor</logId>
-              <primary>true</primary>
-            </log>
-            <log>
-              <logId>storm_worker</logId>
-            </log>
-            <log>
-              <logId>storm_logviewer</logId>
-            </log>
-          </logs>
-        </component>
-
-        <component>
-          <name>STORM_UI_SERVER</name>
-          <displayName>Storm UI Server</displayName>
-          <category>MASTER</category>
-          <cardinality>1</cardinality>
-          <versionAdvertised>true</versionAdvertised>
-          <commandScript>
-            <script>scripts/ui_server.py</script>
-            <scriptType>PYTHON</scriptType>
-            <timeout>1200</timeout>
-          </commandScript>
-          <logs>
-            <log>
-              <logId>storm_ui</logId>
-              <primary>true</primary>
-            </log>
-          </logs>
-        </component>
-
-        <component>
-          <name>DRPC_SERVER</name>
-          <displayName>DRPC Server</displayName>
-          <category>MASTER</category>
-          <cardinality>1</cardinality>
-          <versionAdvertised>true</versionAdvertised>
-          <commandScript>
-            <script>scripts/drpc_server.py</script>
-            <scriptType>PYTHON</scriptType>
-            <timeout>1200</timeout>
-          </commandScript>
-          <logs>
-            <log>
-              <logId>storm_drpc</logId>
-              <primary>true</primary>
-            </log>
-          </logs>
-        </component>
-      </components>
-
-      <osSpecifics>
-        <osSpecific>
-          <osFamily>any</osFamily>
-          <packages>
-            <package>
-              <name>storm</name>
-            </package>
-          </packages>
-        </osSpecific>
-      </osSpecifics>
-
-      <commandScript>
-        <script>scripts/service_check.py</script>
-        <scriptType>PYTHON</scriptType>
-        <timeout>300</timeout>
-      </commandScript>
-
-      <requiredServices>
-        <service>ZOOKEEPER</service>
-      </requiredServices>
-
-      <configuration-dependencies>
-        <config-type>storm-site</config-type>
-        <config-type>storm-env</config-type>
-        <config-type>ranger-storm-plugin-properties</config-type>
-        <config-type>ranger-storm-audit</config-type>
-        <config-type>ranger-storm-policymgr-ssl</config-type>
-        <config-type>ranger-storm-security</config-type>
-        <config-type>admin-properties</config-type>
-        <config-type>ranger-ugsync-site</config-type>
-        <config-type>ranger-admin-site</config-type>
-        <config-type>zookeeper-env</config-type>
-        <config-type>zoo.cfg</config-type>
-        <config-type>ams-ssl-client</config-type>
-      </configuration-dependencies>
-      <quickLinksConfigurations>
-        <quickLinksConfiguration>
-          <fileName>quicklinks.json</fileName>
-          <default>true</default>
-        </quickLinksConfiguration>
-      </quickLinksConfigurations>
-    </service>
-  </services>
-</metainfo>

http://git-wip-us.apache.org/repos/asf/ambari/blob/d216c128/ambari-server/src/main/resources/common-services/STORM/0.9.1/metrics.json
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/STORM/0.9.1/metrics.json b/ambari-server/src/main/resources/common-services/STORM/0.9.1/metrics.json
deleted file mode 100644
index 08c2b50..0000000
--- a/ambari-server/src/main/resources/common-services/STORM/0.9.1/metrics.json
+++ /dev/null
@@ -1,829 +0,0 @@
-{
-  "STORM_REST_API": {
-    "Component": [
-      {
-        "type": "org.apache.ambari.server.controller.metrics.RestMetricsPropertyProvider",
-        "properties": {
-          "default_port": "8745",
-          "port_config_type": "storm-site",
-          "port_property_name": "storm.port",
-          "protocol": "http"
-        },
-        "metrics": {
-          "default": {
-            "metrics/api/cluster/summary/tasks.total": {
-              "metric": "/api/cluster/summary##tasks.total",
-              "pointInTime": true,
-              "temporal": false
-            },
-            "metrics/api/cluster/summary/slots.total": {
-              "metric": "/api/cluster/summary##slots.total",
-              "pointInTime": true,
-              "temporal": false
-            },
-            "metrics/api/cluster/summary/slots.free": {
-              "metric": "/api/cluster/summary##slots.free",
-              "pointInTime": true,
-              "temporal": false
-            },
-            "metrics/api/cluster/summary/supervisors": {
-              "metric": "/api/cluster/summary##supervisors",
-              "pointInTime": true,
-              "temporal": false
-            },
-            "metrics/api/cluster/summary/executors.total": {
-              "metric": "/api/cluster/summary##executors.total",
-              "pointInTime": true,
-              "temporal": false
-            },
-            "metrics/api/cluster/summary/slots.used": {
-              "metric": "/api/cluster/summary##slots.used",
-              "pointInTime": true,
-              "temporal": false
-            },
-            "metrics/api/cluster/summary/topologies": {
-              "metric": "/api/cluster/summary##topologies",
-              "pointInTime": true,
-              "temporal": false
-            },
-            "metrics/api/cluster/summary/nimbus.uptime": {
-              "metric": "/api/cluster/summary##nimbus.uptime",
-              "pointInTime": true,
-              "temporal": false
-            }
-          }
-        }
-      }
-    ],
-    "HostComponent": [
-      {
-        "type": "org.apache.ambari.server.controller.metrics.RestMetricsPropertyProvider",
-        "properties": {
-          "default_port": "8745",
-          "port_config_type": "storm-site",
-          "port_property_name": "storm.port",
-          "protocol": "http"
-        },
-        "metrics": {
-          "default": {
-            "metrics/api/cluster/summary/tasks.total": {
-              "metric": "/api/cluster/summary##tasks.total",
-              "pointInTime": true,
-              "temporal": false
-            },
-            "metrics/api/cluster/summary/slots.total": {
-              "metric": "/api/cluster/summary##slots.total",
-              "pointInTime": true,
-              "temporal": false
-            },
-            "metrics/api/cluster/summary/slots.free": {
-              "metric": "/api/cluster/summary##slots.free",
-              "pointInTime": true,
-              "temporal": false
-            },
-            "metrics/api/cluster/summary/supervisors": {
-              "metric": "/api/cluster/summary##supervisors",
-              "pointInTime": true,
-              "temporal": false
-            },
-            "metrics/api/cluster/summary/executors.total": {
-              "metric": "/api/cluster/summary##executors.total",
-              "pointInTime": true,
-              "temporal": false
-            },
-            "metrics/api/cluster/summary/slots.used": {
-              "metric": "/api/cluster/summary##slots.used",
-              "pointInTime": true,
-              "temporal": false
-            },
-            "metrics/api/cluster/summary/topologies": {
-              "metric": "/api/cluster/summary##topologies",
-              "pointInTime": true,
-              "temporal": false
-            },
-            "metrics/api/cluster/summary/nimbus.uptime": {
-              "metric": "/api/cluster/summary##nimbus.uptime",
-              "pointInTime": true,
-              "temporal": false
-            }
-          }
-        }
-      }
-    ]
-  },
-  "NIMBUS": {
-    "Component": [
-      {
-        "type": "ganglia",
-        "metrics": {
-          "default": {
-            "metrics/boottime": {
-              "metric": "boottime",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/network/bytes_in": {
-              "metric": "bytes_in",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/network/bytes_out": {
-              "metric": "bytes_out",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/cpu/cpu_aidle": {
-              "metric": "cpu_aidle",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/load_fifteen": {
-              "metric": "load_fifteen",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/load_five": {
-              "metric": "load_five",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/load_one": {
-              "metric": "load_one",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/memory/mem_buffers": {
-              "metric": "mem_buffers",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/disk/part_max_used": {
-              "metric": "part_max_used",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/network/pkts_in": {
-              "metric": "pkts_in",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/network/pkts_out": {
-              "metric": "pkts_out",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/process/proc_run": {
-              "metric": "proc_run",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/process/proc_total": {
-              "metric": "proc_total",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/memory/swap_total": {
-              "metric": "swap_total",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/nimbus/jvm/memory/heap/committed": {
-              "metric": "Nimbus.JVM.Memory.Heap.committed",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/nimbus/jvm/memory/heap/init": {
-              "metric": "Nimbus.JVM.Memory.Heap.init",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/nimbus/jvm/memory/heap/max": {
-              "metric": "Nimbus.JVM.Memory.Heap.max",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/nimbus/jvm/memory/heap/used": {
-              "metric": "Nimbus.JVM.Memory.Heap.used",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/nimbus/jvm/memory/nonheap/committed": {
-              "metric": "Nimbus.JVM.Memory.NonHeap.committed",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/nimbus/jvm/memory/nonheap/init": {
-              "metric": "Nimbus.JVM.Memory.NonHeap.init",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/nimbus/jvm/memory/nonheap/max": {
-              "metric": "Nimbus.JVM.Memory.NonHeap.max",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/nimbus/jvm/memory/nonheap/used": {
-              "metric": "Nimbus.JVM.Memory.NonHeap.used",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/nimbus/jvm/os/processcputime": {
-              "metric": "Nimbus.JVM.OS.ProcessCpuTime",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/nimbus/jvm/threading/daemonthreadcount": {
-              "metric": "Nimbus.JVM.Threading.DaemonThreadCount",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/nimbus/jvm/threading/threadcount": {
-              "metric": "Nimbus.JVM.Threading.ThreadCount",
-              "pointInTime": true,
-              "temporal": true
-            },
-
-            "metrics/storm/nimbus/freeslots": {
-              "metric": "Free Slots",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/nimbus/supervisors": {
-              "metric": "Supervisors",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/nimbus/topologies": {
-              "metric": "Topologies",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/nimbus/totalexecutors": {
-              "metric": "Total Executors",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/nimbus/totalslots": {
-              "metric": "Total Slots",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/nimbus/totaltasks": {
-              "metric": "Total Tasks",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/nimbus/usedslots": {
-              "metric": "Used Slots",
-              "pointInTime": true,
-              "temporal": true
-            }
-          }
-        }
-      }
-    ],
-    "HostComponent": [
-      {
-        "type": "ganglia",
-        "metrics": {
-          "default": {
-            "metrics/boottime": {
-              "metric": "boottime",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric": true
-            },
-            "metrics/network/bytes_in": {
-              "metric": "bytes_in",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric": true
-            },
-            "metrics/network/bytes_out": {
-              "metric": "bytes_out",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric": true
-            },
-            "metrics/cpu/cpu_aidle": {
-              "metric": "cpu_aidle",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric": true
-            },
-            "metrics/load_fifteen": {
-              "metric": "load_fifteen",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric": true
-            },
-            "metrics/load_five": {
-              "metric": "load_five",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric": true
-            },
-            "metrics/load_one": {
-              "metric": "load_one",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric": true
-            },
-            "metrics/memory/mem_buffers": {
-              "metric": "mem_buffers",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric": true
-            },
-            "metrics/disk/part_max_used": {
-              "metric": "part_max_used",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric": true
-            },
-            "metrics/network/pkts_in": {
-              "metric": "pkts_in",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric": true
-            },
-            "metrics/network/pkts_out": {
-              "metric": "pkts_out",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric": true
-            },
-            "metrics/process/proc_run": {
-              "metric": "proc_run",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric": true
-            },
-            "metrics/process/proc_total": {
-              "metric": "proc_total",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric": true
-            },
-            "metrics/memory/swap_total": {
-              "metric": "swap_total",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric": true
-            },
-            "metrics/storm/nimbus/jvm/memory/heap/committed": {
-              "metric": "Nimbus.JVM.Memory.Heap.committed",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/nimbus/jvm/memory/heap/init": {
-              "metric": "Nimbus.JVM.Memory.Heap.init",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/nimbus/jvm/memory/heap/max": {
-              "metric": "Nimbus.JVM.Memory.Heap.max",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/nimbus/jvm/memory/heap/used": {
-              "metric": "Nimbus.JVM.Memory.Heap.used",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/nimbus/jvm/memory/nonheap/committed": {
-              "metric": "Nimbus.JVM.Memory.NonHeap.committed",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/nimbus/jvm/memory/nonheap/init": {
-              "metric": "Nimbus.JVM.Memory.NonHeap.init",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/nimbus/jvm/memory/nonheap/max": {
-              "metric": "Nimbus.JVM.Memory.NonHeap.max",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/nimbus/jvm/memory/nonheap/used": {
-              "metric": "Nimbus.JVM.Memory.NonHeap.used",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/nimbus/jvm/os/processcputime": {
-              "metric": "Nimbus.JVM.OS.ProcessCpuTime",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/nimbus/jvm/threading/daemonthreadcount": {
-              "metric": "Nimbus.JVM.Threading.DaemonThreadCount",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/nimbus/jvm/threading/threadcount": {
-              "metric": "Nimbus.JVM.Threading.ThreadCount",
-              "pointInTime": true,
-              "temporal": true
-            }
-
-          }
-        }
-      }
-    ]
-  },
-  "SUPERVISOR": {
-    "Component": [
-      {
-        "type": "ganglia",
-        "metrics": {
-          "default": {
-            "metrics/boottime": {
-              "metric": "boottime",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/network/bytes_in": {
-              "metric": "bytes_in",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/network/bytes_out": {
-              "metric": "bytes_out",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/cpu/cpu_aidle": {
-              "metric": "cpu_aidle",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/load_fifteen": {
-              "metric": "load_fifteen",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/load_five": {
-              "metric": "load_five",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/load_one": {
-              "metric": "load_one",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/memory/mem_buffers": {
-              "metric": "mem_buffers",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/disk/part_max_used": {
-              "metric": "part_max_used",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/network/pkts_in": {
-              "metric": "pkts_in",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/network/pkts_out": {
-              "metric": "pkts_out",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/process/proc_run": {
-              "metric": "proc_run",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/process/proc_total": {
-              "metric": "proc_total",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/memory/swap_total": {
-              "metric": "swap_total",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/supervisor/jvm/memory/heap/committed": {
-              "metric": "Supervisor.JVM.Memory.Heap.committed",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/supervisor/jvm/memory/heap/init": {
-              "metric": "Supervisor.JVM.Memory.Heap.init",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/supervisor/jvm/memory/heap/max": {
-              "metric": "Supervisor.JVM.Memory.Heap.max",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/supervisor/jvm/memory/heap/used": {
-              "metric": "Supervisor.JVM.Memory.Heap.used",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/supervisor/jvm/memory/nonheap/committed": {
-              "metric": "Supervisor.JVM.Memory.NonHeap.committed",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/supervisor/jvm/memory/nonheap/init": {
-              "metric": "Supervisor.JVM.Memory.NonHeap.init",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/supervisor/jvm/memory/nonheap/max": {
-              "metric": "Supervisor.JVM.Memory.NonHeap.max",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/supervisor/jvm/memory/nonheap/used": {
-              "metric": "Supervisor.JVM.Memory.NonHeap.used",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/supervisor/jvm/os/processcputime": {
-              "metric": "Supervisor.JVM.OS.ProcessCpuTime",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/supervisor/jvm/threading/daemonthreadcount": {
-              "metric": "Supervisor.JVM.Threading.DaemonThreadCount",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/supervisor/jvm/threading/threadcount": {
-              "metric": "Supervisor.JVM.Threading.ThreadCount",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/worker/$1/jvm/memory/heap/committed": {
-              "metric": "Worker.(.+).JVM.Memory.Heap.committed",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/worker/$1/jvm/memory/heap/init": {
-              "metric": "Worker.(.+).JVM.Memory.Heap.init",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/worker/$1/jvm/memory/heap/max": {
-              "metric": "Worker.(.+).JVM.Memory.Heap.max",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/worker/$1/jvm/memory/heap/used": {
-              "metric": "Worker.(.+).JVM.Memory.Heap.used",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/worker/$1/jvm/memory/nonheap/committed": {
-              "metric": "Worker.(.+).JVM.Memory.NonHeap.committed",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/worker/$1/jvm/memory/nonheap/init": {
-              "metric": "Worker.(.+).JVM.Memory.NonHeap.init",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/worker/$1/jvm/memory/nonheap/max": {
-              "metric": "Worker.(.+).JVM.Memory.NonHeap.max",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/worker/$1/jvm/memory/nonheap/used": {
-              "metric": "Worker.(.+).JVM.Memory.NonHeap.used",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/worker/$1/jvm/os/processcputime": {
-              "metric": "Worker.(.+).JVM.OS.ProcessCpuTime",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/worker/$1/jvm/threading/daemonthreadcount": {
-              "metric": "Worker.(.+).JVM.Threading.DaemonThreadCount",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/worker/$1/jvm/threading/threadcount": {
-              "metric": "Worker.(.+).JVM.Threading.ThreadCount",
-              "pointInTime": true,
-              "temporal": true
-            }
-          }
-        }
-      }
-    ],
-    "HostComponent": [
-      {
-        "type": "ganglia",
-        "metrics": {
-          "default": {
-            "metrics/boottime": {
-              "metric": "boottime",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric":true
-            },
-            "metrics/network/bytes_in": {
-              "metric": "bytes_in",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric":true
-            },
-            "metrics/network/bytes_out": {
-              "metric": "bytes_out",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric":true
-            },
-            "metrics/cpu/cpu_aidle": {
-              "metric": "cpu_aidle",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric":true
-            },
-            "metrics/load_fifteen": {
-              "metric": "load_fifteen",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric":true
-            },
-            "metrics/load_five": {
-              "metric": "load_five",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric":true
-            },
-            "metrics/load_one": {
-              "metric": "load_one",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric":true
-            },
-            "metrics/memory/mem_buffers": {
-              "metric": "mem_buffers",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric":true
-            },
-            "metrics/disk/part_max_used": {
-              "metric": "part_max_used",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric":true
-            },
-            "metrics/network/pkts_in": {
-              "metric": "pkts_in",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric":true
-            },
-            "metrics/network/pkts_out": {
-              "metric": "pkts_out",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric":true
-            },
-            "metrics/process/proc_run": {
-              "metric": "proc_run",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric":true
-            },
-            "metrics/process/proc_total": {
-              "metric": "proc_total",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric":true
-            },
-            "metrics/memory/swap_total": {
-              "metric": "swap_total",
-              "pointInTime": true,
-              "temporal": true,
-              "amsHostMetric":true
-            },
-            "metrics/storm/supervisor/jvm/memory/heap/committed": {
-              "metric": "Supervisor.JVM.Memory.Heap.committed",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/supervisor/jvm/memory/heap/init": {
-              "metric": "Supervisor.JVM.Memory.Heap.init",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/supervisor/jvm/memory/heap/max": {
-              "metric": "Supervisor.JVM.Memory.Heap.max",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/supervisor/jvm/memory/heap/used": {
-              "metric": "Supervisor.JVM.Memory.Heap.used",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/supervisor/jvm/memory/nonheap/committed": {
-              "metric": "Supervisor.JVM.Memory.NonHeap.committed",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/supervisor/jvm/memory/nonheap/init": {
-              "metric": "Supervisor.JVM.Memory.NonHeap.init",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/supervisor/jvm/memory/nonheap/max": {
-              "metric": "Supervisor.JVM.Memory.NonHeap.max",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/supervisor/jvm/memory/nonheap/used": {
-              "metric": "Supervisor.JVM.Memory.NonHeap.used",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/supervisor/jvm/os/processcputime": {
-              "metric": "Supervisor.JVM.OS.ProcessCpuTime",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/supervisor/jvm/threading/daemonthreadcount": {
-              "metric": "Supervisor.JVM.Threading.DaemonThreadCount",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/supervisor/jvm/threading/threadcount": {
-              "metric": "Supervisor.JVM.Threading.ThreadCount",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/worker/$1/jvm/memory/heap/committed": {
-              "metric": "Worker.(.+).JVM.Memory.Heap.committed",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/worker/$1/jvm/memory/heap/init": {
-              "metric": "Worker.(.+).JVM.Memory.Heap.init",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/worker/$1/jvm/memory/heap/max": {
-              "metric": "Worker.(.+).JVM.Memory.Heap.max",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/worker/$1/jvm/memory/heap/used": {
-              "metric": "Worker.(.+).JVM.Memory.Heap.used",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/worker/$1/jvm/memory/nonheap/committed": {
-              "metric": "Worker.(.+).JVM.Memory.NonHeap.committed",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/worker/$1/jvm/memory/nonheap/init": {
-              "metric": "Worker.(.+).JVM.Memory.NonHeap.init",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/worker/$1/jvm/memory/nonheap/max": {
-              "metric": "Worker.(.+).JVM.Memory.NonHeap.max",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/worker/$1/jvm/memory/nonheap/used": {
-              "metric": "Worker.(.+).JVM.Memory.NonHeap.used",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/worker/$1/jvm/os/processcputime": {
-              "metric": "Worker.(.+).JVM.OS.ProcessCpuTime",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/worker/$1/jvm/threading/daemonthreadcount": {
-              "metric": "Worker.(.+).JVM.Threading.DaemonThreadCount",
-              "pointInTime": true,
-              "temporal": true
-            },
-            "metrics/storm/worker/$1/jvm/threading/threadcount": {
-              "metric": "Worker.(.+).JVM.Threading.ThreadCount",
-              "pointInTime": true,
-              "temporal": true
-            }
-          }
-        }
-      }
-    ]
-  }
-}

http://git-wip-us.apache.org/repos/asf/ambari/blob/d216c128/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/alerts/check_supervisor_process_win.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/alerts/check_supervisor_process_win.py b/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/alerts/check_supervisor_process_win.py
deleted file mode 100644
index a698415..0000000
--- a/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/alerts/check_supervisor_process_win.py
+++ /dev/null
@@ -1,50 +0,0 @@
-#!/usr/bin/env python
-
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-"""
-
-from resource_management.libraries.functions import check_windows_service_status
-
-
-RESULT_CODE_OK = 'OK'
-RESULT_CODE_CRITICAL = 'CRITICAL'
-RESULT_CODE_UNKNOWN = 'UNKNOWN'
-
-
-def get_tokens():
-  """
-  Returns a tuple of tokens in the format {{site/property}} that will be used
-  to build the dictionary passed into execute
-  """
-  return ()
-
-def execute(configurations={}, parameters={}, host_name=None):
-  """
-  Returns a tuple containing the result code and a pre-formatted result label
-
-  Keyword arguments:
-  configurations (dictionary): a mapping of configuration key to value
-  parameters (dictionary): a mapping of script parameter key to value
-  host_name (string): the name of this host where the alert is running
-  """
-
-  try:
-    check_windows_service_status("supervisor")
-    return (RESULT_CODE_OK, ["Supervisor is running"])
-  except:
-    return (RESULT_CODE_CRITICAL, ["Supervisor is stopped"])

http://git-wip-us.apache.org/repos/asf/ambari/blob/d216c128/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/files/wordCount.jar
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/files/wordCount.jar b/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/files/wordCount.jar
deleted file mode 100644
index aed64be..0000000
Binary files a/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/files/wordCount.jar and /dev/null differ

http://git-wip-us.apache.org/repos/asf/ambari/blob/d216c128/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/scripts/drpc_server.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/scripts/drpc_server.py b/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/scripts/drpc_server.py
deleted file mode 100644
index b156578..0000000
--- a/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/scripts/drpc_server.py
+++ /dev/null
@@ -1,139 +0,0 @@
-#!/usr/bin/env python
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-"""
-
-import sys
-from resource_management.libraries.functions import check_process_status
-from resource_management.libraries.script import Script
-from resource_management.libraries.functions import conf_select
-from resource_management.libraries.functions import stack_select
-from resource_management.libraries.functions import format
-from resource_management.core.resources.system import Execute
-from resource_management.libraries.functions.stack_features import check_stack_feature
-from resource_management.libraries.functions import StackFeature
-from storm import storm
-from service import service
-from service_check import ServiceCheck
-from resource_management.libraries.functions.security_commons import build_expectations, \
-  cached_kinit_executor, get_params_from_filesystem, validate_security_config_properties, \
-  FILE_TYPE_JAAS_CONF
-
-class DrpcServer(Script):
-
-  def get_component_name(self):
-    return "storm-client"
-
-  def install(self, env):
-    self.install_packages(env)
-    self.configure(env)
-
-  def configure(self, env):
-    import params
-    env.set_params(params)
-
-    storm()
-
-  def pre_upgrade_restart(self, env, upgrade_type=None):
-    import params
-    env.set_params(params)
-
-    if params.version and check_stack_feature(StackFeature.ROLLING_UPGRADE, params.version):
-      conf_select.select(params.stack_name, "storm", params.version)
-      stack_select.select("storm-client", params.version)
-
-  def start(self, env, upgrade_type=None):
-    import params
-    env.set_params(params)
-    self.configure(env)
-
-    service("drpc", action="start")
-
-  def stop(self, env, upgrade_type=None):
-    import params
-    env.set_params(params)
-
-    service("drpc", action="stop")
-
-  def status(self, env):
-    import status_params
-    env.set_params(status_params)
-    check_process_status(status_params.pid_drpc)
-
-  def security_status(self, env):
-    import status_params
-
-    env.set_params(status_params)
-
-    if status_params.security_enabled:
-      # Expect the following files to be available in status_params.config_dir:
-      #   storm_jaas.conf
-
-      try:
-        props_value_check = None
-        props_empty_check = ['StormServer/keyTab', 'StormServer/principal']
-        props_read_check = ['StormServer/keyTab']
-        storm_env_expectations = build_expectations('storm_jaas', props_value_check, props_empty_check,
-                                                 props_read_check)
-
-        storm_expectations = {}
-        storm_expectations.update(storm_env_expectations)
-
-        security_params = get_params_from_filesystem(status_params.conf_dir,
-                                                     {'storm_jaas.conf': FILE_TYPE_JAAS_CONF})
-
-        result_issues = validate_security_config_properties(security_params, storm_expectations)
-        if not result_issues:  # If all validations passed successfully
-          # Double check the dict before calling execute
-          if ( 'storm_jaas' not in security_params
-               or 'StormServer' not in security_params['storm_jaas']
-               or 'keyTab' not in security_params['storm_jaas']['StormServer']
-               or 'principal' not in security_params['storm_jaas']['StormServer']):
-            self.put_structured_out({"securityState": "ERROR"})
-            self.put_structured_out({"securityIssuesFound": "Keytab file or principal are not set property."})
-            return
-
-          cached_kinit_executor(status_params.kinit_path_local,
-                                status_params.storm_user,
-                                security_params['storm_jaas']['StormServer']['keyTab'],
-                                security_params['storm_jaas']['StormServer']['principal'],
-                                status_params.hostname,
-                                status_params.tmp_dir)
-          self.put_structured_out({"securityState": "SECURED_KERBEROS"})
-        else:
-          issues = []
-          for cf in result_issues:
-            issues.append("Configuration file %s did not pass the validation. Reason: %s" % (cf, result_issues[cf]))
-          self.put_structured_out({"securityIssuesFound": ". ".join(issues)})
-          self.put_structured_out({"securityState": "UNSECURED"})
-      except Exception as e:
-        self.put_structured_out({"securityState": "ERROR"})
-        self.put_structured_out({"securityStateErrorInfo": str(e)})
-    else:
-      self.put_structured_out({"securityState": "UNSECURED"})
-      
-  def get_log_folder(self):
-    import params
-    return params.log_dir
-  
-  def get_user(self):
-    import params
-    return params.storm_user
-
-if __name__ == "__main__":
-  DrpcServer().execute()

http://git-wip-us.apache.org/repos/asf/ambari/blob/d216c128/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/scripts/nimbus.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/scripts/nimbus.py b/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/scripts/nimbus.py
deleted file mode 100644
index b2148a5..0000000
--- a/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/scripts/nimbus.py
+++ /dev/null
@@ -1,157 +0,0 @@
-#!/usr/bin/env python
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-"""
-
-import sys
-from resource_management.libraries.functions import check_process_status
-from resource_management.libraries.script import Script
-from resource_management.libraries.functions import format
-from resource_management.libraries.functions import conf_select
-from resource_management.libraries.functions import stack_select
-from resource_management.core.resources.system import Execute
-from resource_management.libraries.functions.stack_features import check_stack_feature
-from resource_management.libraries.functions import StackFeature
-from storm import storm
-from service import service
-from resource_management.libraries.functions.security_commons import build_expectations, \
-  cached_kinit_executor, get_params_from_filesystem, validate_security_config_properties, \
-  FILE_TYPE_JAAS_CONF
-from setup_ranger_storm import setup_ranger_storm
-from ambari_commons import OSConst
-from ambari_commons.os_family_impl import OsFamilyImpl
-from resource_management.core.resources.service import Service
-
-class Nimbus(Script):
-  def get_component_name(self):
-    return "storm-nimbus"
-
-  def install(self, env):
-    self.install_packages(env)
-    self.configure(env)
-
-  def configure(self, env):
-    import params
-    env.set_params(params)
-    storm("nimbus")
-
-
-@OsFamilyImpl(os_family=OsFamilyImpl.DEFAULT)
-class NimbusDefault(Nimbus):
-
-  def pre_upgrade_restart(self, env, upgrade_type=None):
-    import params
-    env.set_params(params)
-    if params.version and check_stack_feature(StackFeature.ROLLING_UPGRADE, params.version):
-      conf_select.select(params.stack_name, "storm", params.version)
-      stack_select.select("storm-client", params.version)
-      stack_select.select("storm-nimbus", params.version)
-
-
-  def start(self, env, upgrade_type=None):
-    import params
-    env.set_params(params)
-    self.configure(env)
-    setup_ranger_storm(upgrade_type=upgrade_type)
-    service("nimbus", action="start")
-
-
-  def stop(self, env, upgrade_type=None):
-    import params
-    env.set_params(params)
-    service("nimbus", action="stop")
-
-
-  def status(self, env):
-    import status_params
-    env.set_params(status_params)
-    check_process_status(status_params.pid_nimbus)
-
-
-  def security_status(self, env):
-    import status_params
-    env.set_params(status_params)
-    if status_params.security_enabled:
-      # Expect the following files to be available in status_params.config_dir:
-      #   storm_jaas.conf
-      try:
-        props_value_check = None
-        props_empty_check = ['StormServer/keyTab', 'StormServer/principal']
-        props_read_check = ['StormServer/keyTab']
-        storm_env_expectations = build_expectations('storm_jaas', props_value_check, props_empty_check,  props_read_check)
-        storm_expectations = {}
-        storm_expectations.update(storm_env_expectations)
-        security_params = get_params_from_filesystem(status_params.conf_dir, {'storm_jaas.conf': FILE_TYPE_JAAS_CONF})
-        result_issues = validate_security_config_properties(security_params, storm_expectations)
-        if not result_issues:  # If all validations passed successfully
-          # Double check the dict before calling execute
-          if ( 'storm_jaas' not in security_params
-               or 'StormServer' not in security_params['storm_jaas']
-               or 'keyTab' not in security_params['storm_jaas']['StormServer']
-               or 'principal' not in security_params['storm_jaas']['StormServer']):
-            self.put_structured_out({"securityState": "ERROR"})
-            self.put_structured_out({"securityIssuesFound": "Keytab file or principal are not set property."})
-            return
-
-          cached_kinit_executor(status_params.kinit_path_local,
-                                status_params.storm_user,
-                                security_params['storm_jaas']['StormServer']['keyTab'],
-                                security_params['storm_jaas']['StormServer']['principal'],
-                                status_params.hostname,
-                                status_params.tmp_dir)
-          self.put_structured_out({"securityState": "SECURED_KERBEROS"})
-        else:
-          issues = []
-          for cf in result_issues:
-            issues.append("Configuration file %s did not pass the validation. Reason: %s" % (cf, result_issues[cf]))
-          self.put_structured_out({"securityIssuesFound": ". ".join(issues)})
-          self.put_structured_out({"securityState": "UNSECURED"})
-      except Exception as e:
-        self.put_structured_out({"securityState": "ERROR"})
-        self.put_structured_out({"securityStateErrorInfo": str(e)})
-    else:
-      self.put_structured_out({"securityState": "UNSECURED"})
-
-  def get_log_folder(self):
-    import params
-    return params.log_dir
-  
-  def get_user(self):
-    import params
-    return params.storm_user
-
-@OsFamilyImpl(os_family=OSConst.WINSRV_FAMILY)
-class NimbusWindows(Nimbus):
-  def start(self, env):
-    import status_params
-    env.set_params(status_params)
-    Service(status_params.nimbus_win_service_name, action="start")
-
-  def stop(self, env):
-    import status_params
-    env.set_params(status_params)
-    Service(status_params.nimbus_win_service_name, action="stop")
-
-  def status(self, env):
-    import status_params
-    from resource_management.libraries.functions.windows_service_utils import check_windows_service_status
-    env.set_params(status_params)
-    check_windows_service_status(status_params.nimbus_win_service_name)
-
-if __name__ == "__main__":
-  Nimbus().execute()

http://git-wip-us.apache.org/repos/asf/ambari/blob/d216c128/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/scripts/nimbus_prod.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/scripts/nimbus_prod.py b/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/scripts/nimbus_prod.py
deleted file mode 100644
index 39bda4d..0000000
--- a/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/scripts/nimbus_prod.py
+++ /dev/null
@@ -1,81 +0,0 @@
-#!/usr/bin/env python
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-"""
-
-import sys
-from resource_management.libraries.script import Script
-from storm import storm
-from supervisord_service import supervisord_service, supervisord_check_status
-from resource_management.libraries.functions import conf_select
-from resource_management.libraries.functions import stack_select
-from resource_management.libraries.functions import format
-from resource_management.core.resources.system import Execute
-from resource_management.libraries.functions.stack_features import check_stack_feature
-from resource_management.libraries.functions import StackFeature
-
-class Nimbus(Script):
-
-  def get_component_name(self):
-    return "storm-nimbus"
-
-  def install(self, env):
-    self.install_packages(env)
-    self.configure(env)
-
-  def configure(self, env):
-    import params
-    env.set_params(params)
-
-    storm()
-
-  def pre_upgrade_restart(self, env, upgrade_type=None):
-    import params
-    env.set_params(params)
-
-    if params.version and check_stack_feature(StackFeature.ROLLING_UPGRADE, params.version):
-      conf_select.select(params.stack_name, "storm", params.version)
-      stack_select.select("storm-client", params.version)
-      stack_select.select("storm-nimbus", params.version)
-
-  def start(self, env, upgrade_type=None):
-    import params
-    env.set_params(params)
-    self.configure(env)
-
-    supervisord_service("nimbus", action="start")
-
-  def stop(self, env, upgrade_type=None):
-    import params
-    env.set_params(params)
-
-    supervisord_service("nimbus", action="stop")
-
-  def status(self, env):
-    supervisord_check_status("nimbus")
-    
-  def get_log_folder(self):
-    import params
-    return params.log_dir
-  
-  def get_user(self):
-    import params
-    return params.storm_user
-
-if __name__ == "__main__":
-  Nimbus().execute()

http://git-wip-us.apache.org/repos/asf/ambari/blob/d216c128/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/scripts/pacemaker.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/scripts/pacemaker.py b/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/scripts/pacemaker.py
deleted file mode 100644
index 2ea02e8..0000000
--- a/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/scripts/pacemaker.py
+++ /dev/null
@@ -1,138 +0,0 @@
-#!/usr/bin/env python
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-"""
-
-import sys
-from resource_management.libraries.functions import check_process_status
-from resource_management.libraries.script import Script
-from resource_management.libraries.functions import conf_select
-from resource_management.libraries.functions import stack_select
-from resource_management.libraries.functions import format
-from resource_management.core.resources.system import Execute
-from resource_management.libraries.functions.stack_features import check_stack_feature
-from resource_management.libraries.functions import StackFeature
-from storm import storm
-from service import service
-from service_check import ServiceCheck
-from resource_management.libraries.functions.security_commons import build_expectations, \
-    cached_kinit_executor, get_params_from_filesystem, validate_security_config_properties, \
-    FILE_TYPE_JAAS_CONF
-
-class PaceMaker(Script):
-
-  def get_component_name(self):
-      return "storm-client"
-
-  def install(self, env):
-      self.install_packages(env)
-      self.configure(env)
-
-  def configure(self, env):
-      import params
-      env.set_params(params)
-      storm()
-
-  def pre_upgrade_restart(self, env, upgrade_type=None):
-      import params
-      env.set_params(params)
-
-      if params.version and check_stack_feature(StackFeature.ROLLING_UPGRADE, params.version):
-        conf_select.select(params.stack_name, "storm", params.version)
-        stack_select.select("storm-client", params.version)
-
-  def start(self, env, upgrade_type=None):
-      import params
-      env.set_params(params)
-      self.configure(env)
-
-      service("pacemaker", action="start")
-
-  def stop(self, env, upgrade_type=None):
-      import params
-      env.set_params(params)
-
-      service("pacemaker", action="stop")
-
-  def status(self, env):
-      import status_params
-      env.set_params(status_params)
-      check_process_status(status_params.pid_pacemaker)
-
-  def security_status(self, env):
-      import status_params
-
-      env.set_params(status_params)
-
-      if status_params.security_enabled:
-          # Expect the following files to be available in status_params.config_dir:
-          #   storm_jaas.conf
-
-          try:
-              props_value_check = None
-              props_empty_check = ['StormServer/keyTab', 'StormServer/principal']
-              props_read_check = ['StormServer/keyTab']
-              storm_env_expectations = build_expectations('storm_jaas', props_value_check, props_empty_check,
-                                                          props_read_check)
-
-              storm_expectations = {}
-              storm_expectations.update(storm_env_expectations)
-
-              security_params = get_params_from_filesystem(status_params.conf_dir,
-                                                     {'storm_jaas.conf': FILE_TYPE_JAAS_CONF})
-
-              result_issues = validate_security_config_properties(security_params, storm_expectations)
-              if not result_issues:  # If all validations passed successfully
-                  # Double check the dict before calling execute
-                  if ( 'storm_jaas' not in security_params
-                       or 'StormServer' not in security_params['storm_jaas']
-                       or 'keyTab' not in security_params['storm_jaas']['StormServer']
-                       or 'principal' not in security_params['storm_jaas']['StormServer']):
-                      self.put_structured_out({"securityState": "ERROR"})
-                      self.put_structured_out({"securityIssuesFound": "Keytab file or principal are not set property."})
-                      return
-
-                  cached_kinit_executor(status_params.kinit_path_local,
-                                        status_params.storm_user,
-                                        security_params['storm_jaas']['StormServer']['keyTab'],
-                                        security_params['storm_jaas']['StormServer']['principal'],
-                                        status_params.hostname,
-                                        status_params.tmp_dir)
-                  self.put_structured_out({"securityState": "SECURED_KERBEROS"})
-              else:
-                  issues = []
-                  for cf in result_issues:
-                      issues.append("Configuration file %s did not pass the validation. Reason: %s" % (cf, result_issues[cf]))
-                      self.put_structured_out({"securityIssuesFound": ". ".join(issues)})
-                      self.put_structured_out({"securityState": "UNSECURED"})
-          except Exception as e:
-              self.put_structured_out({"securityState": "ERROR"})
-              self.put_structured_out({"securityStateErrorInfo": str(e)})
-      else:
-        self.put_structured_out({"securityState": "UNSECURED"})
-
-  def get_log_folder(self):
-      import params
-      return params.log_dir
-
-  def get_user(self):
-      import params
-      return params.storm_user
-
-if __name__ == "__main__":
-    PaceMaker().execute()

http://git-wip-us.apache.org/repos/asf/ambari/blob/d216c128/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/scripts/params.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/scripts/params.py b/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/scripts/params.py
deleted file mode 100644
index f10a3f3..0000000
--- a/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/scripts/params.py
+++ /dev/null
@@ -1,29 +0,0 @@
-#!/usr/bin/env python
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-"""
-from ambari_commons import OSCheck
-from resource_management.libraries.functions.default import default
-
-if OSCheck.is_windows_family():
-  from params_windows import *
-else:
-  from params_linux import *
-
-host_sys_prepped = default("/hostLevelParams/host_sys_prepped", False)
-retryAble = default("/commandParams/command_retry_enabled", False)

http://git-wip-us.apache.org/repos/asf/ambari/blob/d216c128/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/scripts/params_linux.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/scripts/params_linux.py b/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/scripts/params_linux.py
deleted file mode 100644
index d715a25..0000000
--- a/ambari-server/src/main/resources/common-services/STORM/0.9.1/package/scripts/params_linux.py
+++ /dev/null
@@ -1,351 +0,0 @@
-#!/usr/bin/env python
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-"""
-import os
-import re
-import ambari_simplejson as json # simplejson is much faster comparing to Python 2.6 json module and has the same functions set.
-
-import status_params
-
-from ambari_commons.constants import AMBARI_SUDO_BINARY
-from resource_management.libraries.functions.constants import Direction
-from resource_management.libraries.functions import format
-from resource_management.libraries.functions.version import format_stack_version
-from resource_management.libraries.functions.default import default
-from resource_management.libraries.functions.get_bare_principal import get_bare_principal
-from resource_management.libraries.script import Script
-from resource_management.libraries.resources.hdfs_resource import HdfsResource
-from resource_management.libraries.functions import stack_select
-from resource_management.libraries.functions import conf_select
-from resource_management.libraries.functions import get_kinit_path
-from resource_management.libraries.functions.get_not_managed_resources import get_not_managed_resources
-from resource_management.libraries.functions.stack_features import check_stack_feature
-from resource_management.libraries.functions import StackFeature
-from resource_management.libraries.functions.expect import expect
-
-# server configurations
-config = Script.get_config()
-tmp_dir = Script.get_tmp_dir()
-stack_root = status_params.stack_root
-sudo = AMBARI_SUDO_BINARY
-
-cluster_name = config['clusterName']
-
-stack_name = status_params.stack_name
-upgrade_direction = default("/commandParams/upgrade_direction", Direction.UPGRADE)
-version = default("/commandParams/version", None)
-
-agent_stack_retry_on_unavailability = config['hostLevelParams']['agent_stack_retry_on_unavailability']
-agent_stack_retry_count = expect("/hostLevelParams/agent_stack_retry_count", int)
-
-storm_component_home_dir = status_params.storm_component_home_dir
-conf_dir = status_params.conf_dir
-
-stack_version_unformatted = status_params.stack_version_unformatted
-stack_version_formatted = status_params.stack_version_formatted
-stack_supports_ru = stack_version_formatted and check_stack_feature(StackFeature.ROLLING_UPGRADE, stack_version_formatted)
-stack_supports_storm_kerberos = stack_version_formatted and check_stack_feature(StackFeature.STORM_KERBEROS, stack_version_formatted)
-stack_supports_storm_ams = stack_version_formatted and check_stack_feature(StackFeature.STORM_AMS, stack_version_formatted)
-stack_supports_ranger_kerberos = stack_version_formatted and check_stack_feature(StackFeature.RANGER_KERBEROS_SUPPORT, stack_version_formatted)
-
-# default hadoop params
-rest_lib_dir = "/usr/lib/storm/contrib/storm-rest"
-storm_bin_dir = "/usr/bin"
-storm_lib_dir = "/usr/lib/storm/lib/"
-
-# hadoop parameters for 2.2+
-if stack_supports_ru:
-  rest_lib_dir = format("{storm_component_home_dir}/contrib/storm-rest")
-  storm_bin_dir = format("{storm_component_home_dir}/bin")
-  storm_lib_dir = format("{storm_component_home_dir}/lib")
-  log4j_dir = format("{storm_component_home_dir}/log4j2")
-
-storm_user = config['configurations']['storm-env']['storm_user']
-log_dir = config['configurations']['storm-env']['storm_log_dir']
-pid_dir = status_params.pid_dir
-local_dir = config['configurations']['storm-site']['storm.local.dir']
-user_group = config['configurations']['cluster-env']['user_group']
-java64_home = config['hostLevelParams']['java_home']
-jps_binary = format("{java64_home}/bin/jps")
-nimbus_port = config['configurations']['storm-site']['nimbus.thrift.port']
-storm_zookeeper_root_dir = default('/configurations/storm-site/storm.zookeeper.root', None)
-storm_zookeeper_servers = config['configurations']['storm-site']['storm.zookeeper.servers']
-storm_zookeeper_port = config['configurations']['storm-site']['storm.zookeeper.port']
-storm_logs_supported = config['configurations']['storm-env']['storm_logs_supported']
-
-# nimbus.seeds is supported in HDP 2.3.0.0 and higher
-nimbus_seeds_supported = default('/configurations/storm-env/nimbus_seeds_supported', False)
-nimbus_host = default('/configurations/storm-site/nimbus.host', None)
-nimbus_seeds = default('/configurations/storm-site/nimbus.seeds', None)
-default_topology_max_replication_wait_time_sec = default('/configurations/storm-site/topology.max.replication.wait.time.sec.default', -1)
-nimbus_hosts = default("/clusterHostInfo/nimbus_hosts", [])
-default_topology_min_replication_count = default('/configurations/storm-site/topology.min.replication.count.default', 1)
-
-#Calculate topology.max.replication.wait.time.sec and topology.min.replication.count
-if len(nimbus_hosts) > 1:
-  # for HA Nimbus
-  actual_topology_max_replication_wait_time_sec = -1
-  actual_topology_min_replication_count = len(nimbus_hosts) / 2 + 1
-else:
-  # for non-HA Nimbus
-  actual_topology_max_replication_wait_time_sec = default_topology_max_replication_wait_time_sec
-  actual_topology_min_replication_count = default_topology_min_replication_count
-
-if 'topology.max.replication.wait.time.sec.default' in config['configurations']['storm-site']:
-  del config['configurations']['storm-site']['topology.max.replication.wait.time.sec.default']
-if 'topology.min.replication.count.default' in config['configurations']['storm-site']:
-  del config['configurations']['storm-site']['topology.min.replication.count.default']
-
-rest_api_port = "8745"
-rest_api_admin_port = "8746"
-rest_api_conf_file = format("{conf_dir}/config.yaml")
-storm_env_sh_template = config['configurations']['storm-env']['content']
-jmxremote_port = config['configurations']['storm-env']['jmxremote_port']
-
-if 'ganglia_server_host' in config['clusterHostInfo'] and len(config['clusterHostInfo']['ganglia_server_host'])>0:
-  ganglia_installed = True
-  ganglia_server = config['clusterHostInfo']['ganglia_server_host'][0]
-  ganglia_report_interval = 60
-else:
-  ganglia_installed = False
-
-security_enabled = config['configurations']['cluster-env']['security_enabled']
-
-storm_ui_host = default("/clusterHostInfo/storm_ui_server_hosts", [])
-
-storm_user_nofile_limit = default('/configurations/storm-env/storm_user_nofile_limit', 128000)
-storm_user_nproc_limit = default('/configurations/storm-env/storm_user_noproc_limit', 65536)
-
-if security_enabled:
-  _hostname_lowercase = config['hostname'].lower()
-  _storm_principal_name = config['configurations']['storm-env']['storm_principal_name']
-  storm_jaas_principal = _storm_principal_name.replace('_HOST',_hostname_lowercase)
-  storm_keytab_path = config['configurations']['storm-env']['storm_keytab']
-
-  if stack_supports_storm_kerberos:
-    storm_ui_keytab_path = config['configurations']['storm-env']['storm_ui_keytab']
-    _storm_ui_jaas_principal_name = config['configurations']['storm-env']['storm_ui_principal_name']
-    storm_ui_jaas_principal = _storm_ui_jaas_principal_name.replace('_HOST',_hostname_lowercase)
-    storm_bare_jaas_principal = get_bare_principal(_storm_principal_name)
-    _nimbus_principal_name = config['configurations']['storm-env']['nimbus_principal_name']
-    nimbus_jaas_principal = _nimbus_principal_name.replace('_HOST', _hostname_lowercase)
-    nimbus_bare_jaas_principal = get_bare_principal(_nimbus_principal_name)
-    nimbus_keytab_path = config['configurations']['storm-env']['nimbus_keytab']
-
-kafka_bare_jaas_principal = None
-if stack_supports_storm_kerberos:
-  if security_enabled:
-    storm_thrift_transport = config['configurations']['storm-site']['_storm.thrift.secure.transport']
-    # generate KafkaClient jaas config if kafka is kerberoized
-    _kafka_principal_name = default("/configurations/kafka-env/kafka_principal_name", None)
-    kafka_bare_jaas_principal = get_bare_principal(_kafka_principal_name)
-  else:
-    storm_thrift_transport = config['configurations']['storm-site']['_storm.thrift.nonsecure.transport']
-
-ams_collector_hosts = default("/clusterHostInfo/metrics_collector_hosts", [])
-has_metric_collector = not len(ams_collector_hosts) == 0
-if has_metric_collector:
-  if 'cluster-env' in config['configurations'] and \
-      'metrics_collector_vip_host' in config['configurations']['cluster-env']:
-    metric_collector_host = config['configurations']['cluster-env']['metrics_collector_vip_host']
-  else:
-    metric_collector_host = ams_collector_hosts[0]
-  if 'cluster-env' in config['configurations'] and \
-      'metrics_collector_vip_port' in config['configurations']['cluster-env']:
-    metric_collector_port = config['configurations']['cluster-env']['metrics_collector_vip_port']
-  else:
-    metric_collector_web_address = default("/configurations/ams-site/timeline.metrics.service.webapp.address", "localhost:6188")
-    if metric_collector_web_address.find(':') != -1:
-      metric_collector_port = metric_collector_web_address.split(':')[1]
-    else:
-      metric_collector_port = '6188'
-
-  metric_collector_report_interval = 60
-  metric_collector_app_id = "nimbus"
-  if default("/configurations/ams-site/timeline.metrics.service.http.policy", "HTTP_ONLY") == "HTTPS_ONLY":
-    metric_collector_protocol = 'https'
-  else:
-    metric_collector_protocol = 'http'
-  metric_truststore_path= default("/configurations/ams-ssl-client/ssl.client.truststore.location", "")
-  metric_truststore_type= default("/configurations/ams-ssl-client/ssl.client.truststore.type", "")
-  metric_truststore_password= default("/configurations/ams-ssl-client/ssl.client.truststore.password", "")
-  pass
-metrics_report_interval = default("/configurations/ams-site/timeline.metrics.sink.report.interval", 60)
-metrics_collection_period = default("/configurations/ams-site/timeline.metrics.sink.collection.period", 10)
-metric_collector_sink_jar = "/usr/lib/storm/lib/ambari-metrics-storm-sink*.jar"
-
-jar_jvm_opts = ''
-
-# Atlas related params
-atlas_hosts = default('/clusterHostInfo/atlas_server_hosts', [])
-has_atlas = len(atlas_hosts) > 0
-atlas_plugin_package = "atlas-metadata*-hive-plugin"
-atlas_ubuntu_plugin_package = "atlas-metadata.*-hive-plugin"
-
-if has_atlas:
-  atlas_home_dir = os.environ['METADATA_HOME_DIR'] if 'METADATA_HOME_DIR' in os.environ else stack_root + '/current/atlas-server'
-  atlas_conf_dir = os.environ['METADATA_CONF'] if 'METADATA_CONF' in os.environ else '/etc/atlas/conf'
-  jar_jvm_opts = '-Datlas.conf=' + atlas_conf_dir
-
-# ranger host
-stack_supports_ranger_audit_db = stack_version_formatted and check_stack_feature(StackFeature.RANGER_AUDIT_DB_SUPPORT, stack_version_formatted)
-ranger_admin_hosts = default("/clusterHostInfo/ranger_admin_hosts", [])
-has_ranger_admin = not len(ranger_admin_hosts) == 0
-xml_configurations_supported = config['configurations']['ranger-env']['xml_configurations_supported']
-ambari_server_hostname = config['clusterHostInfo']['ambari_server_host'][0]
-
-#ranger storm properties
-policymgr_mgr_url = config['configurations']['admin-properties']['policymgr_external_url']
-xa_audit_db_name = config['configurations']['admin-properties']['audit_db_name']
-xa_audit_db_user = config['configurations']['admin-properties']['audit_db_user']
-xa_db_host = config['configurations']['admin-properties']['db_host']
-repo_name = str(config['clusterName']) + '_storm'
-
-common_name_for_certificate = config['configurations']['ranger-storm-plugin-properties']['common.name.for.certificate']
-
-storm_ui_port = config['configurations']['storm-site']['ui.port']
-
-repo_config_username = config['configurations']['ranger-storm-plugin-properties']['REPOSITORY_CONFIG_USERNAME']
-ranger_env = config['configurations']['ranger-env']
-ranger_plugin_properties = config['configurations']['ranger-storm-plugin-properties']
-policy_user = config['configurations']['ranger-storm-plugin-properties']['policy_user']
-storm_cluster_log4j_content = config['configurations']['storm-cluster-log4j']['content']
-storm_worker_log4j_content = config['configurations']['storm-worker-log4j']['content']
-
-# some commands may need to supply the JAAS location when running as storm
-storm_jaas_file = format("{conf_dir}/storm_jaas.conf")
-
-# For curl command in ranger plugin to get db connector
-jdk_location = config['hostLevelParams']['jdk_location']
-java_share_dir = '/usr/share/java'
-
-if has_ranger_admin:
-  enable_ranger_storm = (config['configurations']['ranger-storm-plugin-properties']['ranger-storm-plugin-enabled'].lower() == 'yes')
-  xa_audit_db_password = unicode(config['configurations']['admin-properties']['audit_db_password']) if stack_supports_ranger_audit_db else None
-  repo_config_password = unicode(config['configurations']['ranger-storm-plugin-properties']['REPOSITORY_CONFIG_PASSWORD'])
-  xa_audit_db_flavor = (config['configurations']['admin-properties']['DB_FLAVOR']).lower()
-
-  if stack_supports_ranger_audit_db:
-    if xa_audit_db_flavor == 'mysql':
-      jdbc_jar_name = default("/hostLevelParams/custom_mysql_jdbc_name", None)
-      audit_jdbc_url = format('jdbc:mysql://{xa_db_host}/{xa_audit_db_name}')
-      jdbc_driver = "com.mysql.jdbc.Driver"
-    elif xa_audit_db_flavor == 'oracle':
-      jdbc_jar_name = default("/hostLevelParams/custom_oracle_jdbc_name", None)
-      colon_count = xa_db_host.count(':')
-      if colon_count == 2 or colon_count == 0:
-        audit_jdbc_url = format('jdbc:oracle:thin:@{xa_db_host}')
-      else:
-        audit_jdbc_url = format('jdbc:oracle:thin:@//{xa_db_host}')
-      jdbc_driver = "oracle.jdbc.OracleDriver"
-    elif xa_audit_db_flavor == 'postgres':
-      jdbc_jar_name = default("/hostLevelParams/custom_postgres_jdbc_name", None)
-      audit_jdbc_url = format('jdbc:postgresql://{xa_db_host}/{xa_audit_db_name}')
-      jdbc_driver = "org.postgresql.Driver"
-    elif xa_audit_db_flavor == 'mssql':
-      jdbc_jar_name = default("/hostLevelParams/custom_mssql_jdbc_name", None)
-      audit_jdbc_url = format('jdbc:sqlserver://{xa_db_host};databaseName={xa_audit_db_name}')
-      jdbc_driver = "com.microsoft.sqlserver.jdbc.SQLServerDriver"
-    elif xa_audit_db_flavor == 'sqla':
-      jdbc_jar_name = default("/hostLevelParams/custom_sqlanywhere_jdbc_name", None)
-      audit_jdbc_url = format('jdbc:sqlanywhere:database={xa_audit_db_name};host={xa_db_host}')
-      jdbc_driver = "sap.jdbc4.sqlanywhere.IDriver"
-
-  downloaded_custom_connector = format("{tmp_dir}/{jdbc_jar_name}") if stack_supports_ranger_audit_db else None
-  driver_curl_source = format("{jdk_location}/{jdbc_jar_name}") if stack_supports_ranger_audit_db else None
-  driver_curl_target = format("{storm_component_home_dir}/lib/{jdbc_jar_name}") if stack_supports_ranger_audit_db else None
-  sql_connector_jar = ''
-
-  storm_ranger_plugin_config = {
-    'username': repo_config_username,
-    'password': repo_config_password,
-    'nimbus.url': 'http://' + storm_ui_host[0].lower() + ':' + str(storm_ui_port),
-    'commonNameForCertificate': common_name_for_certificate
-  }
-
-  storm_ranger_plugin_repo = {
-    'isActive': 'true',
-    'config': json.dumps(storm_ranger_plugin_config),
-    'description': 'storm repo',
-    'name': repo_name,
-    'repositoryType': 'storm',
-    'assetType': '6'
-  }
-
-  if stack_supports_ranger_kerberos and security_enabled:
-    storm_ranger_plugin_config['policy.download.auth.users'] = storm_user
-    storm_ranger_plugin_config['tag.download.auth.users'] = storm_user
-
-    storm_ranger_plugin_repo = {
-      'isEnabled': 'true',
-      'configs': storm_ranger_plugin_config,
-      'description': 'storm repo',
-      'name': repo_name,
-      'type': 'storm'
-    }
-
-  if stack_supports_ranger_kerberos and security_enabled and 'storm-nimbus' in status_params.component_directory.lower():
-    ranger_storm_principal = nimbus_jaas_principal
-    ranger_storm_keytab = nimbus_keytab_path
-  elif stack_supports_ranger_kerberos and security_enabled and 'storm-client' in status_params.component_directory.lower():
-    ranger_storm_principal = storm_ui_jaas_principal
-    ranger_storm_keytab = storm_ui_keytab_path
-
-  xa_audit_db_is_enabled = False
-  ranger_audit_solr_urls = config['configurations']['ranger-admin-site']['ranger.audit.solr.urls']
-  if xml_configurations_supported and stack_supports_ranger_audit_db:
-    xa_audit_db_is_enabled = config['configurations']['ranger-storm-audit']['xasecure.audit.destination.db']
-  xa_audit_hdfs_is_enabled = config['configurations']['ranger-storm-audit']['xasecure.audit.destination.hdfs'] if xml_configurations_supported else None
-  ssl_keystore_password = unicode(config['configurations']['ranger-storm-policymgr-ssl']['xasecure.policymgr.clientssl.keystore.password']) if xml_configurations_supported else None
-  ssl_truststore_password = unicode(config['configurations']['ranger-storm-policymgr-ssl']['xasecure.policymgr.clientssl.truststore.password']) if xml_configurations_supported else None
-  credential_file = format('/etc/ranger/{repo_name}/cred.jceks') if xml_configurations_supported else None
-
-  #For SQLA explicitly disable audit to DB for Ranger
-  if xa_audit_db_flavor == 'sqla':
-    xa_audit_db_is_enabled = False
-
-namenode_hosts = default("/clusterHostInfo/namenode_host", [])
-has_namenode = not len(namenode_hosts) == 0
-
-hdfs_user = config['configurations']['hadoop-env']['hdfs_user'] if has_namenode else None
-hdfs_user_keytab = config['configurations']['hadoop-env']['hdfs_user_keytab'] if has_namenode else None
-hdfs_principal_name = config['configurations']['hadoop-env']['hdfs_principal_name'] if has_namenode else None
-hdfs_site = config['configurations']['hdfs-site'] if has_namenode else None
-default_fs = config['configurations']['core-site']['fs.defaultFS'] if has_namenode else None
-hadoop_bin_dir = stack_select.get_hadoop_dir("bin") if has_namenode else None
-hadoop_conf_dir = conf_select.get_hadoop_conf_dir() if has_namenode else None
-kinit_path_local = get_kinit_path(default('/configurations/kerberos-env/executable_search_paths', None))
-
-import functools
-#create partial functions with common arguments for every HdfsResource call
-#to create/delete hdfs directory/file/copyfromlocal we need to call params.HdfsResource in code
-HdfsResource = functools.partial(
-  HdfsResource,
-  user=hdfs_user,
-  hdfs_resource_ignore_file = "/var/lib/ambari-agent/data/.hdfs_resource_ignore",
-  security_enabled = security_enabled,
-  keytab = hdfs_user_keytab,
-  kinit_path_local = kinit_path_local,
-  hadoop_bin_dir = hadoop_bin_dir,
-  hadoop_conf_dir = hadoop_conf_dir,
-  principal_name = hdfs_principal_name,
-  hdfs_site = hdfs_site,
-  default_fs = default_fs,
-  immutable_paths = get_not_managed_resources()
-)


Mime
View raw message