ambari-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From jlun...@apache.org
Subject [19/37] ambari git commit: AMBARI-8745: Common Services: Refactor HDP2.0.6 FLUME, GANGLIA, HBASE services (Jayush Luniya)
Date Wed, 17 Dec 2014 20:05:51 GMT
http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/metrics.json
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/metrics.json b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/metrics.json
deleted file mode 100644
index 2114c12..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/metrics.json
+++ /dev/null
@@ -1,716 +0,0 @@
-{
-  "FLUME_HANDLER": {
-    "Component": [
-        {
-        "type": "ganglia",
-        "metrics": {
-          "metrics/boottime":{
-            "metric":"boottime",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/cpu/cpu_aidle":{
-            "metric":"cpu_aidle",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/cpu/cpu_idle":{
-            "metric":"cpu_idle",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/cpu/cpu_nice":{
-            "metric":"cpu_nice",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/cpu/cpu_num":{
-            "metric":"cpu_num",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/cpu/cpu_speed":{
-            "metric":"cpu_speed",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/cpu/cpu_system":{
-            "metric":"cpu_system",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/cpu/cpu_user":{
-            "metric":"cpu_user",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/cpu/cpu_wio":{
-            "metric":"cpu_wio",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/disk/disk_free":{
-            "metric":"disk_free",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/disk/disk_total":{
-            "metric":"disk_total",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/disk/part_max_used":{
-            "metric":"part_max_used",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/gcCount":{
-            "metric":"jvm.metrics.gcCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/gcTimeMillis":{
-            "metric":"jvm.metrics.gcTimeMillis",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/logError":{
-            "metric":"jvm.metrics.logError",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/logFatal":{
-            "metric":"jvm.metrics.logFatal",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/logInfo":{
-            "metric":"jvm.metrics.logInfo",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/logWarn":{
-            "metric":"jvm.metrics.logWarn",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/maxMemoryM":{
-            "metric":"jvm.metrics.maxMemoryM",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/memHeapCommittedM":{
-            "metric":"jvm.metrics.memHeapCommittedM",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/memHeapUsedM":{
-            "metric":"jvm.metrics.memHeapUsedM",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/memNonHeapCommittedM":{
-            "metric":"jvm.metrics.memNonHeapCommittedM",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/memNonHeapUsedM":{
-            "metric":"jvm.metrics.memNonHeapUsedM",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/threadsBlocked":{
-            "metric":"jvm.metrics.threadsBlocked",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/threadsNew":{
-            "metric":"jvm.metrics.threadsNew",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/threadsRunnable":{
-            "metric":"jvm.metrics.threadsRunnable",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/threadsTerminated":{
-            "metric":"jvm.metrics.threadsTerminated",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/threadsTimedWaiting":{
-            "metric":"jvm.metrics.threadsTimedWaiting",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/threadsWaiting":{
-            "metric":"jvm.metrics.threadsWaiting",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/CHANNEL/$2/ChannelCapacity":{
-            "metric":"(\\w+).CHANNEL.(\\w+).ChannelCapacity",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/CHANNEL/$2/StartTime":{
-            "metric":"(\\w+).CHANNEL.(\\w+).StartTime",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/CHANNEL/$2/EventTakeAttemptCount":{
-            "metric":"(\\w+).CHANNEL.(\\w+).EventTakeAttemptCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/CHANNEL/$2/EventTakeSuccessCount":{
-            "metric":"(\\w+).CHANNEL.(\\w+).EventTakeSuccessCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/CHANNEL/$2/EventPutAttemptCount":{
-            "metric":"(\\w+).CHANNEL.(\\w+).EventPutAttemptCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/CHANNEL/$2/StopTime":{
-            "metric":"(\\w+).CHANNEL.(\\w+).StopTime",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/CHANNEL/$2/ChannelFillPercentage":{
-            "metric":"(\\w+).CHANNEL.(\\w+).ChannelFillPercentage",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/CHANNEL/$2/ChannelSize":{
-            "metric":"(\\w+).CHANNEL.(\\w+).ChannelSize",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/CHANNEL/$2/EventPutSuccessCount":{
-            "metric":"(\\w+).CHANNEL.(\\w+).EventPutSuccessCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SINK/$2/ConnectionCreatedCount":{
-            "metric":"(\\w+).SINK.(\\w+).ConnectionCreatedCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SINK/$2/BatchCompleteCount":{
-            "metric":"(\\w+).SINK.(\\w+).BatchCompleteCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SINK/$2/EventDrainSuccessCount":{
-            "metric":"(\\w+).SINK.(\\w+).EventDrainSuccessCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SINK/$2/StartTime":{
-            "metric":"(\\w+).SINK.(\\w+).StartTime",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SINK/$2/EventDrainAttemptCount":{
-            "metric":"(\\w+).SINK.(\\w+).EventDrainAttemptCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SINK/$2/ConnectionFailedCount":{
-            "metric":"(\\w+).SINK.(\\w+).ConnectionFailedCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SINK/$2/BatchUnderflowCount":{
-            "metric":"(\\w+).SINK.(\\w+).BatchUnderflowCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SINK/$2/ConnectionClosedCount":{
-            "metric":"(\\w+).SINK.(\\w+).ConnectionClosedCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SINK/$2/StopTime":{
-            "metric":"(\\w+).SINK.(\\w+).StopTime",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SINK/$2/BatchEmptyCount":{
-            "metric":"(\\w+).SINK.(\\w+).BatchEmptyCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SOURCE/$2/AppendBatchReceivedCount":{
-            "metric":"(\\w+).SOURCE.(\\w+).AppendBatchReceivedCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SOURCE/$2/AppendAcceptedCount":{
-            "metric":"(\\w+).SOURCE.(\\w+).AppendAcceptedCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SOURCE/$2/StartTime":{
-            "metric":"(\\w+).SOURCE.(\\w+).StartTime",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SOURCE/$2/OpenConnectionCount":{
-            "metric":"(\\w+).SOURCE.(\\w+).OpenConnectionCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SOURCE/$2/AppendBatchAcceptedCount":{
-            "metric":"(\\w+).SOURCE.(\\w+).AppendBatchAcceptedCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SOURCE/$2/AppendReceivedCount":{
-            "metric":"(\\w+).SOURCE.(\\w+).AppendReceivedCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SOURCE/$2/EventReceivedCount":{
-            "metric":"(\\w+).SOURCE.(\\w+).EventReceivedCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SOURCE/$2/StopTime":{
-            "metric":"(\\w+).SOURCE.(\\w+).StopTime",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SOURCE/$2/EventAcceptedCount":{
-            "metric":"(\\w+).SOURCE.(\\w+).EventAcceptedCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-
-          "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventTakeSuccessCount/rate/min": {
-            "metric":"(\\w+).CHANNEL.(\\w+).EventTakeSuccessCount._rate._min",
-            "pointInTime":false,
-            "temporal":true
-          },
-          "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventTakeSuccessCount/rate/max": {
-            "metric":"(\\w+).CHANNEL.(\\w+).EventTakeSuccessCount._rate._max",
-            "pointInTime":false,
-            "temporal":true
-          },
-          "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventTakeSuccessCount/rate/avg": {
-            "metric":"(\\w+).CHANNEL.(\\w+).EventTakeSuccessCount._rate._avg",
-            "pointInTime":false,
-            "temporal":true
-          },
-          "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventTakeSuccessCount/rate/sum": {
-            "metric":"(\\w+).CHANNEL.(\\w+).EventTakeSuccessCount._rate._sum",
-            "pointInTime":false,
-            "temporal":true
-          },
-
-          "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventPutSuccessCount/rate/avg": {
-            "metric":"(\\w+).CHANNEL.(\\w+).EventPutSuccessCount._rate._avg",
-            "pointInTime":false,
-            "temporal":true
-          },
-          "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventPutSuccessCount/rate/max": {
-            "metric":"(\\w+).CHANNEL.(\\w+).EventPutSuccessCount._rate._max",
-            "pointInTime":false,
-            "temporal":true
-          },
-          "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventPutSuccessCount/rate/min": {
-            "metric":"(\\w+).CHANNEL.(\\w+).EventPutSuccessCount._rate._min",
-            "pointInTime":false,
-            "temporal":true
-          },
-          "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventPutSuccessCount/rate/sum": {
-            "metric":"(\\w+).CHANNEL.(\\w+).EventPutSuccessCount._rate._sum",
-            "pointInTime":false,
-            "temporal":true
-          },
-          
-          "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")ChannelSize/avg": {
-            "metric":"(\\w+).CHANNEL.(\\w+).ChannelSize._avg",
-            "pointInTime":false,
-            "temporal":true
-          },
-          "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")ChannelSize/max": {
-            "metric":"(\\w+).CHANNEL.(\\w+).ChannelSize._max",
-            "pointInTime":false,
-            "temporal":true
-          },
-          "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")ChannelSize/min": {
-            "metric":"(\\w+).CHANNEL.(\\w+).ChannelSize._min",
-            "pointInTime":false,
-            "temporal":true
-          },
-          "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")ChannelSize/sum": {
-            "metric":"(\\w+).CHANNEL.(\\w+).ChannelSize._sum",
-            "pointInTime":false,
-            "temporal":true
-          }
-
-        }
-      }
-    ],
-    "HostComponent": [
-      {
-        "type": "ganglia",
-        "metrics": {
-          "metrics/boottime":{
-            "metric":"boottime",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/cpu/cpu_aidle":{
-            "metric":"cpu_aidle",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/cpu/cpu_idle":{
-            "metric":"cpu_idle",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/cpu/cpu_nice":{
-            "metric":"cpu_nice",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/cpu/cpu_num":{
-            "metric":"cpu_num",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/cpu/cpu_speed":{
-            "metric":"cpu_speed",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/cpu/cpu_system":{
-            "metric":"cpu_system",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/cpu/cpu_user":{
-            "metric":"cpu_user",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/cpu/cpu_wio":{
-            "metric":"cpu_wio",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/disk/disk_free":{
-            "metric":"disk_free",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/disk/disk_total":{
-            "metric":"disk_total",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/disk/part_max_used":{
-            "metric":"part_max_used",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/gcCount":{
-            "metric":"jvm.metrics.gcCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/gcTimeMillis":{
-            "metric":"jvm.metrics.gcTimeMillis",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/logError":{
-            "metric":"jvm.metrics.logError",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/logFatal":{
-            "metric":"jvm.metrics.logFatal",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/logInfo":{
-            "metric":"jvm.metrics.logInfo",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/logWarn":{
-            "metric":"jvm.metrics.logWarn",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/maxMemoryM":{
-            "metric":"jvm.metrics.maxMemoryM",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/memHeapCommittedM":{
-            "metric":"jvm.metrics.memHeapCommittedM",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/memHeapUsedM":{
-            "metric":"jvm.metrics.memHeapUsedM",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/memNonHeapCommittedM":{
-            "metric":"jvm.metrics.memNonHeapCommittedM",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/memNonHeapUsedM":{
-            "metric":"jvm.metrics.memNonHeapUsedM",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/threadsBlocked":{
-            "metric":"jvm.metrics.threadsBlocked",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/threadsNew":{
-            "metric":"jvm.metrics.threadsNew",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/threadsRunnable":{
-            "metric":"jvm.metrics.threadsRunnable",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/threadsTerminated":{
-            "metric":"jvm.metrics.threadsTerminated",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/threadsTimedWaiting":{
-            "metric":"jvm.metrics.threadsTimedWaiting",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/jvm/threadsWaiting":{
-            "metric":"jvm.metrics.threadsWaiting",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/CHANNEL/$2/ChannelCapacity":{
-            "metric":"(\\w+).CHANNEL.(\\w+).ChannelCapacity",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/CHANNEL/$2/StartTime":{
-            "metric":"(\\w+).CHANNEL.(\\w+).StartTime",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/CHANNEL/$2/EventTakeAttemptCount":{
-            "metric":"(\\w+).CHANNEL.(\\w+).EventTakeAttemptCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/CHANNEL/$2/EventTakeSuccessCount":{
-            "metric":"(\\w+).CHANNEL.(\\w+).EventTakeSuccessCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/CHANNEL/$2/EventPutAttemptCount":{
-            "metric":"(\\w+).CHANNEL.(\\w+).EventPutAttemptCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/CHANNEL/$2/StopTime":{
-            "metric":"(\\w+).CHANNEL.(\\w+).StopTime",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/CHANNEL/$2/ChannelFillPercentage":{
-            "metric":"(\\w+).CHANNEL.(\\w+).ChannelFillPercentage",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/CHANNEL/$2/ChannelSize":{
-            "metric":"(\\w+).CHANNEL.(\\w+).ChannelSize",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/CHANNEL/$2/EventPutSuccessCount":{
-            "metric":"(\\w+).CHANNEL.(\\w+).EventPutSuccessCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SINK/$2/ConnectionCreatedCount":{
-            "metric":"(\\w+).SINK.(\\w+).ConnectionCreatedCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SINK/$2/BatchCompleteCount":{
-            "metric":"(\\w+).SINK.(\\w+).BatchCompleteCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SINK/$2/EventDrainSuccessCount":{
-            "metric":"(\\w+).SINK.(\\w+).EventDrainSuccessCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SINK/$2/StartTime":{
-            "metric":"(\\w+).SINK.(\\w+).StartTime",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SINK/$2/EventDrainAttemptCount":{
-            "metric":"(\\w+).SINK.(\\w+).EventDrainAttemptCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SINK/$2/ConnectionFailedCount":{
-            "metric":"(\\w+).SINK.(\\w+).ConnectionFailedCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SINK/$2/BatchUnderflowCount":{
-            "metric":"(\\w+).SINK.(\\w+).BatchUnderflowCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SINK/$2/ConnectionClosedCount":{
-            "metric":"(\\w+).SINK.(\\w+).ConnectionClosedCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SINK/$2/StopTime":{
-            "metric":"(\\w+).SINK.(\\w+).StopTime",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SINK/$2/BatchEmptyCount":{
-            "metric":"(\\w+).SINK.(\\w+).BatchEmptyCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SOURCE/$2/AppendBatchReceivedCount":{
-            "metric":"(\\w+).SOURCE.(\\w+).AppendBatchReceivedCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SOURCE/$2/AppendAcceptedCount":{
-            "metric":"(\\w+).SOURCE.(\\w+).AppendAcceptedCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SOURCE/$2/StartTime":{
-            "metric":"(\\w+).SOURCE.(\\w+).StartTime",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SOURCE/$2/OpenConnectionCount":{
-            "metric":"(\\w+).SOURCE.(\\w+).OpenConnectionCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SOURCE/$2/AppendBatchAcceptedCount":{
-            "metric":"(\\w+).SOURCE.(\\w+).AppendBatchAcceptedCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SOURCE/$2/AppendReceivedCount":{
-            "metric":"(\\w+).SOURCE.(\\w+).AppendReceivedCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SOURCE/$2/EventReceivedCount":{
-            "metric":"(\\w+).SOURCE.(\\w+).EventReceivedCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SOURCE/$2/StopTime":{
-            "metric":"(\\w+).SOURCE.(\\w+).StopTime",
-            "pointInTime":true,
-            "temporal":true
-          },
-          "metrics/flume/$1/SOURCE/$2/EventAcceptedCount":{
-            "metric":"(\\w+).SOURCE.(\\w+).EventAcceptedCount",
-            "pointInTime":true,
-            "temporal":true
-          },
-
-          "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventTakeSuccessCount/rate/avg": {
-            "metric":"(\\w+).CHANNEL.(\\w+).EventTakeSuccessCount._rate._avg",
-            "pointInTime":false,
-            "temporal":true
-          },
-          "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventTakeSuccessCount/rate/max": {
-            "metric":"(\\w+).CHANNEL.(\\w+).EventTakeSuccessCount._rate._max",
-            "pointInTime":false,
-            "temporal":true
-          },
-          "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventTakeSuccessCount/rate/min": {
-            "metric":"(\\w+).CHANNEL.(\\w+).EventTakeSuccessCount._rate._min",
-            "pointInTime":false,
-            "temporal":true
-          },
-          "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventTakeSuccessCount/rate/sum": {
-            "metric":"(\\w+).CHANNEL.(\\w+).EventTakeSuccessCount._rate._sum",
-            "pointInTime":false,
-            "temporal":true
-          },
-
-          "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventPutSuccessCount/rate/avg": {
-            "metric":"(\\w+).CHANNEL.(\\w+).EventPutSuccessCount._rate._avg",
-            "pointInTime":false,
-            "temporal":true
-          },
-          "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventPutSuccessCount/rate/max": {
-            "metric":"(\\w+).CHANNEL.(\\w+).EventPutSuccessCount._rate._max",
-            "pointInTime":false,
-            "temporal":true
-          },
-          "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventPutSuccessCount/rate/min": {
-            "metric":"(\\w+).CHANNEL.(\\w+).EventPutSuccessCount._rate._min",
-            "pointInTime":false,
-            "temporal":true
-          },
-          "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventPutSuccessCount/rate/sum": {
-            "metric":"(\\w+).CHANNEL.(\\w+).EventPutSuccessCount._rate._sum",
-            "pointInTime":false,
-            "temporal":true
-          },
-          
-          "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")ChannelSize/avg": {
-            "metric":"(\\w+).CHANNEL.(\\w+).ChannelSize._avg",
-            "pointInTime":false,
-            "temporal":true
-          },
-          "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")ChannelSize/max": {
-            "metric":"(\\w+).CHANNEL.(\\w+).ChannelSize._max",
-            "pointInTime":false,
-            "temporal":true
-          },
-          "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")ChannelSize/min": {
-            "metric":"(\\w+).CHANNEL.(\\w+).ChannelSize._min",
-            "pointInTime":false,
-            "temporal":true
-          },
-          "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")ChannelSize/sum": {
-            "metric":"(\\w+).CHANNEL.(\\w+).ChannelSize._sum",
-            "pointInTime":false,
-            "temporal":true
-          }
-          
-        }
-      }
-    ]
-  }
-}

http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/files/alert_flume_agent_status.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/files/alert_flume_agent_status.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/files/alert_flume_agent_status.py
deleted file mode 100644
index b183bbc..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/files/alert_flume_agent_status.py
+++ /dev/null
@@ -1,99 +0,0 @@
-#!/usr/bin/env python
-
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-"""
-
-import socket
-
-from resource_management.libraries.functions.flume_agent_helper import find_expected_agent_names
-from resource_management.libraries.functions.flume_agent_helper import get_flume_status
-
-RESULT_CODE_OK = 'OK'
-RESULT_CODE_CRITICAL = 'CRITICAL'
-RESULT_CODE_UNKNOWN = 'UNKNOWN'
-
-FLUME_CONF_DIR_KEY = '{{flume-env/flume_conf_dir}}'
-
-FLUME_RUN_DIR = '/var/run/flume'
-
-def get_tokens():
-  """
-  Returns a tuple of tokens in the format {{site/property}} that will be used
-  to build the dictionary passed into execute
-  """
-  return (FLUME_CONF_DIR_KEY,)
-  
-
-def execute(parameters=None, host_name=None):
-  """
-  Returns a tuple containing the result code and a pre-formatted result label
-
-  Keyword arguments:
-  parameters (dictionary): a mapping of parameter key to value
-  host_name (string): the name of this host where the alert is running
-  """
-
-  if parameters is None:
-    return (RESULT_CODE_UNKNOWN, ['There were no parameters supplied to the script.'])
-
-  flume_conf_directory = None
-  if FLUME_CONF_DIR_KEY in parameters:
-    flume_conf_directory = parameters[FLUME_CONF_DIR_KEY]
-
-  if flume_conf_directory is None:
-    return (RESULT_CODE_UNKNOWN, ['The Flume configuration directory is a required parameter.'])
-
-  if host_name is None:
-    host_name = socket.getfqdn()
-
-  processes = get_flume_status(flume_conf_directory, FLUME_RUN_DIR)
-  expected_agents = find_expected_agent_names(flume_conf_directory)
-
-  alert_label = ''
-  alert_state = RESULT_CODE_OK
-
-  if len(processes) == 0 and len(expected_agents) == 0:
-    alert_label = 'No agents defined on {0}'.format(host_name)
-  else:
-    ok = []
-    critical = []
-    text_arr = []
-
-    for process in processes:
-      if not process.has_key('status') or process['status'] == 'NOT_RUNNING':
-        critical.append(process['name'])
-      else:
-        ok.append(process['name'])
-
-    if len(critical) > 0:
-      text_arr.append("{0} {1} NOT running".format(", ".join(critical),
-        "is" if len(critical) == 1 else "are"))
-
-    if len(ok) > 0:
-      text_arr.append("{0} {1} running".format(", ".join(ok),
-        "is" if len(ok) == 1 else "are"))
-
-    plural = len(critical) > 1 or len(ok) > 1
-    alert_label = "Agent{0} {1} {2}".format(
-      "s" if plural else "",
-      " and ".join(text_arr),
-      "on " + host_name)
-
-    alert_state = RESULT_CODE_CRITICAL if len(critical) > 0 else RESULT_CODE_OK
-
-  return (alert_state, [alert_label])
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/flume.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/flume.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/flume.py
deleted file mode 100644
index f4870ad..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/flume.py
+++ /dev/null
@@ -1,206 +0,0 @@
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-"""
-
-import glob
-import json
-import os
-from resource_management import *
-from resource_management.libraries.functions.flume_agent_helper import is_flume_process_live
-from resource_management.libraries.functions.flume_agent_helper import find_expected_agent_names
-
-def flume(action = None):
-  import params
-
-  if action == 'config':
-    # remove previously defined meta's
-    for n in find_expected_agent_names(params.flume_conf_dir):
-      os.unlink(os.path.join(params.flume_conf_dir, n, 'ambari-meta.json'))
-
-    Directory(params.flume_conf_dir, recursive=True)
-    Directory(params.flume_log_dir, owner=params.flume_user)
-
-    flume_agents = {}
-    if params.flume_conf_content is not None:
-      flume_agents = build_flume_topology(params.flume_conf_content)
-
-    for agent in flume_agents.keys():
-      flume_agent_conf_dir = os.path.join(params.flume_conf_dir, agent)
-      flume_agent_conf_file = os.path.join(flume_agent_conf_dir, 'flume.conf')
-      flume_agent_meta_file = os.path.join(flume_agent_conf_dir, 'ambari-meta.json')
-      flume_agent_log4j_file = os.path.join(flume_agent_conf_dir, 'log4j.properties')
-      flume_agent_env_file = os.path.join(flume_agent_conf_dir, 'flume-env.sh')
-
-      Directory(flume_agent_conf_dir)
-
-      PropertiesFile(flume_agent_conf_file,
-        properties=flume_agents[agent],
-        mode = 0644)
-
-      File(flume_agent_log4j_file,
-        content=Template('log4j.properties.j2', agent_name = agent),
-        mode = 0644)
-
-      File(flume_agent_meta_file,
-        content = json.dumps(ambari_meta(agent, flume_agents[agent])),
-        mode = 0644)
-
-      File(flume_agent_env_file,
-           owner=params.flume_user,
-           content=InlineTemplate(params.flume_env_sh_template)
-      )
-
-      if params.has_metric_collector:
-        File(os.path.join(flume_agent_conf_dir, "flume-metrics2.properties"),
-             owner=params.flume_user,
-             content=Template("flume-metrics2.properties.j2")
-        )
-
-  elif action == 'start':
-    # desired state for service should be STARTED
-    if len(params.flume_command_targets) == 0:
-      _set_desired_state('STARTED')
-
-    # It is important to run this command as a background process.
-    
-    
-    flume_base = as_user(format("{flume_bin} agent --name {{0}} --conf {{1}} --conf-file {{2}} {{3}}"), params.flume_user, env={'JAVA_HOME': params.java_home}) + " &"
-
-    for agent in cmd_target_names():
-      flume_agent_conf_dir = params.flume_conf_dir + os.sep + agent
-      flume_agent_conf_file = flume_agent_conf_dir + os.sep + "flume.conf"
-      flume_agent_pid_file = params.flume_run_dir + os.sep + agent + ".pid"
-
-      if not os.path.isfile(flume_agent_conf_file):
-        continue
-
-      if not is_flume_process_live(flume_agent_pid_file):
-        # TODO someday make the ganglia ports configurable
-        extra_args = ''
-        if params.ganglia_server_host is not None:
-          extra_args = '-Dflume.monitoring.type=ganglia -Dflume.monitoring.hosts={0}:{1}'
-          extra_args = extra_args.format(params.ganglia_server_host, '8655')
-        if params.has_metric_collector:
-          extra_args = '-Dflume.monitoring.type=org.apache.hadoop.metrics2.sink.flume.FlumeTimelineMetricsSink'
-
-        flume_cmd = flume_base.format(agent, flume_agent_conf_dir,
-           flume_agent_conf_file, extra_args)
-
-        Execute(flume_cmd, 
-          wait_for_finish=False,
-          environment={'JAVA_HOME': params.java_home}
-        )
-
-        # sometimes startup spawns a couple of threads - so only the first line may count
-        pid_cmd = format('pgrep -o -u {flume_user} -f ^{java_home}.*{agent}.* > {flume_agent_pid_file}')
-        Execute(pid_cmd, logoutput=True, tries=20, try_sleep=6)
-
-    pass
-  elif action == 'stop':
-    # desired state for service should be INSTALLED
-    if len(params.flume_command_targets) == 0:
-      _set_desired_state('INSTALLED')
-
-    pid_files = glob.glob(params.flume_run_dir + os.sep + "*.pid")
-
-    if 0 == len(pid_files):
-      return
-
-    agent_names = cmd_target_names()
-
-
-    for agent in agent_names:
-      pid_file = params.flume_run_dir + os.sep + agent + '.pid'
-      pid = format('`cat {pid_file}` > /dev/null 2>&1')
-      Execute(format('kill {pid}'), ignore_failures=True)
-      File(pid_file, action = 'delete')
-
-
-def ambari_meta(agent_name, agent_conf):
-  res = {}
-
-  sources = agent_conf[agent_name + '.sources'].split(' ')
-  res['sources_count'] = len(sources)
-
-  sinks = agent_conf[agent_name + '.sinks'].split(' ')
-  res['sinks_count'] = len(sinks)
-
-  channels = agent_conf[agent_name + '.channels'].split(' ')
-  res['channels_count'] = len(channels)
-
-  return res
-
-
-# define a map of dictionaries, where the key is agent name
-# and the dictionary is the name/value pair
-def build_flume_topology(content):
-
-  result = {}
-  agent_names = []
-
-  for line in content.split('\n'):
-    rline = line.strip()
-    if 0 != len(rline) and not rline.startswith('#'):
-      pair = rline.split('=')
-      lhs = pair[0].strip()
-      rhs = pair[1].strip()
-
-      part0 = lhs.split('.')[0]
-
-      if lhs.endswith(".sources"):
-        agent_names.append(part0)
-
-      if not result.has_key(part0):
-        result[part0] = {}
-
-      result[part0][lhs] = rhs
-
-  # trim out non-agents
-  for k in result.keys():
-    if not k in agent_names:
-      del result[k]
-
-  return result
-
-
-def cmd_target_names():
-  import params
-
-  if len(params.flume_command_targets) > 0:
-    return params.flume_command_targets
-  else:
-    return find_expected_agent_names(params.flume_conf_dir)
-
-
-def _set_desired_state(state):
-  import params
-  filename = os.path.join(params.flume_run_dir, 'ambari-state.txt')
-  File(filename,
-    content = state,
-  )
-
-
-def get_desired_state():
-  import params
-
-  try:
-    with open(os.path.join(params.flume_run_dir, 'ambari-state.txt'), 'r') as fp:
-      return fp.read()
-  except:
-    return 'INSTALLED'
-  

http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/flume_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/flume_check.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/flume_check.py
deleted file mode 100644
index b93b8e8..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/flume_check.py
+++ /dev/null
@@ -1,40 +0,0 @@
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-"""
-
-from resource_management import *
-
-
-class FlumeServiceCheck(Script):
-
-  def service_check(self, env):
-    import params
-
-    env.set_params(params)
-    if params.security_enabled:
-      principal_replaced = params.http_principal.replace("_HOST", params.hostname)
-      Execute(format("{kinit_path_local} -kt {http_keytab} {principal_replaced}"),
-              user=params.smoke_user)
-
-    Execute(format('env JAVA_HOME={java_home} {flume_bin} version'),
-            logoutput=True,
-            tries = 3,
-            try_sleep = 20)
-
-if __name__ == "__main__":
-  FlumeServiceCheck().execute()

http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/flume_handler.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/flume_handler.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/flume_handler.py
deleted file mode 100644
index a13f507..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/flume_handler.py
+++ /dev/null
@@ -1,79 +0,0 @@
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-"""
-
-from flume import flume
-from flume import get_desired_state
-
-from resource_management import *
-from resource_management.libraries.functions.flume_agent_helper import find_expected_agent_names
-from resource_management.libraries.functions.flume_agent_helper import get_flume_status
-
-class FlumeHandler(Script):
-  def install(self, env):
-    import params
-
-    self.install_packages(env)
-    env.set_params(params)
-
-  def start(self, env, rolling_restart=False):
-    import params
-
-    env.set_params(params)
-    self.configure(env)
-
-    flume(action='start')
-
-  def stop(self, env, rolling_restart=False):
-    import params
-
-    env.set_params(params)
-
-    flume(action='stop')
-
-  def configure(self, env):
-    import params
-
-    env.set_params(params)
-
-    flume(action='config')
-
-  def status(self, env):
-    import params
-    env.set_params(params)
-
-
-    processes = get_flume_status(params.flume_conf_dir, params.flume_run_dir)
-    expected_agents = find_expected_agent_names(params.flume_conf_dir)
-
-    json = {}
-    json['processes'] = processes
-    self.put_structured_out(json)
-
-    # only throw an exception if there are agents defined and there is a 
-    # problem with the processes; if there are no agents defined, then 
-    # the service should report STARTED (green) ONLY if the desired state is started.  otherwise, INSTALLED (red)
-    if len(expected_agents) > 0:
-      for proc in processes:
-        if not proc.has_key('status') or proc['status'] == 'NOT_RUNNING':
-          raise ComponentIsNotRunning()
-    elif len(expected_agents) == 0 and 'INSTALLED' == get_desired_state():
-      raise ComponentIsNotRunning()
-
-if __name__ == "__main__":
-  FlumeHandler().execute()

http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/params.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/params.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/params.py
deleted file mode 100644
index 2997867..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/params.py
+++ /dev/null
@@ -1,80 +0,0 @@
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-"""
-
-from resource_management.libraries.functions.version import format_hdp_stack_version, compare_versions
-from resource_management import *
-
-config = Script.get_config()
-
-user_group = config['configurations']['cluster-env']['user_group']
-proxyuser_group =  config['configurations']['hadoop-env']['proxyuser_group']
-
-security_enabled = False
-
-hdp_stack_version = str(config['hostLevelParams']['stack_version'])
-hdp_stack_version = format_hdp_stack_version(hdp_stack_version)
-
-#hadoop params
-if hdp_stack_version != "" and compare_versions(hdp_stack_version, '2.2') >= 0:
-  flume_bin = '/usr/hdp/current/flume-server/bin/flume-ng'
-  flume_hive_home = '/usr/hdp/current/hive-metastore'
-  flume_hcat_home = '/usr/hdp/current/hive-webhcat'
-else:
-  flume_bin = '/usr/bin/flume-ng'
-  flume_hive_home = '/usr/lib/hive'
-  flume_hcat_home = '/usr/lib/hive-hcatalog'
-
-flume_conf_dir = '/etc/flume/conf'
-java_home = config['hostLevelParams']['java_home']
-flume_log_dir = '/var/log/flume'
-flume_run_dir = '/var/run/flume'
-flume_user = 'flume'
-flume_group = 'flume'
-
-if 'flume-env' in config['configurations'] and 'flume_user' in config['configurations']['flume-env']:
-  flume_user = config['configurations']['flume-env']['flume_user']
-
-if (('flume-conf' in config['configurations']) and('content' in config['configurations']['flume-conf'])):
-  flume_conf_content = config['configurations']['flume-conf']['content']
-else:
-  flume_conf_content = None
-
-if (('flume-log4j' in config['configurations']) and ('content' in config['configurations']['flume-log4j'])):
-  flume_log4j_content = config['configurations']['flume-log4j']['content']
-else:
-  flume_log4j_content = None
-
-targets = default('/commandParams/flume_handler', None)
-flume_command_targets = [] if targets is None else targets.split(',')
-
-flume_env_sh_template = config['configurations']['flume-env']['content']
-
-ganglia_server_hosts = default('/clusterHostInfo/ganglia_server_host', [])
-ganglia_server_host = None
-if 0 != len(ganglia_server_hosts):
-  ganglia_server_host = ganglia_server_hosts[0]
-
-hostname = None
-if config.has_key('hostname'):
-  hostname = config['hostname']
-
-ams_collector_hosts = default("/clusterHostInfo/metric_collector_hosts", [])
-has_metric_collector = not len(ams_collector_hosts) == 0
-if has_metric_collector:
-  metric_collector_host = ams_collector_hosts[0]

http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/templates/flume-metrics2.properties.j2
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/templates/flume-metrics2.properties.j2 b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/templates/flume-metrics2.properties.j2
deleted file mode 100644
index 7458bf8..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/templates/flume-metrics2.properties.j2
+++ /dev/null
@@ -1,22 +0,0 @@
-{#
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements.  See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership.  The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License.  You may obtain a copy of the License at
-#
-#   http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-#}
-
-collector={{metric_collector_host}}:8188
-collectionFrequency=60000
-maxRowCacheSize=10000
-sendInterval=59000
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/templates/flume.conf.j2
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/templates/flume.conf.j2 b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/templates/flume.conf.j2
deleted file mode 100644
index 4dee67f..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/templates/flume.conf.j2
+++ /dev/null
@@ -1,24 +0,0 @@
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements.  See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership.  The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License.  You may obtain a copy of the License at
-#
-#  http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied.  See the License for the
-# specific language governing permissions and limitations
-# under the License.
-
-
-# flume.conf: Add your flume configuration here and start flume
-#             Note if you are using the Windows srvice or Unix service
-#             provided by the HDP distribution, they will assume the
-#             agent's name in this file to be 'a1'
-#
-{{flume_agent_conf_content}}

http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/templates/log4j.properties.j2
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/templates/log4j.properties.j2 b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/templates/log4j.properties.j2
deleted file mode 100644
index 3b34db8..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/templates/log4j.properties.j2
+++ /dev/null
@@ -1,67 +0,0 @@
-#
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements.  See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership.  The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License.  You may obtain a copy of the License at
-#
-#  http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied.  See the License for the
-# specific language governing permissions and limitations
-# under the License.
-#
-
-# Define some default values that can be overridden by system properties.
-#
-# For testing, it may also be convenient to specify
-# -Dflume.root.logger=DEBUG,console when launching flume.
-
-#flume.root.logger=DEBUG,console
-flume.root.logger=INFO,LOGFILE
-flume.log.dir={{flume_log_dir}}
-flume.log.file=flume-{{agent_name}}.log
-
-log4j.logger.org.apache.flume.lifecycle = INFO
-log4j.logger.org.jboss = WARN
-log4j.logger.org.mortbay = INFO
-log4j.logger.org.apache.avro.ipc.NettyTransceiver = WARN
-log4j.logger.org.apache.hadoop = INFO
-
-# Define the root logger to the system property "flume.root.logger".
-log4j.rootLogger=${flume.root.logger}
-
-
-# Stock log4j rolling file appender
-# Default log rotation configuration
-log4j.appender.LOGFILE=org.apache.log4j.RollingFileAppender
-log4j.appender.LOGFILE.MaxFileSize=100MB
-log4j.appender.LOGFILE.MaxBackupIndex=10
-log4j.appender.LOGFILE.File=${flume.log.dir}/${flume.log.file}
-log4j.appender.LOGFILE.layout=org.apache.log4j.PatternLayout
-log4j.appender.LOGFILE.layout.ConversionPattern=%d{dd MMM yyyy HH:mm:ss,SSS} %-5p [%t] (%C.%M:%L) %x - %m%n
-
-
-# Warning: If you enable the following appender it will fill up your disk if you don't have a cleanup job!
-# This uses the updated rolling file appender from log4j-extras that supports a reliable time-based rolling policy.
-# See http://logging.apache.org/log4j/companions/extras/apidocs/org/apache/log4j/rolling/TimeBasedRollingPolicy.html
-# Add "DAILY" to flume.root.logger above if you want to use this
-log4j.appender.DAILY=org.apache.log4j.rolling.RollingFileAppender
-log4j.appender.DAILY.rollingPolicy=org.apache.log4j.rolling.TimeBasedRollingPolicy
-log4j.appender.DAILY.rollingPolicy.ActiveFileName=${flume.log.dir}/${flume.log.file}
-log4j.appender.DAILY.rollingPolicy.FileNamePattern=${flume.log.dir}/${flume.log.file}.%d{yyyy-MM-dd}
-log4j.appender.DAILY.layout=org.apache.log4j.PatternLayout
-log4j.appender.DAILY.layout.ConversionPattern=%d{dd MMM yyyy HH:mm:ss,SSS} %-5p [%t] (%C.%M:%L) %x - %m%n
-
-
-# console
-# Add "console" to flume.root.logger above if you want to use this
-log4j.appender.console=org.apache.log4j.ConsoleAppender
-log4j.appender.console.target=System.err
-log4j.appender.console.layout=org.apache.log4j.PatternLayout
-log4j.appender.console.layout.ConversionPattern=%d (%t) [%p - %l] %m%n

http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/alerts.json
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/alerts.json b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/alerts.json
deleted file mode 100644
index 7605787..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/alerts.json
+++ /dev/null
@@ -1,137 +0,0 @@
-{
-  "GANGLIA": {
-    "service": [],
-    "GANGLIA_SERVER": [
-      {
-        "name": "ganglia_server_process",
-        "label": "Ganglia Server Process",
-        "description": "This host-level alert is triggered if the Ganglia server process cannot be established to be up and listening on the network.",
-        "interval": 1,
-        "scope": "ANY",
-        "enabled": true,
-        "source": {
-          "type": "PORT",
-          "uri": "8651",
-          "default_port": 8651,
-          "reporting": {
-            "ok": {
-              "text": "TCP OK - {0:.3f}s response on port {1}"
-            },
-            "warning": {
-              "text": "TCP OK - {0:.3f}s response on port {1}",
-              "value": 1.5
-            },
-            "critical": {
-              "text": "Connection failed: {0} to {1}:{2}",
-              "value": 5.0
-            }
-          }
-        }
-      },
-      {
-        "name": "ganglia_monitor_hdfs_namenode",
-        "label": "Ganglia NameNode Process Monitor",
-        "description": "This host-level alert is triggered if the Ganglia gmond process which handles receiving metrics for HDFS NameNode is not up and listening.",
-        "interval": 1,
-        "scope": "ANY",
-        "enabled": true,
-        "source": {
-          "type": "PORT",
-          "uri": "8661",
-          "default_port": 8661,
-          "reporting": {
-            "ok": {
-              "text": "TCP OK - {0:.3f}s response on port {1}"
-            },
-            "warning": {
-              "text": "TCP OK - {0:.3f}s response on port {1}",
-              "value": 1.5
-            },
-            "critical": {
-              "text": "Connection failed: {0} to {1}:{2}",
-              "value": 5.0
-            }
-          }
-        }
-      },
-      {
-        "name": "ganglia_monitor_hbase_master",
-        "label": "Ganglia HBase Master Process Monitor",
-        "description": "This host-level alert is triggered if the Ganglia gmond process which handles receiving metrics for the HBase Master process is not up and listening.",
-        "interval": 1,
-        "scope": "ANY",
-        "enabled": true,
-        "source": {
-          "type": "PORT",
-          "uri": "8663",
-          "default_port": 8663,
-          "reporting": {
-            "ok": {
-              "text": "TCP OK - {0:.3f}s response on port {1}"
-            },
-            "warning": {
-              "text": "TCP OK - {0:.3f}s response on port {1}",
-              "value": 1.5
-            },
-            "critical": {
-              "text": "Connection failed: {0} to {1}:{2}",
-              "value": 5.0
-            }
-          }
-        }
-      },
-      {
-        "name": "ganglia_monitor_yarn_resourcemanager",
-        "label": "Ganglia ResourceManager Process Monitor",
-        "description": "This host-level alert is triggered if the Ganglia gmond process which handles receiving metrics for the YARN ResourceManager process is not up and listening.",
-        "interval": 1,
-        "scope": "ANY",
-        "enabled": true,
-        "source": {
-          "type": "PORT",
-          "uri": "8664",
-          "default_port": 8664,
-          "reporting": {
-            "ok": {
-              "text": "TCP OK - {0:.3f}s response on port {1}"
-            },
-            "warning": {
-              "text": "TCP OK - {0:.3f}s response on port {1}",
-              "value": 1.5
-            },
-            "critical": {
-              "text": "Connection failed: {0} to {1}:{2}",
-              "value": 5.0
-            }
-          }
-        }
-      },
-      {
-        "name": "ganglia_monitor_mapreduce_history_server",
-        "label": "Ganglia History Server Process Monitor",
-        "description": "This host-level alert is triggered if the Ganglia gmond process which handles receiving metrics for the MapReduce History Server process is not up and listening.",
-        "interval": 1,
-        "scope": "ANY",
-        "enabled": true,
-        "source": {
-          "type": "PORT",
-          "uri": "8666",
-          "default_port": 8666,
-          "reporting": {
-            "ok": {
-              "text": "TCP OK - {0:.3f}s response on port {1}"
-            },
-            "warning": {
-              "text": "TCP OK - {0:.3f}s response on port {1}",
-              "value": 1.5
-            },
-            "critical": {
-              "text": "Connection failed: {0} to {1}:{2}",
-              "value": 5.0
-            }
-          }
-        }
-      }
-    ]
-  }
-}
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/configuration/ganglia-env.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/configuration/ganglia-env.xml b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/configuration/ganglia-env.xml
deleted file mode 100644
index 3328acf..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/configuration/ganglia-env.xml
+++ /dev/null
@@ -1,77 +0,0 @@
-<?xml version="1.0"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *     http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
--->
-
-<configuration>
-  <property>
-    <name>ganglia_conf_dir</name>
-    <value>/etc/ganglia/hdp</value>
-    <description>Config directory for Ganglia</description>
-  </property>
-  <property>
-    <name>ganglia_runtime_dir</name>
-    <value>/var/run/ganglia/hdp</value>
-    <description>Run directories for Ganglia</description>
-  </property>
-  <property>
-    <name>gmetad_user</name>
-    <value>nobody</value>
-    <property-type>USER GROUP</property-type>
-    <description>User </description>
-  </property>
-    <property>
-    <name>gmond_user</name>
-    <value>nobody</value>
-    <property-type>USER GROUP</property-type>
-    <description>User </description>
-  </property>
-  <property>
-    <name>rrdcached_base_dir</name>
-    <value>/var/lib/ganglia/rrds</value>
-    <description>Default directory for saving the rrd files on ganglia server</description>
-  </property>
-  <property>
-    <name>rrdcached_timeout</name>
-    <value>3600</value>
-    <description>(-w) Data is written to disk every timeout seconds. If this option is not specified the default interval of 300 seconds will be used.</description>
-  </property>
-  <property>
-    <name>rrdcached_flush_timeout</name>
-    <value>7200</value>
-      <description>(-f) Every timeout seconds the entire cache is searched for old values which are written to disk. This only concerns files to which updates have stopped, so setting this to a high value, such as 3600 seconds, is acceptable in most cases. This timeout defaults to 3600 seconds.</description>
-  </property>
-  <property>
-    <name>rrdcached_delay</name>
-    <value>1800</value>
-    <description>(-z) If specified, rrdcached will delay writing of each RRD for a random number of seconds in the range [0,delay). This will avoid too many writes being queued simultaneously. This value should be no greater than the value specified in -w. By default, there is no delay.</description>
-  </property>
-  <property>
-    <name>rrdcached_write_threads</name>
-    <value>4</value>
-    <description>(-t) Specifies the number of threads used for writing RRD files. The default is 4. Increasing this number will allow rrdcached to have more simultaneous I/O requests into the kernel. This may allow the kernel to re-order disk writes, resulting in better disk throughput.</description>
-  </property>
-  <property>
-    <name>additional_clusters</name>
-    <value> </value>
-    <description>Add additional desired Ganglia metrics cluster in the form "name1:port1,name2:port2". Ensure that the names and ports are unique across all cluster and ports are available on ganglia server host. Ambari has reserved ports 8667-8669 within its own pool.</description>
-  </property>
-
-</configuration>

http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/metainfo.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/metainfo.xml b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/metainfo.xml
index 4e96ade..33bc837 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/metainfo.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/metainfo.xml
@@ -20,108 +20,7 @@
   <services>
     <service>
       <name>GANGLIA</name>
-      <displayName>Ganglia</displayName>
-      <comment>Ganglia Metrics Collection system (&lt;a href=&quot;http://oss.oetiker.ch/rrdtool/&quot; target=&quot;_blank&quot;&gt;RRDTool&lt;/a&gt; will be installed too)</comment>
-      <version>3.5.0</version>
-      <components>
-        <component>
-          <name>GANGLIA_SERVER</name>
-          <displayName>Ganglia Server</displayName>
-          <category>MASTER</category>
-          <cardinality>1</cardinality>
-          <commandScript>
-            <script>scripts/ganglia_server.py</script>
-            <scriptType>PYTHON</scriptType>
-            <timeout>600</timeout>
-          </commandScript>
-        </component>
-
-        <component>
-          <name>GANGLIA_MONITOR</name>
-          <displayName>Ganglia Monitor</displayName>
-          <category>SLAVE</category>
-          <cardinality>ALL</cardinality>
-          <auto-deploy>
-            <enabled>true</enabled>
-          </auto-deploy>
-          <commandScript>
-            <script>scripts/ganglia_monitor.py</script>
-            <scriptType>PYTHON</scriptType>
-            <timeout>600</timeout>
-          </commandScript>
-        </component>
-      </components>
-      <osSpecifics>
-        <osSpecific>
-          <osFamily>redhat5,redhat6,suse11</osFamily>
-          <packages>
-            <package>
-              <name>python-rrdtool-1.4.5</name>
-            </package>
-            <package>
-              <name>libganglia-3.5.0-99</name>
-            </package>
-            <package>
-              <name>ganglia-devel-3.5.0-99</name>
-            </package>
-            <package>
-              <name>ganglia-gmetad-3.5.0-99</name>
-            </package>
-            <package>
-              <name>ganglia-web-3.5.7-99.noarch</name>
-            </package>
-            <package>
-              <name>ganglia-gmond-3.5.0-99</name>
-            </package>
-            <package>
-              <name>ganglia-gmond-modules-python-3.5.0-99</name>
-            </package>
-          </packages>
-        </osSpecific>
-        <osSpecific>
-          <osFamily>ubuntu12</osFamily>
-          <packages>
-            <package>
-              <name>python-rrdtool</name>
-            </package>
-            <package>
-              <name>gmetad</name>
-            </package>
-            <package>
-              <name>ganglia-webfrontend</name>
-            </package>
-            <package>
-              <name>ganglia-monitor-python</name>
-            </package>
-            <package>
-              <name>rrdcached</name>
-            </package>
-          </packages>
-        </osSpecific>
-        <osSpecific>
-          <osFamily>suse11</osFamily>
-          <packages>
-            <package>
-              <name>apache2</name>
-            </package>
-            <package>
-              <name>apache2?mod_php*</name>
-            </package>
-          </packages>
-        </osSpecific>
-        <osSpecific>
-          <osFamily>redhat5,redhat6</osFamily>
-          <packages>
-            <package>
-              <name>httpd</name>
-            </package>
-          </packages>
-        </osSpecific>
-      </osSpecifics>
-      <configuration-dependencies>
-        <config-type>ganglia-env</config-type>
-      </configuration-dependencies>
-      <monitoringService>true</monitoringService>
+      <extends>common-services/GANGLIA/3.5.0</extends>
     </service>
   </services>
 </metainfo>

http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/checkGmetad.sh
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/checkGmetad.sh b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/checkGmetad.sh
deleted file mode 100644
index f45e371..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/checkGmetad.sh
+++ /dev/null
@@ -1,37 +0,0 @@
-#!/usr/bin/env bash
-
-#/*
-# * Licensed to the Apache Software Foundation (ASF) under one
-# * or more contributor license agreements.  See the NOTICE file
-# * distributed with this work for additional information
-# * regarding copyright ownership.  The ASF licenses this file
-# * to you under the Apache License, Version 2.0 (the
-# * "License"); you may not use this file except in compliance
-# * with the License.  You may obtain a copy of the License at
-# *
-# *     http://www.apache.org/licenses/LICENSE-2.0
-# *
-# * Unless required by applicable law or agreed to in writing, software
-# * distributed under the License is distributed on an "AS IS" BASIS,
-# * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# * See the License for the specific language governing permissions and
-# * limitations under the License.
-# */
-
-cd `dirname ${0}`;
-
-# Get all our common constants etc. set up.
-source ./gmetadLib.sh;
-
-# Before checking gmetad, check rrdcached.
-./checkRrdcached.sh;
-
-gmetadRunningPid=`getGmetadRunningPid`;
-
-if [ -n "${gmetadRunningPid}" ]
-then
-  echo "${GMETAD_BIN} running with PID ${gmetadRunningPid}";
-else
-  echo "Failed to find running ${GMETAD_BIN}";
-  exit 1;
-fi

http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/checkGmond.sh
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/checkGmond.sh b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/checkGmond.sh
deleted file mode 100644
index fbf524a..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/checkGmond.sh
+++ /dev/null
@@ -1,62 +0,0 @@
-#!/usr/bin/env bash
-
-#/*
-# * Licensed to the Apache Software Foundation (ASF) under one
-# * or more contributor license agreements.  See the NOTICE file
-# * distributed with this work for additional information
-# * regarding copyright ownership.  The ASF licenses this file
-# * to you under the Apache License, Version 2.0 (the
-# * "License"); you may not use this file except in compliance
-# * with the License.  You may obtain a copy of the License at
-# *
-# *     http://www.apache.org/licenses/LICENSE-2.0
-# *
-# * Unless required by applicable law or agreed to in writing, software
-# * distributed under the License is distributed on an "AS IS" BASIS,
-# * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# * See the License for the specific language governing permissions and
-# * limitations under the License.
-# */
-
-cd `dirname ${0}`;
-
-# Get all our common constants etc. set up.
-# Pulls in gangliaLib.sh as well, so we can skip pulling it in again.
-source ./gmondLib.sh;
-
-function checkGmondForCluster()
-{
-    gmondClusterName=${1};
-
-    gmondCoreConfFileName=`getGmondCoreConfFileName ${gmondClusterName}`;
-
-    # Skip over (purported) Clusters that don't have their core conf file present.
-    if [ -e "${gmondCoreConfFileName}" ]
-    then 
-      gmondRunningPid=`getGmondRunningPid ${gmondClusterName}`;
-
-      if [ -n "${gmondRunningPid}" ]
-      then
-        echo "${GMOND_BIN} for cluster ${gmondClusterName} running with PID ${gmondRunningPid}";
-      else
-        echo "Failed to find running ${GMOND_BIN} for cluster ${gmondClusterName}";
-        exit 1;
-      fi
-    fi
-}
-
-# main()
-gmondClusterName=${1};
-
-if [ "x" == "x${gmondClusterName}" ]
-then
-    # No ${gmondClusterName} passed in as command-line arg, so check
-    # all the gmonds we know about.
-    for gmondClusterName in `getConfiguredGangliaClusterNames`
-    do
-        checkGmondForCluster ${gmondClusterName};
-    done
-else
-    # Just check the one ${gmondClusterName} that was asked for.
-    checkGmondForCluster ${gmondClusterName};
-fi

http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/checkRrdcached.sh
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/checkRrdcached.sh b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/checkRrdcached.sh
deleted file mode 100644
index 1e0c2e2..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/checkRrdcached.sh
+++ /dev/null
@@ -1,34 +0,0 @@
-#!/usr/bin/env bash
-
-#/*
-# * Licensed to the Apache Software Foundation (ASF) under one
-# * or more contributor license agreements.  See the NOTICE file
-# * distributed with this work for additional information
-# * regarding copyright ownership.  The ASF licenses this file
-# * to you under the Apache License, Version 2.0 (the
-# * "License"); you may not use this file except in compliance
-# * with the License.  You may obtain a copy of the License at
-# *
-# *     http://www.apache.org/licenses/LICENSE-2.0
-# *
-# * Unless required by applicable law or agreed to in writing, software
-# * distributed under the License is distributed on an "AS IS" BASIS,
-# * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# * See the License for the specific language governing permissions and
-# * limitations under the License.
-# */
-
-cd `dirname ${0}`;
-
-# Get all our common constants etc. set up.
-source ./rrdcachedLib.sh;
-
-rrdcachedRunningPid=`getRrdcachedRunningPid`;
-
-if [ -n "${rrdcachedRunningPid}" ]
-then
-  echo "${RRDCACHED_BIN} running with PID ${rrdcachedRunningPid}";
-else
-  echo "Failed to find running ${RRDCACHED_BIN}";
-  exit 1;
-fi

http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/gmetad.init
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/gmetad.init b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/gmetad.init
deleted file mode 100644
index 20b388e..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/gmetad.init
+++ /dev/null
@@ -1,73 +0,0 @@
-#!/bin/sh
-# chkconfig: 2345 70 40
-# description: hdp-gmetad startup script
-# processname: hdp-gmetad
-#/*
-# * Licensed to the Apache Software Foundation (ASF) under one
-# * or more contributor license agreements.  See the NOTICE file
-# * distributed with this work for additional information
-# * regarding copyright ownership.  The ASF licenses this file
-# * to you under the Apache License, Version 2.0 (the
-# * "License"); you may not use this file except in compliance
-# * with the License.  You may obtain a copy of the License at
-# *
-# *     http://www.apache.org/licenses/LICENSE-2.0
-# *
-# * Unless required by applicable law or agreed to in writing, software
-# * distributed under the License is distributed on an "AS IS" BASIS,
-# * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# * See the License for the specific language governing permissions and
-# * limitations under the License.
-# */
-
-# Remember to keep this in-sync with the definition of 
-# GANGLIA_RUNTIME_COMPONENTS_UNPACK_DIR in monrpmInstaller.sh.
-HDP_GANGLIA_RUNTIME_COMPONENTS_DIR=/usr/libexec/hdp/ganglia
-HDP_GANLIA_GMETAD_STARTER=${HDP_GANGLIA_RUNTIME_COMPONENTS_DIR}/startGmetad.sh
-HDP_GANLIA_GMETAD_STOPPER=${HDP_GANGLIA_RUNTIME_COMPONENTS_DIR}/stopGmetad.sh
-HDP_GANLIA_GMETAD_CHECKER=${HDP_GANGLIA_RUNTIME_COMPONENTS_DIR}/checkGmetad.sh
-
-RETVAL=0
-
-case "$1" in
-   start)
-      echo "============================="
-      echo "Starting hdp-gmetad..."
-      echo "============================="
-      [ -f ${HDP_GANLIA_GMETAD_STARTER} ] || exit 1
-      eval "${HDP_GANLIA_GMETAD_STARTER}"
-      RETVAL=$?
-      echo
-      [ $RETVAL -eq 0 ] && touch /var/lock/subsys/hdp-gmetad
-      ;;
-
-  stop)
-      echo "=================================="
-      echo "Shutting down hdp-gmetad..."
-      echo "=================================="
-      [ -f ${HDP_GANLIA_GMETAD_STOPPER} ] || exit 1
-      eval "${HDP_GANLIA_GMETAD_STOPPER}"
-      RETVAL=$?
-      echo
-      [ $RETVAL -eq 0 ] && rm -f /var/lock/subsys/hdp-gmetad
-      ;;
-
-  restart|reload)
-   	$0 stop
-   	$0 start
-   	RETVAL=$?
-	;;
-  status)
-      echo "======================================="
-      echo "Checking status of hdp-gmetad..."
-      echo "======================================="
-      [ -f ${HDP_GANLIA_GMETAD_CHECKER} ] || exit 1
-      eval "${HDP_GANLIA_GMETAD_CHECKER}"
-      RETVAL=$?
-      ;;
-  *)
-	echo "Usage: $0 {start|stop|restart|status}"
-	exit 1
-esac
-
-exit $RETVAL

http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/gmetadLib.sh
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/gmetadLib.sh b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/gmetadLib.sh
deleted file mode 100644
index 6a24bed..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/gmetadLib.sh
+++ /dev/null
@@ -1,204 +0,0 @@
-#!/usr/bin/env bash
-
-#/*
-# * Licensed to the Apache Software Foundation (ASF) under one
-# * or more contributor license agreements.  See the NOTICE file
-# * distributed with this work for additional information
-# * regarding copyright ownership.  The ASF licenses this file
-# * to you under the Apache License, Version 2.0 (the
-# * "License"); you may not use this file except in compliance
-# * with the License.  You may obtain a copy of the License at
-# *
-# *     http://www.apache.org/licenses/LICENSE-2.0
-# *
-# * Unless required by applicable law or agreed to in writing, software
-# * distributed under the License is distributed on an "AS IS" BASIS,
-# * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# * See the License for the specific language governing permissions and
-# * limitations under the License.
-# */
-
-cd `dirname ${0}`;
-
-# Slurp in all our user-customizable settings.
-source ./gangliaEnv.sh;
-
-# Get access to Ganglia-wide constants etc.
-source ./gangliaLib.sh;
-
-GMETAD_BIN=/usr/sbin/gmetad;
-GMETAD_CONF_FILE=${GANGLIA_CONF_DIR}/gmetad.conf;
-GMETAD_PID_FILE=${GANGLIA_RUNTIME_DIR}/gmetad.pid;
-
-function getGmetadLoggedPid()
-{
-    if [ -e "${GMETAD_PID_FILE}" ]
-    then
-        echo `cat ${GMETAD_PID_FILE}`;
-    fi
-}
-
-function getGmetadRunningPid()
-{
-    gmetadLoggedPid=`getGmetadLoggedPid`;
-
-    if [ -n "${gmetadLoggedPid}" ]
-    then
-        echo `ps -o pid=MYPID -p ${gmetadLoggedPid} | tail -1 | awk '{print $1}' | grep -v MYPID`;
-    fi
-}
-
-function generateGmetadConf()
-{
-    now=`date`;
-
-    cat <<END_OF_GMETAD_CONF_1
-#################### Generated by ${0} on ${now} ####################
-#
-#-------------------------------------------------------------------------------
-# Setting the debug_level to 1 will keep daemon in the forground and
-# show only error messages. Setting this value higher than 1 will make 
-# gmetad output debugging information and stay in the foreground.
-# default: 0
-# debug_level 10
-#
-#-------------------------------------------------------------------------------
-# What to monitor. The most important section of this file. 
-#
-# The data_source tag specifies either a cluster or a grid to
-# monitor. If we detect the source is a cluster, we will maintain a complete
-# set of RRD databases for it, which can be used to create historical 
-# graphs of the metrics. If the source is a grid (it comes from another gmetad),
-# we will only maintain summary RRDs for it.
-#
-# Format: 
-# data_source "my cluster" [polling interval] address1:port addreses2:port ...
-# 
-# The keyword 'data_source' must immediately be followed by a unique
-# string which identifies the source, then an optional polling interval in 
-# seconds. The source will be polled at this interval on average. 
-# If the polling interval is omitted, 15sec is asssumed. 
-#
-# If you choose to set the polling interval to something other than the default,
-# note that the web frontend determines a host as down if its TN value is less
-# than 4 * TMAX (20sec by default).  Therefore, if you set the polling interval
-# to something around or greater than 80sec, this will cause the frontend to
-# incorrectly display hosts as down even though they are not.
-#
-# A list of machines which service the data source follows, in the 
-# format ip:port, or name:port. If a port is not specified then 8649
-# (the default gmond port) is assumed.
-# default: There is no default value
-#
-# data_source "my cluster" 10 localhost  my.machine.edu:8649  1.2.3.5:8655
-# data_source "my grid" 50 1.3.4.7:8655 grid.org:8651 grid-backup.org:8651
-# data_source "another source" 1.3.4.7:8655  1.3.4.8
-END_OF_GMETAD_CONF_1
-
-    # Get info about all the configured Ganglia clusters.
-    getGangliaClusterInfo | while read gangliaClusterInfoLine
-    do
-        # From each, parse out ${gmondClusterName}, ${gmondMasterIP} and ${gmondPort}... 
-        read gmondClusterName gmondMasterIP gmondPort <<<`echo ${gangliaClusterInfoLine}`;
-        # ...and generate a corresponding data_source line for gmetad.conf. 
-        echo "data_source \"${gmondClusterName}\" ${gmondMasterIP}:${gmondPort}";
-    done
-
-    cat <<END_OF_GMETAD_CONF_2
-#
-# Round-Robin Archives
-# You can specify custom Round-Robin archives here (defaults are listed below)
-#
-# Old Default RRA: Keep 1 hour of metrics at 15 second resolution. 1 day at 6 minute
-# RRAs "RRA:AVERAGE:0.5:1:244" "RRA:AVERAGE:0.5:24:244" "RRA:AVERAGE:0.5:168:244" "RRA:AVERAGE:0.5:672:244" \
-#      "RRA:AVERAGE:0.5:5760:374"
-# New Default RRA
-# Keep 5856 data points at 15 second resolution assuming 15 second (default) polling. That's 1 day
-# Two weeks of data points at 1 minute resolution (average)
-#RRAs "RRA:AVERAGE:0.5:1:5856" "RRA:AVERAGE:0.5:4:20160" "RRA:AVERAGE:0.5:40:52704"
-# Retaining existing resolution
-RRAs "RRA:AVERAGE:0.5:1:244" "RRA:AVERAGE:0.5:24:244" "RRA:AVERAGE:0.5:168:244" "RRA:AVERAGE:0.5:672:244" \
-     "RRA:AVERAGE:0.5:5760:374"
-#
-#-------------------------------------------------------------------------------
-# Scalability mode. If on, we summarize over downstream grids, and respect
-# authority tags. If off, we take on 2.5.0-era behavior: we do not wrap our output
-# in <GRID></GRID> tags, we ignore all <GRID> tags we see, and always assume
-# we are the "authority" on data source feeds. This approach does not scale to
-# large groups of clusters, but is provided for backwards compatibility.
-# default: on
-# scalable off
-#
-#-------------------------------------------------------------------------------
-# The name of this Grid. All the data sources above will be wrapped in a GRID
-# tag with this name.
-# default: unspecified
-gridname "HDP_GRID"
-#
-#-------------------------------------------------------------------------------
-# The authority URL for this grid. Used by other gmetads to locate graphs
-# for our data sources. Generally points to a ganglia/
-# website on this machine.
-# default: "http://hostname/ganglia/",
-#   where hostname is the name of this machine, as defined by gethostname().
-# authority "http://mycluster.org/newprefix/"
-#
-#-------------------------------------------------------------------------------
-# List of machines this gmetad will share XML with. Localhost
-# is always trusted. 
-# default: There is no default value
-# trusted_hosts 127.0.0.1 169.229.50.165 my.gmetad.org
-#
-#-------------------------------------------------------------------------------
-# If you want any host which connects to the gmetad XML to receive
-# data, then set this value to "on"
-# default: off
-# all_trusted on
-#
-#-------------------------------------------------------------------------------
-# If you don't want gmetad to setuid then set this to off
-# default: on
-# setuid off
-#
-#-------------------------------------------------------------------------------
-# User gmetad will setuid to (defaults to "nobody")
-# default: "nobody"
-setuid_username "${GMETAD_USER}"
-#
-#-------------------------------------------------------------------------------
-# Umask to apply to created rrd files and grid directory structure
-# default: 0 (files are public)
-# umask 022
-#
-#-------------------------------------------------------------------------------
-# The port gmetad will answer requests for XML
-# default: 8651
-# xml_port 8651
-#
-#-------------------------------------------------------------------------------
-# The port gmetad will answer queries for XML. This facility allows
-# simple subtree and summation views of the XML tree.
-# default: 8652
-# interactive_port 8652
-#
-#-------------------------------------------------------------------------------
-# The number of threads answering XML requests
-# default: 4
-# server_threads 10
-#
-#-------------------------------------------------------------------------------
-# Where gmetad stores its round-robin databases
-# default: "/var/lib/ganglia/rrds"
-rrd_rootdir "${RRD_ROOTDIR}"
-#
-#-------------------------------------------------------------------------------
-# In earlier versions of gmetad, hostnames were handled in a case
-# sensitive manner
-# If your hostname directories have been renamed to lower case,
-# set this option to 0 to disable backward compatibility.
-# From version 3.2, backwards compatibility will be disabled by default.
-# default: 1   (for gmetad < 3.2)
-# default: 0   (for gmetad >= 3.2)
-case_sensitive_hostnames 1
-END_OF_GMETAD_CONF_2
-}


Mime
View raw message