Return-Path: X-Original-To: apmail-ambari-commits-archive@www.apache.org Delivered-To: apmail-ambari-commits-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id E35FD18C29 for ; Mon, 17 Aug 2015 05:14:38 +0000 (UTC) Received: (qmail 75262 invoked by uid 500); 17 Aug 2015 05:14:38 -0000 Delivered-To: apmail-ambari-commits-archive@ambari.apache.org Received: (qmail 75178 invoked by uid 500); 17 Aug 2015 05:14:38 -0000 Mailing-List: contact commits-help@ambari.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: ambari-dev@ambari.apache.org Delivered-To: mailing list commits@ambari.apache.org Received: (qmail 74866 invoked by uid 99); 17 Aug 2015 05:14:38 -0000 Received: from git1-us-west.apache.org (HELO git1-us-west.apache.org) (140.211.11.23) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 17 Aug 2015 05:14:38 +0000 Received: by git1-us-west.apache.org (ASF Mail Server at git1-us-west.apache.org, from userid 33) id 189EADFC89; Mon, 17 Aug 2015 05:14:38 +0000 (UTC) Content-Type: text/plain; charset="us-ascii" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit From: jluniya@apache.org To: commits@ambari.apache.org Date: Mon, 17 Aug 2015 05:14:43 -0000 Message-Id: In-Reply-To: <64d97738be4a4a108c77f8e2c4b3ecfc@git.apache.org> References: <64d97738be4a4a108c77f8e2c4b3ecfc@git.apache.org> X-Mailer: ASF-Git Admin Mailer Subject: [07/23] ambari git commit: AMBARI-12779: [PluggableStackDefinition] Remove ambari-server/src/main/resources/stacks/PHD (jluniya) http://git-wip-us.apache.org/repos/asf/ambari/blob/6fcd95fe/ambari-server/src/main/resources/stacks/PHD/3.0.0.0/services/NAGIOS/package/templates/hadoop-commands.cfg.j2 ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/PHD/3.0.0.0/services/NAGIOS/package/templates/hadoop-commands.cfg.j2 b/ambari-server/src/main/resources/stacks/PHD/3.0.0.0/services/NAGIOS/package/templates/hadoop-commands.cfg.j2 deleted file mode 100644 index c1a792c..0000000 --- a/ambari-server/src/main/resources/stacks/PHD/3.0.0.0/services/NAGIOS/package/templates/hadoop-commands.cfg.j2 +++ /dev/null @@ -1,166 +0,0 @@ -{# -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -#} - -# -# -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, -# software distributed under the License is distributed on an -# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY -# KIND, either express or implied. See the License for the -# specific language governing permissions and limitations -# under the License. -# -# -# - -{% if check_cpu_on %} -# 'check_cpu' check remote cpu load -define command { - command_name check_cpu - command_line /var/lib/ambari-agent/ambari-python-wrap $USER1$/mm_wrapper.py legacy_check_wrapper -- php $USER1$/check_cpu.php -h $HOSTADDRESS$ -p $ARG1$ -w $ARG2$ -c $ARG3$ -e $ARG4$ -k $ARG5$ -r $ARG6$ -t $ARG7$ -u $ARG8$ - } -define command { - command_name check_cpu_ha - command_line /var/lib/ambari-agent/ambari-python-wrap $USER1$/mm_wrapper.py or $ARG1$ -- php $USER1$/check_cpu_ha.php -h ^^ -p $ARG2$ -w $ARG3$ -c $ARG4$ -e $ARG5$ -k $ARG6$ -r $ARG7$ -t $ARG8$ -u $ARG9$ - } -{% endif %} - -# Check data node storage full -define command { - command_name check_datanode_storage - command_line /var/lib/ambari-agent/ambari-python-wrap $USER1$/mm_wrapper.py legacy_check_wrapper -- php $USER1$/check_datanode_storage.php -h $HOSTADDRESS$ -p $ARG1$ -w $ARG2$ -c $ARG3$ -e $ARG4$ -k $ARG5$ -r $ARG6$ -t $ARG7$ -s $ARG8$ - } - -define command{ - command_name check_hdfs_blocks - command_line /var/lib/ambari-agent/ambari-python-wrap $USER1$/mm_wrapper.py and $ARG1$ -- php $USER1$/check_hdfs_blocks.php -h ^^ -p $ARG2$ -s $ARG3$ -e $ARG4$ -k $ARG5$ -r $ARG6$ -t $ARG7$ -u $ARG8$ - } - -define command{ - command_name check_hdfs_capacity - command_line /var/lib/ambari-agent/ambari-python-wrap $USER1$/mm_wrapper.py and $ARG1$ -- php $USER1$/check_hdfs_capacity.php -h ^^ -p $ARG2$ -w $ARG3$ -c $ARG4$ -e $ARG5$ -k $ARG6$ -r $ARG7$ -t $ARG8$ -s $ARG9$ - } - -define command{ - command_name check_aggregate - command_line php $USER1$/check_aggregate.php -f /var/nagios/status.dat -s 1 -t service -n $ARG1$ -w $ARG2$ -c $ARG3$ - } - -define command{ - command_name check_rpcq_latency - command_line /var/lib/ambari-agent/ambari-python-wrap $USER1$/mm_wrapper.py legacy_check_wrapper -- php $USER1$/check_rpcq_latency.php -h $HOSTADDRESS$ -p $ARG2$ -n $ARG1$ -w $ARG3$ -c $ARG4$ -e $ARG5$ -k $ARG6$ -r $ARG7$ -t $ARG8$ -s $ARG9$ - } - -define command{ - command_name check_rpcq_latency_ha - command_line /var/lib/ambari-agent/ambari-python-wrap $USER1$/mm_wrapper.py or $ARG1$ -- php $USER1$/check_rpcq_latency_ha.php -h ^^ -p $ARG3$ -n $ARG2$ -w $ARG4$ -c $ARG5$ -e $ARG6$ -k $ARG7$ -r $ARG8$ -t $ARG9$ -s $ARG10$ - } - -define command{ - command_name check_nagios - command_line /var/lib/ambari-agent/ambari-python-wrap $USER1$/mm_wrapper.py legacy_check_wrapper -- $USER1$/check_nagios -e $ARG1$ -F $ARG2$ -C $ARG3$ - } - -define command{ - command_name check_webui - command_line /var/lib/ambari-agent/ambari-python-wrap $USER1$/mm_wrapper.py legacy_check_wrapper -- $USER1$/check_webui.sh $ARG1$ $HOSTADDRESS$ $ARG2$ - } - -define command{ - command_name check_webui_ha - command_line /var/lib/ambari-agent/ambari-python-wrap $USER1$/mm_wrapper.py and $ARG2$ -- $USER1$/check_webui_ha.sh $ARG1$ ^^ $ARG3$ - } - -define command{ - command_name check_name_dir_status - command_line /var/lib/ambari-agent/ambari-python-wrap $USER1$/mm_wrapper.py legacy_check_wrapper -- php $USER1$/check_name_dir_status.php -h $HOSTADDRESS$ -p $ARG1$ -e $ARG2$ -k $ARG3$ -r $ARG4$ -t $ARG5$ -s $ARG6$ - } - -define command{ - command_name check_oozie_status - command_line /var/lib/ambari-agent/ambari-python-wrap $USER1$/mm_wrapper.py legacy_check_wrapper -- $USER1$/check_oozie_status.sh $HOSTADDRESS$ $ARG1$ $ARG2$ $ARG3$ $ARG4$ $ARG5$ $ARG6$ $ARG7$ - } - -define command{ - command_name check_templeton_status - command_line /var/lib/ambari-agent/ambari-python-wrap $USER1$/mm_wrapper.py legacy_check_wrapper -- $USER1$/check_templeton_status.sh $HOSTADDRESS$ $ARG1$ $ARG2$ $ARG3$ $ARG4$ $ARG5$ $ARG6$ $ARG7$ - } - -define command{ - command_name check_hive_metastore_status - command_line /var/lib/ambari-agent/ambari-python-wrap $USER1$/mm_wrapper.py legacy_check_wrapper -- $USER1$/check_hive_metastore_status.sh $HOSTADDRESS$ $ARG1$ $ARG2$ $ARG3$ $ARG4$ $ARG5$ $ARG6$ $ARG7$ - } -define command{ - command_name check_hue_status - command_line /var/lib/ambari-agent/ambari-python-wrap $USER1$/mm_wrapper.py legacy_check_wrapper -- $USER1$/check_hue_status.sh - } - -define command{ - command_name check_mapred_local_dir_used_space - command_line /var/lib/ambari-agent/ambari-python-wrap $USER1$/mm_wrapper.py legacy_check_wrapper -- $USER1$/check_mapred_local_dir_used.sh $ARG1$ $ARG2$ - } - -define command{ - command_name check_namenodes_ha - command_line /var/lib/ambari-agent/ambari-python-wrap $USER1$/mm_wrapper.py legacy_check_wrapper -- $USER1$/check_namenodes_ha.sh $ARG1$ $ARG2$ - } - -define command{ - command_name check_nodemanager_health - command_line /var/lib/ambari-agent/ambari-python-wrap $USER1$/mm_wrapper.py legacy_check_wrapper -- $USER1$/check_nodemanager_health.sh $HOSTADDRESS$ $ARG1$ - } - -define command{ - command_name host_sys_logger - command_line $USER1$/sys_logger.py $HOSTSTATETYPE$ $HOSTATTEMPT$ $HOSTSTATE$ "Host::Ping" "Event Host=$HOSTADDRESS$($HOSTSTATE$), $HOSTOUTPUT$ $LONGHOSTOUTPUT$" - } - -define command{ - command_name service_sys_logger - command_line $USER1$/sys_logger.py $SERVICESTATETYPE$ $SERVICEATTEMPT$ $SERVICESTATE$ "$SERVICEDESC$" "Event Host=$HOSTADDRESS$ Service Description=$SERVICEDESC$($SERVICESTATE$), $SERVICEOUTPUT$ $LONGSERVICEOUTPUT$" - } - -define command{ - command_name check_tcp_wrapper - command_line /var/lib/ambari-agent/ambari-python-wrap $USER1$/mm_wrapper.py and $HOSTADDRESS$ -- $USER1$/check_tcp -H ^^ -p $ARG1$ $ARG2$ - } - -define command{ - command_name check_checkpoint_time - command_line /var/lib/ambari-agent/ambari-python-wrap $USER1$/mm_wrapper.py or $ARG1$ -- /var/lib/ambari-agent/ambari-python-wrap $USER1$/check_checkpoint_time.py -H ^^ -p $ARG2$ -w $ARG3$ -c $ARG4$ -t $ARG5$ -x $ARG6$ -s $ARG7$ - } - -define command{ - command_name check_tcp_wrapper_sasl - command_line /var/lib/ambari-agent/ambari-python-wrap $USER1$/mm_wrapper.py and $HOSTADDRESS$ -- /var/lib/ambari-agent/ambari-python-wrap $USER1$/check_hive_thrift_port.py -H ^^ -p $ARG1$ $ARG2$ - } - -define command{ - command_name check_ambari - command_line /var/lib/ambari-agent/ambari-python-wrap $USER1$/mm_wrapper.py legacy_check_wrapper -- /var/lib/ambari-agent/ambari-python-wrap $USER1$/check_ambari_alerts.py -H $HOSTADDRESS$ -f $ARG1$ -n $ARG2$ - } http://git-wip-us.apache.org/repos/asf/ambari/blob/6fcd95fe/ambari-server/src/main/resources/stacks/PHD/3.0.0.0/services/NAGIOS/package/templates/hadoop-hostgroups.cfg.j2 ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/PHD/3.0.0.0/services/NAGIOS/package/templates/hadoop-hostgroups.cfg.j2 b/ambari-server/src/main/resources/stacks/PHD/3.0.0.0/services/NAGIOS/package/templates/hadoop-hostgroups.cfg.j2 deleted file mode 100644 index 05c1252..0000000 --- a/ambari-server/src/main/resources/stacks/PHD/3.0.0.0/services/NAGIOS/package/templates/hadoop-hostgroups.cfg.j2 +++ /dev/null @@ -1,33 +0,0 @@ -{# -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -#} - -{% for name, hosts in hostgroup_defs.iteritems() %} -{% if hosts %} -define hostgroup { - hostgroup_name {{name}} - alias {{name}} - members {{','.join(hosts)}} -} -{% endif %} -{% endfor %} - -define hostgroup { - hostgroup_name all-servers - alias All Servers - members {{','.join(all_hosts)}} -} \ No newline at end of file http://git-wip-us.apache.org/repos/asf/ambari/blob/6fcd95fe/ambari-server/src/main/resources/stacks/PHD/3.0.0.0/services/NAGIOS/package/templates/hadoop-hosts.cfg.j2 ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/PHD/3.0.0.0/services/NAGIOS/package/templates/hadoop-hosts.cfg.j2 b/ambari-server/src/main/resources/stacks/PHD/3.0.0.0/services/NAGIOS/package/templates/hadoop-hosts.cfg.j2 deleted file mode 100644 index 8bcc980..0000000 --- a/ambari-server/src/main/resources/stacks/PHD/3.0.0.0/services/NAGIOS/package/templates/hadoop-hosts.cfg.j2 +++ /dev/null @@ -1,53 +0,0 @@ -{# -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -#} - -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, -# software distributed under the License is distributed on an -# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY -# KIND, either express or implied. See the License for the -# specific language governing permissions and limitations -# under the License. -# -# -{% for host in all_hosts %} -define host { - alias {{host}} - host_name {{host}} - use {{host_template}} - address {{host}} - check_command check_tcp_wrapper!{{all_ping_ports[loop.index-1]}}!-w 1 -c 1 - check_interval 0.25 - retry_interval 0.25 - max_check_attempts 4 - notifications_enabled 1 - first_notification_delay 0 # Send notification soon after change in the hard state - notification_interval 0 # Send the notification once - notification_options d,u,r -} - -{% endfor %} http://git-wip-us.apache.org/repos/asf/ambari/blob/6fcd95fe/ambari-server/src/main/resources/stacks/PHD/3.0.0.0/services/NAGIOS/package/templates/hadoop-servicegroups.cfg.j2 ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/PHD/3.0.0.0/services/NAGIOS/package/templates/hadoop-servicegroups.cfg.j2 b/ambari-server/src/main/resources/stacks/PHD/3.0.0.0/services/NAGIOS/package/templates/hadoop-servicegroups.cfg.j2 deleted file mode 100644 index aee9b15..0000000 --- a/ambari-server/src/main/resources/stacks/PHD/3.0.0.0/services/NAGIOS/package/templates/hadoop-servicegroups.cfg.j2 +++ /dev/null @@ -1,113 +0,0 @@ -{# -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -#} - - - -{% if hostgroup_defs['namenode'] or - hostgroup_defs['snamenode'] or - hostgroup_defs['slaves'] %} - {% if hostgroup_defs['namenode'] != None %} - define servicegroup { - servicegroup_name HDFS - alias HDFS Checks - } - {% endif %} -{% endif %} -{%if hostgroup_defs['jobtracker'] or - hostgroup_defs['historyserver2']-%} -define servicegroup { - servicegroup_name MAPREDUCE - alias MAPREDUCE Checks -} -{% endif %} -{%if hostgroup_defs['resourcemanager'] or - hostgroup_defs['nodemanagers'] %} -define servicegroup { - servicegroup_name YARN - alias YARN Checks -} -{% endif %} -{%if hostgroup_defs['hbasemasters'] %} -define servicegroup { - servicegroup_name HBASE - alias HBASE Checks -} -{% endif %} -{% if hostgroup_defs['oozie-server'] %} -define servicegroup { - servicegroup_name OOZIE - alias OOZIE Checks -} -{% endif %} -{% if hostgroup_defs['nagios-server'] %} -define servicegroup { - servicegroup_name NAGIOS - alias NAGIOS Checks -} -{% endif %} -{% if hostgroup_defs['ganglia-server'] %} -define servicegroup { - servicegroup_name GANGLIA - alias GANGLIA Checks -} -{% endif %} -{% if hostgroup_defs['hiveserver'] or hostgroup_defs['webhcat-server'] %} -define servicegroup { - servicegroup_name HIVE - alias HIVE Checks -} -{% endif %} -{% if hostgroup_defs['zookeeper-servers'] %} -define servicegroup { - servicegroup_name ZOOKEEPER - alias ZOOKEEPER Checks -} -{% endif %} -define servicegroup { - servicegroup_name AMBARI - alias AMBARI Checks -} -{% if hostgroup_defs['hue-server'] %} -define servicegroup { - servicegroup_name HUE - alias HUE Checks -} -{% endif %} -{% if hostgroup_defs['nimbus'] or - hostgroup_defs['drpc-server'] or - hostgroup_defs['storm_ui'] or - hostgroup_defs['supervisors'] or - hostgroup_defs['storm_rest_api']%} -define servicegroup { - servicegroup_name STORM - alias STORM Checks -} -{% endif %} -{% if hostgroup_defs['falcon-server'] %} -define servicegroup { - servicegroup_name FALCON - alias FALCON Checks -} -{% endif %} - -{%if hostgroup_defs['flume-servers'] %} -define servicegroup { - servicegroup_name FLUME - alias FLUME Checks -} -{% endif %} http://git-wip-us.apache.org/repos/asf/ambari/blob/6fcd95fe/ambari-server/src/main/resources/stacks/PHD/3.0.0.0/services/NAGIOS/package/templates/hadoop-services.cfg.j2 ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/PHD/3.0.0.0/services/NAGIOS/package/templates/hadoop-services.cfg.j2 b/ambari-server/src/main/resources/stacks/PHD/3.0.0.0/services/NAGIOS/package/templates/hadoop-services.cfg.j2 deleted file mode 100644 index f278260..0000000 --- a/ambari-server/src/main/resources/stacks/PHD/3.0.0.0/services/NAGIOS/package/templates/hadoop-services.cfg.j2 +++ /dev/null @@ -1,791 +0,0 @@ -{# -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -#} - -# -# -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, -# software distributed under the License is distributed on an -# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY -# KIND, either express or implied. See the License for the -# specific language governing permissions and limitations -# under the License. -# -# - -{# TODO: Look for { or } in created file #} -# NAGIOS SERVER Check (status log update) -{% if hostgroup_defs['nagios-server'] %} -define service { - name hadoop-service - use generic-service - notification_options w,u,c,r,f,s - first_notification_delay 0 - notification_interval 0 # Send the notification once - contact_groups admins - notifications_enabled 1 - event_handler_enabled 1 - register 0 -} - -define service { - hostgroup_name nagios-server - use hadoop-service - service_description NAGIOS::Nagios status log freshness - servicegroups NAGIOS - check_command check_nagios!10!/var/nagios/status.dat!{{nagios_lookup_daemon_str}} - normal_check_interval 5 - retry_check_interval 0.5 - max_check_attempts 2 -} - -# NAGIOS SERVER HDFS Checks -{% if hostgroup_defs['namenode'] != None %} -define service { - hostgroup_name nagios-server - use hadoop-service - service_description HDFS::Percent DataNodes with space available - servicegroups HDFS - check_command check_aggregate!"DATANODE::DataNode space"!10%!30% - normal_check_interval 0.5 - retry_check_interval 1 - max_check_attempts 1 -} - -define service { - hostgroup_name nagios-server - use hadoop-service - service_description HDFS::Percent DataNodes live - servicegroups HDFS - check_command check_aggregate!"DATANODE::DataNode process"!10%!30% - normal_check_interval 0.5 - retry_check_interval 0.25 - max_check_attempts 3 -} -{% endif %} -{# used only for HDP2 #} -{% if hostgroup_defs['namenode'] and hostgroup_defs['namenode'] != None and dfs_ha_enabled %} -define service { - hostgroup_name nagios-server - use hadoop-service - service_description HDFS::NameNode HA Healthy - servicegroups HDFS - check_command check_namenodes_ha!$HOSTGROUPMEMBERS:namenode$!{{ namenode_port }} - normal_check_interval 0.5 - retry_check_interval 0.25 - max_check_attempts 5 -} -{% endif %} - -# AMBARI AGENT Checks -{% for hostname in all_hosts %} -define service { - host_name {{ hostname }} - use hadoop-service - service_description AMBARI::Ambari Agent process - servicegroups AMBARI - check_command check_tcp_wrapper!{{all_ping_ports[loop.index-1]}}!-w 1 -c 1 - normal_check_interval 1 - retry_check_interval 0.25 - max_check_attempts 4 -} - -{% endfor %} - -# NAGIOS SERVER ZOOKEEPER Checks -{% if hostgroup_defs['zookeeper-servers'] %} -define service { - hostgroup_name nagios-server - use hadoop-service - service_description ZOOKEEPER::Percent ZooKeeper Servers live - servicegroups ZOOKEEPER - check_command check_aggregate!"ZOOKEEPER::ZooKeeper Server process"!35%!70% - normal_check_interval 0.5 - retry_check_interval 0.25 - max_check_attempts 3 -} -{% endif %} - -# NAGIOS SERVER HBASE Checks -{% if hostgroup_defs['hbasemasters'] %} -define service { - hostgroup_name nagios-server - use hadoop-service - service_description HBASE::Percent RegionServers live - servicegroups HBASE - check_command check_aggregate!"REGIONSERVER::RegionServer process"!10%!30% - normal_check_interval 0.5 - retry_check_interval 0.25 - max_check_attempts 3 -} -{% endif %} -{% endif %} - - - -# GANGLIA SERVER Checks -{% if hostgroup_defs['ganglia-server'] %} -define service { - hostgroup_name ganglia-server - use hadoop-service - service_description GANGLIA::Ganglia Server process - servicegroups GANGLIA - check_command check_tcp_wrapper!{{ ganglia_port }}!-w 1 -c 1 - normal_check_interval 0.25 - retry_check_interval 0.25 - max_check_attempts 4 -} - -{% if hostgroup_defs['namenode'] %} -define service { - hostgroup_name ganglia-server - use hadoop-service - service_description GANGLIA::Ganglia Monitor process for NameNode - servicegroups GANGLIA - check_command check_tcp_wrapper!{{ ganglia_collector_namenode_port }}!-w 1 -c 1 - normal_check_interval 0.25 - retry_check_interval 0.25 - max_check_attempts 4 -} -{% endif %} - -{% if hostgroup_defs['hbasemasters'] %} -define service { - hostgroup_name ganglia-server - use hadoop-service - service_description GANGLIA::Ganglia Monitor process for HBase Master - servicegroups GANGLIA - check_command check_tcp_wrapper!{{ ganglia_collector_hbase_port }}!-w 1 -c 1 - normal_check_interval 0.25 - retry_check_interval 0.25 - max_check_attempts 4 -} -{% endif %} - -{% if hostgroup_defs['resourcemanager'] %} -define service { - hostgroup_name ganglia-server - use hadoop-service - service_description GANGLIA::Ganglia Monitor process for ResourceManager - servicegroups GANGLIA - check_command check_tcp_wrapper!{{ ganglia_collector_rm_port }}!-w 1 -c 1 - normal_check_interval 0.25 - retry_check_interval 0.25 - max_check_attempts 4 -} -{% endif %} - -{% if hostgroup_defs['historyserver2'] %} -define service { - hostgroup_name ganglia-server - use hadoop-service - service_description GANGLIA::Ganglia Monitor process for HistoryServer - servicegroups GANGLIA - check_command check_tcp_wrapper!{{ ganglia_collector_hs_port }}!-w 1 -c 1 - normal_check_interval 0.25 - retry_check_interval 0.25 - max_check_attempts 4 -} -{% endif %} -{% endif %} - - -{% if hostgroup_defs['snamenode'] and hostgroup_defs['namenode'] != None %} -# Secondary namenode checks -define service { - hostgroup_name snamenode - use hadoop-service - service_description NAMENODE::Secondary NameNode process - servicegroups HDFS - check_command check_tcp_wrapper!{{ snamenode_port }}!-w 1 -c 1 - normal_check_interval 0.5 - retry_check_interval 0.25 - max_check_attempts 3 -} -{% endif %} - -{% if hostgroup_defs['storm_ui'] %} -# STORM UI Checks -define service { - hostgroup_name storm_ui - use hadoop-service - service_description STORM_UI_SERVER::Storm UI on {{ hostgroup_defs['storm_ui'][0] }} - servicegroups STORM - check_command check_webui!storm_ui!{{ storm_ui_port }} - normal_check_interval 1 - retry_check_interval 1 - max_check_attempts 3 -} -{% endif %} - -{% if hostgroup_defs['storm_ui'] %} -# STORM UI Checks -define service { - hostgroup_name storm_ui - use hadoop-service - service_description STORM_UI_SERVER::Storm UI Server process - servicegroups STORM - check_command check_tcp_wrapper!{{ storm_ui_port }}!-w 1 -c 1 - normal_check_interval 0.5 - retry_check_interval 0.25 - max_check_attempts 3 -} -{% endif %} - -{% if hostgroup_defs['nimbus'] %} -# Nimbus Checks -define service { - hostgroup_name nimbus - use hadoop-service - service_description NIMBUS::Nimbus process - servicegroups STORM - check_command check_tcp_wrapper!{{ nimbus_port }}!-w 1 -c 1 - normal_check_interval 0.5 - retry_check_interval 0.25 - max_check_attempts 3 -} -{% endif %} - -{% if hostgroup_defs['drpc-server'] %} -# drpc Checks -define service { - hostgroup_name drpc-server - use hadoop-service - service_description DRPC_SERVER::DRPC Server process - servicegroups STORM - check_command check_tcp_wrapper!{{ drpc_port }}!-w 1 -c 1 - normal_check_interval 0.5 - retry_check_interval 0.25 - max_check_attempts 3 -} -{% endif %} - -{% if hostgroup_defs['storm_rest_api'] %} -# Storm REST API Checks -define service { - hostgroup_name storm_rest_api - use hadoop-service - service_description STORM_REST_API::Storm REST API Server process - servicegroups STORM - check_command check_tcp_wrapper!{{ storm_rest_api_port }}!-w 1 -c 1 - normal_check_interval 0.5 - retry_check_interval 0.25 - max_check_attempts 3 -} -{% endif %} - -# NAGIOS SERVER Supervisor Checks -{% if hostgroup_defs['supervisors'] %} -define service { - hostgroup_name nagios-server - use hadoop-service - service_description SUPERVISOR::Percent Supervisors live - servicegroups STORM - check_command check_aggregate!"SUPERVISOR::Supervisors process"!10%!30% - normal_check_interval 0.5 - retry_check_interval 0.25 - max_check_attempts 3 -} - -define service { - hostgroup_name supervisors - use hadoop-service - service_description SUPERVISOR::Supervisors process - servicegroups STORM - check_command check_tcp_wrapper!{{ supervisor_port }}!-w 1 -c 1 - normal_check_interval 1 - retry_check_interval 0.5 - max_check_attempts 3 -} -{% endif %} - -{% if hostgroup_defs['namenode'] and hostgroup_defs['namenode'] != None %} -# HDFS Checks -{% for namenode_hostname in namenode_host %} -{# TODO: check if we can get rid of str, lower #} -define service { - host_name {{ namenode_hostname }} - use hadoop-service - service_description NAMENODE::NameNode edit logs directory status on {{ namenode_hostname }} - servicegroups HDFS - check_command check_name_dir_status!{{ namenode_port }}!{{ str(hdfs_ssl_enabled).lower() }}!{{ nagios_keytab_path }}!{{ nagios_principal_name }}!{{ kinit_path_local }}!{{ str(security_enabled).lower() }} - normal_check_interval 0.5 - retry_check_interval 0.5 - max_check_attempts 3 -} - -{% if check_cpu_on %} -define service { - host_name {{ namenode_hostname }} - use hadoop-service - service_description NAMENODE::NameNode host CPU utilization on {{ namenode_hostname }} - servicegroups HDFS - check_command check_cpu!{{ namenode_port }}!200%!250%!{{ str(hdfs_ssl_enabled).lower() }}!{{ nagios_keytab_path }}!{{ nagios_principal_name }}!{{ kinit_path_local }}!{{ str(security_enabled).lower() }} - normal_check_interval 5 - retry_check_interval 2 - max_check_attempts 5 -} -{% endif %} - -define service { - host_name {{ namenode_hostname }} - use hadoop-service - service_description NAMENODE::NameNode Web UI on {{ namenode_hostname }} - servicegroups HDFS - check_command check_webui!namenode!{{ namenode_port }} - normal_check_interval 1 - retry_check_interval 1 - max_check_attempts 3 -} - -define service { - host_name {{ namenode_hostname }} - use hadoop-service - service_description NAMENODE::NameNode process on {{ namenode_hostname }} - servicegroups HDFS - check_command check_tcp_wrapper!{{nn_ha_host_port_map[namenode_hostname]}}!-w 1 -c 1 - normal_check_interval 0.5 - retry_check_interval 0.25 - max_check_attempts 3 -} - -define service { - host_name {{ namenode_hostname }} - use hadoop-service - service_description HDFS::NameNode RPC latency on {{ namenode_hostname }} - servicegroups HDFS - check_command check_rpcq_latency!NameNode!{{ namenode_port }}!3000!5000!{{ str(hdfs_ssl_enabled).lower() }}!{{ nagios_keytab_path }}!{{ nagios_principal_name }}!{{ kinit_path_local }}!{{ str(security_enabled).lower() }} - normal_check_interval 5 - retry_check_interval 1 - max_check_attempts 5 -} - -{% endfor %} - -define service { - host_name {{namenode_host[0]}} - use hadoop-service - service_description NAMENODE::Last checkpoint time - servicegroups HDFS - check_command check_checkpoint_time!{{ nn_hosts_string }}!{{ namenode_port }}!200!200!{{ dfs_namenode_checkpoint_period }}!{{dfs_namenode_checkpoint_txns}}!{{str(hdfs_ssl_enabled).lower()}} - normal_check_interval 0.5 - retry_check_interval 0.25 - max_check_attempts 3 -} - -define service { - hostgroup_name nagios-server - use hadoop-service - service_description HDFS::Blocks health - servicegroups HDFS - check_command check_hdfs_blocks!$HOSTGROUPMEMBERS:namenode$!{{ namenode_port }}!{{ nn_metrics_property }}!{{ str(hdfs_ssl_enabled).lower() }}!{{ nagios_keytab_path }}!{{ nagios_principal_name }}!{{ kinit_path_local }}!{{ str(security_enabled).lower() }} - normal_check_interval 2 - retry_check_interval 1 - max_check_attempts 1 -} - -define service { - hostgroup_name nagios-server - use hadoop-service - service_description HDFS::HDFS capacity utilization - servicegroups HDFS - check_command check_hdfs_capacity!$HOSTGROUPMEMBERS:namenode$!{{ namenode_port }}!80%!90%!{{ str(hdfs_ssl_enabled).lower() }}!{{ nagios_keytab_path }}!{{ nagios_principal_name }}!{{ kinit_path_local }}!{{ str(security_enabled).lower() }} - normal_check_interval 2 - retry_check_interval 1 - max_check_attempts 1 -} - -{% endif %} - -{% if hostgroup_defs['resourcemanager'] %} -# YARN::RESOURCEMANAGER Checks -define service { - hostgroup_name nagios-server - use hadoop-service - service_description RESOURCEMANAGER::ResourceManager Web UI - servicegroups YARN - check_command check_webui_ha!resourcemanager!{{ rm_hosts_in_str }}!{{ rm_port }} - normal_check_interval 1 - retry_check_interval 1 - max_check_attempts 3 -} - -{% if check_cpu_on %} -define service { - hostgroup_name nagios-server - use hadoop-service - service_description RESOURCEMANAGER::ResourceManager CPU utilization - servicegroups YARN - check_command check_cpu_ha!{{ rm_hosts_in_str }}!{{ rm_port }}!200%!250%!{{ str(yarn_ssl_enabled).lower() }}!{{ nagios_keytab_path }}!{{ nagios_principal_name }}!{{ kinit_path_local }}!{{ str(security_enabled).lower() }} - normal_check_interval 5 - retry_check_interval 2 - max_check_attempts 5 -} -{% endif %} - -define service { - hostgroup_name nagios-server - use hadoop-service - service_description RESOURCEMANAGER::ResourceManager RPC latency - servicegroups YARN - check_command check_rpcq_latency_ha!{{ rm_hosts_in_str }}!ResourceManager!{{ rm_port }}!3000!5000!{{ str(yarn_ssl_enabled).lower() }}!{{ nagios_keytab_path }}!{{ nagios_principal_name }}!{{ kinit_path_local }}!{{ str(security_enabled).lower() }} - normal_check_interval 5 - retry_check_interval 1 - max_check_attempts 5 -} - -{% for rm_host in _rm_host %} -define service { - host_name {{ rm_host }} - use hadoop-service - service_description RESOURCEMANAGER::ResourceManager process on {{ rm_host }} - servicegroups YARN - check_command check_tcp_wrapper!{{ rm_port }}!-w 1 -c 1 - normal_check_interval 1 - retry_check_interval 0.5 - max_check_attempts 3 -} -{% endfor %} -{% endif %} - -{% if hostgroup_defs['nodemanagers'] %} -# YARN::NODEMANAGER Checks -define service { - hostgroup_name nodemanagers - use hadoop-service - service_description NODEMANAGER::NodeManager process - servicegroups YARN - check_command check_tcp_wrapper!{{ nm_port }}!-w 1 -c 1 - normal_check_interval 1 - retry_check_interval 0.5 - max_check_attempts 3 -} - -define service { - hostgroup_name nodemanagers - use hadoop-service - service_description NODEMANAGER::NodeManager health - servicegroups YARN - check_command check_nodemanager_health!{{ nm_port }}!{{ str(security_enabled).lower() }}!{{ str(yarn_ssl_enabled).lower() }}!{{ nagios_keytab_path }}!{{ nagios_principal_name }}!{{ kinit_path_local }} - normal_check_interval 1 - retry_check_interval 1 - max_check_attempts 3 -} -define service { - hostgroup_name nagios-server - use hadoop-service - service_description NODEMANAGER::Percent NodeManagers live - servicegroups YARN - check_command check_aggregate!"NODEMANAGER::NodeManager process"!10%!30% - normal_check_interval 0.5 - retry_check_interval 0.25 - max_check_attempts 3 -} -{% endif %} - -{% if hostgroup_defs['historyserver2'] %} -# MAPREDUCE::JOBHISTORY Checks -define service { - hostgroup_name historyserver2 - use hadoop-service - service_description JOBHISTORY::HistoryServer Web UI - servicegroups MAPREDUCE - check_command check_webui!historyserver2!{{ hs_port }} - normal_check_interval 1 - retry_check_interval 1 - max_check_attempts 3 -} - -{% if check_cpu_on %} -define service { - hostgroup_name historyserver2 - use hadoop-service - service_description JOBHISTORY::HistoryServer CPU utilization - servicegroups MAPREDUCE - check_command check_cpu!{{ hs_port }}!200%!250%!{{ str(mapreduce_ssl_enabled).lower() }}!{{ nagios_keytab_path }}!{{ nagios_principal_name }}!{{ kinit_path_local }}!{{ str(security_enabled).lower() }} - normal_check_interval 5 - retry_check_interval 2 - max_check_attempts 5 -} -{% endif %} - -define service { - hostgroup_name historyserver2 - use hadoop-service - service_description JOBHISTORY::HistoryServer RPC latency - servicegroups MAPREDUCE - check_command check_rpcq_latency!JobHistoryServer!{{ hs_port }}!3000!5000!{{ str(mapreduce_ssl_enabled).lower() }}!{{ nagios_keytab_path }}!{{ nagios_principal_name }}!{{ kinit_path_local }}!{{ str(security_enabled).lower() }} - normal_check_interval 5 - retry_check_interval 1 - max_check_attempts 5 -} - -define service { - hostgroup_name historyserver2 - use hadoop-service - service_description JOBHISTORY::HistoryServer process - servicegroups MAPREDUCE - check_command check_tcp_wrapper!{{ hs_port }}!-w 1 -c 1 - normal_check_interval 1 - retry_check_interval 0.5 - max_check_attempts 3 -} - -{% endif %} - -{% if hostgroup_defs['journalnodes'] %} -# Journalnode checks -define service { - hostgroup_name journalnodes - use hadoop-service - service_description JOURNALNODE::JournalNode process - servicegroups HDFS - check_command check_tcp_wrapper!{{ journalnode_port }}!-w 1 -c 1 - normal_check_interval 1 - retry_check_interval 0.5 - max_check_attempts 3 -} - -{% if dfs_ha_enabled %} -define service { - hostgroup_name nagios-server - use hadoop-service - service_description HDFS::Percent JournalNodes live - servicegroups HDFS - check_command check_aggregate!"JOURNALNODE::JournalNode process"!33%!50% - normal_check_interval 0.5 - retry_check_interval 0.25 - max_check_attempts 3 -} -{% endif %} -{% endif %} - -{% if hostgroup_defs['slaves'] and hostgroup_defs['namenode'] != None %} -# HDFS::DATANODE Checks -define service { - hostgroup_name slaves - use hadoop-service - service_description DATANODE::DataNode process - servicegroups HDFS - check_command check_tcp_wrapper!{{datanode_port}}!-w 1 -c 1 - normal_check_interval 1 - retry_check_interval 0.5 - max_check_attempts 3 -} - -define service { - hostgroup_name slaves - use hadoop-service - service_description DATANODE::DataNode space - servicegroups HDFS - check_command check_datanode_storage!{{ datanode_port }}!90%!90%!{{ str(hdfs_ssl_enabled).lower() }}!{{ nagios_keytab_path }}!{{ nagios_principal_name }}!{{ kinit_path_local }}!{{ str(security_enabled).lower() }} - normal_check_interval 2 - retry_check_interval 1 - max_check_attempts 2 -} - -{% endif %} - -{% if hostgroup_defs['zookeeper-servers'] %} -# ZOOKEEPER Checks -define service { - hostgroup_name zookeeper-servers - use hadoop-service - service_description ZOOKEEPER::ZooKeeper Server process - servicegroups ZOOKEEPER - check_command check_tcp_wrapper!{{ clientPort }}!-w 1 -c 1 - normal_check_interval 1 - retry_check_interval 0.5 - max_check_attempts 3 -} -{% endif %} - -{% if hostgroup_defs['hbasemasters'] %} -# HBASE::REGIONSERVER Checks -define service { - hostgroup_name region-servers - use hadoop-service - service_description REGIONSERVER::RegionServer process - servicegroups HBASE - check_command check_tcp_wrapper!{{ hbase_rs_port }}!-w 1 -c 1 - normal_check_interval 1 - retry_check_interval 0.5 - max_check_attempts 3 -} - -{% if hostgroup_defs['hbasemasters'] %} -{% if check_cpu_on %} -define service { - hostgroup_name nagios-server - use hadoop-service - service_description HBASEMASTER::HBase Master CPU utilization - servicegroups HBASE - check_command check_cpu_ha!{{ hbase_master_hosts_in_str }}!{{ hbase_master_port }}!200%!250%!false!{{ nagios_keytab_path }}!{{ nagios_principal_name }}!{{ kinit_path_local }}!{{ str(security_enabled).lower() }} - normal_check_interval 5 - retry_check_interval 2 - max_check_attempts 5 -} -{% endif %} -{% endif %} - -{% for hbasemaster in hbase_master_hosts %} -define service { - host_name {{ hbasemaster }} - use hadoop-service - service_description HBASEMASTER::HBase Master process on {{ hbasemaster }} - servicegroups HBASE - check_command check_tcp_wrapper!{{ hbase_master_rpc_port }}!-w 1 -c 1 - normal_check_interval 0.5 - retry_check_interval 0.25 - max_check_attempts 4 -} -{% endfor %} -{% endif %} - -{% if hostgroup_defs['hiveserver'] %} -# HIVE Metastore check -define service { - hostgroup_name hiveserver - use hadoop-service - service_description HIVE-METASTORE::Hive Metastore process - servicegroups HIVE - check_command check_tcp_wrapper!{{ hive_metastore_port }}!-w 1 -c 1 - normal_check_interval 0.5 - retry_check_interval 0.5 - max_check_attempts 3 -} - -# HIVE Server check -define service { - hostgroup_name hiveserver - use hadoop-service - service_description HIVE-SERVER::HiveServer2 process - servicegroups HIVE - check_command check_tcp_wrapper_sasl!{{ hive_server_port }}!{{ '--security-enabled' if security_enabled else '' }}!-w 1 -c 1 - normal_check_interval 0.5 - retry_check_interval 0.5 - max_check_attempts 3 -} -{% endif %} -{% if hostgroup_defs['oozie-server'] %} -# Oozie check -define service { - hostgroup_name oozie-server - use hadoop-service - service_description OOZIE::Oozie Server status - servicegroups OOZIE - {% if security_enabled %} - check_command check_oozie_status!{{ oozie_server_port }}!{{ java64_home }}!true!{{ nagios_keytab_path }}!{{ nagios_principal_name }}!{{ kinit_path_local }} - {% else %} - check_command check_oozie_status!{{ oozie_server_port }}!{{ java64_home }}!false - {% endif %} - normal_check_interval 1 - retry_check_interval 1 - max_check_attempts 3 -} -{% endif %} -{% if hostgroup_defs['webhcat-server'] %} -# WEBHCAT check -define service { - hostgroup_name webhcat-server - use hadoop-service - service_description WEBHCAT::WebHCat Server status - servicegroups HIVE - {% if security_enabled %} - check_command check_templeton_status!{{ templeton_port }}!v1!{{ str(security_enabled).lower() }}!{{ nagios_keytab_path }}!{{ nagios_principal_name }}!{{ kinit_path_local }} - {% else %} - check_command check_templeton_status!{{ templeton_port }}!v1!false - {% endif %} - normal_check_interval 1 - retry_check_interval 0.5 - max_check_attempts 3 -} -{% endif %} - -{% if hostgroup_defs['hue-server'] %} -define service { - hostgroup_name hue-server - use hadoop-service - service_description HUE::Hue Server status - servicegroups HUE - check_command check_hue_status - normal_check_interval 100 - retry_check_interval 0.5 - max_check_attempts 3 -} -{% endif %} - -#FALCON checks -{% if hostgroup_defs['falcon-server'] %} -define service { - hostgroup_name falcon-server - service_description FALCON::Falcon Server process - servicegroups FALCON - check_command check_tcp_wrapper!{{ falcon_port }}!-w 1 -c 1 - normal_check_interval 1 - retry_check_interval 0.5 - max_check_attempts 3 -} -define service { - hostgroup_name falcon-server - service_description FALCON::Falcon Server Web UI - servicegroups FALCON - check_command check_webui!falconserver!{{ falcon_port }} - normal_check_interval 1 - retry_check_interval 0.5 - max_check_attempts 3 -} -{% endif %} -{% if hostgroup_defs['ats-servers'] %} -define service { - hostgroup_name ats-servers - use hadoop-service - service_description APP_TIMELINE_SERVER::App Timeline Server process - servicegroups YARN - check_command check_tcp_wrapper!{{ ahs_port }}!-w 1 -c 1 - normal_check_interval 1 - retry_check_interval 0.5 - max_check_attempts 3 -} -{% endif %} - -{% if hostgroup_defs['flume-servers'] %} -# FLUME Checks -define service { - hostgroup_name flume-servers - use hadoop-service - service_description FLUME::Flume Agent process - servicegroups FLUME - check_command check_ambari!/var/nagios/ambari.json!flume_agent - normal_check_interval 1 - retry_check_interval 0.5 - max_check_attempts 3 -} -{% endif %} - - -