ambari-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From dmitriu...@apache.org
Subject [4/4] ambari git commit: AMBARI-17769. property_description is null for some properties in config page (dlysnichenko)
Date Mon, 18 Jul 2016 14:27:57 GMT
AMBARI-17769. property_description is null for some properties in config page (dlysnichenko)


Project: http://git-wip-us.apache.org/repos/asf/ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/77c2dc6e
Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/77c2dc6e
Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/77c2dc6e

Branch: refs/heads/branch-2.4
Commit: 77c2dc6e5abf5ef7cd975f92a05d25187ebcc190
Parents: bdf8880
Author: Lisnichenko Dmitro <dlysnichenko@hortonworks.com>
Authored: Mon Jul 18 17:26:44 2016 +0300
Committer: Lisnichenko Dmitro <dlysnichenko@hortonworks.com>
Committed: Mon Jul 18 17:27:27 2016 +0300

----------------------------------------------------------------------
 .../services/HDFS/configuration/core-site.xml   |  4 ++-
 .../services/HDFS/configuration/hdfs-site.xml   | 13 +++++++++-
 .../YARN/configuration/capacity-scheduler.xml   |  6 +++++
 .../services/YARN/configuration/yarn-site.xml   | 26 ++++++++++++++++---
 .../YARN/configuration/capacity-scheduler.xml   |  6 +++++
 .../services/YARN/configuration/yarn-site.xml   | 22 +++++++++++++---
 .../services/HDFS/configuration/hdfs-site.xml   |  8 ++++++
 .../services/YARN/configuration/yarn-site.xml   | 19 ++++++++++++++
 .../YARN/configuration/capacity-scheduler.xml   |  1 +
 .../services/YARN/configuration/yarn-site.xml   |  2 ++
 .../services/HDFS/configuration/core-site.xml   |  4 ++-
 .../services/HDFS/configuration/hdfs-site.xml   |  9 ++++++-
 .../YARN/configuration/capacity-scheduler.xml   |  6 +++++
 .../services/YARN/configuration/yarn-site.xml   | 27 +++++++++++++++++---
 .../services/HDFS/configuration/hdfs-site.xml   |  8 ++++++
 .../services/HDFS/configuration/core-site.xml   |  4 ++-
 .../services/HDFS/configuration/hdfs-site.xml   |  9 ++++++-
 17 files changed, 156 insertions(+), 18 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/ambari/blob/77c2dc6e/ambari-server/src/main/resources/stacks/HDP/2.2/services/HDFS/configuration/core-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.2/services/HDFS/configuration/core-site.xml
b/ambari-server/src/main/resources/stacks/HDP/2.2/services/HDFS/configuration/core-site.xml
index fdb7ce3..846b09e 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.2/services/HDFS/configuration/core-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/2.2/services/HDFS/configuration/core-site.xml
@@ -21,7 +21,9 @@
   <property>
     <name>hadoop.http.authentication.simple.anonymous.allowed</name>
     <value>true</value>
-    <description/>
+    <description>
+      Indicates if anonymous requests are allowed when using &apos;simple&apos; authentication.
+    </description>
     <on-ambari-upgrade add="true"/>
   </property>
   <property>

http://git-wip-us.apache.org/repos/asf/ambari/blob/77c2dc6e/ambari-server/src/main/resources/stacks/HDP/2.2/services/HDFS/configuration/hdfs-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.2/services/HDFS/configuration/hdfs-site.xml
b/ambari-server/src/main/resources/stacks/HDP/2.2/services/HDFS/configuration/hdfs-site.xml
index adcd9c8..aad524d 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.2/services/HDFS/configuration/hdfs-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/2.2/services/HDFS/configuration/hdfs-site.xml
@@ -21,7 +21,14 @@
   <property>
     <name>dfs.namenode.startup.delay.block.deletion.sec</name>
     <value>3600</value>
-    <description/>
+    <description>
+      The delay in seconds at which we will pause the blocks deletion
+      after Namenode startup. By default it's disabled.
+      In the case a directory has large number of directories and files are
+      deleted, suggested delay is one hour to give the administrator enough time
+      to notice large number of pending deletion blocks and take corrective
+      action.
+    </description>
     <on-ambari-upgrade add="true"/>
   </property>
   <property>
@@ -44,6 +51,10 @@
   </property>
   <property>
     <name>dfs.encryption.key.provider.uri</name>
+    <description>
+      The KeyProvider to use when interacting with encryption keys used
+      when reading and writing to an encryption zone.
+    </description>
     <value/>
     <value-attributes>
       <empty-value-valid>true</empty-value-valid>

http://git-wip-us.apache.org/repos/asf/ambari/blob/77c2dc6e/ambari-server/src/main/resources/stacks/HDP/2.2/services/YARN/configuration/capacity-scheduler.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.2/services/YARN/configuration/capacity-scheduler.xml
b/ambari-server/src/main/resources/stacks/HDP/2.2/services/YARN/configuration/capacity-scheduler.xml
index 90e64b0..084b3c2 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.2/services/YARN/configuration/capacity-scheduler.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/2.2/services/YARN/configuration/capacity-scheduler.xml
@@ -18,6 +18,12 @@
 <configuration supports_final="false" supports_adding_forbidden="true">
   <property>
     <name>yarn.scheduler.capacity.resource-calculator</name>
+    <description>
+      The ResourceCalculator implementation to be used to compare Resources in the scheduler.
+      The default i.e. org.apache.hadoop.yarn.util.resource.DefaultResourseCalculator only
uses
+      Memory while DominantResourceCalculator uses Dominant-resource to compare multi-dimensional
+      resources such as Memory, CPU etc. A Java ResourceCalculator class name is expected.
+    </description>
     <value>org.apache.hadoop.yarn.util.resource.DefaultResourceCalculator</value>
     <display-name>CPU Scheduling</display-name>
     <value-attributes>

http://git-wip-us.apache.org/repos/asf/ambari/blob/77c2dc6e/ambari-server/src/main/resources/stacks/HDP/2.2/services/YARN/configuration/yarn-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.2/services/YARN/configuration/yarn-site.xml
b/ambari-server/src/main/resources/stacks/HDP/2.2/services/YARN/configuration/yarn-site.xml
index 2d545d7..f68f0fe 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.2/services/YARN/configuration/yarn-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/2.2/services/YARN/configuration/yarn-site.xml
@@ -210,7 +210,14 @@
   <property>
     <name>yarn.nodemanager.resource.cpu-vcores</name>
     <value>8</value>
-    <description/>
+    <description>Number of vcores that can be allocated
+      for containers. This is used by the RM scheduler when allocating
+      resources for containers. This is not used to limit the number of
+      CPUs used by YARN containers. If it is set to -1 and
+      yarn.nodemanager.resource.detect-hardware-capabilities is true, it is
+      automatically determined from the hardware in case of Windows and Linux.
+      In other cases, number of vcores is 8 by default.
+    </description>
     <display-name>Number of virtual cores</display-name>
     <value-attributes>
       <type>int</type>
@@ -247,7 +254,11 @@
   <property>
     <name>yarn.node-labels.fs-store.retry-policy-spec</name>
     <value>2000, 500</value>
-    <description/>
+    <description>
+      Retry policy used for FileSystem node label store. The policy is
+      specified by N pairs of sleep-time in milliseconds and number-of-retries
+      &quot;s1,n1,s2,n2,...&quot;.
+    </description>
     <on-ambari-upgrade add="true"/>
   </property>
   <property>
@@ -292,7 +303,7 @@
   <property>
     <name>yarn.resourcemanager.system-metrics-publisher.dispatcher.pool-size</name>
     <value>10</value>
-    <description/>
+    <description>Number of worker threads that send the yarn system metrics data.</description>
     <on-ambari-upgrade add="true"/>
   </property>
   <property>
@@ -421,7 +432,9 @@
   <property>
     <name>yarn.node-labels.fs-store.root-dir</name>
     <value>/system/yarn/node-labels</value>
-    <description/>
+    <description>
+      URI for NodeLabelManager.
+    </description>
     <on-ambari-upgrade add="true"/>
   </property>
   <property>
@@ -526,6 +539,11 @@
   </property>
   <property>
     <name>yarn.resourcemanager.scheduler.monitor.enable</name>
+    <description>
+      Enable a set of periodic monitors (specified in
+      yarn.resourcemanager.scheduler.monitor.policies) that affect the
+      scheduler.
+    </description>
     <value>false</value>
     <display-name>Pre-emption</display-name>
     <value-attributes>

http://git-wip-us.apache.org/repos/asf/ambari/blob/77c2dc6e/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/YARN/configuration/capacity-scheduler.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/YARN/configuration/capacity-scheduler.xml
b/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/YARN/configuration/capacity-scheduler.xml
index 2ce0d25..40748a2 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/YARN/configuration/capacity-scheduler.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/YARN/configuration/capacity-scheduler.xml
@@ -18,6 +18,12 @@
 <configuration supports_final="false" supports_adding_forbidden="true">
   <property>
     <name>yarn.scheduler.capacity.resource-calculator</name>
+    <description>
+      The ResourceCalculator implementation to be used to compare Resources in the scheduler.
+      The default i.e. org.apache.hadoop.yarn.util.resource.DefaultResourseCalculator only
uses
+      Memory while DominantResourceCalculator uses Dominant-resource to compare multi-dimensional
+      resources such as Memory, CPU etc. A Java ResourceCalculator class name is expected.
+    </description>
     <value>org.apache.hadoop.yarn.util.resource.DefaultResourceCalculator</value>
     <description/>
     <on-ambari-upgrade add="true"/>

http://git-wip-us.apache.org/repos/asf/ambari/blob/77c2dc6e/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/YARN/configuration/yarn-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/YARN/configuration/yarn-site.xml
b/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/YARN/configuration/yarn-site.xml
index 0eb515c..d029694 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/YARN/configuration/yarn-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/YARN/configuration/yarn-site.xml
@@ -206,7 +206,15 @@
   <property>
     <name>yarn.nodemanager.resource.cpu-vcores</name>
     <value>8</value>
-    <description/>
+    <description>
+      Number of vcores that can be allocated
+      for containers. This is used by the RM scheduler when allocating
+      resources for containers. This is not used to limit the number of
+      CPUs used by YARN containers. If it is set to -1 and
+      yarn.nodemanager.resource.detect-hardware-capabilities is true, it is
+      automatically determined from the hardware in case of Windows and Linux.
+      In other cases, number of vcores is 8 by default.
+    </description>
     <display-name>Total NM CPU vCores available to Containers</display-name>
     <value-attributes>
       <type>int</type>
@@ -236,7 +244,11 @@
   <property>
     <name>yarn.node-labels.fs-store.retry-policy-spec</name>
     <value>2000, 500</value>
-    <description/>
+    <description>
+      Retry policy used for FileSystem node label store. The policy is
+      specified by N pairs of sleep-time in milliseconds and number-of-retries
+      &quot;s1,n1,s2,n2,...&quot;.
+    </description>
     <on-ambari-upgrade add="true"/>
   </property>
   <property>
@@ -281,7 +293,7 @@
   <property>
     <name>yarn.resourcemanager.system-metrics-publisher.dispatcher.pool-size</name>
     <value>10</value>
-    <description/>
+    <description>Number of worker threads that send the yarn system metrics data.</description>
     <on-ambari-upgrade add="true"/>
   </property>
   <property>
@@ -381,7 +393,9 @@
   <property>
     <name>yarn.node-labels.fs-store.root-dir</name>
     <value>/system/yarn/node-labels</value>
-    <description/>
+    <description>
+      URI for NodeLabelManager.
+    </description>
     <on-ambari-upgrade add="true"/>
   </property>
   <property>

http://git-wip-us.apache.org/repos/asf/ambari/blob/77c2dc6e/ambari-server/src/main/resources/stacks/HDP/2.3/services/HDFS/configuration/hdfs-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.3/services/HDFS/configuration/hdfs-site.xml
b/ambari-server/src/main/resources/stacks/HDP/2.3/services/HDFS/configuration/hdfs-site.xml
index d221143..1c4076e 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.3/services/HDFS/configuration/hdfs-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/2.3/services/HDFS/configuration/hdfs-site.xml
@@ -36,6 +36,14 @@
   <property>
     <name>nfs.exports.allowed.hosts</name>
     <value>* rw</value>
+    <description>
+      By default, the export can be mounted by any client. To better control the access,
+      users can update the following property. The value string contains machine name and
access privilege,
+      separated by whitespace characters. Machine name format can be single host, wildcards,
and IPv4
+      networks.The access privilege uses rw or ro to specify readwrite or readonly access
of the machines
+      to exports. If the access privilege is not provided, the default is read-only. Entries
are separated
+      by &quot;;&quot;. For example: &quot;192.168.0.0/22 rw ; host*.example.com
; host1.test.org ro;&quot;.
+    </description>
     <display-name>Allowed hosts</display-name>
     <on-ambari-upgrade add="true"/>
   </property>

http://git-wip-us.apache.org/repos/asf/ambari/blob/77c2dc6e/ambari-server/src/main/resources/stacks/HDP/2.3/services/YARN/configuration/yarn-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.3/services/YARN/configuration/yarn-site.xml
b/ambari-server/src/main/resources/stacks/HDP/2.3/services/YARN/configuration/yarn-site.xml
index 463f80e..68f0e04 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.3/services/YARN/configuration/yarn-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/2.3/services/YARN/configuration/yarn-site.xml
@@ -25,6 +25,10 @@
   </property>
   <property>
     <name>yarn.timeline-service.recovery.enabled</name>
+    <description>
+      Enable timeline server to recover state after starting. If
+      true, then yarn.timeline-service.state-store-class must be specified.
+    </description>
     <value>true</value>
     <on-ambari-upgrade add="true"/>
   </property>
@@ -97,24 +101,39 @@
   <!-- advanced ats v1.5 properties-->
   <property>
     <name>yarn.timeline-service.entity-group-fs-store.summary-store</name>
+    <description>Summary storage for ATS v1.5</description>
     <!-- Use rolling leveldb, advanced -->
     <value>org.apache.hadoop.yarn.server.timeline.RollingLevelDBTimelineStore</value>
     <on-ambari-upgrade add="true"/>
   </property>
   <property>
     <name>yarn.timeline-service.entity-group-fs-store.scan-interval-seconds</name>
+    <description>
+      Scan interval for ATS v1.5 entity group file system storage reader.This
+      value controls how frequent the reader will scan the HDFS active directory
+      for application status.
+    </description>
     <!-- Default is 60 seconds, advanced -->
     <value>60</value>
     <on-ambari-upgrade add="true"/>
   </property>
   <property>
     <name>yarn.timeline-service.entity-group-fs-store.cleaner-interval-seconds</name>
+    <description>
+      Scan interval for ATS v1.5 entity group file system storage cleaner.This
+      value controls how frequent the reader will scan the HDFS done directory
+      for stale application data.
+    </description>
     <!-- 3600 is default, advanced -->
     <value>3600</value>
     <on-ambari-upgrade add="true"/>
   </property>
   <property>
     <name>yarn.timeline-service.entity-group-fs-store.retain-seconds</name>
+    <description>
+      How long the ATS v1.5 entity group file system storage will keep an
+      application's data in the done directory.
+    </description>
     <!-- 7 days is default, advanced -->
     <value>604800</value>
     <on-ambari-upgrade add="true"/>

http://git-wip-us.apache.org/repos/asf/ambari/blob/77c2dc6e/ambari-server/src/main/resources/stacks/HDP/2.5/services/YARN/configuration/capacity-scheduler.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.5/services/YARN/configuration/capacity-scheduler.xml
b/ambari-server/src/main/resources/stacks/HDP/2.5/services/YARN/configuration/capacity-scheduler.xml
index adfa4af..9ff8484 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.5/services/YARN/configuration/capacity-scheduler.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/2.5/services/YARN/configuration/capacity-scheduler.xml
@@ -18,6 +18,7 @@
 <configuration supports_final="false" supports_adding_forbidden="true">
   <property>
     <name>capacity-scheduler</name>
+    <description>Enter key=value (one per line) for all properties of capacity-scheduler.xml</description>
     <depends-on>
       <property>
         <type>hive-interactive-env</type>

http://git-wip-us.apache.org/repos/asf/ambari/blob/77c2dc6e/ambari-server/src/main/resources/stacks/HDP/2.5/services/YARN/configuration/yarn-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.5/services/YARN/configuration/yarn-site.xml
b/ambari-server/src/main/resources/stacks/HDP/2.5/services/YARN/configuration/yarn-site.xml
index f1b73dc..ea7a7af 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.5/services/YARN/configuration/yarn-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/2.5/services/YARN/configuration/yarn-site.xml
@@ -32,11 +32,13 @@
   <property>
     <name>yarn.nodemanager.aux-services.spark_shuffle.classpath</name>
     <value>{{stack_root}}/${hdp.version}/spark/aux/*</value>
+    <description>The auxiliary service classpath to use for Spark</description>
     <on-ambari-upgrade add="true"/>
   </property>
   <property>
     <name>yarn.nodemanager.aux-services.spark2_shuffle.classpath</name>
     <value>{{stack_root}}/${hdp.version}/spark2/aux/*</value>
+    <description>The auxiliary service classpath to use for Spark 2</description>
     <on-ambari-upgrade add="true"/>
   </property>
   <property>

http://git-wip-us.apache.org/repos/asf/ambari/blob/77c2dc6e/ambari-server/src/main/resources/stacks/HDPWIN/2.2/services/HDFS/configuration/core-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDPWIN/2.2/services/HDFS/configuration/core-site.xml
b/ambari-server/src/main/resources/stacks/HDPWIN/2.2/services/HDFS/configuration/core-site.xml
index fb46b74..2ff95c0 100644
--- a/ambari-server/src/main/resources/stacks/HDPWIN/2.2/services/HDFS/configuration/core-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDPWIN/2.2/services/HDFS/configuration/core-site.xml
@@ -21,7 +21,9 @@
   <property>
     <name>hadoop.http.authentication.simple.anonymous.allowed</name>
     <value>true</value>
-    <description/>
+    <description>
+      Indicates if anonymous requests are allowed when using &apos;simple&apos; authentication.
+    </description>
     <on-ambari-upgrade add="true"/>
   </property>
 </configuration>

http://git-wip-us.apache.org/repos/asf/ambari/blob/77c2dc6e/ambari-server/src/main/resources/stacks/HDPWIN/2.2/services/HDFS/configuration/hdfs-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDPWIN/2.2/services/HDFS/configuration/hdfs-site.xml
b/ambari-server/src/main/resources/stacks/HDPWIN/2.2/services/HDFS/configuration/hdfs-site.xml
index caa977f..89254b4 100644
--- a/ambari-server/src/main/resources/stacks/HDPWIN/2.2/services/HDFS/configuration/hdfs-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDPWIN/2.2/services/HDFS/configuration/hdfs-site.xml
@@ -21,7 +21,14 @@
   <property>
     <name>dfs.namenode.startup.delay.block.deletion.sec</name>
     <value>3600</value>
-    <description/>
+    <description>
+      The delay in seconds at which we will pause the blocks deletion
+      after Namenode startup. By default it's disabled.
+      In the case a directory has large number of directories and files are
+      deleted, suggested delay is one hour to give the administrator enough time
+      to notice large number of pending deletion blocks and take corrective
+      action.
+    </description>
     <on-ambari-upgrade add="true"/>
   </property>
 </configuration>

http://git-wip-us.apache.org/repos/asf/ambari/blob/77c2dc6e/ambari-server/src/main/resources/stacks/HDPWIN/2.2/services/YARN/configuration/capacity-scheduler.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDPWIN/2.2/services/YARN/configuration/capacity-scheduler.xml
b/ambari-server/src/main/resources/stacks/HDPWIN/2.2/services/YARN/configuration/capacity-scheduler.xml
index 6a957c0..5483c97 100644
--- a/ambari-server/src/main/resources/stacks/HDPWIN/2.2/services/YARN/configuration/capacity-scheduler.xml
+++ b/ambari-server/src/main/resources/stacks/HDPWIN/2.2/services/YARN/configuration/capacity-scheduler.xml
@@ -19,6 +19,12 @@
   <property>
     <name>yarn.scheduler.capacity.resource-calculator</name>
     <value>org.apache.hadoop.yarn.util.resource.DefaultResourceCalculator</value>
+    <description>
+      The ResourceCalculator implementation to be used to compare Resources in the scheduler.
+      The default i.e. org.apache.hadoop.yarn.util.resource.DefaultResourseCalculator only
uses
+      Memory while DominantResourceCalculator uses Dominant-resource to compare multi-dimensional
+      resources such as Memory, CPU etc. A Java ResourceCalculator class name is expected.
+    </description>
     <display-name>CPU Scheduling</display-name>
     <value-attributes>
       <type>value-list</type>

http://git-wip-us.apache.org/repos/asf/ambari/blob/77c2dc6e/ambari-server/src/main/resources/stacks/HDPWIN/2.2/services/YARN/configuration/yarn-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDPWIN/2.2/services/YARN/configuration/yarn-site.xml
b/ambari-server/src/main/resources/stacks/HDPWIN/2.2/services/YARN/configuration/yarn-site.xml
index 62c5693..8b3db83 100644
--- a/ambari-server/src/main/resources/stacks/HDPWIN/2.2/services/YARN/configuration/yarn-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDPWIN/2.2/services/YARN/configuration/yarn-site.xml
@@ -210,7 +210,15 @@
   <property>
     <name>yarn.nodemanager.resource.cpu-vcores</name>
     <value>8</value>
-    <description/>
+    <description>
+      Number of vcores that can be allocated
+      for containers. This is used by the RM scheduler when allocating
+      resources for containers. This is not used to limit the number of
+      CPUs used by YARN containers. If it is set to -1 and
+      yarn.nodemanager.resource.detect-hardware-capabilities is true, it is
+      automatically determined from the hardware in case of Windows and Linux.
+      In other cases, number of vcores is 8 by default.
+    </description>
     <display-name>Number of virtual cores</display-name>
     <value-attributes>
       <type>int</type>
@@ -247,7 +255,11 @@
   <property>
     <name>yarn.node-labels.fs-store.retry-policy-spec</name>
     <value>2000, 500</value>
-    <description/>
+    <description>
+      Retry policy used for FileSystem node label store. The policy is
+      specified by N pairs of sleep-time in milliseconds and number-of-retries
+      &quot;s1,n1,s2,n2,...&quot;.
+    </description>
     <on-ambari-upgrade add="true"/>
   </property>
   <property>
@@ -292,7 +304,7 @@
   <property>
     <name>yarn.resourcemanager.system-metrics-publisher.dispatcher.pool-size</name>
     <value>10</value>
-    <description/>
+    <description>Number of worker threads that send the yarn system metrics data.</description>
     <on-ambari-upgrade add="true"/>
   </property>
   <property>
@@ -392,7 +404,9 @@
   <property>
     <name>yarn.node-labels.fs-store.root-dir</name>
     <value>/system/yarn/node-labels</value>
-    <description/>
+    <description>
+      URI for NodeLabelManager.
+    </description>
     <on-ambari-upgrade add="true"/>
   </property>
   <property>
@@ -458,6 +472,11 @@
   </property>
   <property>
     <name>yarn.resourcemanager.scheduler.monitor.enable</name>
+    <description>
+      Enable a set of periodic monitors (specified in
+      yarn.resourcemanager.scheduler.monitor.policies) that affect the
+      scheduler.
+    </description>
     <value>false</value>
     <display-name>Pre-emption</display-name>
     <value-attributes>

http://git-wip-us.apache.org/repos/asf/ambari/blob/77c2dc6e/ambari-server/src/main/resources/stacks/HDPWIN/2.3/services/HDFS/configuration/hdfs-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDPWIN/2.3/services/HDFS/configuration/hdfs-site.xml
b/ambari-server/src/main/resources/stacks/HDPWIN/2.3/services/HDFS/configuration/hdfs-site.xml
index 41590b6..d7348b0 100644
--- a/ambari-server/src/main/resources/stacks/HDPWIN/2.3/services/HDFS/configuration/hdfs-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDPWIN/2.3/services/HDFS/configuration/hdfs-site.xml
@@ -32,6 +32,14 @@
   </property>
   <property>
     <name>nfs.exports.allowed.hosts</name>
+    <description>
+      By default, the export can be mounted by any client. To better control the access,
+      users can update the following property. The value string contains machine name and
access privilege,
+      separated by whitespace characters. Machine name format can be single host, wildcards,
and IPv4
+      networks.The access privilege uses rw or ro to specify readwrite or readonly access
of the machines
+      to exports. If the access privilege is not provided, the default is read-only. Entries
are separated
+      by &quot;;&quot;. For example: &quot;192.168.0.0/22 rw ; host*.example.com
; host1.test.org ro;&quot;.
+    </description>
     <value>* rw</value>
     <display-name>Allowed hosts</display-name>
     <on-ambari-upgrade add="true"/>

http://git-wip-us.apache.org/repos/asf/ambari/blob/77c2dc6e/ambari-server/src/main/resources/stacks/PHD/3.0/services/HDFS/configuration/core-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/PHD/3.0/services/HDFS/configuration/core-site.xml
b/ambari-server/src/main/resources/stacks/PHD/3.0/services/HDFS/configuration/core-site.xml
index ab67a5b..4f18611 100644
--- a/ambari-server/src/main/resources/stacks/PHD/3.0/services/HDFS/configuration/core-site.xml
+++ b/ambari-server/src/main/resources/stacks/PHD/3.0/services/HDFS/configuration/core-site.xml
@@ -21,7 +21,9 @@
   <property>
     <name>hadoop.http.authentication.simple.anonymous.allowed</name>
     <value>true</value>
-    <description/>
+    <description>
+      Indicates if anonymous requests are allowed when using &apos;simple&apos; authentication.
+    </description>
     <on-ambari-upgrade add="true"/>
   </property>
   <property>

http://git-wip-us.apache.org/repos/asf/ambari/blob/77c2dc6e/ambari-server/src/main/resources/stacks/PHD/3.0/services/HDFS/configuration/hdfs-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/PHD/3.0/services/HDFS/configuration/hdfs-site.xml
b/ambari-server/src/main/resources/stacks/PHD/3.0/services/HDFS/configuration/hdfs-site.xml
index df2c591..0fed58d 100644
--- a/ambari-server/src/main/resources/stacks/PHD/3.0/services/HDFS/configuration/hdfs-site.xml
+++ b/ambari-server/src/main/resources/stacks/PHD/3.0/services/HDFS/configuration/hdfs-site.xml
@@ -21,7 +21,14 @@
   <property>
     <name>dfs.namenode.startup.delay.block.deletion.sec</name>
     <value>3600</value>
-    <description/>
+    <description>
+      The delay in seconds at which we will pause the blocks deletion
+      after Namenode startup. By default it's disabled.
+      In the case a directory has large number of directories and files are
+      deleted, suggested delay is one hour to give the administrator enough time
+      to notice large number of pending deletion blocks and take corrective
+      action.
+    </description>
     <on-ambari-upgrade add="true"/>
   </property>
   <property>


Mime
View raw message