Return-Path: X-Original-To: archive-asf-public-internal@cust-asf2.ponee.io Delivered-To: archive-asf-public-internal@cust-asf2.ponee.io Received: from cust-asf.ponee.io (cust-asf.ponee.io [163.172.22.183]) by cust-asf2.ponee.io (Postfix) with ESMTP id B24DF200ACA for ; Thu, 9 Jun 2016 17:01:31 +0200 (CEST) Received: by cust-asf.ponee.io (Postfix) id B1155160A63; Thu, 9 Jun 2016 15:01:31 +0000 (UTC) Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by cust-asf.ponee.io (Postfix) with SMTP id 6828C160A65 for ; Thu, 9 Jun 2016 17:01:29 +0200 (CEST) Received: (qmail 49296 invoked by uid 500); 9 Jun 2016 15:01:27 -0000 Mailing-List: contact commits-help@ambari.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: ambari-dev@ambari.apache.org Delivered-To: mailing list commits@ambari.apache.org Received: (qmail 48330 invoked by uid 99); 9 Jun 2016 15:01:27 -0000 Received: from git1-us-west.apache.org (HELO git1-us-west.apache.org) (140.211.11.23) by apache.org (qpsmtpd/0.29) with ESMTP; Thu, 09 Jun 2016 15:01:27 +0000 Received: by git1-us-west.apache.org (ASF Mail Server at git1-us-west.apache.org, from userid 33) id 286FAE0991; Thu, 9 Jun 2016 15:01:27 +0000 (UTC) Content-Type: text/plain; charset="us-ascii" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit From: dmitriusan@apache.org To: commits@ambari.apache.org Date: Thu, 09 Jun 2016 15:01:43 -0000 Message-Id: In-Reply-To: <1f690137a0d04f40a98aa60925acb588@git.apache.org> References: <1f690137a0d04f40a98aa60925acb588@git.apache.org> X-Mailer: ASF-Git Admin Mailer Subject: [18/70] ambari git commit: AMBARI-17112. Fixed implementation of on-ambari-upgrade support. Update all stack configuration xmls to pass validation. Change defaults (dlysnichenko) archived-at: Thu, 09 Jun 2016 15:01:31 -0000 http://git-wip-us.apache.org/repos/asf/ambari/blob/fa1b536d/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/HIVE/configuration/hive-site.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/HIVE/configuration/hive-site.xml b/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/HIVE/configuration/hive-site.xml index 1ee7330..0ad559c 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/HIVE/configuration/hive-site.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/HIVE/configuration/hive-site.xml @@ -21,7 +21,7 @@ limitations under the License. hive.heapsize 1024 Hive Java heap size - + ambari.hive.db.schema.name @@ -30,25 +30,25 @@ limitations under the License. database - + javax.jdo.option.ConnectionURL jdbc:mysql://localhost/hive?createDatabaseIfNotExist=true JDBC connect string for a JDBC metastore - + javax.jdo.option.ConnectionDriverName com.mysql.jdbc.Driver Driver class name for a JDBC metastore - + javax.jdo.option.ConnectionUserName hive username to use against metastore database - + javax.jdo.option.ConnectionPassword @@ -58,38 +58,38 @@ limitations under the License. password - + hive.metastore.warehouse.dir /apps/hive/warehouse location of default database for the warehouse - + hive.metastore.sasl.enabled false If true, the metastore thrift interface will be secured with SASL. Clients must authenticate with Kerberos. - + hive.warehouse.subdir.inherit.perms true If true, subdirectories inherit the same permissions - + hive.metastore.cache.pinobjtypes Table,Database,Type,FieldSchema,Order List of comma separated metastore object types that should be pinned in the cache - + hive.metastore.uris thrift://localhost:9083 URI for client to contact metastore server - + hive.metastore.pre.event.listeners @@ -98,7 +98,7 @@ limitations under the License. whenever databases, tables, and partitions are created, altered, or dropped. Set to org.apache.hadoop.hive.ql.security.authorization.AuthorizationPreEventListener if metastore-side authorization is desired. - + hive.metastore.pre.event.listeners @@ -107,32 +107,32 @@ limitations under the License. whenever databases, tables, and partitions are created, altered, or dropped. Set to org.apache.hadoop.hive.ql.security.authorization.AuthorizationPreEventListener if metastore-side authorization is desired. - + hive.metastore.client.socket.timeout 60 MetaStore Client socket timeout in seconds - + hive.metastore.execute.setugi true In unsecure mode, setting this property to true will cause the metastore to execute DFS operations using the client's reported user and group permissions. Note that this property must be set on both the client and server sides. Further note that its best effort. If client sets its to true and server sets it to false, client setting will be ignored. - + hive.security.authorization.enabled false enable or disable the hive client authorization - + hive.security.authorization.manager org.apache.hadoop.hive.ql.security.authorization.StorageBasedAuthorizationProvider the hive client authorization manager class name. The user defined authorization class should implement interface org.apache.hadoop.hive.ql.security.authorization.HiveAuthorizationProvider. - + hive.security.metastore.authorization.manager @@ -142,13 +142,13 @@ limitations under the License. string - + hive.security.authenticator.manager org.apache.hadoop.hive.ql.security.ProxyUserAuthenticator Hive client authenticator manager class name. The user-defined authenticator class should implement interface org.apache.hadoop.hive.ql.security.HiveAuthenticationProvider. - + hive.server2.enable.doAs @@ -157,49 +157,49 @@ limitations under the License. submitted the query. But if the parameter is set to false, the query will run as the user that the hiveserver2 process runs as. - + hive.server2.enable.impersonation Enable user impersonation for HiveServer2 true - + hive.server2.authentication Authentication mode, default NONE. Options are NONE, NOSASL, KERBEROS, LDAP, PAM and CUSTOM NONE - + fs.hdfs.impl.disable.cache true Disable HDFS filesystem cache. - + fs.file.impl.disable.cache true Disable local filesystem cache. - + hive.enforce.bucketing true Whether bucketing is enforced. If true, while inserting into the table, bucketing is enforced. - + hive.enforce.sorting true Whether sorting is enforced. If true, while inserting into the table, sorting is enforced. - + hive.map.aggr true Whether to use map-side aggregation in Hive Group By queries. - + hive.optimize.bucketmapjoin @@ -208,7 +208,7 @@ limitations under the License. is a multiple of the number of buckets in the other table, the buckets can be joined with each other by setting this parameter as true. - + hive.optimize.bucketmapjoin.sortedmerge @@ -216,20 +216,20 @@ limitations under the License. If the tables being joined are sorted and bucketized on the join columns, and they have the same number of buckets, a sort-merge join can be performed by setting this parameter as true. - + hive.mapred.reduce.tasks.speculative.execution false Whether speculative execution for reducers should be turned on. - + hive.auto.convert.join true Whether Hive enable the optimization about converting common join into mapjoin based on the input file size. - + hive.auto.convert.sortmerge.join @@ -237,13 +237,13 @@ limitations under the License. Will the join be automatically converted to a sort-merge join, if the joined tables pass the criteria for sort-merge join. - + hive.auto.convert.sortmerge.join.noconditionaltask true Required to Enable the conversion of an SMB (Sort-Merge-Bucket) to a map-join SMB. - + hive.auto.convert.join.noconditionaltask @@ -252,7 +252,7 @@ limitations under the License. size. If this paramater is on, and the sum of size for n-1 of the tables/partitions for a n-way join is smaller than the specified size, the join is directly converted to a mapjoin (there is no conditional task). - + hive.auto.convert.join.noconditionaltask.size @@ -269,7 +269,7 @@ limitations under the License. B - + hive.optimize.reducededuplication.min.reducer @@ -278,7 +278,7 @@ limitations under the License. That means if reducer-num of the child RS is fixed (order by or forced bucketing) and small, it can make very slow, single MR. The optimization will be disabled if number of reducers is less than specified value. - + hive.optimize.mapjoin.mapreduce @@ -288,7 +288,7 @@ limitations under the License. job (for e.g a group by), each map-only job is merged with the following map-reduce job. - + hive.mapjoin.bucket.cache.size @@ -297,21 +297,21 @@ limitations under the License. Size per reducer.The default is 1G, i.e if the input size is 10G, it will use 10 reducers. - + hive.vectorized.execution.enabled true This flag controls the vectorized mode of query execution as documented in HIVE-4160 (as of Hive 0.13.0) - + hive.optimize.reducededuplication true Remove extra map-reduce jobs if the data is already clustered by the same key which needs to be used again. - + hive.optimize.index.filter @@ -319,7 +319,7 @@ limitations under the License. Whether to enable automatic use of indexes - + hive.server2.thrift.port @@ -327,7 +327,7 @@ limitations under the License. TCP port number to listen on, default 10000. - + hive.server2.support.dynamic.service.discovery @@ -338,7 +338,7 @@ limitations under the License. should use the ZooKeeper ensemble: hive.zookeeper.quorum in their connection string. - + hive.server2.zookeeper.namespace @@ -346,7 +346,7 @@ limitations under the License. The parent node in ZooKeeper used by HiveServer2 when supporting dynamic service discovery. - + hive.server2.transport.mode @@ -355,7 +355,7 @@ limitations under the License. Expects one of [binary, http]. Transport mode of HiveServer2. - + hive.default.fileformat @@ -383,14 +383,14 @@ limitations under the License. - + hive.server2.authentication Authentication mode, default NONE. Options are NONE, NOSASL, KERBEROS, LDAP, PAM and CUSTOM NONE - + hive.server2.authentication.ldap.url @@ -401,7 +401,7 @@ limitations under the License. hive.server2.authentication - + hive.server2.authentication.ldap.baseDN @@ -412,7 +412,7 @@ limitations under the License. hive.server2.authentication - + hive.server2.authentication.kerberos.keytab @@ -423,7 +423,7 @@ limitations under the License. hive.server2.authentication - + hive.server2.authentication.kerberos.principal @@ -434,7 +434,7 @@ limitations under the License. hive.server2.authentication - + hive.server2.authentication.pam.services @@ -445,7 +445,7 @@ limitations under the License. hive.server2.authentication - + hive.server2.custom.authentication.class @@ -456,6 +456,6 @@ limitations under the License. hive.server2.authentication - + http://git-wip-us.apache.org/repos/asf/ambari/blob/fa1b536d/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/HIVE/configuration/webhcat-site.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/HIVE/configuration/webhcat-site.xml b/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/HIVE/configuration/webhcat-site.xml index f233df3..7d8858d 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/HIVE/configuration/webhcat-site.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/HIVE/configuration/webhcat-site.xml @@ -24,108 +24,108 @@ limitations under the License. templeton.port 50111 The HTTP port for the main server. - + templeton.hadoop.conf.dir /usr/hdp/current/hadoop-client/conf The path to the Hadoop configuration. - + templeton.jar /usr/lib/hcatalog/share/webhcat/svr/webhcat.jar The path to the Templeton jar file. - + templeton.libjars /usr/lib/zookeeper/zookeeper.jar Jars to add the the classpath. - + templeton.hadoop /usr/bin/hadoop The path to the Hadoop executable. - + templeton.pig.archive glusterfs:///apps/webhcat/pig.tar.gz The path to the Pig archive. - + templeton.pig.path pig.tar.gz/pig/bin/pig The path to the Pig executable. - + templeton.hcat /usr/bin/hcat The path to the hcatalog executable. - + templeton.hive.archive glusterfs:///apps/webhcat/hive.tar.gz The path to the Hive archive. - + templeton.hive.home hive.tar.gz/hive The path to the Hive home within the tar. Has no effect if templeton.hive.archive is not set. - + templeton.hcat.home hive.tar.gz/hive/hcatalog The path to the HCat home within the tar. Has no effect if templeton.hive.archive is not set. - + templeton.hive.path hive.tar.gz/hive/bin/hive The path to the Hive executable. - + templeton.hive.properties hive.metastore.local=false,hive.metastore.uris=thrift://localhost:9083,hive.metastore.sasl.enabled=false Properties to set when running hive. - + templeton.zookeeper.hosts localhost:2181 ZooKeeper servers, as comma separated host:port pairs - + templeton.storage.class org.apache.hive.hcatalog.templeton.tool.ZooKeeperStorage The class to use as storage - + templeton.override.enabled false Enable the override path in templeton.override.jars - + templeton.streaming.jar glusterfs:///apps/webhcat/hadoop-streaming.jar The hdfs path to the Hadoop streaming jar file. - + templeton.exec.timeout 60000 Time out for templeton api - + http://git-wip-us.apache.org/repos/asf/ambari/blob/fa1b536d/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/OOZIE/configuration/oozie-site.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/OOZIE/configuration/oozie-site.xml b/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/OOZIE/configuration/oozie-site.xml index 7e59010..58188cf 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/OOZIE/configuration/oozie-site.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/OOZIE/configuration/oozie-site.xml @@ -25,7 +25,7 @@ oozie.base.url http://localhost:11000/oozie Base Oozie URL. - + oozie.system.id @@ -33,7 +33,7 @@ The Oozie system ID. - + oozie.systemmode @@ -41,7 +41,7 @@ System mode for Oozie at startup. - + oozie.service.AuthorizationService.security.enabled @@ -50,7 +50,7 @@ Specifies whether security (user name/admin role) is enabled or not. If disabled any user can manage Oozie system and manage any job. - + oozie.service.PurgeService.older.than @@ -58,7 +58,7 @@ Jobs older than this value, in days, will be purged by the PurgeService. - + oozie.service.PurgeService.purge.interval @@ -66,19 +66,19 @@ Interval at which the purge service will run, in seconds. - + oozie.service.CallableQueueService.queue.size 1000 Max callable queue size - + oozie.service.CallableQueueService.threads 10 Number of threads used for executing callables - + oozie.service.CallableQueueService.callable.concurrency @@ -90,7 +90,7 @@ All commands that use action executors (action-start, action-end, action-kill and action-check) use the action type as the callable type. - + oozie.service.coord.normal.default.timeout @@ -98,13 +98,13 @@ Default timeout for a coordinator action input check (in minutes) for normal job. -1 means infinite timeout - + oozie.service.HadoopAccessorService.supported.filesystems glusterfs List of other HCFS supported filesystems - + oozie.db.schema.name @@ -115,7 +115,7 @@ database - + oozie.authentication.type @@ -124,7 +124,7 @@ Authentication used for Oozie HTTP endpoint, the supported values are: simple | kerberos | #AUTHENTICATION_HANDLER_CLASSNAME#. - + oozie.service.WorkflowAppService.system.libpath @@ -134,7 +134,7 @@ This path is added to workflow application if their job properties sets the property 'oozie.use.system.libpath' to true. - + use.system.libpath.for.mapreduce.and.pig.jobs @@ -145,7 +145,7 @@ specify where the Pig JAR files are. Instead, the ones from the system library path are used. - + oozie.authentication.kerberos.name.rules @@ -157,7 +157,7 @@ The mapping from kerberos principal names to local OS user names. - + oozie.service.HadoopAccessorService.hadoop.configurations @@ -170,7 +170,7 @@ the Oozie configuration directory; though the path can be absolute (i.e. to point to Hadoop client conf/ directories in the local filesystem. - + oozie.service.ActionService.executor.ext.classes @@ -186,7 +186,7 @@ be used in workflows. This property is a convenience property to add extensions to the built in executors without having to include all the built in ones. - + oozie.service.SchemaService.wf.ext.schemas @@ -195,7 +195,7 @@ Schemas for additional actions types. IMPORTANT: if there are no schemas leave a 1 space string, the service trims the value, if empty Configuration assumes it is NULL. - + oozie.service.JPAService.create.db.schema @@ -206,7 +206,7 @@ If set to true, it creates the DB schema if it does not exist. If the DB schema exists is a NOP. If set to false, it does not create the DB schema. If the DB schema does not exist it fails start up. - + oozie.service.JPAService.jdbc.driver @@ -214,7 +214,7 @@ JDBC driver class. - + oozie.service.JPAService.jdbc.url @@ -222,7 +222,7 @@ JDBC URL. - + oozie.service.JPAService.jdbc.username @@ -230,7 +230,7 @@ Database user name to use to connect to the database - + oozie.service.JPAService.jdbc.password @@ -247,7 +247,7 @@ password false - + oozie.service.JPAService.pool.max.active.conn @@ -255,7 +255,7 @@ Max number of connections. - + oozie.services @@ -293,7 +293,7 @@ org.apache.oozie.service.ProxyUserService List of Oozie services - + oozie.service.HadoopAccessorService.kerberos.enabled @@ -301,7 +301,7 @@ Indicates if Oozie is configured to use Kerberos. - + oozie.service.URIHandlerService.uri.handlers @@ -309,7 +309,7 @@ Enlist the different uri handlers supported for data availability checks. - + oozie.services.ext @@ -319,7 +319,7 @@ To add/replace services defined in 'oozie.services' with custom implementations. Class names must be separated by commas. - + oozie.service.coord.push.check.requeue.interval @@ -327,7 +327,7 @@ Command re-queue interval for push dependencies (in millisecond). - + oozie.credentials.credentialclasses @@ -335,6 +335,6 @@ Credential Class to be used for HCat. - + http://git-wip-us.apache.org/repos/asf/ambari/blob/fa1b536d/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/TEZ/configuration/tez-site.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/TEZ/configuration/tez-site.xml b/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/TEZ/configuration/tez-site.xml index 1332ded..b5161ac 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/TEZ/configuration/tez-site.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/TEZ/configuration/tez-site.xml @@ -23,7 +23,7 @@ AM view ACLs. This allows the specified users/groups to view the status of the AM and all DAGs that run within this AM. Value format: Comma separated list of users, followed by whitespace, followed by a comma separated list of groups. - + tez.runtime.optimize.local.fetch @@ -32,7 +32,7 @@ boolean - + tez.task.generate.counters.per.io @@ -41,7 +41,7 @@ boolean - + tez.runtime.sorter.class @@ -61,7 +61,7 @@ 1 - + tez.runtime.pipelined.sorter.sort.threads @@ -76,7 +76,7 @@ tez.runtime.sorter.class - + tez.runtime.io.sort.mb @@ -92,6 +92,6 @@ tez.runtime.sorter.class - + http://git-wip-us.apache.org/repos/asf/ambari/blob/fa1b536d/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/YARN/configuration-mapred/mapred-site.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/YARN/configuration-mapred/mapred-site.xml b/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/YARN/configuration-mapred/mapred-site.xml index e8a4a55..f098328 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/YARN/configuration-mapred/mapred-site.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/YARN/configuration-mapred/mapred-site.xml @@ -25,7 +25,7 @@ Directory where history files are written by MapReduce jobs. - + mapreduce.jobhistory.done-dir @@ -33,7 +33,7 @@ Directory where history files are managed by the MR JobHistory Server. - + yarn.app.mapreduce.am.staging-dir @@ -41,46 +41,46 @@ The staging dir used while submitting jobs. - + mapred.healthChecker.script.path glusterfs:///mapred/jobstatus - + mapred.job.tracker.history.completed.location glusterfs:///mapred/history/done - + mapred.system.dir glusterfs:///mapred/system - + mapreduce.jobtracker.staging.root.dir glusterfs:///user - + mapred.healthChecker.script.path glusterfs:///mapred/jobstatus - + mapred.job.tracker.history.completed.location glusterfs:///mapred/history/done - + mapred.system.dir glusterfs:///mapred/system - + mapreduce.jobtracker.staging.root.dir glusterfs:///user - + http://git-wip-us.apache.org/repos/asf/ambari/blob/fa1b536d/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/YARN/configuration/capacity-scheduler.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/YARN/configuration/capacity-scheduler.xml b/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/YARN/configuration/capacity-scheduler.xml index 1975b79..2ce0d25 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/YARN/configuration/capacity-scheduler.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/YARN/configuration/capacity-scheduler.xml @@ -20,18 +20,18 @@ yarn.scheduler.capacity.resource-calculator org.apache.hadoop.yarn.util.resource.DefaultResourceCalculator - + yarn.scheduler.capacity.root.accessible-node-labels * - + yarn.scheduler.capacity.root.default-node-label-expression - + http://git-wip-us.apache.org/repos/asf/ambari/blob/fa1b536d/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/YARN/configuration/yarn-site.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/YARN/configuration/yarn-site.xml b/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/YARN/configuration/yarn-site.xml index 43e2473..0eb515c 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/YARN/configuration/yarn-site.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.3.GlusterFS/services/YARN/configuration/yarn-site.xml @@ -22,7 +22,7 @@ yarn.application.classpath $HADOOP_CONF_DIR,/usr/hdp/current/hadoop-client/*,/usr/hdp/current/hadoop-client/lib/*,/usr/hdp/current/hadoop-hdfs-client/*,/usr/hdp/current/hadoop-hdfs-client/lib/*,/usr/hdp/current/hadoop-yarn-client/*,/usr/hdp/current/hadoop-yarn-client/lib/* Classpath for typical applications. - + hadoop.registry.rm.enabled @@ -30,7 +30,7 @@ Is the registry enabled: does the RM start it up, create the user and system paths, and purge service records when containers, application attempts and applications complete - + hadoop.registry.zk.quorum @@ -38,13 +38,13 @@ List of hostname:port pairs defining the zookeeper quorum binding for the registry - + yarn.nodemanager.recovery.enabled true Enable the node manager to recover after starting - + yarn.nodemanager.recovery.dir @@ -53,19 +53,19 @@ The local filesystem directory in which the node manager will store state when recovery is enabled. - + yarn.client.nodemanager-connect.retry-interval-ms 10000 Time interval between each attempt to connect to NM - + yarn.client.nodemanager-connect.max-wait-ms 60000 Max time to wait to establish a connection to NM - + yarn.resourcemanager.recovery.enabled @@ -74,7 +74,7 @@ Enable RM to recover state after starting. If true, then yarn.resourcemanager.store.class must be specified. - + yarn.resourcemanager.work-preserving-recovery.enabled @@ -82,7 +82,7 @@ Enable RM work preserving recovery. This configuration is private to YARN for experimenting the feature. - + yarn.resourcemanager.store.class @@ -93,7 +93,7 @@ the store is implicitly fenced; meaning a single ResourceManager is able to use the store at any point in time. - + yarn.resourcemanager.zk-address @@ -101,37 +101,37 @@ List Host:Port of the ZooKeeper servers to be used by the RM. comma separated host:port pairs, each corresponding to a zk server. e.g. "127.0.0.1:3000,127.0.0.1:3001,127.0.0.1:3002" If the optional chroot suffix is used the example would look like: "127.0.0.1:3000,127.0.0.1:3001,127.0.0.1:3002/app/a" where the client would be rooted at "/app/a" and all paths would be relative to this root - ie getting/setting/etc... "/foo/bar" would result in operations being run on "/app/a/foo/bar" (from the server perspective). - + yarn.resourcemanager.zk-state-store.parent-path /rmstore Full path of the ZooKeeper znode where RM state will be stored. This must be supplied when using org.apache.hadoop.yarn.server.resourcemanager.recovery.ZKRMStateStore as the value for yarn.resourcemanager.store.class - + yarn.resourcemanager.zk-acl world:anyone:rwcda ACL's to be used for ZooKeeper znodes. - + yarn.resourcemanager.work-preserving-recovery.scheduling-wait-ms 10000 Set the amount of time RM waits before allocating new containers on work-preserving-recovery. Such wait period gives RM a chance to settle down resyncing with NMs in the cluster on recovery, before assigning new containers to applications. - + yarn.resourcemanager.connect.retry-interval.ms 30000 How often to try connecting to the ResourceManager. - + yarn.resourcemanager.connect.max-wait.ms 900000 Maximum time to wait to establish connection to ResourceManager - + yarn.resourcemanager.zk-retry-interval-ms @@ -141,67 +141,67 @@ automatically from yarn.resourcemanager.zk-timeout-ms and yarn.resourcemanager.zk-num-retries." - + yarn.resourcemanager.zk-num-retries 1000 Number of times RM tries to connect to ZooKeeper. - + yarn.resourcemanager.zk-timeout-ms 10000 ZooKeeper session timeout in milliseconds. Session expiration is managed by the ZooKeeper cluster itself, not by the client. This value is used by the cluster to determine when the client's session expires. Expirations happens when the cluster does not hear from the client within the specified session timeout period (i.e. no heartbeat). - + yarn.resourcemanager.state-store.max-completed-applications ${yarn.resourcemanager.max-completed-applications} The maximum number of completed applications RM state store keeps, less than or equals to ${yarn.resourcemanager.max-completed-applications}. By default, it equals to ${yarn.resourcemanager.max-completed-applications}. This ensures that the applications kept in the state store are consistent with the applications remembered in RM memory. Any values larger than ${yarn.resourcemanager.max-completed-applications} will be reset to ${yarn.resourcemanager.max-completed-applications}. Note that this value impacts the RM recovery performance.Typically, a smaller value indicates better performance on RM recovery. - + yarn.resourcemanager.fs.state-store.retry-policy-spec 2000, 500 hdfs client retry policy specification. hdfs client retry is always enabled. Specified in pairs of sleep-time and number-of-retries and (t0, n0), (t1, n1), ..., the first n0 retries sleep t0 milliseconds on average, the following n1 retries sleep t1 milliseconds on average, and so on. - + yarn.resourcemanager.fs.state-store.uri RI pointing to the location of the FileSystem path where RM state will be stored. This must be supplied when using org.apache.hadoop.yarn.server.resourcemanager.recovery.FileSystemRMStateStore as the value for yarn.resourcemanager.store.class - + yarn.resourcemanager.ha.enabled false enable RM HA or not - + yarn.nodemanager.linux-container-executor.resources-handler.class org.apache.hadoop.yarn.server.nodemanager.util.DefaultLCEResourcesHandler Pre-requisite to use CGroups - + yarn.nodemanager.linux-container-executor.cgroups.hierarchy hadoop-yarn Name of the Cgroups hierarchy under which all YARN jobs will be launched - + yarn.nodemanager.linux-container-executor.cgroups.mount false If true, YARN will automount the CGroup, however the directory needs to already exist; else, the cgroup should be mounted by the admin - + yarn.nodemanager.linux-container-executor.cgroups.strict-resource-usage false Strictly limit CPU resource usage to allocated usage even if spare CPU is available - + yarn.nodemanager.resource.cpu-vcores @@ -213,7 +213,7 @@ 0 32 - + yarn.nodemanager.resource.percentage-physical-cpu-limit @@ -225,37 +225,37 @@ 0 100 - + yarn.node-labels.manager-class org.apache.hadoop.yarn.server.resourcemanager.nodelabels.MemoryRMNodeLabelsManager If user want to enable this feature, specify it to "org.apache.hadoop.yarn.server.resourcemanager.nodelabels.RMNodeLabelsManager - + yarn.node-labels.fs-store.retry-policy-spec 2000, 500 - + yarn.nodemanager.disk-health-checker.min-free-space-per-disk-mb 1000 This is related to disk size on the machines, admins should set one of yarn.nodemanager.disk-health-checker.min-free-space-per-disk-mb or yarn.nodemanager.disk-health-checker.max-disk-utilization-per-disk-percentage but not both. If both are set, the more conservative value will be used - + yarn.nodemanager.disk-health-checker.max-disk-utilization-per-disk-percentage 90 This is related to disk size on the machines, admins should set one of yarn.nodemanager.disk-health-checker.min-free-space-per-disk-mb or yarn.nodemanager.disk-health-checker.max-disk-utilization-per-disk-percentage but not both. If both are set, the more conservative value will be used - + yarn.nodemanager.log-aggregation.roll-monitoring-interval-seconds -1 Defines how often NMs wake up to upload log files. The default value is -1. By default, the logs will be uploaded whenthe application is finished. By setting this configure, logs can be uploaded periodically when the application is running. The minimum rolling-interval-seconds can be set is 3600. - + yarn.nodemanager.log-aggregation.debug-enabled @@ -264,37 +264,37 @@ This configuration is for debug and test purpose. By setting this configuration as true. We can break the lower bound of yarn.nodemanager.log-aggregation.roll-monitoring-interval-seconds - + yarn.nodemanager.log-aggregation.num-log-files-per-app 30 This is temporary solution. The configuration will be deleted once, we find a more scalable method to only write a single log file per LRS. - + yarn.resourcemanager.system-metrics-publisher.enabled true - + yarn.resourcemanager.system-metrics-publisher.dispatcher.pool-size 10 - + yarn.timeline-service.client.max-retries 30 - + yarn.timeline-service.client.retry-interval-ms 1000 - + yarn.timeline-service.ttl-enable @@ -302,13 +302,13 @@ Enable age off of timeline store data. - + yarn.timeline-service.leveldb-timeline-store.path /hadoop/yarn/timeline Store file name for leveldb timeline store. - + yarn.timeline-service.leveldb-timeline-store.read-cache-size @@ -316,7 +316,7 @@ Size of read cache for uncompressed blocks for leveldb timeline store in bytes. - + yarn.timeline-service.leveldb-timeline-store.start-time-read-cache-size @@ -324,7 +324,7 @@ Size of cache for recently read entity start times for leveldb timeline store in number of entities. - + yarn.timeline-service.leveldb-timeline-store.start-time-write-cache-size @@ -332,7 +332,7 @@ Size of cache for recently written entity start times for leveldb timeline store in number of entities. - + yarn.timeline-service.http-authentication.type @@ -341,13 +341,13 @@ Defines authentication used for the Timeline Server HTTP endpoint. Supported values are: simple | kerberos | $AUTHENTICATION_HANDLER_CLASSNAME - + yarn.timeline-service.http-authentication.simple.anonymous.allowed true - + yarn.resourcemanager.webapp.delegation-token-auth-filter.enabled @@ -358,31 +358,31 @@ tokens(fallback to kerberos if the tokens are missing). Only applicable when the http authentication type is kerberos. - + yarn.resourcemanager.bind-host 0.0.0.0 Default value is 0.0.0.0, when this is set the service will bind on all interfaces. I think these two options (blank, "0.0.0.0" sans quotes) should be the two available values, with blank as the default. - + yarn.nodemanager.bind-host 0.0.0.0 Default value is 0.0.0.0, when this is set the service will bind on all interfaces. I think these two options (blank, "0.0.0.0" sans quotes) should be the two available values, with blank as the default. - + yarn.timeline-service.bind-host 0.0.0.0 Default value is 0.0.0.0, when this is set the service will bind on all interfaces. I think these two options (blank, "0.0.0.0" sans quotes) should be the two available values, with blank as the default. - + yarn.node-labels.fs-store.root-dir /system/yarn/node-labels - + yarn.scheduler.minimum-allocation-vcores @@ -400,7 +400,7 @@ yarn.nodemanager.resource.cpu-vcores - + yarn.scheduler.maximum-allocation-vcores @@ -418,6 +418,6 @@ yarn.nodemanager.resource.cpu-vcores - + http://git-wip-us.apache.org/repos/asf/ambari/blob/fa1b536d/ambari-server/src/main/resources/stacks/HDP/2.3/services/ACCUMULO/configuration/accumulo-env.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.3/services/ACCUMULO/configuration/accumulo-env.xml b/ambari-server/src/main/resources/stacks/HDP/2.3/services/ACCUMULO/configuration/accumulo-env.xml index 46dae90..d9f3b7c 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.3/services/ACCUMULO/configuration/accumulo-env.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.3/services/ACCUMULO/configuration/accumulo-env.xml @@ -29,6 +29,6 @@ true false - + http://git-wip-us.apache.org/repos/asf/ambari/blob/fa1b536d/ambari-server/src/main/resources/stacks/HDP/2.3/services/ACCUMULO/configuration/accumulo-log4j.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.3/services/ACCUMULO/configuration/accumulo-log4j.xml b/ambari-server/src/main/resources/stacks/HDP/2.3/services/ACCUMULO/configuration/accumulo-log4j.xml index 5db38c3..4f25f31 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.3/services/ACCUMULO/configuration/accumulo-log4j.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.3/services/ACCUMULO/configuration/accumulo-log4j.xml @@ -24,38 +24,38 @@ audit_log_level OFF Log level for audit logging - + monitor_forwarding_log_level WARN Log level for logging forwarded to the Accumulo Monitor - + debug_log_size 512M Size of each debug rolling log file - + debug_num_logs 10 Number of rolling debug log files to keep - + info_log_size 512M Size of each info rolling log file - + info_num_logs 10 Number of rolling info log files to keep - + content @@ -110,6 +110,6 @@ log4j.appender.A1.layout=org.apache.log4j.PatternLayout content false - + http://git-wip-us.apache.org/repos/asf/ambari/blob/fa1b536d/ambari-server/src/main/resources/stacks/HDP/2.3/services/ACCUMULO/configuration/accumulo-site.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.3/services/ACCUMULO/configuration/accumulo-site.xml b/ambari-server/src/main/resources/stacks/HDP/2.3/services/ACCUMULO/configuration/accumulo-site.xml index a035100..5a89d29 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.3/services/ACCUMULO/configuration/accumulo-site.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.3/services/ACCUMULO/configuration/accumulo-site.xml @@ -44,6 +44,6 @@ $HADOOP_CONF_DIR, content - + http://git-wip-us.apache.org/repos/asf/ambari/blob/fa1b536d/ambari-server/src/main/resources/stacks/HDP/2.3/services/FALCON/configuration/falcon-env.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.3/services/FALCON/configuration/falcon-env.xml b/ambari-server/src/main/resources/stacks/HDP/2.3/services/FALCON/configuration/falcon-env.xml index 8592bdc..6c0d84d 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.3/services/FALCON/configuration/falcon-env.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.3/services/FALCON/configuration/falcon-env.xml @@ -24,6 +24,6 @@ supports_hive_dr true Enable support of Falcon Hive DR - + http://git-wip-us.apache.org/repos/asf/ambari/blob/fa1b536d/ambari-server/src/main/resources/stacks/HDP/2.3/services/FALCON/configuration/falcon-startup.properties.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.3/services/FALCON/configuration/falcon-startup.properties.xml b/ambari-server/src/main/resources/stacks/HDP/2.3/services/FALCON/configuration/falcon-startup.properties.xml index 00a6855..ae4ecfd 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.3/services/FALCON/configuration/falcon-startup.properties.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.3/services/FALCON/configuration/falcon-startup.properties.xml @@ -24,6 +24,6 @@ *.shared.libs activemq-core,ant,geronimo-j2ee-management,jms,json-simple,oozie-client,spring-jms,commons-lang3,commons-el - + http://git-wip-us.apache.org/repos/asf/ambari/blob/fa1b536d/ambari-server/src/main/resources/stacks/HDP/2.3/services/HBASE/configuration/hbase-env.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.3/services/HBASE/configuration/hbase-env.xml b/ambari-server/src/main/resources/stacks/HDP/2.3/services/HBASE/configuration/hbase-env.xml index 019c973..b1f8491 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.3/services/HBASE/configuration/hbase-env.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.3/services/HBASE/configuration/hbase-env.xml @@ -105,6 +105,6 @@ export HBASE_REGIONSERVER_OPTS="$HBASE_REGIONSERVER_OPTS {% if hbase_max_direct_ content - + http://git-wip-us.apache.org/repos/asf/ambari/blob/fa1b536d/ambari-server/src/main/resources/stacks/HDP/2.3/services/HBASE/configuration/hbase-site.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.3/services/HBASE/configuration/hbase-site.xml b/ambari-server/src/main/resources/stacks/HDP/2.3/services/HBASE/configuration/hbase-site.xml index 49b7382..1016af5 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.3/services/HBASE/configuration/hbase-site.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.3/services/HBASE/configuration/hbase-site.xml @@ -29,36 +29,36 @@ false int - + hbase.master.info.port 16010 The port for the HBase Master web UI. - + hbase.regionserver.port 16020 The port the HBase RegionServer binds to. - + hbase.regionserver.info.port 16030 The port for the HBase RegionServer web UI. - + hbase.regionserver.global.memstore.upperLimit 0.4 true - + hbase.regionserver.global.memstore.lowerLimit 0.38 true - + http://git-wip-us.apache.org/repos/asf/ambari/blob/fa1b536d/ambari-server/src/main/resources/stacks/HDP/2.3/services/HBASE/configuration/ranger-hbase-audit.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.3/services/HBASE/configuration/ranger-hbase-audit.xml b/ambari-server/src/main/resources/stacks/HDP/2.3/services/HBASE/configuration/ranger-hbase-audit.xml index e9a9ff3..f670d7e 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.3/services/HBASE/configuration/ranger-hbase-audit.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.3/services/HBASE/configuration/ranger-hbase-audit.xml @@ -23,7 +23,7 @@ xasecure.audit.is.enabled true Is Audit enabled? - + xasecure.audit.destination.db @@ -39,19 +39,19 @@ xasecure.audit.destination.db - + xasecure.audit.destination.db.jdbc.url {{audit_jdbc_url}} Audit DB JDBC URL - + xasecure.audit.destination.db.user {{xa_audit_db_user}} Audit DB JDBC User - + xasecure.audit.destination.db.password @@ -61,25 +61,25 @@ password - + xasecure.audit.destination.db.jdbc.driver {{jdbc_driver}} Audit DB JDBC Driver - + xasecure.audit.credential.provider.file jceks://file{{credential_file}} Credential file store - + xasecure.audit.destination.db.batch.filespool.dir /var/log/hbase/audit/db/spool /var/log/hbase/audit/db/spool - + xasecure.audit.destination.hdfs @@ -95,7 +95,7 @@ xasecure.audit.destination.hdfs - + xasecure.audit.destination.hdfs.dir @@ -107,13 +107,13 @@ xasecure.audit.destination.hdfs.dir - + xasecure.audit.destination.hdfs.batch.filespool.dir /var/log/hbase/audit/hdfs/spool /var/log/hbase/audit/hdfs/spool - + xasecure.audit.destination.solr @@ -129,7 +129,7 @@ xasecure.audit.destination.solr - + xasecure.audit.destination.solr.urls @@ -144,7 +144,7 @@ ranger.audit.solr.urls - + xasecure.audit.destination.solr.zookeepers @@ -156,13 +156,13 @@ ranger.audit.solr.zookeepers - + xasecure.audit.destination.solr.batch.filespool.dir /var/log/hbase/audit/solr/spool /var/log/hbase/audit/solr/spool - + xasecure.audit.provider.summary.enabled @@ -172,6 +172,6 @@ boolean - + http://git-wip-us.apache.org/repos/asf/ambari/blob/fa1b536d/ambari-server/src/main/resources/stacks/HDP/2.3/services/HBASE/configuration/ranger-hbase-plugin-properties.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.3/services/HBASE/configuration/ranger-hbase-plugin-properties.xml b/ambari-server/src/main/resources/stacks/HDP/2.3/services/HBASE/configuration/ranger-hbase-plugin-properties.xml index 6b2caab..81ca4ca 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.3/services/HBASE/configuration/ranger-hbase-plugin-properties.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.3/services/HBASE/configuration/ranger-hbase-plugin-properties.xml @@ -22,131 +22,131 @@ XAAUDIT.DB.IS_ENABLED true - + XAAUDIT.HDFS.IS_ENABLED true - + XAAUDIT.HDFS.DESTINATION_DIRECTORY true - + XAAUDIT.HDFS.LOCAL_BUFFER_DIRECTORY true - + XAAUDIT.HDFS.LOCAL_ARCHIVE_DIRECTORY true - + XAAUDIT.HDFS.DESTINTATION_FILE true - + XAAUDIT.HDFS.DESTINTATION_FLUSH_INTERVAL_SECONDS true - + XAAUDIT.HDFS.DESTINTATION_ROLLOVER_INTERVAL_SECONDS true - + XAAUDIT.HDFS.DESTINTATION_OPEN_RETRY_INTERVAL_SECONDS true - + XAAUDIT.HDFS.LOCAL_BUFFER_FILE true - + XAAUDIT.HDFS.LOCAL_BUFFER_FLUSH_INTERVAL_SECONDS true - + XAAUDIT.HDFS.LOCAL_BUFFER_ROLLOVER_INTERVAL_SECONDS true - + XAAUDIT.HDFS.LOCAL_ARCHIVE_MAX_FILE_COUNT true - + SSL_KEYSTORE_FILE_PATH true - + SSL_KEYSTORE_PASSWORD true - + SSL_TRUSTSTORE_FILE_PATH true - + SSL_TRUSTSTORE_PASSWORD true - + UPDATE_XAPOLICIES_ON_GRANT_REVOKE true - + POLICY_MGR_URL true - + SQL_CONNECTOR_JAR true - + XAAUDIT.DB.FLAVOUR true - + XAAUDIT.DB.DATABASE_NAME true - + XAAUDIT.DB.USER_NAME true - + XAAUDIT.DB.PASSWORD true - + XAAUDIT.DB.HOSTNAME true - + REPOSITORY_NAME true - + http://git-wip-us.apache.org/repos/asf/ambari/blob/fa1b536d/ambari-server/src/main/resources/stacks/HDP/2.3/services/HBASE/configuration/ranger-hbase-policymgr-ssl.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.3/services/HBASE/configuration/ranger-hbase-policymgr-ssl.xml b/ambari-server/src/main/resources/stacks/HDP/2.3/services/HBASE/configuration/ranger-hbase-policymgr-ssl.xml index 570364a..79370bc 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.3/services/HBASE/configuration/ranger-hbase-policymgr-ssl.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.3/services/HBASE/configuration/ranger-hbase-policymgr-ssl.xml @@ -23,7 +23,7 @@ xasecure.policymgr.clientssl.keystore /usr/hdp/current/hbase-client/conf/ranger-plugin-keystore.jks Java Keystore files - + xasecure.policymgr.clientssl.keystore.password @@ -33,13 +33,13 @@ password - + xasecure.policymgr.clientssl.truststore /usr/hdp/current/hbase-client/conf/ranger-plugin-truststore.jks java truststore file - + xasecure.policymgr.clientssl.truststore.password @@ -49,18 +49,18 @@ password - + xasecure.policymgr.clientssl.keystore.credential.file jceks://file{{credential_file}} java keystore credential file - + xasecure.policymgr.clientssl.truststore.credential.file jceks://file{{credential_file}} java truststore credential file - +