Return-Path: X-Original-To: archive-asf-public-internal@cust-asf2.ponee.io Delivered-To: archive-asf-public-internal@cust-asf2.ponee.io Received: from cust-asf.ponee.io (cust-asf.ponee.io [163.172.22.183]) by cust-asf2.ponee.io (Postfix) with ESMTP id A1117200AF5 for ; Thu, 2 Jun 2016 17:02:38 +0200 (CEST) Received: by cust-asf.ponee.io (Postfix) id 9F69B160A51; Thu, 2 Jun 2016 15:02:38 +0000 (UTC) Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by cust-asf.ponee.io (Postfix) with SMTP id 5A21D160A68 for ; Thu, 2 Jun 2016 17:02:35 +0200 (CEST) Received: (qmail 63035 invoked by uid 500); 2 Jun 2016 15:02:34 -0000 Mailing-List: contact commits-help@ambari.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: ambari-dev@ambari.apache.org Delivered-To: mailing list commits@ambari.apache.org Received: (qmail 61008 invoked by uid 99); 2 Jun 2016 15:02:32 -0000 Received: from git1-us-west.apache.org (HELO git1-us-west.apache.org) (140.211.11.23) by apache.org (qpsmtpd/0.29) with ESMTP; Thu, 02 Jun 2016 15:02:32 +0000 Received: by git1-us-west.apache.org (ASF Mail Server at git1-us-west.apache.org, from userid 33) id 01646E0105; Thu, 2 Jun 2016 15:02:31 +0000 (UTC) Content-Type: text/plain; charset="us-ascii" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit From: dmitriusan@apache.org To: commits@ambari.apache.org Date: Thu, 02 Jun 2016 15:03:02 -0000 Message-Id: In-Reply-To: References: X-Mailer: ASF-Git Admin Mailer Subject: [32/47] ambari git commit: AMBARI-16272. Ambari Upgrade shouldn't automatically add stack configs (dlysnichenko) archived-at: Thu, 02 Jun 2016 15:02:38 -0000 http://git-wip-us.apache.org/repos/asf/ambari/blob/6919aa50/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/YARN/configuration-mapred/mapred-site.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/YARN/configuration-mapred/mapred-site.xml b/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/YARN/configuration-mapred/mapred-site.xml index 671f328..0544cf3 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/YARN/configuration-mapred/mapred-site.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/YARN/configuration-mapred/mapred-site.xml @@ -1,6 +1,5 @@ - - - - - - - + + mapreduce.jobhistory.intermediate-done-dir glusterfs:///mr-history/tmp Directory where history files are written by MapReduce jobs. + + - mapreduce.jobhistory.done-dir glusterfs:///mr-history/done Directory where history files are managed by the MR JobHistory Server. + + - yarn.app.mapreduce.am.staging-dir - glusterfs:///user - + yarn.app.mapreduce.am.staging-dir + glusterfs:///user + The staging dir used while submitting jobs. + + - mapred.healthChecker.script.path - glusterfs:///mapred/jobstatus - + mapred.healthChecker.script.path + glusterfs:///mapred/jobstatus + + + - mapred.job.tracker.history.completed.location - glusterfs:///mapred/history/done + mapred.job.tracker.history.completed.location + glusterfs:///mapred/history/done + + - mapred.system.dir glusterfs:///mapred/system + + - mapreduce.jobtracker.staging.root.dir glusterfs:///user + + + + + mapred.healthChecker.script.path + glusterfs:///mapred/jobstatus + + + + + mapred.job.tracker.history.completed.location + glusterfs:///mapred/history/done + + + + + mapred.system.dir + glusterfs:///mapred/system + + + + + mapreduce.jobtracker.staging.root.dir + glusterfs:///user + + - - -mapred.healthChecker.script.path -glusterfs:///mapred/jobstatus - - - -mapred.job.tracker.history.completed.location -glusterfs:///mapred/history/done - - - -mapred.system.dir -glusterfs:///mapred/system - - - -mapreduce.jobtracker.staging.root.dir -glusterfs:///user - - http://git-wip-us.apache.org/repos/asf/ambari/blob/6919aa50/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/YARN/configuration/capacity-scheduler.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/YARN/configuration/capacity-scheduler.xml b/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/YARN/configuration/capacity-scheduler.xml index b99849b..6f597a2 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/YARN/configuration/capacity-scheduler.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/YARN/configuration/capacity-scheduler.xml @@ -1,3 +1,4 @@ + - - yarn.scheduler.capacity.maximum-applications 10000 Maximum number of applications that can be pending and running. + + - yarn.scheduler.capacity.maximum-am-resource-percent 0.2 @@ -33,16 +33,18 @@ application masters i.e. controls number of concurrent running applications. + + - yarn.scheduler.capacity.root.queues default The queues at the this level (root is the root queue). + + - yarn.scheduler.capacity.root.capacity 100 @@ -52,54 +54,61 @@ The child queues capacity should add up to their parent queue's capacity or less. + + - yarn.scheduler.capacity.root.default.capacity 100 Default queue target capacity. + + - yarn.scheduler.capacity.root.default.user-limit-factor 1 Default queue user limit a percentage from 0.0 to 1.0. + + - yarn.scheduler.capacity.root.default.maximum-capacity 100 The maximum capacity of the default queue. + + - yarn.scheduler.capacity.root.default.state RUNNING The state of the default queue. State can be one of RUNNING or STOPPED. + + - yarn.scheduler.capacity.root.default.acl_submit_applications * The ACL of who can submit jobs to the default queue. + + - yarn.scheduler.capacity.root.default.acl_administer_jobs * The ACL of who can administer jobs on the default queue. + + - yarn.scheduler.capacity.root.acl_administer_queue * @@ -107,8 +116,9 @@ The ACL for who can administer this queue i.e. change sub-queue allocations. + + - yarn.scheduler.capacity.node-locality-delay 40 @@ -118,15 +128,16 @@ Typically this should be set to number of nodes in the cluster, By default is setting approximately number of nodes in one rack which is 40. + + - yarn.scheduler.capacity.default.minimum-user-limit-percent 100 Default minimum queue resource limit depends on the number of users who have submitted applications. + + - - http://git-wip-us.apache.org/repos/asf/ambari/blob/6919aa50/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/YARN/configuration/yarn-env.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/YARN/configuration/yarn-env.xml b/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/YARN/configuration/yarn-env.xml index 2cf21fa..c37e994 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/YARN/configuration/yarn-env.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/YARN/configuration/yarn-env.xml @@ -19,48 +19,63 @@ * limitations under the License. */ --> - yarn_log_dir_prefix /var/log/hadoop-yarn YARN Log Dir Prefix + + yarn_pid_dir_prefix /var/run/hadoop-yarn YARN PID Dir Prefix + + yarn_user Yarn User yarn YARN User + + yarn_heapsize 1024 Max heapsize for all YARN components using a numerical value in the scale of MB + + resourcemanager_heapsize 1024 Max heapsize for ResourceManager using a numerical value in the scale of MB + + nodemanager_heapsize 1024 Max heapsize for NodeManager using a numerical value in the scale of MB + + apptimelineserver_heapsize 1024 Max heapsize for AppTimelineServer using a numerical value in the scale of MB + + namenode_heapsize 1024 Max heapsize for NameNode using a numerical value in the scale of MB + + @@ -182,5 +197,7 @@ YARN_OPTS="$YARN_OPTS -Dyarn.policy.file=$YARN_POLICYFILE" content + + http://git-wip-us.apache.org/repos/asf/ambari/blob/6919aa50/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/YARN/configuration/yarn-site.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/YARN/configuration/yarn-site.xml b/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/YARN/configuration/yarn-site.xml index 893ccd8..4b236db 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/YARN/configuration/yarn-site.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6.GlusterFS/services/YARN/configuration/yarn-site.xml @@ -16,31 +16,30 @@ See the License for the specific language governing permissions and limitations under the License. --> - - - - + - yarn.resourcemanager.hostname localhost The hostname of the RM. + + - yarn.resourcemanager.resource-tracker.address localhost:8025 The address of ResourceManager. + + - yarn.resourcemanager.scheduler.address localhost:8030 The address of the scheduler interface. + + - yarn.resourcemanager.address localhost:8050 @@ -48,20 +47,23 @@ The address of the applications manager interface in the RM. + + - yarn.resourcemanager.admin.address localhost:8141 The address of the RM admin interface. + + - yarn.resourcemanager.scheduler.class org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler The class to use as the resource scheduler. + + - yarn.scheduler.minimum-allocation-mb 512 @@ -70,8 +72,9 @@ in MBs. Memory requests lower than this won't take effect, and the specified value will get allocated at minimum. + + - yarn.scheduler.maximum-allocation-mb 2048 @@ -80,41 +83,46 @@ in MBs. Memory requests higher than this won't take effect, and will get capped to this value. + + - yarn.acl.enable false Are acls enabled. + + - yarn.admin.acl - + ACL of who can be admin of the YARN cluster. + + - - yarn.nodemanager.address 0.0.0.0:45454 The address of the container manager in the NM. + + - yarn.nodemanager.resource.memory-mb 5120 Amount of physical memory, in MB, that can be allocated for containers. + + - yarn.application.classpath /etc/hadoop/conf,/usr/lib/hadoop/*,/usr/lib/hadoop/lib/*,/usr/lib/hadoop-hdfs/*,/usr/lib/hadoop-hdfs/lib/*,/usr/lib/hadoop-yarn/*,/usr/lib/hadoop-yarn/lib/*,/usr/lib/hadoop-mapreduce/*,/usr/lib/hadoop-mapreduce/lib/* Classpath for typical applications. + + - yarn.nodemanager.vmem-pmem-ratio 2.1 @@ -123,33 +131,38 @@ expressed in terms of physical memory, and virtual memory usage is allowed to exceed this allocation by this ratio. + + - yarn.nodemanager.container-executor.class org.apache.hadoop.yarn.server.nodemanager.GlusterContainerExecutor ContainerExecutor for launching containers + + - yarn.nodemanager.linux-container-executor.group hadoop Unix group of the NodeManager + + - yarn.nodemanager.aux-services mapreduce_shuffle Auxilliary services of NodeManager. A valid service name should only contain a-zA-Z0-9_ and can not start with numbers + + - yarn.nodemanager.aux-services.mapreduce_shuffle.class org.apache.hadoop.mapred.ShuffleHandler The auxiliary service class to use + + - yarn.nodemanager.log-dirs /hadoop/yarn/log @@ -160,8 +173,9 @@ named container_{$contid}. Each container directory will contain the files stderr, stdin, and syslog generated by that container. + + - yarn.nodemanager.local-dirs /hadoop/yarn/local @@ -172,8 +186,9 @@ Individual containers' work directories, called container_${contid}, will be subdirectories of this. + + - yarn.nodemanager.container-monitor.interval-ms 3000 @@ -181,8 +196,9 @@ The interval, in milliseconds, for which the node manager waits between two cycles of monitoring its containers' memory usage. + + - - yarn.nodemanager.health-checker.interval-ms 135000 Frequency of running node health script. + + - yarn.nodemanager.health-checker.script.timeout-ms 60000 Script time out period. + + - yarn.nodemanager.log.retain-second 604800 @@ -210,20 +227,23 @@ Time in seconds to retain user logs. Only applicable if log aggregation is disabled. + + - yarn.log-aggregation-enable true Whether to enable log aggregation. + + - yarn.nodemanager.remote-app-log-dir /app-logs Location to aggregate logs to. + + - yarn.nodemanager.remote-app-log-dir-suffix logs @@ -231,16 +251,18 @@ The remote log dir will be created at {yarn.nodemanager.remote-app-log-dir}/${user}/{thisParam}. + + - yarn.nodemanager.log-aggregation.compression-type gz T-file compression types used to compress aggregated logs. + + - yarn.nodemanager.delete.debug-delay-sec 0 @@ -259,8 +281,9 @@ of the Yarn applications' log directories is configurable with the yarn.nodemanager.log-dirs property (see also below). + + - yarn.log-aggregation.retain-seconds 2592000 @@ -268,8 +291,9 @@ How long to keep aggregation logs before deleting them. -1 disables. Be careful set this too small and you will spam the name node. + + - yarn.nodemanager.admin-env MALLOC_ARENA_MAX=$MALLOC_ARENA_MAX @@ -277,8 +301,9 @@ Environment variables that should be forwarded from the NodeManager's environment to the container's. + + - yarn.nodemanager.disk-health-checker.min-healthy-disks 0.25 @@ -289,8 +314,9 @@ If there are less number of healthy local-dirs (or log-dirs) available, then new containers will not be launched on this node. + + - yarn.resourcemanager.am.max-attempts 2 @@ -302,32 +328,36 @@ the resourcemanager will override it. The default number is set to 2, to allow at least one retry for AM. + + - yarn.resourcemanager.webapp.address localhost:8088 The address of the RM web application. + + - yarn.nodemanager.vmem-check-enabled false Whether virtual memory limits will be enforced for containers. + + - yarn.log.server.url http://localhost:19888/jobhistory/logs URI for the HistoryServer's log resource + + - yarn.resourcemanager.nodes.exclude-path /etc/hadoop/conf/yarn.exclude @@ -337,56 +367,63 @@ file must be specified. If the value is empty, no hosts are excluded. + + - yarn.timeline-service.enabled true Indicate to clients whether timeline service is enabled or not. If enabled, clients will put entities and events to the timeline server. + + - yarn.timeline-service.store-class org.apache.hadoop.yarn.server.timeline.LeveldbTimelineStore Store class name for timeline store + + - yarn.timeline-service.generic-application-history.store-class org.apache.hadoop.yarn.server.applicationhistoryservice.NullApplicationHistoryStore Store class name for history store, defaulting to file system store + + - yarn.timeline-service.leveldb-timeline-store.path /mnt/glusterfs/hadoop/yarn/timeline Store file name for leveldb timeline store + + - yarn.timeline-service.webapp.address 0.0.0.0:8188 The http address of the timeline service web application. + + - yarn.timeline-service.webapp.https.address 0.0.0.0:8190 The http address of the timeline service web application. + + - yarn.timeline-service.address 0.0.0.0:10200 @@ -394,20 +431,28 @@ This is default address for the timeline server to start the RPC server. + + Enable age off of timeline store data. yarn.timeline-service.ttl-enable true + + Time to live for timeline store data in milliseconds. yarn.timeline-service.ttl-ms 2678400000 + + Length of time to wait between deletion cycles of leveldb timeline store in milliseconds. yarn.timeline-service.leveldb-timeline-store.ttl-interval-ms 300000 + + http://git-wip-us.apache.org/repos/asf/ambari/blob/6919aa50/ambari-server/src/main/resources/stacks/HDP/2.0.6/configuration/cluster-env.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/configuration/cluster-env.xml b/ambari-server/src/main/resources/stacks/HDP/2.0.6/configuration/cluster-env.xml index 31556cb..2ae5b72 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/configuration/cluster-env.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/configuration/cluster-env.xml @@ -19,189 +19,225 @@ * limitations under the License. */ --> - - - recovery_enabled - true - Auto start enabled or not for this cluster. - - - recovery_type - AUTO_START - Auto start type. - - - recovery_lifetime_max_count - 1024 - Auto start lifetime maximum count of recovery attempt allowed per host component. This is reset when agent is restarted. - - - recovery_max_count - 6 - Auto start maximum count of recovery attempt allowed per host component in a window. This is reset when agent is restarted. - - - recovery_window_in_minutes - 60 - Auto start recovery window size in minutes. - - - recovery_retry_interval - 5 - Auto start recovery retry gap between tries per host component. - - - security_enabled - false - Hadoop Security - - - kerberos_domain - EXAMPLE.COM - Kerberos realm. - - - ignore_groupsusers_create - Skip group modifications during install - false - ADDITIONAL_USER_PROPERTY - Whether to ignore failures on users and group creation - - false - boolean - - - - smokeuser - Smoke User - ambari-qa - USER - User executing service checks - - user - false - - - - smokeuser_keytab - /etc/security/keytabs/smokeuser.headless.keytab - Path to smoke test user keytab file - - - user_group - Hadoop Group - hadoop - GROUP - Hadoop user group. - - user - false - - - - repo_suse_rhel_template - [{{repo_id}}] + + recovery_enabled + true + Auto start enabled or not for this cluster. + + + + + recovery_type + AUTO_START + Auto start type. + + + + + recovery_lifetime_max_count + 1024 + Auto start lifetime maximum count of recovery attempt allowed per host component. This is reset when agent is restarted. + + + + + recovery_max_count + 6 + Auto start maximum count of recovery attempt allowed per host component in a window. This is reset when agent is restarted. + + + + + recovery_window_in_minutes + 60 + Auto start recovery window size in minutes. + + + + + recovery_retry_interval + 5 + Auto start recovery retry gap between tries per host component. + + + + + security_enabled + false + Hadoop Security + + + + + kerberos_domain + EXAMPLE.COM + Kerberos realm. + + + + + ignore_groupsusers_create + Skip group modifications during install + false + ADDITIONAL_USER_PROPERTY + Whether to ignore failures on users and group creation + + false + boolean + + + + + + smokeuser + Smoke User + ambari-qa + USER + User executing service checks + + user + false + + + + + + smokeuser_keytab + /etc/security/keytabs/smokeuser.headless.keytab + Path to smoke test user keytab file + + + + + user_group + Hadoop Group + hadoop + GROUP + Hadoop user group. + + user + false + + + + + + repo_suse_rhel_template + [{{repo_id}}] name={{repo_id}} {% if mirror_list %}mirrorlist={{mirror_list}}{% else %}baseurl={{base_url}}{% endif %} path=/ enabled=1 gpgcheck=0 - Template of repositories for rhel and suse. - - - repo_ubuntu_template - {{package_type}} {{base_url}} {{components}} - Template of repositories for ubuntu. - - - - override_uid - true - ADDITIONAL_USER_PROPERTY - Have Ambari manage UIDs - Have Ambari manage UIDs - - false - boolean - - - - - fetch_nonlocal_groups - true - Ambari fetch nonlocal groups - Ambari requires fetching all the groups. This can be slow + Template of repositories for rhel and suse. + + + + + repo_ubuntu_template + {{package_type}} {{base_url}} {{components}} + Template of repositories for ubuntu. + + + + + override_uid + true + ADDITIONAL_USER_PROPERTY + Have Ambari manage UIDs + Have Ambari manage UIDs + + false + boolean + + + + + + fetch_nonlocal_groups + true + Ambari fetch nonlocal groups + Ambari requires fetching all the groups. This can be slow on envs with enabled ldap. Setting this option to false will enable Ambari, to skip user/group management connected with ldap groups. - - false - boolean - - - - - managed_hdfs_resource_property_names - - Comma separated list of property names with HDFS resource paths. + + false + boolean + + + + + + managed_hdfs_resource_property_names + + Comma separated list of property names with HDFS resource paths. Resource from this list will be managed even if it is marked as not managed in the stack - - false - true - - - - - stack_tools - - Stack specific tools - VALUE_FROM_PROPERTY_FILE - - stack_tools.json - json - - - - - stack_features - - List of features supported by the stack - VALUE_FROM_PROPERTY_FILE - - stack_features.json - json - - - - - stack_root - /usr/hdp - Stack root folder - - - - alerts_repeat_tolerance - 1 - The number of consecutive alerts required to transition an alert from the SOFT to the HARD state. - - - - ignore_bad_mounts - false - For properties handled by handle_mounted_dirs this will make Ambari not to create any directories. - - - - create_dirs_on_root - true - For properties handled by handle_mounted_dirs this will make Ambari to create not-existent unknown directories on / partition - - - - one_dir_per_partition - true - For properties handled by handle_mounted_dirs this will make Ambari - - + + false + true + + + + + + stack_tools + + Stack specific tools + VALUE_FROM_PROPERTY_FILE + + stack_tools.json + json + + + + + + stack_features + + List of features supported by the stack + VALUE_FROM_PROPERTY_FILE + + stack_features.json + json + + + + + + stack_root + /usr/hdp + Stack root folder + + + + + alerts_repeat_tolerance + 1 + The number of consecutive alerts required to transition an alert from the SOFT to the HARD state. + + + + + ignore_bad_mounts + false + For properties handled by handle_mounted_dirs this will make Ambari not to create any directories. + + + + + create_dirs_on_root + true + For properties handled by handle_mounted_dirs this will make Ambari to create not-existent unknown directories on / partition + + + + + one_dir_per_partition + true + For properties handled by handle_mounted_dirs this will make Ambari + + + http://git-wip-us.apache.org/repos/asf/ambari/blob/6919aa50/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/OOZIE/configuration/oozie-env.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/OOZIE/configuration/oozie-env.xml b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/OOZIE/configuration/oozie-env.xml index 87a885f..efd97c9 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/OOZIE/configuration/oozie-env.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/OOZIE/configuration/oozie-env.xml @@ -19,9 +19,7 @@ * limitations under the License. */ --> - - content @@ -96,9 +94,10 @@ export JAVA_LIBRARY_PATH={{hadoop_lib_home}}/native/Linux-amd64-64 export HADOOP_OPTS="-Dhdp.version=${HDP_VERSION} ${HADOOP_OPTS}" - - content - + + content + + + - http://git-wip-us.apache.org/repos/asf/ambari/blob/6919aa50/ambari-server/src/main/resources/stacks/HDP/2.1.GlusterFS/services/FALCON/configuration/falcon-env.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.1.GlusterFS/services/FALCON/configuration/falcon-env.xml b/ambari-server/src/main/resources/stacks/HDP/2.1.GlusterFS/services/FALCON/configuration/falcon-env.xml index a48ede3..e0df663 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.1.GlusterFS/services/FALCON/configuration/falcon-env.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.1.GlusterFS/services/FALCON/configuration/falcon-env.xml @@ -1,3 +1,4 @@ + falcon.embeddedmq.data /hadoop/falcon/embeddedmq/data Directory in which embeddedmq data is stored. + + falcon.embeddedmq true Whether embeddedmq is enabled or not. + + falcon.emeddedmq.port 61616 Port that embeddedmq will listen on. + + http://git-wip-us.apache.org/repos/asf/ambari/blob/6919aa50/ambari-server/src/main/resources/stacks/HDP/2.1.GlusterFS/services/FALCON/configuration/falcon-runtime.properties.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.1.GlusterFS/services/FALCON/configuration/falcon-runtime.properties.xml b/ambari-server/src/main/resources/stacks/HDP/2.1.GlusterFS/services/FALCON/configuration/falcon-runtime.properties.xml index 94c8755..ddb7f69 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.1.GlusterFS/services/FALCON/configuration/falcon-runtime.properties.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.1.GlusterFS/services/FALCON/configuration/falcon-runtime.properties.xml @@ -1,3 +1,4 @@ + *.domain ${falcon.app.type} - + + + *.config.store.uri file:///hadoop/falcon/store Location to store user entity configurations + + *.system.lib.location ${falcon.home}/server/webapp/${falcon.app.type}/WEB-INF/lib Location of libraries that is shipped to Hadoop + + *.retry.recorder.path ${falcon.log.dir}/retry - + + + *.falcon.cleanup.service.frequency days(1) - + + + *.broker.url tcp://localhost:61616 Default Active MQ url + + *.broker.ttlInMins 4320 - + + + *.entity.topic FALCON.ENTITY.TOPIC - + + + *.max.retry.failure.count 1 - + + + *.internal.queue.size 1000 - + + + *.falcon.http.authentication.cookie.domain EXAMPLE.COM - + + + *.falcon.http.authentication.blacklisted.users - + Comma separated list of black listed users + + *.falcon.authentication.type simple - + + + *.falcon.http.authentication.type simple - + + + *.falcon.http.authentication.token.validity 36000 - + + + *.falcon.http.authentication.signature.secret falcon - + + + *.falcon.http.authentication.simple.anonymous.allowed true Indicates if anonymous requests are allowed when using 'simple' authentication + + *.falcon.http.authentication.kerberos.name.rules DEFAULT The kerberos names rules is to resolve kerberos principal names, refer to Hadoop's KerberosName for more details. + + *.falcon.service.authentication.kerberos.principal falcon/_HOST@EXAMPLE.COM - + + + *.falcon.service.authentication.kerberos.keytab /etc/security/keytabs/falcon.service.keytab - + + + *.dfs.namenode.kerberos.principal nn/_HOST@EXAMPLE.COM name node principal to talk to config store + + *.falcon.http.authentication.kerberos.principal HTTP/_HOST@EXAMPLE.COM Indicates the Kerberos principal to be used for HTTP endpoint + + *.falcon.http.authentication.kerberos.keytab /etc/security/keytabs/spnego.service.keytab Location of the keytab file with the credentials for the HTTP principal + + http://git-wip-us.apache.org/repos/asf/ambari/blob/6919aa50/ambari-server/src/main/resources/stacks/HDP/2.1.GlusterFS/services/FALCON/configuration/oozie-site.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.1.GlusterFS/services/FALCON/configuration/oozie-site.xml b/ambari-server/src/main/resources/stacks/HDP/2.1.GlusterFS/services/FALCON/configuration/oozie-site.xml index 89e914b..0874115 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.1.GlusterFS/services/FALCON/configuration/oozie-site.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.1.GlusterFS/services/FALCON/configuration/oozie-site.xml @@ -16,7 +16,6 @@ See the License for the specific language governing permissions and limitations under the License. --> - oozie.service.ELService.ext.functions.coord-job-submit-instances @@ -40,8 +39,9 @@ custom + + - oozie.service.ELService.ext.functions.coord-action-create-inst @@ -65,8 +65,9 @@ custom + + - oozie.service.ELService.ext.functions.coord-action-create @@ -90,8 +91,9 @@ custom + + - oozie.service.ELService.ext.functions.coord-job-submit-data @@ -116,8 +118,9 @@ custom + + - oozie.service.ELService.ext.functions.coord-action-start @@ -144,8 +147,9 @@ custom + + - oozie.service.ELService.ext.functions.coord-sla-submit @@ -158,8 +162,9 @@ custom + + - oozie.service.ELService.ext.functions.coord-sla-create @@ -172,17 +177,18 @@ custom + + - - - + + + - - - - + + + http://git-wip-us.apache.org/repos/asf/ambari/blob/6919aa50/ambari-server/src/main/resources/stacks/HDP/2.1.GlusterFS/services/GLUSTERFS/configuration/core-site.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.1.GlusterFS/services/GLUSTERFS/configuration/core-site.xml b/ambari-server/src/main/resources/stacks/HDP/2.1.GlusterFS/services/GLUSTERFS/configuration/core-site.xml index 12dfafc..9cc000f 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.1.GlusterFS/services/GLUSTERFS/configuration/core-site.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.1.GlusterFS/services/GLUSTERFS/configuration/core-site.xml @@ -1,7 +1,6 @@ - - - - - - - - + + - fs.AbstractFileSystem.glusterfs.impl - GlusterFS fs impl - org.apache.hadoop.fs.local.GlusterFs + fs.AbstractFileSystem.glusterfs.impl + GlusterFS fs impl + org.apache.hadoop.fs.local.GlusterFs + + - - fs.glusterfs.impl - org.apache.hadoop.fs.glusterfs.GlusterFileSystem - GlusterFS Abstract File System Implementation + fs.glusterfs.impl + org.apache.hadoop.fs.glusterfs.GlusterFileSystem + GlusterFS Abstract File System Implementation + + - fs.defaultFS glusterfs:///localhost:8020 - + + + http://git-wip-us.apache.org/repos/asf/ambari/blob/6919aa50/ambari-server/src/main/resources/stacks/HDP/2.1.GlusterFS/services/GLUSTERFS/configuration/hadoop-env.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.1.GlusterFS/services/GLUSTERFS/configuration/hadoop-env.xml b/ambari-server/src/main/resources/stacks/HDP/2.1.GlusterFS/services/GLUSTERFS/configuration/hadoop-env.xml index cd3ea79..b5c31ed 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.1.GlusterFS/services/GLUSTERFS/configuration/hadoop-env.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.1.GlusterFS/services/GLUSTERFS/configuration/hadoop-env.xml @@ -19,7 +19,6 @@ * limitations under the License. */ --> - hadoop_pid_dir_prefix @@ -31,15 +30,19 @@ true false + + - + hadoop_heapsize 1024 - Hadoop maximum Java heap size - Hadoop maximum Java heap size - - MB - + Hadoop maximum Java heap size + Hadoop maximum Java heap size + + MB + + + glusterfs_user @@ -49,6 +52,8 @@ false + + hdfs_log_dir_prefix @@ -60,6 +65,8 @@ true false + + namenode_heapsize @@ -71,16 +78,22 @@ true false + + namenode_host - + NameNode Host. + + snamenode_host - + Secondary NameNode. + + proxyuser_group @@ -90,12 +103,16 @@ false + + hdfs_user HDFS User hdfs User to run HDFS as + + - + content hadoop-env template @@ -198,12 +215,12 @@ export HADOOP_IDENT_STRING=$USER # Use libraries from standard classpath JAVA_JDBC_LIBS="" #Add libraries required by mysql connector -for jarFile in `ls /usr/share/java/*mysql* 2>/dev/null` +for jarFile in `ls /usr/share/java/*mysql* 2>/dev/null` do JAVA_JDBC_LIBS=${JAVA_JDBC_LIBS}:$jarFile done #Add libraries required by oracle connector -for jarFile in `ls /usr/share/java/*ojdbc* 2>/dev/null` +for jarFile in `ls /usr/share/java/*ojdbc* 2>/dev/null` do JAVA_JDBC_LIBS=${JAVA_JDBC_LIBS}:$jarFile done @@ -222,8 +239,9 @@ export HADOOP_LIBEXEC_DIR={{hadoop_libexec_dir}} export JAVA_LIBRARY_PATH=${JAVA_LIBRARY_PATH}:/usr/lib/hadoop/lib/native/Linux-amd64-64 - content + content + + -