ambari-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Hudson (JIRA)" <>
Subject [jira] [Commented] (AMBARI-8244) Ambari HDP 2.0.6+ stacks do not work with fs.defaultFS not being hdfs
Date Tue, 31 Mar 2015 22:22:53 GMT


Hudson commented on AMBARI-8244:

SUCCESS: Integrated in Ambari-trunk-Commit #2161 (See [])
AMBARI-8244. Ambari HDP 2.0.6+ stacks do not work with fs.defaultFS not being hdfs (with Ivan
Mitic and Yusaku Sako) (smohanty:
* ambari-web/app/controllers/main/service/reassign/step4_controller.js
* ambari-server/src/main/resources/common-services/HDFS/
* ambari-server/src/test/python/stacks/2.0.6/HDFS/
* ambari-server/src/test/python/stacks/2.0.6/HDFS/
* ambari-web/app/models/service_config.js
* ambari-server/src/test/python/stacks/2.0.6/configs/altfs_plus_hdfs.json
* ambari-server/src/main/resources/common-services/HDFS/
* ambari-server/src/main/resources/common-services/HDFS/
* ambari-server/src/main/java/org/apache/ambari/server/controller/internal/
* ambari-server/src/main/resources/common-services/HDFS/

> Ambari HDP 2.0.6+ stacks do not work with fs.defaultFS not being hdfs
> ---------------------------------------------------------------------
>                 Key: AMBARI-8244
>                 URL:
>             Project: Ambari
>          Issue Type: Bug
>          Components: stacks
>    Affects Versions: 2.0.0
>            Reporter: Ivan Mitic
>            Assignee: Ivan Mitic
>              Labels: HDP
>             Fix For: 2.1.0
>         Attachments: AMBARI-8244.2.patch, AMBARI-8244.3.patch, AMBARI-8244.4.patch, AMBARI-8244.5.patch,
AMBARI-8244.6.patch, AMBARI-8244.7.combined.patch, AMBARI-8244.patch
> Right now changing the default file system does not work with the HDP 2.0.6+ stacks.
Given that it might be common to run HDP against some other file system in the cloud, adding
support for this will be super useful. One alternative is to consider a separate stack definition
for other file systems, however, given that I noticed just 2 minor bugs needed to support
this, I would rather extend on the existing code.
> Bugs:
>  - One issue is in Nagios install scripts, where it is assumed that fs.defaultFS has
the namenode port number.
>  - Another issue is in HDFS install scripts, where {{hadoop dfsadmin}} command only works
when hdfs is the default file system.
> Fix for both places is to extract the namenode address/port from {{dfs.namenode.rpc-address}}
if one is defined and use it instead of relying on {{fs.defaultFS}}. 
> Haven't included any tests yet (my first Ambari patch, not sure what is appropriate,
so please comment).

This message was sent by Atlassian JIRA

View raw message