ambari-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From nc...@apache.org
Subject [10/50] ambari git commit: AMBARI-14773. Stack advisor error for recommendations request (aonishuk)
Date Tue, 26 Jan 2016 21:45:13 GMT
AMBARI-14773. Stack advisor error for recommendations request (aonishuk)


Project: http://git-wip-us.apache.org/repos/asf/ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/6f6a7ce7
Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/6f6a7ce7
Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/6f6a7ce7

Branch: refs/heads/branch-dev-patch-upgrade
Commit: 6f6a7ce72986b20538fbf0578615ebc71ca0813b
Parents: 034cb32
Author: Andrew Onishuk <aonishuk@hortonworks.com>
Authored: Fri Jan 22 15:50:20 2016 +0200
Committer: Andrew Onishuk <aonishuk@hortonworks.com>
Committed: Fri Jan 22 15:50:20 2016 +0200

----------------------------------------------------------------------
 .../main/resources/stacks/HDP/2.0.6/services/stack_advisor.py  | 6 ++++--
 1 file changed, 4 insertions(+), 2 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/ambari/blob/6f6a7ce7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/stack_advisor.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/stack_advisor.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/stack_advisor.py
index 8e6dcf3..5bcebf7 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/stack_advisor.py
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/stack_advisor.py
@@ -241,7 +241,10 @@ class HDP206StackAdvisor(DefaultStackAdvisor):
 
     #Initialize default 'dfs.datanode.data.dir' if needed
     if (not hdfsSiteProperties) or ('dfs.datanode.data.dir' not in hdfsSiteProperties):
-      putHDFSSiteProperty('dfs.datanode.data.dir', '/hadoop/hdfs/data')
+      dataDirs = '/hadoop/hdfs/data'
+      putHDFSSiteProperty('dfs.datanode.data.dir', dataDirs)
+    else:
+      dataDirs = hdfsSiteProperties['dfs.datanode.data.dir'].split(",")
     #dfs.datanode.du.reserved should be set to 10-15% of volume size
     mountPoints = []
     mountPointDiskAvailableSpace = [] #kBytes
@@ -250,7 +253,6 @@ class HDP206StackAdvisor(DefaultStackAdvisor):
         mountPoints.append(diskInfo["mountpoint"])
         mountPointDiskAvailableSpace.append(long(diskInfo["size"]))
     maxFreeVolumeSize = 0l #kBytes
-    dataDirs = hdfsSiteProperties['dfs.datanode.data.dir'].split(",")
     for dataDir in dataDirs:
       mp = getMountPointForDir(dataDir, mountPoints)
       for i in range(len(mountPoints)):


Mime
View raw message