ambari-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Siddharth Wagle <swa...@hortonworks.com>
Subject Re: Re:Re: Re:Re: Re:How to get datanode numbers in stack_advisor.py#recommendHDFSConfigurations?
Date Fri, 17 Apr 2015 17:36:57 GMT
Hi Qi Yun Li,


I do not see any downstream changes in this patch. How does this help you?


Let me elaborate more on what I mean by heterogeneous envs are not supported:

When you have different host configurations the way to configure them differently is by using
the Config Group feature of Ambari. This is currently user driven since even if we change
stack_advisor to suggest Config Groups for different types of Host, there will be a need for
some UI work to make it usable on the configuration page for Ambari user. Having said that
I believe this is something in the work.


I looked at the patch and it does look harmless to me, but I would like to get opinions from
devs for this feature: sguturi@apache and dsen@apache.


Could you please open a Review Board for this at, reviews.apache.org?

https://cwiki.apache.org/confluence/display/AMBARI/How+to+Contribute (Review Process)


BR,

Sid


________________________________
From: Qi Yun Liu <amari_liuqiyun@163.com>
Sent: Friday, April 17, 2015 4:38 AM
To: user@ambari.apache.org
Subject: Re:Re: Re:Re: Re:How to get datanode numbers in stack_advisor.py#recommendHDFSConfigurations?

Hi Siddharth,

Yes, I agree that it makes sense to get only 1 host when we were not supporting heterogeneous
environments. At present, we might support heterogeneous environments, so it also make sense
to update this method accordingly. In addition, I really need this method to return all hosts
and it's also possible that others might also have similar requirements. Therefore, I created
a patch for jira AMBARI-10515 and did some tests to ensure no regression bug is brought. Could
you please help take a review on it?

Thanks a lot!





At 2015-04-17 00:34:51, "Siddharth Wagle" <swagle@hortonworks.com> wrote:

By we, I meant the stack advisor feature only.


-Sid


________________________________
From: Siddharth Wagle
Sent: Thursday, April 16, 2015 9:34 AM
To: user@ambari.apache.org<mailto:user@ambari.apache.org>
Subject: Re: Re:Re: Re:How to get datanode numbers in stack_advisor.py#recommendHDFSConfigurations?


Hi Qi Yun Li,


I believe the 1 host is intended behavior since a the moment we are not supporting heterogeneous
environments that is what any 1 of candidate hosts is chose to represent what cpu / memory
/ disk characteristics to use for recommending configurations for a component.


Srimanth, can attest to this.


BR,

Sid


________________________________
From: Qi Yun Liu <amari_liuqiyun@163.com<mailto:amari_liuqiyun@163.com>>
Sent: Wednesday, April 15, 2015 11:34 PM
To: user@ambari.apache.org<mailto:user@ambari.apache.org>
Subject: Re:Re: Re:How to get datanode numbers in stack_advisor.py#recommendHDFSConfigurations?

Hi Siddharth,

Thanks a lot for your comments! According to your suggestions´╝î I did a test:
1. Launch Ambari server GUI and start a brand new cluster installation
2. In 'Assign Slaves and Clients' page, select two hosts(hostname0.com, hostname1.com) as
the datanodes
3. After clicking Next button, I found in the ambari-server/resources/stacks/HDP/2.2/services/stack_advisor.py#HDP22StackAdvisor#recommendHDFSConfigurations,
the value of datanodeHosts got by line 'datanodeHosts = self.getHostsWithComponent("HDFS",
"DATANODE", services, hosts)' only includes 1 host info as below and its length is 1 not 2:
datanodeHosts= [{u'href': u'/api/v1/hosts/hostname1.com', u'Hosts': {u'last_heartbeat_time':
1429155431532, u'recovery_summary': u'DISABLED', u'host_health_report': u'', u'disk_info':
[{u'available': u'83808976', u'used': u'7951800', u'percent': u'9%', u'device': u'/dev/mapper/vg_sdsvm923094-lv_root',
u'mountpoint': u'/', u'type': u'ext4', u'size': u'96671468'}, {u'available': u'4031416', u'used':
u'0', u'percent': u'0%', u'device': u'tmpfs', u'mountpoint': u'/dev/shm', u'type': u'tmpfs',
u'size': u'4031416'}, {u'available': u'378216', u'used': u'92028', u'percent': u'20%', u'device':
u'/dev/sda1', u'mountpoint': u'/boot', u'type': u'ext4', u'size': u'495844'}], u'desired_configs':
None, u'cpu_count': 2, u'recovery_report': {u'component_reports': [], u'summary': u'DISABLED'},
u'host_state': u'HEALTHY', u'os_arch': u'x86_64', u'total_mem': 8062836, u'host_status': u'HEALTHY',
u'last_registration_time': 1429153847302, u'os_family': u'redhat6', u'host_name': u'hostname1.com',
u'ip': u'9.30.249.190', u'rack_info': u'/default-rack', u'os_type': u'redhat6', u'last_agent_env':
{u'transparentHugePage': u'never', u'hostHealth': {u'agentTimeStampAtReporting': 1429155391291,
u'activeJavaProcs': [], u'serverTimeStampAtReporting': 1429155391348, u'liveServices': [{u'status':
u'Healthy', u'name': u'ntpd', u'desc': u''}]}, u'umask': 18, u'reverseLookup': True, u'alternatives':
[], u'existingUsers': [], u'firewallName': u'iptables', u'stackFoldersAndFiles': [{u'type':
u'directory', u'name': u'/etc/slider'}], u'existingRepos': [], u'installedPackages': [], u'firewallRunning':
False}, u'public_host_name': u'hostname1.com', u'ph_cpu_count': 2}}]

I think the correct value of datanodeHosts should not only include 'hostname1.com' but also
include 'hostname0.com', and the length of datanodeHosts should be 2, but not 1, because they
were all selected as datanode host.

Is it a bug?



At 2015-04-11 13:09:33, "Siddharth Wagle" <swagle@hortonworks.com<mailto:swagle@hortonworks.com>>
wrote:

Hi Qi Yun Liu,


This method is what you are looking for : stacks/HDP/2.0.6/services

HDP206StackAdvisor#getHostsWithComponent


Example:

datanodeHosts = self.getHostsWithComponent("HDFS", "DATANODE", services, hosts)


In: stacks/HDP/2.2/services/

HDP22StackAdvisor#recommendHDFSConfigurations


BR,

Sid


________________________________
From: Qi Yun Liu <amari_liuqiyun@163.com<mailto:amari_liuqiyun@163.com>>
Sent: Friday, April 10, 2015 7:13 PM
To: user@ambari.apache.org<mailto:user@ambari.apache.org>
Subject: Re:How to get datanode numbers in stack_advisor.py#recommendHDFSConfigurations?

I just wanna to get datanode numbers in stack_advisor.py#recommendHDFSConfigurations(self,
configurations, clusterData)?

Anyone could help me?

Thanks in advance!


At 2015-04-10 15:09:18, "Qi Yun Liu" <amari_liuqiyun@163.com<mailto:amari_liuqiyun@163.com>>
wrote:
Hi Experts,

How to get datanode numbers in stack_advisor.py#recommendHDFSConfigurations(self, configurations,
clusterData)? using its input parameters 'self', 'configurations' or 'clusterData'. At the
same time, another method 'def getComponentLayoutValidations(self, services, hosts)' has a
input parameter 'services', so it could get nameNodeHosts using 'services', but I failed to
get parameter 'services' in method recommendHDFSConfigurations.

Any comments?

Thanks a lot!









Mime
View raw message