ambari-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Siddharth Wagle <swa...@hortonworks.com>
Subject Re: Re:Re: Re:How to get datanode numbers in stack_advisor.py#recommendHDFSConfigurations?
Date Thu, 16 Apr 2015 16:34:51 GMT
By we, I meant the stack advisor feature only.


-Sid


________________________________
From: Siddharth Wagle
Sent: Thursday, April 16, 2015 9:34 AM
To: user@ambari.apache.org
Subject: Re: Re:Re: Re:How to get datanode numbers in stack_advisor.py#recommendHDFSConfigurations?


Hi Qi Yun Li,


I believe the 1 host is intended behavior since a the moment we are not supporting heterogeneous
environments that is what any 1 of candidate hosts is chose to represent what cpu / memory
/ disk characteristics to use for recommending configurations for a component.


Srimanth, can attest to this.


BR,

Sid


________________________________
From: Qi Yun Liu <amari_liuqiyun@163.com>
Sent: Wednesday, April 15, 2015 11:34 PM
To: user@ambari.apache.org
Subject: Re:Re: Re:How to get datanode numbers in stack_advisor.py#recommendHDFSConfigurations?

Hi Siddharth,

Thanks a lot for your comments! According to your suggestions´╝î I did a test:
1. Launch Ambari server GUI and start a brand new cluster installation
2. In 'Assign Slaves and Clients' page, select two hosts(hostname0.com, hostname1.com) as
the datanodes
3. After clicking Next button, I found in the ambari-server/resources/stacks/HDP/2.2/services/stack_advisor.py#HDP22StackAdvisor#recommendHDFSConfigurations,
the value of datanodeHosts got by line 'datanodeHosts = self.getHostsWithComponent("HDFS",
"DATANODE", services, hosts)' only includes 1 host info as below and its length is 1 not 2:
datanodeHosts= [{u'href': u'/api/v1/hosts/hostname1.com', u'Hosts': {u'last_heartbeat_time':
1429155431532, u'recovery_summary': u'DISABLED', u'host_health_report': u'', u'disk_info':
[{u'available': u'83808976', u'used': u'7951800', u'percent': u'9%', u'device': u'/dev/mapper/vg_sdsvm923094-lv_root',
u'mountpoint': u'/', u'type': u'ext4', u'size': u'96671468'}, {u'available': u'4031416', u'used':
u'0', u'percent': u'0%', u'device': u'tmpfs', u'mountpoint': u'/dev/shm', u'type': u'tmpfs',
u'size': u'4031416'}, {u'available': u'378216', u'used': u'92028', u'percent': u'20%', u'device':
u'/dev/sda1', u'mountpoint': u'/boot', u'type': u'ext4', u'size': u'495844'}], u'desired_configs':
None, u'cpu_count': 2, u'recovery_report': {u'component_reports': [], u'summary': u'DISABLED'},
u'host_state': u'HEALTHY', u'os_arch': u'x86_64', u'total_mem': 8062836, u'host_status': u'HEALTHY',
u'last_registration_time': 1429153847302, u'os_family': u'redhat6', u'host_name': u'hostname1.com',
u'ip': u'9.30.249.190', u'rack_info': u'/default-rack', u'os_type': u'redhat6', u'last_agent_env':
{u'transparentHugePage': u'never', u'hostHealth': {u'agentTimeStampAtReporting': 1429155391291,
u'activeJavaProcs': [], u'serverTimeStampAtReporting': 1429155391348, u'liveServices': [{u'status':
u'Healthy', u'name': u'ntpd', u'desc': u''}]}, u'umask': 18, u'reverseLookup': True, u'alternatives':
[], u'existingUsers': [], u'firewallName': u'iptables', u'stackFoldersAndFiles': [{u'type':
u'directory', u'name': u'/etc/slider'}], u'existingRepos': [], u'installedPackages': [], u'firewallRunning':
False}, u'public_host_name': u'hostname1.com', u'ph_cpu_count': 2}}]

I think the correct value of datanodeHosts should not only include 'hostname1.com' but also
include 'hostname0.com', and the length of datanodeHosts should be 2, but not 1, because they
were all selected as datanode host.

Is it a bug?



At 2015-04-11 13:09:33, "Siddharth Wagle" <swagle@hortonworks.com<mailto:swagle@hortonworks.com>>
wrote:

Hi Qi Yun Liu,


This method is what you are looking for : stacks/HDP/2.0.6/services

HDP206StackAdvisor#getHostsWithComponent


Example:

datanodeHosts = self.getHostsWithComponent("HDFS", "DATANODE", services, hosts)


In: stacks/HDP/2.2/services/

HDP22StackAdvisor#recommendHDFSConfigurations


BR,

Sid


________________________________
From: Qi Yun Liu <amari_liuqiyun@163.com<mailto:amari_liuqiyun@163.com>>
Sent: Friday, April 10, 2015 7:13 PM
To: user@ambari.apache.org<mailto:user@ambari.apache.org>
Subject: Re:How to get datanode numbers in stack_advisor.py#recommendHDFSConfigurations?

I just wanna to get datanode numbers in stack_advisor.py#recommendHDFSConfigurations(self,
configurations, clusterData)?

Anyone could help me?

Thanks in advance!


At 2015-04-10 15:09:18, "Qi Yun Liu" <amari_liuqiyun@163.com<mailto:amari_liuqiyun@163.com>>
wrote:
Hi Experts,

How to get datanode numbers in stack_advisor.py#recommendHDFSConfigurations(self, configurations,
clusterData)? using its input parameters 'self', 'configurations' or 'clusterData'. At the
same time, another method 'def getComponentLayoutValidations(self, services, hosts)' has a
input parameter 'services', so it could get nameNodeHosts using 'services', but I failed to
get parameter 'services' in method recommendHDFSConfigurations.

Any comments?

Thanks a lot!







Mime
View raw message