Hi Siddharth,

Yes, I agree that it makes sense to get only 1 host when we were not supporting 
heterogeneous environments. At present, we might support heterogeneous 
environments, so it also make sense to update this method accordingly. In 
addition, I really need this method to return all hosts and it's also possible 
that others might also have similar requirements. Therefore, I created a patch 
for jira AMBARI-10515 and did some tests to ensure no regression bug is 
brought. Could you please help take a review on it?


Thanks a lot!






At 2015-04-17 00:34:51, "Siddharth Wagle" <[email protected]> wrote:


By we, I meant the stack advisor feature only.




-Sid




From: Siddharth Wagle
Sent: Thursday, April 16, 2015 9:34 AM
To:[email protected]
Subject: Re: Re:Re: Re:How to get datanode numbers in 
stack_advisor.py#recommendHDFSConfigurations?
 

Hi Qi Yun Li,




I believe the 1 host is intended behavior since a the moment we are not 
supporting heterogeneous environments that is what any 1 of candidate hosts is 
chose to represent what cpu / memory / disk characteristics to use for 
recommending configurations for a component.





Srimanth, can attest to this.




BR,

Sid





From: Qi Yun Liu <[email protected]>
Sent: Wednesday, April 15, 2015 11:34 PM
To:[email protected]
Subject: Re:Re: Re:How to get datanode numbers in 
stack_advisor.py#recommendHDFSConfigurations?
 
Hi Siddharth,


Thanks a lot for your comments! According to your suggestions, I did a test:
1. Launch Ambari server GUI and start a brand new cluster installation
2. In 'Assign Slaves and Clients' page, select two hosts(hostname0.com, 
hostname1.com) as the datanodes
3. After clicking Next button, I found in the 
ambari-server/resources/stacks/HDP/2.2/services/stack_advisor.py#HDP22StackAdvisor#recommendHDFSConfigurations,
 the value of datanodeHosts got by line 'datanodeHosts = 
self.getHostsWithComponent("HDFS", "DATANODE", services, hosts)' only includes 
1 host info as below and its length is 1 not 2:
datanodeHosts= [{u'href': u'/api/v1/hosts/hostname1.com', u'Hosts': 
{u'last_heartbeat_time': 1429155431532, u'recovery_summary': u'DISABLED', 
u'host_health_report': u'', u'disk_info': [{u'available': u'83808976', u'used': 
u'7951800', u'percent': u'9%', u'device': 
u'/dev/mapper/vg_sdsvm923094-lv_root', u'mountpoint': u'/', u'type': u'ext4', 
u'size': u'96671468'}, {u'available': u'4031416', u'used': u'0', u'percent': 
u'0%', u'device': u'tmpfs', u'mountpoint': u'/dev/shm', u'type': u'tmpfs', 
u'size': u'4031416'}, {u'available': u'378216', u'used': u'92028', u'percent': 
u'20%', u'device': u'/dev/sda1', u'mountpoint': u'/boot', u'type': u'ext4', 
u'size': u'495844'}], u'desired_configs': None, u'cpu_count': 2, 
u'recovery_report': {u'component_reports': [], u'summary': u'DISABLED'}, 
u'host_state': u'HEALTHY', u'os_arch': u'x86_64', u'total_mem': 8062836, 
u'host_status': u'HEALTHY', u'last_registration_time': 1429153847302, 
u'os_family': u'redhat6', u'host_name': u'hostname1.com', u'ip': 
u'9.30.249.190', u'rack_info': u'/default-rack', u'os_type': u'redhat6', 
u'last_agent_env': {u'transparentHugePage': u'never', u'hostHealth': 
{u'agentTimeStampAtReporting': 1429155391291, u'activeJavaProcs': [], 
u'serverTimeStampAtReporting': 1429155391348, u'liveServices': [{u'status': 
u'Healthy', u'name': u'ntpd', u'desc': u''}]}, u'umask': 18, u'reverseLookup': 
True, u'alternatives': [], u'existingUsers': [], u'firewallName': u'iptables', 
u'stackFoldersAndFiles': [{u'type': u'directory', u'name': u'/etc/slider'}], 
u'existingRepos': [], u'installedPackages': [], u'firewallRunning': False}, 
u'public_host_name': u'hostname1.com', u'ph_cpu_count': 2}}]


I think the correct value of datanodeHosts should not only include 
'hostname1.com' but also include 'hostname0.com', and the length of 
datanodeHosts should be 2, but not 1, because they were all selected as 
datanode host.


Is it a bug? 




At 2015-04-11 13:09:33, "Siddharth Wagle" <[email protected]> wrote:


Hi Qi Yun Liu,




This method is what you are looking for : stacks/HDP/2.0.6/services

HDP206StackAdvisor#getHostsWithComponent





Example:

datanodeHosts = self.getHostsWithComponent("HDFS", "DATANODE", services, hosts)




In: stacks/HDP/2.2/services/

HDP22StackAdvisor#recommendHDFSConfigurations




BR,

Sid





From: Qi Yun Liu <[email protected]>
Sent: Friday, April 10, 2015 7:13 PM
To:[email protected]
Subject: Re:How to get datanode numbers in 
stack_advisor.py#recommendHDFSConfigurations?
 
I just wanna to get datanode numbers in 
stack_advisor.py#recommendHDFSConfigurations(self, configurations, clusterData)?


Anyone could help me?

Thanks in advance!



At 2015-04-10 15:09:18, "Qi Yun Liu" <[email protected]> wrote:

Hi Experts,


How to get datanode numbers in 
stack_advisor.py#recommendHDFSConfigurations(self, configurations, 
clusterData)? using its input parameters 'self', 'configurations' or 
'clusterData'. At the same time, another method 'def 
getComponentLayoutValidations(self, services, hosts)' has a input parameter 
'services', so it could get nameNodeHosts using 'services', but I failed to get 
parameter 'services' in method recommendHDFSConfigurations.


Any comments?


Thanks a lot!








Reply via email to