mirror of https://github.com/ansible/ansible.git
VMware: new module : vmware_cluster_facts (#37105)
Signed-off-by: Abhijeet Kasurde <akasurde@redhat.com>pull/39000/merge
parent
87a07e81a9
commit
268c682a05
@ -0,0 +1,225 @@
|
|||||||
|
#!/usr/bin/python
|
||||||
|
# -*- coding: utf-8 -*-
|
||||||
|
# Copyright: (c) 2018, Ansible Project
|
||||||
|
# Copyright: (c) 2018, Abhijeet Kasurde <akasurde@redhat.com>
|
||||||
|
# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt)
|
||||||
|
|
||||||
|
from __future__ import absolute_import, division, print_function
|
||||||
|
__metaclass__ = type
|
||||||
|
|
||||||
|
ANSIBLE_METADATA = {
|
||||||
|
'metadata_version': '1.1',
|
||||||
|
'status': ['preview'],
|
||||||
|
'supported_by': 'community'
|
||||||
|
}
|
||||||
|
|
||||||
|
DOCUMENTATION = '''
|
||||||
|
---
|
||||||
|
module: vmware_cluster_facts
|
||||||
|
short_description: Gather facts about clusters available in given vCenter
|
||||||
|
description:
|
||||||
|
- This module can be used to gather facts about clusters in VMWare infrastructure.
|
||||||
|
- All values and VMware object names are case sensitive.
|
||||||
|
version_added: '2.6'
|
||||||
|
author:
|
||||||
|
- Abhijeet Kasurde (@akasurde)
|
||||||
|
notes:
|
||||||
|
- Tested on vSphere 6.5
|
||||||
|
requirements:
|
||||||
|
- "python >= 2.6"
|
||||||
|
- PyVmomi
|
||||||
|
options:
|
||||||
|
datacenter:
|
||||||
|
description:
|
||||||
|
- Datacenter to search for cluster/s.
|
||||||
|
- This parameter is required, if C(cluster_name) is not supplied.
|
||||||
|
required: False
|
||||||
|
cluster_name:
|
||||||
|
description:
|
||||||
|
- Name of the cluster.
|
||||||
|
- If set, facts of this cluster will be returned.
|
||||||
|
- This parameter is required, if C(datacenter) is not supplied.
|
||||||
|
required: False
|
||||||
|
extends_documentation_fragment: vmware.documentation
|
||||||
|
'''
|
||||||
|
|
||||||
|
EXAMPLES = '''
|
||||||
|
- name: Gather cluster facts from given datacenter
|
||||||
|
vmware_cluster_facts:
|
||||||
|
hostname: 192.168.1.209
|
||||||
|
username: administrator@vsphere.local
|
||||||
|
password: vmware
|
||||||
|
datacenter: ha-datacenter
|
||||||
|
validate_certs: False
|
||||||
|
delegate_to: localhost
|
||||||
|
register: cluster_facts
|
||||||
|
|
||||||
|
- name: Gather facts from datacenter about specific cluster
|
||||||
|
vmware_cluster_facts:
|
||||||
|
hostname: 192.168.1.209
|
||||||
|
username: administrator@vsphere.local
|
||||||
|
password: vmware
|
||||||
|
cluster_name: DC0_C0
|
||||||
|
validate_certs: False
|
||||||
|
delegate_to: localhost
|
||||||
|
register: cluster_facts
|
||||||
|
'''
|
||||||
|
|
||||||
|
RETURN = """
|
||||||
|
clusters:
|
||||||
|
description: metadata about the available clusters
|
||||||
|
returned: always
|
||||||
|
type: dict
|
||||||
|
sample: {
|
||||||
|
"DC0_C0": {
|
||||||
|
"drs_default_vm_behavior": null,
|
||||||
|
"drs_enable_vm_behavior_overrides": null,
|
||||||
|
"drs_vmotion_rate": null,
|
||||||
|
"enable_ha": null,
|
||||||
|
"enabled_drs": true,
|
||||||
|
"enabled_vsan": false,
|
||||||
|
"ha_admission_control_enabled": null,
|
||||||
|
"ha_failover_level": null,
|
||||||
|
"ha_host_monitoring": null,
|
||||||
|
"ha_restart_priority": null,
|
||||||
|
"ha_vm_failure_interval": null,
|
||||||
|
"ha_vm_max_failure_window": null,
|
||||||
|
"ha_vm_max_failures": null,
|
||||||
|
"ha_vm_min_up_time": null,
|
||||||
|
"ha_vm_monitoring": null,
|
||||||
|
"ha_vm_tools_monitoring": null,
|
||||||
|
"vsan_auto_claim_storage": false
|
||||||
|
},
|
||||||
|
}
|
||||||
|
"""
|
||||||
|
|
||||||
|
try:
|
||||||
|
from pyVmomi import vim
|
||||||
|
except ImportError:
|
||||||
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
from ansible.module_utils.basic import AnsibleModule
|
||||||
|
from ansible.module_utils.vmware import PyVmomi, vmware_argument_spec, find_datacenter_by_name, find_cluster_by_name
|
||||||
|
|
||||||
|
|
||||||
|
class VmwreClusterFactsManager(PyVmomi):
|
||||||
|
def __init__(self, module):
|
||||||
|
super(VmwreClusterFactsManager, self).__init__(module)
|
||||||
|
datacenter = self.params.get('datacenter')
|
||||||
|
cluster_name = self.params.get('cluster_name')
|
||||||
|
self.cluster_objs = []
|
||||||
|
if datacenter:
|
||||||
|
datacenter_obj = find_datacenter_by_name(self.content, datacenter_name=datacenter)
|
||||||
|
if datacenter_obj is None:
|
||||||
|
self.module.fail_json(msg="Failed to find datacenter '%s'" % datacenter)
|
||||||
|
self.cluster_objs = self.get_all_cluster_objs(parent=datacenter_obj)
|
||||||
|
elif cluster_name:
|
||||||
|
cluster_obj = find_cluster_by_name(self.content, cluster_name=cluster_name)
|
||||||
|
if cluster_obj is None:
|
||||||
|
self.module.fail_json(msg="Failed to find cluster '%s'" % cluster_name)
|
||||||
|
|
||||||
|
self.cluster_objs = [cluster_obj]
|
||||||
|
|
||||||
|
def get_all_cluster_objs(self, parent):
|
||||||
|
"""
|
||||||
|
Function to get all cluster managed objects from given parent object
|
||||||
|
Args:
|
||||||
|
parent: Managed objected of datacenter or host folder
|
||||||
|
|
||||||
|
Returns: List of host managed objects
|
||||||
|
|
||||||
|
"""
|
||||||
|
cluster_objs = []
|
||||||
|
if isinstance(parent, vim.Datacenter):
|
||||||
|
folder = parent.hostFolder
|
||||||
|
else:
|
||||||
|
folder = parent
|
||||||
|
|
||||||
|
for child in folder.childEntity:
|
||||||
|
if isinstance(child, vim.Folder):
|
||||||
|
cluster_objs = cluster_objs + self.get_all_cluster_objs(child)
|
||||||
|
if isinstance(child, vim.ClusterComputeResource):
|
||||||
|
cluster_objs.append(child)
|
||||||
|
return cluster_objs
|
||||||
|
|
||||||
|
def gather_cluster_facts(self):
|
||||||
|
"""
|
||||||
|
Function to gather facts about cluster
|
||||||
|
|
||||||
|
"""
|
||||||
|
results = dict(changed=False, clusters=dict())
|
||||||
|
for cluster in self.cluster_objs:
|
||||||
|
# Default values
|
||||||
|
ha_failover_level = None
|
||||||
|
ha_restart_priority = None
|
||||||
|
ha_vm_tools_monitoring = None
|
||||||
|
ha_vm_min_up_time = None
|
||||||
|
ha_vm_max_failures = None
|
||||||
|
ha_vm_max_failure_window = None
|
||||||
|
ha_vm_failure_interval = None
|
||||||
|
enabled_vsan = False
|
||||||
|
vsan_auto_claim_storage = False
|
||||||
|
|
||||||
|
# HA
|
||||||
|
das_config = cluster.configurationEx.dasConfig
|
||||||
|
if das_config.admissionControlPolicy:
|
||||||
|
ha_failover_level = das_config.admissionControlPolicy.failoverLevel
|
||||||
|
if das_config.defaultVmSettings:
|
||||||
|
ha_restart_priority = das_config.defaultVmSettings.restartPriority,
|
||||||
|
ha_vm_tools_monitoring = das_config.defaultVmSettings.vmToolsMonitoringSettings.vmMonitoring,
|
||||||
|
ha_vm_min_up_time = das_config.defaultVmSettings.vmToolsMonitoringSettings.minUpTime,
|
||||||
|
ha_vm_max_failures = das_config.defaultVmSettings.vmToolsMonitoringSettings.maxFailures,
|
||||||
|
ha_vm_max_failure_window = das_config.defaultVmSettings.vmToolsMonitoringSettings.maxFailureWindow,
|
||||||
|
ha_vm_failure_interval = das_config.defaultVmSettings.vmToolsMonitoringSettings.failureInterval,
|
||||||
|
|
||||||
|
# DRS
|
||||||
|
drs_config = cluster.configurationEx.drsConfig
|
||||||
|
|
||||||
|
# VSAN
|
||||||
|
if hasattr(cluster.configurationEx, 'vsanConfig'):
|
||||||
|
vsan_config = cluster.configurationEx.vsanConfig
|
||||||
|
enabled_vsan = vsan_config.enabled,
|
||||||
|
vsan_auto_claim_storage = vsan_config.defaultConfig.autoClaimStorage,
|
||||||
|
|
||||||
|
results['clusters'][cluster.name] = dict(
|
||||||
|
enable_ha=das_config.enabled,
|
||||||
|
ha_failover_level=ha_failover_level,
|
||||||
|
ha_vm_monitoring=das_config.vmMonitoring,
|
||||||
|
ha_host_monitoring=das_config.hostMonitoring,
|
||||||
|
ha_admission_control_enabled=das_config.admissionControlEnabled,
|
||||||
|
ha_restart_priority=ha_restart_priority,
|
||||||
|
ha_vm_tools_monitoring=ha_vm_tools_monitoring,
|
||||||
|
ha_vm_min_up_time=ha_vm_min_up_time,
|
||||||
|
ha_vm_max_failures=ha_vm_max_failures,
|
||||||
|
ha_vm_max_failure_window=ha_vm_max_failure_window,
|
||||||
|
ha_vm_failure_interval=ha_vm_failure_interval,
|
||||||
|
enabled_drs=drs_config.enabled,
|
||||||
|
drs_enable_vm_behavior_overrides=drs_config.enableVmBehaviorOverrides,
|
||||||
|
drs_default_vm_behavior=drs_config.defaultVmBehavior,
|
||||||
|
drs_vmotion_rate=drs_config.vmotionRate,
|
||||||
|
enabled_vsan=enabled_vsan,
|
||||||
|
vsan_auto_claim_storage=vsan_auto_claim_storage,
|
||||||
|
)
|
||||||
|
|
||||||
|
self.module.exit_json(**results)
|
||||||
|
|
||||||
|
|
||||||
|
def main():
|
||||||
|
argument_spec = vmware_argument_spec()
|
||||||
|
argument_spec.update(
|
||||||
|
datacenter=dict(type='str'),
|
||||||
|
cluster_name=dict(type='str')
|
||||||
|
)
|
||||||
|
module = AnsibleModule(
|
||||||
|
argument_spec=argument_spec,
|
||||||
|
required_one_of=[
|
||||||
|
['cluster_name', 'datacenter'],
|
||||||
|
],
|
||||||
|
)
|
||||||
|
pyv = VmwreClusterFactsManager(module)
|
||||||
|
pyv.gather_cluster_facts()
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == '__main__':
|
||||||
|
main()
|
@ -0,0 +1,3 @@
|
|||||||
|
posix/ci/cloud/group4/vcenter
|
||||||
|
cloud/vcenter
|
||||||
|
|
@ -0,0 +1,91 @@
|
|||||||
|
# Test code for the vmware_cluster_facts module.
|
||||||
|
# Copyright: (c) 2018, Abhijeet Kasurde <akasurde@redhat.com>
|
||||||
|
# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt)
|
||||||
|
|
||||||
|
- name: make sure pyvmomi is installed
|
||||||
|
pip:
|
||||||
|
name: pyvmomi
|
||||||
|
state: latest
|
||||||
|
when: "{{ ansible_user_id == 'root' }}"
|
||||||
|
|
||||||
|
- name: store the vcenter container ip
|
||||||
|
set_fact:
|
||||||
|
vcsim: "{{ lookup('env', 'vcenter_host') }}"
|
||||||
|
|
||||||
|
- debug: var=vcsim
|
||||||
|
|
||||||
|
- name: Wait for Flask controller to come up online
|
||||||
|
wait_for:
|
||||||
|
host: "{{ vcsim }}"
|
||||||
|
port: 5000
|
||||||
|
state: started
|
||||||
|
|
||||||
|
- name: kill vcsim
|
||||||
|
uri:
|
||||||
|
url: http://{{ vcsim }}:5000/killall
|
||||||
|
|
||||||
|
- name: start vcsim
|
||||||
|
uri:
|
||||||
|
url: http://{{ vcsim }}:5000/spawn?cluster=2
|
||||||
|
register: vcsim_instance
|
||||||
|
|
||||||
|
- debug:
|
||||||
|
var: vcsim_instance
|
||||||
|
|
||||||
|
- name: Wait for vcsim server to come up online
|
||||||
|
wait_for:
|
||||||
|
host: "{{ vcsim }}"
|
||||||
|
port: 443
|
||||||
|
state: started
|
||||||
|
|
||||||
|
- name: get a list of Cluster from vcsim
|
||||||
|
uri:
|
||||||
|
url: http://{{ vcsim }}:5000/govc_find?filter=CCR
|
||||||
|
register: clusters
|
||||||
|
|
||||||
|
- name: get a cluster
|
||||||
|
set_fact:
|
||||||
|
ccr1: "{{ clusters.json[0] | basename }}"
|
||||||
|
|
||||||
|
- debug: var=ccr1
|
||||||
|
|
||||||
|
- name: get a list of datacenter
|
||||||
|
uri:
|
||||||
|
url: http://{{ vcsim }}:5000/govc_find?filter=DC
|
||||||
|
register: datacenters
|
||||||
|
|
||||||
|
- name: get a datacenter
|
||||||
|
set_fact:
|
||||||
|
dc1: "{{ datacenters.json[0] | basename }}"
|
||||||
|
|
||||||
|
- debug: var=dc1
|
||||||
|
|
||||||
|
- name: gather facts about all clusters in given datacenter
|
||||||
|
vmware_cluster_facts:
|
||||||
|
hostname: "{{ vcsim }}"
|
||||||
|
username: "{{ vcsim_instance.json.username }}"
|
||||||
|
password: "{{ vcsim_instance.json.password }}"
|
||||||
|
datacenter: "{{ dc1 }}"
|
||||||
|
validate_certs: False
|
||||||
|
register: all_cluster_result
|
||||||
|
|
||||||
|
- name: ensure facts are gathered for all cluster
|
||||||
|
assert:
|
||||||
|
that:
|
||||||
|
- all_cluster_result.clusters
|
||||||
|
- not all_cluster_result.changed
|
||||||
|
|
||||||
|
- name: gather facts about given cluster
|
||||||
|
vmware_cluster_facts:
|
||||||
|
hostname: "{{ vcsim }}"
|
||||||
|
username: "{{ vcsim_instance.json.username }}"
|
||||||
|
password: "{{ vcsim_instance.json.password }}"
|
||||||
|
cluster_name: "{{ ccr1 }}"
|
||||||
|
validate_certs: False
|
||||||
|
register: cluster_result
|
||||||
|
|
||||||
|
- name: ensure facts are gathered for the given cluster
|
||||||
|
assert:
|
||||||
|
that:
|
||||||
|
- cluster_result.clusters
|
||||||
|
- not cluster_result.changed
|
Loading…
Reference in New Issue