From c57a7f05e1b37084110d2b1bc219b40fc9023ffa Mon Sep 17 00:00:00 2001 From: John Barker Date: Thu, 7 Dec 2017 16:27:06 +0000 Subject: Bulk autopep8 (modules) As agreed in 2017-12-07 Core meeting bulk fix pep8 issues Generated using: autopep8 1.3.3 (pycodestyle: 2.3.1) autopep8 -r --max-line-length 160 --in-place --ignore E305,E402,E722,E741 lib/ansible/modules Manually fix issues that autopep8 has introduced --- lib/ansible/modules/cloud/amazon/_ec2_ami_find.py | 52 +-- .../modules/cloud/amazon/_ec2_ami_search.py | 14 +- .../modules/cloud/amazon/_ec2_remote_facts.py | 76 ++--- lib/ansible/modules/cloud/amazon/_ec2_vpc.py | 66 ++-- lib/ansible/modules/cloud/amazon/aws_kms.py | 33 +- lib/ansible/modules/cloud/amazon/cloudformation.py | 28 +- .../modules/cloud/amazon/cloudfront_facts.py | 56 ++-- lib/ansible/modules/cloud/amazon/data_pipeline.py | 4 +- lib/ansible/modules/cloud/amazon/dynamodb_table.py | 9 +- lib/ansible/modules/cloud/amazon/ec2_ami_copy.py | 5 +- lib/ansible/modules/cloud/amazon/ec2_elb.py | 6 +- lib/ansible/modules/cloud/amazon/ec2_elb_lb.py | 36 ++- lib/ansible/modules/cloud/amazon/ec2_eni_facts.py | 10 +- lib/ansible/modules/cloud/amazon/ec2_lc.py | 2 +- lib/ansible/modules/cloud/amazon/ec2_lc_facts.py | 16 +- .../modules/cloud/amazon/ec2_metric_alarm.py | 49 +-- .../modules/cloud/amazon/ec2_scaling_policy.py | 24 +- lib/ansible/modules/cloud/amazon/ec2_snapshot.py | 30 +- .../modules/cloud/amazon/ec2_snapshot_facts.py | 4 +- lib/ansible/modules/cloud/amazon/ec2_tag.py | 14 +- lib/ansible/modules/cloud/amazon/ec2_vol.py | 46 +-- lib/ansible/modules/cloud/amazon/ec2_vol_facts.py | 7 +- .../modules/cloud/amazon/ec2_vpc_dhcp_option.py | 11 +- lib/ansible/modules/cloud/amazon/ec2_vpc_nacl.py | 12 +- lib/ansible/modules/cloud/amazon/ec2_vpc_net.py | 54 ++-- lib/ansible/modules/cloud/amazon/ec2_vpc_peer.py | 2 +- lib/ansible/modules/cloud/amazon/ec2_vpc_vgw.py | 23 +- .../modules/cloud/amazon/ec2_vpc_vgw_facts.py | 18 +- .../modules/cloud/amazon/ec2_win_password.py | 16 +- lib/ansible/modules/cloud/amazon/ecs_cluster.py | 29 +- lib/ansible/modules/cloud/amazon/ecs_service.py | 42 +-- .../modules/cloud/amazon/ecs_service_facts.py | 15 +- lib/ansible/modules/cloud/amazon/ecs_task.py | 33 +- .../modules/cloud/amazon/ecs_taskdefinition.py | 5 +- .../cloud/amazon/elasticache_subnet_group.py | 27 +- .../modules/cloud/amazon/elb_application_lb.py | 1 - lib/ansible/modules/cloud/amazon/elb_classic_lb.py | 36 ++- lib/ansible/modules/cloud/amazon/elb_instance.py | 6 +- lib/ansible/modules/cloud/amazon/execute_lambda.py | 32 +- lib/ansible/modules/cloud/amazon/iam.py | 19 +- lib/ansible/modules/cloud/amazon/iam_policy.py | 42 +-- .../modules/cloud/amazon/rds_subnet_group.py | 28 +- lib/ansible/modules/cloud/amazon/redshift.py | 94 +++--- .../modules/cloud/amazon/route53_health_check.py | 48 +-- lib/ansible/modules/cloud/amazon/s3_lifecycle.py | 37 +-- lib/ansible/modules/cloud/amazon/s3_logging.py | 8 +- lib/ansible/modules/cloud/amazon/s3_sync.py | 1 - lib/ansible/modules/cloud/amazon/s3_website.py | 8 +- lib/ansible/modules/cloud/amazon/sns_topic.py | 39 +-- .../modules/cloud/amazon/sts_assume_role.py | 15 +- .../modules/cloud/amazon/sts_session_token.py | 8 +- .../modules/cloud/azure/azure_rm_deployment.py | 20 +- .../modules/cloud/azure/azure_rm_managed_disk.py | 1 + .../cloud/azure/azure_rm_networkinterface.py | 13 +- .../cloud/azure/azure_rm_publicipaddress.py | 2 +- .../cloud/azure/azure_rm_publicipaddress_facts.py | 1 - .../modules/cloud/azure/azure_rm_storageaccount.py | 6 +- lib/ansible/modules/cloud/azure/azure_rm_subnet.py | 1 - .../modules/cloud/azure/azure_rm_virtualmachine.py | 29 +- .../modules/cloud/azure/azure_rm_virtualnetwork.py | 3 +- .../cloud/azure/azure_rm_virtualnetwork_facts.py | 1 + lib/ansible/modules/cloud/cloudstack/cs_host.py | 2 +- .../modules/cloud/docker/docker_container.py | 7 +- lib/ansible/modules/cloud/docker/docker_image.py | 2 +- .../modules/cloud/docker/docker_image_facts.py | 2 +- lib/ansible/modules/cloud/docker/docker_network.py | 21 +- lib/ansible/modules/cloud/google/gc_storage.py | 79 +++-- lib/ansible/modules/cloud/google/gcdns_record.py | 190 +++++------ lib/ansible/modules/cloud/google/gcdns_zone.py | 73 +++-- lib/ansible/modules/cloud/lxc/lxc_container.py | 10 +- lib/ansible/modules/cloud/lxd/lxd_container.py | 7 +- lib/ansible/modules/cloud/misc/rhevm.py | 139 ++++---- lib/ansible/modules/cloud/misc/serverless.py | 18 +- lib/ansible/modules/cloud/misc/virt_net.py | 99 +++--- lib/ansible/modules/cloud/misc/virt_pool.py | 130 ++++---- lib/ansible/modules/cloud/misc/xenserver_facts.py | 3 + lib/ansible/modules/cloud/openstack/os_image.py | 32 +- lib/ansible/modules/cloud/openstack/os_keypair.py | 10 +- .../modules/cloud/openstack/os_nova_flavor.py | 30 +- lib/ansible/modules/cloud/openstack/os_quota.py | 37 ++- .../cloud/openstack/os_security_group_rule.py | 32 +- lib/ansible/modules/cloud/openstack/os_server.py | 51 +-- lib/ansible/modules/cloud/openstack/os_stack.py | 19 +- .../cloud/ovh/ovh_ip_loadbalancing_backend.py | 2 +- lib/ansible/modules/cloud/packet/packet_sshkey.py | 28 +- .../modules/cloud/profitbricks/profitbricks.py | 16 +- .../cloud/profitbricks/profitbricks_datacenter.py | 8 +- .../modules/cloud/profitbricks/profitbricks_nic.py | 8 +- .../cloud/profitbricks/profitbricks_volume.py | 4 +- .../profitbricks_volume_attachments.py | 9 +- .../modules/cloud/rackspace/rax_cbs_attachments.py | 2 +- lib/ansible/modules/cloud/rackspace/rax_cdb.py | 2 +- lib/ansible/modules/cloud/rackspace/rax_clb.py | 2 +- lib/ansible/modules/cloud/rackspace/rax_clb_ssl.py | 6 +- lib/ansible/modules/cloud/rackspace/rax_dns.py | 2 +- .../modules/cloud/rackspace/rax_dns_record.py | 2 +- lib/ansible/modules/cloud/rackspace/rax_facts.py | 2 +- lib/ansible/modules/cloud/rackspace/rax_keypair.py | 2 +- .../modules/cloud/rackspace/rax_mon_alarm.py | 1 + .../modules/cloud/rackspace/rax_mon_check.py | 1 + .../modules/cloud/rackspace/rax_mon_entity.py | 1 + .../cloud/rackspace/rax_mon_notification.py | 1 + .../cloud/rackspace/rax_mon_notification_plan.py | 1 + .../modules/cloud/univention/udm_dns_record.py | 38 +-- .../modules/cloud/univention/udm_dns_zone.py | 102 +++--- lib/ansible/modules/cloud/univention/udm_group.py | 44 +-- lib/ansible/modules/cloud/univention/udm_share.py | 302 +++++++++--------- lib/ansible/modules/cloud/univention/udm_user.py | 262 +++++++-------- .../modules/cloud/vmware/vmware_dns_config.py | 4 +- .../modules/cloud/vmware/vmware_dvs_host.py | 6 +- lib/ansible/modules/cloud/vmware/vmware_guest.py | 3 +- .../modules/cloud/vmware/vmware_guest_find.py | 1 - .../modules/cloud/vmware/vmware_vmkernel.py | 18 +- .../cloud/vmware/vmware_vmkernel_ip_config.py | 4 +- lib/ansible/modules/cloud/vmware/vsphere_guest.py | 222 ++++++------- .../modules/cloud/webfaction/webfaction_app.py | 31 +- .../modules/cloud/webfaction/webfaction_db.py | 29 +- .../modules/cloud/webfaction/webfaction_domain.py | 22 +- .../modules/cloud/webfaction/webfaction_site.py | 36 +-- lib/ansible/modules/clustering/consul_acl.py | 5 + lib/ansible/modules/database/misc/kibana_plugin.py | 21 +- lib/ansible/modules/database/misc/riak.py | 19 +- .../modules/database/mongodb/mongodb_parameter.py | 2 +- .../modules/database/mongodb/mongodb_user.py | 16 +- lib/ansible/modules/database/mssql/mssql_db.py | 3 +- .../modules/database/postgresql/postgresql_ext.py | 17 +- .../modules/database/postgresql/postgresql_lang.py | 21 +- .../database/postgresql/postgresql_schema.py | 20 +- .../database/vertica/vertica_configuration.py | 11 +- .../modules/database/vertica/vertica_facts.py | 22 +- .../modules/database/vertica/vertica_role.py | 19 +- .../modules/database/vertica/vertica_schema.py | 21 +- .../modules/database/vertica/vertica_user.py | 31 +- lib/ansible/modules/monitoring/bigpanda.py | 2 +- lib/ansible/modules/monitoring/datadog_event.py | 2 +- lib/ansible/modules/monitoring/icinga2_feature.py | 2 +- .../modules/monitoring/librato_annotation.py | 23 +- lib/ansible/modules/monitoring/logentries.py | 16 +- lib/ansible/modules/monitoring/logicmonitor.py | 49 ++- .../modules/monitoring/logicmonitor_facts.py | 10 +- lib/ansible/modules/monitoring/nagios.py | 22 +- .../modules/monitoring/newrelic_deployment.py | 5 +- lib/ansible/modules/monitoring/pagerduty.py | 16 +- lib/ansible/modules/monitoring/pingdom.py | 4 +- lib/ansible/modules/monitoring/sensu_check.py | 48 +-- .../modules/monitoring/sensu_subscription.py | 6 +- lib/ansible/modules/monitoring/stackdriver.py | 4 +- lib/ansible/modules/monitoring/uptimerobot.py | 8 +- .../modules/monitoring/zabbix/zabbix_proxy.py | 4 +- lib/ansible/modules/net_tools/cloudflare_dns.py | 198 ++++++------ lib/ansible/modules/net_tools/dnsimple.py | 34 +- lib/ansible/modules/net_tools/dnsmadeeasy.py | 5 +- lib/ansible/modules/net_tools/ipinfoio_facts.py | 4 +- lib/ansible/modules/net_tools/nmcli.py | 355 ++++++++++----------- lib/ansible/modules/net_tools/omapi_host.py | 2 +- lib/ansible/modules/net_tools/snmp_facts.py | 45 +-- .../modules/network/a10/a10_server_axapi3.py | 7 +- .../modules/network/a10/a10_virtual_server.py | 3 +- lib/ansible/modules/network/aos/aos_asn_pool.py | 66 ++-- lib/ansible/modules/network/aos/aos_blueprint.py | 15 +- .../modules/network/aos/aos_blueprint_param.py | 20 +- .../modules/network/aos/aos_blueprint_virtnet.py | 22 +- lib/ansible/modules/network/aos/aos_device.py | 18 +- .../modules/network/aos/aos_external_router.py | 57 ++-- lib/ansible/modules/network/aos/aos_ip_pool.py | 54 ++-- .../modules/network/aos/aos_logical_device.py | 36 ++- .../modules/network/aos/aos_logical_device_map.py | 40 ++- lib/ansible/modules/network/aos/aos_rack_type.py | 40 ++- lib/ansible/modules/network/aos/aos_template.py | 38 ++- lib/ansible/modules/network/asa/asa_acl.py | 2 + lib/ansible/modules/network/asa/asa_command.py | 1 - lib/ansible/modules/network/asa/asa_config.py | 4 +- .../modules/network/bigswitch/bigmon_chain.py | 5 +- .../modules/network/bigswitch/bigmon_policy.py | 23 +- lib/ansible/modules/network/cnos/cnos_bgp.py | 2 +- lib/ansible/modules/network/cnos/cnos_command.py | 2 +- .../network/cnos/cnos_conditional_command.py | 2 +- lib/ansible/modules/network/cnos/cnos_image.py | 2 +- lib/ansible/modules/network/cnos/cnos_interface.py | 2 +- .../modules/network/cnos/cnos_portchannel.py | 2 +- lib/ansible/modules/network/cnos/cnos_rollback.py | 2 +- lib/ansible/modules/network/cnos/cnos_vlan.py | 2 +- lib/ansible/modules/network/cumulus/_cl_bond.py | 1 + .../modules/network/cumulus/_cl_img_install.py | 3 +- lib/ansible/modules/network/cumulus/_cl_license.py | 2 +- lib/ansible/modules/network/cumulus/_cl_ports.py | 5 +- lib/ansible/modules/network/cumulus/nclu.py | 18 +- lib/ansible/modules/network/eos/eos_banner.py | 3 + lib/ansible/modules/network/eos/eos_command.py | 4 + lib/ansible/modules/network/eos/eos_config.py | 1 - lib/ansible/modules/network/eos/eos_eapi.py | 14 +- lib/ansible/modules/network/eos/eos_facts.py | 4 + lib/ansible/modules/network/eos/eos_system.py | 17 +- lib/ansible/modules/network/eos/eos_user.py | 5 +- .../modules/network/f5/bigip_monitor_tcp.py | 6 +- lib/ansible/modules/network/f5/bigip_ucs.py | 1 + lib/ansible/modules/network/f5/bigip_wait.py | 1 + .../modules/network/fortios/fortios_config.py | 41 ++- .../modules/network/fortios/fortios_ipv4_policy.py | 46 +-- lib/ansible/modules/network/illumos/dladm_iptun.py | 1 - .../modules/network/illumos/dladm_linkprop.py | 1 - lib/ansible/modules/network/ios/ios_banner.py | 4 + lib/ansible/modules/network/ios/ios_command.py | 4 +- lib/ansible/modules/network/ios/ios_facts.py | 3 +- lib/ansible/modules/network/ios/ios_system.py | 15 +- lib/ansible/modules/network/ios/ios_vrf.py | 9 +- .../network/netscaler/netscaler_lb_vserver.py | 4 +- lib/ansible/modules/network/netvisor/pn_cluster.py | 4 +- .../modules/network/netvisor/pn_ospfarea.py | 4 +- lib/ansible/modules/network/netvisor/pn_vlag.py | 6 +- lib/ansible/modules/network/netvisor/pn_vlan.py | 6 +- lib/ansible/modules/network/netvisor/pn_vrouter.py | 4 +- .../modules/network/netvisor/pn_vrouterbgp.py | 2 +- .../modules/network/netvisor/pn_vrouterif.py | 4 +- .../modules/network/netvisor/pn_vrouterlbif.py | 4 +- lib/ansible/modules/network/nxos/_nxos_mtu.py | 1 + .../modules/network/nxos/nxos_aaa_server_host.py | 4 +- lib/ansible/modules/network/nxos/nxos_config.py | 7 +- lib/ansible/modules/network/nxos/nxos_gir.py | 42 +-- .../network/nxos/nxos_gir_profile_management.py | 5 +- lib/ansible/modules/network/nxos/nxos_igmp.py | 4 +- .../modules/network/nxos/nxos_igmp_interface.py | 12 +- .../modules/network/nxos/nxos_igmp_snooping.py | 4 +- .../modules/network/nxos/nxos_ip_interface.py | 14 +- lib/ansible/modules/network/nxos/nxos_logging.py | 2 +- lib/ansible/modules/network/nxos/nxos_ntp_auth.py | 5 +- .../modules/network/nxos/nxos_ntp_options.py | 3 +- lib/ansible/modules/network/nxos/nxos_nxapi.py | 12 +- .../modules/network/nxos/nxos_overlay_global.py | 6 +- lib/ansible/modules/network/nxos/nxos_ping.py | 6 +- lib/ansible/modules/network/nxos/nxos_smu.py | 3 +- lib/ansible/modules/network/nxos/nxos_snapshot.py | 3 +- .../modules/network/nxos/nxos_static_route.py | 2 +- lib/ansible/modules/network/nxos/nxos_system.py | 18 +- lib/ansible/modules/network/nxos/nxos_udld.py | 6 +- .../modules/network/nxos/nxos_udld_interface.py | 16 +- lib/ansible/modules/network/nxos/nxos_user.py | 19 +- lib/ansible/modules/network/nxos/nxos_vrf.py | 4 +- .../modules/network/nxos/nxos_vtp_domain.py | 2 +- .../modules/network/nxos/nxos_vtp_password.py | 4 +- .../modules/network/nxos/nxos_vtp_version.py | 2 +- .../modules/network/ordnance/ordnance_config.py | 9 +- .../modules/network/ordnance/ordnance_facts.py | 1 + .../modules/network/ovs/openvswitch_bridge.py | 8 +- .../modules/network/ovs/openvswitch_port.py | 11 +- lib/ansible/modules/network/panos/panos_admin.py | 2 +- lib/ansible/modules/network/panos/panos_admpwd.py | 10 +- .../modules/network/panos/panos_cert_gen_ssh.py | 12 +- lib/ansible/modules/network/panos/panos_check.py | 4 +- lib/ansible/modules/network/panos/panos_dag.py | 2 +- lib/ansible/modules/network/panos/panos_import.py | 4 +- .../modules/network/panos/panos_interface.py | 10 +- lib/ansible/modules/network/panos/panos_loadcfg.py | 2 +- .../modules/network/panos/panos_mgtconfig.py | 10 +- lib/ansible/modules/network/panos/panos_pg.py | 2 +- lib/ansible/modules/network/sros/sros_command.py | 3 +- lib/ansible/modules/network/sros/sros_config.py | 1 + lib/ansible/modules/network/sros/sros_rollback.py | 9 + lib/ansible/modules/network/vyos/vyos_command.py | 1 + lib/ansible/modules/network/vyos/vyos_system.py | 1 + lib/ansible/modules/notification/campfire.py | 8 +- lib/ansible/modules/notification/flowdock.py | 8 +- lib/ansible/modules/notification/grove.py | 16 +- lib/ansible/modules/notification/hall.py | 4 +- lib/ansible/modules/notification/irc.py | 10 +- lib/ansible/modules/notification/jabber.py | 12 +- lib/ansible/modules/notification/mattermost.py | 46 +-- lib/ansible/modules/notification/mqtt.py | 54 ++-- lib/ansible/modules/notification/osx_say.py | 4 +- lib/ansible/modules/notification/pushbullet.py | 32 +- lib/ansible/modules/notification/pushover.py | 10 +- lib/ansible/modules/notification/rocketchat.py | 30 +- lib/ansible/modules/notification/sendgrid.py | 25 +- lib/ansible/modules/notification/slack.py | 34 +- lib/ansible/modules/notification/twilio.py | 10 +- lib/ansible/modules/packaging/language/bundler.py | 6 +- lib/ansible/modules/packaging/language/cpanm.py | 21 +- lib/ansible/modules/packaging/language/gem.py | 50 +-- .../modules/packaging/language/maven_artifact.py | 37 ++- lib/ansible/modules/packaging/language/pear.py | 19 +- lib/ansible/modules/packaging/os/apk.py | 10 + .../modules/packaging/os/dpkg_selections.py | 1 + lib/ansible/modules/packaging/os/homebrew.py | 6 +- lib/ansible/modules/packaging/os/homebrew_cask.py | 8 +- lib/ansible/modules/packaging/os/layman.py | 16 +- lib/ansible/modules/packaging/os/macports.py | 9 +- lib/ansible/modules/packaging/os/opkg.py | 1 + lib/ansible/modules/packaging/os/pkgin.py | 39 ++- lib/ansible/modules/packaging/os/pkgng.py | 53 +-- lib/ansible/modules/packaging/os/pkgutil.py | 37 ++- lib/ansible/modules/packaging/os/portage.py | 4 +- lib/ansible/modules/packaging/os/portinstall.py | 14 +- lib/ansible/modules/packaging/os/slackpkg.py | 6 +- lib/ansible/modules/packaging/os/sorcery.py | 20 +- lib/ansible/modules/packaging/os/svr4pkg.py | 38 ++- lib/ansible/modules/packaging/os/swdepot.py | 24 +- lib/ansible/modules/packaging/os/zypper.py | 26 +- .../modules/packaging/os/zypper_repository.py | 18 +- .../modules/storage/infinidat/infini_export.py | 14 +- .../storage/infinidat/infini_export_client.py | 22 +- lib/ansible/modules/storage/infinidat/infini_fs.py | 14 +- .../modules/storage/infinidat/infini_host.py | 14 +- .../modules/storage/infinidat/infini_pool.py | 26 +- .../modules/storage/infinidat/infini_vol.py | 12 +- .../modules/storage/netapp/na_cdot_qtree.py | 2 +- .../modules/storage/netapp/na_cdot_volume.py | 2 +- lib/ansible/modules/storage/netapp/netapp_e_amg.py | 2 +- .../modules/storage/netapp/netapp_e_storagepool.py | 2 +- .../storage/netapp/sf_snapshot_schedule_manager.py | 4 +- .../netapp/sf_volume_access_group_manager.py | 2 +- .../modules/storage/netapp/sf_volume_manager.py | 6 +- lib/ansible/modules/system/cron.py | 1 + .../modules/web_infrastructure/letsencrypt.py | 3 + 313 files changed, 3462 insertions(+), 3134 deletions(-) (limited to 'lib/ansible') diff --git a/lib/ansible/modules/cloud/amazon/_ec2_ami_find.py b/lib/ansible/modules/cloud/amazon/_ec2_ami_find.py index fa67add809..324405647a 100644 --- a/lib/ansible/modules/cloud/amazon/_ec2_ami_find.py +++ b/lib/ansible/modules/cloud/amazon/_ec2_ami_find.py @@ -303,7 +303,7 @@ def get_block_device_mapping(image): """ bdm_dict = dict() - bdm = getattr(image,'block_device_mapping') + bdm = getattr(image, 'block_device_mapping') for device_name in bdm.keys(): bdm_dict[device_name] = { 'size': bdm[device_name].size, @@ -319,28 +319,28 @@ def get_block_device_mapping(image): def main(): argument_spec = ec2_argument_spec() argument_spec.update(dict( - owner = dict(required=False, default=None), - ami_id = dict(required=False), - ami_tags = dict(required=False, type='dict', - aliases = ['search_tags', 'image_tags']), - architecture = dict(required=False), - hypervisor = dict(required=False), - is_public = dict(required=False, type='bool'), - name = dict(required=False), - platform = dict(required=False), - product_code = dict(required=False), - sort = dict(required=False, default=None, - choices=['name', 'description', 'tag', 'architecture', 'block_device_mapping', 'creationDate', 'hypervisor', 'is_public', 'location', - 'owner_id', 'platform', 'root_device_name', 'root_device_type', 'state', 'virtualization_type']), - sort_tag = dict(required=False), - sort_order = dict(required=False, default='ascending', - choices=['ascending', 'descending']), - sort_start = dict(required=False), - sort_end = dict(required=False), - state = dict(required=False, default='available'), - virtualization_type = dict(required=False), - no_result_action = dict(required=False, default='success', - choices = ['success', 'fail']), + owner=dict(required=False, default=None), + ami_id=dict(required=False), + ami_tags=dict(required=False, type='dict', + aliases=['search_tags', 'image_tags']), + architecture=dict(required=False), + hypervisor=dict(required=False), + is_public=dict(required=False, type='bool'), + name=dict(required=False), + platform=dict(required=False), + product_code=dict(required=False), + sort=dict(required=False, default=None, + choices=['name', 'description', 'tag', 'architecture', 'block_device_mapping', 'creationDate', 'hypervisor', 'is_public', 'location', + 'owner_id', 'platform', 'root_device_name', 'root_device_type', 'state', 'virtualization_type']), + sort_tag=dict(required=False), + sort_order=dict(required=False, default='ascending', + choices=['ascending', 'descending']), + sort_start=dict(required=False), + sort_end=dict(required=False), + state=dict(required=False, default='available'), + virtualization_type=dict(required=False), + no_result_action=dict(required=False, default='success', + choices=['success', 'fail']), ) ) @@ -379,7 +379,7 @@ def main(): filter['image_id'] = ami_id if ami_tags: for tag in ami_tags: - filter['tag:'+tag] = ami_tags[tag] + filter['tag:' + tag] = ami_tags[tag] if architecture: filter['architecture'] = architecture if hypervisor: @@ -435,9 +435,9 @@ def main(): if sort == 'tag': if not sort_tag: module.fail_json(msg="'sort_tag' option must be given with 'sort=tag'") - results.sort(key=lambda e: e['tags'][sort_tag], reverse=(sort_order=='descending')) + results.sort(key=lambda e: e['tags'][sort_tag], reverse=(sort_order == 'descending')) elif sort: - results.sort(key=lambda e: e[sort], reverse=(sort_order=='descending')) + results.sort(key=lambda e: e[sort], reverse=(sort_order == 'descending')) try: if sort and sort_start and sort_end: diff --git a/lib/ansible/modules/cloud/amazon/_ec2_ami_search.py b/lib/ansible/modules/cloud/amazon/_ec2_ami_search.py index 3235b1b84c..d270e159c3 100644 --- a/lib/ansible/modules/cloud/amazon/_ec2_ami_search.py +++ b/lib/ansible/modules/cloud/amazon/_ec2_ami_search.py @@ -136,9 +136,9 @@ def ubuntu(module): reader = csv.reader(req, delimiter='\t') try: ami, aki, ari, tag, serial = lookup_ubuntu_ami(reader, release, stream, - store, arch, region, virt) + store, arch, region, virt) module.exit_json(changed=False, ami=ami, aki=aki, ari=ari, tag=tag, - serial=serial) + serial=serial) except KeyError: module.fail_json(msg="No matching AMI found") @@ -163,7 +163,7 @@ def lookup_ubuntu_ami(table, release, stream, store, arch, region, virt): actual_store, actual_arch, actual_region, ami, aki, ari, actual_virt) = row actual = (actual_release, actual_stream, actual_store, actual_arch, - actual_region, actual_virt) + actual_region, actual_virt) if actual == expected: # aki and ari are sometimes blank if aki == '': @@ -185,14 +185,14 @@ def main(): distro=dict(required=True, choices=SUPPORTED_DISTROS), release=dict(required=True), stream=dict(required=False, default='server', - choices=['desktop', 'server']), + choices=['desktop', 'server']), store=dict(required=False, default='ebs', - choices=['ebs', 'ebs-io1', 'ebs-ssd', 'instance-store']), + choices=['ebs', 'ebs-io1', 'ebs-ssd', 'instance-store']), arch=dict(required=False, default='amd64', - choices=['i386', 'amd64']), + choices=['i386', 'amd64']), region=dict(required=False, default='us-east-1', choices=AWS_REGIONS), virt=dict(required=False, default='paravirtual', - choices=['paravirtual', 'hvm']), + choices=['paravirtual', 'hvm']), ) module = AnsibleModule(argument_spec=arg_spec) distro = module.params['distro'] diff --git a/lib/ansible/modules/cloud/amazon/_ec2_remote_facts.py b/lib/ansible/modules/cloud/amazon/_ec2_remote_facts.py index f07cbe88cc..57e342e3d7 100644 --- a/lib/ansible/modules/cloud/amazon/_ec2_remote_facts.py +++ b/lib/ansible/modules/cloud/amazon/_ec2_remote_facts.py @@ -74,12 +74,12 @@ def get_instance_info(instance): # Get groups groups = [] for group in instance.groups: - groups.append({ 'id': group.id, 'name': group.name }.copy()) + groups.append({'id': group.id, 'name': group.name}.copy()) # Get interfaces interfaces = [] for interface in instance.interfaces: - interfaces.append({ 'id': interface.id, 'mac_address': interface.mac_address }.copy()) + interfaces.append({'id': interface.id, 'mac_address': interface.mac_address}.copy()) # If an instance is terminated, sourceDestCheck is no longer returned try: @@ -104,41 +104,41 @@ def get_instance_info(instance): instance_profile = dict(instance.instance_profile) if instance.instance_profile is not None else None - instance_info = { 'id': instance.id, - 'kernel': instance.kernel, - 'instance_profile': instance_profile, - 'root_device_type': instance.root_device_type, - 'private_dns_name': instance.private_dns_name, - 'public_dns_name': instance.public_dns_name, - 'ebs_optimized': instance.ebs_optimized, - 'client_token': instance.client_token, - 'virtualization_type': instance.virtualization_type, - 'architecture': instance.architecture, - 'ramdisk': instance.ramdisk, - 'tags': instance.tags, - 'key_name': instance.key_name, - 'source_destination_check': source_dest_check, - 'image_id': instance.image_id, - 'groups': groups, - 'interfaces': interfaces, - 'spot_instance_request_id': instance.spot_instance_request_id, - 'requester_id': instance.requester_id, - 'monitoring_state': instance.monitoring_state, - 'placement': { - 'tenancy': instance._placement.tenancy, - 'zone': instance._placement.zone - }, - 'ami_launch_index': instance.ami_launch_index, - 'launch_time': instance.launch_time, - 'hypervisor': instance.hypervisor, - 'region': instance.region.name, - 'persistent': instance.persistent, - 'private_ip_address': instance.private_ip_address, - 'public_ip_address': instance.ip_address, - 'state': instance._state.name, - 'vpc_id': instance.vpc_id, - 'block_device_mapping': bdm_dict, - } + instance_info = {'id': instance.id, + 'kernel': instance.kernel, + 'instance_profile': instance_profile, + 'root_device_type': instance.root_device_type, + 'private_dns_name': instance.private_dns_name, + 'public_dns_name': instance.public_dns_name, + 'ebs_optimized': instance.ebs_optimized, + 'client_token': instance.client_token, + 'virtualization_type': instance.virtualization_type, + 'architecture': instance.architecture, + 'ramdisk': instance.ramdisk, + 'tags': instance.tags, + 'key_name': instance.key_name, + 'source_destination_check': source_dest_check, + 'image_id': instance.image_id, + 'groups': groups, + 'interfaces': interfaces, + 'spot_instance_request_id': instance.spot_instance_request_id, + 'requester_id': instance.requester_id, + 'monitoring_state': instance.monitoring_state, + 'placement': { + 'tenancy': instance._placement.tenancy, + 'zone': instance._placement.zone + }, + 'ami_launch_index': instance.ami_launch_index, + 'launch_time': instance.launch_time, + 'hypervisor': instance.hypervisor, + 'region': instance.region.name, + 'persistent': instance.persistent, + 'private_ip_address': instance.private_ip_address, + 'public_ip_address': instance.ip_address, + 'state': instance._state.name, + 'vpc_id': instance.vpc_id, + 'block_device_mapping': bdm_dict, + } return instance_info @@ -163,7 +163,7 @@ def main(): argument_spec = ec2_argument_spec() argument_spec.update( dict( - filters = dict(default=None, type='dict') + filters=dict(default=None, type='dict') ) ) diff --git a/lib/ansible/modules/cloud/amazon/_ec2_vpc.py b/lib/ansible/modules/cloud/amazon/_ec2_vpc.py index ab6f3a71d1..2b1b88862e 100644 --- a/lib/ansible/modules/cloud/amazon/_ec2_vpc.py +++ b/lib/ansible/modules/cloud/amazon/_ec2_vpc.py @@ -189,6 +189,7 @@ def get_vpc_info(vpc): 'state': vpc.state, }) + def find_vpc(module, vpc_conn, vpc_id=None, cidr=None): """ Finds a VPC that matches a specific id or cidr + tags @@ -211,7 +212,7 @@ def find_vpc(module, vpc_conn, vpc_id=None, cidr=None): # Check for existing VPC by cidr_block or id if vpc_id is not None: - found_vpcs = vpc_conn.get_all_vpcs(None, {'vpc-id': vpc_id, 'state': 'available',}) + found_vpcs = vpc_conn.get_all_vpcs(None, {'vpc-id': vpc_id, 'state': 'available', }) else: previous_vpcs = vpc_conn.get_all_vpcs(None, {'cidr': cidr, 'state': 'available'}) @@ -234,8 +235,8 @@ def find_vpc(module, vpc_conn, vpc_id=None, cidr=None): return (found_vpc) -def routes_match(rt_list=None, rt=None, igw=None): +def routes_match(rt_list=None, rt=None, igw=None): """ Check if the route table has all routes as in given list @@ -284,6 +285,7 @@ def routes_match(rt_list=None, rt=None, igw=None): else: return True + def rtb_changed(route_tables=None, vpc_conn=None, module=None, vpc=None, igw=None): """ Checks if the remote routes match the local routes. @@ -299,7 +301,7 @@ def rtb_changed(route_tables=None, vpc_conn=None, module=None, vpc=None, igw=Non False when both routes and subnet associations matched. """ - #We add a one for the main table + # We add a one for the main table rtb_len = len(route_tables) + 1 remote_rtb_len = len(vpc_conn.get_all_route_tables(filters={'vpc_id': vpc.id})) if remote_rtb_len != rtb_len: @@ -307,13 +309,13 @@ def rtb_changed(route_tables=None, vpc_conn=None, module=None, vpc=None, igw=Non for rt in route_tables: rt_id = None for sn in rt['subnets']: - rsn = vpc_conn.get_all_subnets(filters={'cidr': sn, 'vpc_id': vpc.id }) + rsn = vpc_conn.get_all_subnets(filters={'cidr': sn, 'vpc_id': vpc.id}) if len(rsn) != 1: module.fail_json( - msg='The subnet {0} to associate with route_table {1} ' \ + msg='The subnet {0} to associate with route_table {1} ' 'does not exist, aborting'.format(sn, rt) ) - nrt = vpc_conn.get_all_route_tables(filters={'vpc_id': vpc.id, 'association.subnet-id': rsn[0].id}) + nrt = vpc_conn.get_all_route_tables(filters={'vpc_id': vpc.id, 'association.subnet-id': rsn[0].id}) if not nrt: return True else: @@ -388,10 +390,10 @@ def create_vpc(module, vpc_conn): time.sleep(5) if wait and wait_timeout <= time.time(): # waiting took too long - module.fail_json(msg = "wait for vpc availability timeout on %s" % time.asctime()) + module.fail_json(msg="wait for vpc availability timeout on %s" % time.asctime()) except boto.exception.BotoServerError as e: - module.fail_json(msg = "%s: %s" % (e.error_code, e.error_message)) + module.fail_json(msg="%s: %s" % (e.error_code, e.error_message)) # Done with base VPC, now change to attributes and features. @@ -408,7 +410,6 @@ def create_vpc(module, vpc_conn): if new_tags: vpc_conn.create_tags(vpc.id, new_tags) - # boto doesn't appear to have a way to determine the existing # value of the dns attributes, so we just set them. # It also must be done one at a time. @@ -420,7 +421,7 @@ def create_vpc(module, vpc_conn): if not isinstance(subnets, list): module.fail_json(msg='subnets needs to be a list of cidr blocks') - current_subnets = vpc_conn.get_all_subnets(filters={ 'vpc_id': vpc.id }) + current_subnets = vpc_conn.get_all_subnets(filters={'vpc_id': vpc.id}) # First add all new subnets for subnet in subnets: @@ -468,7 +469,7 @@ def create_vpc(module, vpc_conn): # to create tags results in exception. # boto doesn't seem to refresh 'state' of the newly created subnet, i.e.: it's always 'pending' # so i resorted to polling vpc_conn.get_all_subnets with the id of the newly added subnet - while len(vpc_conn.get_all_subnets(filters={ 'subnet-id': new_subnet.id })) == 0: + while len(vpc_conn.get_all_subnets(filters={'subnet-id': new_subnet.id})) == 0: time.sleep(0.1) vpc_conn.create_tags(new_subnet.id, new_subnet_tags) @@ -548,7 +549,7 @@ def create_vpc(module, vpc_conn): if route['gw'] == 'igw': if not internet_gateway: module.fail_json( - msg='You asked for an Internet Gateway ' \ + msg='You asked for an Internet Gateway ' '(igw) route, but you have no Internet Gateway' ) route_kwargs['gateway_id'] = igw.id @@ -564,10 +565,10 @@ def create_vpc(module, vpc_conn): # Associate with subnets for sn in rt['subnets']: - rsn = vpc_conn.get_all_subnets(filters={'cidr': sn, 'vpc_id': vpc.id }) + rsn = vpc_conn.get_all_subnets(filters={'cidr': sn, 'vpc_id': vpc.id}) if len(rsn) != 1: module.fail_json( - msg='The subnet {0} to associate with route_table {1} ' \ + msg='The subnet {0} to associate with route_table {1} ' 'does not exist, aborting'.format(sn, rt) ) rsn = rsn[0] @@ -576,7 +577,7 @@ def create_vpc(module, vpc_conn): old_rt = vpc_conn.get_all_route_tables( filters={'association.subnet_id': rsn.id, 'vpc_id': vpc.id} ) - old_rt = [ x for x in old_rt if x.id is not None ] + old_rt = [x for x in old_rt if x.id is not None] if len(old_rt) == 1: old_rt = old_rt[0] association_id = None @@ -591,7 +592,7 @@ def create_vpc(module, vpc_conn): changed = True except EC2ResponseError as e: module.fail_json( - msg='Unable to create and associate route table {0}, error: ' \ + msg='Unable to create and associate route table {0}, error: ' '{1}'.format(rt, e) ) @@ -625,7 +626,7 @@ def create_vpc(module, vpc_conn): created_vpc_id = vpc.id returned_subnets = [] - current_subnets = vpc_conn.get_all_subnets(filters={ 'vpc_id': vpc.id }) + current_subnets = vpc_conn.get_all_subnets(filters={'vpc_id': vpc.id}) for sn in current_subnets: returned_subnets.append({ @@ -647,6 +648,7 @@ def create_vpc(module, vpc_conn): return (vpc_dict, created_vpc_id, returned_subnets, igw_id, changed) + def terminate_vpc(module, vpc_conn, vpc_id=None, cidr=None): """ Terminates a VPC @@ -671,8 +673,8 @@ def terminate_vpc(module, vpc_conn, vpc_id=None, cidr=None): if vpc is not None: if vpc.state == 'available': - terminated_vpc_id=vpc.id - vpc_dict=get_vpc_info(vpc) + terminated_vpc_id = vpc.id + vpc_dict = get_vpc_info(vpc) try: subnets = vpc_conn.get_all_subnets(filters={'vpc_id': vpc.id}) for sn in subnets: @@ -709,18 +711,18 @@ def terminate_vpc(module, vpc_conn, vpc_id=None, cidr=None): def main(): argument_spec = ec2_argument_spec() argument_spec.update(dict( - cidr_block = dict(), - instance_tenancy = dict(choices=['default', 'dedicated'], default='default'), - wait = dict(type='bool', default=False), - wait_timeout = dict(default=300), - dns_support = dict(type='bool', default=True), - dns_hostnames = dict(type='bool', default=True), - subnets = dict(type='list'), - vpc_id = dict(), - internet_gateway = dict(type='bool', default=False), - resource_tags = dict(type='dict', required=True), - route_tables = dict(type='list'), - state = dict(choices=['present', 'absent'], default='present'), + cidr_block=dict(), + instance_tenancy=dict(choices=['default', 'dedicated'], default='default'), + wait=dict(type='bool', default=False), + wait_timeout=dict(default=300), + dns_support=dict(type='bool', default=True), + dns_hostnames=dict(type='bool', default=True), + subnets=dict(type='list'), + vpc_id=dict(), + internet_gateway=dict(type='bool', default=False), + resource_tags=dict(type='dict', required=True), + route_tables=dict(type='list'), + state=dict(choices=['present', 'absent'], default='present'), ) ) @@ -740,7 +742,7 @@ def main(): try: vpc_conn = connect_to_aws(boto.vpc, region, **aws_connect_kwargs) except boto.exception.NoAuthHandlerFound as e: - module.fail_json(msg = str(e)) + module.fail_json(msg=str(e)) else: module.fail_json(msg="region must be specified") diff --git a/lib/ansible/modules/cloud/amazon/aws_kms.py b/lib/ansible/modules/cloud/amazon/aws_kms.py index 1c6ebd21af..27750f60d7 100644 --- a/lib/ansible/modules/cloud/amazon/aws_kms.py +++ b/lib/ansible/modules/cloud/amazon/aws_kms.py @@ -120,6 +120,7 @@ try: except ImportError: HAS_BOTO3 = False + def get_arn_from_kms_alias(kms, aliasname): ret = kms.list_aliases() key_id = None @@ -138,12 +139,14 @@ def get_arn_from_kms_alias(kms, aliasname): return k['KeyArn'] raise Exception('could not find key from id: {}'.format(key_id)) + def get_arn_from_role_name(iam, rolename): ret = iam.get_role(RoleName=rolename) if ret.get('Role') and ret['Role'].get('Arn'): return ret['Role']['Arn'] raise Exception('could not find arn for name {}.'.format(rolename)) + def do_grant(kms, keyarn, role_arn, granttypes, mode='grant', dry_run=True, clean_invalid_entries=True): ret = {} keyret = kms.get_key_policy(KeyId=keyarn, PolicyName='default') @@ -179,10 +182,10 @@ def do_grant(kms, keyarn, role_arn, granttypes, mode='grant', dry_run=True, clea statement['Principal']['AWS'] = valid_entries had_invalid_entries = True - if not role_arn in statement['Principal']['AWS']: # needs to be added. + if not role_arn in statement['Principal']['AWS']: # needs to be added. changes_needed[granttype] = 'add' statement['Principal']['AWS'].append(role_arn) - elif role_arn in statement['Principal']['AWS']: # not one the places the role should be + elif role_arn in statement['Principal']['AWS']: # not one the places the role should be changes_needed[granttype] = 'remove' statement['Principal']['AWS'].remove(role_arn) @@ -210,6 +213,7 @@ def do_grant(kms, keyarn, role_arn, granttypes, mode='grant', dry_run=True, clea return ret + def assert_policy_shape(policy): '''Since the policy seems a little, uh, fragile, make sure we know approximately what we're looking at.''' errors = [] @@ -218,7 +222,7 @@ def assert_policy_shape(policy): found_statement_type = {} for statement in policy['Statement']: - for label,sidlabel in statement_label.items(): + for label, sidlabel in statement_label.items(): if statement['Sid'] == sidlabel: found_statement_type[label] = True @@ -230,16 +234,17 @@ def assert_policy_shape(policy): raise Exception('Problems asserting policy shape. Cowardly refusing to modify it: {}'.format(' '.join(errors))) return None + def main(): argument_spec = ansible.module_utils.ec2.ec2_argument_spec() argument_spec.update(dict( - mode = dict(choices=['grant', 'deny'], default='grant'), - key_alias = dict(required=False, type='str'), - key_arn = dict(required=False, type='str'), - role_name = dict(required=False, type='str'), - role_arn = dict(required=False, type='str'), - grant_types = dict(required=False, type='list'), - clean_invalid_entries = dict(type='bool', default=True), + mode=dict(choices=['grant', 'deny'], default='grant'), + key_alias=dict(required=False, type='str'), + key_arn=dict(required=False, type='str'), + role_name=dict(required=False, type='str'), + role_arn=dict(required=False, type='str'), + grant_types=dict(required=False, type='list'), + clean_invalid_entries=dict(type='bool', default=True), ) ) @@ -255,7 +260,6 @@ def main(): result = {} mode = module.params['mode'] - try: region, ec2_url, aws_connect_kwargs = ansible.module_utils.ec2.get_aws_connection_info(module, boto3=True) kms = ansible.module_utils.ec2.boto3_conn(module, conn_type='client', resource='kms', region=region, endpoint=ec2_url, **aws_connect_kwargs) @@ -263,7 +267,6 @@ def main(): except botocore.exceptions.NoCredentialsError as e: module.fail_json(msg='cannot connect to AWS', exception=traceback.format_exc()) - try: if module.params['key_alias'] and not module.params['key_arn']: module.params['key_arn'] = get_arn_from_kms_alias(kms, module.params['key_alias']) @@ -282,9 +285,9 @@ def main(): module.fail_json(msg='{} is an unknown grant type.'.format(g)) ret = do_grant(kms, module.params['key_arn'], module.params['role_arn'], module.params['grant_types'], - mode=mode, - dry_run=module.check_mode, - clean_invalid_entries=module.params['clean_invalid_entries']) + mode=mode, + dry_run=module.check_mode, + clean_invalid_entries=module.params['clean_invalid_entries']) result.update(ret) except Exception as err: diff --git a/lib/ansible/modules/cloud/amazon/cloudformation.py b/lib/ansible/modules/cloud/amazon/cloudformation.py index 87f6281688..920a4d7aaa 100644 --- a/lib/ansible/modules/cloud/amazon/cloudformation.py +++ b/lib/ansible/modules/cloud/amazon/cloudformation.py @@ -263,7 +263,7 @@ from ansible.module_utils._text import to_bytes, to_native def get_stack_events(cfn, stack_name, token_filter=None): '''This event data was never correct, it worked as a side effect. So the v2.3 format is different.''' - ret = {'events':[], 'log':[]} + ret = {'events': [], 'log': []} try: pg = cfn.get_paginator( @@ -348,8 +348,8 @@ def create_changeset(module, stack_params, cfn): cs = cfn.create_change_set(**stack_params) result = stack_operation(cfn, stack_params['StackName'], 'CREATE_CHANGESET') result['warnings'] = ['Created changeset named %s for stack %s' % (changeset_name, stack_params['StackName']), - 'You can execute it using: aws cloudformation execute-change-set --change-set-name %s' % cs['Id'], - 'NOTE that dependencies on this stack might fail due to pending changes!'] + 'You can execute it using: aws cloudformation execute-change-set --change-set-name %s' % cs['Id'], + 'NOTE that dependencies on this stack might fail due to pending changes!'] except Exception as err: error_msg = boto_exception(err) if 'No updates are to be performed.' in error_msg: @@ -413,7 +413,7 @@ def stack_operation(cfn, stack_name, operation, op_token=None): except: # If the stack previously existed, and now can't be found then it's # been deleted successfully. - if 'yes' in existed or operation == 'DELETE': # stacks may delete fast, look in a few ways. + if 'yes' in existed or operation == 'DELETE': # stacks may delete fast, look in a few ways. ret = get_stack_events(cfn, stack_name, op_token) ret.update({'changed': True, 'output': 'Stack Deleted'}) return ret @@ -421,12 +421,12 @@ def stack_operation(cfn, stack_name, operation, op_token=None): return {'changed': True, 'failed': True, 'output': 'Stack Not Found', 'exception': traceback.format_exc()} ret = get_stack_events(cfn, stack_name, op_token) if not stack: - if 'yes' in existed or operation == 'DELETE': # stacks may delete fast, look in a few ways. + if 'yes' in existed or operation == 'DELETE': # stacks may delete fast, look in a few ways. ret = get_stack_events(cfn, stack_name, op_token) ret.update({'changed': True, 'output': 'Stack Deleted'}) return ret else: - ret.update({'changed': False, 'failed': True, 'output' : 'Stack not found.'}) + ret.update({'changed': False, 'failed': True, 'output': 'Stack not found.'}) return ret # it covers ROLLBACK_COMPLETE and UPDATE_ROLLBACK_COMPLETE # Possible states: https://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/using-cfn-describing-stacks.html#w1ab2c15c17c21c13 @@ -435,7 +435,7 @@ def stack_operation(cfn, stack_name, operation, op_token=None): return ret # note the ordering of ROLLBACK_COMPLETE and COMPLETE, because otherwise COMPLETE will match both cases. elif stack['StackStatus'].endswith('_COMPLETE'): - ret.update({'changed': True, 'output' : 'Stack %s complete' % operation }) + ret.update({'changed': True, 'output': 'Stack %s complete' % operation}) return ret elif stack['StackStatus'].endswith('_ROLLBACK_FAILED'): ret.update({'changed': True, 'failed': True, 'output': 'Stack %s rollback failed' % operation}) @@ -447,7 +447,7 @@ def stack_operation(cfn, stack_name, operation, op_token=None): else: # this can loop forever :/ time.sleep(5) - return {'failed': True, 'output':'Failed for unknown reasons.'} + return {'failed': True, 'output': 'Failed for unknown reasons.'} def build_changeset_name(stack_params): @@ -470,7 +470,7 @@ def check_mode_changeset(module, stack_params, cfn): try: change_set = cfn.create_change_set(**stack_params) - for i in range(60): # total time 5 min + for i in range(60): # total time 5 min description = cfn.describe_change_set(ChangeSetName=change_set['Id']) if description['Status'] in ('CREATE_COMPLETE', 'FAILED'): break @@ -496,7 +496,7 @@ def get_stack_facts(cfn, stack_name): try: stack_response = cfn.describe_stacks(StackName=stack_name) stack_info = stack_response['Stacks'][0] - except (botocore.exceptions.ValidationError,botocore.exceptions.ClientError) as err: + except (botocore.exceptions.ValidationError, botocore.exceptions.ClientError) as err: error_msg = boto_exception(err) if 'does not exist' in error_msg: # missing stack, don't bail. @@ -567,7 +567,7 @@ def main(): stack_params['StackPolicyBody'] = open(module.params['stack_policy'], 'r').read() template_parameters = module.params['template_parameters'] - stack_params['Parameters'] = [{'ParameterKey':k, 'ParameterValue':str(v)} for k, v in template_parameters.items()] + stack_params['Parameters'] = [{'ParameterKey': k, 'ParameterValue': str(v)} for k, v in template_parameters.items()] if isinstance(module.params.get('tags'), dict): stack_params['Tags'] = ansible.module_utils.ec2.ansible_dict_to_boto3_tag_list(module.params['tags']) @@ -637,7 +637,7 @@ def main(): "resource_type": res['ResourceType'], "last_updated_time": res['LastUpdatedTimestamp'], "status": res['ResourceStatus'], - "status_reason": res.get('ResourceStatusReason') # can be blank, apparently + "status_reason": res.get('ResourceStatusReason') # can be blank, apparently }) result['stack_resources'] = stack_resources @@ -658,8 +658,8 @@ def main(): if module.params['template_format'] is not None: result['warnings'] = [('Argument `template_format` is deprecated ' - 'since Ansible 2.3, JSON and YAML templates are now passed ' - 'directly to the CloudFormation API.')] + 'since Ansible 2.3, JSON and YAML templates are now passed ' + 'directly to the CloudFormation API.')] module.exit_json(**result) diff --git a/lib/ansible/modules/cloud/amazon/cloudfront_facts.py b/lib/ansible/modules/cloud/amazon/cloudfront_facts.py index 3bc68f719d..9b0f97c732 100644 --- a/lib/ansible/modules/cloud/amazon/cloudfront_facts.py +++ b/lib/ansible/modules/cloud/amazon/cloudfront_facts.py @@ -259,7 +259,7 @@ class CloudFrontServiceManager: def get_distribution(self, distribution_id): try: - func = partial(self.client.get_distribution,Id=distribution_id) + func = partial(self.client.get_distribution, Id=distribution_id) return self.paginated_response(func) except botocore.exceptions.ClientError as e: self.module.fail_json(msg="Error describing distribution - " + str(e), @@ -268,7 +268,7 @@ class CloudFrontServiceManager: def get_distribution_config(self, distribution_id): try: - func = partial(self.client.get_distribution_config,Id=distribution_id) + func = partial(self.client.get_distribution_config, Id=distribution_id) return self.paginated_response(func) except botocore.exceptions.ClientError as e: self.module.fail_json(msg="Error describing distribution configuration - " + str(e), @@ -277,7 +277,7 @@ class CloudFrontServiceManager: def get_origin_access_identity(self, origin_access_identity_id): try: - func = partial(self.client.get_cloud_front_origin_access_identity,Id=origin_access_identity_id) + func = partial(self.client.get_cloud_front_origin_access_identity, Id=origin_access_identity_id) return self.paginated_response(func) except botocore.exceptions.ClientError as e: self.module.fail_json(msg="Error describing origin access identity - " + str(e), @@ -286,7 +286,7 @@ class CloudFrontServiceManager: def get_origin_access_identity_config(self, origin_access_identity_id): try: - func = partial(self.client.get_cloud_front_origin_access_identity_config,Id=origin_access_identity_id) + func = partial(self.client.get_cloud_front_origin_access_identity_config, Id=origin_access_identity_id) return self.paginated_response(func) except botocore.exceptions.ClientError as e: self.module.fail_json(msg="Error describing origin access identity configuration - " + str(e), @@ -295,7 +295,7 @@ class CloudFrontServiceManager: def get_invalidation(self, distribution_id, invalidation_id): try: - func = partial(self.client.get_invalidation,DistributionId=distribution_id,Id=invalidation_id) + func = partial(self.client.get_invalidation, DistributionId=distribution_id, Id=invalidation_id) return self.paginated_response(func) except botocore.exceptions.ClientError as e: self.module.fail_json(msg="Error describing invalidation - " + str(e), @@ -304,7 +304,7 @@ class CloudFrontServiceManager: def get_streaming_distribution(self, distribution_id): try: - func = partial(self.client.get_streaming_distribution,Id=distribution_id) + func = partial(self.client.get_streaming_distribution, Id=distribution_id) return self.paginated_response(func) except botocore.exceptions.ClientError as e: self.module.fail_json(msg="Error describing streaming distribution - " + str(e), @@ -313,7 +313,7 @@ class CloudFrontServiceManager: def get_streaming_distribution_config(self, distribution_id): try: - func = partial(self.client.get_streaming_distribution_config,Id=distribution_id) + func = partial(self.client.get_streaming_distribution_config, Id=distribution_id) return self.paginated_response(func) except botocore.exceptions.ClientError as e: self.module.fail_json(msg="Error describing streaming distribution - " + str(e), @@ -399,13 +399,13 @@ class CloudFrontServiceManager: def summary_get_origin_access_identity_list(self): try: - origin_access_identity_list = { 'origin_access_identities': [] } + origin_access_identity_list = {'origin_access_identities': []} origin_access_identities = self.list_origin_access_identities() for origin_access_identity in origin_access_identities: oai_id = origin_access_identity['Id'] oai_full_response = self.get_origin_access_identity(oai_id) - oai_summary = { 'Id': oai_id, 'ETag': oai_full_response['ETag'] } - origin_access_identity_list['origin_access_identities'].append( oai_summary ) + oai_summary = {'Id': oai_id, 'ETag': oai_full_response['ETag']} + origin_access_identity_list['origin_access_identities'].append(oai_summary) return origin_access_identity_list except botocore.exceptions.ClientError as e: self.module.fail_json(msg="Error generating summary of origin access identities - " + str(e), @@ -415,8 +415,8 @@ class CloudFrontServiceManager: def summary_get_distribution_list(self, streaming=False): try: list_name = 'streaming_distributions' if streaming else 'distributions' - key_list = ['Id', 'ARN', 'Status', 'LastModifiedTime', 'DomainName', 'Comment', 'PriceClass', 'Enabled' ] - distribution_list = { list_name: [] } + key_list = ['Id', 'ARN', 'Status', 'LastModifiedTime', 'DomainName', 'Comment', 'PriceClass', 'Enabled'] + distribution_list = {list_name: []} distributions = self.list_streaming_distributions(False) if streaming else self.list_distributions(False) for dist in distributions: temp_distribution = {} @@ -520,16 +520,18 @@ class CloudFrontServiceManager: if 'Items' in item['Aliases']: aliases = item['Aliases']['Items'] for alias in aliases: - keyed_list.update( { alias: item } ) - keyed_list.update( { distribution_id: item } ) + keyed_list.update({alias: item}) + keyed_list.update({distribution_id: item}) return keyed_list + def set_facts_for_distribution_id_and_alias(details, facts, distribution_id, aliases): facts[distribution_id].update(details) for alias in aliases: facts[alias].update(details) return facts + def main(): argument_spec = ec2_argument_spec() argument_spec.update(dict( @@ -581,17 +583,17 @@ def main(): summary = module.params.get('summary') aliases = [] - result = { 'cloudfront': {} } + result = {'cloudfront': {}} facts = {} require_distribution_id = (distribution or distribution_config or invalidation or streaming_distribution or - streaming_distribution_config or list_invalidations) + streaming_distribution_config or list_invalidations) # set default to summary if no option specified summary = summary or not (distribution or distribution_config or origin_access_identity or - origin_access_identity_config or invalidation or streaming_distribution or streaming_distribution_config or - list_origin_access_identities or list_distributions_by_web_acl_id or list_invalidations or - list_streaming_distributions or list_distributions) + origin_access_identity_config or invalidation or streaming_distribution or streaming_distribution_config or + list_origin_access_identities or list_distributions_by_web_acl_id or list_invalidations or + list_streaming_distributions or list_distributions) # validations if require_distribution_id and distribution_id is None and domain_name_alias is None: @@ -611,21 +613,21 @@ def main(): # set appropriate cloudfront id if distribution_id and not list_invalidations: - facts = { distribution_id: {} } + facts = {distribution_id: {}} aliases = service_mgr.get_aliases_from_distribution_id(distribution_id) for alias in aliases: - facts.update( { alias: {} } ) + facts.update({alias: {}}) if invalidation_id: - facts.update( { invalidation_id: {} } ) + facts.update({invalidation_id: {}}) elif distribution_id and list_invalidations: - facts = { distribution_id: {} } + facts = {distribution_id: {}} aliases = service_mgr.get_aliases_from_distribution_id(distribution_id) for alias in aliases: - facts.update( { alias: {} } ) + facts.update({alias: {}}) elif origin_access_identity_id: - facts = { origin_access_identity_id: {} } + facts = {origin_access_identity_id: {}} elif web_acl_id: - facts = { web_acl_id: {} } + facts = {web_acl_id: {}} # get details based on options if distribution: @@ -644,7 +646,7 @@ def main(): if streaming_distribution_config: facts_to_set = service_mgr.get_streaming_distribution_config(distribution_id) if list_invalidations: - facts_to_set = {'invalidations': service_mgr.list_invalidations(distribution_id) } + facts_to_set = {'invalidations': service_mgr.list_invalidations(distribution_id)} if 'facts_to_set' in vars(): facts = set_facts_for_distribution_id_and_alias(facts_to_set, facts, distribution_id, aliases) diff --git a/lib/ansible/modules/cloud/amazon/data_pipeline.py b/lib/ansible/modules/cloud/amazon/data_pipeline.py index f17e092548..168623cf41 100644 --- a/lib/ansible/modules/cloud/amazon/data_pipeline.py +++ b/lib/ansible/modules/cloud/amazon/data_pipeline.py @@ -332,7 +332,7 @@ def activate_pipeline(client, module): pass else: module.fail_json(msg=('Data Pipeline {0} failed to activate ' - 'within timeout {1} seconds').format(dp_name, timeout)) + 'within timeout {1} seconds').format(dp_name, timeout)) changed = True data_pipeline = get_result(client, dp_id) @@ -477,7 +477,7 @@ def diff_pipeline(client, module, objects, unique_id, dp_name): result = {'data_pipeline': data_pipeline, 'msg': msg} except DataPipelineNotFound: - create_dp = True + create_dp = True return create_dp, changed, result diff --git a/lib/ansible/modules/cloud/amazon/dynamodb_table.py b/lib/ansible/modules/cloud/amazon/dynamodb_table.py index 58386ff047..838af01878 100644 --- a/lib/ansible/modules/cloud/amazon/dynamodb_table.py +++ b/lib/ansible/modules/cloud/amazon/dynamodb_table.py @@ -223,7 +223,6 @@ def create_or_update_dynamo_table(connection, module, boto3_dynamodb=None, boto3 try: table = Table(table_name, connection=connection) - if dynamo_table_exists(table): result['changed'] = update_dynamo_table(table, throughput=throughput, check_mode=module.check_mode, global_indexes=global_indexes) else: @@ -349,7 +348,7 @@ def has_throughput_changed(table, new_throughput): return False return new_throughput['read'] != table.throughput['read'] or \ - new_throughput['write'] != table.throughput['write'] + new_throughput['write'] != table.throughput['write'] def get_schema_param(hash_key_name, hash_key_type, range_key_name, range_key_type): @@ -397,6 +396,7 @@ def validate_index(index, module): if index['type'] not in INDEX_TYPE_OPTIONS: module.fail_json(msg='%s is not a valid index type, must be one of %s' % (index['type'], INDEX_TYPE_OPTIONS)) + def get_indexes(all_indexes): indexes = [] global_indexes = [] @@ -429,7 +429,6 @@ def get_indexes(all_indexes): return indexes, global_indexes - def main(): argument_spec = ec2_argument_spec() argument_spec.update(dict( @@ -442,8 +441,8 @@ def main(): read_capacity=dict(default=1, type='int'), write_capacity=dict(default=1, type='int'), indexes=dict(default=[], type='list'), - tags = dict(type='dict'), - wait_for_active_timeout = dict(default=60, type='int'), + tags=dict(type='dict'), + wait_for_active_timeout=dict(default=60, type='int'), )) module = AnsibleModule( diff --git a/lib/ansible/modules/cloud/amazon/ec2_ami_copy.py b/lib/ansible/modules/cloud/amazon/ec2_ami_copy.py index 91e8b98e72..e753b369b9 100644 --- a/lib/ansible/modules/cloud/amazon/ec2_ami_copy.py +++ b/lib/ansible/modules/cloud/amazon/ec2_ami_copy.py @@ -158,7 +158,6 @@ except ImportError: HAS_BOTO3 = False - def copy_image(module, ec2): """ Copies an AMI @@ -185,8 +184,8 @@ def copy_image(module, ec2): if module.params.get('tags'): ec2.create_tags( Resources=[image_id], - Tags=[{'Key' : k, 'Value': v} for k,v in module.params.get('tags').items()] - ) + Tags=[{'Key': k, 'Value': v} for k, v in module.params.get('tags').items()] + ) module.exit_json(changed=True, image_id=image_id) except WaiterError as we: diff --git a/lib/ansible/modules/cloud/amazon/ec2_elb.py b/lib/ansible/modules/cloud/amazon/ec2_elb.py index 56e8adcdd6..7133c1f4bf 100644 --- a/lib/ansible/modules/cloud/amazon/ec2_elb.py +++ b/lib/ansible/modules/cloud/amazon/ec2_elb.py @@ -171,7 +171,7 @@ class ElbManager: found = False for lb in self.lbs: if lb.name == lbtest: - found=True + found = True break return found @@ -330,7 +330,7 @@ def main(): argument_spec.update(dict( state={'required': True}, instance_id={'required': True}, - ec2_elbs={'default': None, 'required': False, 'type':'list'}, + ec2_elbs={'default': None, 'required': False, 'type': 'list'}, enable_availability_zone={'default': True, 'required': False, 'type': 'bool'}, wait={'required': False, 'default': True, 'type': 'bool'}, wait_timeout={'required': False, 'default': 0, 'type': 'int'} @@ -363,7 +363,7 @@ def main(): if ec2_elbs is not None: for elb in ec2_elbs: if not elb_man.exists(elb): - msg="ELB %s does not exist" % elb + msg = "ELB %s does not exist" % elb module.fail_json(msg=msg) if module.params['state'] == 'present': diff --git a/lib/ansible/modules/cloud/amazon/ec2_elb_lb.py b/lib/ansible/modules/cloud/amazon/ec2_elb_lb.py index c983a602f5..116a9898bf 100644 --- a/lib/ansible/modules/cloud/amazon/ec2_elb_lb.py +++ b/lib/ansible/modules/cloud/amazon/ec2_elb_lb.py @@ -425,6 +425,7 @@ def _throttleable_operation(max_retries): return _do_op return _operation_wrapper + def _get_vpc_connection(module, region, aws_connect_params): try: return connect_to_aws(boto.vpc, region, **aws_connect_params) @@ -434,6 +435,7 @@ def _get_vpc_connection(module, region, aws_connect_params): _THROTTLING_RETRIES = 5 + class ElbManager(object): """Handles ELB creation and destruction""" @@ -579,10 +581,10 @@ class ElbManager(object): # status of instances behind the ELB if info['instances']: - info['instance_health'] = [ dict( - instance_id = instance_state.instance_id, - reason_code = instance_state.reason_code, - state = instance_state.state + info['instance_health'] = [dict( + instance_id=instance_state.instance_id, + reason_code=instance_state.reason_code, + state=instance_state.state ) for instance_state in self.elb_conn.describe_instance_health(self.name)] else: info['instance_health'] = [] @@ -663,7 +665,7 @@ class ElbManager(object): elb_interfaces = self.ec2_conn.get_all_network_interfaces( filters={'attachment.instance-owner-id': 'amazon-elb', - 'description': 'ELB {0}'.format(self.name) }) + 'description': 'ELB {0}'.format(self.name)}) for x in range(0, max_retries): for interface in elb_interfaces: @@ -888,13 +890,13 @@ class ElbManager(object): if self.zones: if self.purge_zones: zones_to_disable = list(set(self.elb.availability_zones) - - set(self.zones)) + set(self.zones)) zones_to_enable = list(set(self.zones) - - set(self.elb.availability_zones)) + set(self.elb.availability_zones)) else: zones_to_disable = None zones_to_enable = list(set(self.zones) - - set(self.elb.availability_zones)) + set(self.elb.availability_zones)) if zones_to_enable: self._enable_zones(zones_to_enable) # N.B. This must come second, in case it would have removed all zones @@ -962,7 +964,7 @@ class ElbManager(object): "enabled": True, "s3_bucket_name": self.access_logs['s3_location'], "s3_bucket_prefix": self.access_logs.get('s3_prefix', ''), - "emit_interval": self.access_logs.get('interval', 60), + "emit_interval": self.access_logs.get('interval', 60), } update_access_logs_config = False @@ -1002,10 +1004,10 @@ class ElbManager(object): self.elb_conn.modify_lb_attribute(self.name, 'ConnectingSettings', attributes.connecting_settings) def _policy_name(self, policy_type): - return __file__.split('/')[-1].split('.')[0].replace('_', '-') + '-' + policy_type + return __file__.split('/')[-1].split('.')[0].replace('_', '-') + '-' + policy_type def _create_policy(self, policy_param, policy_meth, policy): - getattr(self.elb_conn, policy_meth )(policy_param, self.elb.name, policy) + getattr(self.elb_conn, policy_meth)(policy_param, self.elb.name, policy) def _delete_policy(self, elb_name, policy): self.elb_conn.delete_lb_policy(elb_name, policy) @@ -1223,7 +1225,7 @@ class ElbManager(object): params['Tags.member.%d.Value' % (i + 1)] = dictact[key] self.elb_conn.make_request('AddTags', params) - self.changed=True + self.changed = True # Remove extra tags dictact = dict(set(tagdict.items()) - set(self.tags.items())) @@ -1232,7 +1234,7 @@ class ElbManager(object): params['Tags.member.%d.Key' % (i + 1)] = key self.elb_conn.make_request('RemoveTags', params) - self.changed=True + self.changed = True def _get_health_check_target(self): """Compose target string from healthcheck parameters""" @@ -1275,7 +1277,7 @@ def main(): module = AnsibleModule( argument_spec=argument_spec, - mutually_exclusive = [['security_group_ids', 'security_group_names']] + mutually_exclusive=[['security_group_ids', 'security_group_names']] ) if not HAS_BOTO: @@ -1321,7 +1323,7 @@ def main(): security_group_ids = [] try: ec2 = connect_to_aws(boto.ec2, region, **aws_connect_params) - if subnets: # We have at least one subnet, ergo this is a VPC + if subnets: # We have at least one subnet, ergo this is a VPC vpc_conn = _get_vpc_connection(module=module, region=region, aws_connect_params=aws_connect_params) vpc_id = vpc_conn.get_all_subnets([subnets[0]])[0].vpc_id filters = {'vpc_id': vpc_id} @@ -1333,10 +1335,10 @@ def main(): if isinstance(group_name, string_types): group_name = [group_name] - group_id = [ str(grp.id) for grp in grp_details if str(grp.name) in group_name ] + group_id = [str(grp.id) for grp in grp_details if str(grp.name) in group_name] security_group_ids.extend(group_id) except boto.exception.NoAuthHandlerFound as e: - module.fail_json(msg = str(e)) + module.fail_json(msg=str(e)) elb_man = ElbManager(module, name, listeners, purge_listeners, zones, purge_zones, security_group_ids, health_check, diff --git a/lib/ansible/modules/cloud/amazon/ec2_eni_facts.py b/lib/ansible/modules/cloud/amazon/ec2_eni_facts.py index e3d86e2a3d..adf80cfac9 100644 --- a/lib/ansible/modules/cloud/amazon/ec2_eni_facts.py +++ b/lib/ansible/modules/cloud/amazon/ec2_eni_facts.py @@ -68,9 +68,9 @@ except ImportError: from ansible.module_utils.basic import AnsibleModule from ansible.module_utils.ec2 import (AnsibleAWSError, - ansible_dict_to_boto3_filter_list, boto3_conn, - boto3_tag_list_to_ansible_dict, camel_dict_to_snake_dict, - connect_to_aws, ec2_argument_spec, get_aws_connection_info) + ansible_dict_to_boto3_filter_list, boto3_conn, + boto3_tag_list_to_ansible_dict, camel_dict_to_snake_dict, + connect_to_aws, ec2_argument_spec, get_aws_connection_info) def list_ec2_eni_boto3(connection, module): @@ -99,7 +99,7 @@ def get_eni_info(interface): # Private addresses private_addresses = [] for ip in interface.private_ip_addresses: - private_addresses.append({ 'private_ip_address': ip.private_ip_address, 'primary_address': ip.primary }) + private_addresses.append({'private_ip_address': ip.private_ip_address, 'primary_address': ip.primary}) interface_info = {'id': interface.id, 'subnet_id': interface.subnet_id, @@ -152,7 +152,7 @@ def main(): argument_spec = ec2_argument_spec() argument_spec.update( dict( - filters = dict(default=None, type='dict') + filters=dict(default=None, type='dict') ) ) diff --git a/lib/ansible/modules/cloud/amazon/ec2_lc.py b/lib/ansible/modules/cloud/amazon/ec2_lc.py index 74787d1a74..67875c4434 100755 --- a/lib/ansible/modules/cloud/amazon/ec2_lc.py +++ b/lib/ansible/modules/cloud/amazon/ec2_lc.py @@ -339,7 +339,7 @@ def create_launch_config(connection, module): module.fail_json(msg="Failed to create launch configuration", exception=traceback.format_exc(), **camel_dict_to_snake_dict(e.response)) result = (dict((k, v) for k, v in launch_config.items() - if k not in ['Connection', 'CreatedTime', 'InstanceMonitoring', 'BlockDeviceMappings'])) + if k not in ['Connection', 'CreatedTime', 'InstanceMonitoring', 'BlockDeviceMappings'])) result['CreatedTime'] = to_text(launch_config.get('CreatedTime')) diff --git a/lib/ansible/modules/cloud/amazon/ec2_lc_facts.py b/lib/ansible/modules/cloud/amazon/ec2_lc_facts.py index 50ad935fe8..bff65108e2 100644 --- a/lib/ansible/modules/cloud/amazon/ec2_lc_facts.py +++ b/lib/ansible/modules/cloud/amazon/ec2_lc_facts.py @@ -191,7 +191,7 @@ def list_launch_configs(connection, module): launch_config['CreatedTime'] = str(launch_config['CreatedTime']) if sort: - snaked_launch_configs.sort(key=lambda e: e[sort], reverse=(sort_order=='descending')) + snaked_launch_configs.sort(key=lambda e: e[sort], reverse=(sort_order == 'descending')) try: if sort and sort_start and sort_end: @@ -210,13 +210,13 @@ def main(): argument_spec = ec2_argument_spec() argument_spec.update( dict( - name = dict(required=False, default=[], type='list'), - sort = dict(required=False, default=None, - choices=['launch_configuration_name', 'image_id', 'created_time', 'instance_type', 'kernel_id', 'ramdisk_id', 'key_name']), - sort_order = dict(required=False, default='ascending', - choices=['ascending', 'descending']), - sort_start = dict(required=False), - sort_end = dict(required=False), + name=dict(required=False, default=[], type='list'), + sort=dict(required=False, default=None, + choices=['launch_configuration_name', 'image_id', 'created_time', 'instance_type', 'kernel_id', 'ramdisk_id', 'key_name']), + sort_order=dict(required=False, default='ascending', + choices=['ascending', 'descending']), + sort_start=dict(required=False), + sort_end=dict(required=False), ) ) diff --git a/lib/ansible/modules/cloud/amazon/ec2_metric_alarm.py b/lib/ansible/modules/cloud/amazon/ec2_metric_alarm.py index 6c8cd247ef..cbccfe662c 100644 --- a/lib/ansible/modules/cloud/amazon/ec2_metric_alarm.py +++ b/lib/ansible/modules/cloud/amazon/ec2_metric_alarm.py @@ -199,13 +199,13 @@ def create_metric_alarm(connection, module): alarm = alarms[0] changed = False - for attr in ('comparison','metric','namespace','statistic','threshold','period','evaluation_periods','unit','description'): + for attr in ('comparison', 'metric', 'namespace', 'statistic', 'threshold', 'period', 'evaluation_periods', 'unit', 'description'): if getattr(alarm, attr) != module.params.get(attr): changed = True setattr(alarm, attr, module.params.get(attr)) - #this is to deal with a current bug where you cannot assign '<=>' to the comparator when modifying an existing alarm + # this is to deal with a current bug where you cannot assign '<=>' to the comparator when modifying an existing alarm comparison = alarm.comparison - comparisons = {'<=' : 'LessThanOrEqualToThreshold', '<' : 'LessThanThreshold', '>=' : 'GreaterThanOrEqualToThreshold', '>' : 'GreaterThanThreshold'} + comparisons = {'<=': 'LessThanOrEqualToThreshold', '<': 'LessThanThreshold', '>=': 'GreaterThanOrEqualToThreshold', '>': 'GreaterThanThreshold'} alarm.comparison = comparisons[comparison] dim1 = module.params.get('dimensions') @@ -215,10 +215,10 @@ def create_metric_alarm(connection, module): if not isinstance(dim1[keys], list): dim1[keys] = [dim1[keys]] if keys not in dim2 or dim1[keys] != dim2[keys]: - changed=True + changed = True setattr(alarm, 'dimensions', dim1) - for attr in ('alarm_actions','insufficient_data_actions','ok_actions'): + for attr in ('alarm_actions', 'insufficient_data_actions', 'ok_actions'): action = module.params.get(attr) or [] # Boto and/or ansible may provide same elements in lists but in different order. # Compare on sets since they do not need any order. @@ -233,24 +233,25 @@ def create_metric_alarm(connection, module): module.fail_json(msg=str(e)) result = alarms[0] module.exit_json(changed=changed, name=result.name, - actions_enabled=result.actions_enabled, - alarm_actions=result.alarm_actions, - alarm_arn=result.alarm_arn, - comparison=result.comparison, - description=result.description, - dimensions=result.dimensions, - evaluation_periods=result.evaluation_periods, - insufficient_data_actions=result.insufficient_data_actions, - last_updated=result.last_updated, - metric=result.metric, - namespace=result.namespace, - ok_actions=result.ok_actions, - period=result.period, - state_reason=result.state_reason, - state_value=result.state_value, - statistic=result.statistic, - threshold=result.threshold, - unit=result.unit) + actions_enabled=result.actions_enabled, + alarm_actions=result.alarm_actions, + alarm_arn=result.alarm_arn, + comparison=result.comparison, + description=result.description, + dimensions=result.dimensions, + evaluation_periods=result.evaluation_periods, + insufficient_data_actions=result.insufficient_data_actions, + last_updated=result.last_updated, + metric=result.metric, + namespace=result.namespace, + ok_actions=result.ok_actions, + period=result.period, + state_reason=result.state_reason, + state_value=result.state_value, + statistic=result.statistic, + threshold=result.threshold, + unit=result.unit) + def delete_metric_alarm(connection, module): name = module.params.get('name') @@ -289,7 +290,7 @@ def main(): insufficient_data_actions=dict(type='list'), ok_actions=dict(type='list'), state=dict(default='present', choices=['present', 'absent']), - ) + ) ) module = AnsibleModule(argument_spec=argument_spec) diff --git a/lib/ansible/modules/cloud/amazon/ec2_scaling_policy.py b/lib/ansible/modules/cloud/amazon/ec2_scaling_policy.py index e4005145f1..708e111be6 100644 --- a/lib/ansible/modules/cloud/amazon/ec2_scaling_policy.py +++ b/lib/ansible/modules/cloud/amazon/ec2_scaling_policy.py @@ -88,7 +88,7 @@ def create_scaling_policy(connection, module): min_adjustment_step = module.params.get('min_adjustment_step') cooldown = module.params.get('cooldown') - scalingPolicies = connection.get_all_policies(as_group=asg_name,policy_names=[sp_name]) + scalingPolicies = connection.get_all_policies(as_group=asg_name, policy_names=[sp_name]) if not scalingPolicies: sp = ScalingPolicy( @@ -101,7 +101,7 @@ def create_scaling_policy(connection, module): try: connection.create_scaling_policy(sp) - policy = connection.get_all_policies(as_group=asg_name,policy_names=[sp_name])[0] + policy = connection.get_all_policies(as_group=asg_name, policy_names=[sp_name])[0] module.exit_json(changed=True, name=policy.name, arn=policy.policy_arn, as_name=policy.as_name, scaling_adjustment=policy.scaling_adjustment, cooldown=policy.cooldown, adjustment_type=policy.adjustment_type, min_adjustment_step=policy.min_adjustment_step) except BotoServerError as e: @@ -121,7 +121,7 @@ def create_scaling_policy(connection, module): setattr(policy, 'min_adjustment_step', module.params.get('min_adjustment_step')) # check the remaining attributes - for attr in ('adjustment_type','scaling_adjustment','cooldown'): + for attr in ('adjustment_type', 'scaling_adjustment', 'cooldown'): if getattr(policy, attr) != module.params.get(attr): changed = True setattr(policy, attr, module.params.get(attr)) @@ -129,7 +129,7 @@ def create_scaling_policy(connection, module): try: if changed: connection.create_scaling_policy(policy) - policy = connection.get_all_policies(as_group=asg_name,policy_names=[sp_name])[0] + policy = connection.get_all_policies(as_group=asg_name, policy_names=[sp_name])[0] module.exit_json(changed=changed, name=policy.name, arn=policy.policy_arn, as_name=policy.as_name, scaling_adjustment=policy.scaling_adjustment, cooldown=policy.cooldown, adjustment_type=policy.adjustment_type, min_adjustment_step=policy.min_adjustment_step) except BotoServerError as e: @@ -140,7 +140,7 @@ def delete_scaling_policy(connection, module): sp_name = module.params.get('name') asg_name = module.params.get('asg_name') - scalingPolicies = connection.get_all_policies(as_group=asg_name,policy_names=[sp_name]) + scalingPolicies = connection.get_all_policies(as_group=asg_name, policy_names=[sp_name]) if scalingPolicies: try: @@ -156,12 +156,12 @@ def main(): argument_spec = ec2_argument_spec() argument_spec.update( dict( - name = dict(required=True, type='str'), - adjustment_type = dict(type='str', choices=['ChangeInCapacity','ExactCapacity','PercentChangeInCapacity']), - asg_name = dict(required=True, type='str'), - scaling_adjustment = dict(type='int'), - min_adjustment_step = dict(type='int'), - cooldown = dict(type='int'), + name=dict(required=True, type='str'), + adjustment_type=dict(type='str', choices=['ChangeInCapacity', 'ExactCapacity', 'PercentChangeInCapacity']), + asg_name=dict(required=True, type='str'), + scaling_adjustment=dict(type='int'), + min_adjustment_step=dict(type='int'), + cooldown=dict(type='int'), state=dict(default='present', choices=['present', 'absent']), ) ) @@ -178,7 +178,7 @@ def main(): try: connection = connect_to_aws(boto.ec2.autoscale, region, **aws_connect_params) except (boto.exception.NoAuthHandlerFound, AnsibleAWSError) as e: - module.fail_json(msg = str(e)) + module.fail_json(msg=str(e)) if state == 'present': create_scaling_policy(connection, module) diff --git a/lib/ansible/modules/cloud/amazon/ec2_snapshot.py b/lib/ansible/modules/cloud/amazon/ec2_snapshot.py index eca2d3aeef..fed3ebe87e 100644 --- a/lib/ansible/modules/cloud/amazon/ec2_snapshot.py +++ b/lib/ansible/modules/cloud/amazon/ec2_snapshot.py @@ -184,7 +184,7 @@ def create_snapshot(module, ec2, state=None, description=None, wait=None, changed = False required = [volume_id, snapshot_id, instance_id] - if required.count(None) != len(required) - 1: # only 1 must be set + if required.count(None) != len(required) - 1: # only 1 must be set module.fail_json(msg='One and only one of volume_id or instance_id or snapshot_id must be specified') if instance_id and not device_name or device_name and not instance_id: module.fail_json(msg='Instance ID and device name must both be specified') @@ -193,7 +193,7 @@ def create_snapshot(module, ec2, state=None, description=None, wait=None, try: volumes = ec2.get_all_volumes(filters={'attachment.instance-id': instance_id, 'attachment.device': device_name}) except boto.exception.BotoServerError as e: - module.fail_json(msg = "%s: %s" % (e.error_code, e.error_message)) + module.fail_json(msg="%s: %s" % (e.error_code, e.error_message)) if not volumes: module.fail_json(msg="Could not find volume with name %s attached to instance %s" % (device_name, instance_id)) @@ -202,7 +202,7 @@ def create_snapshot(module, ec2, state=None, description=None, wait=None, if state == 'absent': if not snapshot_id: - module.fail_json(msg = 'snapshot_id must be set when state is absent') + module.fail_json(msg='snapshot_id must be set when state is absent') try: ec2.delete_snapshot(snapshot_id) except boto.exception.BotoServerError as e: @@ -210,7 +210,7 @@ def create_snapshot(module, ec2, state=None, description=None, wait=None, if e.error_code == 'InvalidSnapshot.NotFound': module.exit_json(changed=False) else: - module.fail_json(msg = "%s: %s" % (e.error_code, e.error_message)) + module.fail_json(msg="%s: %s" % (e.error_code, e.error_message)) # successful delete module.exit_json(changed=True) @@ -221,7 +221,7 @@ def create_snapshot(module, ec2, state=None, description=None, wait=None, except boto.exception.BotoServerError as e: module.fail_json(msg="%s: %s" % (e.error_code, e.error_message)) - last_snapshot_min_age = last_snapshot_min_age * 60 # Convert to seconds + last_snapshot_min_age = last_snapshot_min_age * 60 # Convert to seconds snapshot = _get_most_recent_snapshot(current_snapshots, max_snapshot_age_secs=last_snapshot_min_age) try: @@ -249,16 +249,16 @@ def create_snapshot_ansible_module(): argument_spec = ec2_argument_spec() argument_spec.update( dict( - volume_id = dict(), - description = dict(), - instance_id = dict(), - snapshot_id = dict(), - device_name = dict(), - wait = dict(type='bool', default=True), - wait_timeout = dict(type='int', default=0), - last_snapshot_min_age = dict(type='int', default=0), - snapshot_tags = dict(type='dict', default=dict()), - state = dict(choices=['absent', 'present'], default='present'), + volume_id=dict(), + description=dict(), + instance_id=dict(), + snapshot_id=dict(), + device_name=dict(), + wait=dict(type='bool', default=True), + wait_timeout=dict(type='int', default=0), + last_snapshot_min_age=dict(type='int', default=0), + snapshot_tags=dict(type='dict', default=dict()), + state=dict(choices=['absent', 'present'], default='present'), ) ) module = AnsibleModule(argument_spec=argument_spec) diff --git a/lib/ansible/modules/cloud/amazon/ec2_snapshot_facts.py b/lib/ansible/modules/cloud/amazon/ec2_snapshot_facts.py index 96a70e559e..7972885747 100644 --- a/lib/ansible/modules/cloud/amazon/ec2_snapshot_facts.py +++ b/lib/ansible/modules/cloud/amazon/ec2_snapshot_facts.py @@ -179,8 +179,8 @@ except ImportError: from ansible.module_utils.basic import AnsibleModule from ansible.module_utils.ec2 import (ansible_dict_to_boto3_filter_list, - boto3_conn, boto3_tag_list_to_ansible_dict, camel_dict_to_snake_dict, - ec2_argument_spec, get_aws_connection_info) + boto3_conn, boto3_tag_list_to_ansible_dict, camel_dict_to_snake_dict, + ec2_argument_spec, get_aws_connection_info) def list_ec2_snapshots(connection, module): diff --git a/lib/ansible/modules/cloud/amazon/ec2_tag.py b/lib/ansible/modules/cloud/amazon/ec2_tag.py index 0a3992fc73..beb32eeda2 100644 --- a/lib/ansible/modules/cloud/amazon/ec2_tag.py +++ b/lib/ansible/modules/cloud/amazon/ec2_tag.py @@ -126,9 +126,9 @@ from ansible.module_utils.ec2 import HAS_BOTO, ec2_argument_spec, ec2_connect def main(): argument_spec = ec2_argument_spec() argument_spec.update(dict( - resource = dict(required=True), - tags = dict(type='dict'), - state = dict(default='present', choices=['present', 'absent', 'list']), + resource=dict(required=True), + tags=dict(type='dict'), + state=dict(default='present', choices=['present', 'absent', 'list']), ) ) module = AnsibleModule(argument_spec=argument_spec, supports_check_mode=True) @@ -144,7 +144,7 @@ def main(): # We need a comparison here so that we can accurately report back changed status. # Need to expand the gettags return format and compare with "tags" and then tag or detag as appropriate. - filters = {'resource-id' : resource} + filters = {'resource-id': resource} gettags = ec2.get_all_tags(filters=filters) dictadd = {} @@ -158,14 +158,14 @@ def main(): if not tags: module.fail_json(msg="tags argument is required when state is present") if set(tags.items()).issubset(set(tagdict.items())): - module.exit_json(msg="Tags already exists in %s." %resource, changed=False) + module.exit_json(msg="Tags already exists in %s." % resource, changed=False) else: for (key, value) in set(tags.items()): if (key, value) not in set(tagdict.items()): dictadd[key] = value if not module.check_mode: ec2.create_tags(resource, dictadd) - module.exit_json(msg="Tags %s created for resource %s." % (dictadd,resource), changed=True) + module.exit_json(msg="Tags %s created for resource %s." % (dictadd, resource), changed=True) if state == 'absent': if not tags: @@ -180,7 +180,7 @@ def main(): dictremove[key] = value if not module.check_mode: ec2.delete_tags(resource, dictremove) - module.exit_json(msg="Tags %s removed for resource %s." % (dictremove,resource), changed=True) + module.exit_json(msg="Tags %s removed for resource %s." % (dictremove, resource), changed=True) if state == 'list': module.exit_json(changed=False, tags=tagdict) diff --git a/lib/ansible/modules/cloud/amazon/ec2_vol.py b/lib/ansible/modules/cloud/amazon/ec2_vol.py index 054fcdade8..d0ab999b69 100644 --- a/lib/ansible/modules/cloud/amazon/ec2_vol.py +++ b/lib/ansible/modules/cloud/amazon/ec2_vol.py @@ -280,7 +280,7 @@ def get_volume(module, ec2): try: vols = ec2.get_all_volumes(volume_ids=volume_ids, filters=filters) except boto.exception.BotoServerError as e: - module.fail_json(msg = "%s: %s" % (e.error_code, e.error_message)) + module.fail_json(msg="%s: %s" % (e.error_code, e.error_message)) if not vols: if id: @@ -306,7 +306,7 @@ def get_volumes(module, ec2): else: vols = ec2.get_all_volumes(filters={'attachment.instance-id': instance}) except boto.exception.BotoServerError as e: - module.fail_json(msg = "%s: %s" % (e.error_code, e.error_message)) + module.fail_json(msg="%s: %s" % (e.error_code, e.error_message)) return vols @@ -330,6 +330,7 @@ def boto_supports_volume_encryption(): """ return hasattr(boto, 'Version') and LooseVersion(boto.Version) >= LooseVersion('2.29.0') + def boto_supports_kms_key_id(): """ Check if Boto library supports kms_key_ids (added in 2.39.0) @@ -339,6 +340,7 @@ def boto_supports_kms_key_id(): """ return hasattr(boto, 'Version') and LooseVersion(boto.Version) >= LooseVersion('2.39.0') + def create_volume(module, ec2, zone): changed = False name = module.params.get('name') @@ -375,7 +377,7 @@ def create_volume(module, ec2, zone): if tags: ec2.create_tags([volume.id], tags) except boto.exception.BotoServerError as e: - module.fail_json(msg = "%s: %s" % (e.error_code, e.error_message)) + module.fail_json(msg="%s: %s" % (e.error_code, e.error_message)) return volume, changed @@ -400,12 +402,12 @@ def attach_volume(module, ec2, volume, instance): else: device_name = '/dev/xvdf' except boto.exception.BotoServerError as e: - module.fail_json(msg = "%s: %s" % (e.error_code, e.error_message)) + module.fail_json(msg="%s: %s" % (e.error_code, e.error_message)) if volume.attachment_state() is not None: adata = volume.attach_data if adata.instance_id != instance.id: - module.fail_json(msg = "Volume %s is already attached to another instance: %s" + module.fail_json(msg="Volume %s is already attached to another instance: %s" % (volume.id, adata.instance_id)) else: # Volume is already attached to right instance @@ -418,7 +420,7 @@ def attach_volume(module, ec2, volume, instance): volume.update() changed = True except boto.exception.BotoServerError as e: - module.fail_json(msg = "%s: %s" % (e.error_code, e.error_message)) + module.fail_json(msg="%s: %s" % (e.error_code, e.error_message)) modify_dot_attribute(module, ec2, instance, device_name) @@ -435,7 +437,7 @@ def modify_dot_attribute(module, ec2, instance, device_name): instance.update() dot = instance.block_device_mapping[device_name].delete_on_termination except boto.exception.BotoServerError as e: - module.fail_json(msg = "%s: %s" % (e.error_code, e.error_message)) + module.fail_json(msg="%s: %s" % (e.error_code, e.error_message)) if delete_on_termination != dot: try: @@ -450,7 +452,7 @@ def modify_dot_attribute(module, ec2, instance, device_name): instance.update() changed = True except boto.exception.BotoServerError as e: - module.fail_json(msg = "%s: %s" % (e.error_code, e.error_message)) + module.fail_json(msg="%s: %s" % (e.error_code, e.error_message)) return changed @@ -506,20 +508,20 @@ def get_volume_info(volume, state): def main(): argument_spec = ec2_argument_spec() argument_spec.update(dict( - instance = dict(), - id = dict(), - name = dict(), - volume_size = dict(), - volume_type = dict(choices=['standard', 'gp2', 'io1', 'st1', 'sc1'], default='standard'), - iops = dict(), - encrypted = dict(type='bool', default=False), - kms_key_id = dict(), - device_name = dict(), - delete_on_termination = dict(type='bool', default=False), - zone = dict(aliases=['availability_zone', 'aws_zone', 'ec2_zone']), - snapshot = dict(), - state = dict(choices=['absent', 'present', 'list'], default='present'), - tags = dict(type='dict', default={}) + instance=dict(), + id=dict(), + name=dict(), + volume_size=dict(), + volume_type=dict(choices=['standard', 'gp2', 'io1', 'st1', 'sc1'], default='standard'), + iops=dict(), + encrypted=dict(type='bool', default=False), + kms_key_id=dict(), + device_name=dict(), + delete_on_termination=dict(type='bool', default=False), + zone=dict(aliases=['availability_zone', 'aws_zone', 'ec2_zone']), + snapshot=dict(), + state=dict(choices=['absent', 'present', 'list'], default='present'), + tags=dict(type='dict', default={}) ) ) module = AnsibleModule(argument_spec=argument_spec) diff --git a/lib/ansible/modules/cloud/amazon/ec2_vol_facts.py b/lib/ansible/modules/cloud/amazon/ec2_vol_facts.py index 9d95b75b72..275602890a 100644 --- a/lib/ansible/modules/cloud/amazon/ec2_vol_facts.py +++ b/lib/ansible/modules/cloud/amazon/ec2_vol_facts.py @@ -92,12 +92,13 @@ def get_volume_info(volume): 'device': attachment.device, 'instance_id': attachment.instance_id, 'status': attachment.status - }, + }, 'tags': volume.tags - } + } return volume_info + def list_ec2_volumes(connection, module): filters = module.params.get("filters") @@ -118,7 +119,7 @@ def main(): argument_spec = ec2_argument_spec() argument_spec.update( dict( - filters = dict(default=None, type='dict') + filters=dict(default=None, type='dict') ) ) diff --git a/lib/ansible/modules/cloud/amazon/ec2_vpc_dhcp_option.py b/lib/ansible/modules/cloud/amazon/ec2_vpc_dhcp_option.py index 4c36fb4ff0..d33e9cb5dd 100644 --- a/lib/ansible/modules/cloud/amazon/ec2_vpc_dhcp_option.py +++ b/lib/ansible/modules/cloud/amazon/ec2_vpc_dhcp_option.py @@ -235,6 +235,7 @@ def ensure_tags(module, vpc_conn, resource_id, tags, add_only, check_mode): except EC2ResponseError as e: module.fail_json(msg="Failed to modify tags: %s" % e.message, exception=traceback.format_exc()) + def fetch_dhcp_options_for_vpc(vpc_conn, vpc_id): """ Returns the DHCP options object currently associated with the requested VPC ID using the VPC @@ -284,7 +285,7 @@ def main(): inherit_existing=dict(type='bool', default=False), tags=dict(type='dict', default=None, aliases=['resource_tags']), state=dict(type='str', default='present', choices=['present', 'absent']) - ) + ) ) module = AnsibleModule(argument_spec=argument_spec, supports_check_mode=True) @@ -312,17 +313,17 @@ def main(): new_options['ntp-servers'] = params['ntp_servers'] if params['domain_name'] is not None: # needs to be a list for comparison with boto objects later - new_options['domain-name'] = [ params['domain_name'] ] + new_options['domain-name'] = [params['domain_name']] if params['netbios_node_type'] is not None: # needs to be a list for comparison with boto objects later - new_options['netbios-node-type'] = [ str(params['netbios_node_type']) ] + new_options['netbios-node-type'] = [str(params['netbios_node_type'])] # If we were given a vpc_id then we need to look at the options on that if params['vpc_id']: existing_options = fetch_dhcp_options_for_vpc(connection, params['vpc_id']) # if we've been asked to inherit existing options, do that now if params['inherit_existing']: if existing_options: - for option in [ 'domain-name-servers', 'netbios-name-servers', 'ntp-servers', 'domain-name', 'netbios-node-type']: + for option in ['domain-name-servers', 'netbios-name-servers', 'ntp-servers', 'domain-name', 'netbios-node-type']: if existing_options.options.get(option) and new_options[option] != [] and (not new_options[option] or [''] == new_options[option]): new_options[option] = existing_options.options.get(option) @@ -336,7 +337,7 @@ def main(): # Now let's cover the case where there are existing options that we were told about by id # If a dhcp_options_id was supplied we don't look at options inside, just set tags (if given) else: - supplied_options = connection.get_all_dhcp_options(filters={'dhcp-options-id':params['dhcp_options_id']}) + supplied_options = connection.get_all_dhcp_options(filters={'dhcp-options-id': params['dhcp_options_id']}) if len(supplied_options) != 1: if params['state'] != 'absent': module.fail_json(msg=" a dhcp_options_id was supplied, but does not exist") diff --git a/lib/ansible/modules/cloud/amazon/ec2_vpc_nacl.py b/lib/ansible/modules/cloud/amazon/ec2_vpc_nacl.py index 0f26aac7f9..fceff34b65 100644 --- a/lib/ansible/modules/cloud/amazon/ec2_vpc_nacl.py +++ b/lib/ansible/modules/cloud/amazon/ec2_vpc_nacl.py @@ -147,7 +147,7 @@ from ansible.module_utils.ec2 import boto3_conn, ec2_argument_spec, get_aws_conn DEFAULT_RULE_FIELDS = { 'RuleNumber': 32767, 'RuleAction': 'deny', - 'CidrBlock': '0.0.0.0/0', + 'CidrBlock': '0.0.0.0/0', 'Protocol': '-1' } @@ -159,7 +159,7 @@ DEFAULT_EGRESS = dict(list(DEFAULT_RULE_FIELDS.items()) + [('Egress', True)]) PROTOCOL_NUMBERS = {'all': -1, 'icmp': 1, 'tcp': 6, 'udp': 17, } -#Utility methods +# Utility methods def icmp_present(entry): if len(entry) == 6 and entry[1] == 'icmp' or entry[1] == 1: return True @@ -225,7 +225,7 @@ def nacls_changed(nacl, client, module): nacl_id = nacl['NetworkAcls'][0]['NetworkAclId'] nacl = describe_network_acl(client, module) entries = nacl['NetworkAcls'][0]['Entries'] - tmp_egress = [entry for entry in entries if entry['Egress'] is True and DEFAULT_EGRESS !=entry] + tmp_egress = [entry for entry in entries if entry['Egress'] is True and DEFAULT_EGRESS != entry] tmp_ingress = [entry for entry in entries if entry['Egress'] is False] egress = [rule for rule in tmp_egress if DEFAULT_EGRESS != rule] ingress = [rule for rule in tmp_ingress if DEFAULT_INGRESS != rule] @@ -321,7 +321,7 @@ def construct_acl_entries(nacl, client, module): create_network_acl_entry(params, client, module) -## Module invocations +# Module invocations def setup_network_acl(client, module): changed = False nacl = describe_network_acl(client, module) @@ -372,7 +372,7 @@ def remove_network_acl(client, module): return changed, result -#Boto3 client methods +# Boto3 client methods def create_network_acl(vpc_id, client, module): try: if module.check_mode: @@ -546,7 +546,7 @@ def main(): ingress=dict(required=False, type='list', default=list()), egress=dict(required=False, type='list', default=list(),), state=dict(default='present', choices=['present', 'absent']), - ), + ), ) module = AnsibleModule(argument_spec=argument_spec, supports_check_mode=True, diff --git a/lib/ansible/modules/cloud/amazon/ec2_vpc_net.py b/lib/ansible/modules/cloud/amazon/ec2_vpc_net.py index a1a54d676b..520424bfb4 100644 --- a/lib/ansible/modules/cloud/amazon/ec2_vpc_net.py +++ b/lib/ansible/modules/cloud/amazon/ec2_vpc_net.py @@ -154,9 +154,9 @@ def vpc_exists(module, vpc, name, cidr_block, multi): matched_vpc = None try: - matching_vpcs=vpc.get_all_vpcs(filters={'tag:Name' : name, 'cidr-block' : cidr_block}) + matching_vpcs = vpc.get_all_vpcs(filters={'tag:Name': name, 'cidr-block': cidr_block}) except Exception as e: - e_msg=boto_exception(e) + e_msg = boto_exception(e) module.fail_json(msg=e_msg) if multi: @@ -186,7 +186,7 @@ def update_vpc_tags(vpc, module, vpc_obj, tags, name): else: return False except Exception as e: - e_msg=boto_exception(e) + e_msg = boto_exception(e) module.fail_json(msg=e_msg) @@ -199,6 +199,7 @@ def update_dhcp_opts(connection, module, vpc_obj, dhcp_id): else: return False + def get_vpc_values(vpc_obj): if vpc_obj is not None: @@ -213,18 +214,19 @@ def get_vpc_values(vpc_obj): else: return None + def main(): - argument_spec=ec2_argument_spec() + argument_spec = ec2_argument_spec() argument_spec.update(dict( - name = dict(type='str', default=None, required=True), - cidr_block = dict(type='str', default=None, required=True), - tenancy = dict(choices=['default', 'dedicated'], default='default'), - dns_support = dict(type='bool', default=True), - dns_hostnames = dict(type='bool', default=True), - dhcp_opts_id = dict(type='str', default=None, required=False), - tags = dict(type='dict', required=False, default=None, aliases=['resource_tags']), - state = dict(choices=['present', 'absent'], default='present'), - multi_ok = dict(type='bool', default=False) + name=dict(type='str', default=None, required=True), + cidr_block=dict(type='str', default=None, required=True), + tenancy=dict(choices=['default', 'dedicated'], default='default'), + dns_support=dict(type='bool', default=True), + dns_hostnames=dict(type='bool', default=True), + dhcp_opts_id=dict(type='str', default=None, required=False), + tags=dict(type='dict', required=False, default=None, aliases=['resource_tags']), + state=dict(choices=['present', 'absent'], default='present'), + multi_ok=dict(type='bool', default=False) ) ) @@ -236,17 +238,17 @@ def main(): if not HAS_BOTO: module.fail_json(msg='boto is required for this module') - name=module.params.get('name') - cidr_block=module.params.get('cidr_block') - tenancy=module.params.get('tenancy') - dns_support=module.params.get('dns_support') - dns_hostnames=module.params.get('dns_hostnames') - dhcp_id=module.params.get('dhcp_opts_id') - tags=module.params.get('tags') - state=module.params.get('state') - multi=module.params.get('multi_ok') + name = module.params.get('name') + cidr_block = module.params.get('cidr_block') + tenancy = module.params.get('tenancy') + dns_support = module.params.get('dns_support') + dns_hostnames = module.params.get('dns_hostnames') + dhcp_id = module.params.get('dhcp_opts_id') + tags = module.params.get('tags') + state = module.params.get('state') + multi = module.params.get('multi_ok') - changed=False + changed = False region, ec2_url, aws_connect_params = get_aws_connection_info(module) @@ -298,7 +300,7 @@ def main(): connection.modify_vpc_attribute(vpc_obj.id, enable_dns_support=dns_support) connection.modify_vpc_attribute(vpc_obj.id, enable_dns_hostnames=dns_hostnames) except BotoServerError as e: - e_msg=boto_exception(e) + e_msg = boto_exception(e) module.fail_json(msg=e_msg) if not module.check_mode: @@ -306,7 +308,7 @@ def main(): try: vpc_obj = connection.get_all_vpcs(vpc_obj.id)[0] except BotoServerError as e: - e_msg=boto_exception(e) + e_msg = boto_exception(e) module.fail_json(msg=e_msg) module.exit_json(changed=changed, vpc=get_vpc_values(vpc_obj)) @@ -325,7 +327,7 @@ def main(): except BotoServerError as e: e_msg = boto_exception(e) module.fail_json(msg="%s. You may want to use the ec2_vpc_subnet, ec2_vpc_igw, " - "and/or ec2_vpc_route_table modules to ensure the other components are absent." % e_msg) + "and/or ec2_vpc_route_table modules to ensure the other components are absent." % e_msg) module.exit_json(changed=changed, vpc=get_vpc_values(vpc_obj)) diff --git a/lib/ansible/modules/cloud/amazon/ec2_vpc_peer.py b/lib/ansible/modules/cloud/amazon/ec2_vpc_peer.py index 73d9dfc2ff..d0e106a2d7 100644 --- a/lib/ansible/modules/cloud/amazon/ec2_vpc_peer.py +++ b/lib/ansible/modules/cloud/amazon/ec2_vpc_peer.py @@ -373,7 +373,7 @@ def main(): client = boto3_conn(module, conn_type='client', resource='ec2', region=region, endpoint=ec2_url, **aws_connect_kwargs) except botocore.exceptions.NoCredentialsError as e: - module.fail_json(msg="Can't authorize connection - "+str(e)) + module.fail_json(msg="Can't authorize connection - " + str(e)) if state == 'present': (changed, results) = create_peer_connection(client, module) diff --git a/lib/ansible/modules/cloud/amazon/ec2_vpc_vgw.py b/lib/ansible/modules/cloud/amazon/ec2_vpc_vgw.py index 700b692d88..53c954acc4 100644 --- a/lib/ansible/modules/cloud/amazon/ec2_vpc_vgw.py +++ b/lib/ansible/modules/cloud/amazon/ec2_vpc_vgw.py @@ -147,6 +147,7 @@ def get_vgw_info(vgws): return vgw_info + def wait_for_status(client, module, vpn_gateway_id, status): polling_increment_secs = 15 max_retries = (module.params.get('wait_timeout') // polling_increment_secs) @@ -227,7 +228,7 @@ def delete_vgw(client, module, vpn_gateway_id): except botocore.exceptions.ClientError as e: module.fail_json(msg=to_native(e), exception=traceback.format_exc()) - #return the deleted VpnGatewayId as this is not included in the above response + # return the deleted VpnGatewayId as this is not included in the above response result = vpn_gateway_id return result @@ -236,7 +237,7 @@ def create_tags(client, module, vpn_gateway_id): params = dict() try: - response = client.create_tags(Resources=[vpn_gateway_id],Tags=load_tags(module)) + response = client.create_tags(Resources=[vpn_gateway_id], Tags=load_tags(module)) except botocore.exceptions.ClientError as e: module.fail_json(msg=to_native(e), exception=traceback.format_exc()) @@ -280,7 +281,7 @@ def find_tags(client, module, resource_id=None): try: response = client.describe_tags(Filters=[ {'Name': 'resource-id', 'Values': [resource_id]} - ]) + ]) except botocore.exceptions.ClientError as e: module.fail_json(msg=to_native(e), exception=traceback.format_exc()) @@ -295,7 +296,7 @@ def check_tags(client, module, existing_vgw, vpn_gateway_id): changed = False tags_list = {} - #format tags for comparison + # format tags for comparison for tags in existing_vgw[0]['Tags']: if tags['Key'] != 'Name': tags_list[tags['Key']] = tags['Value'] @@ -307,7 +308,7 @@ def check_tags(client, module, existing_vgw, vpn_gateway_id): vgw = find_vgw(client, module) changed = True - #if no tag args are supplied, delete any existing tags with the exception of the name tag + # if no tag args are supplied, delete any existing tags with the exception of the name tag if params['Tags'] is None and tags_list != {}: tags_to_delete = [] for tags in existing_vgw[0]['Tags']: @@ -346,7 +347,7 @@ def find_vgw(client, module, vpn_gateway_id=None): response = client.describe_vpn_gateways(Filters=[ {'Name': 'type', 'Values': [params['Type']]}, {'Name': 'tag:Name', 'Values': [params['Name']]} - ]) + ]) except botocore.exceptions.ClientError as e: module.fail_json(msg=to_native(e), exception=traceback.format_exc()) @@ -362,7 +363,7 @@ def find_vgw(client, module, vpn_gateway_id=None): response = client.describe_vpn_gateways(Filters=[ {'Name': 'type', 'Values': [params['Type']]}, {'Name': 'tag:Name', 'Values': [params['Name']]} - ]) + ]) except botocore.exceptions.ClientError as e: module.fail_json(msg=to_native(e), exception=traceback.format_exc()) @@ -502,7 +503,7 @@ def ensure_vgw_absent(client, module): deleted_vgw = "Nothing to do" else: - #Check that a name and type argument has been supplied if no vgw-id + # Check that a name and type argument has been supplied if no vgw-id if not module.params.get('name') or not module.params.get('type'): module.fail_json(msg='A name and type is required when no vgw-id and a status of \'absent\' is suppled') @@ -518,7 +519,7 @@ def ensure_vgw_absent(client, module): # detach the vpc from the vgw detach_vgw(client, module, vpn_gateway_id, params['VpcId']) - #now that the vpc has been detached, delete the vgw + # now that the vpc has been detached, delete the vgw deleted_vgw = delete_vgw(client, module, vpn_gateway_id) changed = True @@ -528,7 +529,7 @@ def ensure_vgw_absent(client, module): detach_vgw(client, module, vpn_gateway_id, vpc_to_detach) changed = True - #now that the vpc has been detached, delete the vgw + # now that the vpc has been detached, delete the vgw deleted_vgw = delete_vgw(client, module, vpn_gateway_id) else: @@ -555,7 +556,7 @@ def main(): wait_timeout=dict(type='int', default=320), type=dict(default='ipsec.1', choices=['ipsec.1']), tags=dict(default=None, required=False, type='dict', aliases=['resource_tags']), - ) + ) ) module = AnsibleModule(argument_spec=argument_spec) diff --git a/lib/ansible/modules/cloud/amazon/ec2_vpc_vgw_facts.py b/lib/ansible/modules/cloud/amazon/ec2_vpc_vgw_facts.py index 208dc04c69..a6cf9313eb 100644 --- a/lib/ansible/modules/cloud/amazon/ec2_vpc_vgw_facts.py +++ b/lib/ansible/modules/cloud/amazon/ec2_vpc_vgw_facts.py @@ -107,10 +107,10 @@ from ansible.module_utils.ec2 import (ec2_argument_spec, get_aws_connection_info def get_virtual_gateway_info(virtual_gateway): virtual_gateway_info = {'VpnGatewayId': virtual_gateway['VpnGatewayId'], - 'State': virtual_gateway['State'], - 'Type': virtual_gateway['Type'], - 'VpcAttachments': virtual_gateway['VpcAttachments'], - 'Tags': virtual_gateway['Tags']} + 'State': virtual_gateway['State'], + 'Type': virtual_gateway['Type'], + 'VpcAttachments': virtual_gateway['VpcAttachments'], + 'Tags': virtual_gateway['Tags']} return virtual_gateway_info @@ -126,10 +126,10 @@ def list_virtual_gateways(client, module): try: all_virtual_gateways = client.describe_vpn_gateways(**params) except botocore.exceptions.ClientError as e: - module.fail_json(msg=str(e),exception=traceback.format_exc()) + module.fail_json(msg=str(e), exception=traceback.format_exc()) snaked_vgws = [camel_dict_to_snake_dict(get_virtual_gateway_info(vgw)) - for vgw in all_virtual_gateways['VpnGateways']] + for vgw in all_virtual_gateways['VpnGateways']] module.exit_json(virtual_gateways=snaked_vgws) @@ -138,8 +138,8 @@ def main(): argument_spec = ec2_argument_spec() argument_spec.update( dict( - filters = dict(type='dict', default=dict()), - vpn_gateway_ids = dict(type='list', default=None) + filters=dict(type='dict', default=dict()), + vpn_gateway_ids=dict(type='list', default=None) ) ) @@ -153,7 +153,7 @@ def main(): region, ec2_url, aws_connect_kwargs = get_aws_connection_info(module, boto3=True) connection = boto3_conn(module, conn_type='client', resource='ec2', region=region, endpoint=ec2_url, **aws_connect_kwargs) except botocore.exceptions.NoCredentialsError as e: - module.fail_json(msg="Can't authorize connection - "+str(e)) + module.fail_json(msg="Can't authorize connection - " + str(e)) # call your function here results = list_virtual_gateways(connection, module) diff --git a/lib/ansible/modules/cloud/amazon/ec2_win_password.py b/lib/ansible/modules/cloud/amazon/ec2_win_password.py index ad7715defa..b28a59f89c 100644 --- a/lib/ansible/modules/cloud/amazon/ec2_win_password.py +++ b/lib/ansible/modules/cloud/amazon/ec2_win_password.py @@ -119,11 +119,11 @@ BACKEND = default_backend() def main(): argument_spec = ec2_argument_spec() argument_spec.update(dict( - instance_id = dict(required=True), - key_file = dict(required=True, type='path'), - key_passphrase = dict(no_log=True, default=None, required=False), - wait = dict(type='bool', default=False, required=False), - wait_timeout = dict(default=120, required=False), + instance_id=dict(required=True), + key_file=dict(required=True, type='path'), + key_passphrase=dict(no_log=True, default=None, required=False), + wait=dict(type='bool', default=False, required=False), + wait_timeout=dict(default=120, required=False), ) ) module = AnsibleModule(argument_spec=argument_spec) @@ -158,18 +158,18 @@ def main(): decoded = b64decode(data) if wait and datetime.datetime.now() >= end: - module.fail_json(msg = "wait for password timeout after %d seconds" % wait_timeout) + module.fail_json(msg="wait for password timeout after %d seconds" % wait_timeout) try: f = open(key_file, 'rb') except IOError as e: - module.fail_json(msg = "I/O error (%d) opening key file: %s" % (e.errno, e.strerror)) + module.fail_json(msg="I/O error (%d) opening key file: %s" % (e.errno, e.strerror)) else: try: with f: key = load_pem_private_key(f.read(), b_key_passphrase, BACKEND) except (ValueError, TypeError) as e: - module.fail_json(msg = "unable to parse key file") + module.fail_json(msg="unable to parse key file") try: decrypted = key.decrypt(decoded, PKCS1v15()) diff --git a/lib/ansible/modules/cloud/amazon/ecs_cluster.py b/lib/ansible/modules/cloud/amazon/ecs_cluster.py index 6cd1e58d8b..c50eedc57b 100644 --- a/lib/ansible/modules/cloud/amazon/ecs_cluster.py +++ b/lib/ansible/modules/cloud/amazon/ecs_cluster.py @@ -144,34 +144,35 @@ class EcsClusterManager: response = self.ecs.describe_clusters(clusters=[ cluster_name ]) - if len(response['failures'])>0: + if len(response['failures']) > 0: c = self.find_in_array(response['failures'], cluster_name, 'arn') - if c and c['reason']=='MISSING': + if c and c['reason'] == 'MISSING': return None # fall thru and look through found ones - if len(response['clusters'])>0: + if len(response['clusters']) > 0: c = self.find_in_array(response['clusters'], cluster_name) if c: return c raise Exception("Unknown problem describing cluster %s." % cluster_name) - def create_cluster(self, clusterName = 'default'): + def create_cluster(self, clusterName='default'): response = self.ecs.create_cluster(clusterName=clusterName) return response['cluster'] def delete_cluster(self, clusterName): return self.ecs.delete_cluster(cluster=clusterName) + def main(): argument_spec = ec2_argument_spec() argument_spec.update(dict( - state=dict(required=True, choices=['present', 'absent', 'has_instances'] ), - name=dict(required=True, type='str' ), + state=dict(required=True, choices=['present', 'absent', 'has_instances']), + name=dict(required=True, type='str'), delay=dict(required=False, type='int', default=10), repeat=dict(required=False, type='int', default=10) )) - required_together = ( ['state', 'name'] ) + required_together = (['state', 'name']) module = AnsibleModule(argument_spec=argument_spec, supports_check_mode=True, required_together=required_together) @@ -185,12 +186,12 @@ def main(): try: existing = cluster_mgr.describe_cluster(module.params['name']) except Exception as e: - module.fail_json(msg="Exception describing cluster '"+module.params['name']+"': "+str(e)) + module.fail_json(msg="Exception describing cluster '" + module.params['name'] + "': " + str(e)) results = dict(changed=False) if module.params['state'] == 'present': - if existing and 'status' in existing and existing['status']=="ACTIVE": - results['cluster']=existing + if existing and 'status' in existing and existing['status'] == "ACTIVE": + results['cluster'] = existing else: if not module.check_mode: # doesn't exist. create it. @@ -205,7 +206,7 @@ def main(): # it exists, so we should delete it and mark changed. # return info about the cluster deleted results['cluster'] = existing - if 'status' in existing and existing['status']=="INACTIVE": + if 'status' in existing and existing['status'] == "INACTIVE": results['changed'] = False else: if not module.check_mode: @@ -213,7 +214,7 @@ def main(): results['changed'] = True elif module.params['state'] == 'has_instances': if not existing: - module.fail_json(msg="Cluster '"+module.params['name']+" not found.") + module.fail_json(msg="Cluster '" + module.params['name'] + " not found.") return # it exists, so we should delete it and mark changed. # return info about the cluster deleted @@ -228,8 +229,8 @@ def main(): results['changed'] = True break time.sleep(delay) - if count == 0 and i is repeat-1: - module.fail_json(msg="Cluster instance count still zero after "+str(repeat)+" tries of "+str(delay)+" seconds each.") + if count == 0 and i is repeat - 1: + module.fail_json(msg="Cluster instance count still zero after " + str(repeat) + " tries of " + str(delay) + " seconds each.") return module.exit_json(**results) diff --git a/lib/ansible/modules/cloud/amazon/ecs_service.py b/lib/ansible/modules/cloud/amazon/ecs_service.py index 68d8905b0e..c0cb0e2d1e 100644 --- a/lib/ansible/modules/cloud/amazon/ecs_service.py +++ b/lib/ansible/modules/cloud/amazon/ecs_service.py @@ -308,13 +308,13 @@ class EcsServiceManager: cluster=cluster_name, services=[service_name]) msg = '' - if len(response['failures'])>0: + if len(response['failures']) > 0: c = self.find_in_array(response['failures'], service_name, 'arn') msg += ", failure reason is " + c['reason'] - if c and c['reason']=='MISSING': + if c and c['reason'] == 'MISSING': return None # fall thru and look through found ones - if len(response['services'])>0: + if len(response['services']) > 0: c = self.find_in_array(response['services'], service_name) if c: return c @@ -426,7 +426,7 @@ def main(): matching = False update = False - if existing and 'status' in existing and existing['status']=="ACTIVE": + if existing and 'status' in existing and existing['status'] == "ACTIVE": if service_mgr.is_matching_service(module.params, existing): matching = True results['service'] = service_mgr.jsonize(existing) @@ -446,25 +446,25 @@ def main(): if update: # update required response = service_mgr.update_service(module.params['name'], - module.params['cluster'], - module.params['task_definition'], - loadBalancers, - module.params['desired_count'], - clientToken, - role, - deploymentConfiguration) + module.params['cluster'], + module.params['task_definition'], + loadBalancers, + module.params['desired_count'], + clientToken, + role, + deploymentConfiguration) else: # doesn't exist. create it. response = service_mgr.create_service(module.params['name'], - module.params['cluster'], - module.params['task_definition'], - loadBalancers, - module.params['desired_count'], - clientToken, - role, - deploymentConfiguration, - module.params['placement_constraints'], - module.params['placement_strategy']) + module.params['cluster'], + module.params['task_definition'], + loadBalancers, + module.params['desired_count'], + clientToken, + role, + deploymentConfiguration, + module.params['placement_constraints'], + module.params['placement_strategy']) results['service'] = response @@ -479,7 +479,7 @@ def main(): del existing['deployments'] del existing['events'] results['ansible_facts'] = existing - if 'status' in existing and existing['status']=="INACTIVE": + if 'status' in existing and existing['status'] == "INACTIVE": results['changed'] = False else: if not module.check_mode: diff --git a/lib/ansible/modules/cloud/amazon/ecs_service_facts.py b/lib/ansible/modules/cloud/amazon/ecs_service_facts.py index 001f0a8944..d35e54dd66 100644 --- a/lib/ansible/modules/cloud/amazon/ecs_service_facts.py +++ b/lib/ansible/modules/cloud/amazon/ecs_service_facts.py @@ -170,17 +170,17 @@ class EcsServiceManager: if cluster and cluster is not None: fn_args['cluster'] = cluster response = self.ecs.list_services(**fn_args) - relevant_response = dict(services = response['serviceArns']) + relevant_response = dict(services=response['serviceArns']) return relevant_response def describe_services(self, cluster, services): fn_args = dict() if cluster and cluster is not None: fn_args['cluster'] = cluster - fn_args['services']=services.split(",") + fn_args['services'] = services.split(",") response = self.ecs.describe_services(**fn_args) - relevant_response = dict(services = map(self.extract_service_from, response['services'])) - if 'failures' in response and len(response['failures'])>0: + relevant_response = dict(services=map(self.extract_service_from, response['services'])) + if 'failures' in response and len(response['failures']) > 0: relevant_response['services_not_running'] = response['failures'] return relevant_response @@ -199,13 +199,14 @@ class EcsServiceManager: e['createdAt'] = str(e['createdAt']) return service + def main(): argument_spec = ec2_argument_spec() argument_spec.update(dict( - details=dict(required=False, type='bool', default=False ), - cluster=dict(required=False, type='str' ), - service=dict(required=False, type='str' ) + details=dict(required=False, type='bool', default=False), + cluster=dict(required=False, type='str'), + service=dict(required=False, type='str') )) module = AnsibleModule(argument_spec=argument_spec, supports_check_mode=True) diff --git a/lib/ansible/modules/cloud/amazon/ecs_task.py b/lib/ansible/modules/cloud/amazon/ecs_task.py index 804890b633..aa4cb6ad90 100644 --- a/lib/ansible/modules/cloud/amazon/ecs_task.py +++ b/lib/ansible/modules/cloud/amazon/ecs_task.py @@ -186,7 +186,7 @@ class EcsExecManager: family=service_name, desiredStatus=status ) - if len(response['taskArns'])>0: + if len(response['taskArns']) > 0: for c in response['taskArns']: if c.endswith(service_name): return c @@ -209,13 +209,13 @@ class EcsExecManager: if cluster: args['cluster'] = cluster if task_definition: - args['taskDefinition']=task_definition + args['taskDefinition'] = task_definition if overrides: - args['overrides']=overrides + args['overrides'] = overrides if container_instances: - args['containerInstances']=container_instances + args['containerInstances'] = container_instances if startedBy: - args['startedBy']=startedBy + args['startedBy'] = startedBy response = self.ecs.start_task(**args) # include tasks and failures return response['tasks'] @@ -224,17 +224,18 @@ class EcsExecManager: response = self.ecs.stop_task(cluster=cluster, task=task) return response['task'] + def main(): argument_spec = ec2_argument_spec() argument_spec.update(dict( - operation=dict(required=True, choices=['run', 'start', 'stop'] ), - cluster=dict(required=False, type='str' ), # R S P - task_definition=dict(required=False, type='str' ), # R* S* - overrides=dict(required=False, type='dict'), # R S - count=dict(required=False, type='int' ), # R - task=dict(required=False, type='str' ), # P* - container_instances=dict(required=False, type='list'), # S* - started_by=dict(required=False, type='str' ) # R S + operation=dict(required=True, choices=['run', 'start', 'stop']), + cluster=dict(required=False, type='str'), # R S P + task_definition=dict(required=False, type='str'), # R* S* + overrides=dict(required=False, type='dict'), # R S + count=dict(required=False, type='int'), # R + task=dict(required=False, type='str'), # P* + container_instances=dict(required=False, type='list'), # S* + started_by=dict(required=False, type='str') # R S )) module = AnsibleModule(argument_spec=argument_spec, supports_check_mode=True) @@ -276,7 +277,7 @@ def main(): if module.params['operation'] == 'run': if existing: # TBD - validate the rest of the details - results['task']=existing + results['task'] = existing else: if not module.check_mode: results['task'] = service_mgr.run_task( @@ -290,7 +291,7 @@ def main(): elif module.params['operation'] == 'start': if existing: # TBD - validate the rest of the details - results['task']=existing + results['task'] = existing else: if not module.check_mode: results['task'] = service_mgr.start_task( @@ -304,7 +305,7 @@ def main(): elif module.params['operation'] == 'stop': if existing: - results['task']=existing + results['task'] = existing else: if not module.check_mode: # it exists, so we should delete it and mark changed. diff --git a/lib/ansible/modules/cloud/amazon/ecs_taskdefinition.py b/lib/ansible/modules/cloud/amazon/ecs_taskdefinition.py index bcbffc6fdd..31f7f395dd 100644 --- a/lib/ansible/modules/cloud/amazon/ecs_taskdefinition.py +++ b/lib/ansible/modules/cloud/amazon/ecs_taskdefinition.py @@ -131,6 +131,7 @@ from ansible.module_utils.basic import AnsibleModule from ansible.module_utils.ec2 import boto3_conn, camel_dict_to_snake_dict, ec2_argument_spec, get_aws_connection_info from ansible.module_utils._text import to_text + class EcsTaskManager: """Handles ECS Tasks""" @@ -183,7 +184,7 @@ class EcsTaskManager: def describe_task_definitions(self, family): data = { "taskDefinitionArns": [], - "nextToken": None + "nextToken": None } def fetch(): @@ -371,7 +372,7 @@ def main(): if 'arn' in module.params and module.params['arn'] is not None: task_to_describe = module.params['arn'] elif 'family' in module.params and module.params['family'] is not None and 'revision' in module.params and \ - module.params['revision'] is not None: + module.params['revision'] is not None: task_to_describe = module.params['family'] + ":" + str(module.params['revision']) else: module.fail_json(msg="To use task definitions, an arn or family and revision must be specified") diff --git a/lib/ansible/modules/cloud/amazon/elasticache_subnet_group.py b/lib/ansible/modules/cloud/amazon/elasticache_subnet_group.py index ccf9ccdbb2..867f0a70fb 100644 --- a/lib/ansible/modules/cloud/amazon/elasticache_subnet_group.py +++ b/lib/ansible/modules/cloud/amazon/elasticache_subnet_group.py @@ -76,10 +76,10 @@ from ansible.module_utils.ec2 import HAS_BOTO, ec2_argument_spec, get_aws_connec def main(): argument_spec = ec2_argument_spec() argument_spec.update(dict( - state = dict(required=True, choices=['present', 'absent']), - name = dict(required=True), - description = dict(required=False), - subnets = dict(required=False, type='list'), + state=dict(required=True, choices=['present', 'absent']), + name=dict(required=True), + description=dict(required=False), + subnets=dict(required=False, type='list'), ) ) module = AnsibleModule(argument_spec=argument_spec) @@ -87,26 +87,25 @@ def main(): if not HAS_BOTO: module.fail_json(msg='boto required for this module') - state = module.params.get('state') - group_name = module.params.get('name').lower() - group_description = module.params.get('description') - group_subnets = module.params.get('subnets') or {} + state = module.params.get('state') + group_name = module.params.get('name').lower() + group_description = module.params.get('description') + group_subnets = module.params.get('subnets') or {} if state == 'present': for required in ['name', 'description', 'subnets']: if not module.params.get(required): - module.fail_json(msg = str("Parameter %s required for state='present'" % required)) + module.fail_json(msg=str("Parameter %s required for state='present'" % required)) else: for not_allowed in ['description', 'subnets']: if module.params.get(not_allowed): - module.fail_json(msg = str("Parameter %s not allowed for state='absent'" % not_allowed)) + module.fail_json(msg=str("Parameter %s not allowed for state='absent'" % not_allowed)) # Retrieve any AWS settings from the environment. region, ec2_url, aws_connect_kwargs = get_aws_connection_info(module) if not region: - module.fail_json(msg = str("Either region or AWS_REGION or EC2_REGION environment variable or boto config aws_region or ec2_region must be set.")) - + module.fail_json(msg=str("Either region or AWS_REGION or EC2_REGION environment variable or boto config aws_region or ec2_region must be set.")) """Get an elasticache connection""" try: @@ -123,7 +122,7 @@ def main(): exists = len(matching_groups) > 0 except BotoServerError as e: if e.error_code != 'CacheSubnetGroupNotFoundFault': - module.fail_json(msg = e.error_message) + module.fail_json(msg=e.error_message) if state == 'absent': if exists: @@ -139,7 +138,7 @@ def main(): except BotoServerError as e: if e.error_message != 'No modifications were requested.': - module.fail_json(msg = e.error_message) + module.fail_json(msg=e.error_message) else: changed = False diff --git a/lib/ansible/modules/cloud/amazon/elb_application_lb.py b/lib/ansible/modules/cloud/amazon/elb_application_lb.py index 5fabfd3464..2878dceb24 100644 --- a/lib/ansible/modules/cloud/amazon/elb_application_lb.py +++ b/lib/ansible/modules/cloud/amazon/elb_application_lb.py @@ -657,7 +657,6 @@ def compare_listeners(connection, module, current_listeners, new_listeners, purg def compare_rules(connection, module, current_listeners, listener): - """ Compare rules and return rules to add, rules to modify and rules to remove Rules are compared based on priority diff --git a/lib/ansible/modules/cloud/amazon/elb_classic_lb.py b/lib/ansible/modules/cloud/amazon/elb_classic_lb.py index 81c354cb88..4db1b735f0 100644 --- a/lib/ansible/modules/cloud/amazon/elb_classic_lb.py +++ b/lib/ansible/modules/cloud/amazon/elb_classic_lb.py @@ -425,6 +425,7 @@ def _throttleable_operation(max_retries): return _do_op return _operation_wrapper + def _get_vpc_connection(module, region, aws_connect_params): try: return connect_to_aws(boto.vpc, region, **aws_connect_params) @@ -434,6 +435,7 @@ def _get_vpc_connection(module, region, aws_connect_params): _THROTTLING_RETRIES = 5 + class ElbManager(object): """Handles ELB creation and destruction""" @@ -579,10 +581,10 @@ class ElbManager(object): # status of instances behind the ELB if info['instances']: - info['instance_health'] = [ dict( - instance_id = instance_state.instance_id, - reason_code = instance_state.reason_code, - state = instance_state.state + info['instance_health'] = [dict( + instance_id=instance_state.instance_id, + reason_code=instance_state.reason_code, + state=instance_state.state ) for instance_state in self.elb_conn.describe_instance_health(self.name)] else: info['instance_health'] = [] @@ -663,7 +665,7 @@ class ElbManager(object): elb_interfaces = self.ec2_conn.get_all_network_interfaces( filters={'attachment.instance-owner-id': 'amazon-elb', - 'description': 'ELB {0}'.format(self.name) }) + 'description': 'ELB {0}'.format(self.name)}) for x in range(0, max_retries): for interface in elb_interfaces: @@ -888,13 +890,13 @@ class ElbManager(object): if self.zones: if self.purge_zones: zones_to_disable = list(set(self.elb.availability_zones) - - set(self.zones)) + set(self.zones)) zones_to_enable = list(set(self.zones) - - set(self.elb.availability_zones)) + set(self.elb.availability_zones)) else: zones_to_disable = None zones_to_enable = list(set(self.zones) - - set(self.elb.availability_zones)) + set(self.elb.availability_zones)) if zones_to_enable: self._enable_zones(zones_to_enable) # N.B. This must come second, in case it would have removed all zones @@ -962,7 +964,7 @@ class ElbManager(object): "enabled": True, "s3_bucket_name": self.access_logs['s3_location'], "s3_bucket_prefix": self.access_logs.get('s3_prefix', ''), - "emit_interval": self.access_logs.get('interval', 60), + "emit_interval": self.access_logs.get('interval', 60), } update_access_logs_config = False @@ -1002,10 +1004,10 @@ class ElbManager(object): self.elb_conn.modify_lb_attribute(self.name, 'ConnectingSettings', attributes.connecting_settings) def _policy_name(self, policy_type): - return __file__.split('/')[-1].split('.')[0].replace('_', '-') + '-' + policy_type + return __file__.split('/')[-1].split('.')[0].replace('_', '-') + '-' + policy_type def _create_policy(self, policy_param, policy_meth, policy): - getattr(self.elb_conn, policy_meth )(policy_param, self.elb.name, policy) + getattr(self.elb_conn, policy_meth)(policy_param, self.elb.name, policy) def _delete_policy(self, elb_name, policy): self.elb_conn.delete_lb_policy(elb_name, policy) @@ -1223,7 +1225,7 @@ class ElbManager(object): params['Tags.member.%d.Value' % (i + 1)] = dictact[key] self.elb_conn.make_request('AddTags', params) - self.changed=True + self.changed = True # Remove extra tags dictact = dict(set(tagdict.items()) - set(self.tags.items())) @@ -1232,7 +1234,7 @@ class ElbManager(object): params['Tags.member.%d.Key' % (i + 1)] = key self.elb_conn.make_request('RemoveTags', params) - self.changed=True + self.changed = True def _get_health_check_target(self): """Compose target string from healthcheck parameters""" @@ -1275,7 +1277,7 @@ def main(): module = AnsibleModule( argument_spec=argument_spec, - mutually_exclusive = [['security_group_ids', 'security_group_names']] + mutually_exclusive=[['security_group_ids', 'security_group_names']] ) if not HAS_BOTO: @@ -1321,7 +1323,7 @@ def main(): security_group_ids = [] try: ec2 = connect_to_aws(boto.ec2, region, **aws_connect_params) - if subnets: # We have at least one subnet, ergo this is a VPC + if subnets: # We have at least one subnet, ergo this is a VPC vpc_conn = _get_vpc_connection(module=module, region=region, aws_connect_params=aws_connect_params) vpc_id = vpc_conn.get_all_subnets([subnets[0]])[0].vpc_id filters = {'vpc_id': vpc_id} @@ -1333,10 +1335,10 @@ def main(): if isinstance(group_name, string_types): group_name = [group_name] - group_id = [ str(grp.id) for grp in grp_details if str(grp.name) in group_name ] + group_id = [str(grp.id) for grp in grp_details if str(grp.name) in group_name] security_group_ids.extend(group_id) except boto.exception.NoAuthHandlerFound as e: - module.fail_json(msg = str(e)) + module.fail_json(msg=str(e)) elb_man = ElbManager(module, name, listeners, purge_listeners, zones, purge_zones, security_group_ids, health_check, diff --git a/lib/ansible/modules/cloud/amazon/elb_instance.py b/lib/ansible/modules/cloud/amazon/elb_instance.py index dea3a00c14..c6193f5938 100644 --- a/lib/ansible/modules/cloud/amazon/elb_instance.py +++ b/lib/ansible/modules/cloud/amazon/elb_instance.py @@ -167,7 +167,7 @@ class ElbManager: found = False for lb in self.lbs: if lb.name == lbtest: - found=True + found = True break return found @@ -326,7 +326,7 @@ def main(): argument_spec.update(dict( state={'required': True}, instance_id={'required': True}, - ec2_elbs={'default': None, 'required': False, 'type':'list'}, + ec2_elbs={'default': None, 'required': False, 'type': 'list'}, enable_availability_zone={'default': True, 'required': False, 'type': 'bool'}, wait={'required': False, 'default': True, 'type': 'bool'}, wait_timeout={'required': False, 'default': 0, 'type': 'int'} @@ -359,7 +359,7 @@ def main(): if ec2_elbs is not None: for elb in ec2_elbs: if not elb_man.exists(elb): - msg="ELB %s does not exist" % elb + msg = "ELB %s does not exist" % elb module.fail_json(msg=msg) if module.params['state'] == 'present': diff --git a/lib/ansible/modules/cloud/amazon/execute_lambda.py b/lib/ansible/modules/cloud/amazon/execute_lambda.py index d2530d8284..e0c9be5ac1 100644 --- a/lib/ansible/modules/cloud/amazon/execute_lambda.py +++ b/lib/ansible/modules/cloud/amazon/execute_lambda.py @@ -153,13 +153,13 @@ from ansible.module_utils._text import to_native def main(): argument_spec = ec2_argument_spec() argument_spec.update(dict( - name = dict(), - function_arn = dict(), - wait = dict(default=True, type='bool'), - tail_log = dict(default=False, type='bool'), - dry_run = dict(default=False, type='bool'), - version_qualifier = dict(), - payload = dict(default={}, type='dict'), + name=dict(), + function_arn=dict(), + wait=dict(default=True, type='bool'), + tail_log=dict(default=False, type='bool'), + dry_run=dict(default=False, type='bool'), + version_qualifier=dict(), + payload=dict(default={}, type='dict'), )) module = AnsibleModule( argument_spec=argument_spec, @@ -172,13 +172,13 @@ def main(): if not HAS_BOTO3: module.fail_json(msg='boto3 required for this module') - name = module.params.get('name') - function_arn = module.params.get('function_arn') - await_return = module.params.get('wait') - dry_run = module.params.get('dry_run') - tail_log = module.params.get('tail_log') - version_qualifier = module.params.get('version_qualifier') - payload = module.params.get('payload') + name = module.params.get('name') + function_arn = module.params.get('function_arn') + await_return = module.params.get('wait') + dry_run = module.params.get('dry_run') + tail_log = module.params.get('tail_log') + version_qualifier = module.params.get('version_qualifier') + payload = module.params.get('payload') if not HAS_BOTO3: module.fail_json(msg='Python module "boto3" is missing, please install it') @@ -247,7 +247,7 @@ def main(): module.fail_json(msg="Unexpected failure while invoking Lambda function", exception=traceback.format_exc()) - results ={ + results = { 'logs': '', 'status': response['StatusCode'], 'output': '', @@ -276,7 +276,7 @@ def main(): # format the stacktrace sent back as an array into a multiline string 'trace': '\n'.join( [' '.join([ - str(x) for x in line # cast line numbers to strings + str(x) for x in line # cast line numbers to strings ]) for line in results.get('output', {}).get('stackTrace', [])] ), 'errmsg': results['output'].get('errorMessage'), diff --git a/lib/ansible/modules/cloud/amazon/iam.py b/lib/ansible/modules/cloud/amazon/iam.py index 0c2770815b..346ac2095e 100644 --- a/lib/ansible/modules/cloud/amazon/iam.py +++ b/lib/ansible/modules/cloud/amazon/iam.py @@ -240,8 +240,8 @@ def create_user(module, iam, name, pwd, path, key_state, key_count): if key_count: while key_count > key_qty: keys.append(iam.create_access_key( - user_name=name).create_access_key_response.\ - create_access_key_result.\ + user_name=name).create_access_key_response. + create_access_key_result. access_key) key_qty += 1 else: @@ -258,7 +258,7 @@ def delete_dependencies_first(module, iam, name): # try to delete any keys try: current_keys = [ck['access_key_id'] for ck in - iam.get_all_access_keys(name).list_access_keys_result.access_key_metadata] + iam.get_all_access_keys(name).list_access_keys_result.access_key_metadata] for key in current_keys: iam.delete_access_key(key, name) changed = True @@ -447,7 +447,7 @@ def update_user(module, iam, name, new_name, new_path, key_state, key_count, key def set_users_groups(module, iam, name, groups, updated=None, -new_name=None): + new_name=None): """ Sets groups for a user, will purge groups not explicitly passed, while retaining pre-existing groups that also are in the new list. """ @@ -526,6 +526,7 @@ def delete_group(module=None, iam=None, name=None): changed = True return changed, name + def update_group(module=None, iam=None, name=None, new_name=None, new_path=None): changed = False try: @@ -554,12 +555,12 @@ def create_role(module, iam, name, path, role_list, prof_list, trust_policy_doc) if name not in role_list: changed = True iam_role_result = iam.create_role(name, - assume_role_policy_document=trust_policy_doc, - path=path).create_role_response.create_role_result.role + assume_role_policy_document=trust_policy_doc, + path=path).create_role_response.create_role_result.role if name not in prof_list: instance_profile_result = iam.create_instance_profile(name, - path=path).create_instance_profile_response.create_instance_profile_result.instance_profile + path=path).create_instance_profile_response.create_instance_profile_result.instance_profile iam.add_role_to_instance_profile(name, name) else: instance_profile_result = iam.get_instance_profile(name).get_instance_profile_response.get_instance_profile_result.instance_profile @@ -685,7 +686,7 @@ def main(): if iam_type == 'role' and state == 'update': module.fail_json(changed=False, msg="iam_type: role, cannot currently be updated, " - "please specify present or absent") + "please specify present or absent") # check if trust_policy is present -- it can be inline JSON or a file path to a JSON file if trust_policy_filepath: @@ -865,7 +866,7 @@ def main(): module.fail_json( changed=False, msg='Role update not currently supported by boto.') module.exit_json(changed=changed, roles=role_list, role_result=role_result, - instance_profile_result=instance_profile_result) + instance_profile_result=instance_profile_result) if __name__ == '__main__': diff --git a/lib/ansible/modules/cloud/amazon/iam_policy.py b/lib/ansible/modules/cloud/amazon/iam_policy.py index 06268514c0..b4c1eafe16 100644 --- a/lib/ansible/modules/cloud/amazon/iam_policy.py +++ b/lib/ansible/modules/cloud/amazon/iam_policy.py @@ -138,15 +138,15 @@ def user_action(module, iam, name, policy_name, skip, pdoc, state): changed = False try: current_policies = [cp for cp in iam.get_all_user_policies(name). - list_user_policies_result. - policy_names] + list_user_policies_result. + policy_names] matching_policies = [] for pol in current_policies: ''' urllib is needed here because boto returns url encoded strings instead ''' if urllib.parse.unquote(iam.get_user_policy(name, pol). - get_user_policy_result.policy_document) == pdoc: + get_user_policy_result.policy_document) == pdoc: policy_match = True matching_policies.append(pol) @@ -168,8 +168,8 @@ def user_action(module, iam, name, policy_name, skip, pdoc, state): module.exit_json(changed=changed, msg="%s policy is already absent" % policy_name) updated_policies = [cp for cp in iam.get_all_user_policies(name). - list_user_policies_result. - policy_names] + list_user_policies_result. + policy_names] except boto.exception.BotoServerError as err: error_msg = boto_exception(err) module.fail_json(changed=changed, msg=error_msg) @@ -182,8 +182,8 @@ def role_action(module, iam, name, policy_name, skip, pdoc, state): changed = False try: current_policies = [cp for cp in iam.list_role_policies(name). - list_role_policies_result. - policy_names] + list_role_policies_result. + policy_names] except boto.exception.BotoServerError as e: if e.error_code == "NoSuchEntity": # Role doesn't exist so it's safe to assume the policy doesn't either @@ -195,7 +195,7 @@ def role_action(module, iam, name, policy_name, skip, pdoc, state): matching_policies = [] for pol in current_policies: if urllib.parse.unquote(iam.get_role_policy(name, pol). - get_role_policy_result.policy_document) == pdoc: + get_role_policy_result.policy_document) == pdoc: policy_match = True matching_policies.append(pol) @@ -220,8 +220,8 @@ def role_action(module, iam, name, policy_name, skip, pdoc, state): module.fail_json(msg=err.message) updated_policies = [cp for cp in iam.list_role_policies(name). - list_role_policies_result. - policy_names] + list_role_policies_result. + policy_names] except boto.exception.BotoServerError as err: error_msg = boto_exception(err) module.fail_json(changed=changed, msg=error_msg) @@ -232,19 +232,19 @@ def role_action(module, iam, name, policy_name, skip, pdoc, state): def group_action(module, iam, name, policy_name, skip, pdoc, state): policy_match = False changed = False - msg='' + msg = '' try: current_policies = [cp for cp in iam.get_all_group_policies(name). - list_group_policies_result. - policy_names] + list_group_policies_result. + policy_names] matching_policies = [] for pol in current_policies: if urllib.parse.unquote(iam.get_group_policy(name, pol). - get_group_policy_result.policy_document) == pdoc: + get_group_policy_result.policy_document) == pdoc: policy_match = True matching_policies.append(pol) - msg=("The policy document you specified already exists " - "under the name %s." % pol) + msg = ("The policy document you specified already exists " + "under the name %s." % pol) if state == 'present': # If policy document does not already exist (either it's changed # or the policy is not present) or if we're not skipping dupes then @@ -264,8 +264,8 @@ def group_action(module, iam, name, policy_name, skip, pdoc, state): msg="%s policy is already absent" % policy_name) updated_policies = [cp for cp in iam.get_all_group_policies(name). - list_group_policies_result. - policy_names] + list_group_policies_result. + policy_names] except boto.exception.BotoServerError as err: error_msg = boto_exception(err) module.fail_json(changed=changed, msg=error_msg) @@ -317,7 +317,7 @@ def main(): except Exception as e: module.fail_json(msg='Failed to convert the policy into valid JSON: %s' % str(e)) else: - pdoc=None + pdoc = None region, ec2_url, aws_connect_kwargs = get_aws_connection_info(module) @@ -343,8 +343,8 @@ def main(): module.exit_json(changed=changed, role_name=name, policies=current_policies) elif iam_type == 'group': changed, group_name, current_policies, msg = group_action(module, iam, name, - policy_name, skip, pdoc, - state) + policy_name, skip, pdoc, + state) module.exit_json(changed=changed, group_name=name, policies=current_policies, msg=msg) diff --git a/lib/ansible/modules/cloud/amazon/rds_subnet_group.py b/lib/ansible/modules/cloud/amazon/rds_subnet_group.py index 3b28468292..e415319f00 100644 --- a/lib/ansible/modules/cloud/amazon/rds_subnet_group.py +++ b/lib/ansible/modules/cloud/amazon/rds_subnet_group.py @@ -80,10 +80,10 @@ from ansible.module_utils.ec2 import HAS_BOTO, connect_to_aws, ec2_argument_spec def main(): argument_spec = ec2_argument_spec() argument_spec.update(dict( - state = dict(required=True, choices=['present', 'absent']), - name = dict(required=True), - description = dict(required=False), - subnets = dict(required=False, type='list'), + state=dict(required=True, choices=['present', 'absent']), + name=dict(required=True), + description=dict(required=False), + subnets=dict(required=False, type='list'), ) ) module = AnsibleModule(argument_spec=argument_spec) @@ -91,30 +91,30 @@ def main(): if not HAS_BOTO: module.fail_json(msg='boto required for this module') - state = module.params.get('state') - group_name = module.params.get('name').lower() - group_description = module.params.get('description') - group_subnets = module.params.get('subnets') or {} + state = module.params.get('state') + group_name = module.params.get('name').lower() + group_description = module.params.get('description') + group_subnets = module.params.get('subnets') or {} if state == 'present': for required in ['name', 'description', 'subnets']: if not module.params.get(required): - module.fail_json(msg = str("Parameter %s required for state='present'" % required)) + module.fail_json(msg=str("Parameter %s required for state='present'" % required)) else: for not_allowed in ['description', 'subnets']: if module.params.get(not_allowed): - module.fail_json(msg = str("Parameter %s not allowed for state='absent'" % not_allowed)) + module.fail_json(msg=str("Parameter %s not allowed for state='absent'" % not_allowed)) # Retrieve any AWS settings from the environment. region, ec2_url, aws_connect_kwargs = get_aws_connection_info(module) if not region: - module.fail_json(msg = str("Either region or AWS_REGION or EC2_REGION environment variable or boto config aws_region or ec2_region must be set.")) + module.fail_json(msg=str("Either region or AWS_REGION or EC2_REGION environment variable or boto config aws_region or ec2_region must be set.")) try: conn = connect_to_aws(boto.rds, region, **aws_connect_kwargs) except BotoServerError as e: - module.fail_json(msg = e.error_message) + module.fail_json(msg=e.error_message) try: changed = False @@ -125,7 +125,7 @@ def main(): exists = len(matching_groups) > 0 except BotoServerError as e: if e.error_code != 'DBSubnetGroupNotFoundFault': - module.fail_json(msg = e.error_message) + module.fail_json(msg=e.error_message) if state == 'absent': if exists: @@ -145,7 +145,7 @@ def main(): changed_group = conn.modify_db_subnet_group(group_name, description=group_description, subnet_ids=group_subnets) changed = True except BotoServerError as e: - module.fail_json(msg = e.error_message) + module.fail_json(msg=e.error_message) module.exit_json(changed=changed) diff --git a/lib/ansible/modules/cloud/amazon/redshift.py b/lib/ansible/modules/cloud/amazon/redshift.py index f035d17505..6d7a942752 100644 --- a/lib/ansible/modules/cloud/amazon/redshift.py +++ b/lib/ansible/modules/cloud/amazon/redshift.py @@ -237,15 +237,15 @@ from ansible.module_utils.ec2 import HAS_BOTO, connect_to_aws, ec2_argument_spec def _collect_facts(resource): """Transfrom cluster information to dict.""" facts = { - 'identifier' : resource['ClusterIdentifier'], - 'create_time' : resource['ClusterCreateTime'], - 'status' : resource['ClusterStatus'], - 'username' : resource['MasterUsername'], - 'db_name' : resource['DBName'], - 'availability_zone' : resource['AvailabilityZone'], + 'identifier': resource['ClusterIdentifier'], + 'create_time': resource['ClusterCreateTime'], + 'status': resource['ClusterStatus'], + 'username': resource['MasterUsername'], + 'db_name': resource['DBName'], + 'availability_zone': resource['AvailabilityZone'], 'maintenance_window': resource['PreferredMaintenanceWindow'], - 'url' : resource['Endpoint']['Address'], - 'port' : resource['Endpoint']['Port'] + 'url': resource['Endpoint']['Address'], + 'port': resource['Endpoint']['Port'] } for node in resource['ClusterNodes']: @@ -267,11 +267,11 @@ def create_cluster(module, redshift): Returns: """ - identifier = module.params.get('identifier') - node_type = module.params.get('node_type') - username = module.params.get('username') - password = module.params.get('password') - wait = module.params.get('wait') + identifier = module.params.get('identifier') + node_type = module.params.get('node_type') + username = module.params.get('username') + password = module.params.get('password') + wait = module.params.get('wait') wait_timeout = module.params.get('wait_timeout') changed = True @@ -286,7 +286,7 @@ def create_cluster(module, redshift): 'number_of_nodes', 'publicly_accessible', 'encrypted', 'elastic_ip'): if p in module.params: - params[ p ] = module.params.get( p ) + params[p] = module.params.get(p) try: redshift.describe_clusters(identifier)['DescribeClustersResponse']['DescribeClustersResult']['Clusters'][0] @@ -310,7 +310,7 @@ def create_cluster(module, redshift): while wait_timeout > time.time() and resource['ClusterStatus'] != 'available': time.sleep(5) if wait_timeout <= time.time(): - module.fail_json(msg = "Timeout waiting for resource %s" % resource.id) + module.fail_json(msg="Timeout waiting for resource %s" % resource.id) resource = redshift.describe_clusters(identifier)['DescribeClustersResponse']['DescribeClustersResult']['Clusters'][0] @@ -368,7 +368,7 @@ def delete_cluster(module, redshift): while wait_timeout > time.time() and resource['ClusterStatus'] != 'deleting': time.sleep(5) if wait_timeout <= time.time(): - module.fail_json(msg = "Timeout waiting for resource %s" % resource.id) + module.fail_json(msg="Timeout waiting for resource %s" % resource.id) resource = redshift.describe_clusters(identifier)['DescribeClustersResponse']['DescribeClustersResult']['Clusters'][0] @@ -386,8 +386,8 @@ def modify_cluster(module, redshift): redshift: authenticated redshift connection object """ - identifier = module.params.get('identifier') - wait = module.params.get('wait') + identifier = module.params.get('identifier') + wait = module.params.get('wait') wait_timeout = module.params.get('wait_timeout') # Package up the optional parameters @@ -422,7 +422,7 @@ def modify_cluster(module, redshift): while wait_timeout > time.time() and resource['ClusterStatus'] != 'available': time.sleep(5) if wait_timeout <= time.time(): - module.fail_json(msg = "Timeout waiting for resource %s" % resource.id) + module.fail_json(msg="Timeout waiting for resource %s" % resource.id) resource = redshift.describe_clusters(identifier)['DescribeClustersResponse']['DescribeClustersResult']['Clusters'][0] @@ -436,34 +436,34 @@ def modify_cluster(module, redshift): def main(): argument_spec = ec2_argument_spec() argument_spec.update(dict( - command = dict(choices=['create', 'facts', 'delete', 'modify'], required=True), - identifier = dict(required=True), - node_type = dict(choices=['ds1.xlarge', 'ds1.8xlarge', 'ds2.xlarge', 'ds2.8xlarge', 'dc1.large', - 'dc2.large','dc1.8xlarge', 'dw1.xlarge', 'dw1.8xlarge', 'dw2.large', - 'dw2.8xlarge'], required=False), - username = dict(required=False), - password = dict(no_log=True, required=False), - db_name = dict(require=False), - cluster_type = dict(choices=['multi-node', 'single-node', ], default='single-node'), - cluster_security_groups = dict(aliases=['security_groups'], type='list'), - vpc_security_group_ids = dict(aliases=['vpc_security_groups'], type='list'), - skip_final_cluster_snapshot = dict(aliases=['skip_final_snapshot'], type='bool', default=False), - final_cluster_snapshot_identifier = dict(aliases=['final_snapshot_id'], required=False), - cluster_subnet_group_name = dict(aliases=['subnet']), - availability_zone = dict(aliases=['aws_zone', 'zone']), - preferred_maintenance_window = dict(aliases=['maintance_window', 'maint_window']), - cluster_parameter_group_name = dict(aliases=['param_group_name']), - automated_snapshot_retention_period = dict(aliases=['retention_period']), - port = dict(type='int'), - cluster_version = dict(aliases=['version'], choices=['1.0']), - allow_version_upgrade = dict(aliases=['version_upgrade'], type='bool', default=True), - number_of_nodes = dict(type='int'), - publicly_accessible = dict(type='bool', default=False), - encrypted = dict(type='bool', default=False), - elastic_ip = dict(required=False), - new_cluster_identifier = dict(aliases=['new_identifier']), - wait = dict(type='bool', default=False), - wait_timeout = dict(type='int', default=300), + command=dict(choices=['create', 'facts', 'delete', 'modify'], required=True), + identifier=dict(required=True), + node_type=dict(choices=['ds1.xlarge', 'ds1.8xlarge', 'ds2.xlarge', 'ds2.8xlarge', 'dc1.large', + 'dc2.large', 'dc1.8xlarge', 'dw1.xlarge', 'dw1.8xlarge', 'dw2.large', + 'dw2.8xlarge'], required=False), + username=dict(required=False), + password=dict(no_log=True, required=False), + db_name=dict(require=False), + cluster_type=dict(choices=['multi-node', 'single-node', ], default='single-node'), + cluster_security_groups=dict(aliases=['security_groups'], type='list'), + vpc_security_group_ids=dict(aliases=['vpc_security_groups'], type='list'), + skip_final_cluster_snapshot=dict(aliases=['skip_final_snapshot'], type='bool', default=False), + final_cluster_snapshot_identifier=dict(aliases=['final_snapshot_id'], required=False), + cluster_subnet_group_name=dict(aliases=['subnet']), + availability_zone=dict(aliases=['aws_zone', 'zone']), + preferred_maintenance_window=dict(aliases=['maintance_window', 'maint_window']), + cluster_parameter_group_name=dict(aliases=['param_group_name']), + automated_snapshot_retention_period=dict(aliases=['retention_period']), + port=dict(type='int'), + cluster_version=dict(aliases=['version'], choices=['1.0']), + allow_version_upgrade=dict(aliases=['version_upgrade'], type='bool', default=True), + number_of_nodes=dict(type='int'), + publicly_accessible=dict(type='bool', default=False), + encrypted=dict(type='bool', default=False), + elastic_ip=dict(required=False), + new_cluster_identifier=dict(aliases=['new_identifier']), + wait=dict(type='bool', default=False), + wait_timeout=dict(type='int', default=300), )) required_if = [ diff --git a/lib/ansible/modules/cloud/amazon/route53_health_check.py b/lib/ansible/modules/cloud/amazon/route53_health_check.py index ea9949a9b0..babfde324b 100644 --- a/lib/ansible/modules/cloud/amazon/route53_health_check.py +++ b/lib/ansible/modules/cloud/amazon/route53_health_check.py @@ -166,6 +166,7 @@ def find_health_check(conn, wanted): return check return None + def to_health_check(config): return HealthCheck( config.get('IPAddress'), @@ -178,6 +179,7 @@ def to_health_check(config): failure_threshold=int(config.get('FailureThreshold')), ) + def health_check_diff(a, b): a = a.__dict__ b = b.__dict__ @@ -189,6 +191,7 @@ def health_check_diff(a, b): diff[key] = b.get(key) return diff + def to_template_params(health_check): params = { 'ip_addr_part': '', @@ -240,7 +243,8 @@ UPDATEHCXMLBody = """ """ -def create_health_check(conn, health_check, caller_ref = None): + +def create_health_check(conn, health_check, caller_ref=None): if caller_ref is None: caller_ref = str(uuid.uuid4()) uri = '/%s/healthcheck' % conn.Version @@ -259,6 +263,7 @@ def create_health_check(conn, health_check, caller_ref = None): else: raise exception.DNSServerError(response.status, response.reason, body) + def update_health_check(conn, health_check_id, health_check_version, health_check): uri = '/%s/healthcheck/%s' % (conn.Version, health_check_id) params = to_template_params(health_check) @@ -279,18 +284,19 @@ def update_health_check(conn, health_check_id, health_check_version, health_chec h.parse(body) return e + def main(): argument_spec = ec2_argument_spec() argument_spec.update(dict( - state = dict(choices=['present', 'absent'], default='present'), - ip_address = dict(), - port = dict(type='int'), - type = dict(required=True, choices=['HTTP', 'HTTPS', 'HTTP_STR_MATCH', 'HTTPS_STR_MATCH', 'TCP']), - resource_path = dict(), - fqdn = dict(), - string_match = dict(), - request_interval = dict(type='int', choices=[10, 30], default=30), - failure_threshold = dict(type='int', choices=[1, 2, 3, 4, 5, 6, 7, 8, 9, 10], default=3), + state=dict(choices=['present', 'absent'], default='present'), + ip_address=dict(), + port=dict(type='int'), + type=dict(required=True, choices=['HTTP', 'HTTPS', 'HTTP_STR_MATCH', 'HTTPS_STR_MATCH', 'TCP']), + resource_path=dict(), + fqdn=dict(), + string_match=dict(), + request_interval=dict(type='int', choices=[10, 30], default=30), + failure_threshold=dict(type='int', choices=[1, 2, 3, 4, 5, 6, 7, 8, 9, 10], default=3), ) ) module = AnsibleModule(argument_spec=argument_spec) @@ -298,15 +304,15 @@ def main(): if not HAS_BOTO: module.fail_json(msg='boto 2.27.0+ required for this module') - state_in = module.params.get('state') - ip_addr_in = module.params.get('ip_address') - port_in = module.params.get('port') - type_in = module.params.get('type') - resource_path_in = module.params.get('resource_path') - fqdn_in = module.params.get('fqdn') - string_match_in = module.params.get('string_match') - request_interval_in = module.params.get('request_interval') - failure_threshold_in = module.params.get('failure_threshold') + state_in = module.params.get('state') + ip_addr_in = module.params.get('ip_address') + port_in = module.params.get('port') + type_in = module.params.get('type') + resource_path_in = module.params.get('resource_path') + fqdn_in = module.params.get('fqdn') + string_match_in = module.params.get('string_match') + request_interval_in = module.params.get('request_interval') + failure_threshold_in = module.params.get('failure_threshold') if ip_addr_in is None and fqdn_in is None: module.fail_json(msg="parameter 'ip_address' or 'fqdn' is required") @@ -334,7 +340,7 @@ def main(): try: conn = Route53Connection(**aws_connect_kwargs) except boto.exception.BotoServerError as e: - module.fail_json(msg = e.error_message) + module.fail_json(msg=e.error_message) changed = False action = None @@ -362,7 +368,7 @@ def main(): conn.delete_health_check(check_id) changed = True else: - module.fail_json(msg = "Logic Error: Unknown state") + module.fail_json(msg="Logic Error: Unknown state") module.exit_json(changed=changed, health_check=dict(id=check_id), action=action) diff --git a/lib/ansible/modules/cloud/amazon/s3_lifecycle.py b/lib/ansible/modules/cloud/amazon/s3_lifecycle.py index 3cd87c6a12..04f70b3924 100644 --- a/lib/ansible/modules/cloud/amazon/s3_lifecycle.py +++ b/lib/ansible/modules/cloud/amazon/s3_lifecycle.py @@ -256,6 +256,7 @@ def create_lifecycle_rule(connection, module): module.exit_json(changed=changed) + def compare_rule(rule_a, rule_b): # Copy objects @@ -364,27 +365,27 @@ def main(): argument_spec = ec2_argument_spec() argument_spec.update( dict( - name = dict(required=True, type='str'), - expiration_days = dict(default=None, required=False, type='int'), - expiration_date = dict(default=None, required=False, type='str'), - prefix = dict(default=None, required=False), - requester_pays = dict(default='no', type='bool'), - rule_id = dict(required=False, type='str'), - state = dict(default='present', choices=['present', 'absent']), - status = dict(default='enabled', choices=['enabled', 'disabled']), - storage_class = dict(default='glacier', type='str', choices=['glacier', 'standard_ia']), - transition_days = dict(default=None, required=False, type='int'), - transition_date = dict(default=None, required=False, type='str') + name=dict(required=True, type='str'), + expiration_days=dict(default=None, required=False, type='int'), + expiration_date=dict(default=None, required=False, type='str'), + prefix=dict(default=None, required=False), + requester_pays=dict(default='no', type='bool'), + rule_id=dict(required=False, type='str'), + state=dict(default='present', choices=['present', 'absent']), + status=dict(default='enabled', choices=['enabled', 'disabled']), + storage_class=dict(default='glacier', type='str', choices=['glacier', 'standard_ia']), + transition_days=dict(default=None, required=False, type='int'), + transition_date=dict(default=None, required=False, type='str') ) ) module = AnsibleModule(argument_spec=argument_spec, - mutually_exclusive = [ - [ 'expiration_days', 'expiration_date' ], - [ 'expiration_days', 'transition_date' ], - [ 'transition_days', 'transition_date' ], - [ 'transition_days', 'expiration_date' ] - ] + mutually_exclusive=[ + ['expiration_days', 'expiration_date'], + ['expiration_days', 'transition_date'], + ['transition_days', 'transition_date'], + ['transition_days', 'expiration_date'] + ] ) if not HAS_BOTO: @@ -428,7 +429,7 @@ def main(): except ValueError as e: module.fail_json(msg="expiration_date is not a valid ISO-8601 format. The time must be midnight and a timezone of GMT must be included") - boto_required_version = (2,40,0) + boto_required_version = (2, 40, 0) if storage_class == 'standard_ia' and tuple(map(int, (boto.__version__.split(".")))) < boto_required_version: module.fail_json(msg="'standard_ia' class requires boto >= 2.40.0") diff --git a/lib/ansible/modules/cloud/amazon/s3_logging.py b/lib/ansible/modules/cloud/amazon/s3_logging.py index 3fab08e628..429915ab2c 100644 --- a/lib/ansible/modules/cloud/amazon/s3_logging.py +++ b/lib/ansible/modules/cloud/amazon/s3_logging.py @@ -137,10 +137,10 @@ def main(): argument_spec = ec2_argument_spec() argument_spec.update( dict( - name = dict(required=True), - target_bucket = dict(required=False, default=None), - target_prefix = dict(required=False, default=""), - state = dict(required=False, default='present', choices=['present', 'absent']) + name=dict(required=True), + target_bucket=dict(required=False, default=None), + target_prefix=dict(required=False, default=""), + state=dict(required=False, default='present', choices=['present', 'absent']) ) ) diff --git a/lib/ansible/modules/cloud/amazon/s3_sync.py b/lib/ansible/modules/cloud/amazon/s3_sync.py index 54274e88af..9bebc19455 100644 --- a/lib/ansible/modules/cloud/amazon/s3_sync.py +++ b/lib/ansible/modules/cloud/amazon/s3_sync.py @@ -256,7 +256,6 @@ DEFAULT_CHUNK_SIZE = 5 * 1024 * 1024 def calculate_multipart_etag(source_path, chunk_size=DEFAULT_CHUNK_SIZE): - """ calculates a multipart upload etag for amazon s3 diff --git a/lib/ansible/modules/cloud/amazon/s3_website.py b/lib/ansible/modules/cloud/amazon/s3_website.py index 43658181f1..5cb0dc9b9f 100644 --- a/lib/ansible/modules/cloud/amazon/s3_website.py +++ b/lib/ansible/modules/cloud/amazon/s3_website.py @@ -180,10 +180,10 @@ def _create_website_configuration(suffix, error_key, redirect_all_requests): website_configuration = {} if error_key is not None: - website_configuration['ErrorDocument'] = { 'Key': error_key } + website_configuration['ErrorDocument'] = {'Key': error_key} if suffix is not None: - website_configuration['IndexDocument'] = { 'Suffix': suffix } + website_configuration['IndexDocument'] = {'Suffix': suffix} if redirect_all_requests is not None: website_configuration['RedirectAllRequestsTo'] = _create_redirect_dict(redirect_all_requests) @@ -288,10 +288,10 @@ def main(): module = AnsibleModule( argument_spec=argument_spec, - mutually_exclusive = [ + mutually_exclusive=[ ['redirect_all_requests', 'suffix'], ['redirect_all_requests', 'error_key'] - ]) + ]) if not HAS_BOTO3: module.fail_json(msg='boto3 required for this module') diff --git a/lib/ansible/modules/cloud/amazon/sns_topic.py b/lib/ansible/modules/cloud/amazon/sns_topic.py index e0e3ae6a52..fe7c1115ee 100644 --- a/lib/ansible/modules/cloud/amazon/sns_topic.py +++ b/lib/ansible/modules/cloud/amazon/sns_topic.py @@ -176,7 +176,7 @@ class SnsTopicManager(object): def _get_boto_connection(self): try: return connect_to_aws(boto.sns, self.region, - **self.aws_connect_params) + **self.aws_connect_params) except BotoServerError as err: self.module.fail_json(msg=err.message) @@ -194,7 +194,6 @@ class SnsTopicManager(object): break return [t['TopicArn'] for t in topics] - def _arn_topic_lookup(self): # topic names cannot have colons, so this captures the full topic name all_topics = self._get_all_topics() @@ -203,7 +202,6 @@ class SnsTopicManager(object): if topic.endswith(lookup_topic): return topic - def _create_topic(self): self.changed = True self.topic_created = True @@ -214,57 +212,51 @@ class SnsTopicManager(object): time.sleep(3) self.arn_topic = self._arn_topic_lookup() - def _set_topic_attrs(self): - topic_attributes = self.connection.get_topic_attributes(self.arn_topic) \ - ['GetTopicAttributesResponse'] ['GetTopicAttributesResult'] \ - ['Attributes'] + topic_attributes = self.connection.get_topic_attributes(self.arn_topic)['GetTopicAttributesResponse']['GetTopicAttributesResult']['Attributes'] if self.display_name and self.display_name != topic_attributes['DisplayName']: self.changed = True self.attributes_set.append('display_name') if not self.check_mode: self.connection.set_topic_attributes(self.arn_topic, 'DisplayName', - self.display_name) + self.display_name) if self.policy and self.policy != json.loads(topic_attributes['Policy']): self.changed = True self.attributes_set.append('policy') if not self.check_mode: self.connection.set_topic_attributes(self.arn_topic, 'Policy', - json.dumps(self.policy)) + json.dumps(self.policy)) - if self.delivery_policy and ('DeliveryPolicy' not in topic_attributes or \ - self.delivery_policy != json.loads(topic_attributes['DeliveryPolicy'])): + if self.delivery_policy and ('DeliveryPolicy' not in topic_attributes or + self.delivery_policy != json.loads(topic_attributes['DeliveryPolicy'])): self.changed = True self.attributes_set.append('delivery_policy') if not self.check_mode: self.connection.set_topic_attributes(self.arn_topic, 'DeliveryPolicy', - json.dumps(self.delivery_policy)) - + json.dumps(self.delivery_policy)) def _canonicalize_endpoint(self, protocol, endpoint): if protocol == 'sms': return re.sub('[^0-9]*', '', endpoint) return endpoint - def _get_topic_subs(self): next_token = None while True: response = self.connection.get_all_subscriptions_by_topic(self.arn_topic, next_token) - self.subscriptions_existing.extend(response['ListSubscriptionsByTopicResponse'] \ - ['ListSubscriptionsByTopicResult']['Subscriptions']) - next_token = response['ListSubscriptionsByTopicResponse'] \ - ['ListSubscriptionsByTopicResult']['NextToken'] + self.subscriptions_existing.extend(response['ListSubscriptionsByTopicResponse'] + ['ListSubscriptionsByTopicResult']['Subscriptions']) + next_token = response['ListSubscriptionsByTopicResponse']['ListSubscriptionsByTopicResult']['NextToken'] if not next_token: break def _set_topic_subs(self): subscriptions_existing_list = [] desired_subscriptions = [(sub['protocol'], - self._canonicalize_endpoint(sub['protocol'], sub['endpoint'])) for sub in - self.subscriptions] + self._canonicalize_endpoint(sub['protocol'], sub['endpoint'])) for sub in + self.subscriptions] if self.subscriptions_existing: for sub in self.subscriptions_existing: @@ -284,7 +276,6 @@ class SnsTopicManager(object): if not self.check_mode: self.connection.subscribe(self.arn_topic, protocol, endpoint) - def _delete_subscriptions(self): # NOTE: subscriptions in 'PendingConfirmation' timeout in 3 days # https://forums.aws.amazon.com/thread.jspa?threadID=85993 @@ -295,14 +286,12 @@ class SnsTopicManager(object): if not self.check_mode: self.connection.unsubscribe(sub['SubscriptionArn']) - def _delete_topic(self): self.topic_deleted = True self.changed = True if not self.check_mode: self.connection.delete_topic(self.arn_topic) - def ensure_ok(self): self.arn_topic = self._arn_topic_lookup() if not self.arn_topic: @@ -319,7 +308,6 @@ class SnsTopicManager(object): self._delete_subscriptions() self._delete_topic() - def get_info(self): info = { 'name': self.name, @@ -341,14 +329,13 @@ class SnsTopicManager(object): return info - def main(): argument_spec = ec2_argument_spec() argument_spec.update( dict( name=dict(type='str', required=True), state=dict(type='str', default='present', choices=['present', - 'absent']), + 'absent']), display_name=dict(type='str', required=False), policy=dict(type='dict', required=False), delivery_policy=dict(type='dict', required=False), diff --git a/lib/ansible/modules/cloud/amazon/sts_assume_role.py b/lib/ansible/modules/cloud/amazon/sts_assume_role.py index 33099de864..f4bf58bea4 100644 --- a/lib/ansible/modules/cloud/amazon/sts_assume_role.py +++ b/lib/ansible/modules/cloud/amazon/sts_assume_role.py @@ -113,17 +113,18 @@ def assume_role_policy(connection, module): module.exit_json(changed=changed, sts_creds=assumed_role.credentials.__dict__, sts_user=assumed_role.user.__dict__) + def main(): argument_spec = ec2_argument_spec() argument_spec.update( dict( - role_arn = dict(required=True, default=None), - role_session_name = dict(required=True, default=None), - duration_seconds = dict(required=False, default=None, type='int'), - external_id = dict(required=False, default=None), - policy = dict(required=False, default=None), - mfa_serial_number = dict(required=False, default=None), - mfa_token = dict(required=False, default=None) + role_arn=dict(required=True, default=None), + role_session_name=dict(required=True, default=None), + duration_seconds=dict(required=False, default=None, type='int'), + external_id=dict(required=False, default=None), + policy=dict(required=False, default=None), + mfa_serial_number=dict(required=False, default=None), + mfa_token=dict(required=False, default=None) ) ) diff --git a/lib/ansible/modules/cloud/amazon/sts_session_token.py b/lib/ansible/modules/cloud/amazon/sts_session_token.py index 3d766683d7..12eda6e857 100644 --- a/lib/ansible/modules/cloud/amazon/sts_session_token.py +++ b/lib/ansible/modules/cloud/amazon/sts_session_token.py @@ -108,6 +108,7 @@ def normalize_credentials(credentials): 'expiration': expiration } + def get_session_token(connection, module): duration_seconds = module.params.get('duration_seconds') mfa_serial_number = module.params.get('mfa_serial_number') @@ -131,13 +132,14 @@ def get_session_token(connection, module): credentials = normalize_credentials(response.get('Credentials', {})) module.exit_json(changed=changed, sts_creds=credentials) + def main(): argument_spec = ec2_argument_spec() argument_spec.update( dict( - duration_seconds = dict(required=False, default=None, type='int'), - mfa_serial_number = dict(required=False, default=None), - mfa_token = dict(required=False, default=None) + duration_seconds=dict(required=False, default=None, type='int'), + mfa_serial_number=dict(required=False, default=None), + mfa_token=dict(required=False, default=None) ) ) diff --git a/lib/ansible/modules/cloud/azure/azure_rm_deployment.py b/lib/ansible/modules/cloud/azure/azure_rm_deployment.py index 00d8e0a97d..81df15ca59 100644 --- a/lib/ansible/modules/cloud/azure/azure_rm_deployment.py +++ b/lib/ansible/modules/cloud/azure/azure_rm_deployment.py @@ -512,7 +512,7 @@ class AzureRMDeploymentManager(AzureRMModuleBase): if self.wait_for_deployment_completion: deployment_result = self.get_poller_result(result) while deployment_result.properties is None or deployment_result.properties.provisioning_state not in ['Canceled', 'Failed', 'Deleted', - 'Succeeded']: + 'Succeeded']: time.sleep(self.wait_for_deployment_polling_period) deployment_result = self.rm_client.deployments.get(self.resource_group_name, self.deployment_name) except CloudError as exc: @@ -535,7 +535,7 @@ class AzureRMDeploymentManager(AzureRMModuleBase): """ try: result = self.rm_client.resource_groups.delete(self.resource_group_name) - result.wait() # Blocking wait till the delete is finished + result.wait() # Blocking wait till the delete is finished except CloudError as e: if e.status_code == 404 or e.status_code == 204: return @@ -569,7 +569,7 @@ class AzureRMDeploymentManager(AzureRMModuleBase): nested_deployment) except CloudError as exc: self.fail("List nested deployment operations failed with status code: %s and message: %s" % - (exc.status_code, exc.message)) + (exc.status_code, exc.message)) new_nested_operations = self._get_failed_nested_operations(nested_operations) new_operations += new_nested_operations return new_operations @@ -642,10 +642,10 @@ class AzureRMDeploymentManager(AzureRMModuleBase): def _get_ip_dict(self, ip): ip_dict = dict(name=ip.name, - id=ip.id, - public_ip=ip.ip_address, - public_ip_allocation_method=str(ip.public_ip_allocation_method) - ) + id=ip.id, + public_ip=ip.ip_address, + public_ip_allocation_method=str(ip.public_ip_allocation_method) + ) if ip.dns_settings: ip_dict['dns_settings'] = { 'domain_name_label': ip.dns_settings.domain_name_label, @@ -657,9 +657,9 @@ class AzureRMDeploymentManager(AzureRMModuleBase): return [self.network_client.public_ip_addresses.get(public_ip_id.split('/')[4], public_ip_id.split('/')[-1]) for nic_obj in (self.network_client.network_interfaces.get(self.resource_group_name, nic['dep'].resource_name) for nic in nics) - for public_ip_id in [ip_conf_instance.public_ip_address.id - for ip_conf_instance in nic_obj.ip_configurations - if ip_conf_instance.public_ip_address]] + for public_ip_id in [ip_conf_instance.public_ip_address.id + for ip_conf_instance in nic_obj.ip_configurations + if ip_conf_instance.public_ip_address]] def main(): diff --git a/lib/ansible/modules/cloud/azure/azure_rm_managed_disk.py b/lib/ansible/modules/cloud/azure/azure_rm_managed_disk.py index 81213908cb..13f027d176 100644 --- a/lib/ansible/modules/cloud/azure/azure_rm_managed_disk.py +++ b/lib/ansible/modules/cloud/azure/azure_rm_managed_disk.py @@ -154,6 +154,7 @@ def managed_disk_to_dict(managed_disk): class AzureRMManagedDisk(AzureRMModuleBase): """Configuration class for an Azure RM Managed Disk resource""" + def __init__(self): self.module_arg_spec = dict( resource_group=dict( diff --git a/lib/ansible/modules/cloud/azure/azure_rm_networkinterface.py b/lib/ansible/modules/cloud/azure/azure_rm_networkinterface.py index 9dc666a55c..942d6b0082 100644 --- a/lib/ansible/modules/cloud/azure/azure_rm_networkinterface.py +++ b/lib/ansible/modules/cloud/azure/azure_rm_networkinterface.py @@ -219,7 +219,7 @@ state: try: from msrestazure.azure_exceptions import CloudError from azure.mgmt.network.models import NetworkInterface, NetworkInterfaceIPConfiguration, Subnet, \ - PublicIPAddress, NetworkSecurityGroup + PublicIPAddress, NetworkSecurityGroup except ImportError: # This is handled in azure_rm_common pass @@ -442,7 +442,7 @@ class AzureRMNetworkInterface(AzureRMModuleBase): if not pip and self.public_ip: # create a default public_ip pip = self.create_default_pip(self.resource_group, self.location, self.name, - self.public_ip_allocation_method) + self.public_ip_allocation_method) nic = NetworkInterface( location=self.location, @@ -475,8 +475,7 @@ class AzureRMNetworkInterface(AzureRMModuleBase): tags=results['tags'], ip_configurations=[ NetworkInterfaceIPConfiguration( - private_ip_allocation_method= - results['ip_configuration']['private_ip_allocation_method'] + private_ip_allocation_method=results['ip_configuration']['private_ip_allocation_method'] ) ] ) @@ -496,7 +495,7 @@ class AzureRMNetworkInterface(AzureRMModuleBase): id=pip.id, location=pip.location, resource_guid=pip.resource_guid) - #name=pip.name, + # name=pip.name, if results['network_security_group'].get('id'): nsg = self.get_security_group(results['network_security_group']['name']) @@ -549,8 +548,8 @@ class AzureRMNetworkInterface(AzureRMModuleBase): subnet = self.network_client.subnets.get(self.resource_group, vnet_name, subnet_name) except Exception as exc: self.fail("Error: fetching subnet {0} in virtual network {1} - {2}".format(subnet_name, - vnet_name, - str(exc))) + vnet_name, + str(exc))) return subnet def get_security_group(self, name): diff --git a/lib/ansible/modules/cloud/azure/azure_rm_publicipaddress.py b/lib/ansible/modules/cloud/azure/azure_rm_publicipaddress.py index cfaa21dea0..2d5a8d7039 100644 --- a/lib/ansible/modules/cloud/azure/azure_rm_publicipaddress.py +++ b/lib/ansible/modules/cloud/azure/azure_rm_publicipaddress.py @@ -193,7 +193,7 @@ class AzureRMPublicIPAddress(AzureRMModuleBase): if self.domain_name != results['dns_settings'].get('domain_name_label'): self.log('CHANGED: domain_name_label') changed = True - results['dns_settings']['domain_name_label'] =self.domain_name + results['dns_settings']['domain_name_label'] = self.domain_name if self.allocation_method != results['public_ip_allocation_method']: self.log("CHANGED: allocation_method") diff --git a/lib/ansible/modules/cloud/azure/azure_rm_publicipaddress_facts.py b/lib/ansible/modules/cloud/azure/azure_rm_publicipaddress_facts.py index f427c8a562..49b2743e7c 100644 --- a/lib/ansible/modules/cloud/azure/azure_rm_publicipaddress_facts.py +++ b/lib/ansible/modules/cloud/azure/azure_rm_publicipaddress_facts.py @@ -183,7 +183,6 @@ class AzureRMPublicIPFacts(AzureRMModuleBase): return results - def main(): AzureRMPublicIPFacts() diff --git a/lib/ansible/modules/cloud/azure/azure_rm_storageaccount.py b/lib/ansible/modules/cloud/azure/azure_rm_storageaccount.py index 4ab2fcc25d..ff3948adef 100644 --- a/lib/ansible/modules/cloud/azure/azure_rm_storageaccount.py +++ b/lib/ansible/modules/cloud/azure/azure_rm_storageaccount.py @@ -155,7 +155,7 @@ try: from azure.common import AzureMissingResourceHttpError from azure.mgmt.storage.models import ProvisioningState, SkuName, SkuTier, Kind from azure.mgmt.storage.models import StorageAccountUpdateParameters, CustomDomain, \ - StorageAccountCreateParameters, Sku + StorageAccountCreateParameters, Sku except ImportError: # This is handled in azure_rm_common pass @@ -226,7 +226,7 @@ class AzureRMStorageAccount(AzureRMModuleBase): self.account_dict = self.get_account() if self.state == 'present' and self.account_dict and \ - self.account_dict['provisioning_state'] != AZURE_SUCCESS_STATE : + self.account_dict['provisioning_state'] != AZURE_SUCCESS_STATE: self.fail("Error: storage account {0} has not completed provisioning. State is {1}. Expecting state " "to be {2}.".format(self.name, self.account_dict['provisioning_state'], AZURE_SUCCESS_STATE)) @@ -280,7 +280,7 @@ class AzureRMStorageAccount(AzureRMModuleBase): resource_group=self.resource_group, type=account_obj.type, access_tier=(account_obj.access_tier.value - if account_obj.access_tier is not None else None), + if account_obj.access_tier is not None else None), sku_tier=account_obj.sku.tier.value, sku_name=account_obj.sku.name.value, provisioning_state=account_obj.provisioning_state.value, diff --git a/lib/ansible/modules/cloud/azure/azure_rm_subnet.py b/lib/ansible/modules/cloud/azure/azure_rm_subnet.py index 507ba2de9f..7e35e48351 100644 --- a/lib/ansible/modules/cloud/azure/azure_rm_subnet.py +++ b/lib/ansible/modules/cloud/azure/azure_rm_subnet.py @@ -132,7 +132,6 @@ except ImportError: pass - def subnet_to_dict(subnet): result = dict( id=subnet.id, diff --git a/lib/ansible/modules/cloud/azure/azure_rm_virtualmachine.py b/lib/ansible/modules/cloud/azure/azure_rm_virtualmachine.py index 23f3d7f871..9ac237ed5f 100644 --- a/lib/ansible/modules/cloud/azure/azure_rm_virtualmachine.py +++ b/lib/ansible/modules/cloud/azure/azure_rm_virtualmachine.py @@ -598,14 +598,14 @@ try: from msrestazure.azure_exceptions import CloudError from msrestazure.tools import parse_resource_id from azure.mgmt.compute.models import NetworkInterfaceReference, \ - VirtualMachine, HardwareProfile, \ - StorageProfile, OSProfile, OSDisk, DataDisk, \ - VirtualHardDisk, ManagedDiskParameters, \ - ImageReference, NetworkProfile, LinuxConfiguration, \ - SshConfiguration, SshPublicKey, VirtualMachineSizeTypes, \ - DiskCreateOptionTypes, Plan, SubResource + VirtualMachine, HardwareProfile, \ + StorageProfile, OSProfile, OSDisk, DataDisk, \ + VirtualHardDisk, ManagedDiskParameters, \ + ImageReference, NetworkProfile, LinuxConfiguration, \ + SshConfiguration, SshPublicKey, VirtualMachineSizeTypes, \ + DiskCreateOptionTypes, Plan, SubResource from azure.mgmt.network.models import PublicIPAddress, NetworkSecurityGroup, NetworkInterface, \ - NetworkInterfaceIPConfiguration, Subnet + NetworkInterfaceIPConfiguration, Subnet from azure.mgmt.storage.models import StorageAccountCreateParameters, Sku from azure.mgmt.storage.models import Kind, SkuTier, SkuName except ImportError: @@ -659,7 +659,7 @@ class AzureRMVirtualMachine(AzureRMModuleBase): open_ports=dict(type='list'), network_interface_names=dict(type='list', aliases=['network_interfaces']), remove_on_absent=dict(type='list', default=['all']), - virtual_network_resource_group=dict(type = 'str'), + virtual_network_resource_group=dict(type='str'), virtual_network_name=dict(type='str', aliases=['virtual_network']), subnet_name=dict(type='str', aliases=['subnet']), allocated=dict(type='bool', default=True), @@ -1297,7 +1297,7 @@ class AzureRMVirtualMachine(AzureRMModuleBase): nic_names = [] pip_names = [] - if self.remove_on_absent.intersection(set(['all','virtual_storage'])): + if self.remove_on_absent.intersection(set(['all', 'virtual_storage'])): # store the attached vhd info so we can nuke it after the VM is gone if(vm.storage_profile.os_disk.managed_disk): self.log('Storing managed disk ID for deletion') @@ -1319,7 +1319,7 @@ class AzureRMVirtualMachine(AzureRMModuleBase): self.log("Managed disk IDs to delete: {0}".format(', '.join(managed_disk_ids))) self.results['deleted_managed_disk_ids'] = managed_disk_ids - if self.remove_on_absent.intersection(set(['all','network_interfaces'])): + if self.remove_on_absent.intersection(set(['all', 'network_interfaces'])): # store the attached nic info so we can nuke them after the VM is gone self.log('Storing NIC names for deletion.') for interface in vm.network_profile.network_interfaces: @@ -1327,7 +1327,7 @@ class AzureRMVirtualMachine(AzureRMModuleBase): nic_names.append(id_dict['networkInterfaces']) self.log('NIC names to delete {0}'.format(', '.join(nic_names))) self.results['deleted_network_interfaces'] = nic_names - if self.remove_on_absent.intersection(set(['all','public_ips'])): + if self.remove_on_absent.intersection(set(['all', 'public_ips'])): # also store each nic's attached public IPs and delete after the NIC is gone for name in nic_names: nic = self.get_network_interface(name) @@ -1349,18 +1349,18 @@ class AzureRMVirtualMachine(AzureRMModuleBase): # TODO: parallelize nic, vhd, and public ip deletions with begin_deleting # TODO: best-effort to keep deleting other linked resources if we encounter an error - if self.remove_on_absent.intersection(set(['all','virtual_storage'])): + if self.remove_on_absent.intersection(set(['all', 'virtual_storage'])): self.log('Deleting VHDs') self.delete_vm_storage(vhd_uris) self.log('Deleting managed disks') self.delete_managed_disks(managed_disk_ids) - if self.remove_on_absent.intersection(set(['all','network_interfaces'])): + if self.remove_on_absent.intersection(set(['all', 'network_interfaces'])): self.log('Deleting network interfaces') for name in nic_names: self.delete_nic(name) - if self.remove_on_absent.intersection(set(['all','public_ips'])): + if self.remove_on_absent.intersection(set(['all', 'public_ips'])): self.log('Deleting public IPs') for name in pip_names: self.delete_pip(name) @@ -1461,6 +1461,7 @@ class AzureRMVirtualMachine(AzureRMModuleBase): return ImageReference(id=vm_image.id) self.fail("Error could not find image with name {0}".format(name)) + def get_availability_set(self, resource_group, name): try: return self.compute_client.availability_sets.get(resource_group, name) diff --git a/lib/ansible/modules/cloud/azure/azure_rm_virtualnetwork.py b/lib/ansible/modules/cloud/azure/azure_rm_virtualnetwork.py index 674743d222..07c92180e0 100644 --- a/lib/ansible/modules/cloud/azure/azure_rm_virtualnetwork.py +++ b/lib/ansible/modules/cloud/azure/azure_rm_virtualnetwork.py @@ -195,7 +195,7 @@ class AzureRMVirtualNetwork(AzureRMModuleBase): self.dns_servers = None self.purge_dns_servers = None - self.results=dict( + self.results = dict( changed=False, state=dict() ) @@ -327,7 +327,6 @@ class AzureRMVirtualNetwork(AzureRMModuleBase): self.delete_virtual_network() self.results['state']['status'] = 'Deleted' - return self.results def create_or_update_vnet(self, vnet): diff --git a/lib/ansible/modules/cloud/azure/azure_rm_virtualnetwork_facts.py b/lib/ansible/modules/cloud/azure/azure_rm_virtualnetwork_facts.py index 50dd1a95a9..1d78ad0522 100644 --- a/lib/ansible/modules/cloud/azure/azure_rm_virtualnetwork_facts.py +++ b/lib/ansible/modules/cloud/azure/azure_rm_virtualnetwork_facts.py @@ -178,6 +178,7 @@ class AzureRMNetworkInterfaceFacts(AzureRMModuleBase): results.append(self.serialize_obj(item, AZURE_OBJECT_CLASS)) return results + def main(): AzureRMNetworkInterfaceFacts() diff --git a/lib/ansible/modules/cloud/cloudstack/cs_host.py b/lib/ansible/modules/cloud/cloudstack/cs_host.py index 41d38248a8..3f3fdc9170 100644 --- a/lib/ansible/modules/cloud/cloudstack/cs_host.py +++ b/lib/ansible/modules/cloud/cloudstack/cs_host.py @@ -464,7 +464,7 @@ class AnsibleCloudStackHost(AnsibleCloudStack): # Set host allocationstate to be disabled/enabled elif host['resourcestate'].lower() in list(self.allocation_states_for_update.keys()): - host['allocationstate'] = self.allocation_states_for_update[host['resourcestate'].lower()] + host['allocationstate'] = self.allocation_states_for_update[host['resourcestate'].lower()] else: host['allocationstate'] = host['resourcestate'] diff --git a/lib/ansible/modules/cloud/docker/docker_container.py b/lib/ansible/modules/cloud/docker/docker_container.py index 6d41a7ca14..05fda7dc08 100644 --- a/lib/ansible/modules/cloud/docker/docker_container.py +++ b/lib/ansible/modules/cloud/docker/docker_container.py @@ -951,7 +951,7 @@ class TaskParameters(DockerBaseClass): Returns parameters used to create a HostConfig object ''' - host_config_params=dict( + host_config_params = dict( port_bindings='published_ports', publish_all_ports='publish_all_ports', links='links', @@ -1163,7 +1163,7 @@ class TaskParameters(DockerBaseClass): options = dict( Type=self.log_driver, - Config = dict() + Config=dict() ) if self.log_options is not None: @@ -1217,7 +1217,6 @@ class TaskParameters(DockerBaseClass): return network_id - class Container(DockerBaseClass): def __init__(self, container, parameters): @@ -1570,7 +1569,7 @@ class Container(DockerBaseClass): CgroupPermissions=parts[2], PathInContainer=parts[1], PathOnHost=parts[0] - )) + )) return expected_devices def _get_expected_entrypoint(self): diff --git a/lib/ansible/modules/cloud/docker/docker_image.py b/lib/ansible/modules/cloud/docker/docker_image.py index a5978fc545..c2e2d78cd1 100644 --- a/lib/ansible/modules/cloud/docker/docker_image.py +++ b/lib/ansible/modules/cloud/docker/docker_image.py @@ -430,7 +430,7 @@ class ImageManager(DockerBaseClass): if not self.check_mode: status = None try: - for line in self.client.push(repository, tag=tag, stream=True, decode=True): + for line in self.client.push(repository, tag=tag, stream=True, decode=True): self.log(line, pretty_print=True) if line.get('errorDetail'): raise Exception(line['errorDetail']['message']) diff --git a/lib/ansible/modules/cloud/docker/docker_image_facts.py b/lib/ansible/modules/cloud/docker/docker_image_facts.py index 3f1d6afe2b..c6dff60ab9 100644 --- a/lib/ansible/modules/cloud/docker/docker_image_facts.py +++ b/lib/ansible/modules/cloud/docker/docker_image_facts.py @@ -217,7 +217,7 @@ class ImageManager(DockerBaseClass): def main(): argument_spec = dict( name=dict(type='list'), - ) + ) client = AnsibleDockerClient( argument_spec=argument_spec diff --git a/lib/ansible/modules/cloud/docker/docker_network.py b/lib/ansible/modules/cloud/docker/docker_network.py index 1cf2c52108..e4a99b45db 100644 --- a/lib/ansible/modules/cloud/docker/docker_network.py +++ b/lib/ansible/modules/cloud/docker/docker_network.py @@ -184,6 +184,7 @@ class TaskParameters(DockerBaseClass): def container_names_in_network(network): return [c['Name'] for c in network['Containers'].values()] if network['Containers'] else [] + class DockerNetworkManager(object): def __init__(self, client): @@ -362,16 +363,16 @@ class DockerNetworkManager(object): def main(): argument_spec = dict( - network_name = dict(type='str', required=True, aliases=['name']), - connected = dict(type='list', default=[], aliases=['containers']), - state = dict(type='str', default='present', choices=['present', 'absent']), - driver = dict(type='str', default='bridge'), - driver_options = dict(type='dict', default={}), - force = dict(type='bool', default=False), - appends = dict(type='bool', default=False, aliases=['incremental']), - ipam_driver = dict(type='str', default=None), - ipam_options = dict(type='dict', default={}), - debug = dict(type='bool', default=False) + network_name=dict(type='str', required=True, aliases=['name']), + connected=dict(type='list', default=[], aliases=['containers']), + state=dict(type='str', default='present', choices=['present', 'absent']), + driver=dict(type='str', default='bridge'), + driver_options=dict(type='dict', default={}), + force=dict(type='bool', default=False), + appends=dict(type='bool', default=False, aliases=['incremental']), + ipam_driver=dict(type='str', default=None), + ipam_options=dict(type='dict', default={}), + debug=dict(type='bool', default=False) ) client = AnsibleDockerClient( diff --git a/lib/ansible/modules/cloud/google/gc_storage.py b/lib/ansible/modules/cloud/google/gc_storage.py index ee3f4edfc0..d190ca8b7f 100644 --- a/lib/ansible/modules/cloud/google/gc_storage.py +++ b/lib/ansible/modules/cloud/google/gc_storage.py @@ -179,27 +179,26 @@ def grant_check(module, gs, obj): try: acp = obj.get_acl() if module.params.get('permission') == 'public-read': - grant = [ x for x in acp.entries.entry_list if x.scope.type == 'AllUsers'] + grant = [x for x in acp.entries.entry_list if x.scope.type == 'AllUsers'] if not grant: obj.set_acl('public-read') module.exit_json(changed=True, result="The objects permission as been set to public-read") if module.params.get('permission') == 'authenticated-read': - grant = [ x for x in acp.entries.entry_list if x.scope.type == 'AllAuthenticatedUsers'] + grant = [x for x in acp.entries.entry_list if x.scope.type == 'AllAuthenticatedUsers'] if not grant: obj.set_acl('authenticated-read') module.exit_json(changed=True, result="The objects permission as been set to authenticated-read") except gs.provider.storage_response_error as e: - module.fail_json(msg= str(e)) + module.fail_json(msg=str(e)) return True - def key_check(module, gs, bucket, obj): try: bucket = gs.lookup(bucket) key_check = bucket.get_key(obj) except gs.provider.storage_response_error as e: - module.fail_json(msg= str(e)) + module.fail_json(msg=str(e)) if key_check: grant_check(module, gs, key_check) return True @@ -213,7 +212,7 @@ def keysum(module, gs, bucket, obj): if not key_check: return None md5_remote = key_check.etag[1:-1] - etag_multipart = '-' in md5_remote # Check for multipart, etag is not md5 + etag_multipart = '-' in md5_remote # Check for multipart, etag is not md5 if etag_multipart is True: module.fail_json(msg="Files uploaded with multipart of gs are not supported with checksum, unable to compute checksum.") return md5_remote @@ -223,7 +222,7 @@ def bucket_check(module, gs, bucket): try: result = gs.lookup(bucket) except gs.provider.storage_response_error as e: - module.fail_json(msg= str(e)) + module.fail_json(msg=str(e)) if result: grant_check(module, gs, result) return True @@ -237,7 +236,7 @@ def create_bucket(module, gs, bucket): bucket.set_acl(module.params.get('permission')) bucket.configure_versioning(module.params.get('versioning')) except gs.provider.storage_response_error as e: - module.fail_json(msg= str(e)) + module.fail_json(msg=str(e)) if bucket: return True @@ -251,7 +250,7 @@ def delete_bucket(module, gs, bucket): bucket.delete() return True except gs.provider.storage_response_error as e: - module.fail_json(msg= str(e)) + module.fail_json(msg=str(e)) def delete_key(module, gs, bucket, obj): @@ -260,7 +259,7 @@ def delete_key(module, gs, bucket, obj): bucket.delete_key(obj) module.exit_json(msg="Object deleted from bucket ", changed=True) except gs.provider.storage_response_error as e: - module.fail_json(msg= str(e)) + module.fail_json(msg=str(e)) def create_dirkey(module, gs, bucket, obj): @@ -270,7 +269,7 @@ def create_dirkey(module, gs, bucket, obj): key.set_contents_from_string('') module.exit_json(msg="Virtual directory %s created in bucket %s" % (obj, bucket.name), changed=True) except gs.provider.storage_response_error as e: - module.fail_json(msg= str(e)) + module.fail_json(msg=str(e)) def path_check(path): @@ -308,7 +307,7 @@ def upload_gsfile(module, gs, bucket, obj, src, expiry): url = key.generate_url(expiry) module.exit_json(msg="PUT operation complete", url=url, changed=True) except gs.provider.storage_copy_error as e: - module.fail_json(msg= str(e)) + module.fail_json(msg=str(e)) def download_gsfile(module, gs, bucket, obj, dest): @@ -318,7 +317,7 @@ def download_gsfile(module, gs, bucket, obj, dest): key.get_contents_to_filename(dest) module.exit_json(msg="GET operation complete", changed=True) except gs.provider.storage_copy_error as e: - module.fail_json(msg= str(e)) + module.fail_json(msg=str(e)) def download_gsstr(module, gs, bucket, obj): @@ -328,7 +327,7 @@ def download_gsstr(module, gs, bucket, obj): contents = key.get_contents_as_string() module.exit_json(msg="GET operation complete", contents=contents, changed=True) except gs.provider.storage_copy_error as e: - module.fail_json(msg= str(e)) + module.fail_json(msg=str(e)) def get_download_url(module, gs, bucket, obj, expiry): @@ -338,7 +337,7 @@ def get_download_url(module, gs, bucket, obj, expiry): url = key.generate_url(expiry) module.exit_json(msg="Download url:", url=url, expiration=expiry, changed=True) except gs.provider.storage_response_error as e: - module.fail_json(msg= str(e)) + module.fail_json(msg=str(e)) def handle_get(module, gs, bucket, obj, overwrite, dest): @@ -355,7 +354,7 @@ def handle_get(module, gs, bucket, obj, overwrite, dest): def handle_put(module, gs, bucket, obj, overwrite, src, expiration): # Lets check to see if bucket exists to get ground truth. bucket_rc = bucket_check(module, gs, bucket) - key_rc = key_check(module, gs, bucket, obj) + key_rc = key_check(module, gs, bucket, obj) # Lets check key state. Does it exist and if it does, compute the etag md5sum. if bucket_rc and key_rc: @@ -380,7 +379,7 @@ def handle_put(module, gs, bucket, obj, overwrite, src, expiration): def handle_delete(module, gs, bucket, obj): if bucket and not obj: if bucket_check(module, gs, bucket): - module.exit_json(msg="Bucket %s and all keys have been deleted."%bucket, changed=delete_bucket(module, gs, bucket)) + module.exit_json(msg="Bucket %s and all keys have been deleted." % bucket, changed=delete_bucket(module, gs, bucket)) else: module.exit_json(msg="Bucket does not exist.", changed=False) if bucket and obj: @@ -409,7 +408,7 @@ def handle_create(module, gs, bucket, obj): if bucket_check(module, gs, bucket): if key_check(module, gs, bucket, dirobj): - module.exit_json(msg="Bucket %s and key %s already exists."% (bucket, obj), changed=False) + module.exit_json(msg="Bucket %s and key %s already exists." % (bucket, obj), changed=False) else: create_dirkey(module, gs, bucket, dirobj) else: @@ -419,35 +418,35 @@ def handle_create(module, gs, bucket, obj): def main(): module = AnsibleModule( - argument_spec = dict( - bucket = dict(required=True), - object = dict(default=None, type='path'), - src = dict(default=None), - dest = dict(default=None, type='path'), - expiration = dict(type='int', default=600, aliases=['expiry']), - mode = dict(choices=['get', 'put', 'delete', 'create', 'get_url', 'get_str'], required=True), - permission = dict(choices=['private', 'public-read', 'authenticated-read'], default='private'), - headers = dict(type='dict', default={}), - gs_secret_key = dict(no_log=True, required=True), - gs_access_key = dict(required=True), - overwrite = dict(default=True, type='bool', aliases=['force']), - region = dict(default='US', type='str'), - versioning = dict(default='no', type='bool') + argument_spec=dict( + bucket=dict(required=True), + object=dict(default=None, type='path'), + src=dict(default=None), + dest=dict(default=None, type='path'), + expiration=dict(type='int', default=600, aliases=['expiry']), + mode=dict(choices=['get', 'put', 'delete', 'create', 'get_url', 'get_str'], required=True), + permission=dict(choices=['private', 'public-read', 'authenticated-read'], default='private'), + headers=dict(type='dict', default={}), + gs_secret_key=dict(no_log=True, required=True), + gs_access_key=dict(required=True), + overwrite=dict(default=True, type='bool', aliases=['force']), + region=dict(default='US', type='str'), + versioning=dict(default='no', type='bool') ), ) if not HAS_BOTO: module.fail_json(msg='boto 2.9+ required for this module') - bucket = module.params.get('bucket') - obj = module.params.get('object') - src = module.params.get('src') - dest = module.params.get('dest') - mode = module.params.get('mode') - expiry = module.params.get('expiration') + bucket = module.params.get('bucket') + obj = module.params.get('object') + src = module.params.get('src') + dest = module.params.get('dest') + mode = module.params.get('mode') + expiry = module.params.get('expiration') gs_secret_key = module.params.get('gs_secret_key') gs_access_key = module.params.get('gs_access_key') - overwrite = module.params.get('overwrite') + overwrite = module.params.get('overwrite') if mode == 'put': if not src or not object: @@ -459,7 +458,7 @@ def main(): try: gs = boto.connect_gs(gs_access_key, gs_secret_key) except boto.exception.NoAuthHandlerFound as e: - module.fail_json(msg = str(e)) + module.fail_json(msg=str(e)) if mode == 'get': if not bucket_check(module, gs, bucket) or not key_check(module, gs, bucket, obj): diff --git a/lib/ansible/modules/cloud/google/gcdns_record.py b/lib/ansible/modules/cloud/google/gcdns_record.py index d21807fd8f..a96e016c97 100644 --- a/lib/ansible/modules/cloud/google/gcdns_record.py +++ b/lib/ansible/modules/cloud/google/gcdns_record.py @@ -348,7 +348,7 @@ PROVIDER = Provider.GOOGLE # I'm hard-coding the supported record types here, because they (hopefully!) # shouldn't change much, and it allows me to use it as a "choices" parameter # in an AnsibleModule argument_spec. -SUPPORTED_RECORD_TYPES = [ 'A', 'AAAA', 'CNAME', 'SRV', 'TXT', 'SOA', 'NS', 'MX', 'SPF', 'PTR' ] +SUPPORTED_RECORD_TYPES = ['A', 'AAAA', 'CNAME', 'SRV', 'TXT', 'SOA', 'NS', 'MX', 'SPF', 'PTR'] ################################################################################ @@ -378,8 +378,8 @@ def create_record(module, gcdns, zone, record): # The record doesn't match, so we need to check if we can overwrite it. if not overwrite: module.fail_json( - msg = 'cannot overwrite existing record, overwrite protection enabled', - changed = False + msg='cannot overwrite existing record, overwrite protection enabled', + changed=False ) # The record either doesn't exist, or it exists and we can overwrite it. @@ -393,9 +393,9 @@ def create_record(module, gcdns, zone, record): # not when combined (e.g., an 'A' record with "www.example.com" # as its value). module.fail_json( - msg = 'value is invalid for the given type: ' + - "%s, got value: %s" % (record_type, record_data), - changed = False + msg='value is invalid for the given type: ' + + "%s, got value: %s" % (record_type, record_data), + changed=False ) elif error.code == 'cnameResourceRecordSetConflict': @@ -403,8 +403,8 @@ def create_record(module, gcdns, zone, record): # already have another type of resource record with the name # domain name. module.fail_json( - msg = "non-CNAME resource record already exists: %s" % record_name, - changed = False + msg="non-CNAME resource record already exists: %s" % record_name, + changed=False ) else: @@ -428,8 +428,8 @@ def create_record(module, gcdns, zone, record): try: gcdns.create_record(record.name, record.zone, record.type, record.data) module.fail_json( - msg = 'error updating record, the original record was restored', - changed = False + msg='error updating record, the original record was restored', + changed=False ) except LibcloudError: # We deleted the old record, couldn't create the new record, and @@ -437,12 +437,12 @@ def create_record(module, gcdns, zone, record): # record to the failure output so the user can resore it if # necessary. module.fail_json( - msg = 'error updating record, and could not restore original record, ' + - "original name: %s " % record.name + - "original zone: %s " % record.zone + - "original type: %s " % record.type + - "original data: %s" % record.data, - changed = True) + msg='error updating record, and could not restore original record, ' + + "original name: %s " % record.name + + "original zone: %s " % record.zone + + "original type: %s " % record.type + + "original data: %s" % record.data, + changed=True) return True @@ -450,8 +450,8 @@ def create_record(module, gcdns, zone, record): def remove_record(module, gcdns, record): """Remove a resource record.""" - overwrite = module.boolean(module.params['overwrite']) - ttl = module.params['ttl'] + overwrite = module.boolean(module.params['overwrite']) + ttl = module.params['ttl'] record_data = module.params['record_data'] # If there is no record, we're obviously done. @@ -463,10 +463,10 @@ def remove_record(module, gcdns, record): if not overwrite: if not _records_match(record.data['ttl'], record.data['rrdatas'], ttl, record_data): module.fail_json( - msg = 'cannot delete due to non-matching ttl or record_data: ' + - "ttl: %d, record_data: %s " % (ttl, record_data) + - "original ttl: %d, original record_data: %s" % (record.data['ttl'], record.data['rrdatas']), - changed = False + msg='cannot delete due to non-matching ttl or record_data: ' + + "ttl: %d, record_data: %s " % (ttl, record_data) + + "original ttl: %d, original record_data: %s" % (record.data['ttl'], record.data['rrdatas']), + changed=False ) # If we got to this point, we're okay to delete the record. @@ -529,30 +529,30 @@ def _records_match(old_ttl, old_record_data, new_ttl, new_record_data): def _sanity_check(module): """Run sanity checks that don't depend on info from the zone/record.""" - overwrite = module.params['overwrite'] + overwrite = module.params['overwrite'] record_name = module.params['record'] record_type = module.params['type'] - state = module.params['state'] - ttl = module.params['ttl'] + state = module.params['state'] + ttl = module.params['ttl'] record_data = module.params['record_data'] # Apache libcloud needs to be installed and at least the minimum version. if not HAS_LIBCLOUD: module.fail_json( - msg = 'This module requires Apache libcloud %s or greater' % MINIMUM_LIBCLOUD_VERSION, - changed = False + msg='This module requires Apache libcloud %s or greater' % MINIMUM_LIBCLOUD_VERSION, + changed=False ) elif LooseVersion(LIBCLOUD_VERSION) < MINIMUM_LIBCLOUD_VERSION: module.fail_json( - msg = 'This module requires Apache libcloud %s or greater' % MINIMUM_LIBCLOUD_VERSION, - changed = False + msg='This module requires Apache libcloud %s or greater' % MINIMUM_LIBCLOUD_VERSION, + changed=False ) # A negative TTL is not permitted (how would they even work?!). if ttl < 0: module.fail_json( - msg = 'TTL cannot be less than zero, got: %d' % ttl, - changed = False + msg='TTL cannot be less than zero, got: %d' % ttl, + changed=False ) # Deleting SOA records is not permitted. @@ -572,8 +572,8 @@ def _sanity_check(module): socket.inet_aton(value) except socket.error: module.fail_json( - msg = 'invalid A record value, got: %s' % value, - changed = False + msg='invalid A record value, got: %s' % value, + changed=False ) # AAAA records must contain valid IPv6 addresses. @@ -583,23 +583,23 @@ def _sanity_check(module): socket.inet_pton(socket.AF_INET6, value) except socket.error: module.fail_json( - msg = 'invalid AAAA record value, got: %s' % value, - changed = False + msg='invalid AAAA record value, got: %s' % value, + changed=False ) # CNAME and SOA records can't have multiple values. if record_type in ['CNAME', 'SOA'] and len(record_data) > 1: module.fail_json( - msg = 'CNAME or SOA records cannot have more than one value, ' + - "got: %s" % record_data, - changed = False + msg='CNAME or SOA records cannot have more than one value, ' + + "got: %s" % record_data, + changed=False ) # Google Cloud DNS does not support wildcard NS records. if record_type == 'NS' and record_name[0] == '*': module.fail_json( - msg = "wildcard NS records not allowed, got: %s" % record_name, - changed = False + msg="wildcard NS records not allowed, got: %s" % record_name, + changed=False ) # Values for txt records must begin and end with a double quote. @@ -607,32 +607,32 @@ def _sanity_check(module): for value in record_data: if value[0] != '"' and value[-1] != '"': module.fail_json( - msg = 'TXT record_data must be enclosed in double quotes, ' + - 'got: %s' % value, - changed = False + msg='TXT record_data must be enclosed in double quotes, ' + + 'got: %s' % value, + changed=False ) def _additional_sanity_checks(module, zone): """Run input sanity checks that depend on info from the zone/record.""" - overwrite = module.params['overwrite'] + overwrite = module.params['overwrite'] record_name = module.params['record'] record_type = module.params['type'] - state = module.params['state'] + state = module.params['state'] # CNAME records are not allowed to have the same name as the root domain. if record_type == 'CNAME' and record_name == zone.domain: module.fail_json( - msg = 'CNAME records cannot match the zone name', - changed = False + msg='CNAME records cannot match the zone name', + changed=False ) # The root domain must always have an NS record. if record_type == 'NS' and record_name == zone.domain and state == 'absent': module.fail_json( - msg = 'cannot delete root NS records', - changed = False + msg='cannot delete root NS records', + changed=False ) # Updating NS records with the name as the root domain is not allowed @@ -640,16 +640,16 @@ def _additional_sanity_checks(module, zone): # records cannot be removed. if record_type == 'NS' and record_name == zone.domain and overwrite: module.fail_json( - msg = 'cannot update existing root NS records', - changed = False + msg='cannot update existing root NS records', + changed=False ) # SOA records with names that don't match the root domain are not permitted # (and wouldn't make sense anyway). if record_type == 'SOA' and record_name != zone.domain: module.fail_json( - msg = 'non-root SOA records are not permitted, got: %s' % record_name, - changed = False + msg='non-root SOA records are not permitted, got: %s' % record_name, + changed=False ) @@ -661,46 +661,46 @@ def main(): """Main function""" module = AnsibleModule( - argument_spec = dict( - state = dict(default='present', choices=['present', 'absent'], type='str'), - record = dict(required=True, aliases=['name'], type='str'), - zone = dict(type='str'), - zone_id = dict(type='str'), - type = dict(required=True, choices=SUPPORTED_RECORD_TYPES, type='str'), - record_data = dict(aliases=['value'], type='list'), - ttl = dict(default=300, type='int'), - overwrite = dict(default=False, type='bool'), - service_account_email = dict(type='str'), - pem_file = dict(type='path'), - credentials_file = dict(type='path'), - project_id = dict(type='str') + argument_spec=dict( + state=dict(default='present', choices=['present', 'absent'], type='str'), + record=dict(required=True, aliases=['name'], type='str'), + zone=dict(type='str'), + zone_id=dict(type='str'), + type=dict(required=True, choices=SUPPORTED_RECORD_TYPES, type='str'), + record_data=dict(aliases=['value'], type='list'), + ttl=dict(default=300, type='int'), + overwrite=dict(default=False, type='bool'), + service_account_email=dict(type='str'), + pem_file=dict(type='path'), + credentials_file=dict(type='path'), + project_id=dict(type='str') ), - required_if = [ + required_if=[ ('state', 'present', ['record_data']), ('overwrite', False, ['record_data']) ], - required_one_of = [['zone', 'zone_id']], - supports_check_mode = True + required_one_of=[['zone', 'zone_id']], + supports_check_mode=True ) _sanity_check(module) record_name = module.params['record'] record_type = module.params['type'] - state = module.params['state'] - ttl = module.params['ttl'] - zone_name = module.params['zone'] - zone_id = module.params['zone_id'] + state = module.params['state'] + ttl = module.params['ttl'] + zone_name = module.params['zone'] + zone_id = module.params['zone_id'] json_output = dict( - state = state, - record = record_name, - zone = zone_name, - zone_id = zone_id, - type = record_type, - record_data = module.params['record_data'], - ttl = ttl, - overwrite = module.boolean(module.params['overwrite']) + state=state, + record=record_name, + zone=zone_name, + zone_id=zone_id, + type=record_type, + record_data=module.params['record_data'], + ttl=ttl, + overwrite=module.boolean(module.params['overwrite']) ) # Google Cloud DNS wants the trailing dot on all DNS names. @@ -718,13 +718,13 @@ def main(): zone = _get_zone(gcdns, zone_name, zone_id) if zone is None and zone_name is not None: module.fail_json( - msg = 'zone name was not found: %s' % zone_name, - changed = False + msg='zone name was not found: %s' % zone_name, + changed=False ) elif zone is None and zone_id is not None: module.fail_json( - msg = 'zone id was not found: %s' % zone_id, - changed = False + msg='zone id was not found: %s' % zone_id, + changed=False ) # Populate the returns with the actual zone information. @@ -738,8 +738,8 @@ def main(): except InvalidRequestError: # We gave Google Cloud DNS an invalid DNS record name. module.fail_json( - msg = 'record name is invalid: %s' % record_name, - changed = False + msg='record name is invalid: %s' % record_name, + changed=False ) _additional_sanity_checks(module, zone) @@ -752,20 +752,20 @@ def main(): diff['before_header'] = '' else: diff['before'] = dict( - record = record.data['name'], - type = record.data['type'], - record_data = record.data['rrdatas'], - ttl = record.data['ttl'] + record=record.data['name'], + type=record.data['type'], + record_data=record.data['rrdatas'], + ttl=record.data['ttl'] ) diff['before_header'] = "%s:%s" % (record_type, record_name) # Create, remove, or modify the record. if state == 'present': diff['after'] = dict( - record = record_name, - type = record_type, - record_data = module.params['record_data'], - ttl = ttl + record=record_name, + type=record_type, + record_data=module.params['record_data'], + ttl=ttl ) diff['after_header'] = "%s:%s" % (record_type, record_name) diff --git a/lib/ansible/modules/cloud/google/gcdns_zone.py b/lib/ansible/modules/cloud/google/gcdns_zone.py index a2511f82cd..cd7f89ea7d 100644 --- a/lib/ansible/modules/cloud/google/gcdns_zone.py +++ b/lib/ansible/modules/cloud/google/gcdns_zone.py @@ -145,18 +145,19 @@ MINIMUM_LIBCLOUD_VERSION = '0.19.0' PROVIDER = Provider.GOOGLE # The URL used to verify ownership of a zone in Google Cloud DNS. -ZONE_VERIFICATION_URL= 'https://www.google.com/webmasters/verification/' +ZONE_VERIFICATION_URL = 'https://www.google.com/webmasters/verification/' ################################################################################ # Functions ################################################################################ + def create_zone(module, gcdns, zone): """Creates a new Google Cloud DNS zone.""" description = module.params['description'] - extra = dict(description = description) - zone_name = module.params['zone'] + extra = dict(description=description) + zone_name = module.params['zone'] # Google Cloud DNS wants the trailing dot on the domain name. if zone_name[-1] != '.': @@ -184,8 +185,8 @@ def create_zone(module, gcdns, zone): # The zone name or a parameter might be completely invalid. This is # typically caused by an illegal DNS name (e.g. foo..com). module.fail_json( - msg = "zone name is not a valid DNS name: %s" % zone_name, - changed = False + msg="zone name is not a valid DNS name: %s" % zone_name, + changed=False ) elif error.code == 'managedZoneDnsNameNotAvailable': @@ -193,8 +194,8 @@ def create_zone(module, gcdns, zone): # names, such as TLDs, ccTLDs, or special domain names such as # example.com. module.fail_json( - msg = "zone name is reserved or already in use: %s" % zone_name, - changed = False + msg="zone name is reserved or already in use: %s" % zone_name, + changed=False ) elif error.code == 'verifyManagedZoneDnsNameOwnership': @@ -202,8 +203,8 @@ def create_zone(module, gcdns, zone): # it. This occurs when a user attempts to create a zone which shares # a domain name with a zone hosted elsewhere in Google Cloud DNS. module.fail_json( - msg = "ownership of zone %s needs to be verified at %s" % (zone_name, ZONE_VERIFICATION_URL), - changed = False + msg="ownership of zone %s needs to be verified at %s" % (zone_name, ZONE_VERIFICATION_URL), + changed=False ) else: @@ -226,8 +227,8 @@ def remove_zone(module, gcdns, zone): # refuse to remove the zone. if len(zone.list_records()) > 2: module.fail_json( - msg = "zone is not empty and cannot be removed: %s" % zone.domain, - changed = False + msg="zone is not empty and cannot be removed: %s" % zone.domain, + changed=False ) try: @@ -246,8 +247,8 @@ def remove_zone(module, gcdns, zone): # the milliseconds between the check and the removal command, # records were added to the zone. module.fail_json( - msg = "zone is not empty and cannot be removed: %s" % zone.domain, - changed = False + msg="zone is not empty and cannot be removed: %s" % zone.domain, + changed=False ) else: @@ -273,6 +274,7 @@ def _get_zone(gcdns, zone_name): return found_zone + def _sanity_check(module): """Run module sanity checks.""" @@ -281,40 +283,41 @@ def _sanity_check(module): # Apache libcloud needs to be installed and at least the minimum version. if not HAS_LIBCLOUD: module.fail_json( - msg = 'This module requires Apache libcloud %s or greater' % MINIMUM_LIBCLOUD_VERSION, - changed = False + msg='This module requires Apache libcloud %s or greater' % MINIMUM_LIBCLOUD_VERSION, + changed=False ) elif LooseVersion(LIBCLOUD_VERSION) < MINIMUM_LIBCLOUD_VERSION: module.fail_json( - msg = 'This module requires Apache libcloud %s or greater' % MINIMUM_LIBCLOUD_VERSION, - changed = False + msg='This module requires Apache libcloud %s or greater' % MINIMUM_LIBCLOUD_VERSION, + changed=False ) # Google Cloud DNS does not support the creation of TLDs. if '.' not in zone_name or len([label for label in zone_name.split('.') if label]) == 1: module.fail_json( - msg = 'cannot create top-level domain: %s' % zone_name, - changed = False + msg='cannot create top-level domain: %s' % zone_name, + changed=False ) ################################################################################ # Main ################################################################################ + def main(): """Main function""" module = AnsibleModule( - argument_spec = dict( - state = dict(default='present', choices=['present', 'absent'], type='str'), - zone = dict(required=True, aliases=['name'], type='str'), - description = dict(default='', type='str'), - service_account_email = dict(type='str'), - pem_file = dict(type='path'), - credentials_file = dict(type='path'), - project_id = dict(type='str') + argument_spec=dict( + state=dict(default='present', choices=['present', 'absent'], type='str'), + zone=dict(required=True, aliases=['name'], type='str'), + description=dict(default='', type='str'), + service_account_email=dict(type='str'), + pem_file=dict(type='path'), + credentials_file=dict(type='path'), + project_id=dict(type='str') ), - supports_check_mode = True + supports_check_mode=True ) _sanity_check(module) @@ -327,9 +330,9 @@ def main(): zone_name = zone_name + '.' json_output = dict( - state = state, - zone = zone_name, - description = module.params['description'] + state=state, + zone=zone_name, + description=module.params['description'] ) # Build a connection object that was can use to connect with Google @@ -347,16 +350,16 @@ def main(): diff['before_header'] = '' else: diff['before'] = dict( - zone = zone.domain, - description = zone.extra['description'] + zone=zone.domain, + description=zone.extra['description'] ) diff['before_header'] = zone_name # Create or remove the zone. if state == 'present': diff['after'] = dict( - zone = zone_name, - description = module.params['description'] + zone=zone_name, + description=module.params['description'] ) diff['after_header'] = zone_name diff --git a/lib/ansible/modules/cloud/lxc/lxc_container.py b/lib/ansible/modules/cloud/lxc/lxc_container.py index 15bcbc7617..c17cb8b076 100644 --- a/lib/ansible/modules/cloud/lxc/lxc_container.py +++ b/lib/ansible/modules/cloud/lxc/lxc_container.py @@ -377,7 +377,7 @@ EXAMPLES = """ - test-container-new-archive-destroyed-clone """ -RETURN=""" +RETURN = """ lxc_container: description: container information returned: success @@ -579,7 +579,7 @@ def create_script(command): f.close() # Ensure the script is executable. - os.chmod(script_file, int('0700',8)) + os.chmod(script_file, int('0700', 8)) # Output log file. stdout_file = os.fdopen(tempfile.mkstemp(prefix='lxc-attach-script-log')[0], 'ab') @@ -915,7 +915,7 @@ class LxcContainerManagement(object): 'ips': self.container.get_ips(), 'state': self._get_state(), 'init_pid': int(self.container.init_pid), - 'name' : self.container_name, + 'name': self.container_name, } def _unfreeze(self): @@ -1365,7 +1365,7 @@ class LxcContainerManagement(object): :type source_dir: ``str`` """ - old_umask = os.umask(int('0077',8)) + old_umask = os.umask(int('0077', 8)) archive_path = self.module.params.get('archive_path') if not os.path.isdir(archive_path): @@ -1750,7 +1750,7 @@ def main(): ) ), supports_check_mode=False, - required_if = ([ + required_if=([ ('archive', True, ['archive_path']) ]), ) diff --git a/lib/ansible/modules/cloud/lxd/lxd_container.py b/lib/ansible/modules/cloud/lxd/lxd_container.py index bf185c9d3b..12d1bc8965 100644 --- a/lib/ansible/modules/cloud/lxd/lxd_container.py +++ b/lib/ansible/modules/cloud/lxd/lxd_container.py @@ -216,7 +216,7 @@ EXAMPLES = ''' flat: true ''' -RETURN=''' +RETURN = ''' addresses: description: Mapping from the network device name to a list of IPv4 addresses in the container returned: when state is started or restarted @@ -328,7 +328,7 @@ class LXDContainerManagement(object): return ANSIBLE_LXD_STATES[resp_json['metadata']['status']] def _change_state(self, action, force_stop=False): - body_json={'action': action, 'timeout': self.timeout} + body_json = {'action': action, 'timeout': self.timeout} if force_stop: body_json['force'] = True return self.client.do('PUT', '/1.0/containers/{0}/state'.format(self.name), body_json=body_json) @@ -527,6 +527,7 @@ class LXDContainerManagement(object): fail_params['logs'] = e.kwargs['logs'] self.module.fail_json(**fail_params) + def main(): """Ansible Main module.""" @@ -585,7 +586,7 @@ def main(): type='str', default='{}/.config/lxc/client.crt'.format(os.environ['HOME']) ), - trust_password=dict( type='str', no_log=True ) + trust_password=dict(type='str', no_log=True) ), supports_check_mode=False, ) diff --git a/lib/ansible/modules/cloud/misc/rhevm.py b/lib/ansible/modules/cloud/misc/rhevm.py index 098ae76e0e..39e846397a 100644 --- a/lib/ansible/modules/cloud/misc/rhevm.py +++ b/lib/ansible/modules/cloud/misc/rhevm.py @@ -347,6 +347,7 @@ failed = False class RHEVConn(object): 'Connection to RHEV-M' + def __init__(self, module): self.module = module @@ -726,11 +727,11 @@ class RHEVConn(object): bond.append(ifacelist[slave]) try: tmpiface = params.Bonding( - slaves = params.Slaves(host_nic = bond), - options = params.Options( - option = [ - params.Option(name = 'miimon', value = '100'), - params.Option(name = 'mode', value = '4') + slaves=params.Slaves(host_nic=bond), + options=params.Options( + option=[ + params.Option(name='miimon', value='100'), + params.Option(name='mode', value='4') ] ) ) @@ -741,16 +742,16 @@ class RHEVConn(object): return False try: tmpnetwork = params.HostNIC( - network = params.Network(name = iface['network']), - name = iface['name'], - boot_protocol = iface['boot_protocol'], - ip = params.IP( - address = iface['ip'], - netmask = iface['netmask'], - gateway = iface['gateway'] + network=params.Network(name=iface['network']), + name=iface['name'], + boot_protocol=iface['boot_protocol'], + ip=params.IP( + address=iface['ip'], + netmask=iface['netmask'], + gateway=iface['gateway'] ), - override_configuration = True, - bonding = tmpiface) + override_configuration=True, + bonding=tmpiface) networklist.append(tmpnetwork) setMsg('Applying network ' + iface['name']) except Exception as e: @@ -760,13 +761,13 @@ class RHEVConn(object): return False else: tmpnetwork = params.HostNIC( - network = params.Network(name = iface['network']), - name = iface['name'], - boot_protocol = iface['boot_protocol'], - ip = params.IP( - address = iface['ip'], - netmask = iface['netmask'], - gateway = iface['gateway'] + network=params.Network(name=iface['network']), + name=iface['name'], + boot_protocol=iface['boot_protocol'], + ip=params.IP( + address=iface['ip'], + netmask=iface['netmask'], + gateway=iface['gateway'] )) networklist.append(tmpnetwork) setMsg('Applying network ' + iface['name']) @@ -828,8 +829,8 @@ class RHEVConn(object): try: HOST.nics.setupnetworks(params.Action( force=True, - check_connectivity = False, - host_nics = params.HostNics(host_nic = networklist) + check_connectivity=False, + host_nics=params.HostNics(host_nic=networklist) )) setMsg('nics are set') except Exception as e: @@ -1008,24 +1009,24 @@ class RHEV(object): VM = self.conn.get_VM(name) if VM: vminfo = dict() - vminfo['uuid'] = VM.id - vminfo['name'] = VM.name - vminfo['status'] = VM.status.state - vminfo['cpu_cores'] = VM.cpu.topology.cores + vminfo['uuid'] = VM.id + vminfo['name'] = VM.name + vminfo['status'] = VM.status.state + vminfo['cpu_cores'] = VM.cpu.topology.cores vminfo['cpu_sockets'] = VM.cpu.topology.sockets - vminfo['cpu_shares'] = VM.cpu_shares - vminfo['memory'] = (int(VM.memory) // 1024 // 1024 // 1024) - vminfo['mem_pol'] = (int(VM.memory_policy.guaranteed) // 1024 // 1024 // 1024) - vminfo['os'] = VM.get_os().type_ - vminfo['del_prot'] = VM.delete_protected + vminfo['cpu_shares'] = VM.cpu_shares + vminfo['memory'] = (int(VM.memory) // 1024 // 1024 // 1024) + vminfo['mem_pol'] = (int(VM.memory_policy.guaranteed) // 1024 // 1024 // 1024) + vminfo['os'] = VM.get_os().type_ + vminfo['del_prot'] = VM.delete_protected try: - vminfo['host'] = str(self.conn.get_Host_byid(str(VM.host.id)).name) + vminfo['host'] = str(self.conn.get_Host_byid(str(VM.host.id)).name) except Exception: - vminfo['host'] = None - vminfo['boot_order'] = [] + vminfo['host'] = None + vminfo['boot_order'] = [] for boot_dev in VM.os.get_boot(): vminfo['boot_order'].append(str(boot_dev.dev)) - vminfo['disks'] = [] + vminfo['disks'] = [] for DISK in VM.disks.list(): disk = dict() disk['name'] = DISK.name @@ -1033,7 +1034,7 @@ class RHEV(object): disk['domain'] = str((self.conn.get_domain_byid(DISK.get_storage_domains().get_storage_domain()[0].id)).name) disk['interface'] = DISK.interface vminfo['disks'].append(disk) - vminfo['ifaces'] = [] + vminfo['ifaces'] = [] for NIC in VM.nics.list(): iface = dict() iface['name'] = str(NIC.name) @@ -1083,17 +1084,17 @@ class RHEV(object): bootselect = True for disk in disks: - diskname = name + "_Disk" + str(counter) + "_" + disk.get('name', '').replace('/', '_') - disksize = disk.get('size', 1) - diskdomain = disk.get('domain', None) + diskname = name + "_Disk" + str(counter) + "_" + disk.get('name', '').replace('/', '_') + disksize = disk.get('size', 1) + diskdomain = disk.get('domain', None) if diskdomain is None: setMsg("`domain` is a required disk key.") setFailed() return False - diskinterface = disk.get('interface', 'virtio') - diskformat = disk.get('format', 'raw') + diskinterface = disk.get('interface', 'virtio') + diskformat = disk.get('format', 'raw') diskallocationtype = disk.get('thin', False) - diskboot = disk.get('bootable', False) + diskboot = disk.get('bootable', False) if bootselect is False and counter == 0: diskboot = True @@ -1175,7 +1176,7 @@ class RHEV(object): def setBootOrder(self, vmname, boot_order): self.__get_conn() VM = self.conn.get_VM(vmname) - bootorder = [] + bootorder = [] for boot_dev in VM.os.get_boot(): bootorder.append(str(boot_dev.dev)) @@ -1469,31 +1470,31 @@ def core(module): def main(): global module module = AnsibleModule( - argument_spec = dict( - state = dict(default='present', choices=['ping', 'present', 'absent', 'up', 'down', 'restarted', 'cd', 'info']), - user = dict(default="admin@internal"), - password = dict(required=True, no_log=True), - server = dict(default="127.0.0.1"), - port = dict(default="443"), - insecure_api = dict(default=False, type='bool'), - name = dict(), - image = dict(default=False), - datacenter = dict(default="Default"), - type = dict(default="server", choices=['server', 'desktop', 'host']), - cluster = dict(default=''), - vmhost = dict(default=False), - vmcpu = dict(default="2"), - vmmem = dict(default="1"), - disks = dict(), - osver = dict(default="rhel_6x64"), - ifaces = dict(aliases=['nics', 'interfaces']), - timeout = dict(default=False), - mempol = dict(default="1"), - vm_ha = dict(default=True), - cpu_share = dict(default="0"), - boot_order = dict(default=["network", "hd"]), - del_prot = dict(default=True, type="bool"), - cd_drive = dict(default=False) + argument_spec=dict( + state=dict(default='present', choices=['ping', 'present', 'absent', 'up', 'down', 'restarted', 'cd', 'info']), + user=dict(default="admin@internal"), + password=dict(required=True, no_log=True), + server=dict(default="127.0.0.1"), + port=dict(default="443"), + insecure_api=dict(default=False, type='bool'), + name=dict(), + image=dict(default=False), + datacenter=dict(default="Default"), + type=dict(default="server", choices=['server', 'desktop', 'host']), + cluster=dict(default=''), + vmhost=dict(default=False), + vmcpu=dict(default="2"), + vmmem=dict(default="1"), + disks=dict(), + osver=dict(default="rhel_6x64"), + ifaces=dict(aliases=['nics', 'interfaces']), + timeout=dict(default=False), + mempol=dict(default="1"), + vm_ha=dict(default=True), + cpu_share=dict(default="0"), + boot_order=dict(default=["network", "hd"]), + del_prot=dict(default=True, type="bool"), + cd_drive=dict(default=False) ), ) diff --git a/lib/ansible/modules/cloud/misc/serverless.py b/lib/ansible/modules/cloud/misc/serverless.py index a46ab7ac09..044e788242 100644 --- a/lib/ansible/modules/cloud/misc/serverless.py +++ b/lib/ansible/modules/cloud/misc/serverless.py @@ -154,13 +154,13 @@ def get_service_name(module, stage): def main(): module = AnsibleModule( argument_spec=dict( - service_path = dict(required=True, type='path'), - state = dict(default='present', choices=['present', 'absent'], required=False), - functions = dict(type='list', required=False), - region = dict(default='', required=False), - stage = dict(default='', required=False), - deploy = dict(default=True, type='bool', required=False), - serverless_bin_path = dict(required=False, type='path') + service_path=dict(required=True, type='path'), + state=dict(default='present', choices=['present', 'absent'], required=False), + functions=dict(type='list', required=False), + region=dict(default='', required=False), + stage=dict(default='', required=False), + deploy=dict(default=True, type='bool', required=False), + serverless_bin_path=dict(required=False, type='path') ), ) @@ -198,13 +198,13 @@ def main(): if rc != 0: if state == 'absent' and "-{}' does not exist".format(stage) in out: module.exit_json(changed=False, state='absent', command=command, - out=out, service_name=get_service_name(module, stage)) + out=out, service_name=get_service_name(module, stage)) module.fail_json(msg="Failure when executing Serverless command. Exited {}.\nstdout: {}\nstderr: {}".format(rc, out, err)) # gather some facts about the deployment module.exit_json(changed=True, state='present', out=out, command=command, - service_name=get_service_name(module, stage)) + service_name=get_service_name(module, stage)) if __name__ == '__main__': diff --git a/lib/ansible/modules/cloud/misc/virt_net.py b/lib/ansible/modules/cloud/misc/virt_net.py index 6759fc347f..96d61e6101 100644 --- a/lib/ansible/modules/cloud/misc/virt_net.py +++ b/lib/ansible/modules/cloud/misc/virt_net.py @@ -151,31 +151,32 @@ from ansible.module_utils._text import to_native VIRT_FAILED = 1 VIRT_SUCCESS = 0 -VIRT_UNAVAILABLE=2 +VIRT_UNAVAILABLE = 2 ALL_COMMANDS = [] ENTRY_COMMANDS = ['create', 'status', 'start', 'stop', 'undefine', 'destroy', 'get_xml', 'define', - 'modify' ] -HOST_COMMANDS = [ 'list_nets', 'facts', 'info' ] + 'modify'] +HOST_COMMANDS = ['list_nets', 'facts', 'info'] ALL_COMMANDS.extend(ENTRY_COMMANDS) ALL_COMMANDS.extend(HOST_COMMANDS) ENTRY_STATE_ACTIVE_MAP = { - 0 : "inactive", - 1 : "active" + 0: "inactive", + 1: "active" } ENTRY_STATE_AUTOSTART_MAP = { - 0 : "no", - 1 : "yes" + 0: "no", + 1: "yes" } ENTRY_STATE_PERSISTENT_MAP = { - 0 : "no", - 1 : "yes" + 0: "no", + 1: "yes" } + class EntryNotFound(Exception): pass @@ -245,9 +246,9 @@ class LibvirtConnection(object): if host is None: # add the host if not self.module.check_mode: - res = network.update (libvirt.VIR_NETWORK_UPDATE_COMMAND_ADD_LAST, - libvirt.VIR_NETWORK_SECTION_IP_DHCP_HOST, - -1, xml, libvirt.VIR_NETWORK_UPDATE_AFFECT_CURRENT) + res = network.update(libvirt.VIR_NETWORK_UPDATE_COMMAND_ADD_LAST, + libvirt.VIR_NETWORK_SECTION_IP_DHCP_HOST, + -1, xml, libvirt.VIR_NETWORK_UPDATE_AFFECT_CURRENT) else: # pretend there was a change res = 0 @@ -259,9 +260,9 @@ class LibvirtConnection(object): return False else: if not self.module.check_mode: - res = network.update (libvirt.VIR_NETWORK_UPDATE_COMMAND_MODIFY, - libvirt.VIR_NETWORK_SECTION_IP_DHCP_HOST, - -1, xml, libvirt.VIR_NETWORK_UPDATE_AFFECT_CURRENT) + res = network.update(libvirt.VIR_NETWORK_UPDATE_COMMAND_MODIFY, + libvirt.VIR_NETWORK_SECTION_IP_DHCP_HOST, + -1, xml, libvirt.VIR_NETWORK_UPDATE_AFFECT_CURRENT) else: # pretend there was a change res = 0 @@ -286,18 +287,18 @@ class LibvirtConnection(object): def get_status2(self, entry): state = entry.isActive() - return ENTRY_STATE_ACTIVE_MAP.get(state,"unknown") + return ENTRY_STATE_ACTIVE_MAP.get(state, "unknown") def get_status(self, entryid): if not self.module.check_mode: state = self.find_entry(entryid).isActive() - return ENTRY_STATE_ACTIVE_MAP.get(state,"unknown") + return ENTRY_STATE_ACTIVE_MAP.get(state, "unknown") else: try: state = self.find_entry(entryid).isActive() - return ENTRY_STATE_ACTIVE_MAP.get(state,"unknown") + return ENTRY_STATE_ACTIVE_MAP.get(state, "unknown") except: - return ENTRY_STATE_ACTIVE_MAP.get("inactive","unknown") + return ENTRY_STATE_ACTIVE_MAP.get("inactive", "unknown") def get_uuid(self, entryid): return self.find_entry(entryid).UUIDString() @@ -331,7 +332,7 @@ class LibvirtConnection(object): def get_autostart(self, entryid): state = self.find_entry(entryid).autostart() - return ENTRY_STATE_AUTOSTART_MAP.get(state,"unknown") + return ENTRY_STATE_AUTOSTART_MAP.get(state, "unknown") def get_autostart2(self, entryid): if not self.module.check_mode: @@ -358,7 +359,7 @@ class LibvirtConnection(object): def get_persistent(self, entryid): state = self.find_entry(entryid).isPersistent() - return ENTRY_STATE_PERSISTENT_MAP.get(state,"unknown") + return ENTRY_STATE_PERSISTENT_MAP.get(state, "unknown") def get_dhcp_leases(self, entryid): network = self.find_entry(entryid) @@ -398,7 +399,7 @@ class VirtNetwork(object): results = [] for entry in self.list_nets(): state_blurb = self.conn.get_status(entry) - results.append("%s %s" % (entry,state_blurb)) + results.append("%s %s" % (entry, state_blurb)) return results def autostart(self, entryid): @@ -481,11 +482,11 @@ class VirtNetwork(object): def core(module): - state = module.params.get('state', None) - name = module.params.get('name', None) - command = module.params.get('command', None) - uri = module.params.get('uri', None) - xml = module.params.get('xml', None) + state = module.params.get('state', None) + name = module.params.get('name', None) + command = module.params.get('command', None) + uri = module.params.get('uri', None) + xml = module.params.get('xml', None) autostart = module.params.get('autostart', None) v = VirtNetwork(uri, module) @@ -494,33 +495,33 @@ def core(module): if state and command == 'list_nets': res = v.list_nets(state=state) if not isinstance(res, dict): - res = { command: res } + res = {command: res} return VIRT_SUCCESS, res if state: if not name: - module.fail_json(msg = "state change requires a specified name") + module.fail_json(msg="state change requires a specified name") res['changed'] = False - if state in [ 'active' ]: + if state in ['active']: if v.status(name) is not 'active': res['changed'] = True res['msg'] = v.start(name) - elif state in [ 'present' ]: + elif state in ['present']: try: v.get_net(name) except EntryNotFound: if not xml: - module.fail_json(msg = "network '" + name + "' not present, but xml not specified") + module.fail_json(msg="network '" + name + "' not present, but xml not specified") v.define(name, xml) res = {'changed': True, 'created': name} - elif state in [ 'inactive' ]: + elif state in ['inactive']: entries = v.list_nets() if name in entries: if v.status(name) is not 'inactive': res['changed'] = True res['msg'] = v.destroy(name) - elif state in [ 'undefined', 'absent' ]: + elif state in ['undefined', 'absent']: entries = v.list_nets() if name in entries: if v.status(name) is not 'inactive': @@ -535,10 +536,10 @@ def core(module): if command: if command in ENTRY_COMMANDS: if not name: - module.fail_json(msg = "%s requires 1 argument: name" % command) + module.fail_json(msg="%s requires 1 argument: name" % command) if command in ('define', 'modify'): if not xml: - module.fail_json(msg = command+" requires xml argument") + module.fail_json(msg=command + " requires xml argument") try: v.get_net(name) except EntryNotFound: @@ -551,13 +552,13 @@ def core(module): return VIRT_SUCCESS, res res = getattr(v, command)(name) if not isinstance(res, dict): - res = { command: res } + res = {command: res} return VIRT_SUCCESS, res elif hasattr(v, command): res = getattr(v, command)() if not isinstance(res, dict): - res = { command: res } + res = {command: res} return VIRT_SUCCESS, res else: @@ -565,7 +566,7 @@ def core(module): if autostart is not None: if not name: - module.fail_json(msg = "state change requires a specified name") + module.fail_json(msg="state change requires a specified name") res['changed'] = False if autostart: @@ -584,16 +585,16 @@ def core(module): def main(): - module = AnsibleModule ( - argument_spec = dict( - name = dict(aliases=['network']), - state = dict(choices=['active', 'inactive', 'present', 'absent']), - command = dict(choices=ALL_COMMANDS), - uri = dict(default='qemu:///system'), - xml = dict(), - autostart = dict(type='bool') + module = AnsibleModule( + argument_spec=dict( + name=dict(aliases=['network']), + state=dict(choices=['active', 'inactive', 'present', 'absent']), + command=dict(choices=ALL_COMMANDS), + uri=dict(default='qemu:///system'), + xml=dict(), + autostart=dict(type='bool') ), - supports_check_mode = True + supports_check_mode=True ) if not HAS_VIRT: @@ -612,7 +613,7 @@ def main(): except Exception as e: module.fail_json(msg=str(e)) - if rc != 0: # something went wrong emit the msg + if rc != 0: # something went wrong emit the msg module.fail_json(rc=rc, msg=result) else: module.exit_json(**result) diff --git a/lib/ansible/modules/cloud/misc/virt_pool.py b/lib/ansible/modules/cloud/misc/virt_pool.py index 23a96a31e7..a3994a4ab2 100644 --- a/lib/ansible/modules/cloud/misc/virt_pool.py +++ b/lib/ansible/modules/cloud/misc/virt_pool.py @@ -165,49 +165,49 @@ from ansible.module_utils.basic import AnsibleModule VIRT_FAILED = 1 VIRT_SUCCESS = 0 -VIRT_UNAVAILABLE=2 +VIRT_UNAVAILABLE = 2 ALL_COMMANDS = [] ENTRY_COMMANDS = ['create', 'status', 'start', 'stop', 'build', 'delete', 'undefine', 'destroy', 'get_xml', 'define', 'refresh'] -HOST_COMMANDS = [ 'list_pools', 'facts', 'info' ] +HOST_COMMANDS = ['list_pools', 'facts', 'info'] ALL_COMMANDS.extend(ENTRY_COMMANDS) ALL_COMMANDS.extend(HOST_COMMANDS) ENTRY_STATE_ACTIVE_MAP = { - 0 : "inactive", - 1 : "active" + 0: "inactive", + 1: "active" } ENTRY_STATE_AUTOSTART_MAP = { - 0 : "no", - 1 : "yes" + 0: "no", + 1: "yes" } ENTRY_STATE_PERSISTENT_MAP = { - 0 : "no", - 1 : "yes" + 0: "no", + 1: "yes" } ENTRY_STATE_INFO_MAP = { - 0 : "inactive", - 1 : "building", - 2 : "running", - 3 : "degraded", - 4 : "inaccessible" + 0: "inactive", + 1: "building", + 2: "running", + 3: "degraded", + 4: "inaccessible" } ENTRY_BUILD_FLAGS_MAP = { - "new" : 0, - "repair" : 1, - "resize" : 2, - "no_overwrite" : 4, - "overwrite" : 8 + "new": 0, + "repair": 1, + "resize": 2, + "no_overwrite": 4, + "overwrite": 8 } ENTRY_DELETE_FLAGS_MAP = { - "normal" : 0, - "zeroed" : 1 + "normal": 0, + "zeroed": 1 } ALL_MODES = [] @@ -283,18 +283,18 @@ class LibvirtConnection(object): def get_status2(self, entry): state = entry.isActive() - return ENTRY_STATE_ACTIVE_MAP.get(state,"unknown") + return ENTRY_STATE_ACTIVE_MAP.get(state, "unknown") def get_status(self, entryid): if not self.module.check_mode: state = self.find_entry(entryid).isActive() - return ENTRY_STATE_ACTIVE_MAP.get(state,"unknown") + return ENTRY_STATE_ACTIVE_MAP.get(state, "unknown") else: try: state = self.find_entry(entryid).isActive() - return ENTRY_STATE_ACTIVE_MAP.get(state,"unknown") + return ENTRY_STATE_ACTIVE_MAP.get(state, "unknown") except: - return ENTRY_STATE_ACTIVE_MAP.get("inactive","unknown") + return ENTRY_STATE_ACTIVE_MAP.get("inactive", "unknown") def get_uuid(self, entryid): return self.find_entry(entryid).UUIDString() @@ -378,7 +378,7 @@ class LibvirtConnection(object): def get_autostart(self, entryid): state = self.find_entry(entryid).autostart() - return ENTRY_STATE_AUTOSTART_MAP.get(state,"unknown") + return ENTRY_STATE_AUTOSTART_MAP.get(state, "unknown") def get_autostart2(self, entryid): if not self.module.check_mode: @@ -405,7 +405,7 @@ class LibvirtConnection(object): def get_persistent(self, entryid): state = self.find_entry(entryid).isPersistent() - return ENTRY_STATE_PERSISTENT_MAP.get(state,"unknown") + return ENTRY_STATE_PERSISTENT_MAP.get(state, "unknown") def define_from_xml(self, entryid, xml): if not self.module.check_mode: @@ -441,7 +441,7 @@ class VirtStoragePool(object): results = [] for entry in self.list_pools(): state_blurb = self.conn.get_status(entry) - results.append("%s %s" % (entry,state_blurb)) + results.append("%s %s" % (entry, state_blurb)) return results def autostart(self, entryid): @@ -478,10 +478,10 @@ class VirtStoragePool(object): return self.conn.define_from_xml(entryid, xml) def build(self, entryid, flags): - return self.conn.build(entryid, ENTRY_BUILD_FLAGS_MAP.get(flags,0)) + return self.conn.build(entryid, ENTRY_BUILD_FLAGS_MAP.get(flags, 0)) def delete(self, entryid, flags): - return self.conn.delete(entryid, ENTRY_DELETE_FLAGS_MAP.get(flags,0)) + return self.conn.delete(entryid, ENTRY_DELETE_FLAGS_MAP.get(flags, 0)) def refresh(self, entryid): return self.conn.refresh(entryid) @@ -501,10 +501,10 @@ class VirtStoragePool(object): # assume the other end of the xmlrpc connection can figure things # out or doesn't care. results[entry] = { - "status" : ENTRY_STATE_INFO_MAP.get(data[0],"unknown"), - "size_total" : str(data[1]), - "size_used" : str(data[2]), - "size_available" : str(data[3]), + "status": ENTRY_STATE_INFO_MAP.get(data[0], "unknown"), + "size_total": str(data[1]), + "size_used": str(data[2]), + "size_available": str(data[3]), } results[entry]["autostart"] = self.conn.get_autostart(entry) results[entry]["persistent"] = self.conn.get_persistent(entry) @@ -555,13 +555,13 @@ class VirtStoragePool(object): def core(module): - state = module.params.get('state', None) - name = module.params.get('name', None) - command = module.params.get('command', None) - uri = module.params.get('uri', None) - xml = module.params.get('xml', None) + state = module.params.get('state', None) + name = module.params.get('name', None) + command = module.params.get('command', None) + uri = module.params.get('uri', None) + xml = module.params.get('xml', None) autostart = module.params.get('autostart', None) - mode = module.params.get('mode', None) + mode = module.params.get('mode', None) v = VirtStoragePool(uri, module) res = {} @@ -569,40 +569,40 @@ def core(module): if state and command == 'list_pools': res = v.list_pools(state=state) if not isinstance(res, dict): - res = { command: res } + res = {command: res} return VIRT_SUCCESS, res if state: if not name: - module.fail_json(msg = "state change requires a specified name") + module.fail_json(msg="state change requires a specified name") res['changed'] = False - if state in [ 'active' ]: + if state in ['active']: if v.status(name) is not 'active': res['changed'] = True res['msg'] = v.start(name) - elif state in [ 'present' ]: + elif state in ['present']: try: v.get_pool(name) except EntryNotFound: if not xml: - module.fail_json(msg = "storage pool '" + name + "' not present, but xml not specified") + module.fail_json(msg="storage pool '" + name + "' not present, but xml not specified") v.define(name, xml) res = {'changed': True, 'created': name} - elif state in [ 'inactive' ]: + elif state in ['inactive']: entries = v.list_pools() if name in entries: if v.status(name) is not 'inactive': res['changed'] = True res['msg'] = v.destroy(name) - elif state in [ 'undefined', 'absent' ]: + elif state in ['undefined', 'absent']: entries = v.list_pools() if name in entries: if v.status(name) is not 'inactive': v.destroy(name) res['changed'] = True res['msg'] = v.undefine(name) - elif state in [ 'deleted' ]: + elif state in ['deleted']: entries = v.list_pools() if name in entries: if v.status(name) is not 'inactive': @@ -618,10 +618,10 @@ def core(module): if command: if command in ENTRY_COMMANDS: if not name: - module.fail_json(msg = "%s requires 1 argument: name" % command) + module.fail_json(msg="%s requires 1 argument: name" % command) if command == 'define': if not xml: - module.fail_json(msg = "define requires xml argument") + module.fail_json(msg="define requires xml argument") try: v.get_pool(name) except EntryNotFound: @@ -631,22 +631,22 @@ def core(module): elif command == 'build': res = v.build(name, mode) if not isinstance(res, dict): - res = { 'changed': True, command: res } + res = {'changed': True, command: res} return VIRT_SUCCESS, res elif command == 'delete': res = v.delete(name, mode) if not isinstance(res, dict): - res = { 'changed': True, command: res } + res = {'changed': True, command: res} return VIRT_SUCCESS, res res = getattr(v, command)(name) if not isinstance(res, dict): - res = { command: res } + res = {command: res} return VIRT_SUCCESS, res elif hasattr(v, command): res = getattr(v, command)() if not isinstance(res, dict): - res = { command: res } + res = {command: res} return VIRT_SUCCESS, res else: @@ -654,7 +654,7 @@ def core(module): if autostart is not None: if not name: - module.fail_json(msg = "state change requires a specified name") + module.fail_json(msg="state change requires a specified name") res['changed'] = False if autostart: @@ -673,17 +673,17 @@ def core(module): def main(): - module = AnsibleModule ( - argument_spec = dict( - name = dict(aliases=['pool']), - state = dict(choices=['active', 'inactive', 'present', 'absent', 'undefined', 'deleted']), - command = dict(choices=ALL_COMMANDS), - uri = dict(default='qemu:///system'), - xml = dict(), - autostart = dict(type='bool'), - mode = dict(choices=ALL_MODES), + module = AnsibleModule( + argument_spec=dict( + name=dict(aliases=['pool']), + state=dict(choices=['active', 'inactive', 'present', 'absent', 'undefined', 'deleted']), + command=dict(choices=ALL_COMMANDS), + uri=dict(default='qemu:///system'), + xml=dict(), + autostart=dict(type='bool'), + mode=dict(choices=ALL_MODES), ), - supports_check_mode = True + supports_check_mode=True ) if not HAS_VIRT: @@ -702,7 +702,7 @@ def main(): except Exception as e: module.fail_json(msg=str(e)) - if rc != 0: # something went wrong emit the msg + if rc != 0: # something went wrong emit the msg module.fail_json(rc=rc, msg=result) else: module.exit_json(**result) diff --git a/lib/ansible/modules/cloud/misc/xenserver_facts.py b/lib/ansible/modules/cloud/misc/xenserver_facts.py index b1704f0c88..ba995381d7 100644 --- a/lib/ansible/modules/cloud/misc/xenserver_facts.py +++ b/lib/ansible/modules/cloud/misc/xenserver_facts.py @@ -137,12 +137,14 @@ def change_keys(recs, key='uuid', filter_func=None): return new_recs + def get_host(session): """Get the host""" host_recs = session.xenapi.host.get_all() # We only have one host, so just return its entry return session.xenapi.host.get_record(host_recs[0]) + def get_vms(session): xs_vms = {} recs = session.xenapi.VM.get_all() @@ -165,6 +167,7 @@ def get_srs(session): xs_srs[sr['name_label']] = sr return xs_srs + def main(): module = AnsibleModule({}) diff --git a/lib/ansible/modules/cloud/openstack/os_image.py b/lib/ansible/modules/cloud/openstack/os_image.py index f9b6c7539f..faf090de3c 100644 --- a/lib/ansible/modules/cloud/openstack/os_image.py +++ b/lib/ansible/modules/cloud/openstack/os_image.py @@ -137,20 +137,20 @@ from ansible.module_utils.openstack import openstack_full_argument_spec, opensta def main(): argument_spec = openstack_full_argument_spec( - name = dict(required=True), - id = dict(default=None), - checksum = dict(default=None), - disk_format = dict(default='qcow2', choices=['ami', 'ari', 'aki', 'vhd', 'vmdk', 'raw', 'qcow2', 'vdi', 'iso', 'vhdx', 'ploop']), - container_format = dict(default='bare', choices=['ami', 'aki', 'ari', 'bare', 'ovf', 'ova', 'docker']), - owner = dict(default=None), - min_disk = dict(type='int', default=0), - min_ram = dict(type='int', default=0), - is_public = dict(type='bool', default=False), - filename = dict(default=None), - ramdisk = dict(default=None), - kernel = dict(default=None), - properties = dict(type='dict', default={}), - state = dict(default='present', choices=['absent', 'present']), + name=dict(required=True), + id=dict(default=None), + checksum=dict(default=None), + disk_format=dict(default='qcow2', choices=['ami', 'ari', 'aki', 'vhd', 'vmdk', 'raw', 'qcow2', 'vdi', 'iso', 'vhdx', 'ploop']), + container_format=dict(default='bare', choices=['ami', 'aki', 'ari', 'bare', 'ovf', 'ova', 'docker']), + owner=dict(default=None), + min_disk=dict(type='int', default=0), + min_ram=dict(type='int', default=0), + is_public=dict(type='bool', default=False), + filename=dict(default=None), + ramdisk=dict(default=None), + kernel=dict(default=None), + properties=dict(type='dict', default={}), + state=dict(default='present', choices=['absent', 'present']), ) module_kwargs = openstack_module_kwargs() module = AnsibleModule(argument_spec, **module_kwargs) @@ -163,13 +163,13 @@ def main(): changed = False if module.params['checksum']: - image = cloud.get_image(name_or_id=None,filters={'checksum': module.params['checksum']}) + image = cloud.get_image(name_or_id=None, filters={'checksum': module.params['checksum']}) else: image = cloud.get_image(name_or_id=module.params['name']) if module.params['state'] == 'present': if not image: - kwargs={} + kwargs = {} if module.params['id'] is not None: kwargs['id'] = module.params['id'] image = cloud.create_image( diff --git a/lib/ansible/modules/cloud/openstack/os_keypair.py b/lib/ansible/modules/cloud/openstack/os_keypair.py index 8b4e8db0b5..bc9eff7dfe 100644 --- a/lib/ansible/modules/cloud/openstack/os_keypair.py +++ b/lib/ansible/modules/cloud/openstack/os_keypair.py @@ -109,11 +109,11 @@ def _system_state_change(module, keypair): def main(): argument_spec = openstack_full_argument_spec( - name = dict(required=True), - public_key = dict(default=None), - public_key_file = dict(default=None), - state = dict(default='present', - choices=['absent', 'present']), + name=dict(required=True), + public_key=dict(default=None), + public_key_file=dict(default=None), + state=dict(default='present', + choices=['absent', 'present']), ) module_kwargs = openstack_module_kwargs( diff --git a/lib/ansible/modules/cloud/openstack/os_nova_flavor.py b/lib/ansible/modules/cloud/openstack/os_nova_flavor.py index 03a2f646b7..6d2d5f6ebc 100644 --- a/lib/ansible/modules/cloud/openstack/os_nova_flavor.py +++ b/lib/ansible/modules/cloud/openstack/os_nova_flavor.py @@ -194,21 +194,21 @@ def _system_state_change(module, flavor): def main(): argument_spec = openstack_full_argument_spec( - state = dict(required=False, default='present', - choices=['absent', 'present']), - name = dict(required=False), + state=dict(required=False, default='present', + choices=['absent', 'present']), + name=dict(required=False), # required when state is 'present' - ram = dict(required=False, type='int'), - vcpus = dict(required=False, type='int'), - disk = dict(required=False, type='int'), - - ephemeral = dict(required=False, default=0, type='int'), - swap = dict(required=False, default=0, type='int'), - rxtx_factor = dict(required=False, default=1.0, type='float'), - is_public = dict(required=False, default=True, type='bool'), - flavorid = dict(required=False, default="auto"), - extra_specs = dict(required=False, default=None, type='dict'), + ram=dict(required=False, type='int'), + vcpus=dict(required=False, type='int'), + disk=dict(required=False, type='int'), + + ephemeral=dict(required=False, default=0, type='int'), + swap=dict(required=False, default=0, type='int'), + rxtx_factor=dict(required=False, default=1.0, type='float'), + is_public=dict(required=False, default=True, type='bool'), + flavorid=dict(required=False, default="auto"), + extra_specs=dict(required=False, default=None, type='dict'), ) module_kwargs = openstack_module_kwargs() @@ -247,9 +247,9 @@ def main(): rxtx_factor=module.params['rxtx_factor'], is_public=module.params['is_public'] ) - changed=True + changed = True else: - changed=False + changed = False old_extra_specs = flavor['extra_specs'] new_extra_specs = dict([(k, str(v)) for k, v in extra_specs.items()]) diff --git a/lib/ansible/modules/cloud/openstack/os_quota.py b/lib/ansible/modules/cloud/openstack/os_quota.py index 7a6a6c1b54..c319b12ddb 100644 --- a/lib/ansible/modules/cloud/openstack/os_quota.py +++ b/lib/ansible/modules/cloud/openstack/os_quota.py @@ -306,14 +306,17 @@ def _get_volume_quotas(cloud, project): return cloud.get_volume_quotas(project) + def _get_network_quotas(cloud, project): return cloud.get_network_quotas(project) + def _get_compute_quotas(cloud, project): return cloud.get_compute_quotas(project) + def _get_quotas(module, cloud, project): quota = {} @@ -334,6 +337,7 @@ def _get_quotas(module, cloud, project): return quota + def _scrub_results(quota): filter_attr = [ @@ -350,6 +354,7 @@ def _scrub_results(quota): return quota + def _system_state_change_details(module, project_quota_output): quota_change_request = {} @@ -368,6 +373,7 @@ def _system_state_change_details(module, project_quota_output): return (changes_required, quota_change_request) + def _system_state_change(module, project_quota_output): """ Determine if changes are required to the current project quota. @@ -386,6 +392,7 @@ def _system_state_change(module, project_quota_output): else: return False + def main(): argument_spec = openstack_full_argument_spec( @@ -427,8 +434,8 @@ def main(): ) module = AnsibleModule(argument_spec, - supports_check_mode=True - ) + supports_check_mode=True + ) if not HAS_SHADE: module.fail_json(msg='shade is required for this module') @@ -437,7 +444,7 @@ def main(): cloud_params = dict(module.params) cloud = shade.operator_cloud(**cloud_params) - #In order to handle the different volume types we update module params after. + # In order to handle the different volume types we update module params after. dynamic_types = [ 'gigabytes_types', 'snapshots_types', @@ -448,22 +455,22 @@ def main(): for k, v in module.params[dynamic_type].items(): module.params[k] = int(v) - #Get current quota values + # Get current quota values project_quota_output = _get_quotas(module, cloud, cloud_params['name']) changes_required = False if module.params['state'] == "absent": - #If a quota state is set to absent we should assume there will be changes. - #The default quota values are not accessible so we can not determine if - #no changes will occur or not. + # If a quota state is set to absent we should assume there will be changes. + # The default quota values are not accessible so we can not determine if + # no changes will occur or not. if module.check_mode: module.exit_json(changed=True) - #Calling delete_network_quotas when a quota has not been set results - #in an error, according to the shade docs it should return the - #current quota. - #The following error string is returned: - #network client call failed: Quota for tenant 69dd91d217e949f1a0b35a4b901741dc could not be found. + # Calling delete_network_quotas when a quota has not been set results + # in an error, according to the shade docs it should return the + # current quota. + # The following error string is returned: + # network client call failed: Quota for tenant 69dd91d217e949f1a0b35a4b901741dc could not be found. neutron_msg1 = "network client call failed: Quota for tenant" neutron_msg2 = "could not be found" @@ -495,7 +502,7 @@ def main(): quota_call = getattr(cloud, 'set_%s_quotas' % (quota_type)) quota_call(cloud_params['name'], **quota_change_request[quota_type]) - #Get quota state post changes for validation + # Get quota state post changes for validation project_quota_update = _get_quotas(module, cloud, cloud_params['name']) if project_quota_output == project_quota_update: @@ -504,8 +511,8 @@ def main(): project_quota_output = project_quota_update module.exit_json(changed=changes_required, - openstack_quotas=project_quota_output - ) + openstack_quotas=project_quota_output + ) except shade.OpenStackCloudException as e: module.fail_json(msg=str(e), extra_data=e.extra_data) diff --git a/lib/ansible/modules/cloud/openstack/os_security_group_rule.py b/lib/ansible/modules/cloud/openstack/os_security_group_rule.py index 7a9d771cb8..e6c6fac150 100644 --- a/lib/ansible/modules/cloud/openstack/os_security_group_rule.py +++ b/lib/ansible/modules/cloud/openstack/os_security_group_rule.py @@ -212,8 +212,8 @@ def _ports_match(protocol, module_min, module_max, rule_min, rule_max): if ((module_min is None and module_max is None) and (rule_min and int(rule_min) == 1 and rule_max and int(rule_max) == 65535)): - # (None, None) == (1, 65535) - return True + # (None, None) == (1, 65535) + return True # Sanity check to make sure we don't have type comparison issues. if module_min: @@ -270,21 +270,21 @@ def _system_state_change(module, secgroup, remotegroup): def main(): argument_spec = openstack_full_argument_spec( - security_group = dict(required=True), + security_group=dict(required=True), # NOTE(Shrews): None is an acceptable protocol value for # Neutron, but Nova will balk at this. - protocol = dict(default=None, - choices=[None, 'tcp', 'udp', 'icmp', '112']), - port_range_min = dict(required=False, type='int'), - port_range_max = dict(required=False, type='int'), - remote_ip_prefix = dict(required=False, default=None), - remote_group = dict(required=False, default=None), - ethertype = dict(default='IPv4', - choices=['IPv4', 'IPv6']), - direction = dict(default='ingress', - choices=['egress', 'ingress']), - state = dict(default='present', - choices=['absent', 'present']), + protocol=dict(default=None, + choices=[None, 'tcp', 'udp', 'icmp', '112']), + port_range_min=dict(required=False, type='int'), + port_range_max=dict(required=False, type='int'), + remote_ip_prefix=dict(required=False, default=None), + remote_group=dict(required=False, default=None), + ethertype=dict(default='IPv4', + choices=['IPv4', 'IPv6']), + direction=dict(default='ingress', + choices=['egress', 'ingress']), + state=dict(default='present', + choices=['absent', 'present']), ) module_kwargs = openstack_module_kwargs( @@ -312,7 +312,7 @@ def main(): if remote_group: remotegroup = cloud.get_security_group(remote_group) else: - remotegroup = { 'id' : None } + remotegroup = {'id': None} if module.check_mode: module.exit_json(changed=_system_state_change(module, secgroup, remotegroup)) diff --git a/lib/ansible/modules/cloud/openstack/os_server.py b/lib/ansible/modules/cloud/openstack/os_server.py index 098e17959a..751546b745 100644 --- a/lib/ansible/modules/cloud/openstack/os_server.py +++ b/lib/ansible/modules/cloud/openstack/os_server.py @@ -433,6 +433,7 @@ def _parse_nics(nics): else: yield net + def _network_args(module, cloud): args = [] nics = module.params['nics'] @@ -685,31 +686,31 @@ def _get_server_state(module, cloud): def main(): argument_spec = openstack_full_argument_spec( - name = dict(required=True), - image = dict(default=None), - image_exclude = dict(default='(deprecated)'), - flavor = dict(default=None), - flavor_ram = dict(default=None, type='int'), - flavor_include = dict(default=None), - key_name = dict(default=None), - security_groups = dict(default=['default'], type='list'), - network = dict(default=None), - nics = dict(default=[], type='list'), - meta = dict(default=None, type='raw'), - userdata = dict(default=None, aliases=['user_data']), - config_drive = dict(default=False, type='bool'), - auto_ip = dict(default=True, type='bool', aliases=['auto_floating_ip', 'public_ip']), - floating_ips = dict(default=None, type='list'), - floating_ip_pools = dict(default=None, type='list'), - volume_size = dict(default=False, type='int'), - boot_from_volume = dict(default=False, type='bool'), - boot_volume = dict(default=None, aliases=['root_volume']), - terminate_volume = dict(default=False, type='bool'), - volumes = dict(default=[], type='list'), - scheduler_hints = dict(default=None, type='dict'), - state = dict(default='present', choices=['absent', 'present']), - delete_fip = dict(default=False, type='bool'), - reuse_ips = dict(default=True, type='bool'), + name=dict(required=True), + image=dict(default=None), + image_exclude=dict(default='(deprecated)'), + flavor=dict(default=None), + flavor_ram=dict(default=None, type='int'), + flavor_include=dict(default=None), + key_name=dict(default=None), + security_groups=dict(default=['default'], type='list'), + network=dict(default=None), + nics=dict(default=[], type='list'), + meta=dict(default=None, type='raw'), + userdata=dict(default=None, aliases=['user_data']), + config_drive=dict(default=False, type='bool'), + auto_ip=dict(default=True, type='bool', aliases=['auto_floating_ip', 'public_ip']), + floating_ips=dict(default=None, type='list'), + floating_ip_pools=dict(default=None, type='list'), + volume_size=dict(default=False, type='int'), + boot_from_volume=dict(default=False, type='bool'), + boot_volume=dict(default=None, aliases=['root_volume']), + terminate_volume=dict(default=False, type='bool'), + volumes=dict(default=[], type='list'), + scheduler_hints=dict(default=None, type='dict'), + state=dict(default='present', choices=['absent', 'present']), + delete_fip=dict(default=False, type='bool'), + reuse_ips=dict(default=True, type='bool'), ) module_kwargs = openstack_module_kwargs( mutually_exclusive=[ diff --git a/lib/ansible/modules/cloud/openstack/os_stack.py b/lib/ansible/modules/cloud/openstack/os_stack.py index 5f2f9fff5b..e250b5bde8 100644 --- a/lib/ansible/modules/cloud/openstack/os_stack.py +++ b/lib/ansible/modules/cloud/openstack/os_stack.py @@ -1,5 +1,5 @@ #!/usr/bin/python -#coding: utf-8 -*- +# coding: utf-8 -*- # (c) 2016, Mathieu Bultel # (c) 2016, Steve Baker @@ -166,12 +166,12 @@ from ansible.module_utils.openstack import openstack_full_argument_spec, opensta def _create_stack(module, stack, cloud): try: stack = cloud.create_stack(module.params['name'], - template_file=module.params['template'], - environment_files=module.params['environment'], - timeout=module.params['timeout'], - wait=True, - rollback=module.params['rollback'], - **module.params['parameters']) + template_file=module.params['template'], + environment_files=module.params['environment'], + timeout=module.params['timeout'], + wait=True, + rollback=module.params['rollback'], + **module.params['parameters']) stack = cloud.get_stack(stack.id, None) if stack.stack_status == 'CREATE_COMPLETE': @@ -181,6 +181,7 @@ def _create_stack(module, stack, cloud): except shade.OpenStackCloudException as e: module.fail_json(msg=str(e)) + def _update_stack(module, stack, cloud): try: stack = cloud.update_stack( @@ -195,11 +196,12 @@ def _update_stack(module, stack, cloud): if stack['stack_status'] == 'UPDATE_COMPLETE': return stack else: - module.fail_json(msg = "Failure in updating stack: %s" % + module.fail_json(msg="Failure in updating stack: %s" % stack['stack_status_reason']) except shade.OpenStackCloudException as e: module.fail_json(msg=str(e)) + def _system_state_change(module, stack, cloud): state = module.params['state'] if state == 'present': @@ -209,6 +211,7 @@ def _system_state_change(module, stack, cloud): return True return False + def main(): argument_spec = openstack_full_argument_spec( diff --git a/lib/ansible/modules/cloud/ovh/ovh_ip_loadbalancing_backend.py b/lib/ansible/modules/cloud/ovh/ovh_ip_loadbalancing_backend.py index 7774549b27..7353e555c8 100644 --- a/lib/ansible/modules/cloud/ovh/ovh_ip_loadbalancing_backend.py +++ b/lib/ansible/modules/cloud/ovh/ovh_ip_loadbalancing_backend.py @@ -162,7 +162,7 @@ def main(): if not HAS_OVH: module.fail_json(msg='ovh-api python module' - 'is required to run this module ') + 'is required to run this module ') # Get parameters name = module.params.get('name') diff --git a/lib/ansible/modules/cloud/packet/packet_sshkey.py b/lib/ansible/modules/cloud/packet/packet_sshkey.py index 8b1f076e9c..f649e62fee 100644 --- a/lib/ansible/modules/cloud/packet/packet_sshkey.py +++ b/lib/ansible/modules/cloud/packet/packet_sshkey.py @@ -113,7 +113,7 @@ PACKET_API_TOKEN_ENV_VAR = "PACKET_API_TOKEN" def serialize_sshkey(sshkey): sshkey_data = {} - copy_keys = ['id', 'key', 'label','fingerprint'] + copy_keys = ['id', 'key', 'label', 'fingerprint'] for name in copy_keys: sshkey_data[name] = getattr(sshkey, name) return sshkey_data @@ -132,7 +132,7 @@ def load_key_string(key_str): key_str = key_str.strip() ret_dict['key'] = key_str cut_key = key_str.split() - if len(cut_key) in [2,3]: + if len(cut_key) in [2, 3]: if len(cut_key) == 3: ret_dict['label'] = cut_key[2] else: @@ -165,7 +165,7 @@ def get_sshkey_selector(module): return k.key == select_dict['key'] else: # if key string not specified, all the fields must match - return all([select_dict[f] == getattr(k,f) for f in select_dict]) + return all([select_dict[f] == getattr(k, f) for f in select_dict]) return selector @@ -188,10 +188,10 @@ def act_on_sshkeys(target_state, module, packet_conn): newkey['label'] = module.params.get('label') for param in ('label', 'key'): if param not in newkey: - _msg=("If you want to ensure a key is present, you must " - "supply both a label and a key string, either in " - "module params, or in a key file. %s is missing" - % param) + _msg = ("If you want to ensure a key is present, you must " + "supply both a label and a key string, either in " + "module params, or in a key file. %s is missing" + % param) raise Exception(_msg) matching_sshkeys = [] new_key_response = packet_conn.create_ssh_key( @@ -208,7 +208,7 @@ def act_on_sshkeys(target_state, module, packet_conn): except Exception as e: _msg = ("while trying to remove sshkey %s, id %s %s, " "got error: %s" % - (k.label, k.id, target_state, e)) + (k.label, k.id, target_state, e)) raise Exception(_msg) return { @@ -220,7 +220,7 @@ def act_on_sshkeys(target_state, module, packet_conn): def main(): module = AnsibleModule( argument_spec=dict( - state = dict(choices=['present', 'absent'], default='present'), + state=dict(choices=['present', 'absent'], default='present'), auth_token=dict(default=os.environ.get(PACKET_API_TOKEN_ENV_VAR), no_log=True), label=dict(type='str', aliases=['name'], default=None), @@ -236,16 +236,16 @@ def main(): ('key', 'fingerprint'), ('key', 'id'), ('key_file', 'key'), - ] + ] ) if not HAS_PACKET_SDK: module.fail_json(msg='packet required for this module') if not module.params.get('auth_token'): - _fail_msg = ( "if Packet API token is not in environment variable %s, " - "the auth_token parameter is required" % - PACKET_API_TOKEN_ENV_VAR) + _fail_msg = ("if Packet API token is not in environment variable %s, " + "the auth_token parameter is required" % + PACKET_API_TOKEN_ENV_VAR) module.fail_json(msg=_fail_msg) auth_token = module.params.get('auth_token') @@ -254,7 +254,7 @@ def main(): state = module.params.get('state') - if state in ['present','absent']: + if state in ['present', 'absent']: try: module.exit_json(**act_on_sshkeys(state, module, packet_conn)) except Exception as e: diff --git a/lib/ansible/modules/cloud/profitbricks/profitbricks.py b/lib/ansible/modules/cloud/profitbricks/profitbricks.py index d965167d4b..cd11f57cb1 100644 --- a/lib/ansible/modules/cloud/profitbricks/profitbricks.py +++ b/lib/ansible/modules/cloud/profitbricks/profitbricks.py @@ -238,7 +238,7 @@ def _wait_for_completion(profitbricks, promise, wait_timeout, msg): raise Exception( 'Timed out waiting for async operation ' + msg + ' "' + str( promise['requestId'] - ) + '" to complete.') + ) + '" to complete.') def _create_machine(module, profitbricks, datacenter, name): @@ -290,7 +290,7 @@ def _create_machine(module, profitbricks, datacenter, name): n = NIC( lan=int(lan) - ) + ) s = Server( name=name, @@ -299,7 +299,7 @@ def _create_machine(module, profitbricks, datacenter, name): cpu_family=cpu_family, create_volumes=[v], nics=[n], - ) + ) try: create_server_response = profitbricks.create_server( @@ -341,7 +341,7 @@ def _create_datacenter(module, profitbricks): i = Datacenter( name=datacenter, location=location - ) + ) try: datacenter_response = profitbricks.create_datacenter(datacenter=i) @@ -624,7 +624,7 @@ def main(): if state == 'absent': if not module.params.get('datacenter'): module.fail_json(msg='datacenter parameter is required ' + - 'for running or stopping machines.') + 'for running or stopping machines.') try: (changed) = remove_virtual_machine(module, profitbricks) @@ -635,7 +635,7 @@ def main(): elif state in ('running', 'stopped'): if not module.params.get('datacenter'): module.fail_json(msg='datacenter parameter is required for ' + - 'running or stopping machines.') + 'running or stopping machines.') try: (changed) = startstop_machine(module, profitbricks, state) module.exit_json(changed=changed) @@ -649,10 +649,10 @@ def main(): module.fail_json(msg='image parameter is required for new instance') if not module.params.get('subscription_user'): module.fail_json(msg='subscription_user parameter is ' + - 'required for new instance') + 'required for new instance') if not module.params.get('subscription_password'): module.fail_json(msg='subscription_password parameter is ' + - 'required for new instance') + 'required for new instance') try: (machine_dict_array) = create_virtual_machine(module, profitbricks) diff --git a/lib/ansible/modules/cloud/profitbricks/profitbricks_datacenter.py b/lib/ansible/modules/cloud/profitbricks/profitbricks_datacenter.py index bebff23c8c..e320544379 100644 --- a/lib/ansible/modules/cloud/profitbricks/profitbricks_datacenter.py +++ b/lib/ansible/modules/cloud/profitbricks/profitbricks_datacenter.py @@ -118,7 +118,8 @@ def _wait_for_completion(profitbricks, promise, wait_timeout, msg): raise Exception( 'Timed out waiting for async operation ' + msg + ' "' + str( promise['requestId'] - ) + '" to complete.') + ) + '" to complete.') + def _remove_datacenter(module, profitbricks, datacenter): try: @@ -126,6 +127,7 @@ def _remove_datacenter(module, profitbricks, datacenter): except Exception as e: module.fail_json(msg="failed to remove the datacenter: %s" % str(e)) + def create_datacenter(module, profitbricks): """ Creates a Datacenter @@ -148,7 +150,7 @@ def create_datacenter(module, profitbricks): name=name, location=location, description=description - ) + ) try: datacenter_response = profitbricks.create_datacenter(datacenter=i) @@ -166,6 +168,7 @@ def create_datacenter(module, profitbricks): except Exception as e: module.fail_json(msg="failed to create the new datacenter: %s" % str(e)) + def remove_datacenter(module, profitbricks): """ Removes a Datacenter. @@ -197,6 +200,7 @@ def remove_datacenter(module, profitbricks): return changed + def main(): module = AnsibleModule( argument_spec=dict( diff --git a/lib/ansible/modules/cloud/profitbricks/profitbricks_nic.py b/lib/ansible/modules/cloud/profitbricks/profitbricks_nic.py index ccf863a659..5d01618b3e 100644 --- a/lib/ansible/modules/cloud/profitbricks/profitbricks_nic.py +++ b/lib/ansible/modules/cloud/profitbricks/profitbricks_nic.py @@ -121,7 +121,8 @@ def _wait_for_completion(profitbricks, promise, wait_timeout, msg): raise Exception( 'Timed out waiting for async operation ' + msg + ' "' + str( promise['requestId'] - ) + '" to complete.') + ) + '" to complete.') + def create_nic(module, profitbricks): """ @@ -173,6 +174,7 @@ def create_nic(module, profitbricks): except Exception as e: module.fail_json(msg="failed to create the NIC: %s" % str(e)) + def delete_nic(module, profitbricks): """ Removes a NIC @@ -228,12 +230,13 @@ def delete_nic(module, profitbricks): except Exception as e: module.fail_json(msg="failed to remove the NIC: %s" % str(e)) + def main(): module = AnsibleModule( argument_spec=dict( datacenter=dict(), server=dict(), - name=dict(default=str(uuid.uuid4()).replace('-','')[:10]), + name=dict(default=str(uuid.uuid4()).replace('-', '')[:10]), lan=dict(), subscription_user=dict(), subscription_password=dict(no_log=True), @@ -255,7 +258,6 @@ def main(): if not module.params.get('server'): module.fail_json(msg='server parameter is required') - subscription_user = module.params.get('subscription_user') subscription_password = module.params.get('subscription_password') diff --git a/lib/ansible/modules/cloud/profitbricks/profitbricks_volume.py b/lib/ansible/modules/cloud/profitbricks/profitbricks_volume.py index 1d1980e078..c65939e7fa 100644 --- a/lib/ansible/modules/cloud/profitbricks/profitbricks_volume.py +++ b/lib/ansible/modules/cloud/profitbricks/profitbricks_volume.py @@ -170,7 +170,7 @@ def _wait_for_completion(profitbricks, promise, wait_timeout, msg): raise Exception( 'Timed out waiting for async operation ' + msg + ' "' + str( promise['requestId'] - ) + '" to complete.') + ) + '" to complete.') def _create_volume(module, profitbricks, datacenter, name): @@ -194,7 +194,7 @@ def _create_volume(module, profitbricks, datacenter, name): ssh_keys=ssh_keys, disk_type=disk_type, licence_type=licence_type - ) + ) volume_response = profitbricks.create_volume(datacenter, v) diff --git a/lib/ansible/modules/cloud/profitbricks/profitbricks_volume_attachments.py b/lib/ansible/modules/cloud/profitbricks/profitbricks_volume_attachments.py index 9a7bfec3bf..581e76a1ff 100644 --- a/lib/ansible/modules/cloud/profitbricks/profitbricks_volume_attachments.py +++ b/lib/ansible/modules/cloud/profitbricks/profitbricks_volume_attachments.py @@ -118,7 +118,8 @@ def _wait_for_completion(profitbricks, promise, wait_timeout, msg): raise Exception( 'Timed out waiting for async operation ' + msg + ' "' + str( promise['requestId'] - ) + '" to complete.') + ) + '" to complete.') + def attach_volume(module, profitbricks): """ @@ -150,7 +151,7 @@ def attach_volume(module, profitbricks): server_list = profitbricks.list_servers(datacenter) for s in server_list['items']: if server == s['properties']['name']: - server= s['id'] + server = s['id'] break # Locate UUID for Volume @@ -163,6 +164,7 @@ def attach_volume(module, profitbricks): return profitbricks.attach_volume(datacenter, server, volume) + def detach_volume(module, profitbricks): """ Detaches a volume. @@ -193,7 +195,7 @@ def detach_volume(module, profitbricks): server_list = profitbricks.list_servers(datacenter) for s in server_list['items']: if server == s['properties']['name']: - server= s['id'] + server = s['id'] break # Locate UUID for Volume @@ -206,6 +208,7 @@ def detach_volume(module, profitbricks): return profitbricks.detach_volume(datacenter, server, volume) + def main(): module = AnsibleModule( argument_spec=dict( diff --git a/lib/ansible/modules/cloud/rackspace/rax_cbs_attachments.py b/lib/ansible/modules/cloud/rackspace/rax_cbs_attachments.py index 9bafd5e591..7cb1abe398 100644 --- a/lib/ansible/modules/cloud/rackspace/rax_cbs_attachments.py +++ b/lib/ansible/modules/cloud/rackspace/rax_cbs_attachments.py @@ -93,7 +93,7 @@ from ansible.module_utils.rax import (NON_CALLABLES, rax_required_together, rax_to_dict, setup_rax_module, - ) + ) def cloud_block_storage_attachments(module, state, volume, server, device, diff --git a/lib/ansible/modules/cloud/rackspace/rax_cdb.py b/lib/ansible/modules/cloud/rackspace/rax_cdb.py index eb03328ffb..ced4369676 100644 --- a/lib/ansible/modules/cloud/rackspace/rax_cdb.py +++ b/lib/ansible/modules/cloud/rackspace/rax_cdb.py @@ -148,7 +148,7 @@ def save_instance(module, name, flavor, volume, cdb_type, cdb_version, wait, module.fail_json(changed=False, action=action, msg='The new volume size must be larger than ' 'the current volume size', - cdb=rax_to_dict(instance)) + cdb=rax_to_dict(instance)) instance.resize_volume(volume) changed = True diff --git a/lib/ansible/modules/cloud/rackspace/rax_clb.py b/lib/ansible/modules/cloud/rackspace/rax_clb.py index 86b750edf7..f00bd5e63e 100644 --- a/lib/ansible/modules/cloud/rackspace/rax_clb.py +++ b/lib/ansible/modules/cloud/rackspace/rax_clb.py @@ -141,7 +141,7 @@ from ansible.module_utils.rax import (CLB_ALGORITHMS, rax_required_together, rax_to_dict, setup_rax_module, - ) + ) def cloud_load_balancer(module, state, name, meta, algorithm, port, protocol, diff --git a/lib/ansible/modules/cloud/rackspace/rax_clb_ssl.py b/lib/ansible/modules/cloud/rackspace/rax_clb_ssl.py index d6f89417c2..6dfee78817 100644 --- a/lib/ansible/modules/cloud/rackspace/rax_clb_ssl.py +++ b/lib/ansible/modules/cloud/rackspace/rax_clb_ssl.py @@ -11,7 +11,7 @@ ANSIBLE_METADATA = {'metadata_version': '1.1', 'supported_by': 'community'} -DOCUMENTATION=''' +DOCUMENTATION = ''' module: rax_clb_ssl short_description: Manage SSL termination for a Rackspace Cloud Load Balancer. description: @@ -100,7 +100,8 @@ from ansible.module_utils.rax import (rax_argument_spec, rax_required_together, rax_to_dict, setup_rax_module, - ) + ) + def cloud_load_balancer_ssl(module, loadbalancer, state, enabled, private_key, certificate, intermediate_certificate, secure_port, @@ -222,6 +223,7 @@ def cloud_load_balancer_ssl(module, loadbalancer, state, enabled, private_key, else: module.fail_json(**result) + def main(): argument_spec = rax_argument_spec() argument_spec.update(dict( diff --git a/lib/ansible/modules/cloud/rackspace/rax_dns.py b/lib/ansible/modules/cloud/rackspace/rax_dns.py index 76ce9ec625..1a0aaba6f6 100644 --- a/lib/ansible/modules/cloud/rackspace/rax_dns.py +++ b/lib/ansible/modules/cloud/rackspace/rax_dns.py @@ -72,7 +72,7 @@ from ansible.module_utils.rax import (rax_argument_spec, rax_required_together, rax_to_dict, setup_rax_module, - ) + ) def rax_dns(module, comment, email, name, state, ttl): diff --git a/lib/ansible/modules/cloud/rackspace/rax_dns_record.py b/lib/ansible/modules/cloud/rackspace/rax_dns_record.py index 6c04d68c9d..6ed9b4fdc8 100644 --- a/lib/ansible/modules/cloud/rackspace/rax_dns_record.py +++ b/lib/ansible/modules/cloud/rackspace/rax_dns_record.py @@ -128,7 +128,7 @@ from ansible.module_utils.rax import (rax_argument_spec, rax_required_together, rax_to_dict, setup_rax_module, - ) + ) def rax_dns_record_ptr(module, data=None, comment=None, loadbalancer=None, diff --git a/lib/ansible/modules/cloud/rackspace/rax_facts.py b/lib/ansible/modules/cloud/rackspace/rax_facts.py index 42d8201691..4c40a550a0 100644 --- a/lib/ansible/modules/cloud/rackspace/rax_facts.py +++ b/lib/ansible/modules/cloud/rackspace/rax_facts.py @@ -61,7 +61,7 @@ from ansible.module_utils.rax import (rax_argument_spec, rax_required_together, rax_to_dict, setup_rax_module, - ) + ) def rax_facts(module, address, name, server_id): diff --git a/lib/ansible/modules/cloud/rackspace/rax_keypair.py b/lib/ansible/modules/cloud/rackspace/rax_keypair.py index 57988a75e5..0a6010662c 100644 --- a/lib/ansible/modules/cloud/rackspace/rax_keypair.py +++ b/lib/ansible/modules/cloud/rackspace/rax_keypair.py @@ -91,7 +91,7 @@ from ansible.module_utils.rax import (rax_argument_spec, rax_required_together, rax_to_dict, setup_rax_module, - ) + ) def rax_keypair(module, name, public_key, state): diff --git a/lib/ansible/modules/cloud/rackspace/rax_mon_alarm.py b/lib/ansible/modules/cloud/rackspace/rax_mon_alarm.py index 996f5f849d..04e5db258f 100644 --- a/lib/ansible/modules/cloud/rackspace/rax_mon_alarm.py +++ b/lib/ansible/modules/cloud/rackspace/rax_mon_alarm.py @@ -180,6 +180,7 @@ def alarm(module, state, label, entity_id, check_id, notification_plan_id, crite else: module.exit_json(changed=changed) + def main(): argument_spec = rax_argument_spec() argument_spec.update( diff --git a/lib/ansible/modules/cloud/rackspace/rax_mon_check.py b/lib/ansible/modules/cloud/rackspace/rax_mon_check.py index 48a5156b99..a6e791db96 100644 --- a/lib/ansible/modules/cloud/rackspace/rax_mon_check.py +++ b/lib/ansible/modules/cloud/rackspace/rax_mon_check.py @@ -256,6 +256,7 @@ def cloud_check(module, state, entity_id, label, check_type, else: module.exit_json(changed=changed) + def main(): argument_spec = rax_argument_spec() argument_spec.update( diff --git a/lib/ansible/modules/cloud/rackspace/rax_mon_entity.py b/lib/ansible/modules/cloud/rackspace/rax_mon_entity.py index 0ed3aac18f..21d917c301 100644 --- a/lib/ansible/modules/cloud/rackspace/rax_mon_entity.py +++ b/lib/ansible/modules/cloud/rackspace/rax_mon_entity.py @@ -152,6 +152,7 @@ def cloud_monitoring(module, state, label, agent_id, named_ip_addresses, else: module.exit_json(changed=changed) + def main(): argument_spec = rax_argument_spec() argument_spec.update( diff --git a/lib/ansible/modules/cloud/rackspace/rax_mon_notification.py b/lib/ansible/modules/cloud/rackspace/rax_mon_notification.py index d4da95b756..6aee351b96 100644 --- a/lib/ansible/modules/cloud/rackspace/rax_mon_notification.py +++ b/lib/ansible/modules/cloud/rackspace/rax_mon_notification.py @@ -138,6 +138,7 @@ def notification(module, state, label, notification_type, details): else: module.exit_json(changed=changed) + def main(): argument_spec = rax_argument_spec() argument_spec.update( diff --git a/lib/ansible/modules/cloud/rackspace/rax_mon_notification_plan.py b/lib/ansible/modules/cloud/rackspace/rax_mon_notification_plan.py index fb5822aa6e..486599bcbd 100644 --- a/lib/ansible/modules/cloud/rackspace/rax_mon_notification_plan.py +++ b/lib/ansible/modules/cloud/rackspace/rax_mon_notification_plan.py @@ -141,6 +141,7 @@ def notification_plan(module, state, label, critical_state, warning_state, ok_st else: module.exit_json(changed=changed) + def main(): argument_spec = rax_argument_spec() argument_spec.update( diff --git a/lib/ansible/modules/cloud/univention/udm_dns_record.py b/lib/ansible/modules/cloud/univention/udm_dns_record.py index b7916bb7fb..5ad7b90b44 100644 --- a/lib/ansible/modules/cloud/univention/udm_dns_record.py +++ b/lib/ansible/modules/cloud/univention/udm_dns_record.py @@ -94,21 +94,21 @@ from ansible.module_utils.univention_umc import ( def main(): module = AnsibleModule( - argument_spec = dict( - type = dict(required=True, - type='str'), - zone = dict(required=True, - type='str'), - name = dict(required=True, - type='str'), - data = dict(default=[], - type='dict'), - state = dict(default='present', - choices=['present', 'absent'], - type='str') + argument_spec=dict( + type=dict(required=True, + type='str'), + zone=dict(required=True, + type='str'), + name=dict(required=True, + type='str'), + data=dict(default=[], + type='dict'), + state=dict(default='present', + choices=['present', 'absent'], + type='str') ), supports_check_mode=True, - required_if = ([ + required_if=([ ('state', 'present', ['data']) ]) ) @@ -116,12 +116,12 @@ def main(): if not HAVE_UNIVENTION: module.fail_json(msg="This module requires univention python bindings") - type = module.params['type'] - zone = module.params['zone'] - name = module.params['name'] - data = module.params['data'] - state = module.params['state'] - changed = False + type = module.params['type'] + zone = module.params['zone'] + name = module.params['name'] + data = module.params['data'] + state = module.params['state'] + changed = False obj = list(ldap_search( '(&(objectClass=dNSZone)(zoneName={})(relativeDomainName={}))'.format(zone, name), diff --git a/lib/ansible/modules/cloud/univention/udm_dns_zone.py b/lib/ansible/modules/cloud/univention/udm_dns_zone.py index 8d167ff8aa..5b3069f4bd 100644 --- a/lib/ansible/modules/cloud/univention/udm_dns_zone.py +++ b/lib/ansible/modules/cloud/univention/udm_dns_zone.py @@ -111,10 +111,10 @@ from ansible.module_utils.univention_umc import ( def convert_time(time): """Convert a time in seconds into the biggest unit""" units = [ - (24 * 60 * 60 , 'days'), - (60 * 60 , 'hours'), - (60 , 'minutes'), - (1 , 'seconds'), + (24 * 60 * 60, 'days'), + (60 * 60, 'hours'), + (60, 'minutes'), + (1, 'seconds'), ] if time == 0: @@ -126,49 +126,49 @@ def convert_time(time): def main(): module = AnsibleModule( - argument_spec = dict( - type = dict(required=True, - type='str'), - zone = dict(required=True, - aliases=['name'], - type='str'), - nameserver = dict(default=[], - type='list'), - interfaces = dict(default=[], - type='list'), - refresh = dict(default=3600, - type='int'), - retry = dict(default=1800, - type='int'), - expire = dict(default=604800, - type='int'), - ttl = dict(default=600, - type='int'), - contact = dict(default='', - type='str'), - mx = dict(default=[], - type='list'), - state = dict(default='present', - choices=['present', 'absent'], - type='str') + argument_spec=dict( + type=dict(required=True, + type='str'), + zone=dict(required=True, + aliases=['name'], + type='str'), + nameserver=dict(default=[], + type='list'), + interfaces=dict(default=[], + type='list'), + refresh=dict(default=3600, + type='int'), + retry=dict(default=1800, + type='int'), + expire=dict(default=604800, + type='int'), + ttl=dict(default=600, + type='int'), + contact=dict(default='', + type='str'), + mx=dict(default=[], + type='list'), + state=dict(default='present', + choices=['present', 'absent'], + type='str') ), supports_check_mode=True, - required_if = ([ + required_if=([ ('state', 'present', ['nameserver', 'interfaces']) ]) ) - type = module.params['type'] - zone = module.params['zone'] - nameserver = module.params['nameserver'] - interfaces = module.params['interfaces'] - refresh = module.params['refresh'] - retry = module.params['retry'] - expire = module.params['expire'] - ttl = module.params['ttl'] - contact = module.params['contact'] - mx = module.params['mx'] - state = module.params['state'] - changed = False + type = module.params['type'] + zone = module.params['zone'] + nameserver = module.params['nameserver'] + interfaces = module.params['interfaces'] + refresh = module.params['refresh'] + retry = module.params['retry'] + expire = module.params['expire'] + ttl = module.params['ttl'] + contact = module.params['contact'] + mx = module.params['mx'] + state = module.params['state'] + changed = False obj = list(ldap_search( '(&(objectClass=dNSZone)(zoneName={}))'.format(zone), @@ -187,15 +187,15 @@ def main(): obj = umc_module_for_add('dns/{}'.format(type), container) else: obj = umc_module_for_edit('dns/{}'.format(type), dn) - obj['zone'] = zone - obj['nameserver'] = nameserver - obj['a'] = interfaces - obj['refresh'] = convert_time(refresh) - obj['retry'] = convert_time(retry) - obj['expire'] = convert_time(expire) - obj['ttl'] = convert_time(ttl) - obj['contact'] = contact - obj['mx'] = mx + obj['zone'] = zone + obj['nameserver'] = nameserver + obj['a'] = interfaces + obj['refresh'] = convert_time(refresh) + obj['retry'] = convert_time(retry) + obj['expire'] = convert_time(expire) + obj['ttl'] = convert_time(ttl) + obj['contact'] = contact + obj['mx'] = mx diff = obj.diff() if exists: for k in obj.keys(): diff --git a/lib/ansible/modules/cloud/univention/udm_group.py b/lib/ansible/modules/cloud/univention/udm_group.py index be3bb21e59..7e6564591e 100644 --- a/lib/ansible/modules/cloud/univention/udm_group.py +++ b/lib/ansible/modules/cloud/univention/udm_group.py @@ -87,30 +87,30 @@ from ansible.module_utils.univention_umc import ( def main(): module = AnsibleModule( - argument_spec = dict( - name = dict(required=True, - type='str'), - description = dict(default=None, - type='str'), - position = dict(default='', - type='str'), - ou = dict(default='', - type='str'), - subpath = dict(default='cn=groups', - type='str'), - state = dict(default='present', - choices=['present', 'absent'], - type='str') + argument_spec=dict( + name=dict(required=True, + type='str'), + description=dict(default=None, + type='str'), + position=dict(default='', + type='str'), + ou=dict(default='', + type='str'), + subpath=dict(default='cn=groups', + type='str'), + state=dict(default='present', + choices=['present', 'absent'], + type='str') ), supports_check_mode=True ) - name = module.params['name'] + name = module.params['name'] description = module.params['description'] - position = module.params['position'] - ou = module.params['ou'] - subpath = module.params['subpath'] - state = module.params['state'] - changed = False + position = module.params['position'] + ou = module.params['ou'] + subpath = module.params['subpath'] + state = module.params['state'] + changed = False groups = list(ldap_search( '(&(objectClass=posixGroup)(cn={}))'.format(name), @@ -134,8 +134,8 @@ def main(): grp = umc_module_for_add('groups/group', container) else: grp = umc_module_for_edit('groups/group', group_dn) - grp['name'] = name - grp['description'] = description + grp['name'] = name + grp['description'] = description diff = grp.diff() changed = grp.diff() != [] if not module.check_mode: diff --git a/lib/ansible/modules/cloud/univention/udm_share.py b/lib/ansible/modules/cloud/univention/udm_share.py index 4c9abfb95c..68bcad5667 100644 --- a/lib/ansible/modules/cloud/univention/udm_share.py +++ b/lib/ansible/modules/cloud/univention/udm_share.py @@ -380,162 +380,162 @@ from ansible.module_utils.univention_umc import ( def main(): module = AnsibleModule( - argument_spec = dict( - name = dict(required=True, - type='str'), - ou = dict(required=True, - type='str'), - owner = dict(type='str', - default='0'), - group = dict(type='str', - default='0'), - path = dict(type='path', - default=None), - directorymode = dict(type='str', - default='00755'), - host = dict(type='str', - default=None), - root_squash = dict(type='bool', - default=True), - subtree_checking = dict(type='bool', - default=True), - sync = dict(type='str', - default='sync'), - writeable = dict(type='bool', - default=True), - sambaBlockSize = dict(type='str', - aliases=['samba_block_size'], - default=None), - sambaBlockingLocks = dict(type='bool', - aliases=['samba_blocking_locks'], - default=True), - sambaBrowseable = dict(type='bool', - aliases=['samba_browsable'], - default=True), - sambaCreateMode = dict(type='str', - aliases=['samba_create_mode'], - default='0744'), - sambaCscPolicy = dict(type='str', - aliases=['samba_csc_policy'], - default='manual'), - sambaCustomSettings = dict(type='list', - aliases=['samba_custom_settings'], - default=[]), - sambaDirectoryMode = dict(type='str', - aliases=['samba_directory_mode'], - default='0755'), - sambaDirectorySecurityMode = dict(type='str', - aliases=['samba_directory_security_mode'], - default='0777'), - sambaDosFilemode = dict(type='bool', - aliases=['samba_dos_filemode'], - default=False), - sambaFakeOplocks = dict(type='bool', - aliases=['samba_fake_oplocks'], - default=False), - sambaForceCreateMode = dict(type='bool', - aliases=['samba_force_create_mode'], - default=False), - sambaForceDirectoryMode = dict(type='bool', - aliases=['samba_force_directory_mode'], - default=False), - sambaForceDirectorySecurityMode = dict(type='bool', - aliases=['samba_force_directory_security_mode'], - default=False), - sambaForceGroup = dict(type='str', - aliases=['samba_force_group'], - default=None), - sambaForceSecurityMode = dict(type='bool', - aliases=['samba_force_security_mode'], - default=False), - sambaForceUser = dict(type='str', - aliases=['samba_force_user'], - default=None), - sambaHideFiles = dict(type='str', - aliases=['samba_hide_files'], - default=None), - sambaHideUnreadable = dict(type='bool', - aliases=['samba_hide_unreadable'], - default=False), - sambaHostsAllow = dict(type='list', - aliases=['samba_hosts_allow'], - default=[]), - sambaHostsDeny = dict(type='list', - aliases=['samba_hosts_deny'], - default=[]), - sambaInheritAcls = dict(type='bool', - aliases=['samba_inherit_acls'], - default=True), - sambaInheritOwner = dict(type='bool', - aliases=['samba_inherit_owner'], - default=False), - sambaInheritPermissions = dict(type='bool', - aliases=['samba_inherit_permissions'], - default=False), - sambaInvalidUsers = dict(type='str', - aliases=['samba_invalid_users'], - default=None), - sambaLevel2Oplocks = dict(type='bool', - aliases=['samba_level_2_oplocks'], - default=True), - sambaLocking = dict(type='bool', - aliases=['samba_locking'], - default=True), - sambaMSDFSRoot = dict(type='bool', - aliases=['samba_msdfs_root'], - default=False), - sambaName = dict(type='str', - aliases=['samba_name'], - default=None), - sambaNtAclSupport = dict(type='bool', - aliases=['samba_nt_acl_support'], - default=True), - sambaOplocks = dict(type='bool', - aliases=['samba_oplocks'], - default=True), - sambaPostexec = dict(type='str', - aliases=['samba_postexec'], - default=None), - sambaPreexec = dict(type='str', - aliases=['samba_preexec'], - default=None), - sambaPublic = dict(type='bool', - aliases=['samba_public'], - default=False), - sambaSecurityMode = dict(type='str', - aliases=['samba_security_mode'], - default='0777'), - sambaStrictLocking = dict(type='str', - aliases=['samba_strict_locking'], - default='Auto'), - sambaVFSObjects = dict(type='str', - aliases=['samba_vfs_objects'], - default=None), - sambaValidUsers = dict(type='str', - aliases=['samba_valid_users'], - default=None), - sambaWriteList = dict(type='str', - aliases=['samba_write_list'], - default=None), - sambaWriteable = dict(type='bool', - aliases=['samba_writeable'], - default=True), - nfs_hosts = dict(type='list', - default=[]), - nfsCustomSettings = dict(type='list', - aliases=['nfs_custom_settings'], - default=[]), - state = dict(default='present', - choices=['present', 'absent'], - type='str') + argument_spec=dict( + name=dict(required=True, + type='str'), + ou=dict(required=True, + type='str'), + owner=dict(type='str', + default='0'), + group=dict(type='str', + default='0'), + path=dict(type='path', + default=None), + directorymode=dict(type='str', + default='00755'), + host=dict(type='str', + default=None), + root_squash=dict(type='bool', + default=True), + subtree_checking=dict(type='bool', + default=True), + sync=dict(type='str', + default='sync'), + writeable=dict(type='bool', + default=True), + sambaBlockSize=dict(type='str', + aliases=['samba_block_size'], + default=None), + sambaBlockingLocks=dict(type='bool', + aliases=['samba_blocking_locks'], + default=True), + sambaBrowseable=dict(type='bool', + aliases=['samba_browsable'], + default=True), + sambaCreateMode=dict(type='str', + aliases=['samba_create_mode'], + default='0744'), + sambaCscPolicy=dict(type='str', + aliases=['samba_csc_policy'], + default='manual'), + sambaCustomSettings=dict(type='list', + aliases=['samba_custom_settings'], + default=[]), + sambaDirectoryMode=dict(type='str', + aliases=['samba_directory_mode'], + default='0755'), + sambaDirectorySecurityMode=dict(type='str', + aliases=['samba_directory_security_mode'], + default='0777'), + sambaDosFilemode=dict(type='bool', + aliases=['samba_dos_filemode'], + default=False), + sambaFakeOplocks=dict(type='bool', + aliases=['samba_fake_oplocks'], + default=False), + sambaForceCreateMode=dict(type='bool', + aliases=['samba_force_create_mode'], + default=False), + sambaForceDirectoryMode=dict(type='bool', + aliases=['samba_force_directory_mode'], + default=False), + sambaForceDirectorySecurityMode=dict(type='bool', + aliases=['samba_force_directory_security_mode'], + default=False), + sambaForceGroup=dict(type='str', + aliases=['samba_force_group'], + default=None), + sambaForceSecurityMode=dict(type='bool', + aliases=['samba_force_security_mode'], + default=False), + sambaForceUser=dict(type='str', + aliases=['samba_force_user'], + default=None), + sambaHideFiles=dict(type='str', + aliases=['samba_hide_files'], + default=None), + sambaHideUnreadable=dict(type='bool', + aliases=['samba_hide_unreadable'], + default=False), + sambaHostsAllow=dict(type='list', + aliases=['samba_hosts_allow'], + default=[]), + sambaHostsDeny=dict(type='list', + aliases=['samba_hosts_deny'], + default=[]), + sambaInheritAcls=dict(type='bool', + aliases=['samba_inherit_acls'], + default=True), + sambaInheritOwner=dict(type='bool', + aliases=['samba_inherit_owner'], + default=False), + sambaInheritPermissions=dict(type='bool', + aliases=['samba_inherit_permissions'], + default=False), + sambaInvalidUsers=dict(type='str', + aliases=['samba_invalid_users'], + default=None), + sambaLevel2Oplocks=dict(type='bool', + aliases=['samba_level_2_oplocks'], + default=True), + sambaLocking=dict(type='bool', + aliases=['samba_locking'], + default=True), + sambaMSDFSRoot=dict(type='bool', + aliases=['samba_msdfs_root'], + default=False), + sambaName=dict(type='str', + aliases=['samba_name'], + default=None), + sambaNtAclSupport=dict(type='bool', + aliases=['samba_nt_acl_support'], + default=True), + sambaOplocks=dict(type='bool', + aliases=['samba_oplocks'], + default=True), + sambaPostexec=dict(type='str', + aliases=['samba_postexec'], + default=None), + sambaPreexec=dict(type='str', + aliases=['samba_preexec'], + default=None), + sambaPublic=dict(type='bool', + aliases=['samba_public'], + default=False), + sambaSecurityMode=dict(type='str', + aliases=['samba_security_mode'], + default='0777'), + sambaStrictLocking=dict(type='str', + aliases=['samba_strict_locking'], + default='Auto'), + sambaVFSObjects=dict(type='str', + aliases=['samba_vfs_objects'], + default=None), + sambaValidUsers=dict(type='str', + aliases=['samba_valid_users'], + default=None), + sambaWriteList=dict(type='str', + aliases=['samba_write_list'], + default=None), + sambaWriteable=dict(type='bool', + aliases=['samba_writeable'], + default=True), + nfs_hosts=dict(type='list', + default=[]), + nfsCustomSettings=dict(type='list', + aliases=['nfs_custom_settings'], + default=[]), + state=dict(default='present', + choices=['present', 'absent'], + type='str') ), supports_check_mode=True, - required_if = ([ + required_if=([ ('state', 'present', ['path', 'host', 'sambaName']) ]) ) - name = module.params['name'] - state = module.params['state'] + name = module.params['name'] + state = module.params['state'] changed = False obj = list(ldap_search( diff --git a/lib/ansible/modules/cloud/univention/udm_user.py b/lib/ansible/modules/cloud/univention/udm_user.py index 1d021611bb..b0f6ee0a60 100644 --- a/lib/ansible/modules/cloud/univention/udm_user.py +++ b/lib/ansible/modules/cloud/univention/udm_user.py @@ -348,141 +348,141 @@ from ansible.module_utils.univention_umc import ( def main(): expiry = date.strftime(date.today() + timedelta(days=365), "%Y-%m-%d") module = AnsibleModule( - argument_spec = dict( - birthday = dict(default=None, - type='str'), - city = dict(default=None, - type='str'), - country = dict(default=None, - type='str'), - department_number = dict(default=None, - type='str', - aliases=['departmentNumber']), - description = dict(default=None, - type='str'), - display_name = dict(default=None, - type='str', - aliases=['displayName']), - email = dict(default=[''], - type='list'), - employee_number = dict(default=None, - type='str', - aliases=['employeeNumber']), - employee_type = dict(default=None, - type='str', - aliases=['employeeType']), - firstname = dict(default=None, - type='str'), - gecos = dict(default=None, - type='str'), - groups = dict(default=[], - type='list'), - home_share = dict(default=None, - type='str', - aliases=['homeShare']), - home_share_path = dict(default=None, - type='str', - aliases=['homeSharePath']), - home_telephone_number = dict(default=[], - type='list', - aliases=['homeTelephoneNumber']), - homedrive = dict(default=None, - type='str'), - lastname = dict(default=None, - type='str'), - mail_alternative_address= dict(default=[], - type='list', - aliases=['mailAlternativeAddress']), - mail_home_server = dict(default=None, - type='str', - aliases=['mailHomeServer']), - mail_primary_address = dict(default=None, - type='str', - aliases=['mailPrimaryAddress']), - mobile_telephone_number = dict(default=[], - type='list', - aliases=['mobileTelephoneNumber']), - organisation = dict(default=None, - type='str'), - overridePWHistory = dict(default=False, - type='bool', - aliases=['override_pw_history']), - overridePWLength = dict(default=False, - type='bool', - aliases=['override_pw_length']), - pager_telephonenumber = dict(default=[], - type='list', - aliases=['pagerTelephonenumber']), - password = dict(default=None, - type='str', - no_log=True), - phone = dict(default=[], - type='list'), - postcode = dict(default=None, - type='str'), - primary_group = dict(default=None, - type='str', - aliases=['primaryGroup']), - profilepath = dict(default=None, - type='str'), - pwd_change_next_login = dict(default=None, - type='str', - choices=['0', '1'], - aliases=['pwdChangeNextLogin']), - room_number = dict(default=None, - type='str', - aliases=['roomNumber']), - samba_privileges = dict(default=[], - type='list', - aliases=['sambaPrivileges']), - samba_user_workstations = dict(default=[], - type='list', - aliases=['sambaUserWorkstations']), - sambahome = dict(default=None, - type='str'), - scriptpath = dict(default=None, - type='str'), - secretary = dict(default=[], - type='list'), - serviceprovider = dict(default=[''], - type='list'), - shell = dict(default='/bin/bash', - type='str'), - street = dict(default=None, - type='str'), - title = dict(default=None, - type='str'), - unixhome = dict(default=None, - type='str'), - userexpiry = dict(default=expiry, - type='str'), - username = dict(required=True, - aliases=['name'], - type='str'), - position = dict(default='', - type='str'), - update_password = dict(default='always', - choices=['always', 'on_create'], - type='str'), - ou = dict(default='', - type='str'), - subpath = dict(default='cn=users', - type='str'), - state = dict(default='present', - choices=['present', 'absent'], - type='str') + argument_spec=dict( + birthday=dict(default=None, + type='str'), + city=dict(default=None, + type='str'), + country=dict(default=None, + type='str'), + department_number=dict(default=None, + type='str', + aliases=['departmentNumber']), + description=dict(default=None, + type='str'), + display_name=dict(default=None, + type='str', + aliases=['displayName']), + email=dict(default=[''], + type='list'), + employee_number=dict(default=None, + type='str', + aliases=['employeeNumber']), + employee_type=dict(default=None, + type='str', + aliases=['employeeType']), + firstname=dict(default=None, + type='str'), + gecos=dict(default=None, + type='str'), + groups=dict(default=[], + type='list'), + home_share=dict(default=None, + type='str', + aliases=['homeShare']), + home_share_path=dict(default=None, + type='str', + aliases=['homeSharePath']), + home_telephone_number=dict(default=[], + type='list', + aliases=['homeTelephoneNumber']), + homedrive=dict(default=None, + type='str'), + lastname=dict(default=None, + type='str'), + mail_alternative_address=dict(default=[], + type='list', + aliases=['mailAlternativeAddress']), + mail_home_server=dict(default=None, + type='str', + aliases=['mailHomeServer']), + mail_primary_address=dict(default=None, + type='str', + aliases=['mailPrimaryAddress']), + mobile_telephone_number=dict(default=[], + type='list', + aliases=['mobileTelephoneNumber']), + organisation=dict(default=None, + type='str'), + overridePWHistory=dict(default=False, + type='bool', + aliases=['override_pw_history']), + overridePWLength=dict(default=False, + type='bool', + aliases=['override_pw_length']), + pager_telephonenumber=dict(default=[], + type='list', + aliases=['pagerTelephonenumber']), + password=dict(default=None, + type='str', + no_log=True), + phone=dict(default=[], + type='list'), + postcode=dict(default=None, + type='str'), + primary_group=dict(default=None, + type='str', + aliases=['primaryGroup']), + profilepath=dict(default=None, + type='str'), + pwd_change_next_login=dict(default=None, + type='str', + choices=['0', '1'], + aliases=['pwdChangeNextLogin']), + room_number=dict(default=None, + type='str', + aliases=['roomNumber']), + samba_privileges=dict(default=[], + type='list', + aliases=['sambaPrivileges']), + samba_user_workstations=dict(default=[], + type='list', + aliases=['sambaUserWorkstations']), + sambahome=dict(default=None, + type='str'), + scriptpath=dict(default=None, + type='str'), + secretary=dict(default=[], + type='list'), + serviceprovider=dict(default=[''], + type='list'), + shell=dict(default='/bin/bash', + type='str'), + street=dict(default=None, + type='str'), + title=dict(default=None, + type='str'), + unixhome=dict(default=None, + type='str'), + userexpiry=dict(default=expiry, + type='str'), + username=dict(required=True, + aliases=['name'], + type='str'), + position=dict(default='', + type='str'), + update_password=dict(default='always', + choices=['always', 'on_create'], + type='str'), + ou=dict(default='', + type='str'), + subpath=dict(default='cn=users', + type='str'), + state=dict(default='present', + choices=['present', 'absent'], + type='str') ), supports_check_mode=True, - required_if = ([ + required_if=([ ('state', 'present', ['firstname', 'lastname', 'password']) ]) ) - username = module.params['username'] - position = module.params['position'] - ou = module.params['ou'] - subpath = module.params['subpath'] - state = module.params['state'] - changed = False + username = module.params['username'] + position = module.params['position'] + ou = module.params['ou'] + subpath = module.params['subpath'] + state = module.params['state'] + changed = False users = list(ldap_search( '(&(objectClass=posixAccount)(uid={}))'.format(username), @@ -532,7 +532,7 @@ def main(): old_password = obj['password'].split('}', 2)[1] if crypt.crypt(password, old_password) != old_password: obj['overridePWHistory'] = module.params['overridePWHistory'] - obj['overridePWLength'] = module.params['overridePWLength'] + obj['overridePWLength'] = module.params['overridePWLength'] obj['password'] = password diff = obj.diff() diff --git a/lib/ansible/modules/cloud/vmware/vmware_dns_config.py b/lib/ansible/modules/cloud/vmware/vmware_dns_config.py index 0f4de41bb5..a60ac4d691 100644 --- a/lib/ansible/modules/cloud/vmware/vmware_dns_config.py +++ b/lib/ansible/modules/cloud/vmware/vmware_dns_config.py @@ -94,8 +94,8 @@ def main(): argument_spec = vmware_argument_spec() argument_spec.update(dict(change_hostname_to=dict(required=True, type='str'), - domainname=dict(required=True, type='str'), - dns_servers=dict(required=True, type='list'))) + domainname=dict(required=True, type='str'), + dns_servers=dict(required=True, type='list'))) module = AnsibleModule(argument_spec=argument_spec, supports_check_mode=False) diff --git a/lib/ansible/modules/cloud/vmware/vmware_dvs_host.py b/lib/ansible/modules/cloud/vmware/vmware_dvs_host.py index 54b3aedf48..e90d2ff2b6 100644 --- a/lib/ansible/modules/cloud/vmware/vmware_dvs_host.py +++ b/lib/ansible/modules/cloud/vmware/vmware_dvs_host.py @@ -234,9 +234,9 @@ def main(): argument_spec = vmware_argument_spec() argument_spec.update(dict(esxi_hostname=dict(required=True, type='str'), - switch_name=dict(required=True, type='str'), - vmnics=dict(required=True, type='list'), - state=dict(default='present', choices=['present', 'absent'], type='str'))) + switch_name=dict(required=True, type='str'), + vmnics=dict(required=True, type='list'), + state=dict(default='present', choices=['present', 'absent'], type='str'))) module = AnsibleModule(argument_spec=argument_spec, supports_check_mode=True) diff --git a/lib/ansible/modules/cloud/vmware/vmware_guest.py b/lib/ansible/modules/cloud/vmware/vmware_guest.py index 66167ef6e7..185ac9b1fd 100644 --- a/lib/ansible/modules/cloud/vmware/vmware_guest.py +++ b/lib/ansible/modules/cloud/vmware/vmware_guest.py @@ -507,6 +507,7 @@ class PyVmomiDeviceHelper(object): class PyVmomiCache(object): """ This class caches references to objects which are requested multiples times but not modified """ + def __init__(self, content, dc_name=None): self.content = content self.dc_name = dc_name @@ -849,7 +850,7 @@ class PyVmomiHelper(PyVmomi): if (nic.device.backing and not hasattr(nic.device.backing, 'port')): nic_change_detected = True elif (nic.device.backing and (nic.device.backing.port.portgroupKey != pg_obj.key or - nic.device.backing.port.switchUuid != pg_obj.config.distributedVirtualSwitch.uuid)): + nic.device.backing.port.switchUuid != pg_obj.config.distributedVirtualSwitch.uuid)): nic_change_detected = True dvs_port_connection = vim.dvs.PortConnection() diff --git a/lib/ansible/modules/cloud/vmware/vmware_guest_find.py b/lib/ansible/modules/cloud/vmware/vmware_guest_find.py index ad32c8e2a2..f4651fa48d 100644 --- a/lib/ansible/modules/cloud/vmware/vmware_guest_find.py +++ b/lib/ansible/modules/cloud/vmware/vmware_guest_find.py @@ -156,7 +156,6 @@ class PyVmomiHelper(object): return tree def _build_folder_map(self, folder, inpath='/'): - """ Build a searchable index for vms+uuids+folders """ if isinstance(folder, tuple): folder = folder[1] diff --git a/lib/ansible/modules/cloud/vmware/vmware_vmkernel.py b/lib/ansible/modules/cloud/vmware/vmware_vmkernel.py index b16d397990..167df25fd6 100644 --- a/lib/ansible/modules/cloud/vmware/vmware_vmkernel.py +++ b/lib/ansible/modules/cloud/vmware/vmware_vmkernel.py @@ -154,15 +154,15 @@ def main(): argument_spec = vmware_argument_spec() argument_spec.update(dict(portgroup_name=dict(required=True, type='str'), - ip_address=dict(required=True, type='str'), - subnet_mask=dict(required=True, type='str'), - mtu=dict(required=False, type='int'), - enable_vsan=dict(required=False, type='bool'), - enable_vmotion=dict(required=False, type='bool'), - enable_mgmt=dict(required=False, type='bool'), - enable_ft=dict(required=False, type='bool'), - vswitch_name=dict(required=True, type='str'), - vlan_id=dict(required=True, type='int'))) + ip_address=dict(required=True, type='str'), + subnet_mask=dict(required=True, type='str'), + mtu=dict(required=False, type='int'), + enable_vsan=dict(required=False, type='bool'), + enable_vmotion=dict(required=False, type='bool'), + enable_mgmt=dict(required=False, type='bool'), + enable_ft=dict(required=False, type='bool'), + vswitch_name=dict(required=True, type='str'), + vlan_id=dict(required=True, type='int'))) module = AnsibleModule(argument_spec=argument_spec, supports_check_mode=False) diff --git a/lib/ansible/modules/cloud/vmware/vmware_vmkernel_ip_config.py b/lib/ansible/modules/cloud/vmware/vmware_vmkernel_ip_config.py index 215d4d974f..82348fd0fd 100644 --- a/lib/ansible/modules/cloud/vmware/vmware_vmkernel_ip_config.py +++ b/lib/ansible/modules/cloud/vmware/vmware_vmkernel_ip_config.py @@ -87,8 +87,8 @@ def main(): argument_spec = vmware_argument_spec() argument_spec.update(dict(vmk_name=dict(required=True, type='str'), - ip_address=dict(required=True, type='str'), - subnet_mask=dict(required=True, type='str'))) + ip_address=dict(required=True, type='str'), + subnet_mask=dict(required=True, type='str'))) module = AnsibleModule(argument_spec=argument_spec, supports_check_mode=False) diff --git a/lib/ansible/modules/cloud/vmware/vsphere_guest.py b/lib/ansible/modules/cloud/vmware/vsphere_guest.py index 049ea178c7..d7cb384a2a 100644 --- a/lib/ansible/modules/cloud/vmware/vsphere_guest.py +++ b/lib/ansible/modules/cloud/vmware/vsphere_guest.py @@ -627,6 +627,7 @@ def spec_singleton(spec, request, vm): spec = request.new_spec() return spec + def get_cdrom_params(module, s, vm_cdrom): cdrom_type = None cdrom_iso_path = None @@ -648,6 +649,7 @@ def get_cdrom_params(module, s, vm_cdrom): return cdrom_type, cdrom_iso_path + def vmdisk_id(vm, current_datastore_name): id_list = [] for vm_disk in vm._disks: @@ -666,9 +668,9 @@ def deploy_template(vsphere_client, guest, resource_pool, template_src, esxi, mo # Datacenter managed object reference dclist = [k for k, - v in vsphere_client.get_datacenters().items() if v == datacenter] + v in vsphere_client.get_datacenters().items() if v == datacenter] if dclist: - dcmor=dclist[0] + dcmor = dclist[0] else: vsphere_client.disconnect() module.fail_json(msg="Cannot find datacenter named: %s" % datacenter) @@ -744,7 +746,7 @@ def deploy_template(vsphere_client, guest, resource_pool, template_src, esxi, mo cloneArgs = dict(resourcepool=rpmor, power_on=False) if snapshot_to_clone is not None: - #check if snapshot_to_clone is specified, Create a Linked Clone instead of a full clone. + # check if snapshot_to_clone is specified, Create a Linked Clone instead of a full clone. cloneArgs["linked"] = True cloneArgs["snapshot"] = snapshot_to_clone @@ -778,6 +780,8 @@ def deploy_template(vsphere_client, guest, resource_pool, template_src, esxi, mo # example from https://github.com/kalazzerx/pysphere/blob/master/examples/pysphere_create_disk_and_add_to_vm.py # was used. + + def update_disks(vsphere_client, vm, module, vm_disk, changes): request = VI.ReconfigVM_TaskRequestMsg() changed = False @@ -868,7 +872,7 @@ def reconfigure_vm(vsphere_client, vm, module, esxi, resource_pool, cluster_name if vm_extra_config: spec = spec_singleton(spec, request, vm) extra_config = [] - for k,v in vm_extra_config.items(): + for k, v in vm_extra_config.items(): ec = spec.new_extraConfig() ec.set_element_key(str(k)) ec.set_element_value(str(v)) @@ -988,7 +992,7 @@ def reconfigure_vm(vsphere_client, vm, module, esxi, resource_pool, cluster_name spec = spec_singleton(spec, request, vm) # Get a list of the VM's hard drives - dev_list = [d for d in vm.properties.config.hardware.device if d._type=='VirtualDisk'] + dev_list = [d for d in vm.properties.config.hardware.device if d._type == 'VirtualDisk'] if len(vm_disk) > len(dev_list): vsphere_client.disconnect() module.fail_json(msg="Error in vm_disk definition. Too many disks defined in comparison to the VM's disk profile.") @@ -1072,102 +1076,102 @@ def reconfigure_vm(vsphere_client, vm, module, esxi, resource_pool, cluster_name def reconfigure_net(vsphere_client, vm, module, esxi, resource_pool, guest, vm_nic, cluster_name=None): - s = vsphere_client - nics = {} - request = VI.ReconfigVM_TaskRequestMsg() - _this = request.new__this(vm._mor) - _this.set_attribute_type(vm._mor.get_attribute_type()) - request.set_element__this(_this) - nic_changes = [] - datacenter = esxi['datacenter'] - # Datacenter managed object reference - dclist = [k for k, - v in vsphere_client.get_datacenters().items() if v == datacenter] - if dclist: - dcmor=dclist[0] - else: - vsphere_client.disconnect() - module.fail_json(msg="Cannot find datacenter named: %s" % datacenter) - dcprops = VIProperty(vsphere_client, dcmor) - nfmor = dcprops.networkFolder._obj - for k,v in vm_nic.items(): - nicNum = k[len(k) -1] - if vm_nic[k]['network_type'] == 'dvs': - portgroupKey = find_portgroup_key(module, s, nfmor, vm_nic[k]['network']) - todvs = True - elif vm_nic[k]['network_type'] == 'standard': - todvs = False - # Detect cards that need to be changed and network type (and act accordingly) - for dev in vm.properties.config.hardware.device: - if dev._type in ["VirtualE1000", "VirtualE1000e", - "VirtualPCNet32", "VirtualVmxnet", - "VirtualNmxnet2", "VirtualVmxnet3"]: - devNum = dev.deviceInfo.label[len(dev.deviceInfo.label) - 1] - if devNum == nicNum: - fromdvs = dev.deviceInfo.summary.split(':')[0] == 'DVSwitch' - if todvs and fromdvs: - if dev.backing.port._obj.get_element_portgroupKey() != portgroupKey: - nics[k] = (dev, portgroupKey, 1) - elif fromdvs and not todvs: + s = vsphere_client + nics = {} + request = VI.ReconfigVM_TaskRequestMsg() + _this = request.new__this(vm._mor) + _this.set_attribute_type(vm._mor.get_attribute_type()) + request.set_element__this(_this) + nic_changes = [] + datacenter = esxi['datacenter'] + # Datacenter managed object reference + dclist = [k for k, + v in vsphere_client.get_datacenters().items() if v == datacenter] + if dclist: + dcmor = dclist[0] + else: + vsphere_client.disconnect() + module.fail_json(msg="Cannot find datacenter named: %s" % datacenter) + dcprops = VIProperty(vsphere_client, dcmor) + nfmor = dcprops.networkFolder._obj + for k, v in vm_nic.items(): + nicNum = k[len(k) - 1] + if vm_nic[k]['network_type'] == 'dvs': + portgroupKey = find_portgroup_key(module, s, nfmor, vm_nic[k]['network']) + todvs = True + elif vm_nic[k]['network_type'] == 'standard': + todvs = False + # Detect cards that need to be changed and network type (and act accordingly) + for dev in vm.properties.config.hardware.device: + if dev._type in ["VirtualE1000", "VirtualE1000e", + "VirtualPCNet32", "VirtualVmxnet", + "VirtualNmxnet2", "VirtualVmxnet3"]: + devNum = dev.deviceInfo.label[len(dev.deviceInfo.label) - 1] + if devNum == nicNum: + fromdvs = dev.deviceInfo.summary.split(':')[0] == 'DVSwitch' + if todvs and fromdvs: + if dev.backing.port._obj.get_element_portgroupKey() != portgroupKey: + nics[k] = (dev, portgroupKey, 1) + elif fromdvs and not todvs: + nics[k] = (dev, '', 2) + elif not fromdvs and todvs: + nics[k] = (dev, portgroupKey, 3) + elif not fromdvs and not todvs: + if dev.backing._obj.get_element_deviceName() != vm_nic[k]['network']: nics[k] = (dev, '', 2) - elif not fromdvs and todvs: - nics[k] = (dev, portgroupKey, 3) - elif not fromdvs and not todvs: - if dev.backing._obj.get_element_deviceName() != vm_nic[k]['network']: - nics[k] = (dev, '', 2) - else: - pass else: - module.exit_json() - - if len(nics) > 0: - for nic, obj in nics.items(): - """ - 1,2 and 3 are used to mark which action should be taken - 1 = from a distributed switch to a distributed switch - 2 = to a standard switch - 3 = to a distributed switch - """ - dev = obj[0] - pgKey = obj[1] - dvsKey = obj[2] - if dvsKey == 1: - dev.backing.port._obj.set_element_portgroupKey(pgKey) - dev.backing.port._obj.set_element_portKey('') - if dvsKey == 3: - dvswitch_uuid = find_dvswitch_uuid(module, s, nfmor, pgKey) - nic_backing_port = VI.ns0.DistributedVirtualSwitchPortConnection_Def( - "nic_backing_port").pyclass() - nic_backing_port.set_element_switchUuid(dvswitch_uuid) - nic_backing_port.set_element_portgroupKey(pgKey) - nic_backing_port.set_element_portKey('') - nic_backing = VI.ns0.VirtualEthernetCardDistributedVirtualPortBackingInfo_Def( - "nic_backing").pyclass() - nic_backing.set_element_port(nic_backing_port) - dev._obj.set_element_backing(nic_backing) - if dvsKey == 2: - nic_backing = VI.ns0.VirtualEthernetCardNetworkBackingInfo_Def( - "nic_backing").pyclass() - nic_backing.set_element_deviceName(vm_nic[nic]['network']) - dev._obj.set_element_backing(nic_backing) - for nic, obj in nics.items(): - dev = obj[0] - spec = request.new_spec() - nic_change = spec.new_deviceChange() - nic_change.set_element_device(dev._obj) - nic_change.set_element_operation("edit") - nic_changes.append(nic_change) - spec.set_element_deviceChange(nic_changes) - request.set_element_spec(spec) - ret = vsphere_client._proxy.ReconfigVM_Task(request)._returnval - task = VITask(ret, vsphere_client) - status = task.wait_for_state([task.STATE_SUCCESS, task.STATE_ERROR]) - if status == task.STATE_SUCCESS: - return(True) - elif status == task.STATE_ERROR: - module.fail_json(msg="Could not change network %s" % task.get_error_message()) - elif len(nics) == 0: - return(False) + pass + else: + module.exit_json() + + if len(nics) > 0: + for nic, obj in nics.items(): + """ + 1,2 and 3 are used to mark which action should be taken + 1 = from a distributed switch to a distributed switch + 2 = to a standard switch + 3 = to a distributed switch + """ + dev = obj[0] + pgKey = obj[1] + dvsKey = obj[2] + if dvsKey == 1: + dev.backing.port._obj.set_element_portgroupKey(pgKey) + dev.backing.port._obj.set_element_portKey('') + if dvsKey == 3: + dvswitch_uuid = find_dvswitch_uuid(module, s, nfmor, pgKey) + nic_backing_port = VI.ns0.DistributedVirtualSwitchPortConnection_Def( + "nic_backing_port").pyclass() + nic_backing_port.set_element_switchUuid(dvswitch_uuid) + nic_backing_port.set_element_portgroupKey(pgKey) + nic_backing_port.set_element_portKey('') + nic_backing = VI.ns0.VirtualEthernetCardDistributedVirtualPortBackingInfo_Def( + "nic_backing").pyclass() + nic_backing.set_element_port(nic_backing_port) + dev._obj.set_element_backing(nic_backing) + if dvsKey == 2: + nic_backing = VI.ns0.VirtualEthernetCardNetworkBackingInfo_Def( + "nic_backing").pyclass() + nic_backing.set_element_deviceName(vm_nic[nic]['network']) + dev._obj.set_element_backing(nic_backing) + for nic, obj in nics.items(): + dev = obj[0] + spec = request.new_spec() + nic_change = spec.new_deviceChange() + nic_change.set_element_device(dev._obj) + nic_change.set_element_operation("edit") + nic_changes.append(nic_change) + spec.set_element_deviceChange(nic_changes) + request.set_element_spec(spec) + ret = vsphere_client._proxy.ReconfigVM_Task(request)._returnval + task = VITask(ret, vsphere_client) + status = task.wait_for_state([task.STATE_SUCCESS, task.STATE_ERROR]) + if status == task.STATE_SUCCESS: + return(True) + elif status == task.STATE_ERROR: + module.fail_json(msg="Could not change network %s" % task.get_error_message()) + elif len(nics) == 0: + return(False) def _build_folder_tree(nodes, parent): @@ -1218,9 +1222,9 @@ def create_vm(vsphere_client, module, esxi, resource_pool, cluster_name, guest, esxi_hostname = esxi['hostname'] # Datacenter managed object reference dclist = [k for k, - v in vsphere_client.get_datacenters().items() if v == datacenter] + v in vsphere_client.get_datacenters().items() if v == datacenter] if dclist: - dcmor=dclist[0] + dcmor = dclist[0] else: vsphere_client.disconnect() module.fail_json(msg="Cannot find datacenter named: %s" % datacenter) @@ -1419,7 +1423,7 @@ def create_vm(vsphere_client, module, esxi, resource_pool, cluster_name, guest, " to be specified." % vm_hardware['vm_floppy']) # Add a floppy to the VM. add_floppy(module, vsphere_client, config_target, config, devices, - default_devs, floppy_type, floppy_image_path) + default_devs, floppy_type, floppy_image_path) if vm_nic: for nic in sorted(vm_nic): try: @@ -1479,7 +1483,7 @@ def create_vm(vsphere_client, module, esxi, resource_pool, cluster_name, guest, # Power on the VM if it was requested power_state(vm, state, True) - vmfacts=gather_facts(vm) + vmfacts = gather_facts(vm) vsphere_client.disconnect() module.exit_json( ansible_facts=vmfacts, @@ -1579,13 +1583,13 @@ def gather_facts(vm): 'module_hw': True, 'hw_name': vm.properties.name, 'hw_power_status': vm.get_status(), - 'hw_guest_full_name': vm.properties.config.guestFullName, + 'hw_guest_full_name': vm.properties.config.guestFullName, 'hw_guest_id': vm.properties.config.guestId, 'hw_product_uuid': vm.properties.config.uuid, 'hw_instance_uuid': vm.properties.config.instanceUuid, 'hw_processor_count': vm.properties.config.hardware.numCPU, 'hw_memtotal_mb': vm.properties.config.hardware.memoryMB, - 'hw_interfaces':[], + 'hw_interfaces': [], } netInfo = vm.get_property('net') netDict = {} @@ -1608,7 +1612,7 @@ def gather_facts(vm): 'macaddress_dash': entry.macAddress.replace(':', '-'), 'summary': entry.deviceInfo.summary, } - facts['hw_interfaces'].append('eth'+str(ifidx)) + facts['hw_interfaces'].append('eth' + str(ifidx)) ifidx += 1 @@ -1753,7 +1757,7 @@ def main(): ), supports_check_mode=False, - mutually_exclusive=[['state', 'vmware_guest_facts'],['state', 'from_template']], + mutually_exclusive=[['state', 'vmware_guest_facts'], ['state', 'from_template']], required_together=[ ['state', 'force'], [ @@ -1791,7 +1795,6 @@ def main(): power_on_after_clone = module.params['power_on_after_clone'] validate_certs = module.params['validate_certs'] - # CONNECT TO THE SERVER viserver = VIServer() if validate_certs and not hasattr(ssl, 'SSLContext') and not vcenter_hostname.startswith('http://'): @@ -1896,10 +1899,9 @@ def main(): # check if user is trying to perform state operation on a vm which doesn't exists elif state in ['present', 'powered_off', 'powered_on'] and not all((vm_extra_config, - vm_hardware, vm_disk, vm_nic, esxi)): + vm_hardware, vm_disk, vm_nic, esxi)): module.exit_json(changed=False, msg="vm %s not present" % guest) - # Create the VM elif state in ['present', 'powered_off', 'powered_on']: diff --git a/lib/ansible/modules/cloud/webfaction/webfaction_app.py b/lib/ansible/modules/cloud/webfaction/webfaction_app.py index 9f9796160a..d2f6a3a6e7 100644 --- a/lib/ansible/modules/cloud/webfaction/webfaction_app.py +++ b/lib/ansible/modules/cloud/webfaction/webfaction_app.py @@ -110,20 +110,20 @@ webfaction = xmlrpclib.ServerProxy('https://api.webfaction.com/') def main(): module = AnsibleModule( - argument_spec = dict( - name = dict(required=True), - state = dict(required=False, choices=['present', 'absent'], default='present'), - type = dict(required=True), - autostart = dict(required=False, type='bool', default=False), - extra_info = dict(required=False, default=""), - port_open = dict(required=False, type='bool', default=False), - login_name = dict(required=True), - login_password = dict(required=True, no_log=True), - machine = dict(required=False, default=False), + argument_spec=dict( + name=dict(required=True), + state=dict(required=False, choices=['present', 'absent'], default='present'), + type=dict(required=True), + autostart=dict(required=False, type='bool', default=False), + extra_info=dict(required=False, default=""), + port_open=dict(required=False, type='bool', default=False), + login_name=dict(required=True), + login_password=dict(required=True, no_log=True), + machine=dict(required=False, default=False), ), supports_check_mode=True ) - app_name = module.params['name'] + app_name = module.params['name'] app_type = module.params['type'] app_state = module.params['state'] @@ -157,7 +157,7 @@ def main(): # If it exists with the right type, we don't change it # Should check other parameters. module.exit_json( - changed = False, + changed=False, ) if not module.check_mode: @@ -176,7 +176,7 @@ def main(): # If the app's already not there, nothing changed. if not existing_app: module.exit_json( - changed = False, + changed=False, ) if not module.check_mode: @@ -188,10 +188,9 @@ def main(): else: module.fail_json(msg="Unknown state specified: {}".format(app_state)) - module.exit_json( - changed = True, - result = result + changed=True, + result=result ) diff --git a/lib/ansible/modules/cloud/webfaction/webfaction_db.py b/lib/ansible/modules/cloud/webfaction/webfaction_db.py index 35476327d4..deb8fd4e87 100644 --- a/lib/ansible/modules/cloud/webfaction/webfaction_db.py +++ b/lib/ansible/modules/cloud/webfaction/webfaction_db.py @@ -102,21 +102,21 @@ webfaction = xmlrpclib.ServerProxy('https://api.webfaction.com/') def main(): module = AnsibleModule( - argument_spec = dict( - name = dict(required=True), - state = dict(required=False, choices=['present', 'absent'], default='present'), + argument_spec=dict( + name=dict(required=True), + state=dict(required=False, choices=['present', 'absent'], default='present'), # You can specify an IP address or hostname. - type = dict(required=True), - password = dict(required=False, default=None, no_log=True), - login_name = dict(required=True), - login_password = dict(required=True, no_log=True), - machine = dict(required=False, default=False), + type=dict(required=True), + password=dict(required=False, default=None, no_log=True), + login_name=dict(required=True), + login_password=dict(required=True, no_log=True), + machine=dict(required=False, default=False), ), supports_check_mode=True ) - db_name = module.params['name'] + db_name = module.params['name'] db_state = module.params['state'] - db_type = module.params['type'] + db_type = module.params['type'] db_passwd = module.params['password'] if module.params['machine']: @@ -153,10 +153,9 @@ def main(): # If it exists with the right type, we don't change anything. module.exit_json( - changed = False, + changed=False, ) - if not module.check_mode: # If this isn't a dry run, create the db # and default user. @@ -172,7 +171,7 @@ def main(): if not module.check_mode: if not (existing_db or existing_user): - module.exit_json(changed = False,) + module.exit_json(changed=False,) if existing_db: # Delete the db if it exists @@ -190,8 +189,8 @@ def main(): module.fail_json(msg="Unknown state specified: {}".format(db_state)) module.exit_json( - changed = True, - result = result + changed=True, + result=result ) diff --git a/lib/ansible/modules/cloud/webfaction/webfaction_domain.py b/lib/ansible/modules/cloud/webfaction/webfaction_domain.py index 65a7e60beb..2b32de866e 100644 --- a/lib/ansible/modules/cloud/webfaction/webfaction_domain.py +++ b/lib/ansible/modules/cloud/webfaction/webfaction_domain.py @@ -93,16 +93,16 @@ webfaction = xmlrpclib.ServerProxy('https://api.webfaction.com/') def main(): module = AnsibleModule( - argument_spec = dict( - name = dict(required=True), - state = dict(required=False, choices=['present', 'absent'], default='present'), - subdomains = dict(required=False, default=[]), - login_name = dict(required=True), - login_password = dict(required=True, no_log=True), + argument_spec=dict( + name=dict(required=True), + state=dict(required=False, choices=['present', 'absent'], default='present'), + subdomains=dict(required=False, default=[]), + login_name=dict(required=True), + login_password=dict(required=True, no_log=True), ), supports_check_mode=True ) - domain_name = module.params['name'] + domain_name = module.params['name'] domain_state = module.params['state'] domain_subdomains = module.params['subdomains'] @@ -127,7 +127,7 @@ def main(): if set(existing_domain['subdomains']) >= set(domain_subdomains): # If it exists with the right subdomains, we don't change anything. module.exit_json( - changed = False, + changed=False, ) positional_args = [session_id, domain_name] + domain_subdomains @@ -146,7 +146,7 @@ def main(): # If the app's already not there, nothing changed. if not existing_domain: module.exit_json( - changed = False, + changed=False, ) positional_args = [session_id, domain_name] + domain_subdomains @@ -161,8 +161,8 @@ def main(): module.fail_json(msg="Unknown state specified: {}".format(domain_state)) module.exit_json( - changed = True, - result = result + changed=True, + result=result ) diff --git a/lib/ansible/modules/cloud/webfaction/webfaction_site.py b/lib/ansible/modules/cloud/webfaction/webfaction_site.py index ab5e583c5a..bee932419f 100644 --- a/lib/ansible/modules/cloud/webfaction/webfaction_site.py +++ b/lib/ansible/modules/cloud/webfaction/webfaction_site.py @@ -108,20 +108,20 @@ webfaction = xmlrpclib.ServerProxy('https://api.webfaction.com/') def main(): module = AnsibleModule( - argument_spec = dict( - name = dict(required=True), - state = dict(required=False, choices=['present', 'absent'], default='present'), + argument_spec=dict( + name=dict(required=True), + state=dict(required=False, choices=['present', 'absent'], default='present'), # You can specify an IP address or hostname. - host = dict(required=True), - https = dict(required=False, type='bool', default=False), - subdomains = dict(required=False, type='list', default=[]), - site_apps = dict(required=False, type='list', default=[]), - login_name = dict(required=True), - login_password = dict(required=True, no_log=True), + host=dict(required=True), + https=dict(required=False, type='bool', default=False), + subdomains=dict(required=False, type='list', default=[]), + site_apps=dict(required=False, type='list', default=[]), + login_name=dict(required=True), + login_password=dict(required=True, no_log=True), ), supports_check_mode=True ) - site_name = module.params['name'] + site_name = module.params['name'] site_state = module.params['state'] site_host = module.params['host'] site_ip = socket.gethostbyname(site_host) @@ -159,7 +159,7 @@ def main(): (set(existing_site['subdomains']) == set(module.params['subdomains'])) and \ (dict(existing_site['website_apps']) == dict(module.params['site_apps'])): module.exit_json( - changed = False + changed=False ) positional_args = [ @@ -168,16 +168,16 @@ def main(): module.params['subdomains'], ] for a in module.params['site_apps']: - positional_args.append( (a[0], a[1]) ) + positional_args.append((a[0], a[1])) if not module.check_mode: # If this isn't a dry run, create or modify the site result.update( webfaction.create_website( *positional_args - ) if not existing_site else webfaction.update_website ( - *positional_args - ) + ) if not existing_site else webfaction.update_website( + *positional_args + ) ) elif site_state == 'absent': @@ -185,7 +185,7 @@ def main(): # If the site's already not there, nothing changed. if not existing_site: module.exit_json( - changed = False, + changed=False, ) if not module.check_mode: @@ -198,8 +198,8 @@ def main(): module.fail_json(msg="Unknown state specified: {}".format(site_state)) module.exit_json( - changed = True, - result = result + changed=True, + result=result ) diff --git a/lib/ansible/modules/clustering/consul_acl.py b/lib/ansible/modules/clustering/consul_acl.py index 6c304381d3..446e80ee27 100644 --- a/lib/ansible/modules/clustering/consul_acl.py +++ b/lib/ansible/modules/clustering/consul_acl.py @@ -456,6 +456,7 @@ class Configuration: """ Configuration for this module. """ + def __init__(self, management_token=None, host=None, scheme=None, validate_certs=None, name=None, port=None, rules=None, state=None, token=None, token_type=None): self.management_token = management_token # type: str @@ -474,6 +475,7 @@ class Output: """ Output of an action of this module. """ + def __init__(self, changed=None, token=None, rules=None, operation=None): self.changed = changed # type: bool self.token = token # type: str @@ -485,6 +487,7 @@ class ACL: """ Consul ACL. See: https://www.consul.io/docs/guides/acl.html. """ + def __init__(self, rules, token_type, token, name): self.rules = rules self.token_type = token_type @@ -507,6 +510,7 @@ class Rule: """ ACL rule. See: https://www.consul.io/docs/guides/acl.html#acl-rules-and-scope. """ + def __init__(self, scope, policy, pattern=None): self.scope = scope self.policy = policy @@ -532,6 +536,7 @@ class RuleCollection: """ Collection of ACL rules, which are part of a Consul ACL. """ + def __init__(self): self._rules = {} for scope in RULE_SCOPES: diff --git a/lib/ansible/modules/database/misc/kibana_plugin.py b/lib/ansible/modules/database/misc/kibana_plugin.py index f5bb2c0be8..f56f36a87e 100644 --- a/lib/ansible/modules/database/misc/kibana_plugin.py +++ b/lib/ansible/modules/database/misc/kibana_plugin.py @@ -147,9 +147,11 @@ def parse_plugin_repo(string): return repo + def is_plugin_present(plugin_dir, working_dir): return os.path.isdir(os.path.join(working_dir, plugin_dir)) + def parse_error(string): reason = "reason: " try: @@ -157,6 +159,7 @@ def parse_error(string): except ValueError: return string + def install_plugin(module, plugin_bin, plugin_name, url, timeout): cmd_args = [plugin_bin, "plugin", PACKAGE_STATE_MAP["present"], plugin_name] @@ -178,6 +181,7 @@ def install_plugin(module, plugin_bin, plugin_name, url, timeout): return True, cmd, out, err + def remove_plugin(module, plugin_bin, plugin_name): cmd_args = [plugin_bin, "plugin", PACKAGE_STATE_MAP["absent"], plugin_name] @@ -193,6 +197,7 @@ def remove_plugin(module, plugin_bin, plugin_name): return True, cmd, out, err + def main(): module = AnsibleModule( argument_spec=dict( @@ -208,14 +213,14 @@ def main(): supports_check_mode=True, ) - name = module.params["name"] - state = module.params["state"] - url = module.params["url"] - timeout = module.params["timeout"] - plugin_bin = module.params["plugin_bin"] - plugin_dir = module.params["plugin_dir"] - version = module.params["version"] - force = module.params["force"] + name = module.params["name"] + state = module.params["state"] + url = module.params["url"] + timeout = module.params["timeout"] + plugin_bin = module.params["plugin_bin"] + plugin_dir = module.params["plugin_dir"] + version = module.params["version"] + force = module.params["force"] present = is_plugin_present(parse_plugin_repo(name), plugin_dir) diff --git a/lib/ansible/modules/database/misc/riak.py b/lib/ansible/modules/database/misc/riak.py index 03b36b2ca5..13be9cab1f 100644 --- a/lib/ansible/modules/database/misc/riak.py +++ b/lib/ansible/modules/database/misc/riak.py @@ -102,6 +102,7 @@ def ring_check(module, riak_admin_bin): else: return False + def main(): module = AnsibleModule( @@ -115,10 +116,9 @@ def main(): wait_for_ring=dict(default=False, type='int'), wait_for_service=dict( required=False, default=None, choices=['kv']), - validate_certs = dict(default='yes', type='bool')) + validate_certs=dict(default='yes', type='bool')) ) - command = module.params.get('command') http_conn = module.params.get('http_conn') target_node = module.params.get('target_node') @@ -126,8 +126,7 @@ def main(): wait_for_ring = module.params.get('wait_for_ring') wait_for_service = module.params.get('wait_for_service') - - #make sure riak commands are on the path + # make sure riak commands are on the path riak_bin = module.get_bin_path('riak') riak_admin_bin = module.get_bin_path('riak-admin') @@ -150,16 +149,16 @@ def main(): node_name = stats['nodename'] nodes = stats['ring_members'] ring_size = stats['ring_creation_size'] - rc, out, err = module.run_command([riak_bin, 'version'] ) + rc, out, err = module.run_command([riak_bin, 'version']) version = out.strip() result = dict(node_name=node_name, - nodes=nodes, - ring_size=ring_size, - version=version) + nodes=nodes, + ring_size=ring_size, + version=version) if command == 'ping': - cmd = '%s ping %s' % ( riak_bin, target_node ) + cmd = '%s ping %s' % (riak_bin, target_node) rc, out, err = module.run_command(cmd) if rc == 0: result['ping'] = out @@ -219,7 +218,7 @@ def main(): module.fail_json(msg='Timeout waiting for handoffs.') if wait_for_service: - cmd = [riak_admin_bin, 'wait_for_service', 'riak_%s' % wait_for_service, node_name ] + cmd = [riak_admin_bin, 'wait_for_service', 'riak_%s' % wait_for_service, node_name] rc, out, err = module.run_command(cmd) result['service'] = out diff --git a/lib/ansible/modules/database/mongodb/mongodb_parameter.py b/lib/ansible/modules/database/mongodb/mongodb_parameter.py index fdd69ffc9c..28dbf13505 100644 --- a/lib/ansible/modules/database/mongodb/mongodb_parameter.py +++ b/lib/ansible/modules/database/mongodb/mongodb_parameter.py @@ -223,5 +223,5 @@ def main(): after=value) -if __name__ == '__main__': +if __name__ == '__main__': main() diff --git a/lib/ansible/modules/database/mongodb/mongodb_user.py b/lib/ansible/modules/database/mongodb/mongodb_user.py index 00f2620fce..9e00e967e4 100644 --- a/lib/ansible/modules/database/mongodb/mongodb_user.py +++ b/lib/ansible/modules/database/mongodb/mongodb_user.py @@ -270,8 +270,8 @@ def user_find(client, user, db_name): def user_add(module, client, db_name, user, password, roles): - #pymongo's user_add is a _create_or_update_user so we won't know if it was changed or updated - #without reproducing a lot of the logic in database.py of pymongo + # pymongo's user_add is a _create_or_update_user so we won't know if it was changed or updated + # without reproducing a lot of the logic in database.py of pymongo db = client[db_name] if roles is None: @@ -279,6 +279,7 @@ def user_add(module, client, db_name, user, password, roles): else: db.add_user(user, password, None, roles=roles) + def user_remove(module, client, db_name, user): exists = user_find(client, user, db_name) if exists: @@ -289,6 +290,7 @@ def user_remove(module, client, db_name, user): else: module.exit_json(changed=False, user=user) + def load_mongocnf(): config = configparser.RawConfigParser() mongocnf = os.path.expanduser('~/.mongodb.cnf') @@ -305,7 +307,6 @@ def load_mongocnf(): return creds - def check_if_roles_changed(uinfo, roles, db_name): # We must be aware of users which can read the oplog on a replicaset # Such users must have access to the local DB, but since this DB does not store users credentials @@ -327,7 +328,7 @@ def check_if_roles_changed(uinfo, roles, db_name): output = list() for role in roles: if isinstance(role, (binary_type, text_type)): - new_role = { "role": role, "db": db_name } + new_role = {"role": role, "db": db_name} output.append(new_role) else: output.append(role) @@ -341,14 +342,13 @@ def check_if_roles_changed(uinfo, roles, db_name): return True - # ========================================= # Module execution. # def main(): module = AnsibleModule( - argument_spec = dict( + argument_spec=dict( login_user=dict(default=None), login_password=dict(default=None, no_log=True), login_host=dict(default='localhost'), @@ -417,7 +417,7 @@ def main(): elif LooseVersion(PyMongoVersion) >= LooseVersion('3.0'): if db_name != "admin": module.fail_json(msg='The localhost login exception only allows the first admin account to be created') - #else: this has to be the first admin user added + # else: this has to be the first admin user added except Exception as e: module.fail_json(msg='unable to connect to database: %s' % to_native(e), exception=traceback.format_exc()) @@ -443,7 +443,7 @@ def main(): # Here we can check password change if mongo provide a query for that : https://jira.mongodb.org/browse/SERVER-22848 #newuinfo = user_find(client, user, db_name) - #if uinfo['role'] == newuinfo['role'] and CheckPasswordHere: + # if uinfo['role'] == newuinfo['role'] and CheckPasswordHere: # module.exit_json(changed=False, user=user) elif state == 'absent': diff --git a/lib/ansible/modules/database/mssql/mssql_db.py b/lib/ansible/modules/database/mssql/mssql_db.py index 018a68c3a6..f542b44a31 100644 --- a/lib/ansible/modules/database/mssql/mssql_db.py +++ b/lib/ansible/modules/database/mssql/mssql_db.py @@ -91,7 +91,7 @@ EXAMPLES = ''' target: /tmp/dump.sql ''' -RETURN = ''' +RETURN = ''' # ''' @@ -126,6 +126,7 @@ def db_delete(conn, cursor, db): cursor.execute("DROP DATABASE [%s]" % db) return not db_exists(conn, cursor, db) + def db_import(conn, cursor, module, db, target): if os.path.isfile(target): backup = open(target, 'r') diff --git a/lib/ansible/modules/database/postgresql/postgresql_ext.py b/lib/ansible/modules/database/postgresql/postgresql_ext.py index f0e83fb148..97bd549f21 100644 --- a/lib/ansible/modules/database/postgresql/postgresql_ext.py +++ b/lib/ansible/modules/database/postgresql/postgresql_ext.py @@ -99,6 +99,7 @@ def ext_exists(cursor, ext): cursor.execute(query, {'ext': ext}) return cursor.rowcount == 1 + def ext_delete(cursor, ext): if ext_exists(cursor, ext): query = "DROP EXTENSION \"%s\"" % ext @@ -107,6 +108,7 @@ def ext_delete(cursor, ext): else: return False + def ext_create(cursor, ext): if not ext_exists(cursor, ext): query = 'CREATE EXTENSION "%s"' % ext @@ -119,6 +121,7 @@ def ext_create(cursor, ext): # Module execution. # + def main(): module = AnsibleModule( argument_spec=dict( @@ -130,7 +133,7 @@ def main(): ext=dict(required=True, aliases=['name']), state=dict(default="present", choices=["absent", "present"]), ), - supports_check_mode = True + supports_check_mode=True ) if not postgresqldb_found: @@ -145,13 +148,13 @@ def main(): # check which values are empty and don't include in the **kw # dictionary params_map = { - "login_host":"host", - "login_user":"user", - "login_password":"password", - "port":"port" + "login_host": "host", + "login_user": "user", + "login_password": "password", + "port": "port" } - kw = dict( (params_map[k], v) for (k, v) in module.params.items() - if k in params_map and v != '' ) + kw = dict((params_map[k], v) for (k, v) in module.params.items() + if k in params_map and v != '') try: db_connection = psycopg2.connect(database=db, **kw) # Enable autocommit so we can create databases diff --git a/lib/ansible/modules/database/postgresql/postgresql_lang.py b/lib/ansible/modules/database/postgresql/postgresql_lang.py index 1db21f4033..8e5afb6a7a 100644 --- a/lib/ansible/modules/database/postgresql/postgresql_lang.py +++ b/lib/ansible/modules/database/postgresql/postgresql_lang.py @@ -161,18 +161,21 @@ def lang_exists(cursor, lang): cursor.execute(query) return cursor.rowcount > 0 + def lang_istrusted(cursor, lang): """Checks if language is trusted for db""" query = "SELECT lanpltrusted FROM pg_language WHERE lanname='%s'" % lang cursor.execute(query) return cursor.fetchone()[0] + def lang_altertrust(cursor, lang, trust): """Changes if language is trusted for db""" query = "UPDATE pg_language SET lanpltrusted = %s WHERE lanname=%s" cursor.execute(query, (trust, lang)) return True + def lang_add(cursor, lang, trust): """Adds language for db""" if trust: @@ -182,6 +185,7 @@ def lang_add(cursor, lang, trust): cursor.execute(query) return True + def lang_drop(cursor, lang, cascade): """Drops language for db""" cursor.execute("SAVEPOINT ansible_pgsql_lang_drop") @@ -197,6 +201,7 @@ def lang_drop(cursor, lang, cascade): cursor.execute("RELEASE SAVEPOINT ansible_pgsql_lang_drop") return True + def main(): module = AnsibleModule( argument_spec=dict( @@ -212,7 +217,7 @@ def main(): cascade=dict(type='bool', default='no'), fail_on_drop=dict(type='bool', default='yes'), ), - supports_check_mode = True + supports_check_mode=True ) db = module.params["db"] @@ -227,14 +232,14 @@ def main(): module.fail_json(msg="the python psycopg2 module is required") params_map = { - "login_host":"host", - "login_user":"user", - "login_password":"password", - "port":"port", - "db":"database" + "login_host": "host", + "login_user": "user", + "login_password": "password", + "port": "port", + "db": "database" } - kw = dict( (params_map[k], v) for (k, v) in module.params.items() - if k in params_map and v != "" ) + kw = dict((params_map[k], v) for (k, v) in module.params.items() + if k in params_map and v != "") try: db_connection = psycopg2.connect(**kw) cursor = db_connection.cursor() diff --git a/lib/ansible/modules/database/postgresql/postgresql_schema.py b/lib/ansible/modules/database/postgresql/postgresql_schema.py index 0f5673a9f9..d8edee6326 100644 --- a/lib/ansible/modules/database/postgresql/postgresql_schema.py +++ b/lib/ansible/modules/database/postgresql/postgresql_schema.py @@ -125,6 +125,7 @@ def set_owner(cursor, schema, owner): cursor.execute(query) return True + def get_schema_info(cursor, schema): query = """ SELECT schema_owner AS owner @@ -134,11 +135,13 @@ def get_schema_info(cursor, schema): cursor.execute(query, {'schema': schema}) return cursor.fetchone() + def schema_exists(cursor, schema): query = "SELECT schema_name FROM information_schema.schemata WHERE schema_name = %(schema)s" cursor.execute(query, {'schema': schema}) return cursor.rowcount == 1 + def schema_delete(cursor, schema): if schema_exists(cursor, schema): query = "DROP SCHEMA %s" % pg_quote_identifier(schema, 'schema') @@ -147,6 +150,7 @@ def schema_delete(cursor, schema): else: return False + def schema_create(cursor, schema, owner): if not schema_exists(cursor, schema): query_fragments = ['CREATE SCHEMA %s' % pg_quote_identifier(schema, 'schema')] @@ -162,6 +166,7 @@ def schema_create(cursor, schema, owner): else: return False + def schema_matches(cursor, schema, owner): if not schema_exists(cursor, schema): return False @@ -176,6 +181,7 @@ def schema_matches(cursor, schema, owner): # Module execution. # + def main(): module = AnsibleModule( argument_spec=dict( @@ -189,7 +195,7 @@ def main(): database=dict(default="postgres"), state=dict(default="present", choices=["absent", "present"]), ), - supports_check_mode = True + supports_check_mode=True ) if not postgresqldb_found: @@ -205,13 +211,13 @@ def main(): # check which values are empty and don't include in the **kw # dictionary params_map = { - "login_host":"host", - "login_user":"user", - "login_password":"password", - "port":"port" + "login_host": "host", + "login_user": "user", + "login_password": "password", + "port": "port" } - kw = dict( (params_map[k], v) for (k, v) in module.params.items() - if k in params_map and v != '' ) + kw = dict((params_map[k], v) for (k, v) in module.params.items() + if k in params_map and v != '') # If a login_unix_socket is specified, incorporate it here. is_localhost = "host" not in kw or kw["host"] == "" or kw["host"] == "localhost" diff --git a/lib/ansible/modules/database/vertica/vertica_configuration.py b/lib/ansible/modules/database/vertica/vertica_configuration.py index 670e068e24..b4734edf52 100644 --- a/lib/ansible/modules/database/vertica/vertica_configuration.py +++ b/lib/ansible/modules/database/vertica/vertica_configuration.py @@ -86,11 +86,13 @@ from ansible.module_utils._text import to_native class NotSupportedError(Exception): pass + class CannotDropError(Exception): pass # module specific functions + def get_configuration_facts(cursor, parameter_name=''): facts = {} cursor.execute(""" @@ -110,12 +112,14 @@ def get_configuration_facts(cursor, parameter_name=''): 'default_value': row.default_value} return facts + def check(configuration_facts, parameter_name, current_value): parameter_key = parameter_name.lower() if current_value and current_value.lower() != configuration_facts[parameter_key]['current_value'].lower(): return False return True + def present(configuration_facts, cursor, parameter_name, current_value): parameter_key = parameter_name.lower() changed = False @@ -128,6 +132,7 @@ def present(configuration_facts, cursor, parameter_name, current_value): # module logic + def main(): module = AnsibleModule( @@ -139,7 +144,7 @@ def main(): port=dict(default='5433'), login_user=dict(default='dbadmin'), login_password=dict(default=None, no_log=True), - ), supports_check_mode = True) + ), supports_check_mode=True) if not pyodbc_found: module.fail_json(msg="The python pyodbc module is required.") @@ -161,8 +166,8 @@ def main(): "User={3};" "Password={4};" "ConnectionLoadBalance={5}" - ).format(module.params['cluster'], module.params['port'], db, - module.params['login_user'], module.params['login_password'], 'true') + ).format(module.params['cluster'], module.params['port'], db, + module.params['login_user'], module.params['login_password'], 'true') db_conn = pyodbc.connect(dsn, autocommit=True) cursor = db_conn.cursor() except Exception as e: diff --git a/lib/ansible/modules/database/vertica/vertica_facts.py b/lib/ansible/modules/database/vertica/vertica_facts.py index d3b72b6c7e..c471b9413e 100644 --- a/lib/ansible/modules/database/vertica/vertica_facts.py +++ b/lib/ansible/modules/database/vertica/vertica_facts.py @@ -81,6 +81,7 @@ class NotSupportedError(Exception): # module specific functions + def get_schema_facts(cursor, schema=''): facts = {} cursor.execute(""" @@ -121,6 +122,7 @@ def get_schema_facts(cursor, schema=''): facts[schema_key]['usage_roles'].append(row.role_name) return facts + def get_user_facts(cursor, user=''): facts = {} cursor.execute(""" @@ -155,6 +157,7 @@ def get_user_facts(cursor, user=''): facts[user_key]['default_roles'] = row.default_roles.replace(' ', '').split(',') return facts + def get_role_facts(cursor, role=''): facts = {} cursor.execute(""" @@ -175,6 +178,7 @@ def get_role_facts(cursor, role=''): facts[role_key]['assigned_roles'] = row.assigned_roles.replace(' ', '').split(',') return facts + def get_configuration_facts(cursor, parameter=''): facts = {} cursor.execute(""" @@ -194,6 +198,7 @@ def get_configuration_facts(cursor, parameter=''): 'default_value': row.default_value} return facts + def get_node_facts(cursor, schema=''): facts = {} cursor.execute(""" @@ -216,6 +221,7 @@ def get_node_facts(cursor, schema=''): # module logic + def main(): module = AnsibleModule( @@ -225,7 +231,7 @@ def main(): db=dict(default=None), login_user=dict(default='dbadmin'), login_password=dict(default=None, no_log=True), - ), supports_check_mode = True) + ), supports_check_mode=True) if not pyodbc_found: module.fail_json(msg="The python pyodbc module is required.") @@ -243,8 +249,8 @@ def main(): "User=%s;" "Password=%s;" "ConnectionLoadBalance=%s" - ) % (module.params['cluster'], module.params['port'], db, - module.params['login_user'], module.params['login_password'], 'true') + ) % (module.params['cluster'], module.params['port'], db, + module.params['login_user'], module.params['login_password'], 'true') db_conn = pyodbc.connect(dsn, autocommit=True) cursor = db_conn.cursor() except Exception as e: @@ -257,11 +263,11 @@ def main(): configuration_facts = get_configuration_facts(cursor) node_facts = get_node_facts(cursor) module.exit_json(changed=False, - ansible_facts={'vertica_schemas': schema_facts, - 'vertica_users': user_facts, - 'vertica_roles': role_facts, - 'vertica_configuration': configuration_facts, - 'vertica_nodes': node_facts}) + ansible_facts={'vertica_schemas': schema_facts, + 'vertica_users': user_facts, + 'vertica_roles': role_facts, + 'vertica_configuration': configuration_facts, + 'vertica_nodes': node_facts}) except NotSupportedError as e: module.fail_json(msg=to_native(e), exception=traceback.format_exc()) except SystemExit: diff --git a/lib/ansible/modules/database/vertica/vertica_role.py b/lib/ansible/modules/database/vertica/vertica_role.py index 33dbb7ad3c..2385b9f44e 100644 --- a/lib/ansible/modules/database/vertica/vertica_role.py +++ b/lib/ansible/modules/database/vertica/vertica_role.py @@ -98,11 +98,13 @@ from ansible.module_utils._text import to_native class NotSupportedError(Exception): pass + class CannotDropError(Exception): pass # module specific functions + def get_role_facts(cursor, role=''): facts = {} cursor.execute(""" @@ -123,6 +125,7 @@ def get_role_facts(cursor, role=''): facts[role_key]['assigned_roles'] = row.assigned_roles.replace(' ', '').split(',') return facts + def update_roles(role_facts, cursor, role, existing, required): for assigned_role in set(existing) - set(required): @@ -130,6 +133,7 @@ def update_roles(role_facts, cursor, role, for assigned_role in set(required) - set(existing): cursor.execute("grant {0} to {1}".format(assigned_role, role)) + def check(role_facts, role, assigned_roles): role_key = role.lower() if role_key not in role_facts: @@ -138,6 +142,7 @@ def check(role_facts, role, assigned_roles): return False return True + def present(role_facts, cursor, role, assigned_roles): role_key = role.lower() if role_key not in role_facts: @@ -147,19 +152,20 @@ def present(role_facts, cursor, role, assigned_roles): return True else: changed = False - if assigned_roles and (sorted(assigned_roles) != sorted(role_facts[role_key]['assigned_roles'])): + if assigned_roles and (sorted(assigned_roles) != sorted(role_facts[role_key]['assigned_roles'])): update_roles(role_facts, cursor, role, - role_facts[role_key]['assigned_roles'], assigned_roles) + role_facts[role_key]['assigned_roles'], assigned_roles) changed = True if changed: role_facts.update(get_role_facts(cursor, role)) return changed + def absent(role_facts, cursor, role, assigned_roles): role_key = role.lower() if role_key in role_facts: update_roles(role_facts, cursor, role, - role_facts[role_key]['assigned_roles'], []) + role_facts[role_key]['assigned_roles'], []) cursor.execute("drop role {0} cascade".format(role_facts[role_key]['name'])) del role_facts[role_key] return True @@ -168,6 +174,7 @@ def absent(role_facts, cursor, role, assigned_roles): # module logic + def main(): module = AnsibleModule( @@ -180,7 +187,7 @@ def main(): port=dict(default='5433'), login_user=dict(default='dbadmin'), login_password=dict(default=None, no_log=True), - ), supports_check_mode = True) + ), supports_check_mode=True) if not pyodbc_found: module.fail_json(msg="The python pyodbc module is required.") @@ -206,8 +213,8 @@ def main(): "User={3};" "Password={4};" "ConnectionLoadBalance={5}" - ).format(module.params['cluster'], module.params['port'], db, - module.params['login_user'], module.params['login_password'], 'true') + ).format(module.params['cluster'], module.params['port'], db, + module.params['login_user'], module.params['login_password'], 'true') db_conn = pyodbc.connect(dsn, autocommit=True) cursor = db_conn.cursor() except Exception as e: diff --git a/lib/ansible/modules/database/vertica/vertica_schema.py b/lib/ansible/modules/database/vertica/vertica_schema.py index b92b8873de..713435f635 100644 --- a/lib/ansible/modules/database/vertica/vertica_schema.py +++ b/lib/ansible/modules/database/vertica/vertica_schema.py @@ -122,11 +122,13 @@ from ansible.module_utils._text import to_native class NotSupportedError(Exception): pass + class CannotDropError(Exception): pass # module specific functions + def get_schema_facts(cursor, schema=''): facts = {} cursor.execute(""" @@ -167,6 +169,7 @@ def get_schema_facts(cursor, schema=''): facts[schema_key]['usage_roles'].append(row.role_name) return facts + def update_roles(schema_facts, cursor, schema, existing, required, create_existing, create_required): @@ -180,6 +183,7 @@ def update_roles(schema_facts, cursor, schema, for role in set(create_required) - set(create_existing): cursor.execute("grant create on schema {0} to {1}".format(schema, role)) + def check(schema_facts, schema, usage_roles, create_roles, owner): schema_key = schema.lower() if schema_key not in schema_facts: @@ -192,6 +196,7 @@ def check(schema_facts, schema, usage_roles, create_roles, owner): return False return True + def present(schema_facts, cursor, schema, usage_roles, create_roles, owner): schema_key = schema.lower() if schema_key not in schema_facts: @@ -208,23 +213,24 @@ def present(schema_facts, cursor, schema, usage_roles, create_roles, owner): raise NotSupportedError(( "Changing schema owner is not supported. " "Current owner: {0}." - ).format(schema_facts[schema_key]['owner'])) + ).format(schema_facts[schema_key]['owner'])) if sorted(usage_roles) != sorted(schema_facts[schema_key]['usage_roles']) or \ sorted(create_roles) != sorted(schema_facts[schema_key]['create_roles']): update_roles(schema_facts, cursor, schema, - schema_facts[schema_key]['usage_roles'], usage_roles, - schema_facts[schema_key]['create_roles'], create_roles) + schema_facts[schema_key]['usage_roles'], usage_roles, + schema_facts[schema_key]['create_roles'], create_roles) changed = True if changed: schema_facts.update(get_schema_facts(cursor, schema)) return changed + def absent(schema_facts, cursor, schema, usage_roles, create_roles): schema_key = schema.lower() if schema_key in schema_facts: update_roles(schema_facts, cursor, schema, - schema_facts[schema_key]['usage_roles'], [], schema_facts[schema_key]['create_roles'], []) + schema_facts[schema_key]['usage_roles'], [], schema_facts[schema_key]['create_roles'], []) try: cursor.execute("drop schema {0} restrict".format(schema_facts[schema_key]['name'])) except pyodbc.Error: @@ -236,6 +242,7 @@ def absent(schema_facts, cursor, schema, usage_roles, create_roles): # module logic + def main(): module = AnsibleModule( @@ -250,7 +257,7 @@ def main(): port=dict(default='5433'), login_user=dict(default='dbadmin'), login_password=dict(default=None, no_log=True), - ), supports_check_mode = True) + ), supports_check_mode=True) if not pyodbc_found: module.fail_json(msg="The python pyodbc module is required.") @@ -281,8 +288,8 @@ def main(): "User={3};" "Password={4};" "ConnectionLoadBalance={5}" - ).format(module.params['cluster'], module.params['port'], db, - module.params['login_user'], module.params['login_password'], 'true') + ).format(module.params['cluster'], module.params['port'], db, + module.params['login_user'], module.params['login_password'], 'true') db_conn = pyodbc.connect(dsn, autocommit=True) cursor = db_conn.cursor() except Exception as e: diff --git a/lib/ansible/modules/database/vertica/vertica_user.py b/lib/ansible/modules/database/vertica/vertica_user.py index 90cf624caf..d15f1a8e59 100644 --- a/lib/ansible/modules/database/vertica/vertica_user.py +++ b/lib/ansible/modules/database/vertica/vertica_user.py @@ -134,11 +134,13 @@ from ansible.module_utils._text import to_native class NotSupportedError(Exception): pass + class CannotDropError(Exception): pass # module specific functions + def get_user_facts(cursor, user=''): facts = {} cursor.execute(""" @@ -173,6 +175,7 @@ def get_user_facts(cursor, user=''): facts[user_key]['default_roles'] = row.default_roles.replace(' ', '').split(',') return facts + def update_roles(user_facts, cursor, user, existing_all, existing_default, required): del_roles = list(set(existing_all) - set(required)) @@ -184,6 +187,7 @@ def update_roles(user_facts, cursor, user, if required: cursor.execute("alter user {0} default role {1}".format(user, ','.join(required))) + def check(user_facts, user, profile, resource_pool, locked, password, expired, ldap, roles): user_key = user.lower() @@ -198,13 +202,14 @@ def check(user_facts, user, profile, resource_pool, if password and password != user_facts[user_key]['password']: return False if (expired is not None and expired != (user_facts[user_key]['expired'] == 'True') or - ldap is not None and ldap != (user_facts[user_key]['expired'] == 'True')): + ldap is not None and ldap != (user_facts[user_key]['expired'] == 'True')): return False - if roles and (sorted(roles) != sorted(user_facts[user_key]['roles']) or \ - sorted(roles) != sorted(user_facts[user_key]['default_roles'])): + if roles and (sorted(roles) != sorted(user_facts[user_key]['roles']) or + sorted(roles) != sorted(user_facts[user_key]['default_roles'])): return False return True + def present(user_facts, cursor, user, profile, resource_pool, locked, password, expired, ldap, roles): user_key = user.lower() @@ -267,20 +272,21 @@ def present(user_facts, cursor, user, profile, resource_pool, changed = True if changed: cursor.execute(' '.join(query_fragments)) - if roles and (sorted(roles) != sorted(user_facts[user_key]['roles']) or \ - sorted(roles) != sorted(user_facts[user_key]['default_roles'])): + if roles and (sorted(roles) != sorted(user_facts[user_key]['roles']) or + sorted(roles) != sorted(user_facts[user_key]['default_roles'])): update_roles(user_facts, cursor, user, - user_facts[user_key]['roles'], user_facts[user_key]['default_roles'], roles) + user_facts[user_key]['roles'], user_facts[user_key]['default_roles'], roles) changed = True if changed: user_facts.update(get_user_facts(cursor, user)) return changed + def absent(user_facts, cursor, user, roles): user_key = user.lower() if user_key in user_facts: update_roles(user_facts, cursor, user, - user_facts[user_key]['roles'], user_facts[user_key]['default_roles'], []) + user_facts[user_key]['roles'], user_facts[user_key]['default_roles'], []) try: cursor.execute("drop user {0}".format(user_facts[user_key]['name'])) except pyodbc.Error: @@ -292,6 +298,7 @@ def absent(user_facts, cursor, user, roles): # module logic + def main(): module = AnsibleModule( @@ -309,7 +316,7 @@ def main(): port=dict(default='5433'), login_user=dict(default='dbadmin'), login_password=dict(default=None, no_log=True), - ), supports_check_mode = True) + ), supports_check_mode=True) if not pyodbc_found: module.fail_json(msg="The python pyodbc module is required.") @@ -348,8 +355,8 @@ def main(): "User={3};" "Password={4};" "ConnectionLoadBalance={5}" - ).format(module.params['cluster'], module.params['port'], db, - module.params['login_user'], module.params['login_password'], 'true') + ).format(module.params['cluster'], module.params['port'], db, + module.params['login_user'], module.params['login_password'], 'true') db_conn = pyodbc.connect(dsn, autocommit=True) cursor = db_conn.cursor() except Exception as e: @@ -359,7 +366,7 @@ def main(): user_facts = get_user_facts(cursor) if module.check_mode: changed = not check(user_facts, user, profile, resource_pool, - locked, password, expired, ldap, roles) + locked, password, expired, ldap, roles) elif state == 'absent': try: changed = absent(user_facts, cursor, user, roles) @@ -368,7 +375,7 @@ def main(): elif state in ['present', 'locked']: try: changed = present(user_facts, cursor, user, profile, resource_pool, - locked, password, expired, ldap, roles) + locked, password, expired, ldap, roles) except pyodbc.Error as e: module.fail_json(msg=to_native(e), exception=traceback.format_exc()) except NotSupportedError as e: diff --git a/lib/ansible/modules/monitoring/bigpanda.py b/lib/ansible/modules/monitoring/bigpanda.py index 44108608ee..ddd0776b9c 100644 --- a/lib/ansible/modules/monitoring/bigpanda.py +++ b/lib/ansible/modules/monitoring/bigpanda.py @@ -185,7 +185,7 @@ def main(): # Send the data to bigpanda data = json.dumps(body) - headers = {'Authorization':'Bearer %s' % token, 'Content-Type':'application/json'} + headers = {'Authorization': 'Bearer %s' % token, 'Content-Type': 'application/json'} try: response, info = fetch_url(module, request_url, data=data, headers=headers) if info['status'] == 200: diff --git a/lib/ansible/modules/monitoring/datadog_event.py b/lib/ansible/modules/monitoring/datadog_event.py index 8ac6d96e35..5036a647a9 100644 --- a/lib/ansible/modules/monitoring/datadog_event.py +++ b/lib/ansible/modules/monitoring/datadog_event.py @@ -135,7 +135,7 @@ def main(): choices=['error', 'warning', 'info', 'success'] ), aggregation_key=dict(required=False, default=None), - validate_certs = dict(default='yes', type='bool'), + validate_certs=dict(default='yes', type='bool'), ) ) diff --git a/lib/ansible/modules/monitoring/icinga2_feature.py b/lib/ansible/modules/monitoring/icinga2_feature.py index f4e88ec19d..c85af1b108 100644 --- a/lib/ansible/modules/monitoring/icinga2_feature.py +++ b/lib/ansible/modules/monitoring/icinga2_feature.py @@ -103,7 +103,7 @@ class Icinga2FeatureHelper: change_applied = True # RC is not 0 for this already disabled feature, handle it as no change applied elif re.search("Cannot disable feature '%s'. Target file .* does not exist" - % self.module.params["name"]): + % self.module.params["name"]): change_applied = False else: self.module.fail_json(msg="Fail to disable feature. Command returns %s" % out) diff --git a/lib/ansible/modules/monitoring/librato_annotation.py b/lib/ansible/modules/monitoring/librato_annotation.py index 416dc8f6a1..1f1969a12c 100644 --- a/lib/ansible/modules/monitoring/librato_annotation.py +++ b/lib/ansible/modules/monitoring/librato_annotation.py @@ -135,21 +135,22 @@ def post_annotation(module): response = response.read() module.exit_json(changed=True, annotation=response) + def main(): module = AnsibleModule( - argument_spec = dict( - user = dict(required=True), - api_key = dict(required=True), - name = dict(required=False), - title = dict(required=True), - source = dict(required=False), - description = dict(required=False), - start_time = dict(required=False, default=None, type='int'), - end_time = dict(require=False, default=None, type='int'), - links = dict(type='list') - ) + argument_spec=dict( + user=dict(required=True), + api_key=dict(required=True), + name=dict(required=False), + title=dict(required=True), + source=dict(required=False), + description=dict(required=False), + start_time=dict(required=False, default=None, type='int'), + end_time=dict(require=False, default=None, type='int'), + links=dict(type='list') ) + ) post_annotation(module) diff --git a/lib/ansible/modules/monitoring/logentries.py b/lib/ansible/modules/monitoring/logentries.py index bd8f87829f..625c3065fd 100644 --- a/lib/ansible/modules/monitoring/logentries.py +++ b/lib/ansible/modules/monitoring/logentries.py @@ -85,9 +85,9 @@ def follow_log(module, le_path, logs, name=None, logtype=None): cmd = [le_path, 'follow', log] if name: - cmd.extend(['--name',name]) + cmd.extend(['--name', name]) if logtype: - cmd.extend(['--type',logtype]) + cmd.extend(['--type', logtype]) rc, out, err = module.run_command(' '.join(cmd)) if not query_log_status(module, le_path, log): @@ -100,6 +100,7 @@ def follow_log(module, le_path, logs, name=None, logtype=None): module.exit_json(changed=False, msg="logs(s) already followed") + def unfollow_log(module, le_path, logs): """ Unfollows one or more logs if followed. """ @@ -125,13 +126,14 @@ def unfollow_log(module, le_path, logs): module.exit_json(changed=False, msg="logs(s) already unfollowed") + def main(): module = AnsibleModule( - argument_spec = dict( - path = dict(required=True), - state = dict(default="present", choices=["present", "followed", "absent", "unfollowed"]), - name = dict(required=False, default=None, type='str'), - logtype = dict(required=False, default=None, type='str', aliases=['type']) + argument_spec=dict( + path=dict(required=True), + state=dict(default="present", choices=["present", "followed", "absent", "unfollowed"]), + name=dict(required=False, default=None, type='str'), + logtype=dict(required=False, default=None, type='str', aliases=['type']) ), supports_check_mode=True ) diff --git a/lib/ansible/modules/monitoring/logicmonitor.py b/lib/ansible/modules/monitoring/logicmonitor.py index d6885b4baf..21cc2c32a2 100644 --- a/lib/ansible/modules/monitoring/logicmonitor.py +++ b/lib/ansible/modules/monitoring/logicmonitor.py @@ -549,7 +549,6 @@ except ImportError: HAS_LIB_JSON = False - class LogicMonitor(object): def __init__(self, module, **params): @@ -667,7 +666,7 @@ class LogicMonitor(object): for host in hosts: if (host["hostName"] == hostname and - host["agentId"] == collector["id"]): + host["agentId"] == collector["id"]): self.module.debug("Host match found") return host @@ -688,7 +687,7 @@ class LogicMonitor(object): self.module.debug("Looking for displayname " + displayname) self.module.debug("Making RPC call to 'getHost'") host_json = (json.loads(self.rpc("getHost", - {"displayName": displayname}))) + {"displayName": displayname}))) if host_json["status"] == 200: self.module.debug("RPC call succeeded") @@ -1028,18 +1027,18 @@ class Collector(LogicMonitor): else: self.fail(msg="Error: Unable to retrieve timezone offset") - offsetend = offsetstart + datetime.timedelta(0, int(duration)*60) + offsetend = offsetstart + datetime.timedelta(0, int(duration) * 60) h = {"agentId": self.id, "type": 1, "notifyCC": True, "year": offsetstart.year, - "month": offsetstart.month-1, + "month": offsetstart.month - 1, "day": offsetstart.day, "hour": offsetstart.hour, "minute": offsetstart.minute, "endYear": offsetend.year, - "endMonth": offsetend.month-1, + "endMonth": offsetend.month - 1, "endDay": offsetend.day, "endHour": offsetend.hour, "endMinute": offsetend.minute} @@ -1187,7 +1186,7 @@ class Host(LogicMonitor): # Used the host information to grab the collector description # if not provided if (not hasattr(self.params, "collector") and - "agentDescription" in info): + "agentDescription" in info): self.module.debug("Setting collector from host response. " + "Collector " + info["agentDescription"]) self.params["collector"] = info["agentDescription"] @@ -1238,8 +1237,8 @@ class Host(LogicMonitor): if self.info: self.module.debug("Making RPC call to 'getHostProperties'") properties_json = (json.loads(self.rpc("getHostProperties", - {'hostId': self.info["id"], - "filterSystemProperties": True}))) + {'hostId': self.info["id"], + "filterSystemProperties": True}))) if properties_json["status"] == 200: self.module.debug("RPC call succeeded") @@ -1411,8 +1410,8 @@ class Host(LogicMonitor): return True if (self.collector and - hasattr(self.collector, "id") and - hostresp["agentId"] != self.collector["id"]): + hasattr(self.collector, "id") and + hostresp["agentId"] != self.collector["id"]): return True self.module.debug("Comparing groups.") @@ -1469,7 +1468,7 @@ class Host(LogicMonitor): self.fail( msg="Error: Unable to retrieve timezone offset") - offsetend = offsetstart + datetime.timedelta(0, int(duration)*60) + offsetend = offsetstart + datetime.timedelta(0, int(duration) * 60) h = {"hostId": self.info["id"], "type": 1, @@ -1599,7 +1598,7 @@ class Host(LogicMonitor): hgresp = json.loads(self.rpc("getHostGroup", h)) if (hgresp["status"] == 200 and - hgresp["data"]["appliesTo"] == ""): + hgresp["data"]["appliesTo"] == ""): g.append(path[-1]) @@ -1632,7 +1631,7 @@ class Host(LogicMonitor): for prop in propresp: if prop["name"] not in ignore: if ("*******" in prop["value"] and - self._verify_property(prop["name"])): + self._verify_property(prop["name"])): p[prop["name"]] = self.properties[prop["name"]] else: p[prop["name"]] = prop["value"] @@ -1641,7 +1640,7 @@ class Host(LogicMonitor): # Iterate provided properties and compare to received properties for prop in self.properties: if (prop not in p or - p[prop] != self.properties[prop]): + p[prop] != self.properties[prop]): self.module.debug("Properties mismatch") return True self.module.debug("Properties match") @@ -1703,18 +1702,18 @@ class Datasource(LogicMonitor): else: self.fail(msg="Error: Unable to retrieve timezone offset") - offsetend = offsetstart + datetime.timedelta(0, int(duration)*60) + offsetend = offsetstart + datetime.timedelta(0, int(duration) * 60) h = {"hostDataSourceId": self.id, "type": 1, "notifyCC": True, "year": offsetstart.year, - "month": offsetstart.month-1, + "month": offsetstart.month - 1, "day": offsetstart.day, "hour": offsetstart.hour, "minute": offsetstart.minute, "endYear": offsetend.year, - "endMonth": offsetend.month-1, + "endMonth": offsetend.month - 1, "endDay": offsetend.day, "endHour": offsetend.hour, "endMinute": offsetend.minute} @@ -1905,7 +1904,7 @@ class Hostgroup(LogicMonitor): if properties is not None and group is not None: self.module.debug("Comparing simple group properties") if (group["alertEnable"] != self.alertenable or - group["description"] != self.description): + group["description"] != self.description): return True @@ -1915,7 +1914,7 @@ class Hostgroup(LogicMonitor): for prop in properties: if prop["name"] not in ignore: if ("*******" in prop["value"] and - self._verify_property(prop["name"])): + self._verify_property(prop["name"])): p[prop["name"]] = ( self.properties[prop["name"]]) @@ -1965,17 +1964,17 @@ class Hostgroup(LogicMonitor): self.fail( msg="Error: Unable to retrieve timezone offset") - offsetend = offsetstart + datetime.timedelta(0, int(duration)*60) + offsetend = offsetstart + datetime.timedelta(0, int(duration) * 60) h = {"hostGroupId": self.info["id"], "type": 1, "year": offsetstart.year, - "month": offsetstart.month-1, + "month": offsetstart.month - 1, "day": offsetstart.day, "hour": offsetstart.hour, "minute": offsetstart.minute, "endYear": offsetend.year, - "endMonth": offsetend.month-1, + "endMonth": offsetend.month - 1, "endDay": offsetend.day, "endHour": offsetend.hour, "endMinute": offsetend.minute} @@ -2086,8 +2085,8 @@ def selector(module): elif module.params["target"] == "host": # Make sure required parameter collector is specified if ((module.params["action"] == "add" or - module.params["displayname"] is None) and - module.params["collector"] is None): + module.params["displayname"] is None) and + module.params["collector"] is None): module.fail_json( msg="Parameter 'collector' required.") diff --git a/lib/ansible/modules/monitoring/logicmonitor_facts.py b/lib/ansible/modules/monitoring/logicmonitor_facts.py index eda22654d1..0b98cf2705 100644 --- a/lib/ansible/modules/monitoring/logicmonitor_facts.py +++ b/lib/ansible/modules/monitoring/logicmonitor_facts.py @@ -242,7 +242,7 @@ class LogicMonitor(object): for host in hosts: if (host["hostName"] == hostname and - host["agentId"] == collector["id"]): + host["agentId"] == collector["id"]): self.module.debug("Host match found") return host @@ -263,7 +263,7 @@ class LogicMonitor(object): self.module.debug("Looking for displayname " + displayname) self.module.debug("Making RPC call to 'getHost'") host_json = (json.loads(self.rpc("getHost", - {"displayName": displayname}))) + {"displayName": displayname}))) if host_json["status"] == 200: self.module.debug("RPC call succeeded") @@ -431,7 +431,7 @@ class Host(LogicMonitor): # Used the host information to grab the collector description # if not provided if (not hasattr(self.params, "collector") and - "agentDescription" in info): + "agentDescription" in info): self.module.debug("Setting collector from host response. " + "Collector " + info["agentDescription"]) self.params["collector"] = info["agentDescription"] @@ -464,8 +464,8 @@ class Host(LogicMonitor): if self.info: self.module.debug("Making RPC call to 'getHostProperties'") properties_json = (json.loads(self.rpc("getHostProperties", - {'hostId': self.info["id"], - "filterSystemProperties": True}))) + {'hostId': self.info["id"], + "filterSystemProperties": True}))) if properties_json["status"] == 200: self.module.debug("RPC call succeeded") diff --git a/lib/ansible/modules/monitoring/nagios.py b/lib/ansible/modules/monitoring/nagios.py index b851a8c362..1cf7480418 100644 --- a/lib/ansible/modules/monitoring/nagios.py +++ b/lib/ansible/modules/monitoring/nagios.py @@ -231,7 +231,7 @@ def which_cmdfile(): '/etc/icinga/icinga.cfg', # icinga installed from source (default location) '/usr/local/icinga/etc/icinga.cfg', - ] + ] for path in locations: if os.path.exists(path): @@ -257,8 +257,7 @@ def main(): 'command', 'servicegroup_host_downtime', 'servicegroup_service_downtime', - ] - + ] module = AnsibleModule( argument_spec=dict( @@ -271,8 +270,8 @@ def main(): cmdfile=dict(default=which_cmdfile()), services=dict(default=None, aliases=['service']), command=dict(required=False, default=None), - ) ) + ) action = module.params['action'] host = module.params['host'] @@ -595,7 +594,6 @@ class Nagios(object): dt_del_cmd_str = self._fmt_dt_del_str(cmd, host, svc=service, comment=comment) self._write_command(dt_del_cmd_str) - def schedule_hostgroup_host_downtime(self, hostgroup, minutes=30): """ This command is used to schedule downtime for all hosts in a @@ -934,7 +932,7 @@ class Nagios(object): cmd = [ "DISABLE_HOST_SVC_NOTIFICATIONS", "DISABLE_HOST_NOTIFICATIONS" - ] + ] nagios_return = True return_str_list = [] for c in cmd: @@ -962,7 +960,7 @@ class Nagios(object): cmd = [ "ENABLE_HOST_SVC_NOTIFICATIONS", "ENABLE_HOST_NOTIFICATIONS" - ] + ] nagios_return = True return_str_list = [] for c in cmd: @@ -1027,19 +1025,19 @@ class Nagios(object): minutes=self.minutes) elif self.action == 'delete_downtime': - if self.services=='host': + if self.services == 'host': self.delete_host_downtime(self.host) - elif self.services=='all': + elif self.services == 'all': self.delete_host_downtime(self.host, comment='') else: self.delete_host_downtime(self.host, services=self.services) elif self.action == "servicegroup_host_downtime": if self.servicegroup: - self.schedule_servicegroup_host_downtime(servicegroup = self.servicegroup, minutes = self.minutes) + self.schedule_servicegroup_host_downtime(servicegroup=self.servicegroup, minutes=self.minutes) elif self.action == "servicegroup_service_downtime": if self.servicegroup: - self.schedule_servicegroup_svc_downtime(servicegroup = self.servicegroup, minutes = self.minutes) + self.schedule_servicegroup_svc_downtime(servicegroup=self.servicegroup, minutes=self.minutes) # toggle the host AND service alerts elif self.action == 'silence': @@ -1077,7 +1075,7 @@ class Nagios(object): # wtf? else: - self.module.fail_json(msg="unknown action specified: '%s'" % \ + self.module.fail_json(msg="unknown action specified: '%s'" % self.action) self.module.exit_json(nagios_commands=self.command_results, diff --git a/lib/ansible/modules/monitoring/newrelic_deployment.py b/lib/ansible/modules/monitoring/newrelic_deployment.py index 24528100df..92b0e4e42a 100644 --- a/lib/ansible/modules/monitoring/newrelic_deployment.py +++ b/lib/ansible/modules/monitoring/newrelic_deployment.py @@ -86,6 +86,7 @@ from ansible.module_utils.six.moves.urllib.parse import urlencode # Module execution. # + def main(): module = AnsibleModule( @@ -99,7 +100,7 @@ def main(): user=dict(required=False), appname=dict(required=False), environment=dict(required=False), - validate_certs = dict(default='yes', type='bool'), + validate_certs=dict(default='yes', type='bool'), ), required_one_of=[['app_name', 'application_id']], supports_check_mode=True @@ -117,7 +118,7 @@ def main(): else: module.fail_json(msg="you must set one of 'app_name' or 'application_id'") - for item in [ "changelog", "description", "revision", "user", "appname", "environment" ]: + for item in ["changelog", "description", "revision", "user", "appname", "environment"]: if module.params[item]: params[item] = module.params[item] diff --git a/lib/ansible/modules/monitoring/pagerduty.py b/lib/ansible/modules/monitoring/pagerduty.py index 7bf0e6a783..7d9f57af03 100644 --- a/lib/ansible/modules/monitoring/pagerduty.py +++ b/lib/ansible/modules/monitoring/pagerduty.py @@ -112,7 +112,7 @@ options: version_added: 1.5.1 ''' -EXAMPLES=''' +EXAMPLES = ''' # List ongoing maintenance windows using a user/passwd - pagerduty: name: companyabc @@ -171,6 +171,7 @@ import json from ansible.module_utils.basic import AnsibleModule from ansible.module_utils.urls import fetch_url + def auth_header(user, passwd, token): if token: return "Token token=%s" % token @@ -178,6 +179,7 @@ def auth_header(user, passwd, token): auth = base64.encodestring('%s:%s' % (user, passwd)).replace('\n', '') return "Basic %s" % auth + def ongoing(module, name, user, passwd, token): url = "https://" + name + ".pagerduty.com/api/v1/maintenance_windows/ongoing" headers = {"Authorization": auth_header(user, passwd, token)} @@ -203,7 +205,7 @@ def create(module, name, user, passwd, token, requester_id, service, hours, minu url = "https://" + name + ".pagerduty.com/api/v1/maintenance_windows" headers = { 'Authorization': auth_header(user, passwd, token), - 'Content-Type' : 'application/json', + 'Content-Type': 'application/json', } request_data = {'maintenance_window': {'start_time': start, 'end_time': end, 'description': desc, 'service_ids': service}} @@ -225,11 +227,12 @@ def create(module, name, user, passwd, token, requester_id, service, hours, minu return False, json_out, True + def absent(module, name, user, passwd, token, requester_id, service): url = "https://" + name + ".pagerduty.com/api/v1/maintenance_windows/" + service[0] headers = { 'Authorization': auth_header(user, passwd, token), - 'Content-Type' : 'application/json', + 'Content-Type': 'application/json', } request_data = {} @@ -266,7 +269,7 @@ def main(): hours=dict(default='1', required=False), minutes=dict(default='0', required=False), desc=dict(default='Created by Ansible', required=False), - validate_certs = dict(default='yes', type='bool'), + validate_certs=dict(default='yes', type='bool'), ) ) @@ -280,7 +283,7 @@ def main(): minutes = module.params['minutes'] token = module.params['token'] desc = module.params['desc'] - requester_id = module.params['requester_id'] + requester_id = module.params['requester_id'] if not token and not (user or passwd): module.fail_json(msg="neither user and passwd nor token specified") @@ -290,7 +293,7 @@ def main(): module.fail_json(msg="service not specified") (rc, out, changed) = create(module, name, user, passwd, token, requester_id, service, hours, minutes, desc) if rc == 0: - changed=True + changed = True if state == "ongoing": (rc, out, changed) = ongoing(module, name, user, passwd, token) @@ -301,7 +304,6 @@ def main(): if rc != 0: module.fail_json(msg="failed", result=out) - module.exit_json(msg="success", result=out, changed=changed) diff --git a/lib/ansible/modules/monitoring/pingdom.py b/lib/ansible/modules/monitoring/pingdom.py index 68c7dcb137..3fdf696301 100644 --- a/lib/ansible/modules/monitoring/pingdom.py +++ b/lib/ansible/modules/monitoring/pingdom.py @@ -98,7 +98,7 @@ def pause(checkid, uid, passwd, key): check = c.get_check(checkid) name = check.name result = check.status - #if result != "paused": # api output buggy - accept raw exception for now + # if result != "paused": # api output buggy - accept raw exception for now # return (True, name, result) return (False, name, result) @@ -110,7 +110,7 @@ def unpause(checkid, uid, passwd, key): check = c.get_check(checkid) name = check.name result = check.status - #if result != "up": # api output buggy - accept raw exception for now + # if result != "up": # api output buggy - accept raw exception for now # return (True, name, result) return (False, name, result) diff --git a/lib/ansible/modules/monitoring/sensu_check.py b/lib/ansible/modules/monitoring/sensu_check.py index fdc24fb422..afa259f40f 100644 --- a/lib/ansible/modules/monitoring/sensu_check.py +++ b/lib/ansible/modules/monitoring/sensu_check.py @@ -281,12 +281,12 @@ def sensu_check(module, path, name, state='present', backup=False): if module.params['custom']: # Convert to json custom_params = module.params['custom'] - overwrited_fields = set(custom_params.keys()) & set(simple_opts + ['type','subdue','subdue_begin','subdue_end']) + overwrited_fields = set(custom_params.keys()) & set(simple_opts + ['type', 'subdue', 'subdue_begin', 'subdue_end']) if overwrited_fields: msg = 'You can\'t overwriting standard module parameters via "custom". You are trying overwrite: {opt}'.format(opt=list(overwrited_fields)) module.fail_json(msg=msg) - for k,v in custom_params.items(): + for k, v in custom_params.items(): if k in config['checks'][name]: if not config['checks'][name][k] == v: changed = True @@ -298,7 +298,7 @@ def sensu_check(module, path, name, state='present', backup=False): simple_opts += custom_params.keys() # Remove obsolete custom params - for opt in set(config['checks'][name].keys()) - set(simple_opts + ['type','subdue','subdue_begin','subdue_end']): + for opt in set(config['checks'][name].keys()) - set(simple_opts + ['type', 'subdue', 'subdue_begin', 'subdue_end']): changed = True reasons.append('`custom param {opt}\' was deleted'.format(opt=opt)) del check[opt] @@ -345,30 +345,30 @@ def sensu_check(module, path, name, state='present', backup=False): def main(): - arg_spec = {'name': {'type': 'str', 'required': True}, - 'path': {'type': 'str', 'default': '/etc/sensu/conf.d/checks.json'}, - 'state': {'type': 'str', 'default': 'present', 'choices': ['present', 'absent']}, - 'backup': {'type': 'bool', 'default': 'no'}, - 'command': {'type': 'str'}, - 'handlers': {'type': 'list'}, - 'subscribers': {'type': 'list'}, - 'interval': {'type': 'int'}, - 'timeout': {'type': 'int'}, - 'ttl': {'type': 'int'}, - 'handle': {'type': 'bool'}, + arg_spec = {'name': {'type': 'str', 'required': True}, + 'path': {'type': 'str', 'default': '/etc/sensu/conf.d/checks.json'}, + 'state': {'type': 'str', 'default': 'present', 'choices': ['present', 'absent']}, + 'backup': {'type': 'bool', 'default': 'no'}, + 'command': {'type': 'str'}, + 'handlers': {'type': 'list'}, + 'subscribers': {'type': 'list'}, + 'interval': {'type': 'int'}, + 'timeout': {'type': 'int'}, + 'ttl': {'type': 'int'}, + 'handle': {'type': 'bool'}, 'subdue_begin': {'type': 'str'}, - 'subdue_end': {'type': 'str'}, + 'subdue_end': {'type': 'str'}, 'dependencies': {'type': 'list'}, - 'metric': {'type': 'bool', 'default': 'no'}, - 'standalone': {'type': 'bool'}, - 'publish': {'type': 'bool'}, - 'occurrences': {'type': 'int'}, - 'refresh': {'type': 'int'}, - 'aggregate': {'type': 'bool'}, - 'low_flap_threshold': {'type': 'int'}, + 'metric': {'type': 'bool', 'default': 'no'}, + 'standalone': {'type': 'bool'}, + 'publish': {'type': 'bool'}, + 'occurrences': {'type': 'int'}, + 'refresh': {'type': 'int'}, + 'aggregate': {'type': 'bool'}, + 'low_flap_threshold': {'type': 'int'}, 'high_flap_threshold': {'type': 'int'}, - 'custom': {'type': 'dict'}, - 'source': {'type': 'str'}, + 'custom': {'type': 'dict'}, + 'source': {'type': 'str'}, } required_together = [['subdue_begin', 'subdue_end']] diff --git a/lib/ansible/modules/monitoring/sensu_subscription.py b/lib/ansible/modules/monitoring/sensu_subscription.py index 3e2ac652ae..8a6808a87b 100644 --- a/lib/ansible/modules/monitoring/sensu_subscription.py +++ b/lib/ansible/modules/monitoring/sensu_subscription.py @@ -135,9 +135,9 @@ def sensu_subscription(module, path, name, state='present', backup=False): def main(): - arg_spec = {'name': {'type': 'str', 'required': True}, - 'path': {'type': 'str', 'default': '/etc/sensu/conf.d/subscriptions.json'}, - 'state': {'type': 'str', 'default': 'present', 'choices': ['present', 'absent']}, + arg_spec = {'name': {'type': 'str', 'required': True}, + 'path': {'type': 'str', 'default': '/etc/sensu/conf.d/subscriptions.json'}, + 'state': {'type': 'str', 'default': 'present', 'choices': ['present', 'absent']}, 'backup': {'type': 'bool', 'default': 'no'}, } diff --git a/lib/ansible/modules/monitoring/stackdriver.py b/lib/ansible/modules/monitoring/stackdriver.py index e530fea2dd..80e8b5dd71 100644 --- a/lib/ansible/modules/monitoring/stackdriver.py +++ b/lib/ansible/modules/monitoring/stackdriver.py @@ -124,6 +124,7 @@ def send_deploy_event(module, key, revision_id, deployed_by='Ansible', deployed_ return do_send_request(module, deploy_api, params, key) + def send_annotation_event(module, key, msg, annotated_by='Ansible', level=None, instance_id=None, event_epoch=None): """Send an annotation event to Stackdriver""" annotation_api = "https://event-gateway.stackdriver.com/v1/annotationevent" @@ -141,13 +142,14 @@ def send_annotation_event(module, key, msg, annotated_by='Ansible', level=None, return do_send_request(module, annotation_api, params, key) + def do_send_request(module, url, params, key): data = json.dumps(params) headers = { 'Content-Type': 'application/json', 'x-stackdriver-apikey': key } - response, info = fetch_url(module, url, headers=headers, data=data, method='POST') + response, info = fetch_url(module, url, headers=headers, data=data, method='POST') if info['status'] != 200: module.fail_json(msg="Unable to send msg: %s" % info['msg']) diff --git a/lib/ansible/modules/monitoring/uptimerobot.py b/lib/ansible/modules/monitoring/uptimerobot.py index e0c87b0ceb..22426cfb76 100644 --- a/lib/ansible/modules/monitoring/uptimerobot.py +++ b/lib/ansible/modules/monitoring/uptimerobot.py @@ -120,10 +120,10 @@ def pauseMonitor(module, params): def main(): module = AnsibleModule( - argument_spec = dict( - state = dict(required=True, choices=['started', 'paused']), - apikey = dict(required=True), - monitorid = dict(required=True) + argument_spec=dict( + state=dict(required=True, choices=['started', 'paused']), + apikey=dict(required=True), + monitorid=dict(required=True) ), supports_check_mode=SUPPORTS_CHECK_MODE ) diff --git a/lib/ansible/modules/monitoring/zabbix/zabbix_proxy.py b/lib/ansible/modules/monitoring/zabbix/zabbix_proxy.py index ffd0726fdf..bb2d8ecc89 100644 --- a/lib/ansible/modules/monitoring/zabbix/zabbix_proxy.py +++ b/lib/ansible/modules/monitoring/zabbix/zabbix_proxy.py @@ -186,7 +186,7 @@ class Proxy(object): old_interface = {} if 'interface' in self.existing_data and \ len(self.existing_data['interface']) > 0: - old_interface = self.existing_data['interface'] + old_interface = self.existing_data['interface'] final_interface = old_interface.copy() final_interface.update(new_interface) @@ -206,7 +206,7 @@ class Proxy(object): for item in data: if data[item] and item in self.existing_data and \ self.existing_data[item] != data[item]: - parameters[item] = data[item] + parameters[item] = data[item] if 'interface' in parameters: parameters.pop('interface') diff --git a/lib/ansible/modules/net_tools/cloudflare_dns.py b/lib/ansible/modules/net_tools/cloudflare_dns.py index e2ee91c854..d26f01508d 100644 --- a/lib/ansible/modules/net_tools/cloudflare_dns.py +++ b/lib/ansible/modules/net_tools/cloudflare_dns.py @@ -278,35 +278,34 @@ from ansible.module_utils._text import to_native, to_text from ansible.module_utils.urls import fetch_url - class CloudflareAPI(object): cf_api_endpoint = 'https://api.cloudflare.com/client/v4' changed = False def __init__(self, module): - self.module = module + self.module = module self.account_api_token = module.params['account_api_token'] - self.account_email = module.params['account_email'] - self.port = module.params['port'] - self.priority = module.params['priority'] - self.proto = module.params['proto'] - self.proxied = module.params['proxied'] - self.record = module.params['record'] - self.service = module.params['service'] - self.is_solo = module.params['solo'] - self.state = module.params['state'] - self.timeout = module.params['timeout'] - self.ttl = module.params['ttl'] - self.type = module.params['type'] - self.value = module.params['value'] - self.weight = module.params['weight'] - self.zone = module.params['zone'] + self.account_email = module.params['account_email'] + self.port = module.params['port'] + self.priority = module.params['priority'] + self.proto = module.params['proto'] + self.proxied = module.params['proxied'] + self.record = module.params['record'] + self.service = module.params['service'] + self.is_solo = module.params['solo'] + self.state = module.params['state'] + self.timeout = module.params['timeout'] + self.ttl = module.params['ttl'] + self.type = module.params['type'] + self.value = module.params['value'] + self.weight = module.params['weight'] + self.zone = module.params['zone'] if self.record == '@': self.record = self.zone - if (self.type in ['CNAME','NS','MX','SRV']) and (self.value is not None): + if (self.type in ['CNAME', 'NS', 'MX', 'SRV']) and (self.value is not None): self.value = self.value.rstrip('.') if (self.type == 'SRV'): @@ -318,10 +317,10 @@ class CloudflareAPI(object): if not self.record.endswith(self.zone): self.record = self.record + '.' + self.zone - def _cf_simple_api_call(self,api_call,method='GET',payload=None): - headers = { 'X-Auth-Email': self.account_email, - 'X-Auth-Key': self.account_api_token, - 'Content-Type': 'application/json' } + def _cf_simple_api_call(self, api_call, method='GET', payload=None): + headers = {'X-Auth-Email': self.account_email, + 'X-Auth-Key': self.account_api_token, + 'Content-Type': 'application/json'} data = None if payload: try: @@ -336,28 +335,28 @@ class CloudflareAPI(object): method=method, timeout=self.timeout) - if info['status'] not in [200,304,400,401,403,429,405,415]: - self.module.fail_json(msg="Failed API call {0}; got unexpected HTTP code {1}".format(api_call,info['status'])) + if info['status'] not in [200, 304, 400, 401, 403, 429, 405, 415]: + self.module.fail_json(msg="Failed API call {0}; got unexpected HTTP code {1}".format(api_call, info['status'])) error_msg = '' if info['status'] == 401: # Unauthorized - error_msg = "API user does not have permission; Status: {0}; Method: {1}: Call: {2}".format(info['status'],method,api_call) + error_msg = "API user does not have permission; Status: {0}; Method: {1}: Call: {2}".format(info['status'], method, api_call) elif info['status'] == 403: # Forbidden - error_msg = "API request not authenticated; Status: {0}; Method: {1}: Call: {2}".format(info['status'],method,api_call) + error_msg = "API request not authenticated; Status: {0}; Method: {1}: Call: {2}".format(info['status'], method, api_call) elif info['status'] == 429: # Too many requests - error_msg = "API client is rate limited; Status: {0}; Method: {1}: Call: {2}".format(info['status'],method,api_call) + error_msg = "API client is rate limited; Status: {0}; Method: {1}: Call: {2}".format(info['status'], method, api_call) elif info['status'] == 405: # Method not allowed - error_msg = "API incorrect HTTP method provided; Status: {0}; Method: {1}: Call: {2}".format(info['status'],method,api_call) + error_msg = "API incorrect HTTP method provided; Status: {0}; Method: {1}: Call: {2}".format(info['status'], method, api_call) elif info['status'] == 415: # Unsupported Media Type - error_msg = "API request is not valid JSON; Status: {0}; Method: {1}: Call: {2}".format(info['status'],method,api_call) - elif info ['status'] == 400: + error_msg = "API request is not valid JSON; Status: {0}; Method: {1}: Call: {2}".format(info['status'], method, api_call) + elif info['status'] == 400: # Bad Request - error_msg = "API bad request; Status: {0}; Method: {1}: Call: {2}".format(info['status'],method,api_call) + error_msg = "API bad request; Status: {0}; Method: {1}: Call: {2}".format(info['status'], method, api_call) result = None try: @@ -375,22 +374,22 @@ class CloudflareAPI(object): error_msg += "; Failed to parse API response with error {0}: {1}".format(to_native(e), content) # received an error status but no data with details on what failed - if (info['status'] not in [200,304]) and (result is None): + if (info['status'] not in [200, 304]) and (result is None): self.module.fail_json(msg=error_msg) if not result['success']: error_msg += "; Error details: " for error in result['errors']: - error_msg += "code: {0}, error: {1}; ".format(error['code'],error['message']) + error_msg += "code: {0}, error: {1}; ".format(error['code'], error['message']) if 'error_chain' in error: for chain_error in error['error_chain']: - error_msg += "code: {0}, error: {1}; ".format(chain_error['code'],chain_error['message']) + error_msg += "code: {0}, error: {1}; ".format(chain_error['code'], chain_error['message']) self.module.fail_json(msg=error_msg) return result, info['status'] - def _cf_api_call(self,api_call,method='GET',payload=None): - result, status = self._cf_simple_api_call(api_call,method,payload) + def _cf_api_call(self, api_call, method='GET', payload=None): + result, status = self._cf_simple_api_call(api_call, method, payload) data = result['result'] @@ -401,19 +400,19 @@ class CloudflareAPI(object): parameters = ['page={0}'.format(next_page)] # strip "page" parameter from call parameters (if there are any) if '?' in api_call: - raw_api_call,query = api_call.split('?',1) + raw_api_call, query = api_call.split('?', 1) parameters += [param for param in query.split('&') if not param.startswith('page')] else: raw_api_call = api_call while next_page <= pagination['total_pages']: raw_api_call += '?' + '&'.join(parameters) - result, status = self._cf_simple_api_call(raw_api_call,method,payload) + result, status = self._cf_simple_api_call(raw_api_call, method, payload) data += result['result'] next_page += 1 return data, status - def _get_zone_id(self,zone=None): + def _get_zone_id(self, zone=None): if not zone: zone = self.zone @@ -426,16 +425,16 @@ class CloudflareAPI(object): return zones[0]['id'] - def get_zones(self,name=None): + def get_zones(self, name=None): if not name: name = self.zone param = '' if name: - param = '?' + urlencode({'name' : name}) - zones,status = self._cf_api_call('/zones' + param) + param = '?' + urlencode({'name': name}) + zones, status = self._cf_api_call('/zones' + param) return zones - def get_dns_records(self,zone_name=None,type=None,record=None,value=''): + def get_dns_records(self, zone_name=None, type=None, record=None, value=''): if not zone_name: zone_name = self.zone if not type: @@ -459,16 +458,16 @@ class CloudflareAPI(object): if query: api_call += '?' + urlencode(query) - records,status = self._cf_api_call(api_call) + records, status = self._cf_api_call(api_call) return records - def delete_dns_records(self,**kwargs): + def delete_dns_records(self, **kwargs): params = {} - for param in ['port','proto','service','solo','type','record','value','weight','zone']: + for param in ['port', 'proto', 'service', 'solo', 'type', 'record', 'value', 'weight', 'zone']: if param in kwargs: params[param] = kwargs[param] else: - params[param] = getattr(self,param) + params[param] = getattr(self, param) records = [] content = params['value'] @@ -481,27 +480,27 @@ class CloudflareAPI(object): else: search_value = content - records = self.get_dns_records(params['zone'],params['type'],search_record,search_value) + records = self.get_dns_records(params['zone'], params['type'], search_record, search_value) for rr in records: if params['solo']: if not ((rr['type'] == params['type']) and (rr['name'] == search_record) and (rr['content'] == content)): self.changed = True if not self.module.check_mode: - result, info = self._cf_api_call('/zones/{0}/dns_records/{1}'.format(rr['zone_id'],rr['id']),'DELETE') + result, info = self._cf_api_call('/zones/{0}/dns_records/{1}'.format(rr['zone_id'], rr['id']), 'DELETE') else: self.changed = True if not self.module.check_mode: - result, info = self._cf_api_call('/zones/{0}/dns_records/{1}'.format(rr['zone_id'],rr['id']),'DELETE') + result, info = self._cf_api_call('/zones/{0}/dns_records/{1}'.format(rr['zone_id'], rr['id']), 'DELETE') return self.changed - def ensure_dns_record(self,**kwargs): + def ensure_dns_record(self, **kwargs): params = {} - for param in ['port','priority','proto','proxied','service','ttl','type','record','value','weight','zone']: + for param in ['port', 'priority', 'proto', 'proxied', 'service', 'ttl', 'type', 'record', 'value', 'weight', 'zone']: if param in kwargs: params[param] = kwargs[param] else: - params[param] = getattr(self,param) + params[param] = getattr(self, param) search_value = params['value'] search_record = params['record'] @@ -509,7 +508,7 @@ class CloudflareAPI(object): if (params['type'] is None) or (params['record'] is None): self.module.fail_json(msg="You must provide a type and a record to create a new record") - if (params['type'] in [ 'A','AAAA','CNAME','TXT','MX','NS','SPF']): + if (params['type'] in ['A', 'AAAA', 'CNAME', 'TXT', 'MX', 'NS', 'SPF']): if not params['value']: self.module.fail_json(msg="You must provide a non-empty value to create this record type") @@ -527,11 +526,11 @@ class CloudflareAPI(object): "ttl": params['ttl'] } - if (params['type'] in [ 'A', 'AAAA', 'CNAME' ]): + if (params['type'] in ['A', 'AAAA', 'CNAME']): new_record["proxied"] = params["proxied"] if params['type'] == 'MX': - for attr in [params['priority'],params['value']]: + for attr in [params['priority'], params['value']]: if (attr is None) or (attr == ''): self.module.fail_json(msg="You must provide priority and a value to create this record type") new_record = { @@ -543,7 +542,7 @@ class CloudflareAPI(object): } if params['type'] == 'SRV': - for attr in [params['port'],params['priority'],params['proto'],params['service'],params['weight'],params['value']]: + for attr in [params['port'], params['priority'], params['proto'], params['service'], params['weight'], params['value']]: if (attr is None) or (attr == ''): self.module.fail_json(msg="You must provide port, priority, proto, service, weight and a value to create this record type") srv_data = { @@ -555,12 +554,12 @@ class CloudflareAPI(object): "proto": params['proto'], "service": params['service'] } - new_record = { "type": params['type'], "ttl": params['ttl'], 'data': srv_data } + new_record = {"type": params['type'], "ttl": params['ttl'], 'data': srv_data} search_value = str(params['weight']) + '\t' + str(params['port']) + '\t' + params['value'] search_record = params['service'] + '.' + params['proto'] + '.' + params['record'] zone_id = self._get_zone_id(params['zone']) - records = self.get_dns_records(params['zone'],params['type'],search_record,search_value) + records = self.get_dns_records(params['zone'], params['type'], search_record, search_value) # in theory this should be impossible as cloudflare does not allow # the creation of duplicate records but lets cover it anyways if len(records) > 1: @@ -569,7 +568,7 @@ class CloudflareAPI(object): if len(records) == 1: cur_record = records[0] do_update = False - if (params['ttl'] is not None) and (cur_record['ttl'] != params['ttl'] ): + if (params['ttl'] is not None) and (cur_record['ttl'] != params['ttl']): do_update = True if (params['priority'] is not None) and ('priority' in cur_record) and (cur_record['priority'] != params['priority']): do_update = True @@ -582,53 +581,54 @@ class CloudflareAPI(object): if self.module.check_mode: result = new_record else: - result, info = self._cf_api_call('/zones/{0}/dns_records/{1}'.format(zone_id,records[0]['id']),'PUT',new_record) + result, info = self._cf_api_call('/zones/{0}/dns_records/{1}'.format(zone_id, records[0]['id']), 'PUT', new_record) self.changed = True - return result,self.changed + return result, self.changed else: - return records,self.changed + return records, self.changed if self.module.check_mode: result = new_record else: - result, info = self._cf_api_call('/zones/{0}/dns_records'.format(zone_id),'POST',new_record) + result, info = self._cf_api_call('/zones/{0}/dns_records'.format(zone_id), 'POST', new_record) self.changed = True - return result,self.changed + return result, self.changed + def main(): module = AnsibleModule( - argument_spec = dict( - account_api_token = dict(required=True, no_log=True, type='str'), - account_email = dict(required=True, type='str'), - port = dict(required=False, default=None, type='int'), - priority = dict(required=False, default=1, type='int'), - proto = dict(required=False, default=None, choices=[ 'tcp', 'udp' ], type='str'), - proxied = dict(required=False, default=False, type='bool'), - record = dict(required=False, default='@', aliases=['name'], type='str'), - service = dict(required=False, default=None, type='str'), - solo = dict(required=False, default=None, type='bool'), - state = dict(required=False, default='present', choices=['present', 'absent'], type='str'), - timeout = dict(required=False, default=30, type='int'), - ttl = dict(required=False, default=1, type='int'), - type = dict(required=False, default=None, choices=[ 'A', 'AAAA', 'CNAME', 'TXT', 'SRV', 'MX', 'NS', 'SPF' ], type='str'), - value = dict(required=False, default=None, aliases=['content'], type='str'), - weight = dict(required=False, default=1, type='int'), - zone = dict(required=True, default=None, aliases=['domain'], type='str'), + argument_spec=dict( + account_api_token=dict(required=True, no_log=True, type='str'), + account_email=dict(required=True, type='str'), + port=dict(required=False, default=None, type='int'), + priority=dict(required=False, default=1, type='int'), + proto=dict(required=False, default=None, choices=['tcp', 'udp'], type='str'), + proxied=dict(required=False, default=False, type='bool'), + record=dict(required=False, default='@', aliases=['name'], type='str'), + service=dict(required=False, default=None, type='str'), + solo=dict(required=False, default=None, type='bool'), + state=dict(required=False, default='present', choices=['present', 'absent'], type='str'), + timeout=dict(required=False, default=30, type='int'), + ttl=dict(required=False, default=1, type='int'), + type=dict(required=False, default=None, choices=['A', 'AAAA', 'CNAME', 'TXT', 'SRV', 'MX', 'NS', 'SPF'], type='str'), + value=dict(required=False, default=None, aliases=['content'], type='str'), + weight=dict(required=False, default=1, type='int'), + zone=dict(required=True, default=None, aliases=['domain'], type='str'), ), - supports_check_mode = True, - required_if = ([ - ('state','present',['record','type']), - ('type','MX',['priority','value']), - ('type','SRV',['port','priority','proto','service','value','weight']), - ('type','A',['value']), - ('type','AAAA',['value']), - ('type','CNAME',['value']), - ('type','TXT',['value']), - ('type','NS',['value']), - ('type','SPF',['value']) + supports_check_mode=True, + required_if=([ + ('state', 'present', ['record', 'type']), + ('type', 'MX', ['priority', 'value']), + ('type', 'SRV', ['port', 'priority', 'proto', 'service', 'value', 'weight']), + ('type', 'A', ['value']), + ('type', 'AAAA', ['value']), + ('type', 'CNAME', ['value']), + ('type', 'TXT', ['value']), + ('type', 'NS', ['value']), + ('type', 'SPF', ['value']) ] ), - required_one_of = ( - [['record','value','type']] + required_one_of=( + [['record', 'value', 'type']] ) ) @@ -645,11 +645,11 @@ def main(): # delete all records matching record name + type if cf_api.is_solo: changed = cf_api.delete_dns_records(solo=cf_api.is_solo) - result,changed = cf_api.ensure_dns_record() - if isinstance(result,list): - module.exit_json(changed=changed,result={'record': result[0]}) + result, changed = cf_api.ensure_dns_record() + if isinstance(result, list): + module.exit_json(changed=changed, result={'record': result[0]}) else: - module.exit_json(changed=changed,result={'record': result}) + module.exit_json(changed=changed, result={'record': result}) else: # force solo to False, just to be sure changed = cf_api.delete_dns_records(solo=False) diff --git a/lib/ansible/modules/net_tools/dnsimple.py b/lib/ansible/modules/net_tools/dnsimple.py index 8dfdcaed86..5739ee8a04 100644 --- a/lib/ansible/modules/net_tools/dnsimple.py +++ b/lib/ansible/modules/net_tools/dnsimple.py @@ -188,26 +188,26 @@ def main(): state=dict(required=False, choices=['present', 'absent']), solo=dict(required=False, type='bool'), ), - required_together = ( + required_together=( ['record', 'value'] ), - supports_check_mode = True, + supports_check_mode=True, ) if not HAS_DNSIMPLE: module.fail_json(msg="dnsimple required for this module") - account_email = module.params.get('account_email') + account_email = module.params.get('account_email') account_api_token = module.params.get('account_api_token') - domain = module.params.get('domain') - record = module.params.get('record') - record_ids = module.params.get('record_ids') - record_type = module.params.get('type') - ttl = module.params.get('ttl') - value = module.params.get('value') - priority = module.params.get('priority') - state = module.params.get('state') - is_solo = module.params.get('solo') + domain = module.params.get('domain') + record = module.params.get('record') + record_ids = module.params.get('record_ids') + record_type = module.params.get('type') + ttl = module.params.get('ttl') + value = module.params.get('value') + priority = module.params.get('priority') + state = module.params.get('state') + is_solo = module.params.get('solo') if account_email and account_api_token: client = DNSimple(email=account_email, api_token=account_api_token) @@ -278,7 +278,7 @@ def main(): if rr['ttl'] != ttl or rr['prio'] != priority: data = {} if ttl: - data['ttl'] = ttl + data['ttl'] = ttl if priority: data['prio'] = priority if module.check_mode: @@ -290,12 +290,12 @@ def main(): else: # create it data = { - 'name': record, + 'name': record, 'record_type': record_type, - 'content': value, + 'content': value, } if ttl: - data['ttl'] = ttl + data['ttl'] = ttl if priority: data['prio'] = priority if module.check_mode: @@ -315,7 +315,7 @@ def main(): # Make sure these record_ids either all exist or none if domain and record_ids: current_records = [str(r['record']['id']) for r in client.records(str(domain))] - wanted_records = [str(r) for r in record_ids] + wanted_records = [str(r) for r in record_ids] if state == 'present': difference = list(set(wanted_records) - set(current_records)) if difference: diff --git a/lib/ansible/modules/net_tools/dnsmadeeasy.py b/lib/ansible/modules/net_tools/dnsmadeeasy.py index 57bf8215bc..f560f32d87 100644 --- a/lib/ansible/modules/net_tools/dnsmadeeasy.py +++ b/lib/ansible/modules/net_tools/dnsmadeeasy.py @@ -401,7 +401,7 @@ class DME2(object): self.record_map = None # ["record_name"] => ID self.records = None # ["record_ID"] => self.all_records = None - self.contactList_map = None # ["contactList_name"] => ID + self.contactList_map = None # ["contactList_name"] => ID # Lookup the domain ID if passed as a domain name vs. ID if not self.domain.isdigit(): @@ -551,6 +551,7 @@ class DME2(object): # Module execution. # + def main(): module = AnsibleModule( @@ -581,7 +582,7 @@ def main(): ip3=dict(required=False), ip4=dict(required=False), ip5=dict(required=False), - validate_certs = dict(default='yes', type='bool'), + validate_certs=dict(default='yes', type='bool'), ), required_together=( ['record_value', 'record_ttl', 'record_type'] diff --git a/lib/ansible/modules/net_tools/ipinfoio_facts.py b/lib/ansible/modules/net_tools/ipinfoio_facts.py index 97e8f82a26..f1e9efae1d 100644 --- a/lib/ansible/modules/net_tools/ipinfoio_facts.py +++ b/lib/ansible/modules/net_tools/ipinfoio_facts.py @@ -96,7 +96,7 @@ class IpinfoioFacts(object): self.module = module def get_geo_data(self): - response, info = fetch_url(self.module, self.url, force=True, # NOQA + response, info = fetch_url(self.module, self.url, force=True, # NOQA timeout=self.timeout) try: info['status'] == 200 @@ -116,7 +116,7 @@ class IpinfoioFacts(object): def main(): - module = AnsibleModule( # NOQA + module = AnsibleModule( # NOQA argument_spec=dict( http_agent=dict(default=USER_AGENT), timeout=dict(type='int', default=10), diff --git a/lib/ansible/modules/net_tools/nmcli.py b/lib/ansible/modules/net_tools/nmcli.py index 702fb17684..8b71d603de 100644 --- a/lib/ansible/modules/net_tools/nmcli.py +++ b/lib/ansible/modules/net_tools/nmcli.py @@ -13,7 +13,7 @@ ANSIBLE_METADATA = {'metadata_version': '1.1', 'supported_by': 'community'} -DOCUMENTATION=''' +DOCUMENTATION = ''' --- module: nmcli author: "Chris Long (@alcamie101)" @@ -208,7 +208,7 @@ options: ''' -EXAMPLES=''' +EXAMPLES = ''' # These examples are using the following inventory: # # ## Directory layout: @@ -482,17 +482,17 @@ EXAMPLES=''' # - 10 Connection, device, or access point does not exist. ''' -HAVE_DBUS=False +HAVE_DBUS = False try: import dbus - HAVE_DBUS=True + HAVE_DBUS = True except ImportError: pass -HAVE_NM_CLIENT=False +HAVE_NM_CLIENT = False try: from gi.repository import NetworkManager, NMClient - HAVE_NM_CLIENT=True + HAVE_NM_CLIENT = True except ImportError: pass @@ -512,75 +512,74 @@ class Nmcli(object): All subclasses MUST define platform and distribution (which may be None). """ - platform='Generic' - distribution=None + platform = 'Generic' + distribution = None if HAVE_DBUS: - bus=dbus.SystemBus() + bus = dbus.SystemBus() # The following is going to be used in dbus code - DEVTYPES={1: "Ethernet", - 2: "Wi-Fi", - 5: "Bluetooth", - 6: "OLPC", - 7: "WiMAX", - 8: "Modem", - 9: "InfiniBand", - 10: "Bond", - 11: "VLAN", - 12: "ADSL", - 13: "Bridge", - 14: "Generic", - 15: "Team" + DEVTYPES = {1: "Ethernet", + 2: "Wi-Fi", + 5: "Bluetooth", + 6: "OLPC", + 7: "WiMAX", + 8: "Modem", + 9: "InfiniBand", + 10: "Bond", + 11: "VLAN", + 12: "ADSL", + 13: "Bridge", + 14: "Generic", + 15: "Team" } - STATES={0: "Unknown", - 10: "Unmanaged", - 20: "Unavailable", - 30: "Disconnected", - 40: "Prepare", - 50: "Config", - 60: "Need Auth", - 70: "IP Config", - 80: "IP Check", - 90: "Secondaries", - 100: "Activated", - 110: "Deactivating", - 120: "Failed" - } - + STATES = {0: "Unknown", + 10: "Unmanaged", + 20: "Unavailable", + 30: "Disconnected", + 40: "Prepare", + 50: "Config", + 60: "Need Auth", + 70: "IP Config", + 80: "IP Check", + 90: "Secondaries", + 100: "Activated", + 110: "Deactivating", + 120: "Failed" + } def __init__(self, module): - self.module=module - self.state=module.params['state'] - self.autoconnect=module.params['autoconnect'] - self.conn_name=module.params['conn_name'] - self.master=module.params['master'] - self.ifname=module.params['ifname'] - self.type=module.params['type'] - self.ip4=module.params['ip4'] - self.gw4=module.params['gw4'] - self.dns4=' '.join(module.params['dns4']) - self.ip6=module.params['ip6'] - self.gw6=module.params['gw6'] - self.dns6=module.params['dns6'] - self.mtu=module.params['mtu'] - self.stp=module.params['stp'] - self.priority=module.params['priority'] - self.mode=module.params['mode'] - self.miimon=module.params['miimon'] - self.downdelay=module.params['downdelay'] - self.updelay=module.params['updelay'] - self.arp_interval=module.params['arp_interval'] - self.arp_ip_target=module.params['arp_ip_target'] - self.slavepriority=module.params['slavepriority'] - self.forwarddelay=module.params['forwarddelay'] - self.hellotime=module.params['hellotime'] - self.maxage=module.params['maxage'] - self.ageingtime=module.params['ageingtime'] - self.mac=module.params['mac'] - self.vlanid=module.params['vlanid'] - self.vlandev=module.params['vlandev'] - self.flags=module.params['flags'] - self.ingress=module.params['ingress'] - self.egress=module.params['egress'] + self.module = module + self.state = module.params['state'] + self.autoconnect = module.params['autoconnect'] + self.conn_name = module.params['conn_name'] + self.master = module.params['master'] + self.ifname = module.params['ifname'] + self.type = module.params['type'] + self.ip4 = module.params['ip4'] + self.gw4 = module.params['gw4'] + self.dns4 = ' '.join(module.params['dns4']) + self.ip6 = module.params['ip6'] + self.gw6 = module.params['gw6'] + self.dns6 = module.params['dns6'] + self.mtu = module.params['mtu'] + self.stp = module.params['stp'] + self.priority = module.params['priority'] + self.mode = module.params['mode'] + self.miimon = module.params['miimon'] + self.downdelay = module.params['downdelay'] + self.updelay = module.params['updelay'] + self.arp_interval = module.params['arp_interval'] + self.arp_ip_target = module.params['arp_ip_target'] + self.slavepriority = module.params['slavepriority'] + self.forwarddelay = module.params['forwarddelay'] + self.hellotime = module.params['hellotime'] + self.maxage = module.params['maxage'] + self.ageingtime = module.params['ageingtime'] + self.mac = module.params['mac'] + self.vlanid = module.params['vlanid'] + self.vlandev = module.params['vlandev'] + self.flags = module.params['flags'] + self.ingress = module.params['ingress'] + self.egress = module.params['egress'] def execute_command(self, cmd, use_unsafe_shell=False, data=None): return self.module.run_command(cmd, use_unsafe_shell=use_unsafe_shell, data=data) @@ -589,41 +588,41 @@ class Nmcli(object): try: # returns a dict of dicts mapping name::setting, where setting is a dict # mapping key::value. Each member of the 'setting' dict is a secret - secrets=proxy.GetSecrets(setting_name) + secrets = proxy.GetSecrets(setting_name) # Copy the secrets into our connection config for setting in secrets: for key in secrets[setting]: - config[setting_name][key]=secrets[setting][key] + config[setting_name][key] = secrets[setting][key] except: pass def dict_to_string(self, d): # Try to trivially translate a dictionary's elements into nice string # formatting. - dstr="" + dstr = "" for key in d: - val=d[key] - str_val="" - add_string=True + val = d[key] + str_val = "" + add_string = True if isinstance(val, dbus.Array): for elt in val: if isinstance(elt, dbus.Byte): - str_val+="%s " % int(elt) + str_val += "%s " % int(elt) elif isinstance(elt, dbus.String): - str_val+="%s" % elt + str_val += "%s" % elt elif isinstance(val, dbus.Dictionary): - dstr+=self.dict_to_string(val) - add_string=False + dstr += self.dict_to_string(val) + add_string = False else: - str_val=val + str_val = val if add_string: - dstr+="%s: %s\n" % ( key, str_val) + dstr += "%s: %s\n" % (key, str_val) return dstr def connection_to_string(self, config): # dump a connection configuration to use in list_connection_info - setting_list=[] + setting_list = [] for setting_name in config: setting_list.append(self.dict_to_string(config[setting_name])) return setting_list @@ -637,18 +636,18 @@ class Nmcli(object): def list_connection_info(self): # Ask the settings service for the list of connections it provides - bus=dbus.SystemBus() + bus = dbus.SystemBus() - service_name="org.freedesktop.NetworkManager" - proxy=bus.get_object(service_name, "/org/freedesktop/NetworkManager/Settings") - settings=dbus.Interface(proxy, "org.freedesktop.NetworkManager.Settings") - connection_paths=settings.ListConnections() - connection_list=[] + service_name = "org.freedesktop.NetworkManager" + proxy = bus.get_object(service_name, "/org/freedesktop/NetworkManager/Settings") + settings = dbus.Interface(proxy, "org.freedesktop.NetworkManager.Settings") + connection_paths = settings.ListConnections() + connection_list = [] # List each connection's name, UUID, and type for path in connection_paths: - con_proxy=bus.get_object(service_name, path) - settings_connection=dbus.Interface(con_proxy, "org.freedesktop.NetworkManager.Settings.Connection") - config=settings_connection.GetSettings() + con_proxy = bus.get_object(service_name, path) + settings_connection = dbus.Interface(con_proxy, "org.freedesktop.NetworkManager.Settings.Connection") + config = settings_connection.GetSettings() # Now get secrets too; we grab the secrets for each type of connection # (since there isn't a "get all secrets" call because most of the time @@ -662,7 +661,7 @@ class Nmcli(object): self.merge_secrets(settings_connection, config, 'ppp') # Get the details of the 'connection' setting - s_con=config['connection'] + s_con = config['connection'] connection_list.append(s_con['id']) connection_list.append(s_con['uuid']) connection_list.append(s_con['type']) @@ -671,14 +670,14 @@ class Nmcli(object): def connection_exists(self): # we are going to use name and type in this instance to find if that connection exists and is of type x - connections=self.list_connection_info() + connections = self.list_connection_info() for con_item in connections: - if self.conn_name==con_item: + if self.conn_name == con_item: return True def down_connection(self): - cmd=[self.module.get_bin_path('nmcli', True)] + cmd = [self.module.get_bin_path('nmcli', True)] # if self.connection_exists(): cmd.append('con') cmd.append('down') @@ -686,14 +685,14 @@ class Nmcli(object): return self.execute_command(cmd) def up_connection(self): - cmd=[self.module.get_bin_path('nmcli', True)] + cmd = [self.module.get_bin_path('nmcli', True)] cmd.append('con') cmd.append('up') cmd.append(self.conn_name) return self.execute_command(cmd) def create_connection_team(self): - cmd=[self.module.get_bin_path('nmcli', True)] + cmd = [self.module.get_bin_path('nmcli', True)] # format for creating team interface cmd.append('con') cmd.append('add') @@ -727,7 +726,7 @@ class Nmcli(object): return cmd def modify_connection_team(self): - cmd=[self.module.get_bin_path('nmcli', True)] + cmd = [self.module.get_bin_path('nmcli', True)] # format for modifying team interface cmd.append('con') cmd.append('mod') @@ -757,7 +756,7 @@ class Nmcli(object): return cmd def create_connection_team_slave(self): - cmd=[self.module.get_bin_path('nmcli', True)] + cmd = [self.module.get_bin_path('nmcli', True)] # format for creating team-slave interface cmd.append('connection') cmd.append('add') @@ -782,7 +781,7 @@ class Nmcli(object): return cmd def modify_connection_team_slave(self): - cmd=[self.module.get_bin_path('nmcli', True)] + cmd = [self.module.get_bin_path('nmcli', True)] # format for modifying team-slave interface cmd.append('con') cmd.append('mod') @@ -795,7 +794,7 @@ class Nmcli(object): return cmd def create_connection_bond(self): - cmd=[self.module.get_bin_path('nmcli', True)] + cmd = [self.module.get_bin_path('nmcli', True)] # format for creating bond interface cmd.append('con') cmd.append('add') @@ -847,7 +846,7 @@ class Nmcli(object): return cmd def modify_connection_bond(self): - cmd=[self.module.get_bin_path('nmcli', True)] + cmd = [self.module.get_bin_path('nmcli', True)] # format for modifying bond interface cmd.append('con') cmd.append('mod') @@ -876,7 +875,7 @@ class Nmcli(object): return cmd def create_connection_bond_slave(self): - cmd=[self.module.get_bin_path('nmcli', True)] + cmd = [self.module.get_bin_path('nmcli', True)] # format for creating bond-slave interface cmd.append('connection') cmd.append('add') @@ -898,7 +897,7 @@ class Nmcli(object): return cmd def modify_connection_bond_slave(self): - cmd=[self.module.get_bin_path('nmcli', True)] + cmd = [self.module.get_bin_path('nmcli', True)] # format for modifying bond-slave interface cmd.append('con') cmd.append('mod') @@ -908,7 +907,7 @@ class Nmcli(object): return cmd def create_connection_ethernet(self): - cmd=[self.module.get_bin_path('nmcli', True)] + cmd = [self.module.get_bin_path('nmcli', True)] # format for creating ethernet interface # To add an Ethernet connection with static IP configuration, issue a command as follows # - nmcli: name=add conn_name=my-eth1 ifname=eth1 type=ethernet ip4=192.0.2.100/24 gw4=192.0.2.1 state=present @@ -945,7 +944,7 @@ class Nmcli(object): return cmd def modify_connection_ethernet(self): - cmd=[self.module.get_bin_path('nmcli', True)] + cmd = [self.module.get_bin_path('nmcli', True)] # format for modifying ethernet interface # To add an Ethernet connection with static IP configuration, issue a command as follows # - nmcli: name=add conn_name=my-eth1 ifname=eth1 type=ethernet ip4=192.0.2.100/24 gw4=192.0.2.1 state=present @@ -980,110 +979,110 @@ class Nmcli(object): return cmd def create_connection_bridge(self): - cmd=[self.module.get_bin_path('nmcli', True)] + cmd = [self.module.get_bin_path('nmcli', True)] # format for creating bridge interface return cmd def modify_connection_bridge(self): - cmd=[self.module.get_bin_path('nmcli', True)] + cmd = [self.module.get_bin_path('nmcli', True)] # format for modifying bridge interface return cmd def create_connection_vlan(self): - cmd=[self.module.get_bin_path('nmcli', True)] + cmd = [self.module.get_bin_path('nmcli', True)] # format for creating ethernet interface return cmd def modify_connection_vlan(self): - cmd=[self.module.get_bin_path('nmcli', True)] + cmd = [self.module.get_bin_path('nmcli', True)] # format for modifying ethernet interface return cmd def create_connection(self): - cmd=[] - if self.type=='team': + cmd = [] + if self.type == 'team': # cmd=self.create_connection_team() if (self.dns4 is not None) or (self.dns6 is not None): - cmd=self.create_connection_team() + cmd = self.create_connection_team() self.execute_command(cmd) - cmd=self.modify_connection_team() + cmd = self.modify_connection_team() self.execute_command(cmd) - cmd=self.up_connection() + cmd = self.up_connection() return self.execute_command(cmd) elif (self.dns4 is None) or (self.dns6 is None): - cmd=self.create_connection_team() + cmd = self.create_connection_team() return self.execute_command(cmd) - elif self.type=='team-slave': + elif self.type == 'team-slave': if self.mtu is not None: - cmd=self.create_connection_team_slave() + cmd = self.create_connection_team_slave() self.execute_command(cmd) - cmd=self.modify_connection_team_slave() + cmd = self.modify_connection_team_slave() self.execute_command(cmd) # cmd=self.up_connection() return self.execute_command(cmd) else: - cmd=self.create_connection_team_slave() + cmd = self.create_connection_team_slave() return self.execute_command(cmd) - elif self.type=='bond': + elif self.type == 'bond': if (self.mtu is not None) or (self.dns4 is not None) or (self.dns6 is not None): - cmd=self.create_connection_bond() + cmd = self.create_connection_bond() self.execute_command(cmd) - cmd=self.modify_connection_bond() + cmd = self.modify_connection_bond() self.execute_command(cmd) - cmd=self.up_connection() + cmd = self.up_connection() return self.execute_command(cmd) else: - cmd=self.create_connection_bond() + cmd = self.create_connection_bond() return self.execute_command(cmd) - elif self.type=='bond-slave': - cmd=self.create_connection_bond_slave() - elif self.type=='ethernet': + elif self.type == 'bond-slave': + cmd = self.create_connection_bond_slave() + elif self.type == 'ethernet': if (self.mtu is not None) or (self.dns4 is not None) or (self.dns6 is not None): - cmd=self.create_connection_ethernet() + cmd = self.create_connection_ethernet() self.execute_command(cmd) - cmd=self.modify_connection_ethernet() + cmd = self.modify_connection_ethernet() self.execute_command(cmd) - cmd=self.up_connection() + cmd = self.up_connection() return self.execute_command(cmd) else: - cmd=self.create_connection_ethernet() + cmd = self.create_connection_ethernet() return self.execute_command(cmd) - elif self.type=='bridge': - cmd=self.create_connection_bridge() - elif self.type=='vlan': - cmd=self.create_connection_vlan() + elif self.type == 'bridge': + cmd = self.create_connection_bridge() + elif self.type == 'vlan': + cmd = self.create_connection_vlan() return self.execute_command(cmd) def remove_connection(self): # self.down_connection() - cmd=[self.module.get_bin_path('nmcli', True)] + cmd = [self.module.get_bin_path('nmcli', True)] cmd.append('con') cmd.append('del') cmd.append(self.conn_name) return self.execute_command(cmd) def modify_connection(self): - cmd=[] - if self.type=='team': - cmd=self.modify_connection_team() - elif self.type=='team-slave': - cmd=self.modify_connection_team_slave() - elif self.type=='bond': - cmd=self.modify_connection_bond() - elif self.type=='bond-slave': - cmd=self.modify_connection_bond_slave() - elif self.type=='ethernet': - cmd=self.modify_connection_ethernet() - elif self.type=='bridge': - cmd=self.modify_connection_bridge() - elif self.type=='vlan': - cmd=self.modify_connection_vlan() + cmd = [] + if self.type == 'team': + cmd = self.modify_connection_team() + elif self.type == 'team-slave': + cmd = self.modify_connection_team_slave() + elif self.type == 'bond': + cmd = self.modify_connection_bond() + elif self.type == 'bond-slave': + cmd = self.modify_connection_bond_slave() + elif self.type == 'ethernet': + cmd = self.modify_connection_ethernet() + elif self.type == 'bridge': + cmd = self.modify_connection_bridge() + elif self.type == 'vlan': + cmd = self.modify_connection_vlan() return self.execute_command(cmd) def main(): # Parsing argument file - module=AnsibleModule( + module = AnsibleModule( argument_spec=dict( autoconnect=dict(required=False, default=None, type='bool'), state=dict(required=True, choices=['present', 'absent'], type='str'), @@ -1132,57 +1131,57 @@ def main(): if not HAVE_NM_CLIENT: module.fail_json(msg="This module requires NetworkManager glib API") - nmcli=Nmcli(module) + nmcli = Nmcli(module) - rc=None - out='' - err='' - result={} - result['conn_name']=nmcli.conn_name - result['state']=nmcli.state + rc = None + out = '' + err = '' + result = {} + result['conn_name'] = nmcli.conn_name + result['state'] = nmcli.state # check for issues if nmcli.conn_name is None: nmcli.module.fail_json(msg="You haven't specified a name for the connection") # team-slave checks - if nmcli.type=='team-slave' and nmcli.master is None: + if nmcli.type == 'team-slave' and nmcli.master is None: nmcli.module.fail_json(msg="You haven't specified a name for the master so we're not changing a thing") - if nmcli.type=='team-slave' and nmcli.ifname is None: + if nmcli.type == 'team-slave' and nmcli.ifname is None: nmcli.module.fail_json(msg="You haven't specified a name for the connection") - if nmcli.state=='absent': + if nmcli.state == 'absent': if nmcli.connection_exists(): if module.check_mode: module.exit_json(changed=True) - (rc, out, err)=nmcli.down_connection() - (rc, out, err)=nmcli.remove_connection() - if rc!=0: - module.fail_json(name =('No Connection named %s exists' % nmcli.conn_name), msg=err, rc=rc) + (rc, out, err) = nmcli.down_connection() + (rc, out, err) = nmcli.remove_connection() + if rc != 0: + module.fail_json(name=('No Connection named %s exists' % nmcli.conn_name), msg=err, rc=rc) - elif nmcli.state=='present': + elif nmcli.state == 'present': if nmcli.connection_exists(): # modify connection (note: this function is check mode aware) # result['Connection']=('Connection %s of Type %s is not being added' % (nmcli.conn_name, nmcli.type)) - result['Exists']='Connections do exist so we are modifying them' + result['Exists'] = 'Connections do exist so we are modifying them' if module.check_mode: module.exit_json(changed=True) - (rc, out, err)=nmcli.modify_connection() + (rc, out, err) = nmcli.modify_connection() if not nmcli.connection_exists(): - result['Connection']=('Connection %s of Type %s is being added' % (nmcli.conn_name, nmcli.type)) + result['Connection'] = ('Connection %s of Type %s is being added' % (nmcli.conn_name, nmcli.type)) if module.check_mode: module.exit_json(changed=True) - (rc, out, err)=nmcli.create_connection() - if rc is not None and rc!=0: + (rc, out, err) = nmcli.create_connection() + if rc is not None and rc != 0: module.fail_json(name=nmcli.conn_name, msg=err, rc=rc) if rc is None: - result['changed']=False + result['changed'] = False else: - result['changed']=True + result['changed'] = True if out: - result['stdout']=out + result['stdout'] = out if err: - result['stderr']=err + result['stderr'] = err module.exit_json(**result) diff --git a/lib/ansible/modules/net_tools/omapi_host.py b/lib/ansible/modules/net_tools/omapi_host.py index 468296dfa7..e6f60a00ee 100644 --- a/lib/ansible/modules/net_tools/omapi_host.py +++ b/lib/ansible/modules/net_tools/omapi_host.py @@ -234,7 +234,7 @@ class OmapiHostManager: fields_to_update = {} if to_bytes('ip-address', errors='surrogate_or_strict') not in response_obj or \ - unpack_ip(response_obj[to_bytes('ip-address', errors='surrogate_or_strict')]) != self.module.params['ip']: + unpack_ip(response_obj[to_bytes('ip-address', errors='surrogate_or_strict')]) != self.module.params['ip']: fields_to_update['ip-address'] = pack_ip(self.module.params['ip']) # Name cannot be changed diff --git a/lib/ansible/modules/net_tools/snmp_facts.py b/lib/ansible/modules/net_tools/snmp_facts.py index 8a89f9f806..84c905dcf6 100644 --- a/lib/ansible/modules/net_tools/snmp_facts.py +++ b/lib/ansible/modules/net_tools/snmp_facts.py @@ -102,32 +102,32 @@ from ansible.module_utils._text import to_text class DefineOid(object): - def __init__(self,dotprefix=False): + def __init__(self, dotprefix=False): if dotprefix: dp = "." else: dp = "" # From SNMPv2-MIB - self.sysDescr = dp + "1.3.6.1.2.1.1.1.0" + self.sysDescr = dp + "1.3.6.1.2.1.1.1.0" self.sysObjectId = dp + "1.3.6.1.2.1.1.2.0" - self.sysUpTime = dp + "1.3.6.1.2.1.1.3.0" - self.sysContact = dp + "1.3.6.1.2.1.1.4.0" - self.sysName = dp + "1.3.6.1.2.1.1.5.0" + self.sysUpTime = dp + "1.3.6.1.2.1.1.3.0" + self.sysContact = dp + "1.3.6.1.2.1.1.4.0" + self.sysName = dp + "1.3.6.1.2.1.1.5.0" self.sysLocation = dp + "1.3.6.1.2.1.1.6.0" # From IF-MIB - self.ifIndex = dp + "1.3.6.1.2.1.2.2.1.1" - self.ifDescr = dp + "1.3.6.1.2.1.2.2.1.2" - self.ifMtu = dp + "1.3.6.1.2.1.2.2.1.4" - self.ifSpeed = dp + "1.3.6.1.2.1.2.2.1.5" + self.ifIndex = dp + "1.3.6.1.2.1.2.2.1.1" + self.ifDescr = dp + "1.3.6.1.2.1.2.2.1.2" + self.ifMtu = dp + "1.3.6.1.2.1.2.2.1.4" + self.ifSpeed = dp + "1.3.6.1.2.1.2.2.1.5" self.ifPhysAddress = dp + "1.3.6.1.2.1.2.2.1.6" self.ifAdminStatus = dp + "1.3.6.1.2.1.2.2.1.7" - self.ifOperStatus = dp + "1.3.6.1.2.1.2.2.1.8" - self.ifAlias = dp + "1.3.6.1.2.1.31.1.1.1.18" + self.ifOperStatus = dp + "1.3.6.1.2.1.2.2.1.8" + self.ifAlias = dp + "1.3.6.1.2.1.31.1.1.1.18" # From IP-MIB - self.ipAdEntAddr = dp + "1.3.6.1.2.1.4.20.1.1" + self.ipAdEntAddr = dp + "1.3.6.1.2.1.4.20.1.1" self.ipAdEntIfIndex = dp + "1.3.6.1.2.1.4.20.1.2" self.ipAdEntNetMask = dp + "1.3.6.1.2.1.4.20.1.3" @@ -141,6 +141,7 @@ def decode_hex(hexstring): else: return hexstring + def decode_mac(hexstring): if len(hexstring) != 14: @@ -150,17 +151,19 @@ def decode_mac(hexstring): else: return hexstring + def lookup_adminstatus(int_adminstatus): adminstatus_options = { 1: 'up', 2: 'down', 3: 'testing' - } + } if int_adminstatus in adminstatus_options: return adminstatus_options[int_adminstatus] else: return "" + def lookup_operstatus(int_operstatus): operstatus_options = { 1: 'up', @@ -170,12 +173,13 @@ def lookup_operstatus(int_operstatus): 5: 'dormant', 6: 'notPresent', 7: 'lowerLayerDown' - } + } if int_operstatus in operstatus_options: return operstatus_options[int_operstatus] else: return "" + def main(): module = AnsibleModule( argument_spec=dict( @@ -189,7 +193,7 @@ def main(): authkey=dict(required=False), privkey=dict(required=False), removeplaceholder=dict(required=False)), - required_together = ( ['username','level','integrity','authkey'],['privacy','privkey'],), + required_together=(['username', 'level', 'integrity', 'authkey'], ['privacy', 'privkey'],), supports_check_mode=False) m_args = module.params @@ -211,7 +215,6 @@ def main(): if m_args['level'] == "authPriv" and m_args['privacy'] is None: module.fail_json(msg='Privacy algorithm not set when using authPriv') - if m_args['integrity'] == "sha": integrity_proto = cmdgen.usmHMACSHAAuthProtocol elif m_args['integrity'] == "md5": @@ -240,7 +243,7 @@ def main(): # Use v without a prefix to use with return values v = DefineOid(dotprefix=False) - Tree = lambda: defaultdict(Tree) + def Tree(): return defaultdict(Tree) results = Tree() @@ -256,7 +259,6 @@ def main(): lookupMib=False ) - if errorIndication: module.fail_json(msg=str(errorIndication)) @@ -294,7 +296,6 @@ def main(): lookupMib=False ) - if errorIndication: module.fail_json(msg=str(errorIndication)) @@ -351,9 +352,9 @@ def main(): for ipv4_network in ipv4_networks: current_interface = ipv4_networks[ipv4_network]['interface'] current_network = { - 'address': ipv4_networks[ipv4_network]['address'], - 'netmask': ipv4_networks[ipv4_network]['netmask'] - } + 'address': ipv4_networks[ipv4_network]['address'], + 'netmask': ipv4_networks[ipv4_network]['netmask'] + } if not current_interface in interface_to_ipv4: interface_to_ipv4[current_interface] = [] interface_to_ipv4[current_interface].append(current_network) diff --git a/lib/ansible/modules/network/a10/a10_server_axapi3.py b/lib/ansible/modules/network/a10/a10_server_axapi3.py index 7fe6accbf2..64b5a3540d 100644 --- a/lib/ansible/modules/network/a10/a10_server_axapi3.py +++ b/lib/ansible/modules/network/a10/a10_server_axapi3.py @@ -171,7 +171,6 @@ def main(): # validate the ports data structure validate_ports(module, slb_server_ports) - json_post = { "server-list": [ { @@ -188,7 +187,7 @@ def main(): if slb_server_status: json_post['server-list'][0]['action'] = slb_server_status - slb_server_data = axapi_call_v3(module, axapi_base_url+'slb/server/', method='GET', body='', signature=signature) + slb_server_data = axapi_call_v3(module, axapi_base_url + 'slb/server/', method='GET', body='', signature=signature) # for empty slb server list if axapi_failure(slb_server_data): @@ -203,7 +202,7 @@ def main(): changed = False if operation == 'create': if slb_server_exists is False: - result = axapi_call_v3(module, axapi_base_url+'slb/server/', method='POST', body=json.dumps(json_post), signature=signature) + result = axapi_call_v3(module, axapi_base_url + 'slb/server/', method='POST', body=json.dumps(json_post), signature=signature) if axapi_failure(result): module.fail_json(msg="failed to create the server: %s" % result['response']['err']['msg']) changed = True @@ -234,7 +233,7 @@ def main(): # if the config has changed, save the config unless otherwise requested if changed and write_config: - write_result = axapi_call_v3(module, axapi_base_url+'write/memory/', method='POST', body='', signature=signature) + write_result = axapi_call_v3(module, axapi_base_url + 'write/memory/', method='POST', body='', signature=signature) if axapi_failure(write_result): module.fail_json(msg="failed to save the configuration: %s" % write_result['response']['err']['msg']) diff --git a/lib/ansible/modules/network/a10/a10_virtual_server.py b/lib/ansible/modules/network/a10/a10_virtual_server.py index 560f16a8c2..c007a9128a 100644 --- a/lib/ansible/modules/network/a10/a10_virtual_server.py +++ b/lib/ansible/modules/network/a10/a10_virtual_server.py @@ -102,7 +102,7 @@ content: import json from ansible.module_utils.network.a10.a10 import (axapi_call, a10_argument_spec, axapi_authenticate, axapi_failure, - axapi_enabled_disabled, axapi_get_vport_protocol, AXAPI_VPORT_PROTOCOLS) + axapi_enabled_disabled, axapi_get_vport_protocol, AXAPI_VPORT_PROTOCOLS) from ansible.module_utils.basic import AnsibleModule from ansible.module_utils.urls import url_argument_spec @@ -146,6 +146,7 @@ def validate_ports(module, ports): if 'service_group' not in item: item['service_group'] = '' + def main(): argument_spec = a10_argument_spec() argument_spec.update(url_argument_spec()) diff --git a/lib/ansible/modules/network/aos/aos_asn_pool.py b/lib/ansible/modules/network/aos/aos_asn_pool.py index d204292d2f..1fbc9cd806 100644 --- a/lib/ansible/modules/network/aos/aos_asn_pool.py +++ b/lib/ansible/modules/network/aos/aos_asn_pool.py @@ -142,18 +142,19 @@ import json from ansible.module_utils.basic import AnsibleModule from ansible.module_utils.network.aos.aos import get_aos_session, find_collection_item, do_load_resource, check_aos_version, content_to_dict + def check_ranges_are_valid(module, ranges): i = 1 for range in ranges: - if not isinstance(range, list) : + if not isinstance(range, list): module.fail_json(msg="Range (%i) must be a list not %s" % (i, type(range))) elif len(range) != 2: module.fail_json(msg="Range (%i) must be a list of 2 members, not %i" % (i, len(range))) - elif not isinstance( range[0], int ): - module.fail_json(msg="1st element of range (%i) must be integer instead of %s " % (i,type(range[0]))) - elif not isinstance( range[1], int ): - module.fail_json(msg="2nd element of range (%i) must be integer instead of %s " % (i,type(range[1]))) + elif not isinstance(range[0], int): + module.fail_json(msg="1st element of range (%i) must be integer instead of %s " % (i, type(range[0]))) + elif not isinstance(range[1], int): + module.fail_json(msg="2nd element of range (%i) must be integer instead of %s " % (i, type(range[1]))) elif range[1] <= range[0]: module.fail_json(msg="2nd element of range (%i) must be bigger than 1st " % (i)) @@ -161,24 +162,26 @@ def check_ranges_are_valid(module, ranges): return True + def get_list_of_range(asn_pool): ranges = [] for range in asn_pool.value['ranges']: - ranges.append([ range['first'], range['last']]) + ranges.append([range['first'], range['last']]) return ranges + def create_new_asn_pool(asn_pool, name, ranges): # Create value datum = dict(display_name=name, ranges=[]) for range in ranges: - datum['ranges'].append(dict(first=range[0],last=range[1])) + datum['ranges'].append(dict(first=range[0], last=range[1])) asn_pool.datum = datum - ## Write to AOS + # Write to AOS return asn_pool.write() @@ -190,7 +193,7 @@ def asn_pool_absent(module, aos, my_pool): if my_pool.exists is False: module.exit_json(changed=False, name=margs['name'], id='', value={}) - ## Check if object is currently in Use or Not + # Check if object is currently in Use or Not # If in Use, return an error if my_pool.value: if my_pool.value['status'] != 'not_in_use': @@ -205,10 +208,10 @@ def asn_pool_absent(module, aos, my_pool): except: module.fail_json(msg="An error occurred, while trying to delete the ASN Pool") - module.exit_json( changed=True, - name=my_pool.name, - id=my_pool.id, - value={} ) + module.exit_json(changed=True, + name=my_pool.name, + id=my_pool.id, + value={}) def asn_pool_present(module, aos, my_pool): @@ -236,10 +239,10 @@ def asn_pool_present(module, aos, my_pool): except: module.fail_json(msg="An error occurred while trying to create a new ASN Pool ") - module.exit_json( changed=True, - name=my_pool.name, - id=my_pool.id, - value=my_pool.value ) + module.exit_json(changed=True, + name=my_pool.name, + id=my_pool.id, + value=my_pool.value) # Currently only check if the pool exist or not # if exist return change false @@ -248,14 +251,16 @@ def asn_pool_present(module, aos, my_pool): # if pool already exist, check if list of ASN is the same # if same just return the object and report change false # if set(get_list_of_range(my_pool)) == set(margs['ranges']): - module.exit_json( changed=False, - name=my_pool.name, - id=my_pool.id, - value=my_pool.value ) + module.exit_json(changed=False, + name=my_pool.name, + id=my_pool.id, + value=my_pool.value) # ######################################################## # Main Function # ######################################################## + + def asn_pool(module): margs = module.params @@ -271,7 +276,7 @@ def asn_pool(module): # Check ID / Name and Content if margs['content'] is not None: - content = content_to_dict(module, margs['content'] ) + content = content_to_dict(module, margs['content']) if 'display_name' in content.keys(): item_name = content['display_name'] @@ -293,8 +298,8 @@ def asn_pool(module): # ---------------------------------------------------- try: my_pool = find_collection_item(aos.AsnPools, - item_name=item_name, - item_id=item_id) + item_name=item_name, + item_id=item_id) except: module.fail_json(msg="Unable to find the IP Pool based on name or ID, something went wrong") @@ -309,19 +314,20 @@ def asn_pool(module): asn_pool_present(module, aos, my_pool) + def main(): module = AnsibleModule( argument_spec=dict( session=dict(required=True, type="dict"), - name=dict(required=False ), - id=dict(required=False ), + name=dict(required=False), + id=dict(required=False), content=dict(required=False, type="json"), - state=dict( required=False, - choices=['present', 'absent'], - default="present"), + state=dict(required=False, + choices=['present', 'absent'], + default="present"), ranges=dict(required=False, type="list", default=[]) ), - mutually_exclusive = [('name', 'id', 'content')], + mutually_exclusive=[('name', 'id', 'content')], required_one_of=[('name', 'id', 'content')], supports_check_mode=True ) diff --git a/lib/ansible/modules/network/aos/aos_blueprint.py b/lib/ansible/modules/network/aos/aos_blueprint.py index 7ae32043c1..f61f9ba203 100644 --- a/lib/ansible/modules/network/aos/aos_blueprint.py +++ b/lib/ansible/modules/network/aos/aos_blueprint.py @@ -136,6 +136,7 @@ from ansible.module_utils.basic import AnsibleModule from ansible.module_utils.network.aos.aos import get_aos_session, check_aos_version, find_collection_item from ansible.module_utils.pycompat24 import get_exception + def create_blueprint(module, aos, name): margs = module.params @@ -152,7 +153,7 @@ def create_blueprint(module, aos, name): exc = get_exception() msg = "Unable to create blueprint: %s" % exc.message if 'UNPROCESSABLE ENTITY' in exc.message: - msg+= ' (likely missing dependencies)' + msg += ' (likely missing dependencies)' module.fail_json(msg=msg) @@ -177,6 +178,7 @@ def ensure_absent(module, aos, blueprint): id=blueprint.id, name=blueprint.name) + def ensure_present(module, aos, blueprint): margs = module.params @@ -211,13 +213,14 @@ def ensure_present(module, aos, blueprint): module.exit_json(changed=True, name=margs['name']) + def ensure_build_ready(module, aos, blueprint): margs = module.params if not blueprint.exists: module.fail_json(msg='blueprint %s does not exist' % blueprint.name) - if blueprint.await_build_ready(timeout=margs['timeout']*1000): + if blueprint.await_build_ready(timeout=margs['timeout'] * 1000): module.exit_json(contents=blueprint.contents) else: module.fail_json(msg='blueprint %s has build errors', @@ -247,8 +250,8 @@ def aos_blueprint(module): #---------------------------------------------------- try: my_blueprint = find_collection_item(aos.Blueprints, - item_name=item_name, - item_id=item_id) + item_name=item_name, + item_id=item_id) except: module.fail_json(msg="Unable to find the Blueprint based on name or ID, something went wrong") @@ -273,7 +276,7 @@ def main(): argument_spec=dict( session=dict(required=True, type="dict"), name=dict(required=False), - id=dict(required=False ), + id=dict(required=False), state=dict(choices=[ 'present', 'absent', 'build-ready'], default='present'), @@ -281,7 +284,7 @@ def main(): template=dict(required=False), reference_arch=dict(required=False) ), - mutually_exclusive = [('name', 'id')], + mutually_exclusive=[('name', 'id')], required_one_of=[('name', 'id')], supports_check_mode=True ) diff --git a/lib/ansible/modules/network/aos/aos_blueprint_param.py b/lib/ansible/modules/network/aos/aos_blueprint_param.py index a2417dbe0b..e5d1bd0b42 100644 --- a/lib/ansible/modules/network/aos/aos_blueprint_param.py +++ b/lib/ansible/modules/network/aos/aos_blueprint_param.py @@ -189,6 +189,7 @@ param_map_list = dict( ) ) + def get_collection_from_param_map(module, aos): param_map = None @@ -220,6 +221,7 @@ def get_collection_from_param_map(module, aos): return None + def blueprint_param_present(module, aos, blueprint, param, param_value): margs = module.params @@ -278,6 +280,7 @@ def blueprint_param_absent(module, aos, blueprint, param, param_value): name=param.name, value=param.value) + def blueprint_param(module): margs = module.params @@ -295,15 +298,15 @@ def blueprint_param(module): # -------------------------------------------------------------------- try: blueprint = find_collection_item(aos.Blueprints, - item_name=margs['blueprint'], - item_id=margs['blueprint']) + item_name=margs['blueprint'], + item_id=margs['blueprint']) except: module.fail_json(msg="Unable to find the Blueprint based on name or ID, something went wrong") if blueprint.exists is False: module.fail_json(msg='Blueprint %s does not exist.\n' - 'known blueprints are [%s]'% - (margs['blueprint'],','.join(aos.Blueprints.names))) + 'known blueprints are [%s]' % + (margs['blueprint'], ','.join(aos.Blueprints.names))) # -------------------------------------------------------------------- # If get_param_list is defined, build the list of supported parameters @@ -316,8 +319,8 @@ def blueprint_param(module): params_list[param] = blueprint.params[param].info module.exit_json(changed=False, - blueprint= blueprint.name, - params_list=params_list ) + blueprint=blueprint.name, + params_list=params_list) # -------------------------------------------------------------------- # Check Param name, return an error if not supported by this blueprint @@ -325,7 +328,7 @@ def blueprint_param(module): if margs['name'] in blueprint.params.names: param = blueprint.params[margs['name']] else: - module.fail_json(msg='unable to access param %s' % margs['name'] ) + module.fail_json(msg='unable to access param %s' % margs['name']) # -------------------------------------------------------------------- # Check if param_value needs to be converted to an object @@ -350,6 +353,7 @@ def blueprint_param(module): blueprint_param_present(module, aos, blueprint, param, param_value) + def main(): module = AnsibleModule( argument_spec=dict( @@ -359,7 +363,7 @@ def main(): name=dict(required=False), value=dict(required=False, type="dict"), param_map=dict(required=False), - state=dict( choices=['present', 'absent'], default='present') + state=dict(choices=['present', 'absent'], default='present') ), supports_check_mode=True ) diff --git a/lib/ansible/modules/network/aos/aos_blueprint_virtnet.py b/lib/ansible/modules/network/aos/aos_blueprint_virtnet.py index 12441494b9..62937efbca 100644 --- a/lib/ansible/modules/network/aos/aos_blueprint_virtnet.py +++ b/lib/ansible/modules/network/aos/aos_blueprint_virtnet.py @@ -88,6 +88,7 @@ from ansible.module_utils.pycompat24 import get_exception from ansible.module_utils.basic import AnsibleModule from ansible.module_utils.network.aos.aos import get_aos_session, find_collection_item, do_load_resource, check_aos_version, content_to_dict + def ensure_present(module, aos, blueprint, virtnet): # if exist already return tru @@ -104,7 +105,7 @@ def ensure_present(module, aos, blueprint, virtnet): virtnet.create(module.params['content']) except: e = get_exception() - module.fail_json(msg="unable to create virtual-network : %r" % e ) + module.fail_json(msg="unable to create virtual-network : %r" % e) module.exit_json(changed=True, blueprint=blueprint.name, @@ -121,7 +122,7 @@ def ensure_absent(module, aos, blueprint, virtnet): virtnet.delete() except: e = get_exception() - module.fail_json(msg="unable to delete virtual-network %s : %r" % (virtnet.name, e) ) + module.fail_json(msg="unable to delete virtual-network %s : %r" % (virtnet.name, e)) module.exit_json(changed=True, blueprint=blueprint.name) @@ -130,6 +131,7 @@ def ensure_absent(module, aos, blueprint, virtnet): module.exit_json(changed=False, blueprint=blueprint.name) + def blueprint_virtnet(module): margs = module.params @@ -147,22 +149,22 @@ def blueprint_virtnet(module): # -------------------------------------------------------------------- try: blueprint = find_collection_item(aos.Blueprints, - item_name=margs['blueprint'], - item_id=margs['blueprint']) + item_name=margs['blueprint'], + item_id=margs['blueprint']) except: module.fail_json(msg="Unable to find the Blueprint based on name or ID, something went wrong") if blueprint.exists is False: module.fail_json(msg='Blueprint %s does not exist.\n' - 'known blueprints are [%s]'% - (margs['blueprint'],','.join(aos.Blueprints.names))) + 'known blueprints are [%s]' % + (margs['blueprint'], ','.join(aos.Blueprints.names))) # -------------------------------------------------------------------- # Convert "content" to dict and extract name # -------------------------------------------------------------------- if margs['content'] is not None: - content = content_to_dict(module, margs['content'] ) + content = content_to_dict(module, margs['content']) if 'display_name' in content.keys(): item_name = content['display_name'] @@ -179,7 +181,7 @@ def blueprint_virtnet(module): virtnet = blueprint.VirtualNetworks[item_name] except: module.fail_json(msg="Something went wrong while trying to find Virtual Network %s in blueprint %s" - % ( item_name, blueprint.name )) + % (item_name, blueprint.name)) # -------------------------------------------------------------------- # Proceed based on State value @@ -198,11 +200,11 @@ def main(): argument_spec=dict( session=dict(required=True, type="dict"), blueprint=dict(required=True), - name=dict(required=False ), + name=dict(required=False), content=dict(required=False, type="json"), state=dict(choices=['present', 'absent'], default='present') ), - mutually_exclusive = [('name', 'content')], + mutually_exclusive=[('name', 'content')], required_one_of=[('name', 'content')], supports_check_mode=True ) diff --git a/lib/ansible/modules/network/aos/aos_device.py b/lib/ansible/modules/network/aos/aos_device.py index a527fe6921..717083d774 100644 --- a/lib/ansible/modules/network/aos/aos_device.py +++ b/lib/ansible/modules/network/aos/aos_device.py @@ -132,7 +132,7 @@ def aos_device_normal(module, aos, dev): value=dev.value) else: # Check if the device is online - if dev.state in ('OOS-READY','IS-READY'): + if dev.state in ('OOS-READY', 'IS-READY'): module.exit_json(changed=False, name=dev.name, id=dev.id, @@ -140,6 +140,7 @@ def aos_device_normal(module, aos, dev): else: module.fail_json(msg="Device is in '%s' state" % dev.state) + def aos_device(module): margs = module.params @@ -161,8 +162,8 @@ def aos_device(module): # Find Object if available based on ID or Name #---------------------------------------------------- dev = find_collection_item(aos.Devices, - item_name=item_name, - item_id=item_id) + item_name=item_name, + item_id=item_id) if dev.exists is False: module.fail_json(msg="unknown device '%s'" % margs['name']) @@ -189,6 +190,7 @@ def aos_device(module): if margs['state'] == 'normal': aos_device_normal(module, aos, dev) + def main(): module = AnsibleModule( @@ -196,12 +198,12 @@ def main(): session=dict(required=True, type="dict"), name=dict(required=False), id=dict(required=False), - state=dict( choices=['normal'], - default='normal'), - approve=dict( required=False, type='bool' ), - location=dict( required=False, default='') + state=dict(choices=['normal'], + default='normal'), + approve=dict(required=False, type='bool'), + location=dict(required=False, default='') ), - mutually_exclusive = [('name', 'id')], + mutually_exclusive=[('name', 'id')], required_one_of=[('name', 'id')], supports_check_mode=True ) diff --git a/lib/ansible/modules/network/aos/aos_external_router.py b/lib/ansible/modules/network/aos/aos_external_router.py index fc0abeac5e..e693811aab 100644 --- a/lib/ansible/modules/network/aos/aos_external_router.py +++ b/lib/ansible/modules/network/aos/aos_external_router.py @@ -155,16 +155,18 @@ from ansible.module_utils.network.aos.aos import get_aos_session, find_collectio def create_new_ext_router(module, my_ext_router, name, loopback, asn): # Create value - datum = dict(display_name=name, address=loopback, asn=asn ) + datum = dict(display_name=name, address=loopback, asn=asn) my_ext_router.datum = datum - ## Write to AOS + # Write to AOS return my_ext_router.write() ######################################################### # State Processing ######################################################### + + def ext_router_absent(module, aos, my_ext_router): margs = module.params @@ -174,7 +176,7 @@ def ext_router_absent(module, aos, my_ext_router): module.exit_json(changed=False, name=margs['name'], id=margs['id'], - value={} ) + value={}) # If not in check mode, delete External Router if not module.check_mode: @@ -185,10 +187,11 @@ def ext_router_absent(module, aos, my_ext_router): except: module.fail_json(msg="An error occurred, while trying to delete the External Router") - module.exit_json( changed=True, - name=my_ext_router.name, - id=my_ext_router.id, - value={} ) + module.exit_json(changed=True, + name=my_ext_router.name, + id=my_ext_router.id, + value={}) + def ext_router_present(module, aos, my_ext_router): @@ -210,16 +213,15 @@ def ext_router_present(module, aos, my_ext_router): my_ext_router, margs['name'], margs['loopback'], - margs['asn'] ) + margs['asn']) my_ext_router = my_new_ext_router except: module.fail_json(msg="An error occurred while trying to create a new External Router") - - module.exit_json( changed=True, - name=my_ext_router.name, - id=my_ext_router.id, - value=my_ext_router.value ) + module.exit_json(changed=True, + name=my_ext_router.name, + id=my_ext_router.id, + value=my_ext_router.value) # if external Router already exist, check if loopback and ASN are the same # if same just return the object and report change false @@ -249,14 +251,16 @@ def ext_router_present(module, aos, my_ext_router): if int(asn) != int(my_ext_router.value['asn']): module.fail_json(msg="my_ext_router already exist but ASN is different, currently not supported to update a module") - module.exit_json( changed=False, - name=my_ext_router.name, - id=my_ext_router.id, - value=my_ext_router.value ) + module.exit_json(changed=False, + name=my_ext_router.name, + id=my_ext_router.id, + value=my_ext_router.value) ######################################################### # Main Function ######################################################### + + def ext_router(module): margs = module.params @@ -271,7 +275,7 @@ def ext_router(module): if margs['content'] is not None: - content = content_to_dict(module, margs['content'] ) + content = content_to_dict(module, margs['content']) if 'display_name' in content.keys(): item_name = content['display_name'] @@ -289,8 +293,8 @@ def ext_router(module): #---------------------------------------------------- try: my_ext_router = find_collection_item(aos.ExternalRouters, - item_name=item_name, - item_id=item_id) + item_name=item_name, + item_id=item_id) except: module.fail_json(msg="Unable to find the IP Pool based on name or ID, something went wrong") @@ -305,20 +309,21 @@ def ext_router(module): ext_router_present(module, aos, my_ext_router) + def main(): module = AnsibleModule( argument_spec=dict( session=dict(required=True, type="dict"), - name=dict(required=False ), - id=dict(required=False ), + name=dict(required=False), + id=dict(required=False), content=dict(required=False, type="json"), - state=dict( required=False, - choices=['present', 'absent'], - default="present"), + state=dict(required=False, + choices=['present', 'absent'], + default="present"), loopback=dict(required=False), asn=dict(required=False) ), - mutually_exclusive = [('name', 'id', 'content')], + mutually_exclusive=[('name', 'id', 'content')], required_one_of=[('name', 'id', 'content')], supports_check_mode=True ) diff --git a/lib/ansible/modules/network/aos/aos_ip_pool.py b/lib/ansible/modules/network/aos/aos_ip_pool.py index 8a8e35c681..41c7e16754 100644 --- a/lib/ansible/modules/network/aos/aos_ip_pool.py +++ b/lib/ansible/modules/network/aos/aos_ip_pool.py @@ -167,6 +167,7 @@ import json from ansible.module_utils.basic import AnsibleModule from ansible.module_utils.network.aos.aos import get_aos_session, find_collection_item, do_load_resource, check_aos_version, content_to_dict + def get_list_of_subnets(ip_pool): subnets = [] @@ -175,6 +176,7 @@ def get_list_of_subnets(ip_pool): return subnets + def create_new_ip_pool(ip_pool, name, subnets): # Create value @@ -184,12 +186,14 @@ def create_new_ip_pool(ip_pool, name, subnets): ip_pool.datum = datum - ## Write to AOS + # Write to AOS return ip_pool.write() ######################################################### # State Processing ######################################################### + + def ip_pool_absent(module, aos, my_pool): margs = module.params @@ -198,7 +202,7 @@ def ip_pool_absent(module, aos, my_pool): if my_pool.exists is False: module.exit_json(changed=False, name=margs['name'], id='', value={}) - ## Check if object is currently in Use or Not + # Check if object is currently in Use or Not # If in Use, return an error if my_pool.value: if my_pool.value['status'] != 'not_in_use': @@ -213,10 +217,11 @@ def ip_pool_absent(module, aos, my_pool): except: module.fail_json(msg="An error occurred, while trying to delete the IP Pool") - module.exit_json( changed=True, - name=my_pool.name, - id=my_pool.id, - value={} ) + module.exit_json(changed=True, + name=my_pool.name, + id=my_pool.id, + value={}) + def ip_pool_present(module, aos, my_pool): @@ -248,24 +253,26 @@ def ip_pool_present(module, aos, my_pool): except: module.fail_json(msg="An error occurred while trying to create a new IP Pool ") - module.exit_json( changed=True, - name=my_pool.name, - id=my_pool.id, - value=my_pool.value ) + module.exit_json(changed=True, + name=my_pool.name, + id=my_pool.id, + value=my_pool.value) # if pool already exist, check if list of network is the same # if same just return the object and report change false if set(get_list_of_subnets(my_pool)) == set(margs['subnets']): - module.exit_json( changed=False, - name=my_pool.name, - id=my_pool.id, - value=my_pool.value ) + module.exit_json(changed=False, + name=my_pool.name, + id=my_pool.id, + value=my_pool.value) else: module.fail_json(msg="ip_pool already exist but value is different, currently not supported to update a module") ######################################################### # Main Function ######################################################### + + def ip_pool(module): margs = module.params @@ -280,7 +287,7 @@ def ip_pool(module): if margs['content'] is not None: - content = content_to_dict(module, margs['content'] ) + content = content_to_dict(module, margs['content']) if 'display_name' in content.keys(): item_name = content['display_name'] @@ -298,8 +305,8 @@ def ip_pool(module): #---------------------------------------------------- try: my_pool = find_collection_item(aos.IpPools, - item_name=item_name, - item_id=item_id) + item_name=item_name, + item_id=item_id) except: module.fail_json(msg="Unable to find the IP Pool based on name or ID, something went wrong") @@ -314,19 +321,20 @@ def ip_pool(module): ip_pool_present(module, aos, my_pool) + def main(): module = AnsibleModule( argument_spec=dict( session=dict(required=True, type="dict"), - name=dict(required=False ), - id=dict(required=False ), + name=dict(required=False), + id=dict(required=False), content=dict(required=False, type="json"), - state=dict( required=False, - choices=['present', 'absent'], - default="present"), + state=dict(required=False, + choices=['present', 'absent'], + default="present"), subnets=dict(required=False, type="list") ), - mutually_exclusive = [('name', 'id', 'content')], + mutually_exclusive=[('name', 'id', 'content')], required_one_of=[('name', 'id', 'content')], supports_check_mode=True ) diff --git a/lib/ansible/modules/network/aos/aos_logical_device.py b/lib/ansible/modules/network/aos/aos_logical_device.py index abcf99a333..76412dfc33 100644 --- a/lib/ansible/modules/network/aos/aos_logical_device.py +++ b/lib/ansible/modules/network/aos/aos_logical_device.py @@ -134,6 +134,8 @@ from ansible.module_utils.network.aos.aos import get_aos_session, find_collectio ######################################################### # State Processing ######################################################### + + def logical_device_absent(module, aos, my_logical_dev): margs = module.params @@ -154,10 +156,11 @@ def logical_device_absent(module, aos, my_logical_dev): except: module.fail_json(msg="An error occurred, while trying to delete the Logical Device") - module.exit_json( changed=True, - name=my_logical_dev.name, - id=my_logical_dev.id, - value={} ) + module.exit_json(changed=True, + name=my_logical_dev.name, + id=my_logical_dev.id, + value={}) + def logical_device_present(module, aos, my_logical_dev): @@ -174,14 +177,16 @@ def logical_device_present(module, aos, my_logical_dev): if my_logical_dev.exists is False and 'content' not in margs.keys(): module.fail_json(msg="'content' is mandatory for module that don't exist currently") - module.exit_json( changed=False, - name=my_logical_dev.name, - id=my_logical_dev.id, - value=my_logical_dev.value ) + module.exit_json(changed=False, + name=my_logical_dev.name, + id=my_logical_dev.id, + value=my_logical_dev.value) ######################################################### # Main Function ######################################################### + + def logical_device(module): margs = module.params @@ -196,7 +201,7 @@ def logical_device(module): if margs['content'] is not None: - content = content_to_dict(module, margs['content'] ) + content = content_to_dict(module, margs['content']) if 'display_name' in content.keys(): item_name = content['display_name'] @@ -227,18 +232,19 @@ def logical_device(module): logical_device_present(module, aos, my_logical_dev) + def main(): module = AnsibleModule( argument_spec=dict( session=dict(required=True, type="dict"), - name=dict(required=False ), - id=dict(required=False ), + name=dict(required=False), + id=dict(required=False), content=dict(required=False, type="json"), - state=dict( required=False, - choices=['present', 'absent'], - default="present") + state=dict(required=False, + choices=['present', 'absent'], + default="present") ), - mutually_exclusive = [('name', 'id', 'content')], + mutually_exclusive=[('name', 'id', 'content')], required_one_of=[('name', 'id', 'content')], supports_check_mode=True ) diff --git a/lib/ansible/modules/network/aos/aos_logical_device_map.py b/lib/ansible/modules/network/aos/aos_logical_device_map.py index 474feb6871..44d089b784 100644 --- a/lib/ansible/modules/network/aos/aos_logical_device_map.py +++ b/lib/ansible/modules/network/aos/aos_logical_device_map.py @@ -154,6 +154,8 @@ from ansible.module_utils.network.aos.aos import get_aos_session, find_collectio ######################################################### # State Processing ######################################################### + + def logical_device_map_absent(module, aos, my_log_dev_map): margs = module.params @@ -172,10 +174,11 @@ def logical_device_map_absent(module, aos, my_log_dev_map): except: module.fail_json(msg="An error occurred, while trying to delete the Logical Device Map") - module.exit_json( changed=True, - name=my_log_dev_map.name, - id=my_log_dev_map.id, - value={} ) + module.exit_json(changed=True, + name=my_log_dev_map.name, + id=my_log_dev_map.id, + value={}) + def logical_device_map_present(module, aos, my_log_dev_map): @@ -194,14 +197,16 @@ def logical_device_map_present(module, aos, my_log_dev_map): if my_log_dev_map.exists is False and 'content' not in margs.keys(): module.fail_json(msg="'Content' is mandatory for module that don't exist currently") - module.exit_json( changed=False, - name=my_log_dev_map.name, - id=my_log_dev_map.id, - value=my_log_dev_map.value ) + module.exit_json(changed=False, + name=my_log_dev_map.name, + id=my_log_dev_map.id, + value=my_log_dev_map.value) ######################################################### # Main Function ######################################################### + + def logical_device_map(module): margs = module.params @@ -216,7 +221,7 @@ def logical_device_map(module): if margs['content'] is not None: - content = content_to_dict(module, margs['content'] ) + content = content_to_dict(module, margs['content']) if 'display_name' in content.keys(): item_name = content['display_name'] @@ -234,8 +239,8 @@ def logical_device_map(module): #---------------------------------------------------- try: my_log_dev_map = find_collection_item(aos.LogicalDeviceMaps, - item_name=item_name, - item_id=item_id) + item_name=item_name, + item_id=item_id) except: module.fail_json(msg="Unable to find the Logical Device Map based on name or ID, something went wrong") @@ -250,18 +255,19 @@ def logical_device_map(module): logical_device_map_present(module, aos, my_log_dev_map) + def main(): module = AnsibleModule( argument_spec=dict( session=dict(required=True, type="dict"), - name=dict(required=False ), - id=dict(required=False ), + name=dict(required=False), + id=dict(required=False), content=dict(required=False, type="json"), - state=dict( required=False, - choices=['present', 'absent'], - default="present") + state=dict(required=False, + choices=['present', 'absent'], + default="present") ), - mutually_exclusive = [('name', 'id', 'content')], + mutually_exclusive=[('name', 'id', 'content')], required_one_of=[('name', 'id', 'content')], supports_check_mode=True ) diff --git a/lib/ansible/modules/network/aos/aos_rack_type.py b/lib/ansible/modules/network/aos/aos_rack_type.py index e6b4963f32..3be22595c1 100644 --- a/lib/ansible/modules/network/aos/aos_rack_type.py +++ b/lib/ansible/modules/network/aos/aos_rack_type.py @@ -133,6 +133,8 @@ from ansible.module_utils.network.aos.aos import get_aos_session, find_collectio ######################################################### # State Processing ######################################################### + + def rack_type_absent(module, aos, my_rack_type): margs = module.params @@ -151,10 +153,11 @@ def rack_type_absent(module, aos, my_rack_type): except: module.fail_json(msg="An error occurred, while trying to delete the Rack Type") - module.exit_json( changed=True, - name=my_rack_type.name, - id=my_rack_type.id, - value={} ) + module.exit_json(changed=True, + name=my_rack_type.name, + id=my_rack_type.id, + value={}) + def rack_type_present(module, aos, my_rack_type): @@ -171,14 +174,16 @@ def rack_type_present(module, aos, my_rack_type): if my_rack_type.exists is False and 'content' not in margs.keys(): module.fail_json(msg="'content' is mandatory for module that don't exist currently") - module.exit_json( changed=False, - name=my_rack_type.name, - id=my_rack_type.id, - value=my_rack_type.value ) + module.exit_json(changed=False, + name=my_rack_type.name, + id=my_rack_type.id, + value=my_rack_type.value) ######################################################### # Main Function ######################################################### + + def rack_type(module): margs = module.params @@ -193,7 +198,7 @@ def rack_type(module): if margs['content'] is not None: - content = content_to_dict(module, margs['content'] ) + content = content_to_dict(module, margs['content']) if 'display_name' in content.keys(): item_name = content['display_name'] @@ -210,8 +215,8 @@ def rack_type(module): # Find Object if available based on ID or Name #---------------------------------------------------- my_rack_type = find_collection_item(aos.RackTypes, - item_name=item_name, - item_id=item_id) + item_name=item_name, + item_id=item_id) #---------------------------------------------------- # Proceed based on State value @@ -224,18 +229,19 @@ def rack_type(module): rack_type_present(module, aos, my_rack_type) + def main(): module = AnsibleModule( argument_spec=dict( session=dict(required=True, type="dict"), - name=dict(required=False ), - id=dict(required=False ), + name=dict(required=False), + id=dict(required=False), content=dict(required=False, type="json"), - state=dict( required=False, - choices=['present', 'absent'], - default="present") + state=dict(required=False, + choices=['present', 'absent'], + default="present") ), - mutually_exclusive = [('name', 'id', 'content')], + mutually_exclusive=[('name', 'id', 'content')], required_one_of=[('name', 'id', 'content')], supports_check_mode=True ) diff --git a/lib/ansible/modules/network/aos/aos_template.py b/lib/ansible/modules/network/aos/aos_template.py index 26f19dc5f7..c87dee2ca1 100644 --- a/lib/ansible/modules/network/aos/aos_template.py +++ b/lib/ansible/modules/network/aos/aos_template.py @@ -143,6 +143,8 @@ from ansible.module_utils.network.aos.aos import get_aos_session, find_collectio ######################################################### # State Processing ######################################################### + + def template_absent(module, aos, my_template): margs = module.params @@ -163,10 +165,11 @@ def template_absent(module, aos, my_template): except: module.fail_json(msg="An error occurred, while trying to delete the Template") - module.exit_json( changed=True, - name=my_template.name, - id=my_template.id, - value={} ) + module.exit_json(changed=True, + name=my_template.name, + id=my_template.id, + value={}) + def template_present(module, aos, my_template): @@ -186,10 +189,10 @@ def template_present(module, aos, my_template): module.fail_json(msg="'content' is mandatory for module that don't exist currently") # if module already exist, just return it - module.exit_json( changed=False, - name=my_template.name, - id=my_template.id, - value=my_template.value ) + module.exit_json(changed=False, + name=my_template.name, + id=my_template.id, + value=my_template.value) ######################################################### @@ -209,7 +212,7 @@ def aos_template(module): if margs['content'] is not None: - content = content_to_dict(module, margs['content'] ) + content = content_to_dict(module, margs['content']) if 'display_name' in content.keys(): item_name = content['display_name'] @@ -227,8 +230,8 @@ def aos_template(module): #---------------------------------------------------- try: my_template = find_collection_item(aos.DesignTemplates, - item_name=item_name, - item_id=item_id) + item_name=item_name, + item_id=item_id) except: module.fail_json(msg="Unable to find the IP Pool based on name or ID, something went wrong") @@ -243,18 +246,19 @@ def aos_template(module): template_present(module, aos, my_template) + def main(): module = AnsibleModule( argument_spec=dict( session=dict(required=True, type="dict"), - name=dict(required=False ), - id=dict(required=False ), + name=dict(required=False), + id=dict(required=False), content=dict(required=False, type="json"), - state=dict( required=False, - choices=['present', 'absent'], - default="present") + state=dict(required=False, + choices=['present', 'absent'], + default="present") ), - mutually_exclusive = [('name', 'id', 'content')], + mutually_exclusive=[('name', 'id', 'content')], required_one_of=[('name', 'id', 'content')], supports_check_mode=True ) diff --git a/lib/ansible/modules/network/asa/asa_acl.py b/lib/ansible/modules/network/asa/asa_acl.py index 82ded9985a..0f5c51d5c5 100644 --- a/lib/ansible/modules/network/asa/asa_acl.py +++ b/lib/ansible/modules/network/asa/asa_acl.py @@ -151,6 +151,7 @@ def get_acl_config(module, acl_name): return NetworkConfig(indent=1, contents='\n'.join(filtered_config)) + def parse_acl_name(module): first_line = True for line in module.params['lines']: @@ -168,6 +169,7 @@ def parse_acl_name(module): return acl_name + def main(): argument_spec = dict( diff --git a/lib/ansible/modules/network/asa/asa_command.py b/lib/ansible/modules/network/asa/asa_command.py index 9a79f5137e..06b19c77f0 100644 --- a/lib/ansible/modules/network/asa/asa_command.py +++ b/lib/ansible/modules/network/asa/asa_command.py @@ -187,7 +187,6 @@ def main(): msg = 'One or more conditional statements have not be satisfied' module.fail_json(msg=msg, failed_conditions=failed_conditions) - result.update({ 'changed': False, 'stdout': responses, diff --git a/lib/ansible/modules/network/asa/asa_config.py b/lib/ansible/modules/network/asa/asa_config.py index dca23567ca..4ba96aef4f 100644 --- a/lib/ansible/modules/network/asa/asa_config.py +++ b/lib/ansible/modules/network/asa/asa_config.py @@ -201,7 +201,6 @@ from ansible.module_utils.network.common.config import NetworkConfig, dumps from ansible.module_utils._text import to_native - def get_candidate(module): candidate = NetworkConfig(indent=1) if module.params['src']: @@ -211,6 +210,7 @@ def get_candidate(module): candidate.add(module.params['lines'], parents=parents) return candidate + def run(module, result): match = module.params['match'] replace = module.params['replace'] @@ -251,6 +251,7 @@ def run(module, result): run_commands(module, 'write mem') result['changed'] = True + def main(): """ main entry point for module execution """ @@ -293,7 +294,6 @@ def main(): config = None - if module.params['backup']: result['__backup__'] = get_config(module) diff --git a/lib/ansible/modules/network/bigswitch/bigmon_chain.py b/lib/ansible/modules/network/bigswitch/bigmon_chain.py index 49157506f5..2c6f66525e 100755 --- a/lib/ansible/modules/network/bigswitch/bigmon_chain.py +++ b/lib/ansible/modules/network/bigswitch/bigmon_chain.py @@ -80,8 +80,8 @@ def chain(module): controller = module.params['controller'] rest = Rest(module, - {'content-type': 'application/json', 'Cookie': 'session_cookie='+access_token}, - 'https://'+controller+':8443/api/v1/data/controller/applications/bigchain') + {'content-type': 'application/json', 'Cookie': 'session_cookie=' + access_token}, + 'https://' + controller + ':8443/api/v1/data/controller/applications/bigchain') if None in (name, state, controller): module.fail_json(msg='parameter `name` is missing') @@ -115,6 +115,7 @@ def chain(module): else: module.fail_json(msg="error deleting chain '{}': {}".format(name, response.json['description'])) + def main(): module = AnsibleModule( argument_spec=dict( diff --git a/lib/ansible/modules/network/bigswitch/bigmon_policy.py b/lib/ansible/modules/network/bigswitch/bigmon_policy.py index 7b4f0303eb..7aefc644cf 100644 --- a/lib/ansible/modules/network/bigswitch/bigmon_policy.py +++ b/lib/ansible/modules/network/bigswitch/bigmon_policy.py @@ -113,8 +113,8 @@ def policy(module): controller = module.params['controller'] rest = Rest(module, - {'content-type': 'application/json', 'Cookie': 'session_cookie='+access_token}, - 'https://'+controller+':8443/api/v1/data/controller/applications/bigtap') + {'content-type': 'application/json', 'Cookie': 'session_cookie=' + access_token}, + 'https://' + controller + ':8443/api/v1/data/controller/applications/bigtap') if name is None: module.fail_json(msg='parameter `name` is missing') @@ -127,11 +127,11 @@ def policy(module): matching = [policy for policy in response.json if policy['name'] == name and - policy['duration'] == duration and - policy['delivery-packet-count'] == delivery_packet_count and - policy['policy-description'] == policy_description and - policy['action'] == action and - policy['priority'] == priority] + policy['duration'] == duration and + policy['delivery-packet-count'] == delivery_packet_count and + policy['policy-description'] == policy_description and + policy['action'] == action and + policy['priority'] == priority] if matching: config_present = True @@ -143,9 +143,9 @@ def policy(module): module.exit_json(changed=False) if state in ('present'): - data={'name': name, 'action': action, 'policy-description': policy_description, - 'priority': priority, 'duration': duration, 'start-time': start_time, - 'delivery-packet-count': delivery_packet_count } + data = {'name': name, 'action': action, 'policy-description': policy_description, + 'priority': priority, 'duration': duration, 'start-time': start_time, + 'delivery-packet-count': delivery_packet_count} response = rest.put('policy[name="%s"]' % name, data=data) if response.status_code == 204: @@ -160,6 +160,7 @@ def policy(module): else: module.fail_json(msg="error deleting policy '{}': {}".format(name, response.json['description'])) + def main(): module = AnsibleModule( argument_spec=dict( @@ -168,7 +169,7 @@ def main(): action=dict(choices=['forward', 'drop', 'capture', 'flow-gen'], default='forward'), priority=dict(type='int', default=100), duration=dict(type='int', default=0), - start_time=dict(type='str', default=datetime.datetime.now().isoformat()+'+00:00'), + start_time=dict(type='str', default=datetime.datetime.now().isoformat() + '+00:00'), delivery_packet_count=dict(type='int', default=0), controller=dict(type='str', required=True), state=dict(choices=['present', 'absent'], default='present'), diff --git a/lib/ansible/modules/network/cnos/cnos_bgp.py b/lib/ansible/modules/network/cnos/cnos_bgp.py index f653724e36..eca0d3ec17 100644 --- a/lib/ansible/modules/network/cnos/cnos_bgp.py +++ b/lib/ansible/modules/network/cnos/cnos_bgp.py @@ -469,4 +469,4 @@ def main(): if __name__ == '__main__': - main() + main() diff --git a/lib/ansible/modules/network/cnos/cnos_command.py b/lib/ansible/modules/network/cnos/cnos_command.py index 85646957f7..1ee342ce9d 100644 --- a/lib/ansible/modules/network/cnos/cnos_command.py +++ b/lib/ansible/modules/network/cnos/cnos_command.py @@ -163,4 +163,4 @@ def main(): module.fail_json(msg=errorMsg) if __name__ == '__main__': - main() + main() diff --git a/lib/ansible/modules/network/cnos/cnos_conditional_command.py b/lib/ansible/modules/network/cnos/cnos_conditional_command.py index ad98f85218..8360ac7455 100644 --- a/lib/ansible/modules/network/cnos/cnos_conditional_command.py +++ b/lib/ansible/modules/network/cnos/cnos_conditional_command.py @@ -190,4 +190,4 @@ def main(): module.fail_json(msg=errorMsg) if __name__ == '__main__': - main() + main() diff --git a/lib/ansible/modules/network/cnos/cnos_image.py b/lib/ansible/modules/network/cnos/cnos_image.py index d17f2ef9e3..77a4c7c5e6 100644 --- a/lib/ansible/modules/network/cnos/cnos_image.py +++ b/lib/ansible/modules/network/cnos/cnos_image.py @@ -228,4 +228,4 @@ def main(): module.fail_json(msg=errorMsg) if __name__ == '__main__': - main() + main() diff --git a/lib/ansible/modules/network/cnos/cnos_interface.py b/lib/ansible/modules/network/cnos/cnos_interface.py index 8ee5fb619d..538f2e62f3 100644 --- a/lib/ansible/modules/network/cnos/cnos_interface.py +++ b/lib/ansible/modules/network/cnos/cnos_interface.py @@ -585,4 +585,4 @@ def main(): if __name__ == '__main__': - main() + main() diff --git a/lib/ansible/modules/network/cnos/cnos_portchannel.py b/lib/ansible/modules/network/cnos/cnos_portchannel.py index 883602b254..af4a86b867 100644 --- a/lib/ansible/modules/network/cnos/cnos_portchannel.py +++ b/lib/ansible/modules/network/cnos/cnos_portchannel.py @@ -521,4 +521,4 @@ def main(): module.fail_json(msg=errorMsg) if __name__ == '__main__': - main() + main() diff --git a/lib/ansible/modules/network/cnos/cnos_rollback.py b/lib/ansible/modules/network/cnos/cnos_rollback.py index 2ad78cd7f6..46fea95198 100644 --- a/lib/ansible/modules/network/cnos/cnos_rollback.py +++ b/lib/ansible/modules/network/cnos/cnos_rollback.py @@ -273,4 +273,4 @@ def main(): module.fail_json(msg=errorMsg) if __name__ == '__main__': - main() + main() diff --git a/lib/ansible/modules/network/cnos/cnos_vlan.py b/lib/ansible/modules/network/cnos/cnos_vlan.py index 09ccaaf27a..7ef652ddfc 100644 --- a/lib/ansible/modules/network/cnos/cnos_vlan.py +++ b/lib/ansible/modules/network/cnos/cnos_vlan.py @@ -296,4 +296,4 @@ def main(): module.fail_json(msg=errorMsg) if __name__ == '__main__': - main() + main() diff --git a/lib/ansible/modules/network/cumulus/_cl_bond.py b/lib/ansible/modules/network/cumulus/_cl_bond.py index f9050f47a5..d2db8306af 100644 --- a/lib/ansible/modules/network/cumulus/_cl_bond.py +++ b/lib/ansible/modules/network/cumulus/_cl_bond.py @@ -288,6 +288,7 @@ def conv_array_to_str(_value): return ' '.join(_value) return _value + def build_generic_attr(module, _attr): _value = module.params.get(_attr) _value = conv_bool_to_str(_value) diff --git a/lib/ansible/modules/network/cumulus/_cl_img_install.py b/lib/ansible/modules/network/cumulus/_cl_img_install.py index 25072c90d2..2265847537 100644 --- a/lib/ansible/modules/network/cumulus/_cl_img_install.py +++ b/lib/ansible/modules/network/cumulus/_cl_img_install.py @@ -187,7 +187,6 @@ def check_fw_print_env(module, slot_num): return m0.group(1) - def get_primary_slot_num(module): cmd = None if platform.machine() == 'ppc': @@ -264,7 +263,7 @@ def check_sw_version(module): if 'active' in slot: _msg = "Version %s is installed in the active slot" \ % (_version) - module.exit_json(changed=False, msg=_msg) + module.exit_json(changed=False, msg=_msg) else: _msg = "Version " + _version + \ " is installed in the alternate slot. " diff --git a/lib/ansible/modules/network/cumulus/_cl_license.py b/lib/ansible/modules/network/cumulus/_cl_license.py index a504865a23..175e79fb68 100644 --- a/lib/ansible/modules/network/cumulus/_cl_license.py +++ b/lib/ansible/modules/network/cumulus/_cl_license.py @@ -103,7 +103,7 @@ msg: from ansible.module_utils.basic import AnsibleModule -CL_LICENSE_PATH='/usr/cumulus/bin/cl-license' +CL_LICENSE_PATH = '/usr/cumulus/bin/cl-license' def install_license(module): diff --git a/lib/ansible/modules/network/cumulus/_cl_ports.py b/lib/ansible/modules/network/cumulus/_cl_ports.py index 21c3f3bf05..41e329ae29 100644 --- a/lib/ansible/modules/network/cumulus/_cl_ports.py +++ b/lib/ansible/modules/network/cumulus/_cl_ports.py @@ -99,7 +99,7 @@ def hash_existing_ports_conf(module): except IOError as e: _msg = "Failed to open %s: %s" % (PORTS_CONF, to_native(e)) module.fail_json(msg=_msg) - return # for testing only should return on module.fail_json + return # for testing only should return on module.fail_json for _line in existing_ports_conf: _m0 = re.match(r'^(\d+)=(\w+)', _line) @@ -129,7 +129,7 @@ def generate_new_ports_conf_hash(module): port_setting else: int_range = map(int, port_range_str) - portnum_range = range(int_range[0], int_range[1]+1) + portnum_range = range(int_range[0], int_range[1] + 1) for i in portnum_range: new_ports_conf_hash[i] = port_setting module.new_ports_hash = new_ports_conf_hash @@ -160,6 +160,7 @@ def make_copy_of_orig_ports_conf(module): module.fail_json(msg=_msg) return # for testing only + def write_to_ports_conf(module): """ use tempfile to first write out config in temp file diff --git a/lib/ansible/modules/network/cumulus/nclu.py b/lib/ansible/modules/network/cumulus/nclu.py index d14a825025..0f70a09dea 100644 --- a/lib/ansible/modules/network/cumulus/nclu.py +++ b/lib/ansible/modules/network/cumulus/nclu.py @@ -99,7 +99,7 @@ from ansible.module_utils.basic import AnsibleModule def command_helper(module, command, errmsg=None): """Run a command, catch any nclu errors""" - (_rc, output, _err) = module.run_command("/usr/bin/net %s"%command) + (_rc, output, _err) = module.run_command("/usr/bin/net %s" % command) if _rc or 'ERROR' in output or 'ERROR' in _err: module.fail_json(msg=errmsg or output) return str(output) @@ -141,7 +141,7 @@ def run_nclu(module, command_list, command_string, commit, atomic, abort, descri # Run all of the net commands output_lines = [] for line in commands: - output_lines += [command_helper(module, line.strip(), "Failed on line %s"%line)] + output_lines += [command_helper(module, line.strip(), "Failed on line %s" % line)] output = "\n".join(output_lines) # If pending changes changed, report a change. @@ -153,7 +153,7 @@ def run_nclu(module, command_list, command_string, commit, atomic, abort, descri # Do the commit. if do_commit: - result = command_helper(module, "commit description '%s'"%description) + result = command_helper(module, "commit description '%s'" % description) if "commit ignored" in result: _changed = False command_helper(module, "abort") @@ -165,12 +165,12 @@ def run_nclu(module, command_list, command_string, commit, atomic, abort, descri def main(testing=False): module = AnsibleModule(argument_spec=dict( - commands = dict(required=False, type='list'), - template = dict(required=False, type='str'), - description = dict(required=False, type='str', default="Ansible-originated commit"), - abort = dict(required=False, type='bool', default=False), - commit = dict(required=False, type='bool', default=False), - atomic = dict(required=False, type='bool', default=False)), + commands=dict(required=False, type='list'), + template=dict(required=False, type='str'), + description=dict(required=False, type='str', default="Ansible-originated commit"), + abort=dict(required=False, type='bool', default=False), + commit=dict(required=False, type='bool', default=False), + atomic=dict(required=False, type='bool', default=False)), mutually_exclusive=[('commands', 'template'), ('commit', 'atomic'), ('abort', 'atomic')] diff --git a/lib/ansible/modules/network/eos/eos_banner.py b/lib/ansible/modules/network/eos/eos_banner.py index 0fbaa393af..d080fef06c 100644 --- a/lib/ansible/modules/network/eos/eos_banner.py +++ b/lib/ansible/modules/network/eos/eos_banner.py @@ -126,6 +126,7 @@ def map_obj_to_commands(updates, module): return commands + def map_config_to_obj(module): output = run_commands(module, ['show banner %s' % module.params['banner']]) obj = {'banner': module.params['banner'], 'state': 'absent'} @@ -144,6 +145,7 @@ def map_config_to_obj(module): obj['state'] = 'present' return obj + def map_params_to_obj(module): text = module.params['text'] if text: @@ -155,6 +157,7 @@ def map_params_to_obj(module): 'state': module.params['state'] } + def main(): """ main entry point for module execution """ diff --git a/lib/ansible/modules/network/eos/eos_command.py b/lib/ansible/modules/network/eos/eos_command.py index 3fe90bd72a..7d4c4cd18b 100644 --- a/lib/ansible/modules/network/eos/eos_command.py +++ b/lib/ansible/modules/network/eos/eos_command.py @@ -146,6 +146,7 @@ from ansible.module_utils.network.eos.eos import eos_argument_spec, check_args VALID_KEYS = ['command', 'output', 'prompt', 'response'] + def to_lines(stdout): lines = list() for item in stdout: @@ -154,6 +155,7 @@ def to_lines(stdout): lines.append(item) return lines + def parse_commands(module, warnings): spec = dict( command=dict(key=True), @@ -176,12 +178,14 @@ def parse_commands(module, warnings): return commands + def to_cli(obj): cmd = obj['command'] if obj.get('output') == 'json': cmd += ' | json' return cmd + def main(): """entry point for module execution """ diff --git a/lib/ansible/modules/network/eos/eos_config.py b/lib/ansible/modules/network/eos/eos_config.py index f34334e104..88b3fbc4f0 100644 --- a/lib/ansible/modules/network/eos/eos_config.py +++ b/lib/ansible/modules/network/eos/eos_config.py @@ -403,7 +403,6 @@ def main(): result['changed'] = True - running_config = None startup_config = None diff --git a/lib/ansible/modules/network/eos/eos_eapi.py b/lib/ansible/modules/network/eos/eos_eapi.py index f65f4403c6..22d4a8d1d8 100644 --- a/lib/ansible/modules/network/eos/eos_eapi.py +++ b/lib/ansible/modules/network/eos/eos_eapi.py @@ -187,6 +187,7 @@ from ansible.module_utils.network.eos.eos import run_commands, load_config from ansible.module_utils.six import iteritems from ansible.module_utils.network.eos.eos import eos_argument_spec, check_args + def check_transport(module): transport = module.params['transport'] provider_transport = (module.params['provider'] or {}).get('transport') @@ -194,18 +195,22 @@ def check_transport(module): if 'eapi' in (transport, provider_transport): module.fail_json(msg='eos_eapi module is only supported over cli transport') + def validate_http_port(value, module): if not 1 <= value <= 65535: module.fail_json(msg='http_port must be between 1 and 65535') + def validate_https_port(value, module): if not 1 <= value <= 65535: module.fail_json(msg='http_port must be between 1 and 65535') + def validate_local_http_port(value, module): if not 1 <= value <= 65535: module.fail_json(msg='http_port must be between 1 and 65535') + def validate_vrf(value, module): out = run_commands(module, ['show vrf']) configured_vrfs = re.findall(r'^\s+(\w+)(?=\s)', out[0], re.M) @@ -213,11 +218,12 @@ def validate_vrf(value, module): if value not in configured_vrfs: module.fail_json(msg='vrf `%s` is not configured on the system' % value) + def map_obj_to_commands(updates, module, warnings): commands = list() want, have = updates - needs_update = lambda x: want.get(x) is not None and (want.get(x) != have.get(x)) + def needs_update(x): return want.get(x) is not None and (want.get(x) != have.get(x)) def add(cmd): if 'management api http-commands' not in commands: @@ -266,7 +272,6 @@ def map_obj_to_commands(updates, module, warnings): elif want['state'] == 'started': add('no shutdown') - if needs_update('vrf'): add('vrf %s' % want['vrf']) # switching operational vrfs here @@ -278,6 +283,7 @@ def map_obj_to_commands(updates, module, warnings): return commands + def parse_state(data): if data[0]['enabled']: return 'started' @@ -299,6 +305,7 @@ def map_config_to_obj(module): 'state': parse_state(out) } + def map_params_to_obj(module): obj = { 'http': module.params['http'], @@ -320,6 +327,7 @@ def map_params_to_obj(module): return obj + def verify_state(updates, module): want, have = updates @@ -348,6 +356,7 @@ def verify_state(updates, module): if timeout == 0: module.fail_json(msg='timeout expired before eapi running state changed') + def collect_facts(module, result): out = run_commands(module, ['show management api http-commands | json']) facts = dict(eos_eapi_urls=dict()) @@ -359,6 +368,7 @@ def collect_facts(module, result): facts['eos_eapi_urls'][key].append(str(url).strip()) result['ansible_facts'] = facts + def main(): """ main entry point for module execution """ diff --git a/lib/ansible/modules/network/eos/eos_facts.py b/lib/ansible/modules/network/eos/eos_facts.py index e0961abad0..5dc7b20312 100644 --- a/lib/ansible/modules/network/eos/eos_facts.py +++ b/lib/ansible/modules/network/eos/eos_facts.py @@ -142,6 +142,7 @@ from ansible.module_utils.six import iteritems from ansible.module_utils.network.eos.eos import run_commands from ansible.module_utils.network.eos.eos import eos_argument_spec, check_args + class FactsBase(object): COMMANDS = frozenset() @@ -190,6 +191,7 @@ class Default(FactsBase): value = None return dict(image=value) + class Hardware(FactsBase): COMMANDS = [ @@ -218,6 +220,7 @@ class Hardware(FactsBase): memtotal_mb=int(values['memTotal']) / 1024 ) + class Config(FactsBase): COMMANDS = ['show running-config'] @@ -312,6 +315,7 @@ FACT_SUBSETS = dict( VALID_SUBSETS = frozenset(FACT_SUBSETS.keys()) + def main(): """main entry point for module execution """ diff --git a/lib/ansible/modules/network/eos/eos_system.py b/lib/ansible/modules/network/eos/eos_system.py index d8449e5fc2..44fdd1725e 100644 --- a/lib/ansible/modules/network/eos/eos_system.py +++ b/lib/ansible/modules/network/eos/eos_system.py @@ -135,6 +135,7 @@ from ansible.module_utils.network.eos.eos import eos_argument_spec _CONFIGURED_VRFS = None + def has_vrf(module, vrf): global _CONFIGURED_VRFS if _CONFIGURED_VRFS is not None: @@ -144,11 +145,12 @@ def has_vrf(module, vrf): _CONFIGURED_VRFS.append('default') return vrf in _CONFIGURED_VRFS + def map_obj_to_commands(want, have, module): commands = list() state = module.params['state'] - needs_update = lambda x: want.get(x) and (want.get(x) != have.get(x)) + def needs_update(x): return want.get(x) and (want.get(x) != have.get(x)) if state == 'absent': if have['domain_name']: @@ -205,9 +207,9 @@ def map_obj_to_commands(want, have, module): module.fail_json(msg='vrf %s is not configured' % item['vrf']) if item['vrf'] not in ('default', None): values = (item['vrf'], item['server']) - commands.append('no ip name-server vrf %s %s' % values) + commands.append('no ip name-server vrf %s %s' % values) else: - commands.append('no ip name-server %s' % item['server']) + commands.append('no ip name-server %s' % item['server']) # handle name_servers items to be added for item in want['name_servers']: @@ -222,31 +224,36 @@ def map_obj_to_commands(want, have, module): return commands + def parse_hostname(config): match = re.search(r'^hostname (\S+)', config, re.M) if match: return match.group(1) + def parse_domain_name(config): match = re.search(r'^ip domain-name (\S+)', config, re.M) if match: return match.group(1) + def parse_lookup_source(config): objects = list() regex = r'ip domain lookup (?:vrf (\S+) )*source-interface (\S+)' for vrf, intf in re.findall(regex, config, re.M): if len(vrf) == 0: - vrf= None + vrf = None objects.append({'interface': intf, 'vrf': vrf}) return objects + def parse_name_servers(config): objects = list() for vrf, addr in re.findall(r'ip name-server vrf (\S+) (\S+)', config, re.M): objects.append({'server': addr, 'vrf': vrf}) return objects + def map_config_to_obj(module): config = get_config(module) return { @@ -257,6 +264,7 @@ def map_config_to_obj(module): 'name_servers': parse_name_servers(config) } + def map_params_to_obj(module): obj = { 'hostname': module.params['hostname'], @@ -282,6 +290,7 @@ def map_params_to_obj(module): return obj + def main(): """ main entry point for module execution """ diff --git a/lib/ansible/modules/network/eos/eos_user.py b/lib/ansible/modules/network/eos/eos_user.py index 9bdec10838..be4b895000 100644 --- a/lib/ansible/modules/network/eos/eos_user.py +++ b/lib/ansible/modules/network/eos/eos_user.py @@ -174,8 +174,9 @@ def map_obj_to_commands(updates, module): for update in updates: want, have = update - needs_update = lambda x: want.get(x) and (want.get(x) != have.get(x)) - add = lambda x: commands.append('username %s %s' % (want['name'], x)) + def needs_update(x): return want.get(x) and (want.get(x) != have.get(x)) + + def add(x): return commands.append('username %s %s' % (want['name'], x)) if want['state'] == 'absent': commands.append('no username %s' % want['name']) diff --git a/lib/ansible/modules/network/f5/bigip_monitor_tcp.py b/lib/ansible/modules/network/f5/bigip_monitor_tcp.py index 2923c53428..b0abb48396 100644 --- a/lib/ansible/modules/network/f5/bigip_monitor_tcp.py +++ b/lib/ansible/modules/network/f5/bigip_monitor_tcp.py @@ -548,9 +548,9 @@ class Difference(object): ) elif self.want.timeout is not None: if self.have.interval >= self.want.timeout: - raise F5ModuleError( - "Parameter 'interval' must be less than 'timeout'." - ) + raise F5ModuleError( + "Parameter 'interval' must be less than 'timeout'." + ) elif self.want.interval is not None: if self.want.interval >= self.have.timeout: raise F5ModuleError( diff --git a/lib/ansible/modules/network/f5/bigip_ucs.py b/lib/ansible/modules/network/f5/bigip_ucs.py index ee6919b531..568ace4153 100644 --- a/lib/ansible/modules/network/f5/bigip_ucs.py +++ b/lib/ansible/modules/network/f5/bigip_ucs.py @@ -466,6 +466,7 @@ class V1Manager(BaseManager): * No API to upload UCS files """ + def create_on_device(self): remote_path = "/var/local/ucs" tpath_name = '/var/config/rest/downloads' diff --git a/lib/ansible/modules/network/f5/bigip_wait.py b/lib/ansible/modules/network/f5/bigip_wait.py index 8708ec27a6..3494fe3653 100644 --- a/lib/ansible/modules/network/f5/bigip_wait.py +++ b/lib/ansible/modules/network/f5/bigip_wait.py @@ -121,6 +121,7 @@ class AnsibleF5ClientStub(AnsibleF5Client): the result will replace this work here. """ + def __init__(self, argument_spec=None, supports_check_mode=False, mutually_exclusive=None, required_together=None, required_if=None, required_one_of=None, add_file_common_args=False, diff --git a/lib/ansible/modules/network/fortios/fortios_config.py b/lib/ansible/modules/network/fortios/fortios_config.py index 4761c6c0bc..367046740c 100644 --- a/lib/ansible/modules/network/fortios/fortios_config.py +++ b/lib/ansible/modules/network/fortios/fortios_config.py @@ -76,26 +76,26 @@ from ansible.module_utils.basic import AnsibleModule from ansible.module_utils.network.fortios.fortios import fortios_argument_spec, fortios_required_if from ansible.module_utils.network.fortios.fortios import backup -#check for pyFG lib +# check for pyFG lib try: from pyFG import FortiOS, FortiConfig from pyFG.fortios import logger from pyFG.exceptions import CommandExecutionException, FailedCommit, ForcedCommit - HAS_PYFG=True + HAS_PYFG = True except: - HAS_PYFG=False + HAS_PYFG = False # some blocks don't support update, so remove them -NOT_UPDATABLE_CONFIG_OBJECTS=[ +NOT_UPDATABLE_CONFIG_OBJECTS = [ "vpn certificate local", ] def main(): argument_spec = dict( - src = dict(type='str', default=None), - filter = dict(type='str', default=""), + src=dict(type='str', default=None), + filter=dict(type='str', default=""), ) argument_spec.update(fortios_argument_spec) @@ -114,20 +114,20 @@ def main(): if not HAS_PYFG: module.fail_json(msg='Could not import the python library pyFG required by this module') - #define device - f = FortiOS( module.params['host'], - username=module.params['username'], - password=module.params['password'], - timeout=module.params['timeout'], - vdom=module.params['vdom']) + # define device + f = FortiOS(module.params['host'], + username=module.params['username'], + password=module.params['password'], + timeout=module.params['timeout'], + vdom=module.params['vdom']) - #connect + # connect try: f.open() except: module.fail_json(msg='Error connecting device') - #get config + # get config try: f.load_config(path=module.params['filter']) result['running_config'] = f.running_config.to_text() @@ -135,24 +135,23 @@ def main(): except: module.fail_json(msg='Error reading running config') - #backup config + # backup config if module.params['backup']: backup(module, f.running_config.to_text()) - - #update config + # update config if module.params['src'] is not None: - #store config in str + # store config in str try: conf_str = module.params['src'] f.load_config(in_candidate=True, config_text=conf_str) except: module.fail_json(msg="Can't open configuration file, or configuration invalid") - #get updates lines + # get updates lines change_string = f.compare_config() - #remove not updatable parts + # remove not updatable parts c = FortiConfig() c.parse_config_output(change_string) @@ -165,7 +164,7 @@ def main(): result['change_string'] = change_string result['changed'] = True - #Commit if not check mode + # Commit if not check mode if module.check_mode is False and change_string != "": try: f.commit(change_string) diff --git a/lib/ansible/modules/network/fortios/fortios_ipv4_policy.py b/lib/ansible/modules/network/fortios/fortios_ipv4_policy.py index 0f641c300a..49f2f2fd6d 100644 --- a/lib/ansible/modules/network/fortios/fortios_ipv4_policy.py +++ b/lib/ansible/modules/network/fortios/fortios_ipv4_policy.py @@ -202,28 +202,28 @@ from ansible.module_utils.network.fortios.fortios import backup, AnsibleFortios def main(): argument_spec = dict( - comment = dict(type='str'), - id = dict(type='int', required=True), - src_intf = dict(type='list', default='any'), - dst_intf = dict(type='list', default='any'), - state = dict(choices=['present', 'absent'], default='present'), - src_addr = dict(type='list'), - dst_addr = dict(type='list'), - src_addr_negate = dict(type='bool', default=False), - dst_addr_negate = dict(type='bool', default=False), - policy_action = dict(choices=['accept','deny'], aliases=['action']), - service = dict(aliases=['services'], type='list'), - service_negate = dict(type='bool', default=False), - schedule = dict(type='str', default='always'), - nat = dict(type='bool', default=False), - fixedport = dict(type='bool', default=False), - poolname = dict(type='str'), - av_profile = dict(type='str'), - webfilter_profile = dict(type='str'), - ips_sensor = dict(type='str'), - application_list = dict(type='str'), - logtraffic = dict(choices=['disable','all','utm'], default='utm'), - logtraffic_start = dict(type='bool', default=False), + comment=dict(type='str'), + id=dict(type='int', required=True), + src_intf=dict(type='list', default='any'), + dst_intf=dict(type='list', default='any'), + state=dict(choices=['present', 'absent'], default='present'), + src_addr=dict(type='list'), + dst_addr=dict(type='list'), + src_addr_negate=dict(type='bool', default=False), + dst_addr_negate=dict(type='bool', default=False), + policy_action=dict(choices=['accept', 'deny'], aliases=['action']), + service=dict(aliases=['services'], type='list'), + service_negate=dict(type='bool', default=False), + schedule=dict(type='str', default='always'), + nat=dict(type='bool', default=False), + fixedport=dict(type='bool', default=False), + poolname=dict(type='str'), + av_profile=dict(type='str'), + webfilter_profile=dict(type='str'), + ips_sensor=dict(type='str'), + application_list=dict(type='str'), + logtraffic=dict(choices=['disable', 'all', 'utm'], default='utm'), + logtraffic_start=dict(type='bool', default=False), ) # merge global required_if & argument_spec from module_utils/fortios.py @@ -236,7 +236,7 @@ def main(): module = AnsibleModule( argument_spec=argument_spec, supports_check_mode=True, - required_if=fortios_required_if + ipv4_policy_required_if , + required_if=fortios_required_if + ipv4_policy_required_if, ) # init forti object diff --git a/lib/ansible/modules/network/illumos/dladm_iptun.py b/lib/ansible/modules/network/illumos/dladm_iptun.py index e116fbcf71..74b0e197bf 100644 --- a/lib/ansible/modules/network/illumos/dladm_iptun.py +++ b/lib/ansible/modules/network/illumos/dladm_iptun.py @@ -206,7 +206,6 @@ class IPTun(object): rc=rc) - def main(): module = AnsibleModule( argument_spec=dict( diff --git a/lib/ansible/modules/network/illumos/dladm_linkprop.py b/lib/ansible/modules/network/illumos/dladm_linkprop.py index 6c87315ee2..c4c0050b61 100644 --- a/lib/ansible/modules/network/illumos/dladm_linkprop.py +++ b/lib/ansible/modules/network/illumos/dladm_linkprop.py @@ -231,7 +231,6 @@ def main(): supports_check_mode=True ) - linkprop = LinkProp(module) rc = None diff --git a/lib/ansible/modules/network/ios/ios_banner.py b/lib/ansible/modules/network/ios/ios_banner.py index 36882b5dea..f3401eaf4d 100644 --- a/lib/ansible/modules/network/ios/ios_banner.py +++ b/lib/ansible/modules/network/ios/ios_banner.py @@ -96,6 +96,7 @@ from ansible.module_utils.network.ios.ios import load_config, run_commands from ansible.module_utils.network.ios.ios import ios_argument_spec, check_args import re + def map_obj_to_commands(updates, module): commands = list() want, have = updates @@ -114,6 +115,7 @@ def map_obj_to_commands(updates, module): return commands + def map_config_to_obj(module): rc, out, err = exec_command(module, 'show banner %s' % module.params['banner']) if rc == 0: @@ -132,6 +134,7 @@ def map_config_to_obj(module): obj['state'] = 'present' return obj + def map_params_to_obj(module): text = module.params['text'] if text: @@ -143,6 +146,7 @@ def map_params_to_obj(module): 'state': module.params['state'] } + def main(): """ main entry point for module execution """ diff --git a/lib/ansible/modules/network/ios/ios_command.py b/lib/ansible/modules/network/ios/ios_command.py index 43941e5fef..a5c79d2194 100644 --- a/lib/ansible/modules/network/ios/ios_command.py +++ b/lib/ansible/modules/network/ios/ios_command.py @@ -140,12 +140,14 @@ from ansible.module_utils.network.common.utils import ComplexList from ansible.module_utils.network.common.parsing import Conditional from ansible.module_utils.six import string_types + def to_lines(stdout): for item in stdout: if isinstance(item, string_types): item = str(item).split('\n') yield item + def parse_commands(module, warnings): command = ComplexList(dict( command=dict(key=True), @@ -167,6 +169,7 @@ def parse_commands(module, warnings): ) return commands + def main(): """main entry point for module execution """ @@ -220,7 +223,6 @@ def main(): msg = 'One or more conditional statements have not be satisfied' module.fail_json(msg=msg, failed_conditions=failed_conditions) - result.update({ 'changed': False, 'stdout': responses, diff --git a/lib/ansible/modules/network/ios/ios_facts.py b/lib/ansible/modules/network/ios/ios_facts.py index f89c415ac8..6645bb85d7 100644 --- a/lib/ansible/modules/network/ios/ios_facts.py +++ b/lib/ansible/modules/network/ios/ios_facts.py @@ -162,13 +162,13 @@ class FactsBase(object): self.facts = dict() self.responses = None - def populate(self): self.responses = run_commands(self.module, self.COMMANDS, check_rc=False) def run(self, cmd): return run_commands(self.module, cmd, check_rc=False) + class Default(FactsBase): COMMANDS = ['show version'] @@ -440,6 +440,7 @@ FACT_SUBSETS = dict( VALID_SUBSETS = frozenset(FACT_SUBSETS.keys()) + def main(): """main entry point for module execution """ diff --git a/lib/ansible/modules/network/ios/ios_system.py b/lib/ansible/modules/network/ios/ios_system.py index ac1e6778ad..e0baf07c4b 100644 --- a/lib/ansible/modules/network/ios/ios_system.py +++ b/lib/ansible/modules/network/ios/ios_system.py @@ -125,6 +125,7 @@ from ansible.module_utils.network.common.utils import ComplexList _CONFIGURED_VRFS = None + def has_vrf(module, vrf): global _CONFIGURED_VRFS if _CONFIGURED_VRFS is not None: @@ -133,20 +134,23 @@ def has_vrf(module, vrf): _CONFIGURED_VRFS = re.findall(r'vrf definition (\S+)', config) return vrf in _CONFIGURED_VRFS + def requires_vrf(module, vrf): if not has_vrf(module, vrf): module.fail_json(msg='vrf %s is not configured' % vrf) + def diff_list(want, have): adds = [w for w in want if w not in have] removes = [h for h in have if h not in want] return (adds, removes) + def map_obj_to_commands(want, have, module): commands = list() state = module.params['state'] - needs_update = lambda x: want.get(x) is not None and (want.get(x) != have.get(x)) + def needs_update(x): return want.get(x) is not None and (want.get(x) != have.get(x)) if state == 'absent': if have['hostname'] != 'Router': @@ -226,7 +230,6 @@ def map_obj_to_commands(want, have, module): else: commands.append('ip domain list %s' % item['name']) - if want['name_servers']: adds, removes = diff_list(want['name_servers'], have['name_servers']) for item in removes: @@ -243,10 +246,12 @@ def map_obj_to_commands(want, have, module): return commands + def parse_hostname(config): match = re.search(r'^hostname (\S+)', config, re.M) return match.group(1) + def parse_domain_name(config): match = re.findall(r'^ip domain name (?:vrf (\S+) )*(\S+)', config, re.M) matches = list() @@ -256,6 +261,7 @@ def parse_domain_name(config): matches.append({'name': name, 'vrf': vrf}) return matches + def parse_domain_search(config): match = re.findall(r'^ip domain list (?:vrf (\S+) )*(\S+)', config, re.M) matches = list() @@ -265,6 +271,7 @@ def parse_domain_search(config): matches.append({'name': name, 'vrf': vrf}) return matches + def parse_name_servers(config): match = re.findall(r'^ip name-server (?:vrf (\S+) )*(.*)', config, re.M) matches = list() @@ -275,11 +282,13 @@ def parse_name_servers(config): matches.append({'server': server, 'vrf': vrf}) return matches + def parse_lookup_source(config): match = re.search(r'ip domain lookup source-interface (\S+)', config, re.M) if match: return match.group(1) + def map_config_to_obj(module): config = get_config(module) return { @@ -291,6 +300,7 @@ def map_config_to_obj(module): 'name_servers': parse_name_servers(config) } + def map_params_to_obj(module): obj = { 'hostname': module.params['hostname'], @@ -324,6 +334,7 @@ def map_params_to_obj(module): return obj + def main(): """ Main entry point for Ansible module execution """ diff --git a/lib/ansible/modules/network/ios/ios_vrf.py b/lib/ansible/modules/network/ios/ios_vrf.py index 24c721efd3..e53b33b377 100644 --- a/lib/ansible/modules/network/ios/ios_vrf.py +++ b/lib/ansible/modules/network/ios/ios_vrf.py @@ -175,9 +175,10 @@ def add_command_to_vrf(name, cmd, commands): ]) commands.append(cmd) + def map_obj_to_commands(updates, module): commands = list() - state = module.params['state'] # FIXME NOT USED + state = module.params['state'] # FIXME NOT USED for update in updates: want, have = update @@ -226,6 +227,7 @@ def map_obj_to_commands(updates, module): return commands + def parse_description(configobj, name): cfg = configobj['vrf definition %s' % name] cfg = '\n'.join(cfg.children) @@ -233,6 +235,7 @@ def parse_description(configobj, name): if match: return match.group(1) + def parse_rd(configobj, name): cfg = configobj['vrf definition %s' % name] cfg = '\n'.join(cfg.children) @@ -240,6 +243,7 @@ def parse_rd(configobj, name): if match: return match.group(1) + def parse_interfaces(configobj, name): vrf_cfg = 'vrf forwarding %s' % name interfaces = list() @@ -249,6 +253,7 @@ def parse_interfaces(configobj, name): interfaces.append(intf.split(' ')[1]) return interfaces + def map_config_to_obj(module): config = get_config(module) configobj = NetworkConfig(indent=1, contents=config) @@ -290,6 +295,7 @@ def get_param_value(key, item, module): return value + def map_params_to_obj(module): vrfs = module.params.get('vrfs') if not vrfs: @@ -320,6 +326,7 @@ def map_params_to_obj(module): return objects + def update_objects(want, have): updates = list() for entry in want: diff --git a/lib/ansible/modules/network/netscaler/netscaler_lb_vserver.py b/lib/ansible/modules/network/netscaler/netscaler_lb_vserver.py index c5b26b5e69..ca78849dd9 100644 --- a/lib/ansible/modules/network/netscaler/netscaler_lb_vserver.py +++ b/lib/ansible/modules/network/netscaler/netscaler_lb_vserver.py @@ -1900,10 +1900,10 @@ def main(): module.fail_json(msg=msg, diff=lb_vserver_diff(client, module, lbvserver_proxy), **module_result) if not service_bindings_identical(client, module): - module.fail_json(msg='service bindings are not identical', **module_result) + module.fail_json(msg='service bindings are not identical', **module_result) if not servicegroup_bindings_identical(client, module): - module.fail_json(msg='servicegroup bindings are not identical', **module_result) + module.fail_json(msg='servicegroup bindings are not identical', **module_result) if module.params['servicetype'] == 'SSL': if not ssl_certkey_bindings_identical(client, module): diff --git a/lib/ansible/modules/network/netvisor/pn_cluster.py b/lib/ansible/modules/network/netvisor/pn_cluster.py index c50117e296..dada1d16dc 100644 --- a/lib/ansible/modules/network/netvisor/pn_cluster.py +++ b/lib/ansible/modules/network/netvisor/pn_cluster.py @@ -243,8 +243,8 @@ def main(): pn_cliusername=dict(required=False, type='str'), pn_clipassword=dict(required=False, type='str', no_log=True), pn_cliswitch=dict(required=False, type='str', default='local'), - state =dict(required=True, type='str', - choices=['present', 'absent']), + state=dict(required=True, type='str', + choices=['present', 'absent']), pn_name=dict(required=True, type='str'), pn_cluster_node1=dict(type='str'), pn_cluster_node2=dict(type='str'), diff --git a/lib/ansible/modules/network/netvisor/pn_ospfarea.py b/lib/ansible/modules/network/netvisor/pn_ospfarea.py index d0baf731c8..34739d451d 100644 --- a/lib/ansible/modules/network/netvisor/pn_ospfarea.py +++ b/lib/ansible/modules/network/netvisor/pn_ospfarea.py @@ -140,8 +140,8 @@ def main(): pn_cliusername=dict(required=True, type='str'), pn_clipassword=dict(required=True, type='str', no_log=True), pn_cliswitch=dict(required=False, type='str'), - state =dict(required=True, type='str', - choices=['present', 'absent', 'update']), + state=dict(required=True, type='str', + choices=['present', 'absent', 'update']), pn_vrouter_name=dict(required=True, type='str'), pn_ospf_area=dict(required=True, type='str'), pn_stub_type=dict(type='str', choices=['none', 'stub', 'nssa', diff --git a/lib/ansible/modules/network/netvisor/pn_vlag.py b/lib/ansible/modules/network/netvisor/pn_vlag.py index a8f02f66b9..47962d702c 100644 --- a/lib/ansible/modules/network/netvisor/pn_vlag.py +++ b/lib/ansible/modules/network/netvisor/pn_vlag.py @@ -251,8 +251,8 @@ def main(): pn_cliusername=dict(required=False, type='str'), pn_clipassword=dict(required=False, type='str', no_log=True), pn_cliswitch=dict(required=False, type='str', default='local'), - state =dict(required=True, type='str', - choices=['present', 'absent', 'update']), + state=dict(required=True, type='str', + choices=['present', 'absent', 'update']), pn_name=dict(required=True, type='str'), pn_port=dict(type='str'), pn_peer_port=dict(type='str'), @@ -269,7 +269,7 @@ def main(): ), required_if=( ["state", "present", ["pn_name", "pn_port", "pn_peer_port", - "pn_peer_switch"]], + "pn_peer_switch"]], ["state", "absent", ["pn_name"]], ["state", "update", ["pn_name"]] ) diff --git a/lib/ansible/modules/network/netvisor/pn_vlan.py b/lib/ansible/modules/network/netvisor/pn_vlan.py index 80fd8f7860..b455a5b814 100644 --- a/lib/ansible/modules/network/netvisor/pn_vlan.py +++ b/lib/ansible/modules/network/netvisor/pn_vlan.py @@ -175,7 +175,7 @@ def run_cli(module, cli): :param module: The Ansible module to fetch command """ cliswitch = module.params['pn_cliswitch'] - state= module.params['state'] + state = module.params['state'] command = get_command_from_state(state) cmd = shlex.split(cli) @@ -232,8 +232,8 @@ def main(): pn_cliusername=dict(required=False, type='str'), pn_clipassword=dict(required=False, type='str', no_log=True), pn_cliswitch=dict(required=False, type='str', default='local'), - state =dict(required=True, type='str', - choices=['present', 'absent']), + state=dict(required=True, type='str', + choices=['present', 'absent']), pn_vlanid=dict(required=True, type='int'), pn_scope=dict(type='str', choices=['fabric', 'local']), pn_description=dict(type='str'), diff --git a/lib/ansible/modules/network/netvisor/pn_vrouter.py b/lib/ansible/modules/network/netvisor/pn_vrouter.py index 9d3a968bd7..b5271b2c64 100644 --- a/lib/ansible/modules/network/netvisor/pn_vrouter.py +++ b/lib/ansible/modules/network/netvisor/pn_vrouter.py @@ -289,8 +289,8 @@ def main(): pn_cliusername=dict(required=False, type='str'), pn_clipassword=dict(required=False, type='str', no_log=True), pn_cliswitch=dict(required=False, type='str', default='local'), - state =dict(required=True, type='str', - choices=['present', 'absent', 'update']), + state=dict(required=True, type='str', + choices=['present', 'absent', 'update']), pn_name=dict(required=True, type='str'), pn_vnet=dict(type='str'), pn_service_type=dict(type='str', choices=['dedicated', 'shared']), diff --git a/lib/ansible/modules/network/netvisor/pn_vrouterbgp.py b/lib/ansible/modules/network/netvisor/pn_vrouterbgp.py index 17d511c32d..ef3fe495d8 100644 --- a/lib/ansible/modules/network/netvisor/pn_vrouterbgp.py +++ b/lib/ansible/modules/network/netvisor/pn_vrouterbgp.py @@ -333,7 +333,7 @@ def main(): ) # Accessing the arguments - state= module.params['state'] + state = module.params['state'] vrouter_name = module.params['pn_vrouter_name'] neighbor = module.params['pn_neighbor'] remote_as = module.params['pn_remote_as'] diff --git a/lib/ansible/modules/network/netvisor/pn_vrouterif.py b/lib/ansible/modules/network/netvisor/pn_vrouterif.py index 4b55c37554..7f17778186 100644 --- a/lib/ansible/modules/network/netvisor/pn_vrouterif.py +++ b/lib/ansible/modules/network/netvisor/pn_vrouterif.py @@ -347,8 +347,8 @@ def main(): pn_cliusername=dict(required=False, type='str'), pn_clipassword=dict(required=False, type='str', no_log=True), pn_cliswitch=dict(required=False, type='str', default='local'), - state =dict(required=True, type='str', - choices=['present', 'absent']), + state=dict(required=True, type='str', + choices=['present', 'absent']), pn_vrouter_name=dict(required=True, type='str'), pn_vlan=dict(type='int'), pn_interface_ip=dict(required=True, type='str'), diff --git a/lib/ansible/modules/network/netvisor/pn_vrouterlbif.py b/lib/ansible/modules/network/netvisor/pn_vrouterlbif.py index 82594b44d8..94c55cb54d 100644 --- a/lib/ansible/modules/network/netvisor/pn_vrouterlbif.py +++ b/lib/ansible/modules/network/netvisor/pn_vrouterlbif.py @@ -240,8 +240,8 @@ def main(): pn_cliusername=dict(required=False, type='str'), pn_clipassword=dict(required=False, type='str', no_log=True), pn_cliswitch=dict(required=False, type='str', default='local'), - state =dict(required=True, type='str', - choices=['present', 'absent']), + state=dict(required=True, type='str', + choices=['present', 'absent']), pn_vrouter_name=dict(required=True, type='str'), pn_interface_ip=dict(type='str'), pn_index=dict(type='int') diff --git a/lib/ansible/modules/network/nxos/_nxos_mtu.py b/lib/ansible/modules/network/nxos/_nxos_mtu.py index 50fbdcb283..f2a52127a6 100644 --- a/lib/ansible/modules/network/nxos/_nxos_mtu.py +++ b/lib/ansible/modules/network/nxos/_nxos_mtu.py @@ -125,6 +125,7 @@ from ansible.module_utils.network.nxos.nxos import load_config, run_commands from ansible.module_utils.network.nxos.nxos import nxos_argument_spec, check_args from ansible.module_utils.basic import AnsibleModule + def execute_show_command(command, module): if 'show run' not in command: output = 'json' diff --git a/lib/ansible/modules/network/nxos/nxos_aaa_server_host.py b/lib/ansible/modules/network/nxos/nxos_aaa_server_host.py index f12a6f91aa..55dceed384 100644 --- a/lib/ansible/modules/network/nxos/nxos_aaa_server_host.py +++ b/lib/ansible/modules/network/nxos/nxos_aaa_server_host.py @@ -276,12 +276,11 @@ def main(): argument_spec.update(nxos_argument_spec) module = AnsibleModule(argument_spec=argument_spec, - supports_check_mode=True) + supports_check_mode=True) warnings = list() check_args(module, warnings) - server_type = module.params['server_type'] address = module.params['address'] key = module.params['key'] @@ -311,7 +310,6 @@ def main(): module.fail_json(msg='auth_port and acct_port can only be used' 'when server_type=radius') - existing = get_aaa_host_info(module, server_type, address) end_state = existing diff --git a/lib/ansible/modules/network/nxos/nxos_config.py b/lib/ansible/modules/network/nxos/nxos_config.py index a8a9e9ff7f..0232819a49 100644 --- a/lib/ansible/modules/network/nxos/nxos_config.py +++ b/lib/ansible/modules/network/nxos/nxos_config.py @@ -300,9 +300,9 @@ def get_candidate(module): def execute_show_commands(module, commands, output='text'): cmds = [] for command in to_list(commands): - cmd = { 'command': command, - 'output': output, - } + cmd = {'command': command, + 'output': output, + } cmds.append(cmd) body = run_commands(module, cmds) return body @@ -460,7 +460,6 @@ def main(): 'diff': {'before': str(base_config), 'after': str(running_config)} }) - module.exit_json(**result) diff --git a/lib/ansible/modules/network/nxos/nxos_gir.py b/lib/ansible/modules/network/nxos/nxos_gir.py index 549d83e73e..373ee56c60 100644 --- a/lib/ansible/modules/network/nxos/nxos_gir.py +++ b/lib/ansible/modules/network/nxos/nxos_gir.py @@ -168,6 +168,7 @@ from ansible.module_utils.network.nxos.nxos import get_config, load_config, run_ from ansible.module_utils.network.nxos.nxos import nxos_argument_spec, check_args from ansible.module_utils.basic import AnsibleModule + def execute_show_command(command, module, command_type='cli_show_ascii'): cmds = [command] provider = module.params['provider'] @@ -260,40 +261,39 @@ def main(): argument_spec = dict( system_mode_maintenance=dict(required=False, type='bool'), system_mode_maintenance_dont_generate_profile=dict(required=False, - type='bool'), + type='bool'), system_mode_maintenance_timeout=dict(required=False, type='str'), system_mode_maintenance_shutdown=dict(required=False, type='bool'), system_mode_maintenance_on_reload_reset_reason=dict(required=False, - choices=['hw_error','svc_failure','kern_failure', - 'wdog_timeout','fatal_error','lc_failure', - 'match_any','manual_reload']), + choices=['hw_error', 'svc_failure', 'kern_failure', + 'wdog_timeout', 'fatal_error', 'lc_failure', + 'match_any', 'manual_reload']), state=dict(choices=['absent', 'present', 'default'], - default='present', required=False) + default='present', required=False) ) argument_spec.update(nxos_argument_spec) module = AnsibleModule(argument_spec=argument_spec, - mutually_exclusive=[[ - 'system_mode_maintenance', - 'system_mode_maintenance_dont_generate_profile', - 'system_mode_maintenance_timeout', - 'system_mode_maintenance_shutdown', - 'system_mode_maintenance_on_reload_reset_reason' - ]], - required_one_of=[[ - 'system_mode_maintenance', - 'system_mode_maintenance_dont_generate_profile', - 'system_mode_maintenance_timeout', - 'system_mode_maintenance_shutdown', - 'system_mode_maintenance_on_reload_reset_reason' - ]], - supports_check_mode=True) + mutually_exclusive=[[ + 'system_mode_maintenance', + 'system_mode_maintenance_dont_generate_profile', + 'system_mode_maintenance_timeout', + 'system_mode_maintenance_shutdown', + 'system_mode_maintenance_on_reload_reset_reason' + ]], + required_one_of=[[ + 'system_mode_maintenance', + 'system_mode_maintenance_dont_generate_profile', + 'system_mode_maintenance_timeout', + 'system_mode_maintenance_shutdown', + 'system_mode_maintenance_on_reload_reset_reason' + ]], + supports_check_mode=True) warnings = list() check_args(module, warnings) - state = module.params['state'] mode = get_system_mode(module) commands = get_commands(module, state, mode) diff --git a/lib/ansible/modules/network/nxos/nxos_gir_profile_management.py b/lib/ansible/modules/network/nxos/nxos_gir_profile_management.py index a1058cfc79..d5d02fb764 100644 --- a/lib/ansible/modules/network/nxos/nxos_gir_profile_management.py +++ b/lib/ansible/modules/network/nxos/nxos_gir_profile_management.py @@ -176,7 +176,7 @@ def main(): commands=dict(required=False, type='list'), mode=dict(required=True, choices=['maintenance', 'normal']), state=dict(choices=['absent', 'present'], - default='present'), + default='present'), include_defaults=dict(default=False), config=dict() ) @@ -184,12 +184,11 @@ def main(): argument_spec.update(nxos_argument_spec) module = AnsibleModule(argument_spec=argument_spec, - supports_check_mode=True) + supports_check_mode=True) warnings = list() check_args(module, warnings) - state = module.params['state'] commands = module.params['commands'] or [] diff --git a/lib/ansible/modules/network/nxos/nxos_igmp.py b/lib/ansible/modules/network/nxos/nxos_igmp.py index e0c89e828b..6812eba1a5 100644 --- a/lib/ansible/modules/network/nxos/nxos_igmp.py +++ b/lib/ansible/modules/network/nxos/nxos_igmp.py @@ -95,6 +95,7 @@ from ansible.module_utils.network.nxos.nxos import load_config, run_commands from ansible.module_utils.network.nxos.nxos import nxos_argument_spec, check_args from ansible.module_utils.basic import AnsibleModule + def get_current(module): output = run_commands(module, {'command': 'show running-config', 'output': 'text'}) return { @@ -126,12 +127,11 @@ def main(): argument_spec.update(nxos_argument_spec) module = AnsibleModule(argument_spec=argument_spec, - supports_check_mode=True) + supports_check_mode=True) warnings = list() check_args(module, warnings) - current = get_current(module) desired = get_desired(module) diff --git a/lib/ansible/modules/network/nxos/nxos_igmp_interface.py b/lib/ansible/modules/network/nxos/nxos_igmp_interface.py index ec32eec561..fe01bac3d0 100644 --- a/lib/ansible/modules/network/nxos/nxos_igmp_interface.py +++ b/lib/ansible/modules/network/nxos/nxos_igmp_interface.py @@ -242,6 +242,7 @@ from ansible.module_utils.basic import AnsibleModule import re + def execute_show_command(command, module, command_type='cli_show'): if command_type == 'cli_show_ascii': cmds = [{ @@ -501,11 +502,11 @@ def config_remove_oif(existing, existing_oif_prefix_source): if each.get('prefix') and each.get('source'): command = 'no ip igmp static-oif {0} source {1} '.format( each.get('prefix'), each.get('source') - ) + ) elif each.get('prefix'): command = 'no ip igmp static-oif {0}'.format( each.get('prefix') - ) + ) if command: commands.append(command) command = None @@ -533,7 +534,7 @@ def main(): oif_source=dict(required=False, type='str'), restart=dict(type='bool', default=False), state=dict(choices=['present', 'absent', 'default'], - default='present'), + default='present'), include_defaults=dict(default=True), config=dict(), save=dict(type='bool', default=False) @@ -542,12 +543,11 @@ def main(): argument_spec.update(nxos_argument_spec) module = AnsibleModule(argument_spec=argument_spec, - supports_check_mode=True) + supports_check_mode=True) warnings = list() check_args(module, warnings) - state = module.params['state'] interface = module.params['interface'] oif_prefix = module.params['oif_prefix'] @@ -607,7 +607,7 @@ def main(): changed = False commands = [] proposed = dict((k, v) for k, v in module.params.items() - if v is not None and k in args) + if v is not None and k in args) CANNOT_ABSENT = ['version', 'startup_query_interval', 'startup_query_count', 'robustness', 'querier_timeout', diff --git a/lib/ansible/modules/network/nxos/nxos_igmp_snooping.py b/lib/ansible/modules/network/nxos/nxos_igmp_snooping.py index 9ebe7ea1e2..6d84d40fb8 100644 --- a/lib/ansible/modules/network/nxos/nxos_igmp_snooping.py +++ b/lib/ansible/modules/network/nxos/nxos_igmp_snooping.py @@ -274,7 +274,7 @@ def main(): if state == 'present': delta = dict( set(proposed.items()).difference(existing.items()) - ) + ) if delta: command = config_igmp_snooping(delta, existing) if command: @@ -283,7 +283,7 @@ def main(): proposed = get_igmp_snooping_defaults() delta = dict( set(proposed.items()).difference(existing.items()) - ) + ) if delta: command = config_igmp_snooping(delta, existing, default=True) if command: diff --git a/lib/ansible/modules/network/nxos/nxos_ip_interface.py b/lib/ansible/modules/network/nxos/nxos_ip_interface.py index e4da09acf4..74c68c6260 100644 --- a/lib/ansible/modules/network/nxos/nxos_ip_interface.py +++ b/lib/ansible/modules/network/nxos/nxos_ip_interface.py @@ -343,13 +343,13 @@ def parse_interface_data(body): splitted_body = body.split('\n') for index in range(0, len(splitted_body) - 1): - if "Encapsulation 802.1Q" in splitted_body[index]: - regex = r'(.+?ID\s(?P\d+).*)?' - match = re.match(regex, splitted_body[index]) - if match: - match_dict = match.groupdict() - if match_dict['dot1q'] is not None: - return int(match_dict['dot1q']) + if "Encapsulation 802.1Q" in splitted_body[index]: + regex = r'(.+?ID\s(?P\d+).*)?' + match = re.match(regex, splitted_body[index]) + if match: + match_dict = match.groupdict() + if match_dict['dot1q'] is not None: + return int(match_dict['dot1q']) return 0 diff --git a/lib/ansible/modules/network/nxos/nxos_logging.py b/lib/ansible/modules/network/nxos/nxos_logging.py index f0821bb739..c198b21dbc 100644 --- a/lib/ansible/modules/network/nxos/nxos_logging.py +++ b/lib/ansible/modules/network/nxos/nxos_logging.py @@ -146,7 +146,7 @@ def map_obj_to_commands(updates, module): pass if w['facility'] is not None: - commands.append('logging level {} {}'.format(w['facility'], w['facility_level'])) + commands.append('logging level {} {}'.format(w['facility'], w['facility_level'])) return commands diff --git a/lib/ansible/modules/network/nxos/nxos_ntp_auth.py b/lib/ansible/modules/network/nxos/nxos_ntp_auth.py index b8cce10b34..23b55ee2c4 100644 --- a/lib/ansible/modules/network/nxos/nxos_ntp_auth.py +++ b/lib/ansible/modules/network/nxos/nxos_ntp_auth.py @@ -198,7 +198,7 @@ def get_ntp_auth_info(key_id, module): def auth_type_to_num(auth_type): - if auth_type == 'encrypt' : + if auth_type == 'encrypt': return '7' else: return '0' @@ -258,12 +258,11 @@ def main(): argument_spec.update(nxos_argument_spec) module = AnsibleModule(argument_spec=argument_spec, - supports_check_mode=True) + supports_check_mode=True) warnings = list() check_args(module, warnings) - key_id = module.params['key_id'] md5string = module.params['md5string'] auth_type = module.params['auth_type'] diff --git a/lib/ansible/modules/network/nxos/nxos_ntp_options.py b/lib/ansible/modules/network/nxos/nxos_ntp_options.py index d412413898..afa535f07d 100644 --- a/lib/ansible/modules/network/nxos/nxos_ntp_options.py +++ b/lib/ansible/modules/network/nxos/nxos_ntp_options.py @@ -95,7 +95,7 @@ def get_current(module): cmd = ('show running-config', 'show ntp logging') output = run_commands(module, ({'command': cmd[0], 'output': 'text'}, - {'command': cmd[1], 'output': 'text'})) + {'command': cmd[1], 'output': 'text'})) match = re.search(r"^ntp master(?: (\d+))", output[0], re.M) if match: @@ -167,7 +167,6 @@ def main(): else: commands.append('no ntp logging') - result['commands'] = commands result['updates'] = commands diff --git a/lib/ansible/modules/network/nxos/nxos_nxapi.py b/lib/ansible/modules/network/nxos/nxos_nxapi.py index 1e2cfcc01f..3f75808c88 100644 --- a/lib/ansible/modules/network/nxos/nxos_nxapi.py +++ b/lib/ansible/modules/network/nxos/nxos_nxapi.py @@ -130,6 +130,7 @@ from ansible.module_utils.network.nxos.nxos import check_args as nxos_check_args from ansible.module_utils.basic import AnsibleModule from ansible.module_utils.six import iteritems + def check_args(module, warnings): provider = module.params['provider'] if provider['transport'] == 'nxapi': @@ -159,10 +160,11 @@ def check_args(module, warnings): return warnings + def map_obj_to_commands(want, have, module): commands = list() - needs_update = lambda x: want.get(x) is not None and (want.get(x) != have.get(x)) + def needs_update(x): return want.get(x) is not None and (want.get(x) != have.get(x)) if needs_update('state'): if want['state'] == 'absent': @@ -191,6 +193,7 @@ def map_obj_to_commands(want, have, module): return commands + def parse_http(data): http_res = [r'nxapi http port (\d+)'] http_port = None @@ -203,6 +206,7 @@ def parse_http(data): return {'http': http_port is not None, 'http_port': http_port} + def parse_https(data): https_res = [r'nxapi https port (\d+)'] https_port = None @@ -215,6 +219,7 @@ def parse_https(data): return {'https': https_port is not None, 'https_port': https_port} + def parse_sandbox(data): sandbox = [item for item in data.split('\n') if re.search(r'.*sandbox.*', item)] value = False @@ -222,6 +227,7 @@ def parse_sandbox(data): value = True return {'sandbox': value} + def map_config_to_obj(module): out = run_commands(module, ['show run all | inc nxapi'], check_rc=False)[0] match = re.search(r'no feature nxapi', out, re.M) @@ -240,6 +246,7 @@ def map_config_to_obj(module): return obj + def map_params_to_obj(module): obj = { 'http': module.params['http'], @@ -252,6 +259,7 @@ def map_params_to_obj(module): return obj + def main(): """ main entry point for module execution """ @@ -275,8 +283,6 @@ def main(): module = AnsibleModule(argument_spec=argument_spec, supports_check_mode=True) - - warnings = list() check_args(module, warnings) diff --git a/lib/ansible/modules/network/nxos/nxos_overlay_global.py b/lib/ansible/modules/network/nxos/nxos_overlay_global.py index 32e8415d6c..abd77a5cba 100644 --- a/lib/ansible/modules/network/nxos/nxos_overlay_global.py +++ b/lib/ansible/modules/network/nxos/nxos_overlay_global.py @@ -134,7 +134,7 @@ def normalize_mac(proposed_mac, module): else: octect_len = len(octect) padding = 4 - octect_len - splitted_mac.append(octect.zfill(padding+1)) + splitted_mac.append(octect.zfill(padding + 1)) elif ':' in proposed_mac: splitted_mac = proposed_mac.split(':') @@ -150,7 +150,7 @@ def normalize_mac(proposed_mac, module): module.fail_json(msg='Invalid MAC address format', proposed_mac=proposed_mac) joined_mac = ''.join(splitted_mac) - mac = [joined_mac[i:i+4] for i in range(0, len(joined_mac), 4)] + mac = [joined_mac[i:i + 4] for i in range(0, len(joined_mac), 4)] return '.'.join(mac).upper() @@ -172,7 +172,7 @@ def main(): existing = get_existing(module, args) proposed = dict((k, v) for k, v in module.params.items() - if v is not None and k in args) + if v is not None and k in args) candidate = CustomNetworkConfig(indent=3) get_commands(module, existing, proposed, candidate) diff --git a/lib/ansible/modules/network/nxos/nxos_ping.py b/lib/ansible/modules/network/nxos/nxos_ping.py index ee24a5fba4..9a9572088d 100644 --- a/lib/ansible/modules/network/nxos/nxos_ping.py +++ b/lib/ansible/modules/network/nxos/nxos_ping.py @@ -110,7 +110,7 @@ from ansible.module_utils.basic import AnsibleModule def get_summary(results_list, reference_point): - summary_string = results_list[reference_point+1] + summary_string = results_list[reference_point + 1] summary_list = summary_string.split(',') summary = dict( @@ -119,7 +119,7 @@ def get_summary(results_list, reference_point): packet_loss=summary_list[2].split('packet')[0].strip(), ) - if 'bytes from' not in results_list[reference_point-2]: + if 'bytes from' not in results_list[reference_point - 2]: ping_pass = False else: ping_pass = True @@ -168,7 +168,7 @@ def get_ping_results(command, module): splitted_ping = ping.split('\n') reference_point = get_statistics_summary_line(splitted_ping) summary, ping_pass = get_summary(splitted_ping, reference_point) - rtt = get_rtt(splitted_ping, summary['packet_loss'], reference_point+2) + rtt = get_rtt(splitted_ping, summary['packet_loss'], reference_point + 2) return (summary, rtt, ping_pass) diff --git a/lib/ansible/modules/network/nxos/nxos_smu.py b/lib/ansible/modules/network/nxos/nxos_smu.py index 427819377d..aeb1bc9281 100644 --- a/lib/ansible/modules/network/nxos/nxos_smu.py +++ b/lib/ansible/modules/network/nxos/nxos_smu.py @@ -131,13 +131,12 @@ def main(): argument_spec.update(nxos_argument_spec) module = AnsibleModule(argument_spec=argument_spec, - supports_check_mode=True) + supports_check_mode=True) warnings = list() check_args(module, warnings) results = {'changed': False, 'commands': [], 'warnings': warnings} - pkg = module.params['pkg'] file_system = module.params['file_system'] remote_exists = remote_file_exists(module, pkg, file_system=file_system) diff --git a/lib/ansible/modules/network/nxos/nxos_snapshot.py b/lib/ansible/modules/network/nxos/nxos_snapshot.py index 07b50e6475..e8f07af70b 100644 --- a/lib/ansible/modules/network/nxos/nxos_snapshot.py +++ b/lib/ansible/modules/network/nxos/nxos_snapshot.py @@ -333,6 +333,7 @@ def write_on_file(content, filename, module): return filepath + def main(): argument_spec = dict( action=dict(required=True, choices=['create', 'add', 'compare', 'delete', 'delete_all']), @@ -421,4 +422,4 @@ def main(): module.exit_json(**result) if __name__ == '__main__': - main() + main() diff --git a/lib/ansible/modules/network/nxos/nxos_static_route.py b/lib/ansible/modules/network/nxos/nxos_static_route.py index 1e44755d71..3989a88b74 100644 --- a/lib/ansible/modules/network/nxos/nxos_static_route.py +++ b/lib/ansible/modules/network/nxos/nxos_static_route.py @@ -182,7 +182,7 @@ def set_route_command(module, prefix): def get_dotted_mask(mask): bits = 0 - for i in range(32-mask, 32): + for i in range(32 - mask, 32): bits |= (1 << i) mask = ("%d.%d.%d.%d" % ((bits & 0xff000000) >> 24, (bits & 0xff0000) >> 16, (bits & 0xff00) >> 8, (bits & 0xff))) return mask diff --git a/lib/ansible/modules/network/nxos/nxos_system.py b/lib/ansible/modules/network/nxos/nxos_system.py index bf1909c61b..41dbfb4a2e 100644 --- a/lib/ansible/modules/network/nxos/nxos_system.py +++ b/lib/ansible/modules/network/nxos/nxos_system.py @@ -120,6 +120,7 @@ from ansible.module_utils.network.common.utils import ComplexList _CONFIGURED_VRFS = None + def has_vrf(module, vrf): global _CONFIGURED_VRFS if _CONFIGURED_VRFS is not None: @@ -128,12 +129,14 @@ def has_vrf(module, vrf): _CONFIGURED_VRFS = re.findall(r'vrf context (\S+)', config) return vrf in _CONFIGURED_VRFS + def map_obj_to_commands(want, have, module): commands = list() state = module.params['state'] - needs_update = lambda x: want.get(x) and (want.get(x) != have.get(x)) - difference = lambda x,y,z: [item for item in x[z] if item not in y[z]] + def needs_update(x): return want.get(x) and (want.get(x) != have.get(x)) + + def difference(x, y, z): return [item for item in x[z] if item not in y[z]] def remove(cmd, commands, vrf=None): if vrf: @@ -195,7 +198,7 @@ def map_obj_to_commands(want, have, module): if want['name_servers']: for item in difference(have, want, 'name_servers'): - cmd = 'no ip name-server %s' % item['server'] + cmd = 'no ip name-server %s' % item['server'] remove(cmd, commands, item['vrf']) for item in difference(want, have, 'name_servers'): cmd = 'ip name-server %s' % item['server'] @@ -206,11 +209,13 @@ def map_obj_to_commands(want, have, module): return commands + def parse_hostname(config): match = re.search(r'^hostname (\S+)', config, re.M) if match: return match.group(1) + def parse_domain_name(config, vrf_config): objects = list() regex = re.compile(r'ip domain-name (\S+)') @@ -226,6 +231,7 @@ def parse_domain_name(config, vrf_config): return objects + def parse_domain_search(config, vrf_config): objects = list() @@ -238,6 +244,7 @@ def parse_domain_search(config, vrf_config): return objects + def parse_name_servers(config, vrf_config, vrfs): objects = list() @@ -256,11 +263,13 @@ def parse_name_servers(config, vrf_config, vrfs): return objects + def parse_system_mtu(config): match = re.search(r'^system jumbomtu (\d+)', config, re.M) if match: return int(match.group(1)) + def map_config_to_obj(module): config = get_config(module) configobj = NetworkConfig(indent=2, contents=config) @@ -281,10 +290,12 @@ def map_config_to_obj(module): 'system_mtu': parse_system_mtu(config) } + def validate_system_mtu(value, module): if not 1500 <= value <= 9216: module.fail_json(msg='system_mtu must be between 1500 and 9216') + def map_params_to_obj(module): obj = { 'hostname': module.params['hostname'], @@ -316,6 +327,7 @@ def map_params_to_obj(module): return obj + def main(): """ main entry point for module execution """ diff --git a/lib/ansible/modules/network/nxos/nxos_udld.py b/lib/ansible/modules/network/nxos/nxos_udld.py index c0f364b86a..2291d92f0d 100644 --- a/lib/ansible/modules/network/nxos/nxos_udld.py +++ b/lib/ansible/modules/network/nxos/nxos_udld.py @@ -157,7 +157,6 @@ def apply_key_map(key_map, table): return new_dict - def get_commands_config_udld_global(delta, reset): config_args = { 'enabled': 'udld aggressive', @@ -224,13 +223,12 @@ def main(): argument_spec.update(nxos_argument_spec) module = AnsibleModule(argument_spec=argument_spec, - required_one_of=[['aggressive', 'msg_time', 'reset']], - supports_check_mode=True) + required_one_of=[['aggressive', 'msg_time', 'reset']], + supports_check_mode=True) warnings = list() check_args(module, warnings) - aggressive = module.params['aggressive'] msg_time = module.params['msg_time'] reset = module.params['reset'] diff --git a/lib/ansible/modules/network/nxos/nxos_udld_interface.py b/lib/ansible/modules/network/nxos/nxos_udld_interface.py index 08e4b4b3f9..3e641d7fe0 100644 --- a/lib/ansible/modules/network/nxos/nxos_udld_interface.py +++ b/lib/ansible/modules/network/nxos/nxos_udld_interface.py @@ -171,7 +171,7 @@ def get_commands_config_udld_interface1(delta, interface, module, existing): if mode == 'aggressive': command = 'udld aggressive' if mode == 'enabled': - command = 'no udld aggressive ; udld enable' + command = 'no udld aggressive ; udld enable' elif mode == 'disabled': command = 'no udld aggressive ; no udld enable' if command: @@ -188,7 +188,7 @@ def get_commands_config_udld_interface2(delta, interface, module, existing): if mode == 'aggressive': command = 'udld aggressive' if mode == 'enabled': - command = 'no udld aggressive ; no udld disable' + command = 'no udld aggressive ; no udld disable' elif mode == 'disabled': command = 'no udld aggressive ; udld disable' if command: @@ -237,7 +237,7 @@ def get_commands_remove_udld_interface2(delta, interface, module, existing): def main(): argument_spec = dict( mode=dict(choices=['enabled', 'disabled', 'aggressive'], - required=True), + required=True), interface=dict(type='str', required=True), state=dict(choices=['absent', 'present'], default='present'), ) @@ -245,7 +245,7 @@ def main(): argument_spec.update(nxos_argument_spec) module = AnsibleModule(argument_spec=argument_spec, - supports_check_mode=True) + supports_check_mode=True) warnings = list() check_args(module, warnings) @@ -265,14 +265,14 @@ def main(): if state == 'present': if delta: command = get_commands_config_udld_interface1(delta, interface, - module, existing) + module, existing) commands.append(command) elif state == 'absent': common = set(proposed.items()).intersection(existing.items()) if common: command = get_commands_remove_udld_interface1( dict(common), interface, module, existing - ) + ) commands.append(command) cmds = flatten_list(commands) @@ -297,11 +297,11 @@ def main(): commands = [] if state == 'present': command = get_commands_config_udld_interface2(delta, interface, - module, existing) + module, existing) elif state == 'absent': command = get_commands_remove_udld_interface2( dict(common), interface, module, existing - ) + ) commands.append(command) cmds = flatten_list(commands) diff --git a/lib/ansible/modules/network/nxos/nxos_user.py b/lib/ansible/modules/network/nxos/nxos_user.py index c75c5439b8..599409f8ab 100644 --- a/lib/ansible/modules/network/nxos/nxos_user.py +++ b/lib/ansible/modules/network/nxos/nxos_user.py @@ -166,6 +166,7 @@ def validate_roles(value, module): if item not in VALID_ROLES: module.fail_json(msg='invalid role specified') + def map_obj_to_commands(updates, module): commands = list() state = module.params['state'] @@ -174,9 +175,11 @@ def map_obj_to_commands(updates, module): for update in updates: want, have = update - needs_update = lambda x: want.get(x) and (want.get(x) != have.get(x)) - add = lambda x: commands.append('username %s %s' % (want['name'], x)) - remove = lambda x: commands.append('no username %s %s' % (want['name'], x)) + def needs_update(x): return want.get(x) and (want.get(x) != have.get(x)) + + def add(x): return commands.append('username %s %s' % (want['name'], x)) + + def remove(x): return commands.append('no username %s %s' % (want['name'], x)) if want['state'] == 'absent': commands.append('no username %s' % want['name']) @@ -192,7 +195,6 @@ def map_obj_to_commands(updates, module): if needs_update('sshkey'): add('sshkey %s' % want['sshkey']) - if want['roles']: if have: for item in set(have['roles']).difference(want['roles']): @@ -204,13 +206,14 @@ def map_obj_to_commands(updates, module): for item in want['roles']: add('role %s' % item) - return commands + def parse_password(data): if not data.get('remote_login'): return '' + def parse_roles(data): configured_roles = data.get('TABLE_role')['ROW_role'] roles = list() @@ -219,6 +222,7 @@ def parse_roles(data): roles.append(item['role']) return roles + def map_config_to_obj(module): out = run_commands(module, ['show user-account | json']) data = out[0] @@ -235,6 +239,7 @@ def map_config_to_obj(module): }) return objects + def get_param_value(key, item, module): # if key doesn't exist in the item, get it from module.params if not item.get(key): @@ -249,6 +254,7 @@ def get_param_value(key, item, module): return value + def map_params_to_obj(module): aggregate = module.params['aggregate'] if not aggregate: @@ -290,6 +296,7 @@ def map_params_to_obj(module): return objects + def update_objects(want, have): updates = list() for entry in want: @@ -302,6 +309,7 @@ def update_objects(want, have): updates.append((entry, item)) return updates + def main(): """ main entry point for module execution """ @@ -328,7 +336,6 @@ def main(): mutually_exclusive=mutually_exclusive, supports_check_mode=True) - result = {'changed': False} warnings = list() diff --git a/lib/ansible/modules/network/nxos/nxos_vrf.py b/lib/ansible/modules/network/nxos/nxos_vrf.py index a4c7e747eb..358278554f 100644 --- a/lib/ansible/modules/network/nxos/nxos_vrf.py +++ b/lib/ansible/modules/network/nxos/nxos_vrf.py @@ -162,7 +162,7 @@ def get_vrf_description(vrf, module): for element in splitted_body: if 'description' in element: match_description = re.match(descr_regex, element, - re.DOTALL) + re.DOTALL) group_description = match_description.groupdict() description = group_description["descr"] @@ -182,7 +182,7 @@ def get_vrf(vrf, module): vrf_key = { 'vrf_name': 'vrf', 'vrf_state': 'admin_state' - } + } try: body = execute_show_command(command, module)[0] diff --git a/lib/ansible/modules/network/nxos/nxos_vtp_domain.py b/lib/ansible/modules/network/nxos/nxos_vtp_domain.py index 7a9fbde556..53c4080207 100644 --- a/lib/ansible/modules/network/nxos/nxos_vtp_domain.py +++ b/lib/ansible/modules/network/nxos/nxos_vtp_domain.py @@ -167,7 +167,7 @@ def main(): argument_spec.update(nxos_argument_spec) module = AnsibleModule(argument_spec=argument_spec, - supports_check_mode=True) + supports_check_mode=True) warnings = list() check_args(module, warnings) diff --git a/lib/ansible/modules/network/nxos/nxos_vtp_password.py b/lib/ansible/modules/network/nxos/nxos_vtp_password.py index e484d95ca5..14c04411c9 100644 --- a/lib/ansible/modules/network/nxos/nxos_vtp_password.py +++ b/lib/ansible/modules/network/nxos/nxos_vtp_password.py @@ -194,13 +194,13 @@ def main(): argument_spec = dict( vtp_password=dict(type='str', no_log=True), state=dict(choices=['absent', 'present'], - default='present'), + default='present'), ) argument_spec.update(nxos_argument_spec) module = AnsibleModule(argument_spec=argument_spec, - supports_check_mode=True) + supports_check_mode=True) warnings = list() check_args(module, warnings) diff --git a/lib/ansible/modules/network/nxos/nxos_vtp_version.py b/lib/ansible/modules/network/nxos/nxos_vtp_version.py index f25866174f..469b2bb69b 100644 --- a/lib/ansible/modules/network/nxos/nxos_vtp_version.py +++ b/lib/ansible/modules/network/nxos/nxos_vtp_version.py @@ -165,7 +165,7 @@ def main(): argument_spec.update(nxos_argument_spec) module = AnsibleModule(argument_spec=argument_spec, - supports_check_mode=True) + supports_check_mode=True) warnings = list() check_args(module, warnings) diff --git a/lib/ansible/modules/network/ordnance/ordnance_config.py b/lib/ansible/modules/network/ordnance/ordnance_config.py index a123b0d998..6f969604fc 100644 --- a/lib/ansible/modules/network/ordnance/ordnance_config.py +++ b/lib/ansible/modules/network/ordnance/ordnance_config.py @@ -206,6 +206,7 @@ def check_args(module, warnings): 'match=none instead. This argument will be ' 'removed in the future') + def extract_banners(config): banners = {} banner_cmds = re.findall(r'^banner (\w+)', config, re.M) @@ -225,6 +226,7 @@ def extract_banners(config): config = re.sub(r'banner \w+ \^C\^C', '!! banner removed', config) return (config, banners) + def diff_banners(want, have): candidate = {} for key, value in iteritems(want): @@ -232,6 +234,7 @@ def diff_banners(want, have): candidate[key] = value return candidate + def load_banners(module, banners): delimiter = module.params['multiline_delimiter'] for key, value in iteritems(banners): @@ -242,6 +245,7 @@ def load_banners(module, banners): time.sleep(1) module.connection.shell.receive() + def get_config(module, result): contents = module.params['config'] if not contents: @@ -251,6 +255,7 @@ def get_config(module, result): contents, banners = extract_banners(contents) return NetworkConfig(indent=1, contents=contents), banners + def get_candidate(module): candidate = NetworkConfig(indent=1) banners = {} @@ -265,6 +270,7 @@ def get_candidate(module): return candidate, banners + def run(module, result): match = module.params['match'] replace = module.params['replace'] @@ -275,7 +281,7 @@ def run(module, result): if match != 'none': config, have_banners = get_config(module, result) path = module.params['parents'] - configobjs = candidate.difference(config, path=path,match=match, + configobjs = candidate.difference(config, path=path, match=match, replace=replace) else: configobjs = candidate.items @@ -311,6 +317,7 @@ def run(module, result): module.config.save_config() result['changed'] = True + def main(): """ main entry point for module execution """ diff --git a/lib/ansible/modules/network/ordnance/ordnance_facts.py b/lib/ansible/modules/network/ordnance/ordnance_facts.py index eccfc5d7a7..ae4742f192 100644 --- a/lib/ansible/modules/network/ordnance/ordnance_facts.py +++ b/lib/ansible/modules/network/ordnance/ordnance_facts.py @@ -223,6 +223,7 @@ FACT_SUBSETS = dict( VALID_SUBSETS = frozenset(FACT_SUBSETS.keys()) + def main(): spec = dict( gather_subset=dict(default=['!config'], type='list') diff --git a/lib/ansible/modules/network/ovs/openvswitch_bridge.py b/lib/ansible/modules/network/ovs/openvswitch_bridge.py index b8009b6b4a..2f437d1ec7 100644 --- a/lib/ansible/modules/network/ovs/openvswitch_bridge.py +++ b/lib/ansible/modules/network/ovs/openvswitch_bridge.py @@ -1,5 +1,5 @@ #!/usr/bin/python -#coding: utf-8 -*- +# coding: utf-8 -*- # (c) 2013, David Stygstra # Portions copyright @ 2015 VMware, Inc. @@ -109,6 +109,7 @@ def _fail_mode_to_str(text): else: return text.strip() + def _external_ids_to_dict(text): if not text: return None @@ -122,6 +123,7 @@ def _external_ids_to_dict(text): return d + def map_obj_to_commands(want, have, module): commands = list() @@ -156,7 +158,7 @@ def map_obj_to_commands(want, have, module): command = templatized_command % module.params if want['parent']: - templatized_command = "%(parent)s %(vlan)s" + templatized_command = "%(parent)s %(vlan)s" command += " " + templatized_command % module.params if want['set']: @@ -175,7 +177,7 @@ def map_obj_to_commands(want, have, module): if want['external_ids']: for k, v in iteritems(want['external_ids']): templatized_command = ("%(ovs-vsctl)s -t %(timeout)s" - " br-set-external-id %(bridge)s") + " br-set-external-id %(bridge)s") command = templatized_command % module.params command += " " + k + " " + v commands.append(command) diff --git a/lib/ansible/modules/network/ovs/openvswitch_port.py b/lib/ansible/modules/network/ovs/openvswitch_port.py index f24d9658fe..ed0673a223 100644 --- a/lib/ansible/modules/network/ovs/openvswitch_port.py +++ b/lib/ansible/modules/network/ovs/openvswitch_port.py @@ -1,5 +1,5 @@ #!/usr/bin/python -#coding: utf-8 -*- +# coding: utf-8 -*- # (c) 2013, David Stygstra # Portions copyright @ 2015 VMware, Inc. @@ -118,6 +118,7 @@ def _external_ids_to_dict(text): return d + def _tag_to_str(text): text = text.strip() @@ -126,6 +127,7 @@ def _tag_to_str(text): else: return text + def map_obj_to_commands(want, have, module): commands = list() @@ -167,7 +169,7 @@ def map_obj_to_commands(want, have, module): command = templatized_command % module.params if want['tag']: - templatized_command = " tag=%(tag)s" + templatized_command = " tag=%(tag)s" command += templatized_command % module.params if want['set']: @@ -181,7 +183,7 @@ def map_obj_to_commands(want, have, module): templatized_command = ("%(ovs-vsctl)s -t %(timeout)s" " set port %(port)s external_ids:") command = templatized_command % module.params - command += k + "=" + v + command += k + "=" + v commands.append(command) return commands @@ -226,9 +228,10 @@ def map_params_to_obj(module): return obj + def main(): """ Entry point. """ - argument_spec={ + argument_spec = { 'bridge': {'required': True}, 'port': {'required': True}, 'state': {'default': 'present', 'choices': ['present', 'absent']}, diff --git a/lib/ansible/modules/network/panos/panos_admin.py b/lib/ansible/modules/network/panos/panos_admin.py index 6add314a02..436a39cd40 100644 --- a/lib/ansible/modules/network/panos/panos_admin.py +++ b/lib/ansible/modules/network/panos/panos_admin.py @@ -137,7 +137,7 @@ def admin_set(xapi, module, admin_username, admin_password, role): element='<%s>%s' % (role, rbval, role)) if admin_password is not None: - xapi.edit(xpath=_ADMIN_XPATH % admin_username+'/phash', + xapi.edit(xpath=_ADMIN_XPATH % admin_username + '/phash', element='%s' % phash) changed = True diff --git a/lib/ansible/modules/network/panos/panos_admpwd.py b/lib/ansible/modules/network/panos/panos_admpwd.py index e9f71fce6a..375df068b0 100644 --- a/lib/ansible/modules/network/panos/panos_admpwd.py +++ b/lib/ansible/modules/network/panos/panos_admpwd.py @@ -84,9 +84,9 @@ import sys try: import paramiko - HAS_LIB=True + HAS_LIB = True except ImportError: - HAS_LIB=False + HAS_LIB = False _PROMPTBUFF = 4096 @@ -101,7 +101,7 @@ def wait_with_timeout(module, shell, prompt, timeout=60): if len(endresult) != 0 and endresult[-1] == prompt: break - if time.time()-now > timeout: + if time.time() - now > timeout: module.fail_json(msg="Timeout waiting for prompt") return result @@ -143,14 +143,14 @@ def set_panwfw_password(module, ip_address, key_filename, newpassword, username) stdout += buff # enter password for the first time - shell.send(newpassword+'\n') + shell.send(newpassword + '\n') # wait for the password prompt buff = wait_with_timeout(module, shell, ":") stdout += buff # enter password for the second time - shell.send(newpassword+'\n') + shell.send(newpassword + '\n') # wait for the config mode prompt buff = wait_with_timeout(module, shell, "#") diff --git a/lib/ansible/modules/network/panos/panos_cert_gen_ssh.py b/lib/ansible/modules/network/panos/panos_cert_gen_ssh.py index 3a6c47830d..7c2516f9c1 100644 --- a/lib/ansible/modules/network/panos/panos_cert_gen_ssh.py +++ b/lib/ansible/modules/network/panos/panos_cert_gen_ssh.py @@ -81,7 +81,7 @@ EXAMPLES = ''' signed_by: "root-ca" ''' -RETURN=''' +RETURN = ''' # Default return values ''' @@ -96,9 +96,9 @@ import time try: import paramiko - HAS_LIB=True + HAS_LIB = True except ImportError: - HAS_LIB=False + HAS_LIB = False _PROMPTBUFF = 4096 @@ -113,14 +113,14 @@ def wait_with_timeout(module, shell, prompt, timeout=60): if len(endresult) != 0 and endresult[-1] == prompt: break - if time.time()-now > timeout: + if time.time() - now > timeout: module.fail_json(msg="Timeout waiting for prompt") return result def generate_cert(module, ip_address, key_filename, password, - cert_cn, cert_friendly_name, signed_by, rsa_nbits ): + cert_cn, cert_friendly_name, signed_by, rsa_nbits): stdout = "" client = paramiko.SSHClient() @@ -154,7 +154,7 @@ def generate_cert(module, ip_address, key_filename, password, shell.send('exit\n') if 'Success' not in buff: - module.fail_json(msg="Error generating self signed certificate: "+stdout) + module.fail_json(msg="Error generating self signed certificate: " + stdout) client.close() return stdout diff --git a/lib/ansible/modules/network/panos/panos_check.py b/lib/ansible/modules/network/panos/panos_check.py index aca02cc6b1..bfbda14e1e 100644 --- a/lib/ansible/modules/network/panos/panos_check.py +++ b/lib/ansible/modules/network/panos/panos_check.py @@ -75,7 +75,7 @@ EXAMPLES = ''' delay: 30 ''' -RETURN=''' +RETURN = ''' # Default return values ''' @@ -131,7 +131,7 @@ def main(): timeout=60 ) - checkpnt = time.time()+timeout + checkpnt = time.time() + timeout while True: try: xapi.op(cmd="show jobs all", cmd_xml=True) diff --git a/lib/ansible/modules/network/panos/panos_dag.py b/lib/ansible/modules/network/panos/panos_dag.py index db77d6bd83..7748c7ccc2 100644 --- a/lib/ansible/modules/network/panos/panos_dag.py +++ b/lib/ansible/modules/network/panos/panos_dag.py @@ -71,7 +71,7 @@ EXAMPLES = ''' dag_filter: "'aws-tag.aws:cloudformation:logical-id.ServerInstance' and 'instanceState.running'" ''' -RETURN=''' +RETURN = ''' # Default return values ''' diff --git a/lib/ansible/modules/network/panos/panos_import.py b/lib/ansible/modules/network/panos/panos_import.py index 7e0f82151d..db4e523fd8 100644 --- a/lib/ansible/modules/network/panos/panos_import.py +++ b/lib/ansible/modules/network/panos/panos_import.py @@ -73,7 +73,7 @@ EXAMPLES = ''' category: software ''' -RETURN=''' +RETURN = ''' # Default return values ''' @@ -118,7 +118,7 @@ def import_file(xapi, module, ip_address, file_, category): ) r = requests.post( - 'https://'+ip_address+'/api/', + 'https://' + ip_address + '/api/', verify=False, params=params, headers={'Content-Type': mef.content_type}, diff --git a/lib/ansible/modules/network/panos/panos_interface.py b/lib/ansible/modules/network/panos/panos_interface.py index 66be2315df..4aea9f1390 100644 --- a/lib/ansible/modules/network/panos/panos_interface.py +++ b/lib/ansible/modules/network/panos/panos_interface.py @@ -78,7 +78,7 @@ EXAMPLES = ''' create_default_route: "yes" ''' -RETURN=''' +RETURN = ''' # Default return values ''' @@ -98,8 +98,8 @@ _IF_XPATH = "/config/devices/entry[@name='localhost.localdomain']" +\ _ZONE_XPATH = "/config/devices/entry[@name='localhost.localdomain']" +\ "/vsys/entry/zone/entry" -_ZONE_XPATH_QUERY = _ZONE_XPATH+"[network/layer3/member/text()='%s']" -_ZONE_XPATH_IF = _ZONE_XPATH+"[@name='%s']/network/layer3/member[text()='%s']" +_ZONE_XPATH_QUERY = _ZONE_XPATH + "[network/layer3/member/text()='%s']" +_ZONE_XPATH_IF = _ZONE_XPATH + "[@name='%s']/network/layer3/member[text()='%s']" _VR_XPATH = "/config/devices/entry[@name='localhost.localdomain']" +\ "/network/virtual-router/entry" @@ -120,9 +120,9 @@ def add_dhcp_if(xapi, if_name, zone_name, create_default_route): if_xml = (''.join(if_xml)) % (if_name, cdr) xapi.edit(xpath=_IF_XPATH % if_name, element=if_xml) - xapi.set(xpath=_ZONE_XPATH+"[@name='%s']/network/layer3" % zone_name, + xapi.set(xpath=_ZONE_XPATH + "[@name='%s']/network/layer3" % zone_name, element='%s' % if_name) - xapi.set(xpath=_VR_XPATH+"[@name='default']/interface", + xapi.set(xpath=_VR_XPATH + "[@name='default']/interface", element='%s' % if_name) return True diff --git a/lib/ansible/modules/network/panos/panos_loadcfg.py b/lib/ansible/modules/network/panos/panos_loadcfg.py index 7f2146e581..bd30e3accb 100644 --- a/lib/ansible/modules/network/panos/panos_loadcfg.py +++ b/lib/ansible/modules/network/panos/panos_loadcfg.py @@ -71,7 +71,7 @@ EXAMPLES = ''' file: "{{result.filename}}" ''' -RETURN=''' +RETURN = ''' # Default return values ''' diff --git a/lib/ansible/modules/network/panos/panos_mgtconfig.py b/lib/ansible/modules/network/panos/panos_mgtconfig.py index db103249bb..fde5b640d2 100644 --- a/lib/ansible/modules/network/panos/panos_mgtconfig.py +++ b/lib/ansible/modules/network/panos/panos_mgtconfig.py @@ -81,7 +81,7 @@ EXAMPLES = ''' panorama_secondary: "1.1.1.4" ''' -RETURN=''' +RETURN = ''' # Default return values ''' @@ -112,11 +112,11 @@ def set_dns_server(xapi, new_dns_server, primary=True): tag = "primary" else: tag = "secondary" - xpath = _XPATH_DNS_SERVERS+"/"+tag + xpath = _XPATH_DNS_SERVERS + "/" + tag # check the current element value xapi.get(xpath) - val = xapi.element_root.find(".//"+tag) + val = xapi.element_root.find(".//" + tag) if val is not None: # element exists val = val.text @@ -135,11 +135,11 @@ def set_panorama_server(xapi, new_panorama_server, primary=True): tag = "panorama-server" else: tag = "panorama-server-2" - xpath = _XPATH_PANORAMA_SERVERS+"/"+tag + xpath = _XPATH_PANORAMA_SERVERS + "/" + tag # check the current element value xapi.get(xpath) - val = xapi.element_root.find(".//"+tag) + val = xapi.element_root.find(".//" + tag) if val is not None: # element exists val = val.text diff --git a/lib/ansible/modules/network/panos/panos_pg.py b/lib/ansible/modules/network/panos/panos_pg.py index 8cc39841cb..555d1f38c0 100644 --- a/lib/ansible/modules/network/panos/panos_pg.py +++ b/lib/ansible/modules/network/panos/panos_pg.py @@ -101,7 +101,7 @@ EXAMPLES = ''' vulnerability: "default" ''' -RETURN=''' +RETURN = ''' # Default return values ''' diff --git a/lib/ansible/modules/network/sros/sros_command.py b/lib/ansible/modules/network/sros/sros_command.py index b647955af6..1a24fd29a8 100644 --- a/lib/ansible/modules/network/sros/sros_command.py +++ b/lib/ansible/modules/network/sros/sros_command.py @@ -148,6 +148,7 @@ def to_lines(stdout): item = str(item).split('\n') yield item + def parse_commands(module, warnings): command = ComplexList(dict( command=dict(key=True), @@ -168,6 +169,7 @@ def parse_commands(module, warnings): ) return commands + def main(): """main entry point for module execution """ @@ -221,7 +223,6 @@ def main(): msg = 'One or more conditional statements have not be satisfied' module.fail_json(msg=msg, failed_conditions=failed_conditions) - result = { 'changed': False, 'stdout': responses, diff --git a/lib/ansible/modules/network/sros/sros_config.py b/lib/ansible/modules/network/sros/sros_config.py index 6a59e25398..7b65ba886d 100644 --- a/lib/ansible/modules/network/sros/sros_config.py +++ b/lib/ansible/modules/network/sros/sros_config.py @@ -275,6 +275,7 @@ def run(module, result): load_config(module, commands) result['changed'] = True + def main(): """ main entry point for module execution """ diff --git a/lib/ansible/modules/network/sros/sros_rollback.py b/lib/ansible/modules/network/sros/sros_rollback.py index e85192d086..41f9e3b095 100644 --- a/lib/ansible/modules/network/sros/sros_rollback.py +++ b/lib/ansible/modules/network/sros/sros_rollback.py @@ -108,6 +108,7 @@ def invoke(name, *args, **kwargs): if func: return func(*args, **kwargs) + def sanitize_config(lines): commands = list() for line in lines: @@ -118,6 +119,7 @@ def sanitize_config(lines): commands.append(line) return commands + def present(module, commands): setters = set() for key, value in module.argument_spec.items(): @@ -127,6 +129,7 @@ def present(module, commands): setters.add(setter) invoke(setter, module, commands) + def absent(module, commands): config = get_config(module) if 'rollback-location' in config: @@ -138,30 +141,36 @@ def absent(module, commands): if 'local-max-checkpoints' in config: commands.append('configure system rollback no remote-max-checkpoints') + def set_rollback_location(module, commands): value = module.params['rollback_location'] commands.append('configure system rollback rollback-location "%s"' % value) + def set_local_max_checkpoints(module, commands): value = module.params['local_max_checkpoints'] if not 1 <= value <= 50: module.fail_json(msg='local_max_checkpoints must be between 1 and 50') commands.append('configure system rollback local-max-checkpoints %s' % value) + def set_remote_max_checkpoints(module, commands): value = module.params['remote_max_checkpoints'] if not 1 <= value <= 50: module.fail_json(msg='remote_max_checkpoints must be between 1 and 50') commands.append('configure system rollback remote-max-checkpoints %s' % value) + def set_rescue_location(module, commands): value = module.params['rescue_location'] commands.append('configure system rollback rescue-location "%s"' % value) + def get_device_config(module): contents = get_config(module) return NetworkConfig(indent=4, contents=contents) + def main(): """ main entry point for module execution """ diff --git a/lib/ansible/modules/network/vyos/vyos_command.py b/lib/ansible/modules/network/vyos/vyos_command.py index f46532dc49..798ec0077a 100644 --- a/lib/ansible/modules/network/vyos/vyos_command.py +++ b/lib/ansible/modules/network/vyos/vyos_command.py @@ -142,6 +142,7 @@ from ansible.module_utils.six import string_types from ansible.module_utils.network.vyos.vyos import run_commands from ansible.module_utils.network.vyos.vyos import vyos_argument_spec + def to_lines(stdout): for item in stdout: if isinstance(item, string_types): diff --git a/lib/ansible/modules/network/vyos/vyos_system.py b/lib/ansible/modules/network/vyos/vyos_system.py index 40b520de5a..12f607afd8 100644 --- a/lib/ansible/modules/network/vyos/vyos_system.py +++ b/lib/ansible/modules/network/vyos/vyos_system.py @@ -162,6 +162,7 @@ def spec_to_commands(want, have): return commands + def map_param_to_obj(module): return { 'host_name': module.params['host_name'], diff --git a/lib/ansible/modules/notification/campfire.py b/lib/ansible/modules/notification/campfire.py index f3b93be60f..cf64381ddc 100644 --- a/lib/ansible/modules/notification/campfire.py +++ b/lib/ansible/modules/notification/campfire.py @@ -122,21 +122,21 @@ def main(): target_url = '%s/room/%s/speak.xml' % (URI, room) headers = {'Content-Type': 'application/xml', - 'User-agent': AGENT} + 'User-agent': AGENT} # Send some audible notification if requested if notify: response, info = fetch_url(module, target_url, data=NSTR % cgi.escape(notify), headers=headers) if info['status'] not in [200, 201]: module.fail_json(msg="unable to send msg: '%s', campfire api" - " returned error code: '%s'" % + " returned error code: '%s'" % (notify, info['status'])) # Send the message - response, info = fetch_url(module, target_url, data=MSTR %cgi.escape(msg), headers=headers) + response, info = fetch_url(module, target_url, data=MSTR % cgi.escape(msg), headers=headers) if info['status'] not in [200, 201]: module.fail_json(msg="unable to send msg: '%s', campfire api" - " returned error code: '%s'" % + " returned error code: '%s'" % (msg, info['status'])) module.exit_json(changed=True, room=room, msg=msg, notify=notify) diff --git a/lib/ansible/modules/notification/flowdock.py b/lib/ansible/modules/notification/flowdock.py index c17c897041..7e41e663b1 100644 --- a/lib/ansible/modules/notification/flowdock.py +++ b/lib/ansible/modules/notification/flowdock.py @@ -115,7 +115,7 @@ def main(): argument_spec=dict( token=dict(required=True, no_log=True), msg=dict(required=True), - type=dict(required=True, choices=["inbox","chat"]), + type=dict(required=True, choices=["inbox", "chat"]), external_user_name=dict(required=False), from_address=dict(required=False), source=dict(required=False), @@ -125,7 +125,7 @@ def main(): project=dict(required=False), tags=dict(required=False), link=dict(required=False), - validate_certs = dict(default='yes', type='bool'), + validate_certs=dict(default='yes', type='bool'), ), supports_check_mode=True ) @@ -152,7 +152,7 @@ def main(): module.fail_json(msg="external_user_name is required for the 'chat' type") # required params for the 'inbox' type - for item in [ 'from_address', 'source', 'subject' ]: + for item in ['from_address', 'source', 'subject']: if module.params[item]: if type == 'chat': module.fail_json(msg="%s is not valid for the 'chat' type" % item) @@ -166,7 +166,7 @@ def main(): params['tags'] = module.params["tags"] # optional params for the 'inbox' type - for item in [ 'from_name', 'reply_to', 'project', 'link' ]: + for item in ['from_name', 'reply_to', 'project', 'link']: if module.params[item]: if type == 'chat': module.fail_json(msg="%s is not valid for the 'chat' type" % item) diff --git a/lib/ansible/modules/notification/grove.py b/lib/ansible/modules/notification/grove.py index 6734039260..d28e780358 100644 --- a/lib/ansible/modules/notification/grove.py +++ b/lib/ansible/modules/notification/grove.py @@ -71,6 +71,7 @@ BASE_URL = 'https://grove.io/api/notice/%s/' # ============================================================== # do_notify_grove + def do_notify_grove(module, channel_token, service, message, url=None, icon_url=None): my_url = BASE_URL % (channel_token,) @@ -88,15 +89,16 @@ def do_notify_grove(module, channel_token, service, message, url=None, icon_url= # ============================================================== # main + def main(): module = AnsibleModule( - argument_spec = dict( - channel_token = dict(type='str', required=True, no_log=True), - message = dict(type='str', required=True), - service = dict(type='str', default='ansible'), - url = dict(type='str', default=None), - icon_url = dict(type='str', default=None), - validate_certs = dict(default='yes', type='bool'), + argument_spec=dict( + channel_token=dict(type='str', required=True, no_log=True), + message=dict(type='str', required=True), + service=dict(type='str', default='ansible'), + url=dict(type='str', default=None), + icon_url=dict(type='str', default=None), + validate_certs=dict(default='yes', type='bool'), ) ) diff --git a/lib/ansible/modules/notification/hall.py b/lib/ansible/modules/notification/hall.py index 69d39d343e..c1f11222b0 100644 --- a/lib/ansible/modules/notification/hall.py +++ b/lib/ansible/modules/notification/hall.py @@ -62,12 +62,12 @@ from ansible.module_utils.basic import AnsibleModule from ansible.module_utils.urls import fetch_url -HALL_API_ENDPOINT = 'https://hall.com/api/1/services/generic/%s' +HALL_API_ENDPOINT = 'https://hall.com/api/1/services/generic/%s' def send_request_to_hall(module, room_token, payload): headers = {'Content-Type': 'application/json'} - payload=module.jsonify(payload) + payload = module.jsonify(payload) api_endpoint = HALL_API_ENDPOINT % (room_token) response, info = fetch_url(module, api_endpoint, data=payload, headers=headers) if info['status'] != 200: diff --git a/lib/ansible/modules/notification/irc.py b/lib/ansible/modules/notification/irc.py index 1bee674bfd..51809d82a0 100644 --- a/lib/ansible/modules/notification/irc.py +++ b/lib/ansible/modules/notification/irc.py @@ -259,11 +259,11 @@ def main(): nick_to=dict(required=False, type='list'), msg=dict(required=True), color=dict(default="none", aliases=['colour'], choices=["white", "black", "blue", - "green", "red", "brown", - "purple", "orange", "yellow", - "light_green", "teal", "light_cyan", - "light_blue", "pink", "gray", - "light_gray", "none"]), + "green", "red", "brown", + "purple", "orange", "yellow", + "light_green", "teal", "light_cyan", + "light_blue", "pink", "gray", + "light_gray", "none"]), style=dict(default="none", choices=["underline", "reverse", "bold", "italic", "none"]), channel=dict(required=False), key=dict(no_log=True), diff --git a/lib/ansible/modules/notification/jabber.py b/lib/ansible/modules/notification/jabber.py index b97815f7fe..f992817674 100644 --- a/lib/ansible/modules/notification/jabber.py +++ b/lib/ansible/modules/notification/jabber.py @@ -105,7 +105,7 @@ def main(): to=dict(required=True), msg=dict(required=True), host=dict(required=False), - port=dict(required=False,default=5222), + port=dict(required=False, default=5222), encoding=dict(required=False), ), supports_check_mode=True @@ -134,15 +134,15 @@ def main(): msg = xmpp.protocol.Message(body=module.params['msg']) try: - conn=xmpp.Client(server, debug=[]) - if not conn.connect(server=(host,port)): + conn = xmpp.Client(server, debug=[]) + if not conn.connect(server=(host, port)): module.fail_json(rc=1, msg='Failed to connect to server: %s' % (server)) - if not conn.auth(user,password,'Ansible'): - module.fail_json(rc=1, msg='Failed to authorize %s on: %s' % (user,server)) + if not conn.auth(user, password, 'Ansible'): + module.fail_json(rc=1, msg='Failed to authorize %s on: %s' % (user, server)) # some old servers require this, also the sleep following send conn.sendInitPresence(requestRoster=0) - if nick: # sending to room instead of user, need to join + if nick: # sending to room instead of user, need to join msg.setType('groupchat') msg.setTag('x', namespace='http://jabber.org/protocol/muc#user') conn.send(xmpp.Presence(to=module.params['to'])) diff --git a/lib/ansible/modules/notification/mattermost.py b/lib/ansible/modules/notification/mattermost.py index 1ecf4d6d5a..6e9b03ba9a 100644 --- a/lib/ansible/modules/notification/mattermost.py +++ b/lib/ansible/modules/notification/mattermost.py @@ -98,54 +98,54 @@ from ansible.module_utils.urls import fetch_url def main(): module = AnsibleModule( supports_check_mode=True, - argument_spec = dict( - url = dict(type='str', required=True), - api_key = dict(type='str', required=True, no_log=True), - text = dict(type='str', required=True), - channel = dict(type='str', default=None), - username = dict(type='str', default='Ansible'), - icon_url = dict(type='str', default='https://www.ansible.com/favicon.ico'), - validate_certs = dict(default='yes', type='bool'), + argument_spec=dict( + url=dict(type='str', required=True), + api_key=dict(type='str', required=True, no_log=True), + text=dict(type='str', required=True), + channel=dict(type='str', default=None), + username=dict(type='str', default='Ansible'), + icon_url=dict(type='str', default='https://www.ansible.com/favicon.ico'), + validate_certs=dict(default='yes', type='bool'), ) ) - #init return dict + # init return dict result = dict(changed=False, msg="OK") - #define webhook - webhook_url = "{0}/hooks/{1}".format(module.params['url'],module.params['api_key']) + # define webhook + webhook_url = "{0}/hooks/{1}".format(module.params['url'], module.params['api_key']) result['webhook_url'] = webhook_url - #define payload - payload = { } + # define payload + payload = {} for param in ['text', 'channel', 'username', 'icon_url']: if module.params[param] is not None: payload[param] = module.params[param] - payload=module.jsonify(payload) + payload = module.jsonify(payload) result['payload'] = payload - #http headers + # http headers headers = { 'Content-Type': 'application/json', 'Accept': 'application/json', } - #notes: - #Nothing is done in check mode - #it'll pass even if your server is down or/and if your token is invalid. - #If someone find good way to check... + # notes: + # Nothing is done in check mode + # it'll pass even if your server is down or/and if your token is invalid. + # If someone find good way to check... - #send request if not in test mode + # send request if not in test mode if module.check_mode is False: response, info = fetch_url(module=module, url=webhook_url, headers=headers, method='POST', data=payload) - #something's wrong + # something's wrong if info['status'] != 200: - #some problem + # some problem result['msg'] = "Failed to send mattermost message, the error was: {0}".format(info['msg']) module.fail_json(**result) - #Looks good + # Looks good module.exit_json(**result) diff --git a/lib/ansible/modules/notification/mqtt.py b/lib/ansible/modules/notification/mqtt.py index 09fbe09ecf..487ecf4f74 100644 --- a/lib/ansible/modules/notification/mqtt.py +++ b/lib/ansible/modules/notification/mqtt.py @@ -144,18 +144,18 @@ def main(): module = AnsibleModule( argument_spec=dict( - server = dict(default = 'localhost'), - port = dict(default = 1883, type='int'), - topic = dict(required = True), - payload = dict(required = True), - client_id = dict(default = None), - qos = dict(default="0", choices=["0", "1", "2"]), - retain = dict(default=False, type='bool'), - username = dict(default = None), - password = dict(default = None, no_log=True), - ca_certs = dict(default = None, type='path'), - certfile = dict(default = None, type='path'), - keyfile = dict(default = None, type='path'), + server=dict(default='localhost'), + port=dict(default=1883, type='int'), + topic=dict(required=True), + payload=dict(required=True), + client_id=dict(default=None), + qos=dict(default="0", choices=["0", "1", "2"]), + retain=dict(default=False, type='bool'), + username=dict(default=None), + password=dict(default=None, no_log=True), + ca_certs=dict(default=None, type='path'), + certfile=dict(default=None, type='path'), + keyfile=dict(default=None, type='path'), ), supports_check_mode=True ) @@ -163,18 +163,18 @@ def main(): if not HAS_PAHOMQTT: module.fail_json(msg="Paho MQTT is not installed") - server = module.params.get("server", 'localhost') - port = module.params.get("port", 1883) - topic = module.params.get("topic") - payload = module.params.get("payload") - client_id = module.params.get("client_id", '') - qos = int(module.params.get("qos", 0)) - retain = module.params.get("retain") - username = module.params.get("username", None) - password = module.params.get("password", None) - ca_certs = module.params.get("ca_certs", None) - certfile = module.params.get("certfile", None) - keyfile = module.params.get("keyfile", None) + server = module.params.get("server", 'localhost') + port = module.params.get("port", 1883) + topic = module.params.get("topic") + payload = module.params.get("payload") + client_id = module.params.get("client_id", '') + qos = int(module.params.get("qos", 0)) + retain = module.params.get("retain") + username = module.params.get("username", None) + password = module.params.get("password", None) + ca_certs = module.params.get("ca_certs", None) + certfile = module.params.get("certfile", None) + keyfile = module.params.get("keyfile", None) if client_id is None: client_id = "%s_%s" % (socket.getfqdn(), os.getpid()) @@ -182,11 +182,11 @@ def main(): if payload and payload == 'None': payload = None - auth=None + auth = None if username is not None: - auth = { 'username' : username, 'password' : password } + auth = {'username': username, 'password': password} - tls=None + tls = None if ca_certs is not None: tls = {'ca_certs': ca_certs, 'certfile': certfile, 'keyfile': keyfile} diff --git a/lib/ansible/modules/notification/osx_say.py b/lib/ansible/modules/notification/osx_say.py index 6920aea15b..bd9ea5de00 100644 --- a/lib/ansible/modules/notification/osx_say.py +++ b/lib/ansible/modules/notification/osx_say.py @@ -48,7 +48,7 @@ import os from ansible.module_utils.basic import AnsibleModule -DEFAULT_VOICE='Alex' +DEFAULT_VOICE = 'Alex' def say(module, msg, voice): @@ -68,7 +68,7 @@ def main(): if not os.path.exists("/usr/bin/say"): module.fail_json(msg="/usr/bin/say is not installed") - msg = module.params['msg'] + msg = module.params['msg'] voice = module.params['voice'] say(module, msg, voice) diff --git a/lib/ansible/modules/notification/pushbullet.py b/lib/ansible/modules/notification/pushbullet.py index b9ff08dc5a..a7d50c7175 100644 --- a/lib/ansible/modules/notification/pushbullet.py +++ b/lib/ansible/modules/notification/pushbullet.py @@ -107,28 +107,28 @@ from ansible.module_utils.basic import AnsibleModule def main(): module = AnsibleModule( - argument_spec = dict( - api_key = dict(type='str', required=True, no_log=True), - channel = dict(type='str', default=None), - device = dict(type='str', default=None), - push_type = dict(type='str', default="note", choices=['note', 'link']), - title = dict(type='str', required=True), - body = dict(type='str', default=None), - url = dict(type='str', default=None), + argument_spec=dict( + api_key=dict(type='str', required=True, no_log=True), + channel=dict(type='str', default=None), + device=dict(type='str', default=None), + push_type=dict(type='str', default="note", choices=['note', 'link']), + title=dict(type='str', required=True), + body=dict(type='str', default=None), + url=dict(type='str', default=None), ), - mutually_exclusive = ( + mutually_exclusive=( ['channel', 'device'], ), supports_check_mode=True ) - api_key = module.params['api_key'] - channel = module.params['channel'] - device = module.params['device'] - push_type = module.params['push_type'] - title = module.params['title'] - body = module.params['body'] - url = module.params['url'] + api_key = module.params['api_key'] + channel = module.params['channel'] + device = module.params['device'] + push_type = module.params['push_type'] + title = module.params['title'] + body = module.params['body'] + url = module.params['url'] if not pushbullet_found: module.fail_json(msg="Python 'pushbullet.py' module is required. Install via: $ pip install pushbullet.py") diff --git a/lib/ansible/modules/notification/pushover.py b/lib/ansible/modules/notification/pushover.py index 8c478c4346..33c775e1d0 100644 --- a/lib/ansible/modules/notification/pushover.py +++ b/lib/ansible/modules/notification/pushover.py @@ -73,12 +73,12 @@ class Pushover(object): # parse config options = dict(user=self.user, - token=self.token, - priority=priority, - message=msg) + token=self.token, + priority=priority, + message=msg) data = urlencode(options) - headers = { "Content-type": "application/x-www-form-urlencoded"} + headers = {"Content-type": "application/x-www-form-urlencoded"} r, info = fetch_url(self.module, url, method='POST', data=data, headers=headers) if info['status'] != 200: raise Exception(info) @@ -93,7 +93,7 @@ def main(): msg=dict(required=True), app_token=dict(required=True, no_log=True), user_key=dict(required=True, no_log=True), - pri=dict(required=False, default='0', choices=['-2','-1','0','1','2']), + pri=dict(required=False, default='0', choices=['-2', '-1', '0', '1', '2']), ), ) diff --git a/lib/ansible/modules/notification/rocketchat.py b/lib/ansible/modules/notification/rocketchat.py index 49d0b77fa2..420b3258b8 100644 --- a/lib/ansible/modules/notification/rocketchat.py +++ b/lib/ansible/modules/notification/rocketchat.py @@ -196,9 +196,10 @@ def build_payload_for_rocketchat(module, text, channel, username, icon_url, icon attachment['fallback'] = attachment['text'] payload['attachments'].append(attachment) - payload="payload=" + module.jsonify(payload) + payload = "payload=" + module.jsonify(payload) return payload + def do_notify_rocketchat(module, domain, token, protocol, payload): if token.count('/') < 1: @@ -210,21 +211,22 @@ def do_notify_rocketchat(module, domain, token, protocol, payload): if info['status'] != 200: module.fail_json(msg="failed to send message, return status=%s" % str(info['status'])) + def main(): module = AnsibleModule( - argument_spec = dict( - domain = dict(type='str', required=True, default=None), - token = dict(type='str', required=True, no_log=True), - protocol = dict(type='str', default='https', choices=['http', 'https']), - msg = dict(type='str', required=False, default=None), - channel = dict(type='str', default=None), - username = dict(type='str', default='Ansible'), - icon_url = dict(type='str', default='https://www.ansible.com/favicon.ico'), - icon_emoji = dict(type='str', default=None), - link_names = dict(type='int', default=1, choices=[0,1]), - validate_certs = dict(default='yes', type='bool'), - color = dict(type='str', default='normal', choices=['normal', 'good', 'warning', 'danger']), - attachments = dict(type='list', required=False, default=None) + argument_spec=dict( + domain=dict(type='str', required=True, default=None), + token=dict(type='str', required=True, no_log=True), + protocol=dict(type='str', default='https', choices=['http', 'https']), + msg=dict(type='str', required=False, default=None), + channel=dict(type='str', default=None), + username=dict(type='str', default='Ansible'), + icon_url=dict(type='str', default='https://www.ansible.com/favicon.ico'), + icon_emoji=dict(type='str', default=None), + link_names=dict(type='int', default=1, choices=[0, 1]), + validate_certs=dict(default='yes', type='bool'), + color=dict(type='str', default='normal', choices=['normal', 'good', 'warning', 'danger']), + attachments=dict(type='list', required=False, default=None) ) ) diff --git a/lib/ansible/modules/notification/sendgrid.py b/lib/ansible/modules/notification/sendgrid.py index 24e48d39fe..88e476a365 100644 --- a/lib/ansible/modules/notification/sendgrid.py +++ b/lib/ansible/modules/notification/sendgrid.py @@ -145,14 +145,14 @@ from ansible.module_utils.urls import fetch_url def post_sendgrid_api(module, username, password, from_address, to_addresses, - subject, body, api_key=None, cc=None, bcc=None, attachments=None, - html_body=False, from_name=None, headers=None): + subject, body, api_key=None, cc=None, bcc=None, attachments=None, + html_body=False, from_name=None, headers=None): if not HAS_SENDGRID: SENDGRID_URI = "https://api.sendgrid.com/api/mail.send.json" AGENT = "Ansible" - data = {'api_user': username, 'api_key':password, - 'from':from_address, 'subject': subject, 'text': body} + data = {'api_user': username, 'api_key': password, + 'from': from_address, 'subject': subject, 'text': body} encoded_data = urlencode(data) to_addresses_api = '' for recipient in to_addresses: @@ -160,9 +160,9 @@ def post_sendgrid_api(module, username, password, from_address, to_addresses, to_addresses_api += '&to[]=%s' % recipient encoded_data += to_addresses_api - headers = { 'User-Agent': AGENT, - 'Content-type': 'application/x-www-form-urlencoded', - 'Accept': 'application/json'} + headers = {'User-Agent': AGENT, + 'Content-type': 'application/x-www-form-urlencoded', + 'Accept': 'application/json'} return fetch_url(module, SENDGRID_URI, data=encoded_data, headers=headers, method='POST') else: @@ -207,6 +207,7 @@ def post_sendgrid_api(module, username, password, from_address, to_addresses, # Main # + def main(): module = AnsibleModule( argument_spec=dict( @@ -225,11 +226,11 @@ def main(): attachments=dict(required=False, type='list') ), supports_check_mode=True, - mutually_exclusive = [ + mutually_exclusive=[ ['api_key', 'password'], ['api_key', 'username'] - ], - required_together = [['username', 'password']], + ], + required_together=[['username', 'password']], ) username = module.params['username'] @@ -253,8 +254,8 @@ def main(): 'api_key, bcc, cc, headers, from_name, html_body, attachments') response, info = post_sendgrid_api(module, username, password, - from_address, to_addresses, subject, body, attachments=attachments, - bcc=bcc, cc=cc, headers=headers, html_body=html_body, api_key=api_key) + from_address, to_addresses, subject, body, attachments=attachments, + bcc=bcc, cc=cc, headers=headers, html_body=html_body, api_key=api_key) if not HAS_SENDGRID: if info['status'] != 200: diff --git a/lib/ansible/modules/notification/slack.py b/lib/ansible/modules/notification/slack.py index 628da6fdfe..c9f7fdedbb 100644 --- a/lib/ansible/modules/notification/slack.py +++ b/lib/ansible/modules/notification/slack.py @@ -189,7 +189,7 @@ escape_table = { def escape_quotes(text): '''Backslash any quotes within text.''' - return "".join(escape_table.get(c,c) for c in text) + return "".join(escape_table.get(c, c) for c in text) def build_payload_for_slack(module, text, channel, username, icon_url, icon_emoji, link_names, parse, color, attachments): @@ -203,7 +203,7 @@ def build_payload_for_slack(module, text, channel, username, icon_url, icon_emoj if (channel[0] == '#') or (channel[0] == '@'): payload['channel'] = channel else: - payload['channel'] = '#'+channel + payload['channel'] = '#' + channel if username is not None: payload['username'] = username if icon_emoji is not None: @@ -237,9 +237,10 @@ def build_payload_for_slack(module, text, channel, username, icon_url, icon_emoj payload['attachments'].append(attachment) - payload=module.jsonify(payload) + payload = module.jsonify(payload) return payload + def do_notify_slack(module, domain, token, payload): if token.count('/') >= 2: # New style token @@ -259,21 +260,22 @@ def do_notify_slack(module, domain, token, payload): obscured_incoming_webhook = SLACK_INCOMING_WEBHOOK % ('[obscured]') module.fail_json(msg=" failed to send %s to %s: %s" % (payload, obscured_incoming_webhook, info['msg'])) + def main(): module = AnsibleModule( - argument_spec = dict( - domain = dict(type='str', required=False, default=None), - token = dict(type='str', required=True, no_log=True), - msg = dict(type='str', required=False, default=None), - channel = dict(type='str', default=None), - username = dict(type='str', default='Ansible'), - icon_url = dict(type='str', default='https://www.ansible.com/favicon.ico'), - icon_emoji = dict(type='str', default=None), - link_names = dict(type='int', default=1, choices=[0,1]), - parse = dict(type='str', default=None, choices=['none', 'full']), - validate_certs = dict(default='yes', type='bool'), - color = dict(type='str', default='normal', choices=['normal', 'good', 'warning', 'danger']), - attachments = dict(type='list', required=False, default=None) + argument_spec=dict( + domain=dict(type='str', required=False, default=None), + token=dict(type='str', required=True, no_log=True), + msg=dict(type='str', required=False, default=None), + channel=dict(type='str', default=None), + username=dict(type='str', default='Ansible'), + icon_url=dict(type='str', default='https://www.ansible.com/favicon.ico'), + icon_emoji=dict(type='str', default=None), + link_names=dict(type='int', default=1, choices=[0, 1]), + parse=dict(type='str', default=None, choices=['none', 'full']), + validate_certs=dict(default='yes', type='bool'), + color=dict(type='str', default='normal', choices=['normal', 'good', 'warning', 'danger']), + attachments=dict(type='list', required=False, default=None) ) ) diff --git a/lib/ansible/modules/notification/twilio.py b/lib/ansible/modules/notification/twilio.py index b9d1c2260a..0fb8f59f30 100644 --- a/lib/ansible/modules/notification/twilio.py +++ b/lib/ansible/modules/notification/twilio.py @@ -112,15 +112,15 @@ def post_twilio_api(module, account_sid, auth_token, msg, from_number, % (account_sid,) AGENT = "Ansible" - data = {'From':from_number, 'To':to_number, 'Body':msg} + data = {'From': from_number, 'To': to_number, 'Body': msg} if media_url: data['MediaUrl'] = media_url encoded_data = urlencode(data) headers = {'User-Agent': AGENT, - 'Content-type': 'application/x-www-form-urlencoded', - 'Accept': 'application/json', - } + 'Content-type': 'application/x-www-form-urlencoded', + 'Accept': 'application/json', + } # Hack module params to have the Basic auth params that fetch_url expects module.params['url_username'] = account_sid.replace('\n', '') @@ -159,7 +159,7 @@ def main(): for number in to_number: r, info = post_twilio_api(module, account_sid, auth_token, msg, - from_number, number, media_url) + from_number, number, media_url) if info['status'] not in [200, 201]: body_message = "unknown error" if 'body' in info: diff --git a/lib/ansible/modules/packaging/language/bundler.py b/lib/ansible/modules/packaging/language/bundler.py index 88013b6ed4..a82394829e 100644 --- a/lib/ansible/modules/packaging/language/bundler.py +++ b/lib/ansible/modules/packaging/language/bundler.py @@ -13,7 +13,7 @@ ANSIBLE_METADATA = {'metadata_version': '1.1', 'supported_by': 'community'} -DOCUMENTATION=''' +DOCUMENTATION = ''' --- module: bundler short_description: Manage Ruby Gem dependencies with Bundler @@ -104,7 +104,7 @@ options: author: "Tim Hoiberg (@thoiberg)" ''' -EXAMPLES=''' +EXAMPLES = ''' # Installs gems from a Gemfile in the current directory - bundler: state: present @@ -159,7 +159,7 @@ def main(): extra_args=dict(default=None, required=False), ), supports_check_mode=True - ) + ) state = module.params.get('state') chdir = module.params.get('chdir') diff --git a/lib/ansible/modules/packaging/language/cpanm.py b/lib/ansible/modules/packaging/language/cpanm.py index 3195775561..fcd22b6812 100644 --- a/lib/ansible/modules/packaging/language/cpanm.py +++ b/lib/ansible/modules/packaging/language/cpanm.py @@ -141,6 +141,7 @@ def _is_package_installed(module, name, locallib, cpanm, version): res, stdout, stderr = module.run_command(cmd, check_rc=False) return res == 0 + def _build_cmd_line(name, from_path, notest, locallib, mirror, mirror_only, installdeps, cpanm, use_sudo): # this code should use "%s" like everything else and just return early but not fixing all of it now. # don't copy stuff like this @@ -197,23 +198,23 @@ def main(): required_one_of=[['name', 'from_path']], ) - cpanm = _get_cpanm_path(module) - name = module.params['name'] - from_path = module.params['from_path'] - notest = module.boolean(module.params.get('notest', False)) - locallib = module.params['locallib'] - mirror = module.params['mirror'] + cpanm = _get_cpanm_path(module) + name = module.params['name'] + from_path = module.params['from_path'] + notest = module.boolean(module.params.get('notest', False)) + locallib = module.params['locallib'] + mirror = module.params['mirror'] mirror_only = module.params['mirror_only'] installdeps = module.params['installdeps'] - use_sudo = module.params['system_lib'] - version = module.params['version'] + use_sudo = module.params['system_lib'] + version = module.params['version'] - changed = False + changed = False installed = _is_package_installed(module, name, locallib, cpanm, version) if not installed: - cmd = _build_cmd_line(name, from_path, notest, locallib, mirror, mirror_only, installdeps, cpanm, use_sudo) + cmd = _build_cmd_line(name, from_path, notest, locallib, mirror, mirror_only, installdeps, cpanm, use_sudo) rc_cpanm, out_cpanm, err_cpanm = module.run_command(cmd, check_rc=False) diff --git a/lib/ansible/modules/packaging/language/gem.py b/lib/ansible/modules/packaging/language/gem.py index 900b3ac696..4e27004024 100644 --- a/lib/ansible/modules/packaging/language/gem.py +++ b/lib/ansible/modules/packaging/language/gem.py @@ -121,8 +121,9 @@ def get_rubygems_path(module): result = [module.get_bin_path('gem', True)] return result + def get_rubygems_version(module): - cmd = get_rubygems_path(module) + [ '--version' ] + cmd = get_rubygems_path(module) + ['--version'] (rc, out, err) = module.run_command(cmd, check_rc=True) match = re.match(r'^(\d+)\.(\d+)\.(\d+)', out) @@ -131,6 +132,7 @@ def get_rubygems_version(module): return tuple(int(x) for x in match.groups()) + def get_installed_versions(module, remote=False): cmd = get_rubygems_path(module) @@ -138,7 +140,7 @@ def get_installed_versions(module, remote=False): if remote: cmd.append('--remote') if module.params['repository']: - cmd.extend([ '--source', module.params['repository'] ]) + cmd.extend(['--source', module.params['repository']]) cmd.append('-n') cmd.append('^%s$' % module.params['name']) (rc, out, err) = module.run_command(cmd, check_rc=True) @@ -151,6 +153,7 @@ def get_installed_versions(module, remote=False): installed_versions.append(version.split()[0]) return installed_versions + def exists(module): if module.params['state'] == 'latest': @@ -166,6 +169,7 @@ def exists(module): return True return False + def uninstall(module): if module.check_mode: @@ -173,13 +177,14 @@ def uninstall(module): cmd = get_rubygems_path(module) cmd.append('uninstall') if module.params['version']: - cmd.extend([ '--version', module.params['version'] ]) + cmd.extend(['--version', module.params['version']]) else: cmd.append('--all') cmd.append('--executable') cmd.append(module.params['name']) module.run_command(cmd, check_rc=True) + def install(module): if module.check_mode: @@ -194,9 +199,9 @@ def install(module): cmd = get_rubygems_path(module) cmd.append('install') if module.params['version']: - cmd.extend([ '--version', module.params['version'] ]) + cmd.extend(['--version', module.params['version']]) if module.params['repository']: - cmd.extend([ '--source', module.params['repository'] ]) + cmd.extend(['--source', module.params['repository']]) if not module.params['include_dependencies']: cmd.append('--ignore-dependencies') else: @@ -218,28 +223,29 @@ def install(module): cmd.append('--env-shebang') cmd.append(module.params['gem_source']) if module.params['build_flags']: - cmd.extend([ '--', module.params['build_flags'] ]) + cmd.extend(['--', module.params['build_flags']]) module.run_command(cmd, check_rc=True) + def main(): module = AnsibleModule( - argument_spec = dict( - executable = dict(required=False, type='path'), - gem_source = dict(required=False, type='path'), - include_dependencies = dict(required=False, default=True, type='bool'), - name = dict(required=True, type='str'), - repository = dict(required=False, aliases=['source'], type='str'), - state = dict(required=False, default='present', choices=['present','absent','latest'], type='str'), - user_install = dict(required=False, default=True, type='bool'), - pre_release = dict(required=False, default=False, type='bool'), - include_doc = dict(required=False, default=False, type='bool'), - env_shebang = dict(required=False, default=False, type='bool'), - version = dict(required=False, type='str'), - build_flags = dict(required=False, type='str'), + argument_spec=dict( + executable=dict(required=False, type='path'), + gem_source=dict(required=False, type='path'), + include_dependencies=dict(required=False, default=True, type='bool'), + name=dict(required=True, type='str'), + repository=dict(required=False, aliases=['source'], type='str'), + state=dict(required=False, default='present', choices=['present', 'absent', 'latest'], type='str'), + user_install=dict(required=False, default=True, type='bool'), + pre_release=dict(required=False, default=False, type='bool'), + include_doc=dict(required=False, default=False, type='bool'), + env_shebang=dict(required=False, default=False, type='bool'), + version=dict(required=False, type='str'), + build_flags=dict(required=False, type='str'), ), - supports_check_mode = True, - mutually_exclusive = [ ['gem_source','repository'], ['gem_source','version'] ], + supports_check_mode=True, + mutually_exclusive=[['gem_source', 'repository'], ['gem_source', 'version']], ) if module.params['version'] and module.params['state'] == 'latest': @@ -252,7 +258,7 @@ def main(): changed = False - if module.params['state'] in [ 'present', 'latest']: + if module.params['state'] in ['present', 'latest']: if not exists(module): install(module) changed = True diff --git a/lib/ansible/modules/packaging/language/maven_artifact.py b/lib/ansible/modules/packaging/language/maven_artifact.py index 540b63a458..2a8cd8bbc2 100644 --- a/lib/ansible/modules/packaging/language/maven_artifact.py +++ b/lib/ansible/modules/packaging/language/maven_artifact.py @@ -315,12 +315,12 @@ class MavenDownloader: def _request(self, url, failmsg, f): url_to_use = url parsed_url = urlparse(url) - if parsed_url.scheme=='s3': + if parsed_url.scheme == 's3': parsed_url = urlparse(url) bucket_name = parsed_url.netloc key_name = parsed_url.path[1:] - client = boto3.client('s3',aws_access_key_id=self.module.params.get('username', ''), aws_secret_access_key=self.module.params.get('password', '')) - url_to_use = client.generate_presigned_url('get_object',Params={'Bucket':bucket_name,'Key':key_name},ExpiresIn=10) + client = boto3.client('s3', aws_access_key_id=self.module.params.get('username', ''), aws_secret_access_key=self.module.params.get('password', '')) + url_to_use = client.generate_presigned_url('get_object', Params={'Bucket': bucket_name, 'Key': key_name}, ExpiresIn=10) req_timeout = self.module.params.get('timeout') @@ -335,7 +335,6 @@ class MavenDownloader: else: return f(response) - def download(self, artifact, filename=None): filename = artifact.get_filename(filename) if not artifact.version or artifact.version == "latest": @@ -402,20 +401,20 @@ class MavenDownloader: def main(): module = AnsibleModule( - argument_spec = dict( - group_id = dict(default=None), - artifact_id = dict(default=None), - version = dict(default="latest"), - classifier = dict(default=''), - extension = dict(default='jar'), - repository_url = dict(default=None), - username = dict(default=None,aliases=['aws_secret_key']), - password = dict(default=None, no_log=True,aliases=['aws_secret_access_key']), - state = dict(default="present", choices=["present","absent"]), # TODO - Implement a "latest" state - timeout = dict(default=10, type='int'), - dest = dict(type="path", default=None), - validate_certs = dict(required=False, default=True, type='bool'), - keep_name = dict(required=False, default=False, type='bool'), + argument_spec=dict( + group_id=dict(default=None), + artifact_id=dict(default=None), + version=dict(default="latest"), + classifier=dict(default=''), + extension=dict(default='jar'), + repository_url=dict(default=None), + username=dict(default=None, aliases=['aws_secret_key']), + password=dict(default=None, no_log=True, aliases=['aws_secret_access_key']), + state=dict(default="present", choices=["present", "absent"]), # TODO - Implement a "latest" state + timeout=dict(default=10, type='int'), + dest=dict(type="path", default=None), + validate_certs=dict(required=False, default=True, type='bool'), + keep_name=dict(required=False, default=False, type='bool'), ), add_file_common_args=True ) @@ -429,7 +428,7 @@ def main(): except AttributeError as e: module.fail_json(msg='url parsing went wrong %s' % e) - if parsed_url.scheme=='s3' and not HAS_BOTO: + if parsed_url.scheme == 's3' and not HAS_BOTO: module.fail_json(msg='boto3 required for this module, when using s3:// repository URLs') group_id = module.params["group_id"] diff --git a/lib/ansible/modules/packaging/language/pear.py b/lib/ansible/modules/packaging/language/pear.py index a4b738d170..0d0f19a8db 100644 --- a/lib/ansible/modules/packaging/language/pear.py +++ b/lib/ansible/modules/packaging/language/pear.py @@ -83,6 +83,7 @@ def get_local_version(pear_output): return installed return None + def _get_pear_path(module): if module.params['executable'] and os.path.isfile(module.params['executable']): result = module.params['executable'] @@ -90,6 +91,7 @@ def _get_pear_path(module): result = module.get_bin_path('pear', True, [module.params['executable']]) return result + def get_repository_version(pear_output): """Take pear remote-info output and get the latest version""" lines = pear_output.split('\n') @@ -98,6 +100,7 @@ def get_repository_version(pear_output): return line.rsplit(None, 1)[-1].strip() return None + def query_package(module, name, state="present"): """Query the package status in both the local system and the repository. Returns a boolean to indicate if the package is installed, @@ -198,18 +201,14 @@ def check_packages(module, packages, state): module.exit_json(change=False, msg="package(s) already %s" % state) - - def main(): module = AnsibleModule( - argument_spec = dict( - name = dict(aliases=['pkg']), - state = dict(default='present', choices=['present', 'installed', "latest", 'absent', 'removed']), - executable = dict(default=None, required=False, type='path')), - required_one_of = [['name']], - supports_check_mode = True) - - + argument_spec=dict( + name=dict(aliases=['pkg']), + state=dict(default='present', choices=['present', 'installed', "latest", 'absent', 'removed']), + executable=dict(default=None, required=False, type='path')), + required_one_of=[['name']], + supports_check_mode=True) p = module.params diff --git a/lib/ansible/modules/packaging/os/apk.py b/lib/ansible/modules/packaging/os/apk.py index 0d1e1f73f2..8015abe642 100644 --- a/lib/ansible/modules/packaging/os/apk.py +++ b/lib/ansible/modules/packaging/os/apk.py @@ -147,6 +147,7 @@ import re # Import module snippets. from ansible.module_utils.basic import AnsibleModule + def parse_for_packages(stdout): packages = [] data = stdout.split('\n') @@ -157,6 +158,7 @@ def parse_for_packages(stdout): packages.append(p.group(1)) return packages + def update_package_db(module, exit): cmd = "%s update" % (APK_PATH) rc, stdout, stderr = module.run_command(cmd, check_rc=False) @@ -167,6 +169,7 @@ def update_package_db(module, exit): else: return True + def query_package(module, name): cmd = "%s -v info --installed %s" % (APK_PATH, name) rc, stdout, stderr = module.run_command(cmd, check_rc=False) @@ -175,6 +178,7 @@ def query_package(module, name): else: return False + def query_latest(module, name): cmd = "%s version %s" % (APK_PATH, name) rc, stdout, stderr = module.run_command(cmd, check_rc=False) @@ -184,6 +188,7 @@ def query_latest(module, name): return False return True + def query_virtual(module, name): cmd = "%s -v info --description %s" % (APK_PATH, name) rc, stdout, stderr = module.run_command(cmd, check_rc=False) @@ -192,6 +197,7 @@ def query_virtual(module, name): return True return False + def get_dependencies(module, name): cmd = "%s -v info --depends %s" % (APK_PATH, name) rc, stdout, stderr = module.run_command(cmd, check_rc=False) @@ -201,6 +207,7 @@ def get_dependencies(module, name): else: return [] + def upgrade_packages(module, available): if module.check_mode: cmd = "%s upgrade --simulate" % (APK_PATH) @@ -216,6 +223,7 @@ def upgrade_packages(module, available): module.exit_json(changed=False, msg="packages already upgraded", stdout=stdout, stderr=stderr, packages=packagelist) module.exit_json(changed=True, msg="upgraded packages", stdout=stdout, stderr=stderr, packages=packagelist) + def install_packages(module, names, state): upgrade = False to_install = [] @@ -254,6 +262,7 @@ def install_packages(module, names, state): module.fail_json(msg="failed to install %s" % (packages), stdout=stdout, stderr=stderr, packages=packagelist) module.exit_json(changed=True, msg="installed %s package(s)" % (packages), stdout=stdout, stderr=stderr, packages=packagelist) + def remove_packages(module, names): installed = [] for name in names: @@ -275,6 +284,7 @@ def remove_packages(module, names): # ========================================== # Main control flow. + def main(): module = AnsibleModule( argument_spec=dict( diff --git a/lib/ansible/modules/packaging/os/dpkg_selections.py b/lib/ansible/modules/packaging/os/dpkg_selections.py index 5a264672b8..b4f1f4f893 100644 --- a/lib/ansible/modules/packaging/os/dpkg_selections.py +++ b/lib/ansible/modules/packaging/os/dpkg_selections.py @@ -41,6 +41,7 @@ EXAMPLES = ''' selection: hold ''' + def main(): module = AnsibleModule( argument_spec=dict( diff --git a/lib/ansible/modules/packaging/os/homebrew.py b/lib/ansible/modules/packaging/os/homebrew.py index 5861eaed96..ba7a039ccd 100644 --- a/lib/ansible/modules/packaging/os/homebrew.py +++ b/lib/ansible/modules/packaging/os/homebrew.py @@ -181,9 +181,9 @@ class Homebrew(object): @ # at-sign ''' - INVALID_PATH_REGEX = _create_regex_group(VALID_PATH_CHARS) - INVALID_BREW_PATH_REGEX = _create_regex_group(VALID_BREW_PATH_CHARS) - INVALID_PACKAGE_REGEX = _create_regex_group(VALID_PACKAGE_CHARS) + INVALID_PATH_REGEX = _create_regex_group(VALID_PATH_CHARS) + INVALID_BREW_PATH_REGEX = _create_regex_group(VALID_BREW_PATH_CHARS) + INVALID_PACKAGE_REGEX = _create_regex_group(VALID_PACKAGE_CHARS) # /class regexes ----------------------------------------------- }}} # class validations -------------------------------------------- {{{ diff --git a/lib/ansible/modules/packaging/os/homebrew_cask.py b/lib/ansible/modules/packaging/os/homebrew_cask.py index 7e43be37b0..0d3ebd590b 100644 --- a/lib/ansible/modules/packaging/os/homebrew_cask.py +++ b/lib/ansible/modules/packaging/os/homebrew_cask.py @@ -137,9 +137,9 @@ class HomebrewCask(object): - # dashes ''' - INVALID_PATH_REGEX = _create_regex_group(VALID_PATH_CHARS) - INVALID_BREW_PATH_REGEX = _create_regex_group(VALID_BREW_PATH_CHARS) - INVALID_CASK_REGEX = _create_regex_group(VALID_CASK_CHARS) + INVALID_PATH_REGEX = _create_regex_group(VALID_PATH_CHARS) + INVALID_BREW_PATH_REGEX = _create_regex_group(VALID_BREW_PATH_CHARS) + INVALID_CASK_REGEX = _create_regex_group(VALID_CASK_CHARS) # /class regexes ----------------------------------------------- }}} # class validations -------------------------------------------- {{{ @@ -589,7 +589,7 @@ def main(): for install_option in p['install_options']] brew_cask = HomebrewCask(module=module, path=path, casks=casks, - state=state, update_homebrew=update_homebrew, + state=state, update_homebrew=update_homebrew, install_options=install_options) (failed, changed, message) = brew_cask.run() if failed: diff --git a/lib/ansible/modules/packaging/os/layman.py b/lib/ansible/modules/packaging/os/layman.py index ef84797be9..7cdf1833e7 100644 --- a/lib/ansible/modules/packaging/os/layman.py +++ b/lib/ansible/modules/packaging/os/layman.py @@ -158,8 +158,8 @@ def install_overlay(module, name, list_url=None): if not layman.is_repo(name): if not list_url: - raise ModuleError("Overlay '%s' is not on the list of known " \ - "overlays and URL of the remote list was not provided." % name) + raise ModuleError("Overlay '%s' is not on the list of known " + "overlays and URL of the remote list was not provided." % name) overlay_defs = layman_conf.get_option('overlay_defs') dest = path.join(overlay_defs, name + '.xml') @@ -209,7 +209,7 @@ def sync_overlay(name): layman = init_layman() if not layman.sync(name): - messages = [ str(item[1]) for item in layman.sync_results[2] ] + messages = [str(item[1]) for item in layman.sync_results[2]] raise ModuleError(messages) @@ -227,11 +227,11 @@ def sync_overlays(): def main(): # define module module = AnsibleModule( - argument_spec = dict( - name = dict(required=True), - list_url = dict(aliases=['url']), - state = dict(default="present", choices=['present', 'absent', 'updated']), - validate_certs = dict(required=False, default=True, type='bool'), + argument_spec=dict( + name=dict(required=True), + list_url=dict(aliases=['url']), + state=dict(default="present", choices=['present', 'absent', 'updated']), + validate_certs=dict(required=False, default=True, type='bool'), ), supports_check_mode=True ) diff --git a/lib/ansible/modules/packaging/os/macports.py b/lib/ansible/modules/packaging/os/macports.py index 34a5683124..fc21ac5ff9 100644 --- a/lib/ansible/modules/packaging/os/macports.py +++ b/lib/ansible/modules/packaging/os/macports.py @@ -68,6 +68,7 @@ EXAMPLES = ''' import pipes + def update_package_db(module, port_path): """ Updates packages list. """ @@ -196,10 +197,10 @@ def deactivate_packages(module, port_path, packages): def main(): module = AnsibleModule( - argument_spec = dict( - name = dict(aliases=["pkg"], required=True), - state = dict(default="present", choices=["present", "installed", "absent", "removed", "active", "inactive"]), - update_cache = dict(default="no", aliases=["update-cache"], type='bool') + argument_spec=dict( + name=dict(aliases=["pkg"], required=True), + state=dict(default="present", choices=["present", "installed", "absent", "removed", "active", "inactive"]), + update_cache=dict(default="no", aliases=["update-cache"], type='bool') ) ) diff --git a/lib/ansible/modules/packaging/os/opkg.py b/lib/ansible/modules/packaging/os/opkg.py index 9933cc807c..7fbe78702c 100644 --- a/lib/ansible/modules/packaging/os/opkg.py +++ b/lib/ansible/modules/packaging/os/opkg.py @@ -89,6 +89,7 @@ EXAMPLES = ''' import pipes + def update_package_db(module, opkg_path): """ Updates packages list. """ diff --git a/lib/ansible/modules/packaging/os/pkgin.py b/lib/ansible/modules/packaging/os/pkgin.py index d9eb723c64..3bccaeb677 100644 --- a/lib/ansible/modules/packaging/os/pkgin.py +++ b/lib/ansible/modules/packaging/os/pkgin.py @@ -132,6 +132,7 @@ EXAMPLES = ''' import re + def query_package(module, name): """Search for the package by name. @@ -201,8 +202,8 @@ def query_package(module, name): def format_action_message(module, action, count): - vars = { "actioned": action, - "count": count } + vars = {"actioned": action, + "count": count} if module.check_mode: message = "would have %(actioned)s %(count)d package" % vars @@ -227,10 +228,10 @@ def format_pkgin_command(module, command, package=None): else: force = "" - vars = { "pkgin": PKGIN_PATH, - "command": command, - "package": package, - "force": force} + vars = {"pkgin": PKGIN_PATH, + "command": command, + "package": package, + "force": force} if module.check_mode: return "%(pkgin)s -n %(command)s %(package)s" % vars @@ -283,6 +284,7 @@ def install_packages(module, packages): module.exit_json(changed=False, msg="package(s) already present") + def update_package_db(module): rc, out, err = module.run_command( format_pkgin_command(module, "update")) @@ -295,6 +297,7 @@ def update_package_db(module): else: module.fail_json(msg="could not update package db") + def do_upgrade_packages(module, full=False): if full: cmd = "full-upgrade" @@ -310,12 +313,15 @@ def do_upgrade_packages(module, full=False): else: module.fail_json(msg="could not %s packages" % cmd) + def upgrade_packages(module): do_upgrade_packages(module) + def full_upgrade_packages(module): do_upgrade_packages(module, True) + def clean_cache(module): rc, out, err = module.run_command( format_pkgin_command(module, "clean")) @@ -327,18 +333,19 @@ def clean_cache(module): else: module.fail_json(msg="could not clean package cache") + def main(): module = AnsibleModule( - argument_spec = dict( - state = dict(default="present", choices=["present","absent"]), - name = dict(aliases=["pkg"], type='list'), - update_cache = dict(default='no', type='bool'), - upgrade = dict(default='no', type='bool'), - full_upgrade = dict(default='no', type='bool'), - clean = dict(default='no', type='bool'), - force = dict(default='no', type='bool')), - required_one_of = [['name', 'update_cache', 'upgrade', 'full_upgrade', 'clean']], - supports_check_mode = True) + argument_spec=dict( + state=dict(default="present", choices=["present", "absent"]), + name=dict(aliases=["pkg"], type='list'), + update_cache=dict(default='no', type='bool'), + upgrade=dict(default='no', type='bool'), + full_upgrade=dict(default='no', type='bool'), + clean=dict(default='no', type='bool'), + force=dict(default='no', type='bool')), + required_one_of=[['name', 'update_cache', 'upgrade', 'full_upgrade', 'clean']], + supports_check_mode=True) global PKGIN_PATH PKGIN_PATH = module.get_bin_path('pkgin', True, ['/opt/local/bin']) diff --git a/lib/ansible/modules/packaging/os/pkgng.py b/lib/ansible/modules/packaging/os/pkgng.py index 2f2f7ec8c1..a6e80ca9d0 100644 --- a/lib/ansible/modules/packaging/os/pkgng.py +++ b/lib/ansible/modules/packaging/os/pkgng.py @@ -111,6 +111,7 @@ EXAMPLES = ''' import re from ansible.module_utils.basic import AnsibleModule + def query_package(module, pkgng_path, name, dir_arg): rc, out, err = module.run_command("%s %s info -g -e %s" % (pkgng_path, dir_arg, name)) @@ -120,6 +121,7 @@ def query_package(module, pkgng_path, name, dir_arg): return False + def pkgng_older_than(module, pkgng_path, compare_version): rc, out, err = module.run_command("%s -v" % pkgng_path) @@ -206,6 +208,7 @@ def install_packages(module, pkgng_path, packages, cached, pkgsite, dir_arg): return (False, "package(s) already present") + def annotation_query(module, pkgng_path, package, tag, dir_arg): rc, out, err = module.run_command("%s %s info -g -A %s" % (pkgng_path, dir_arg, package)) match = re.search(r'^\s*(?P%s)\s*:\s*(?P\w+)' % tag, out, flags=re.MULTILINE) @@ -219,10 +222,10 @@ def annotation_add(module, pkgng_path, package, tag, value, dir_arg): if not _value: # Annotation does not exist, add it. rc, out, err = module.run_command('%s %s annotate -y -A %s %s "%s"' - % (pkgng_path, dir_arg, package, tag, value)) + % (pkgng_path, dir_arg, package, tag, value)) if rc != 0: module.fail_json(msg="could not annotate %s: %s" - % (package, out), stderr=err) + % (package, out), stderr=err) return True elif _value != value: # Annotation exists, but value differs @@ -234,41 +237,43 @@ def annotation_add(module, pkgng_path, package, tag, value, dir_arg): # Annotation exists, nothing to do return False + def annotation_delete(module, pkgng_path, package, tag, value, dir_arg): _value = annotation_query(module, pkgng_path, package, tag, dir_arg) if _value: rc, out, err = module.run_command('%s %s annotate -y -D %s %s' - % (pkgng_path, dir_arg, package, tag)) + % (pkgng_path, dir_arg, package, tag)) if rc != 0: module.fail_json(msg="could not delete annotation to %s: %s" - % (package, out), stderr=err) + % (package, out), stderr=err) return True return False + def annotation_modify(module, pkgng_path, package, tag, value, dir_arg): _value = annotation_query(module, pkgng_path, package, tag, dir_arg) if not value: # No such tag module.fail_json(msg="could not change annotation to %s: tag %s does not exist" - % (package, tag)) + % (package, tag)) elif _value == value: # No change in value return False else: - rc,out,err = module.run_command('%s %s annotate -y -M %s %s "%s"' - % (pkgng_path, dir_arg, package, tag, value)) + rc, out, err = module.run_command('%s %s annotate -y -M %s %s "%s"' + % (pkgng_path, dir_arg, package, tag, value)) if rc != 0: module.fail_json(msg="could not change annotation annotation to %s: %s" - % (package, out), stderr=err) + % (package, out), stderr=err) return True def annotate_packages(module, pkgng_path, packages, annotation, dir_arg): annotate_c = 0 annotations = map(lambda _annotation: - re.match(r'(?P[\+-:])(?P\w+)(=(?P\w+))?', - _annotation).groupdict(), - re.split(r',', annotation)) + re.match(r'(?P[\+-:])(?P\w+)(=(?P\w+))?', + _annotation).groupdict(), + re.split(r',', annotation)) operation = { '+': annotation_add, @@ -285,6 +290,7 @@ def annotate_packages(module, pkgng_path, packages, annotation, dir_arg): return (True, "added %s annotations." % annotate_c) return (False, "changed no annotations") + def autoremove_packages(module, pkgng_path, dir_arg): rc, out, err = module.run_command("%s %s autoremove -n" % (pkgng_path, dir_arg)) @@ -302,20 +308,21 @@ def autoremove_packages(module, pkgng_path, dir_arg): return True, "autoremoved %d package(s)" % (autoremove_c) + def main(): module = AnsibleModule( - argument_spec = dict( - state = dict(default="present", choices=["present","absent"], required=False), - name = dict(aliases=["pkg"], required=True, type='list'), - cached = dict(default=False, type='bool'), - annotation = dict(default="", required=False), - pkgsite = dict(default="", required=False), - rootdir = dict(default="", required=False, type='path'), - chroot = dict(default="", required=False, type='path'), - jail = dict(default="", required=False, type='str'), - autoremove = dict(default=False, type='bool')), - supports_check_mode = True, - mutually_exclusive =[["rootdir", "chroot", "jail"]]) + argument_spec=dict( + state=dict(default="present", choices=["present", "absent"], required=False), + name=dict(aliases=["pkg"], required=True, type='list'), + cached=dict(default=False, type='bool'), + annotation=dict(default="", required=False), + pkgsite=dict(default="", required=False), + rootdir=dict(default="", required=False, type='path'), + chroot=dict(default="", required=False, type='path'), + jail=dict(default="", required=False, type='str'), + autoremove=dict(default=False, type='bool')), + supports_check_mode=True, + mutually_exclusive=[["rootdir", "chroot", "jail"]]) pkgng_path = module.get_bin_path('pkg', True) diff --git a/lib/ansible/modules/packaging/os/pkgutil.py b/lib/ansible/modules/packaging/os/pkgutil.py index 0a3d054ef5..e6fbeb16a7 100644 --- a/lib/ansible/modules/packaging/os/pkgutil.py +++ b/lib/ansible/modules/packaging/os/pkgutil.py @@ -69,6 +69,7 @@ EXAMPLES = ''' import os import pipes + def package_installed(module, name): cmd = ['pkginfo'] cmd.append('-q') @@ -79,11 +80,12 @@ def package_installed(module, name): else: return False + def package_latest(module, name, site): # Only supports one package - cmd = [ 'pkgutil', '-U', '--single', '-c' ] + cmd = ['pkgutil', '-U', '--single', '-c'] if site is not None: - cmd += [ '-t', site] + cmd += ['-t', site] cmd.append(name) rc, out, err = run_command(module, cmd) # replace | tail -1 |grep -v SAME @@ -91,45 +93,50 @@ def package_latest(module, name, site): # at the end of the list return 'SAME' in out.split('\n')[-2] + def run_command(module, cmd, **kwargs): progname = cmd[0] cmd[0] = module.get_bin_path(progname, True, ['/opt/csw/bin']) return module.run_command(cmd, **kwargs) + def package_install(module, state, name, site, update_catalog): - cmd = [ 'pkgutil', '-iy' ] + cmd = ['pkgutil', '-iy'] if update_catalog: - cmd += [ '-U' ] + cmd += ['-U'] if site is not None: - cmd += [ '-t', site ] + cmd += ['-t', site] if state == 'latest': - cmd += [ '-f' ] + cmd += ['-f'] cmd.append(name) (rc, out, err) = run_command(module, cmd) return (rc, out, err) + def package_upgrade(module, name, site, update_catalog): - cmd = [ 'pkgutil', '-ufy' ] + cmd = ['pkgutil', '-ufy'] if update_catalog: - cmd += [ '-U' ] + cmd += ['-U'] if site is not None: - cmd += [ '-t', site ] + cmd += ['-t', site] cmd.append(name) (rc, out, err) = run_command(module, cmd) return (rc, out, err) + def package_uninstall(module, name): - cmd = [ 'pkgutil', '-ry', name] + cmd = ['pkgutil', '-ry', name] (rc, out, err) = run_command(module, cmd) return (rc, out, err) + def main(): module = AnsibleModule( - argument_spec = dict( - name = dict(required = True), - state = dict(required = True, choices=['present', 'absent','latest']), - site = dict(default = None), - update_catalog = dict(required = False, default = False, type='bool'), + argument_spec=dict( + name=dict(required=True), + state=dict(required=True, choices=['present', 'absent', 'latest']), + site=dict(default=None), + update_catalog=dict(required=False, default=False, type='bool'), ), supports_check_mode=True ) diff --git a/lib/ansible/modules/packaging/os/portage.py b/lib/ansible/modules/packaging/os/portage.py index bc4f06bdc5..0cded3b76a 100644 --- a/lib/ansible/modules/packaging/os/portage.py +++ b/lib/ansible/modules/packaging/os/portage.py @@ -289,7 +289,7 @@ def sync_repositories(module, webrsync=False): def emerge_packages(module, packages): p = module.params - if not (p['update'] or p['noreplace'] or p['state']=='latest'): + if not (p['update'] or p['noreplace'] or p['state'] == 'latest'): for package in packages: if not query_package(module, package, 'emerge'): break @@ -319,7 +319,7 @@ def emerge_packages(module, packages): if p[flag]: args.append(arg) - if p['state'] and p['state']=='latest': + if p['state'] and p['state'] == 'latest': args.append("--update") if p['usepkg'] and p['usepkgonly']: diff --git a/lib/ansible/modules/packaging/os/portinstall.py b/lib/ansible/modules/packaging/os/portinstall.py index d937ed2b2d..fd099fe522 100644 --- a/lib/ansible/modules/packaging/os/portinstall.py +++ b/lib/ansible/modules/packaging/os/portinstall.py @@ -138,7 +138,7 @@ def remove_packages(module, packages): name_without_digits = re.sub('[0-9]', '', package) rc, out, err = module.run_command("%s `%s %s`" % (pkg_delete_path, pkg_glob_path, shlex_quote(name_without_digits)), - use_unsafe_shell=True) + use_unsafe_shell=True) if query_package(module, package): module.fail_json(msg="failed to remove %s: %s" % (package, out)) @@ -164,9 +164,9 @@ def install_packages(module, packages, use_packages): portinstall_path = module.get_bin_path('portinstall', True) if use_packages == "yes": - portinstall_params="--use-packages" + portinstall_params = "--use-packages" else: - portinstall_params="" + portinstall_params = "" for package in packages: if query_package(module, package): @@ -193,10 +193,10 @@ def install_packages(module, packages, use_packages): def main(): module = AnsibleModule( - argument_spec = dict( - state = dict(default="present", choices=["present","absent"]), - name = dict(aliases=["pkg"], required=True), - use_packages = dict(type='bool', default='yes'))) + argument_spec=dict( + state=dict(default="present", choices=["present", "absent"]), + name=dict(aliases=["pkg"], required=True), + use_packages=dict(type='bool', default='yes'))) p = module.params diff --git a/lib/ansible/modules/packaging/os/slackpkg.py b/lib/ansible/modules/packaging/os/slackpkg.py index ad65ee50d6..20a22843be 100644 --- a/lib/ansible/modules/packaging/os/slackpkg.py +++ b/lib/ansible/modules/packaging/os/slackpkg.py @@ -96,7 +96,7 @@ def remove_packages(module, slackpkg_path, packages): if not module.check_mode: rc, out, err = module.run_command("%s -default_answer=y -batch=on \ remove %s" % (slackpkg_path, - package)) + package)) if not module.check_mode and query_package(module, slackpkg_path, package): @@ -122,7 +122,7 @@ def install_packages(module, slackpkg_path, packages): if not module.check_mode: rc, out, err = module.run_command("%s -default_answer=y -batch=on \ install %s" % (slackpkg_path, - package)) + package)) if not module.check_mode and not query_package(module, slackpkg_path, package): @@ -145,7 +145,7 @@ def upgrade_packages(module, slackpkg_path, packages): if not module.check_mode: rc, out, err = module.run_command("%s -default_answer=y -batch=on \ upgrade %s" % (slackpkg_path, - package)) + package)) if not module.check_mode and not query_package(module, slackpkg_path, package): diff --git a/lib/ansible/modules/packaging/os/sorcery.py b/lib/ansible/modules/packaging/os/sorcery.py index a148612123..26f37348b7 100644 --- a/lib/ansible/modules/packaging/os/sorcery.py +++ b/lib/ansible/modules/packaging/os/sorcery.py @@ -607,17 +607,17 @@ def manage_spells(module): def main(): module = AnsibleModule( - argument_spec = dict( - name = dict(default=None, aliases=['spell'], type='list'), - state = dict(default='present', choices=['present', 'latest', - 'absent', 'cast', 'dispelled', 'rebuild']), - depends = dict(default=None), - update = dict(default=False, type='bool'), - update_cache = dict(default=False, aliases=['update_codex'], type='bool'), - cache_valid_time = dict(default=0, type='int') + argument_spec=dict( + name=dict(default=None, aliases=['spell'], type='list'), + state=dict(default='present', choices=['present', 'latest', + 'absent', 'cast', 'dispelled', 'rebuild']), + depends=dict(default=None), + update=dict(default=False, type='bool'), + update_cache=dict(default=False, aliases=['update_codex'], type='bool'), + cache_valid_time=dict(default=0, type='int') ), - required_one_of = [['name', 'update', 'update_cache']], - supports_check_mode = True + required_one_of=[['name', 'update', 'update_cache']], + supports_check_mode=True ) if os.geteuid() != 0: diff --git a/lib/ansible/modules/packaging/os/svr4pkg.py b/lib/ansible/modules/packaging/os/svr4pkg.py index b8bcc68400..e4b7d31630 100644 --- a/lib/ansible/modules/packaging/os/svr4pkg.py +++ b/lib/ansible/modules/packaging/os/svr4pkg.py @@ -105,6 +105,7 @@ EXAMPLES = ''' import os import tempfile + def package_installed(module, name, category): cmd = [module.get_bin_path('pkginfo', True)] cmd.append('-q') @@ -117,6 +118,7 @@ def package_installed(module, name, category): else: return False + def create_admin_file(): (desc, filename) = tempfile.mkstemp(prefix='ansible_svr4pkg', text=True) fullauto = ''' @@ -141,48 +143,52 @@ basedir=default os.close(desc) return filename + def run_command(module, cmd): progname = cmd[0] cmd[0] = module.get_bin_path(progname, True) return module.run_command(cmd) + def package_install(module, name, src, proxy, response_file, zone, category): adminfile = create_admin_file() - cmd = [ 'pkgadd', '-n'] + cmd = ['pkgadd', '-n'] if zone == 'current': - cmd += [ '-G' ] - cmd += [ '-a', adminfile, '-d', src ] + cmd += ['-G'] + cmd += ['-a', adminfile, '-d', src] if proxy is not None: - cmd += [ '-x', proxy ] + cmd += ['-x', proxy] if response_file is not None: - cmd += [ '-r', response_file ] + cmd += ['-r', response_file] if category: - cmd += [ '-Y' ] + cmd += ['-Y'] cmd.append(name) (rc, out, err) = run_command(module, cmd) os.unlink(adminfile) return (rc, out, err) + def package_uninstall(module, name, src, category): adminfile = create_admin_file() if category: - cmd = [ 'pkgrm', '-na', adminfile, '-Y', name ] + cmd = ['pkgrm', '-na', adminfile, '-Y', name] else: - cmd = [ 'pkgrm', '-na', adminfile, name] + cmd = ['pkgrm', '-na', adminfile, name] (rc, out, err) = run_command(module, cmd) os.unlink(adminfile) return (rc, out, err) + def main(): module = AnsibleModule( - argument_spec = dict( - name = dict(required = True), - state = dict(required = True, choices=['present', 'absent']), - src = dict(default = None), - proxy = dict(default = None), - response_file = dict(default = None), - zone = dict(required=False, default = 'all', choices=['current','all']), - category = dict(default=False, type='bool') + argument_spec=dict( + name=dict(required=True), + state=dict(required=True, choices=['present', 'absent']), + src=dict(default=None), + proxy=dict(default=None), + response_file=dict(default=None), + zone=dict(required=False, default='all', choices=['current', 'all']), + category=dict(default=False, type='bool') ), supports_check_mode=True ) diff --git a/lib/ansible/modules/packaging/os/swdepot.py b/lib/ansible/modules/packaging/os/swdepot.py index 8b9785202a..f32e43c7e5 100644 --- a/lib/ansible/modules/packaging/os/swdepot.py +++ b/lib/ansible/modules/packaging/os/swdepot.py @@ -108,6 +108,7 @@ def query_package(module, name, depot=None): return rc, version + def remove_package(module, name): """ Uninstall package if installed. """ @@ -119,6 +120,7 @@ def remove_package(module, name): else: return rc, stderr + def install_package(module, depot, name): """ Install package if not already installed """ @@ -129,12 +131,13 @@ def install_package(module, depot, name): else: return rc, stderr + def main(): module = AnsibleModule( - argument_spec = dict( - name = dict(aliases=['pkg'], required=True), - state = dict(choices=['present', 'absent', 'latest'], required=True), - depot = dict(default=None, required=False) + argument_spec=dict( + name=dict(aliases=['pkg'], required=True), + state=dict(choices=['present', 'absent', 'latest'], required=True), + depot=dict(default=None, required=False) ), supports_check_mode=True ) @@ -145,12 +148,11 @@ def main(): changed = False msg = "No changed" rc = 0 - if ( state == 'present' or state == 'latest' ) and depot is None: + if (state == 'present' or state == 'latest') and depot is None: output = "depot parameter is mandatory in present or latest task" module.fail_json(name=name, msg=output, rc=rc) - - #Check local version + # Check local version rc, version_installed = query_package(module, name) if not rc: installed = True @@ -159,7 +161,7 @@ def main(): else: installed = False - if ( state == 'present' or state == 'latest' ) and installed is False: + if (state == 'present' or state == 'latest') and installed is False: if module.check_mode: module.exit_json(changed=True) rc, output = install_package(module, depot, name) @@ -172,14 +174,14 @@ def main(): module.fail_json(name=name, msg=output, rc=rc) elif state == 'latest' and installed is True: - #Check depot version + # Check depot version rc, version_depot = query_package(module, name, depot) if not rc: - if compare_package(version_installed,version_depot) == -1: + if compare_package(version_installed, version_depot) == -1: if module.check_mode: module.exit_json(changed=True) - #Install new version + # Install new version rc, output = install_package(module, depot, name) if not rc: diff --git a/lib/ansible/modules/packaging/os/zypper.py b/lib/ansible/modules/packaging/os/zypper.py index 99d00a0d77..15281e0bce 100644 --- a/lib/ansible/modules/packaging/os/zypper.py +++ b/lib/ansible/modules/packaging/os/zypper.py @@ -204,7 +204,6 @@ class Package: return self.prefix + self.name + self.version - def split_name_version(name): """splits of the package name and desired version @@ -301,7 +300,7 @@ def parse_zypper_xml(m, cmd, fail_not_found=True, packages=None): return parse_zypper_xml(m, cmd, fail_not_found=fail_not_found, packages=packages) return packages, rc, stdout, stderr - m.fail_json(msg='Zypper run command failed with return code %s.'%rc, rc=rc, stdout=stdout, stderr=stderr, cmd=cmd) + m.fail_json(msg='Zypper run command failed with return code %s.' % rc, rc=rc, stdout=stdout, stderr=stderr, cmd=cmd) def get_cmd(m, subcommand): @@ -455,20 +454,21 @@ def repo_refresh(m): # =========================================== # Main control flow + def main(): module = AnsibleModule( - argument_spec = dict( - name = dict(required=True, aliases=['pkg'], type='list'), - state = dict(required=False, default='present', choices=['absent', 'installed', 'latest', 'present', 'removed', 'dist-upgrade']), - type = dict(required=False, default='package', choices=['package', 'patch', 'pattern', 'product', 'srcpackage', 'application']), - disable_gpg_check = dict(required=False, default='no', type='bool'), - disable_recommends = dict(required=False, default='yes', type='bool'), - force = dict(required=False, default='no', type='bool'), - update_cache = dict(required=False, aliases=['refresh'], default='no', type='bool'), - oldpackage = dict(required=False, default='no', type='bool'), - extra_args = dict(required=False, default=None), + argument_spec=dict( + name=dict(required=True, aliases=['pkg'], type='list'), + state=dict(required=False, default='present', choices=['absent', 'installed', 'latest', 'present', 'removed', 'dist-upgrade']), + type=dict(required=False, default='package', choices=['package', 'patch', 'pattern', 'product', 'srcpackage', 'application']), + disable_gpg_check=dict(required=False, default='no', type='bool'), + disable_recommends=dict(required=False, default='yes', type='bool'), + force=dict(required=False, default='no', type='bool'), + update_cache=dict(required=False, aliases=['refresh'], default='no', type='bool'), + oldpackage=dict(required=False, default='no', type='bool'), + extra_args=dict(required=False, default=None), ), - supports_check_mode = True + supports_check_mode=True ) name = module.params['name'] diff --git a/lib/ansible/modules/packaging/os/zypper_repository.py b/lib/ansible/modules/packaging/os/zypper_repository.py index 7ab1afdc61..ca1b04bc10 100644 --- a/lib/ansible/modules/packaging/os/zypper_repository.py +++ b/lib/ansible/modules/packaging/os/zypper_repository.py @@ -146,6 +146,7 @@ REPO_OPTS = ['alias', 'name', 'priority', 'enabled', 'autorefresh', 'gpgcheck'] from distutils.version import LooseVersion + def _get_cmd(*args): """Combines the non-interactive zypper command with arguments/subcommands""" cmd = ['/usr/bin/zypper', '--quiet', '--non-interactive'] @@ -178,6 +179,7 @@ def _parse_repos(module): else: module.fail_json(msg='Failed to execute "%s"' % " ".join(cmd), rc=rc, stdout=stdout, stderr=stderr) + def _repo_changes(realrepo, repocmp): "Check whether the 2 given repos have different settings." for k in repocmp: @@ -194,6 +196,7 @@ def _repo_changes(realrepo, repocmp): return True return False + def repo_exists(module, repodata, overwrite_multiple): """Check whether the repository already exists. @@ -288,6 +291,7 @@ def get_zypper_version(module): return LooseVersion('1.0') return LooseVersion(stdout.split()[1]) + def runrefreshrepo(module, auto_import_keys=False, shortname=None): "Forces zypper to refresh repo metadata." if auto_import_keys: @@ -309,15 +313,15 @@ def main(): state=dict(choices=['present', 'absent'], default='present'), runrefresh=dict(required=False, default='no', type='bool'), description=dict(required=False), - disable_gpg_check = dict(required=False, default=False, type='bool'), - autorefresh = dict(required=False, default=True, type='bool', aliases=['refresh']), - priority = dict(required=False, type='int'), - enabled = dict(required=False, default=True, type='bool'), - overwrite_multiple = dict(required=False, default=False, type='bool'), - auto_import_keys = dict(required=False, default=False, type='bool'), + disable_gpg_check=dict(required=False, default=False, type='bool'), + autorefresh=dict(required=False, default=True, type='bool', aliases=['refresh']), + priority=dict(required=False, type='int'), + enabled=dict(required=False, default=True, type='bool'), + overwrite_multiple=dict(required=False, default=False, type='bool'), + auto_import_keys=dict(required=False, default=False, type='bool'), ), supports_check_mode=False, - required_one_of = [['state','runrefresh']], + required_one_of=[['state', 'runrefresh']], ) repo = module.params['repo'] diff --git a/lib/ansible/modules/storage/infinidat/infini_export.py b/lib/ansible/modules/storage/infinidat/infini_export.py index 827c353163..7edb89a865 100644 --- a/lib/ansible/modules/storage/infinidat/infini_export.py +++ b/lib/ansible/modules/storage/infinidat/infini_export.py @@ -158,10 +158,10 @@ def main(): argument_spec = infinibox_argument_spec() argument_spec.update( dict( - name = dict(required=True), - state = dict(default='present', choices=['present', 'absent']), - filesystem = dict(required=True), - client_list = dict(type='list') + name=dict(required=True), + state=dict(default='present', choices=['present', 'absent']), + filesystem=dict(required=True), + client_list=dict(type='list') ) ) @@ -172,10 +172,10 @@ def main(): if not HAS_MUNCH: module.fail_json(msg='the python munch library is required for this module') - state = module.params['state'] - system = get_system(module) + state = module.params['state'] + system = get_system(module) filesystem = get_filesystem(module, system) - export = get_export(module, filesystem, system) + export = get_export(module, filesystem, system) if filesystem is None: module.fail_json(msg='Filesystem {} not found'.format(module.params['filesystem'])) diff --git a/lib/ansible/modules/storage/infinidat/infini_export_client.py b/lib/ansible/modules/storage/infinidat/infini_export_client.py index dba9b2de0c..80f0a0f479 100644 --- a/lib/ansible/modules/storage/infinidat/infini_export_client.py +++ b/lib/ansible/modules/storage/infinidat/infini_export_client.py @@ -116,11 +116,11 @@ def update_client(module, export): changed = False - client = module.params['client'] - access_mode = module.params['access_mode'] + client = module.params['client'] + access_mode = module.params['access_mode'] no_root_squash = module.params['no_root_squash'] - client_list = export.get_permissions() + client_list = export.get_permissions() client_not_in_list = True for index, item in enumerate(client_list): @@ -154,7 +154,7 @@ def delete_client(module, export): changed = False - client = module.params['client'] + client = module.params['client'] client_list = export.get_permissions() for index, item in enumerate(client_list): @@ -175,11 +175,11 @@ def main(): argument_spec = infinibox_argument_spec() argument_spec.update( dict( - client = dict(required=True), - access_mode = dict(choices=['RO', 'RW'], default='RW'), - no_root_squash = dict(type='bool', default='no'), - state = dict(default='present', choices=['present', 'absent']), - export = dict(required=True) + client=dict(required=True), + access_mode=dict(choices=['RO', 'RW'], default='RW'), + no_root_squash=dict(type='bool', default='no'), + state=dict(default='present', choices=['present', 'absent']), + export=dict(required=True) ) ) @@ -190,8 +190,8 @@ def main(): if not HAS_MUNCH: module.fail_json(msg='the python munch library is required for this module') - system = get_system(module) - export = get_export(module, system) + system = get_system(module) + export = get_export(module, system) if module.params['state'] == 'present': update_client(module, export) diff --git a/lib/ansible/modules/storage/infinidat/infini_fs.py b/lib/ansible/modules/storage/infinidat/infini_fs.py index 68a1749c97..aeddc2dca0 100644 --- a/lib/ansible/modules/storage/infinidat/infini_fs.py +++ b/lib/ansible/modules/storage/infinidat/infini_fs.py @@ -126,10 +126,10 @@ def main(): argument_spec = infinibox_argument_spec() argument_spec.update( dict( - name = dict(required=True), - state = dict(default='present', choices=['present', 'absent']), - pool = dict(required=True), - size = dict() + name=dict(required=True), + state=dict(default='present', choices=['present', 'absent']), + pool=dict(required=True), + size=dict() ) ) @@ -146,9 +146,9 @@ def main(): except: module.fail_json(msg='size (Physical Capacity) should be defined in MB, GB, TB or PB units') - state = module.params['state'] - system = get_system(module) - pool = get_pool(module, system) + state = module.params['state'] + system = get_system(module) + pool = get_pool(module, system) filesystem = get_filesystem(module, system) if pool is None: diff --git a/lib/ansible/modules/storage/infinidat/infini_host.py b/lib/ansible/modules/storage/infinidat/infini_host.py index 6389d22992..516f2303f9 100644 --- a/lib/ansible/modules/storage/infinidat/infini_host.py +++ b/lib/ansible/modules/storage/infinidat/infini_host.py @@ -81,7 +81,7 @@ from ansible.module_utils.infinibox import HAS_INFINISDK, api_wrapper, get_syste @api_wrapper def get_host(module, system): - host = None + host = None for h in system.hosts.to_list(): if h.get_name() == module.params['name']: @@ -124,10 +124,10 @@ def main(): argument_spec = infinibox_argument_spec() argument_spec.update( dict( - name = dict(required=True), - state = dict(default='present', choices=['present', 'absent']), - wwns = dict(type='list'), - volume = dict() + name=dict(required=True), + state=dict(default='present', choices=['present', 'absent']), + wwns=dict(type='list'), + volume=dict() ) ) @@ -136,9 +136,9 @@ def main(): if not HAS_INFINISDK: module.fail_json(msg='infinisdk is required for this module') - state = module.params['state'] + state = module.params['state'] system = get_system(module) - host = get_host(module, system) + host = get_host(module, system) if module.params['volume']: try: diff --git a/lib/ansible/modules/storage/infinidat/infini_pool.py b/lib/ansible/modules/storage/infinidat/infini_pool.py index cf90be4d86..da0068a73b 100644 --- a/lib/ansible/modules/storage/infinidat/infini_pool.py +++ b/lib/ansible/modules/storage/infinidat/infini_pool.py @@ -102,9 +102,9 @@ def get_pool(module, system): @api_wrapper def create_pool(module, system): """Create Pool""" - name = module.params['name'] - size = module.params['size'] - vsize = module.params['vsize'] + name = module.params['name'] + size = module.params['size'] + vsize = module.params['vsize'] ssd_cache = module.params['ssd_cache'] if not module.check_mode: @@ -126,10 +126,10 @@ def create_pool(module, system): @api_wrapper def update_pool(module, system, pool): """Update Pool""" - changed = False + changed = False - size = module.params['size'] - vsize = module.params['vsize'] + size = module.params['size'] + vsize = module.params['vsize'] ssd_cache = module.params['ssd_cache'] # Roundup the capacity to mimic Infinibox behaviour @@ -167,11 +167,11 @@ def main(): argument_spec = infinibox_argument_spec() argument_spec.update( dict( - name = dict(required=True), - state = dict(default='present', choices=['present', 'absent']), - size = dict(), - vsize = dict(), - ssd_cache = dict(type='bool', default=True) + name=dict(required=True), + state=dict(default='present', choices=['present', 'absent']), + size=dict(), + vsize=dict(), + ssd_cache=dict(type='bool', default=True) ) ) @@ -194,9 +194,9 @@ def main(): except: module.fail_json(msg='vsize (Virtual Capacity) should be defined in MB, GB, TB or PB units') - state = module.params['state'] + state = module.params['state'] system = get_system(module) - pool = get_pool(module, system) + pool = get_pool(module, system) if state == 'present' and not pool: create_pool(module, system) diff --git a/lib/ansible/modules/storage/infinidat/infini_vol.py b/lib/ansible/modules/storage/infinidat/infini_vol.py index 729ade6f28..c83a7a6767 100644 --- a/lib/ansible/modules/storage/infinidat/infini_vol.py +++ b/lib/ansible/modules/storage/infinidat/infini_vol.py @@ -126,10 +126,10 @@ def main(): argument_spec = infinibox_argument_spec() argument_spec.update( dict( - name = dict(required=True), - state = dict(default='present', choices=['present', 'absent']), - pool = dict(required=True), - size = dict() + name=dict(required=True), + state=dict(default='present', choices=['present', 'absent']), + pool=dict(required=True), + size=dict() ) ) @@ -144,9 +144,9 @@ def main(): except: module.fail_json(msg='size (Physical Capacity) should be defined in MB, GB, TB or PB units') - state = module.params['state'] + state = module.params['state'] system = get_system(module) - pool = get_pool(module, system) + pool = get_pool(module, system) volume = get_volume(module, system) if pool is None: diff --git a/lib/ansible/modules/storage/netapp/na_cdot_qtree.py b/lib/ansible/modules/storage/netapp/na_cdot_qtree.py index aa9cb7b381..0119327d8c 100644 --- a/lib/ansible/modules/storage/netapp/na_cdot_qtree.py +++ b/lib/ansible/modules/storage/netapp/na_cdot_qtree.py @@ -129,7 +129,7 @@ class NetAppCDOTQTree(object): qtree_list_iter = netapp_utils.zapi.NaElement('qtree-list-iter') query_details = netapp_utils.zapi.NaElement.create_node_with_children( 'qtree-info', **{'vserver': self.vserver, - 'volume':self.flexvol_name, + 'volume': self.flexvol_name, 'qtree': self.name}) query = netapp_utils.zapi.NaElement('query') diff --git a/lib/ansible/modules/storage/netapp/na_cdot_volume.py b/lib/ansible/modules/storage/netapp/na_cdot_volume.py index d62159cfa1..a58c024efb 100644 --- a/lib/ansible/modules/storage/netapp/na_cdot_volume.py +++ b/lib/ansible/modules/storage/netapp/na_cdot_volume.py @@ -246,7 +246,7 @@ class NetAppCDOTVolume(object): else: volume_delete = netapp_utils.zapi.NaElement.create_node_with_children( 'volume-destroy', **{'name': self.name, 'unmount-and-offline': - 'true'}) + 'true'}) try: self.server.invoke_successfully(volume_delete, diff --git a/lib/ansible/modules/storage/netapp/netapp_e_amg.py b/lib/ansible/modules/storage/netapp/netapp_e_amg.py index b5cf67d8cd..476a2b83f9 100644 --- a/lib/ansible/modules/storage/netapp/netapp_e_amg.py +++ b/lib/ansible/modules/storage/netapp/netapp_e_amg.py @@ -168,7 +168,7 @@ def create_async(module, ssid, api_url, api_pwd, api_usr, body): rc, data = request(url, data=post_data, method='POST', url_username=api_usr, url_password=api_pwd, headers=HEADERS) except Exception as e: - module.exit_json(msg="Exception while creating aysnc mirror group. Message: %s" % to_native(e), + module.exit_json(msg="Exception while creating aysnc mirror group. Message: %s" % to_native(e), exception=traceback.format_exc()) return data diff --git a/lib/ansible/modules/storage/netapp/netapp_e_storagepool.py b/lib/ansible/modules/storage/netapp/netapp_e_storagepool.py index 58d3a4723b..994376ecbe 100644 --- a/lib/ansible/modules/storage/netapp/netapp_e_storagepool.py +++ b/lib/ansible/modules/storage/netapp/netapp_e_storagepool.py @@ -125,7 +125,7 @@ class GroupBy(object): # python 2, 3 generic grouping. def __init__(self, iterable, key=None): if key is None: - key = lambda x: x + def key(x): return x self.keyfunc = key self.it = iter(iterable) self.tgtkey = self.currkey = self.currvalue = object() diff --git a/lib/ansible/modules/storage/netapp/sf_snapshot_schedule_manager.py b/lib/ansible/modules/storage/netapp/sf_snapshot_schedule_manager.py index 02298fc895..e609188827 100644 --- a/lib/ansible/modules/storage/netapp/sf_snapshot_schedule_manager.py +++ b/lib/ansible/modules/storage/netapp/sf_snapshot_schedule_manager.py @@ -312,11 +312,11 @@ class SolidFireSnapShotSchedule(object): elif self.state == 'present': # Check if we need to update the account - if self.retention is not None and schedule_detail.schedule_info.retention !=self.retention: + if self.retention is not None and schedule_detail.schedule_info.retention != self.retention: update_schedule = True changed = True - elif schedule_detail.name !=self.name: + elif schedule_detail.name != self.name: update_schedule = True changed = True diff --git a/lib/ansible/modules/storage/netapp/sf_volume_access_group_manager.py b/lib/ansible/modules/storage/netapp/sf_volume_access_group_manager.py index bd6ba5d01d..ef62ebc6dd 100644 --- a/lib/ansible/modules/storage/netapp/sf_volume_access_group_manager.py +++ b/lib/ansible/modules/storage/netapp/sf_volume_access_group_manager.py @@ -225,7 +225,7 @@ class SolidFireVolumeAccessGroup(object): update_group = True changed = True elif self.virtual_network_id is not None or self.virtual_network_tags is not None or \ - self.attributes is not None: + self.attributes is not None: update_group = True changed = True diff --git a/lib/ansible/modules/storage/netapp/sf_volume_manager.py b/lib/ansible/modules/storage/netapp/sf_volume_manager.py index d95c9a1dc9..15455a55fb 100644 --- a/lib/ansible/modules/storage/netapp/sf_volume_manager.py +++ b/lib/ansible/modules/storage/netapp/sf_volume_manager.py @@ -289,7 +289,7 @@ class SolidFireVolume(object): elif volume_detail.total_size is not None and volume_detail.total_size != self.size: size_difference = abs(float(volume_detail.total_size - self.size)) # Change size only if difference is bigger than 0.001 - if size_difference/self.size > 0.001: + if size_difference / self.size > 0.001: update_volume = True changed = True @@ -312,8 +312,8 @@ class SolidFireVolume(object): self.create_volume() result_message = "Volume created" elif update_volume: - self.update_volume() - result_message = "Volume updated" + self.update_volume() + result_message = "Volume updated" elif self.state == 'absent': self.delete_volume() diff --git a/lib/ansible/modules/system/cron.py b/lib/ansible/modules/system/cron.py index 39e8fd9681..2e72720dfb 100644 --- a/lib/ansible/modules/system/cron.py +++ b/lib/ansible/modules/system/cron.py @@ -217,6 +217,7 @@ class CronTab(object): user - the user of the crontab (defaults to root) cron_file - a cron file under /etc/cron.d, or an absolute path """ + def __init__(self, module, user=None, cron_file=None): self.module = module self.user = user diff --git a/lib/ansible/modules/web_infrastructure/letsencrypt.py b/lib/ansible/modules/web_infrastructure/letsencrypt.py index c60832c3a5..de74398abb 100644 --- a/lib/ansible/modules/web_infrastructure/letsencrypt.py +++ b/lib/ansible/modules/web_infrastructure/letsencrypt.py @@ -279,6 +279,7 @@ class ACMEDirectory(object): require authentication). https://tools.ietf.org/html/draft-ietf-acme-acme-02#section-6.2 ''' + def __init__(self, module): self.module = module self.directory_root = module.params['acme_directory'] @@ -304,6 +305,7 @@ class ACMEAccount(object): ACME server. Provides access to account bound information like the currently active authorizations and valid certificates ''' + def __init__(self, module): self.module = module self.key = module.params['account_key'] @@ -513,6 +515,7 @@ class ACMEClient(object): start and validate ACME challenges and download the respective certificates. ''' + def __init__(self, module): self.module = module self.challenge = module.params['challenge'] -- cgit v1.2.1