mirror of
https://github.com/ansible-collections/community.general.git
synced 2026-05-08 14:22:46 +00:00
E501 fixes (#22879)
This commit is contained in:
@@ -119,7 +119,7 @@ msg:
|
||||
returned: success
|
||||
type: string
|
||||
sample: '{"changed": true, "connectionType": "fc", "groupRef": "3700000060080E5000299C24000006E857AC7EEC", "groupState": "optimal", "id": "3700000060080E5000299C24000006E857AC7EEC", "label": "amg_made_by_ansible", "localRole": "primary", "mirrorChannelRemoteTarget": "9000000060080E5000299C24005B06E557AC7EEC", "orphanGroup": false, "recoveryPointAgeAlertThresholdMinutes": 20, "remoteRole": "secondary", "remoteTarget": {"nodeName": {"ioInterfaceType": "fc", "iscsiNodeName": null, "remoteNodeWWN": "20040080E5299F1C"}, "remoteRef": "9000000060080E5000299C24005B06E557AC7EEC", "scsiinitiatorTargetBaseProperties": {"ioInterfaceType": "fc", "iscsiinitiatorTargetBaseParameters": null}}, "remoteTargetId": "ansible2", "remoteTargetName": "Ansible2", "remoteTargetWwn": "60080E5000299F880000000056A25D56", "repositoryUtilizationWarnThreshold": 80, "roleChangeProgress": "none", "syncActivity": "idle", "syncCompletionTimeAlertThresholdMinutes": 10, "syncIntervalMinutes": 10, "worldWideName": "60080E5000299C24000006E857AC7EEC"}'
|
||||
"""
|
||||
""" # NOQA
|
||||
|
||||
import json
|
||||
|
||||
|
||||
@@ -58,7 +58,8 @@ options:
|
||||
state:
|
||||
description:
|
||||
- The synchronization action you'd like to take.
|
||||
- If C(running) then it will begin syncing if there is no active sync or will resume a suspended sync. If there is already a sync in progress, it will return with an OK status.
|
||||
- If C(running) then it will begin syncing if there is no active sync or will resume a suspended sync. If there is already a sync in
|
||||
progress, it will return with an OK status.
|
||||
- If C(suspended) it will suspend any ongoing sync action, but return OK if there is no active sync or if the sync is already suspended
|
||||
choices:
|
||||
- running
|
||||
@@ -67,8 +68,10 @@ options:
|
||||
delete_recovery_point:
|
||||
description:
|
||||
- Indicates whether the failures point can be deleted on the secondary if necessary to achieve the synchronization.
|
||||
- If true, and if the amount of unsynchronized data exceeds the CoW repository capacity on the secondary for any member volume, the last failures point will be deleted and synchronization will continue.
|
||||
- If false, the synchronization will be suspended if the amount of unsynchronized data exceeds the CoW Repository capacity on the secondary and the failures point will be preserved.
|
||||
- If true, and if the amount of unsynchronized data exceeds the CoW repository capacity on the secondary for any member volume, the last
|
||||
failures point will be deleted and synchronization will continue.
|
||||
- If false, the synchronization will be suspended if the amount of unsynchronized data exceeds the CoW Repository capacity on the secondary
|
||||
and the failures point will be preserved.
|
||||
- "NOTE: This only has impact for newly launched syncs."
|
||||
choices:
|
||||
- yes
|
||||
|
||||
@@ -27,7 +27,8 @@ DOCUMENTATION = '''
|
||||
module: netapp_e_auth
|
||||
short_description: Sets or updates the password for a storage array.
|
||||
description:
|
||||
- Sets or updates the password for a storage array. When the password is updated on the storage array, it must be updated on the SANtricity Web Services proxy. Note, all storage arrays do not have a Monitor or RO role.
|
||||
- Sets or updates the password for a storage array. When the password is updated on the storage array, it must be updated on the SANtricity Web
|
||||
Services proxy. Note, all storage arrays do not have a Monitor or RO role.
|
||||
version_added: "2.2"
|
||||
author: Kevin Hulquest (@hulquest)
|
||||
options:
|
||||
@@ -38,7 +39,8 @@ options:
|
||||
- Should https certificates be validated?
|
||||
name:
|
||||
description:
|
||||
- The name of the storage array. Note that if more than one storage array with this name is detected, the task will fail and you'll have to use the ID instead.
|
||||
- The name of the storage array. Note that if more than one storage array with this name is detected, the task will fail and you'll have to use
|
||||
the ID instead.
|
||||
required: False
|
||||
ssid:
|
||||
description:
|
||||
|
||||
@@ -59,7 +59,8 @@ options:
|
||||
required: True
|
||||
host_type_index:
|
||||
description:
|
||||
- The index that maps to host type you wish to create. It is recommended to use the M(netapp_e_facts) module to gather this information. Alternatively you can use the WSP portal to retrieve the information.
|
||||
- The index that maps to host type you wish to create. It is recommended to use the M(netapp_e_facts) module to gather this information.
|
||||
Alternatively you can use the WSP portal to retrieve the information.
|
||||
required: True
|
||||
ports:
|
||||
description:
|
||||
@@ -288,7 +289,9 @@ class Host(object):
|
||||
needs_update = True
|
||||
else:
|
||||
self.module.fail_json(
|
||||
msg="The port you specified:\n%s\n is associated with a different host. Specify force_port as True or try a different port spec" % arg_port)
|
||||
msg="The port you specified:\n%s\n is associated with a different host. Specify force_port as True or try a different "
|
||||
"port spec" % arg_port
|
||||
)
|
||||
|
||||
return needs_update
|
||||
|
||||
|
||||
@@ -93,7 +93,8 @@ clusterRef:
|
||||
type: string
|
||||
sample: "3233343536373839303132333100000000000000"
|
||||
confirmLUNMappingCreation:
|
||||
description: If true, indicates that creation of LUN-to-volume mappings should require careful confirmation from the end-user, since such a mapping will alter the volume access rights of other clusters, in addition to this one.
|
||||
description: If true, indicates that creation of LUN-to-volume mappings should require careful confirmation from the end-user, since such a mapping
|
||||
will alter the volume access rights of other clusters, in addition to this one.
|
||||
returned: always
|
||||
type: boolean
|
||||
sample: false
|
||||
@@ -108,7 +109,8 @@ id:
|
||||
type: string
|
||||
sample: "3233343536373839303132333100000000000000"
|
||||
isSAControlled:
|
||||
description: If true, indicates that I/O accesses from this cluster are subject to the storage array's default LUN-to-volume mappings. If false, indicates that I/O accesses from the cluster are subject to cluster-specific LUN-to-volume mappings.
|
||||
description: If true, indicates that I/O accesses from this cluster are subject to the storage array's default LUN-to-volume mappings. If false,
|
||||
indicates that I/O accesses from the cluster are subject to cluster-specific LUN-to-volume mappings.
|
||||
returned: always except when state is absent
|
||||
type: boolean
|
||||
sample: false
|
||||
|
||||
@@ -78,7 +78,8 @@ options:
|
||||
delete_limit:
|
||||
description:
|
||||
- The automatic deletion indicator.
|
||||
- If non-zero, the oldest snapshot image will be automatically deleted when creating a new snapshot image to keep the total number of snapshot images limited to the number specified.
|
||||
- If non-zero, the oldest snapshot image will be automatically deleted when creating a new snapshot image to keep the total number of
|
||||
snapshot images limited to the number specified.
|
||||
- This value is overridden by the consistency group setting if this snapshot group is associated with a consistency group.
|
||||
required: False
|
||||
default: 30
|
||||
|
||||
@@ -31,7 +31,8 @@ description:
|
||||
version_added: '2.2'
|
||||
author: Kevin Hulquest (@hulquest)
|
||||
notes:
|
||||
- Only I(full_threshold) is supported for update operations. If the snapshot volume already exists and the threshold matches, then an C(ok) status will be returned, no other changes can be made to a pre-existing snapshot volume.
|
||||
- Only I(full_threshold) is supported for update operations. If the snapshot volume already exists and the threshold matches, then an C(ok) status
|
||||
will be returned, no other changes can be made to a pre-existing snapshot volume.
|
||||
options:
|
||||
api_username:
|
||||
required: true
|
||||
|
||||
@@ -62,7 +62,8 @@ options:
|
||||
array_wwn:
|
||||
required: false
|
||||
description:
|
||||
- The WWN of the array to manage. Only necessary if in-band managing multiple arrays on the same agent host. Mutually exclusive of controller_addresses parameter.
|
||||
- The WWN of the array to manage. Only necessary if in-band managing multiple arrays on the same agent host. Mutually exclusive of
|
||||
controller_addresses parameter.
|
||||
array_password:
|
||||
required: false
|
||||
description:
|
||||
@@ -231,7 +232,9 @@ def main():
|
||||
changed = True
|
||||
if array_detail['wwn'] != array_wwn and array_wwn is not None:
|
||||
module.fail_json(
|
||||
msg='It seems you may have specified a bad WWN. The storage system ID you specified, %s, currently has the WWN of %s' % (ssid, array_detail['wwn']))
|
||||
msg='It seems you may have specified a bad WWN. The storage system ID you specified, %s, currently has the WWN of %s' %
|
||||
(ssid, array_detail['wwn'])
|
||||
)
|
||||
elif rc == 404:
|
||||
if state == 'present':
|
||||
changed = True
|
||||
|
||||
@@ -487,7 +487,8 @@ class NetAppESeriesStoragePool(object):
|
||||
# # TODO: this arg appears to be ignored, uncomment if it isn't
|
||||
# #if self.criteria_disk_min_size_gb:
|
||||
# # drives_req['driveCapacityMin'] = self.criteria_disk_min_size_gb * 1024
|
||||
# (rc,drives_resp) = request(self.api_url + "/storage-systems/%s/drives" % (self.ssid), data=json.dumps(drives_req), headers=self.post_headers, method='POST', url_username=self.api_usr, url_password=self.api_pwd, validate_certs=self.validate_certs)
|
||||
# (rc,drives_resp) = request(self.api_url + "/storage-systems/%s/drives" % (self.ssid), data=json.dumps(drives_req), headers=self.post_headers,
|
||||
# method='POST', url_username=self.api_usr, url_password=self.api_pwd, validate_certs=self.validate_certs)
|
||||
#
|
||||
# if rc == 204:
|
||||
# self.module.fail_json(msg='Cannot find disks to match requested criteria for storage pool')
|
||||
|
||||
Reference in New Issue
Block a user