|
|
|
#!/usr/bin/python
|
|
|
|
#
|
|
|
|
# This is a free software: you can redistribute it and/or modify
|
|
|
|
# it under the terms of the GNU General Public License as published by
|
|
|
|
# the Free Software Foundation, either version 3 of the License, or
|
|
|
|
# (at your option) any later version.
|
|
|
|
#
|
|
|
|
# This Ansible library is distributed in the hope that it will be useful,
|
|
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
# GNU General Public License for more details.
|
|
|
|
#
|
|
|
|
# You should have received a copy of the GNU General Public License
|
|
|
|
# along with this library. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
|
|
|
|
DOCUMENTATION = '''
|
|
|
|
---
|
|
|
|
module: s3_bucket
|
|
|
|
short_description: Manage S3 buckets in AWS, Ceph, Walrus and FakeS3
|
|
|
|
description:
|
|
|
|
- Manage S3 buckets in AWS, Ceph, Walrus and FakeS3
|
|
|
|
version_added: "2.0"
|
|
|
|
author: "Rob White (@wimnat)"
|
|
|
|
options:
|
|
|
|
force:
|
|
|
|
description:
|
|
|
|
- When trying to delete a bucket, delete all keys in the bucket first (an s3 bucket must be empty for a successful deletion)
|
|
|
|
required: false
|
|
|
|
default: no
|
|
|
|
choices: [ 'yes', 'no' ]
|
|
|
|
name:
|
|
|
|
description:
|
|
|
|
- Name of the s3 bucket
|
|
|
|
required: true
|
|
|
|
default: null
|
|
|
|
policy:
|
|
|
|
description:
|
|
|
|
- The JSON policy as a string.
|
|
|
|
required: false
|
|
|
|
default: null
|
|
|
|
s3_url:
|
|
|
|
description:
|
|
|
|
- S3 URL endpoint for usage with Ceph, Eucalypus, fakes3, etc. Otherwise assumes AWS
|
|
|
|
default: null
|
|
|
|
aliases: [ S3_URL ]
|
|
|
|
ceph:
|
|
|
|
description:
|
|
|
|
- Enable API compatibility with Ceph. It takes into account the S3 API subset working with Ceph in order to provide the same module behaviour where possible.
|
|
|
|
version_added: "2.2"
|
|
|
|
requester_pays:
|
|
|
|
description:
|
|
|
|
- With Requester Pays buckets, the requester instead of the bucket owner pays the cost of the request and the data download from the bucket.
|
|
|
|
required: false
|
|
|
|
default: no
|
|
|
|
choices: [ 'yes', 'no' ]
|
|
|
|
state:
|
|
|
|
description:
|
|
|
|
- Create or remove the s3 bucket
|
|
|
|
required: false
|
|
|
|
default: present
|
|
|
|
choices: [ 'present', 'absent' ]
|
|
|
|
tags:
|
|
|
|
description:
|
|
|
|
- tags dict to apply to bucket
|
|
|
|
required: false
|
|
|
|
default: null
|
|
|
|
versioning:
|
|
|
|
description:
|
|
|
|
- Whether versioning is enabled or disabled (note that once versioning is enabled, it can only be suspended)
|
|
|
|
required: false
|
|
|
|
default: no
|
|
|
|
choices: [ 'yes', 'no' ]
|
|
|
|
extends_documentation_fragment:
|
|
|
|
- aws
|
|
|
|
- ec2
|
|
|
|
'''
|
|
|
|
|
|
|
|
EXAMPLES = '''
|
|
|
|
# Note: These examples do not set authentication details, see the AWS Guide for details.
|
|
|
|
|
|
|
|
# Create a simple s3 bucket
|
|
|
|
- s3_bucket:
|
|
|
|
name: mys3bucket
|
|
|
|
|
|
|
|
# Create a simple s3 bucket on Ceph Rados Gateway
|
|
|
|
- s3_bucket:
|
|
|
|
name: mys3bucket
|
|
|
|
s3_url: http://your-ceph-rados-gateway-server.xxx
|
|
|
|
ceph: true
|
|
|
|
|
|
|
|
# Remove an s3 bucket and any keys it contains
|
|
|
|
- s3_bucket:
|
|
|
|
name: mys3bucket
|
|
|
|
state: absent
|
|
|
|
force: yes
|
|
|
|
|
|
|
|
# Create a bucket, add a policy from a file, enable requester pays, enable versioning and tag
|
|
|
|
- s3_bucket:
|
|
|
|
name: mys3bucket
|
|
|
|
policy: "{{ lookup('file','policy.json') }}"
|
|
|
|
requester_pays: yes
|
|
|
|
versioning: yes
|
|
|
|
tags:
|
|
|
|
example: tag1
|
|
|
|
another: tag2
|
|
|
|
|
|
|
|
'''
|
|
|
|
|
|
|
|
import xml.etree.ElementTree as ET
|
|
|
|
import urlparse
|
|
|
|
|
|
|
|
try:
|
|
|
|
import boto.ec2
|
|
|
|
from boto.s3.connection import OrdinaryCallingFormat, Location
|
|
|
|
from boto.s3.tagging import Tags, TagSet
|
|
|
|
from boto.exception import BotoServerError, S3CreateError, S3ResponseError
|
|
|
|
HAS_BOTO = True
|
|
|
|
except ImportError:
|
|
|
|
HAS_BOTO = False
|
|
|
|
|
|
|
|
def get_request_payment_status(bucket):
|
|
|
|
|
|
|
|
response = bucket.get_request_payment()
|
|
|
|
root = ET.fromstring(response)
|
|
|
|
for message in root.findall('.//{http://s3.amazonaws.com/doc/2006-03-01/}Payer'):
|
|
|
|
payer = message.text
|
|
|
|
|
|
|
|
if payer == "BucketOwner":
|
|
|
|
return False
|
|
|
|
else:
|
|
|
|
return True
|
|
|
|
|
|
|
|
def create_tags_container(tags):
|
|
|
|
|
|
|
|
tag_set = TagSet()
|
|
|
|
tags_obj = Tags()
|
|
|
|
for key, val in tags.iteritems():
|
|
|
|
tag_set.add_tag(key, val)
|
|
|
|
|
|
|
|
tags_obj.add_tag_set(tag_set)
|
|
|
|
return tags_obj
|
|
|
|
|
|
|
|
def _create_bucket(connection, module, location):
|
|
|
|
|
|
|
|
policy = module.params.get("policy")
|
|
|
|
name = module.params.get("name")
|
|
|
|
requester_pays = module.params.get("requester_pays")
|
|
|
|
tags = module.params.get("tags")
|
|
|
|
versioning = module.params.get("versioning")
|
|
|
|
changed = False
|
|
|
|
|
|
|
|
try:
|
|
|
|
bucket = connection.get_bucket(name)
|
|
|
|
except S3ResponseError, e:
|
|
|
|
try:
|
|
|
|
bucket = connection.create_bucket(name, location=location)
|
|
|
|
changed = True
|
|
|
|
except S3CreateError, e:
|
|
|
|
module.fail_json(msg=e.message)
|
|
|
|
|
|
|
|
# Versioning
|
|
|
|
versioning_status = bucket.get_versioning_status()
|
|
|
|
if not versioning_status and versioning:
|
|
|
|
try:
|
|
|
|
bucket.configure_versioning(versioning)
|
|
|
|
changed = True
|
|
|
|
versioning_status = bucket.get_versioning_status()
|
|
|
|
except S3ResponseError, e:
|
|
|
|
module.fail_json(msg=e.message)
|
|
|
|
elif not versioning_status and not versioning:
|
|
|
|
# do nothing
|
|
|
|
pass
|
|
|
|
else:
|
|
|
|
if versioning_status['Versioning'] == "Enabled" and not versioning:
|
|
|
|
bucket.configure_versioning(versioning)
|
|
|
|
changed = True
|
|
|
|
versioning_status = bucket.get_versioning_status()
|
|
|
|
elif ( (versioning_status['Versioning'] == "Disabled" and versioning) or (versioning_status['Versioning'] == "Suspended" and versioning) ):
|
|
|
|
bucket.configure_versioning(versioning)
|
|
|
|
changed = True
|
|
|
|
versioning_status = bucket.get_versioning_status()
|
|
|
|
|
|
|
|
# Requester pays
|
|
|
|
requester_pays_status = get_request_payment_status(bucket)
|
|
|
|
if requester_pays_status != requester_pays:
|
|
|
|
if requester_pays:
|
|
|
|
bucket.set_request_payment(payer='Requester')
|
|
|
|
changed = True
|
|
|
|
requester_pays_status = get_request_payment_status(bucket)
|
|
|
|
else:
|
|
|
|
bucket.set_request_payment(payer='BucketOwner')
|
|
|
|
changed = True
|
|
|
|
requester_pays_status = get_request_payment_status(bucket)
|
|
|
|
|
|
|
|
# Policy
|
|
|
|
try:
|
|
|
|
current_policy = bucket.get_policy()
|
|
|
|
except S3ResponseError, e:
|
|
|
|
if e.error_code == "NoSuchBucketPolicy":
|
|
|
|
current_policy = None
|
|
|
|
else:
|
|
|
|
module.fail_json(msg=e.message)
|
|
|
|
|
|
|
|
if current_policy is not None and policy is not None:
|
|
|
|
if policy is not None:
|
|
|
|
policy = json.dumps(policy)
|
|
|
|
|
|
|
|
if json.loads(current_policy) != json.loads(policy):
|
|
|
|
try:
|
|
|
|
bucket.set_policy(policy)
|
|
|
|
changed = True
|
|
|
|
current_policy = bucket.get_policy()
|
|
|
|
except S3ResponseError, e:
|
|
|
|
module.fail_json(msg=e.message)
|
|
|
|
|
|
|
|
elif current_policy is None and policy is not None:
|
|
|
|
policy = json.dumps(policy)
|
|
|
|
|
|
|
|
try:
|
|
|
|
bucket.set_policy(policy)
|
|
|
|
changed = True
|
|
|
|
current_policy = bucket.get_policy()
|
|
|
|
except S3ResponseError, e:
|
|
|
|
module.fail_json(msg=e.message)
|
|
|
|
|
|
|
|
elif current_policy is not None and policy is None:
|
|
|
|
try:
|
|
|
|
bucket.delete_policy()
|
|
|
|
changed = True
|
|
|
|
current_policy = bucket.get_policy()
|
|
|
|
except S3ResponseError, e:
|
|
|
|
if e.error_code == "NoSuchBucketPolicy":
|
|
|
|
current_policy = None
|
|
|
|
else:
|
|
|
|
module.fail_json(msg=e.message)
|
|
|
|
|
|
|
|
####
|
|
|
|
## Fix up json of policy so it's not escaped
|
|
|
|
####
|
|
|
|
|
|
|
|
# Tags
|
|
|
|
try:
|
|
|
|
current_tags = bucket.get_tags()
|
|
|
|
tag_set = TagSet()
|
|
|
|
except S3ResponseError, e:
|
|
|
|
if e.error_code == "NoSuchTagSet":
|
|
|
|
current_tags = None
|
|
|
|
else:
|
|
|
|
module.fail_json(msg=e.message)
|
|
|
|
|
|
|
|
if current_tags is not None or tags is not None:
|
|
|
|
|
|
|
|
if current_tags is None:
|
|
|
|
current_tags_dict = {}
|
|
|
|
else:
|
|
|
|
current_tags_dict = dict((t.key, t.value) for t in current_tags[0])
|
|
|
|
|
|
|
|
if current_tags_dict != tags:
|
|
|
|
try:
|
|
|
|
if tags:
|
|
|
|
bucket.set_tags(create_tags_container(tags))
|
|
|
|
else:
|
|
|
|
bucket.delete_tags()
|
|
|
|
current_tags_dict = tags
|
|
|
|
changed = True
|
|
|
|
except S3ResponseError, e:
|
|
|
|
module.fail_json(msg=e.message)
|
|
|
|
|
|
|
|
module.exit_json(changed=changed, name=bucket.name, versioning=versioning_status, requester_pays=requester_pays_status, policy=current_policy, tags=current_tags_dict)
|
|
|
|
|
|
|
|
def _destroy_bucket(connection, module):
|
|
|
|
|
|
|
|
force = module.params.get("force")
|
|
|
|
name = module.params.get("name")
|
|
|
|
changed = False
|
|
|
|
|
|
|
|
try:
|
|
|
|
bucket = connection.get_bucket(name)
|
|
|
|
except S3ResponseError, e:
|
|
|
|
if e.error_code != "NoSuchBucket":
|
|
|
|
module.fail_json(msg=e.message)
|
|
|
|
else:
|
|
|
|
# Bucket already absent
|
|
|
|
module.exit_json(changed=changed)
|
|
|
|
|
|
|
|
if force:
|
|
|
|
try:
|
|
|
|
# Empty the bucket
|
|
|
|
for key in bucket.list():
|
|
|
|
key.delete()
|
|
|
|
|
|
|
|
except BotoServerError, e:
|
|
|
|
module.fail_json(msg=e.message)
|
|
|
|
|
|
|
|
try:
|
|
|
|
bucket = connection.delete_bucket(name)
|
|
|
|
changed = True
|
|
|
|
except S3ResponseError, e:
|
|
|
|
module.fail_json(msg=e.message)
|
|
|
|
|
|
|
|
module.exit_json(changed=changed)
|
|
|
|
|
|
|
|
def _create_bucket_ceph(connection, module, location):
|
|
|
|
|
|
|
|
name = module.params.get("name")
|
|
|
|
|
|
|
|
changed = False
|
|
|
|
|
|
|
|
try:
|
|
|
|
bucket = connection.get_bucket(name)
|
|
|
|
except S3ResponseError, e:
|
|
|
|
try:
|
|
|
|
bucket = connection.create_bucket(name, location=location)
|
|
|
|
changed = True
|
|
|
|
except S3CreateError, e:
|
|
|
|
module.fail_json(msg=e.message)
|
|
|
|
|
|
|
|
module.exit_json(changed=changed)
|
|
|
|
|
|
|
|
def _destroy_bucket_ceph(connection, module):
|
|
|
|
|
|
|
|
_destroy_bucket(connection, module)
|
|
|
|
|
|
|
|
def create_bucket(connection, module, location, flavour='aws'):
|
|
|
|
if flavour == 'ceph':
|
|
|
|
_create_bucket_ceph(connection, module, location)
|
|
|
|
else:
|
|
|
|
_create_bucket(connection, module, location)
|
|
|
|
|
|
|
|
def destroy_bucket(connection, module, flavour='aws'):
|
|
|
|
if flavour == 'ceph':
|
|
|
|
_destroy_bucket_ceph(connection, module)
|
|
|
|
else:
|
|
|
|
_destroy_bucket(connection, module)
|
|
|
|
|
|
|
|
def is_fakes3(s3_url):
|
|
|
|
""" Return True if s3_url has scheme fakes3:// """
|
|
|
|
if s3_url is not None:
|
|
|
|
return urlparse.urlparse(s3_url).scheme in ('fakes3', 'fakes3s')
|
|
|
|
else:
|
|
|
|
return False
|
|
|
|
|
|
|
|
def is_walrus(s3_url):
|
|
|
|
""" Return True if it's Walrus endpoint, not S3
|
|
|
|
|
|
|
|
We assume anything other than *.amazonaws.com is Walrus"""
|
|
|
|
if s3_url is not None:
|
|
|
|
o = urlparse.urlparse(s3_url)
|
|
|
|
return not o.hostname.endswith('amazonaws.com')
|
|
|
|
else:
|
|
|
|
return False
|
|
|
|
|
|
|
|
def main():
|
|
|
|
|
|
|
|
argument_spec = ec2_argument_spec()
|
|
|
|
argument_spec.update(
|
|
|
|
dict(
|
|
|
|
force = dict(required=False, default='no', type='bool'),
|
|
|
|
policy = dict(required=False, default=None),
|
|
|
|
name = dict(required=True),
|
|
|
|
requester_pays = dict(default='no', type='bool'),
|
|
|
|
s3_url = dict(aliases=['S3_URL']),
|
|
|
|
state = dict(default='present', choices=['present', 'absent']),
|
|
|
|
tags = dict(required=None, default={}, type='dict'),
|
|
|
|
versioning = dict(default='no', type='bool'),
|
|
|
|
ceph = dict(default='no', type='bool')
|
|
|
|
)
|
|
|
|
)
|
|
|
|
|
|
|
|
module = AnsibleModule(argument_spec=argument_spec)
|
|
|
|
|
|
|
|
if not HAS_BOTO:
|
|
|
|
module.fail_json(msg='boto required for this module')
|
|
|
|
|
|
|
|
region, ec2_url, aws_connect_params = get_aws_connection_info(module)
|
|
|
|
|
|
|
|
if region in ('us-east-1', '', None):
|
|
|
|
# S3ism for the US Standard region
|
|
|
|
location = Location.DEFAULT
|
|
|
|
else:
|
|
|
|
# Boto uses symbolic names for locations but region strings will
|
|
|
|
# actually work fine for everything except us-east-1 (US Standard)
|
|
|
|
location = region
|
|
|
|
|
|
|
|
s3_url = module.params.get('s3_url')
|
|
|
|
|
|
|
|
# allow eucarc environment variables to be used if ansible vars aren't set
|
|
|
|
if not s3_url and 'S3_URL' in os.environ:
|
|
|
|
s3_url = os.environ['S3_URL']
|
|
|
|
|
|
|
|
ceph = module.params.get('ceph')
|
|
|
|
|
|
|
|
if ceph and not s3_url:
|
|
|
|
module.fail_json(msg='ceph flavour requires s3_url')
|
|
|
|
|
|
|
|
flavour = 'aws'
|
|
|
|
|
|
|
|
# Look at s3_url and tweak connection settings
|
|
|
|
# if connecting to Walrus or fakes3
|
|
|
|
try:
|
|
|
|
if s3_url and ceph:
|
|
|
|
ceph = urlparse.urlparse(s3_url)
|
|
|
|
connection = boto.connect_s3(
|
|
|
|
host=ceph.hostname,
|
|
|
|
port=ceph.port,
|
|
|
|
is_secure=ceph.scheme == 'https',
|
|
|
|
calling_format=OrdinaryCallingFormat(),
|
|
|
|
**aws_connect_params
|
|
|
|
)
|
|
|
|
flavour = 'ceph'
|
|
|
|
elif is_fakes3(s3_url):
|
|
|
|
fakes3 = urlparse.urlparse(s3_url)
|
|
|
|
connection = S3Connection(
|
|
|
|
is_secure=fakes3.scheme == 'fakes3s',
|
|
|
|
host=fakes3.hostname,
|
|
|
|
port=fakes3.port,
|
|
|
|
calling_format=OrdinaryCallingFormat(),
|
|
|
|
**aws_connect_params
|
|
|
|
)
|
|
|
|
elif is_walrus(s3_url):
|
|
|
|
walrus = urlparse.urlparse(s3_url).hostname
|
|
|
|
connection = boto.connect_walrus(walrus, **aws_connect_params)
|
|
|
|
else:
|
|
|
|
connection = boto.s3.connect_to_region(location, is_secure=True, calling_format=OrdinaryCallingFormat(), **aws_connect_params)
|
|
|
|
# use this as fallback because connect_to_region seems to fail in boto + non 'classic' aws accounts in some cases
|
|
|
|
if connection is None:
|
|
|
|
connection = boto.connect_s3(**aws_connect_params)
|
|
|
|
|
|
|
|
except boto.exception.NoAuthHandlerFound, e:
|
|
|
|
module.fail_json(msg='No Authentication Handler found: %s ' % str(e))
|
|
|
|
except Exception, e:
|
|
|
|
module.fail_json(msg='Failed to connect to S3: %s' % str(e))
|
|
|
|
|
|
|
|
if connection is None: # this should never happen
|
|
|
|
module.fail_json(msg ='Unknown error, failed to create s3 connection, no information from boto.')
|
|
|
|
|
|
|
|
state = module.params.get("state")
|
|
|
|
|
|
|
|
if state == 'present':
|
|
|
|
create_bucket(connection, module, location, flavour=flavour)
|
|
|
|
elif state == 'absent':
|
|
|
|
destroy_bucket(connection, module, flavour=flavour)
|
|
|
|
|
|
|
|
from ansible.module_utils.basic import *
|
|
|
|
from ansible.module_utils.ec2 import *
|
|
|
|
import urlparse
|
|
|
|
|
|
|
|
if __name__ == '__main__':
|
|
|
|
main()
|