mirror of
https://github.com/SickGear/SickGear.git
synced 2024-12-05 02:43:37 +00:00
673 lines
26 KiB
Python
673 lines
26 KiB
Python
|
# Copyright (c) 2006-2012 Mitch Garnaat http://garnaat.org/
|
||
|
# Copyright (c) 2012 Amazon.com, Inc. or its affiliates.
|
||
|
# Copyright (c) 2010, Eucalyptus Systems, Inc.
|
||
|
# All rights reserved.
|
||
|
#
|
||
|
# Permission is hereby granted, free of charge, to any person obtaining a
|
||
|
# copy of this software and associated documentation files (the
|
||
|
# "Software"), to deal in the Software without restriction, including
|
||
|
# without limitation the rights to use, copy, modify, merge, publish, dis-
|
||
|
# tribute, sublicense, and/or sell copies of the Software, and to permit
|
||
|
# persons to whom the Software is furnished to do so, subject to the fol-
|
||
|
# lowing conditions:
|
||
|
#
|
||
|
# The above copyright notice and this permission notice shall be included
|
||
|
# in all copies or substantial portions of the Software.
|
||
|
#
|
||
|
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
|
||
|
# OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABIL-
|
||
|
# ITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT
|
||
|
# SHALL THE AUTHOR BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
|
||
|
# WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
||
|
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
|
||
|
# IN THE SOFTWARE.
|
||
|
|
||
|
import xml.sax
|
||
|
import base64
|
||
|
from boto.compat import six, urllib
|
||
|
import time
|
||
|
|
||
|
from boto.auth import detect_potential_s3sigv4
|
||
|
import boto.utils
|
||
|
from boto.connection import AWSAuthConnection
|
||
|
from boto import handler
|
||
|
from boto.s3.bucket import Bucket
|
||
|
from boto.s3.key import Key
|
||
|
from boto.resultset import ResultSet
|
||
|
from boto.exception import BotoClientError, S3ResponseError
|
||
|
|
||
|
|
||
|
def check_lowercase_bucketname(n):
|
||
|
"""
|
||
|
Bucket names must not contain uppercase characters. We check for
|
||
|
this by appending a lowercase character and testing with islower().
|
||
|
Note this also covers cases like numeric bucket names with dashes.
|
||
|
|
||
|
>>> check_lowercase_bucketname("Aaaa")
|
||
|
Traceback (most recent call last):
|
||
|
...
|
||
|
BotoClientError: S3Error: Bucket names cannot contain upper-case
|
||
|
characters when using either the sub-domain or virtual hosting calling
|
||
|
format.
|
||
|
|
||
|
>>> check_lowercase_bucketname("1234-5678-9123")
|
||
|
True
|
||
|
>>> check_lowercase_bucketname("abcdefg1234")
|
||
|
True
|
||
|
"""
|
||
|
if not (n + 'a').islower():
|
||
|
raise BotoClientError("Bucket names cannot contain upper-case " \
|
||
|
"characters when using either the sub-domain or virtual " \
|
||
|
"hosting calling format.")
|
||
|
return True
|
||
|
|
||
|
|
||
|
def assert_case_insensitive(f):
|
||
|
def wrapper(*args, **kwargs):
|
||
|
if len(args) == 3 and check_lowercase_bucketname(args[2]):
|
||
|
pass
|
||
|
return f(*args, **kwargs)
|
||
|
return wrapper
|
||
|
|
||
|
|
||
|
class _CallingFormat(object):
|
||
|
|
||
|
def get_bucket_server(self, server, bucket):
|
||
|
return ''
|
||
|
|
||
|
def build_url_base(self, connection, protocol, server, bucket, key=''):
|
||
|
url_base = '%s://' % protocol
|
||
|
url_base += self.build_host(server, bucket)
|
||
|
url_base += connection.get_path(self.build_path_base(bucket, key))
|
||
|
return url_base
|
||
|
|
||
|
def build_host(self, server, bucket):
|
||
|
if bucket == '':
|
||
|
return server
|
||
|
else:
|
||
|
return self.get_bucket_server(server, bucket)
|
||
|
|
||
|
def build_auth_path(self, bucket, key=''):
|
||
|
key = boto.utils.get_utf8_value(key)
|
||
|
path = ''
|
||
|
if bucket != '':
|
||
|
path = '/' + bucket
|
||
|
return path + '/%s' % urllib.parse.quote(key)
|
||
|
|
||
|
def build_path_base(self, bucket, key=''):
|
||
|
key = boto.utils.get_utf8_value(key)
|
||
|
return '/%s' % urllib.parse.quote(key)
|
||
|
|
||
|
|
||
|
class SubdomainCallingFormat(_CallingFormat):
|
||
|
|
||
|
@assert_case_insensitive
|
||
|
def get_bucket_server(self, server, bucket):
|
||
|
return '%s.%s' % (bucket, server)
|
||
|
|
||
|
|
||
|
class VHostCallingFormat(_CallingFormat):
|
||
|
|
||
|
@assert_case_insensitive
|
||
|
def get_bucket_server(self, server, bucket):
|
||
|
return bucket
|
||
|
|
||
|
|
||
|
class OrdinaryCallingFormat(_CallingFormat):
|
||
|
|
||
|
def get_bucket_server(self, server, bucket):
|
||
|
return server
|
||
|
|
||
|
def build_path_base(self, bucket, key=''):
|
||
|
key = boto.utils.get_utf8_value(key)
|
||
|
path_base = '/'
|
||
|
if bucket:
|
||
|
path_base += "%s/" % bucket
|
||
|
return path_base + urllib.parse.quote(key)
|
||
|
|
||
|
|
||
|
class ProtocolIndependentOrdinaryCallingFormat(OrdinaryCallingFormat):
|
||
|
|
||
|
def build_url_base(self, connection, protocol, server, bucket, key=''):
|
||
|
url_base = '//'
|
||
|
url_base += self.build_host(server, bucket)
|
||
|
url_base += connection.get_path(self.build_path_base(bucket, key))
|
||
|
return url_base
|
||
|
|
||
|
|
||
|
class Location(object):
|
||
|
|
||
|
DEFAULT = '' # US Classic Region
|
||
|
EU = 'EU' # Ireland
|
||
|
EUCentral1 = 'eu-central-1' # Frankfurt
|
||
|
USWest = 'us-west-1'
|
||
|
USWest2 = 'us-west-2'
|
||
|
SAEast = 'sa-east-1'
|
||
|
APNortheast = 'ap-northeast-1'
|
||
|
APSoutheast = 'ap-southeast-1'
|
||
|
APSoutheast2 = 'ap-southeast-2'
|
||
|
CNNorth1 = 'cn-north-1'
|
||
|
|
||
|
|
||
|
class NoHostProvided(object):
|
||
|
# An identifying object to help determine whether the user provided a
|
||
|
# ``host`` or not. Never instantiated.
|
||
|
pass
|
||
|
|
||
|
|
||
|
class HostRequiredError(BotoClientError):
|
||
|
pass
|
||
|
|
||
|
|
||
|
class S3Connection(AWSAuthConnection):
|
||
|
|
||
|
DefaultHost = 's3.amazonaws.com'
|
||
|
DefaultCallingFormat = boto.config.get('s3', 'calling_format', 'boto.s3.connection.SubdomainCallingFormat')
|
||
|
QueryString = 'Signature=%s&Expires=%d&AWSAccessKeyId=%s'
|
||
|
|
||
|
def __init__(self, aws_access_key_id=None, aws_secret_access_key=None,
|
||
|
is_secure=True, port=None, proxy=None, proxy_port=None,
|
||
|
proxy_user=None, proxy_pass=None,
|
||
|
host=NoHostProvided, debug=0, https_connection_factory=None,
|
||
|
calling_format=DefaultCallingFormat, path='/',
|
||
|
provider='aws', bucket_class=Bucket, security_token=None,
|
||
|
suppress_consec_slashes=True, anon=False,
|
||
|
validate_certs=None, profile_name=None):
|
||
|
no_host_provided = False
|
||
|
# Try falling back to the boto config file's value, if present.
|
||
|
if host is NoHostProvided:
|
||
|
host = boto.config.get('s3', 'host')
|
||
|
if host is None:
|
||
|
host = self.DefaultHost
|
||
|
no_host_provided = True
|
||
|
if isinstance(calling_format, six.string_types):
|
||
|
calling_format=boto.utils.find_class(calling_format)()
|
||
|
self.calling_format = calling_format
|
||
|
self.bucket_class = bucket_class
|
||
|
self.anon = anon
|
||
|
super(S3Connection, self).__init__(host,
|
||
|
aws_access_key_id, aws_secret_access_key,
|
||
|
is_secure, port, proxy, proxy_port, proxy_user, proxy_pass,
|
||
|
debug=debug, https_connection_factory=https_connection_factory,
|
||
|
path=path, provider=provider, security_token=security_token,
|
||
|
suppress_consec_slashes=suppress_consec_slashes,
|
||
|
validate_certs=validate_certs, profile_name=profile_name)
|
||
|
# We need to delay until after the call to ``super`` before checking
|
||
|
# to see if SigV4 is in use.
|
||
|
if no_host_provided:
|
||
|
if 'hmac-v4-s3' in self._required_auth_capability():
|
||
|
raise HostRequiredError(
|
||
|
"When using SigV4, you must specify a 'host' parameter."
|
||
|
)
|
||
|
|
||
|
@detect_potential_s3sigv4
|
||
|
def _required_auth_capability(self):
|
||
|
if self.anon:
|
||
|
return ['anon']
|
||
|
else:
|
||
|
return ['s3']
|
||
|
|
||
|
def __iter__(self):
|
||
|
for bucket in self.get_all_buckets():
|
||
|
yield bucket
|
||
|
|
||
|
def __contains__(self, bucket_name):
|
||
|
return not (self.lookup(bucket_name) is None)
|
||
|
|
||
|
def set_bucket_class(self, bucket_class):
|
||
|
"""
|
||
|
Set the Bucket class associated with this bucket. By default, this
|
||
|
would be the boto.s3.key.Bucket class but if you want to subclass that
|
||
|
for some reason this allows you to associate your new class.
|
||
|
|
||
|
:type bucket_class: class
|
||
|
:param bucket_class: A subclass of Bucket that can be more specific
|
||
|
"""
|
||
|
self.bucket_class = bucket_class
|
||
|
|
||
|
def build_post_policy(self, expiration_time, conditions):
|
||
|
"""
|
||
|
Taken from the AWS book Python examples and modified for use with boto
|
||
|
"""
|
||
|
assert isinstance(expiration_time, time.struct_time), \
|
||
|
'Policy document must include a valid expiration Time object'
|
||
|
|
||
|
# Convert conditions object mappings to condition statements
|
||
|
|
||
|
return '{"expiration": "%s",\n"conditions": [%s]}' % \
|
||
|
(time.strftime(boto.utils.ISO8601, expiration_time), ",".join(conditions))
|
||
|
|
||
|
def build_post_form_args(self, bucket_name, key, expires_in=6000,
|
||
|
acl=None, success_action_redirect=None,
|
||
|
max_content_length=None,
|
||
|
http_method='http', fields=None,
|
||
|
conditions=None, storage_class='STANDARD',
|
||
|
server_side_encryption=None):
|
||
|
"""
|
||
|
Taken from the AWS book Python examples and modified for use with boto
|
||
|
This only returns the arguments required for the post form, not the
|
||
|
actual form. This does not return the file input field which also
|
||
|
needs to be added
|
||
|
|
||
|
:type bucket_name: string
|
||
|
:param bucket_name: Bucket to submit to
|
||
|
|
||
|
:type key: string
|
||
|
:param key: Key name, optionally add ${filename} to the end to
|
||
|
attach the submitted filename
|
||
|
|
||
|
:type expires_in: integer
|
||
|
:param expires_in: Time (in seconds) before this expires, defaults
|
||
|
to 6000
|
||
|
|
||
|
:type acl: string
|
||
|
:param acl: A canned ACL. One of:
|
||
|
* private
|
||
|
* public-read
|
||
|
* public-read-write
|
||
|
* authenticated-read
|
||
|
* bucket-owner-read
|
||
|
* bucket-owner-full-control
|
||
|
|
||
|
:type success_action_redirect: string
|
||
|
:param success_action_redirect: URL to redirect to on success
|
||
|
|
||
|
:type max_content_length: integer
|
||
|
:param max_content_length: Maximum size for this file
|
||
|
|
||
|
:type http_method: string
|
||
|
:param http_method: HTTP Method to use, "http" or "https"
|
||
|
|
||
|
:type storage_class: string
|
||
|
:param storage_class: Storage class to use for storing the object.
|
||
|
Valid values: STANDARD | REDUCED_REDUNDANCY
|
||
|
|
||
|
:type server_side_encryption: string
|
||
|
:param server_side_encryption: Specifies server-side encryption
|
||
|
algorithm to use when Amazon S3 creates an object.
|
||
|
Valid values: None | AES256
|
||
|
|
||
|
:rtype: dict
|
||
|
:return: A dictionary containing field names/values as well as
|
||
|
a url to POST to
|
||
|
|
||
|
.. code-block:: python
|
||
|
|
||
|
|
||
|
"""
|
||
|
if fields is None:
|
||
|
fields = []
|
||
|
if conditions is None:
|
||
|
conditions = []
|
||
|
expiration = time.gmtime(int(time.time() + expires_in))
|
||
|
|
||
|
# Generate policy document
|
||
|
conditions.append('{"bucket": "%s"}' % bucket_name)
|
||
|
if key.endswith("${filename}"):
|
||
|
conditions.append('["starts-with", "$key", "%s"]' % key[:-len("${filename}")])
|
||
|
else:
|
||
|
conditions.append('{"key": "%s"}' % key)
|
||
|
if acl:
|
||
|
conditions.append('{"acl": "%s"}' % acl)
|
||
|
fields.append({"name": "acl", "value": acl})
|
||
|
if success_action_redirect:
|
||
|
conditions.append('{"success_action_redirect": "%s"}' % success_action_redirect)
|
||
|
fields.append({"name": "success_action_redirect", "value": success_action_redirect})
|
||
|
if max_content_length:
|
||
|
conditions.append('["content-length-range", 0, %i]' % max_content_length)
|
||
|
|
||
|
if self.provider.security_token:
|
||
|
fields.append({'name': 'x-amz-security-token',
|
||
|
'value': self.provider.security_token})
|
||
|
conditions.append('{"x-amz-security-token": "%s"}' % self.provider.security_token)
|
||
|
|
||
|
if storage_class:
|
||
|
fields.append({'name': 'x-amz-storage-class',
|
||
|
'value': storage_class})
|
||
|
conditions.append('{"x-amz-storage-class": "%s"}' % storage_class)
|
||
|
|
||
|
if server_side_encryption:
|
||
|
fields.append({'name': 'x-amz-server-side-encryption',
|
||
|
'value': server_side_encryption})
|
||
|
conditions.append('{"x-amz-server-side-encryption": "%s"}' % server_side_encryption)
|
||
|
|
||
|
policy = self.build_post_policy(expiration, conditions)
|
||
|
|
||
|
# Add the base64-encoded policy document as the 'policy' field
|
||
|
policy_b64 = base64.b64encode(policy)
|
||
|
fields.append({"name": "policy", "value": policy_b64})
|
||
|
|
||
|
# Add the AWS access key as the 'AWSAccessKeyId' field
|
||
|
fields.append({"name": "AWSAccessKeyId",
|
||
|
"value": self.aws_access_key_id})
|
||
|
|
||
|
# Add signature for encoded policy document as the
|
||
|
# 'signature' field
|
||
|
signature = self._auth_handler.sign_string(policy_b64)
|
||
|
fields.append({"name": "signature", "value": signature})
|
||
|
fields.append({"name": "key", "value": key})
|
||
|
|
||
|
# HTTPS protocol will be used if the secure HTTP option is enabled.
|
||
|
url = '%s://%s/' % (http_method,
|
||
|
self.calling_format.build_host(self.server_name(),
|
||
|
bucket_name))
|
||
|
|
||
|
return {"action": url, "fields": fields}
|
||
|
|
||
|
def generate_url_sigv4(self, expires_in, method, bucket='', key='',
|
||
|
headers=None, force_http=False,
|
||
|
response_headers=None, version_id=None,
|
||
|
iso_date=None):
|
||
|
path = self.calling_format.build_path_base(bucket, key)
|
||
|
auth_path = self.calling_format.build_auth_path(bucket, key)
|
||
|
host = self.calling_format.build_host(self.server_name(), bucket)
|
||
|
|
||
|
# For presigned URLs we should ignore the port if it's HTTPS
|
||
|
if host.endswith(':443'):
|
||
|
host = host[:-4]
|
||
|
|
||
|
params = {}
|
||
|
if version_id is not None:
|
||
|
params['VersionId'] = version_id
|
||
|
|
||
|
if response_headers is not None:
|
||
|
params.update(response_headers)
|
||
|
|
||
|
http_request = self.build_base_http_request(method, path, auth_path,
|
||
|
headers=headers, host=host,
|
||
|
params=params)
|
||
|
|
||
|
return self._auth_handler.presign(http_request, expires_in,
|
||
|
iso_date=iso_date)
|
||
|
|
||
|
def generate_url(self, expires_in, method, bucket='', key='', headers=None,
|
||
|
query_auth=True, force_http=False, response_headers=None,
|
||
|
expires_in_absolute=False, version_id=None):
|
||
|
if self._auth_handler.capability[0] == 'hmac-v4-s3' and query_auth:
|
||
|
# Handle the special sigv4 case
|
||
|
return self.generate_url_sigv4(expires_in, method, bucket=bucket,
|
||
|
key=key, headers=headers, force_http=force_http,
|
||
|
response_headers=response_headers, version_id=version_id)
|
||
|
|
||
|
headers = headers or {}
|
||
|
if expires_in_absolute:
|
||
|
expires = int(expires_in)
|
||
|
else:
|
||
|
expires = int(time.time() + expires_in)
|
||
|
auth_path = self.calling_format.build_auth_path(bucket, key)
|
||
|
auth_path = self.get_path(auth_path)
|
||
|
# optional version_id and response_headers need to be added to
|
||
|
# the query param list.
|
||
|
extra_qp = []
|
||
|
if version_id is not None:
|
||
|
extra_qp.append("versionId=%s" % version_id)
|
||
|
if response_headers:
|
||
|
for k, v in response_headers.items():
|
||
|
extra_qp.append("%s=%s" % (k, urllib.parse.quote(v)))
|
||
|
if self.provider.security_token:
|
||
|
headers['x-amz-security-token'] = self.provider.security_token
|
||
|
if extra_qp:
|
||
|
delimiter = '?' if '?' not in auth_path else '&'
|
||
|
auth_path += delimiter + '&'.join(extra_qp)
|
||
|
self.calling_format.build_path_base(bucket, key)
|
||
|
if query_auth and not self.anon:
|
||
|
c_string = boto.utils.canonical_string(method, auth_path, headers,
|
||
|
expires, self.provider)
|
||
|
b64_hmac = self._auth_handler.sign_string(c_string)
|
||
|
encoded_canonical = urllib.parse.quote(b64_hmac, safe='')
|
||
|
query_part = '?' + self.QueryString % (encoded_canonical, expires,
|
||
|
self.aws_access_key_id)
|
||
|
else:
|
||
|
query_part = ''
|
||
|
if headers:
|
||
|
hdr_prefix = self.provider.header_prefix
|
||
|
for k, v in headers.items():
|
||
|
if k.startswith(hdr_prefix):
|
||
|
# headers used for sig generation must be
|
||
|
# included in the url also.
|
||
|
extra_qp.append("%s=%s" % (k, urllib.parse.quote(v)))
|
||
|
if extra_qp:
|
||
|
delimiter = '?' if not query_part else '&'
|
||
|
query_part += delimiter + '&'.join(extra_qp)
|
||
|
if force_http:
|
||
|
protocol = 'http'
|
||
|
port = 80
|
||
|
else:
|
||
|
protocol = self.protocol
|
||
|
port = self.port
|
||
|
return self.calling_format.build_url_base(self, protocol,
|
||
|
self.server_name(port),
|
||
|
bucket, key) + query_part
|
||
|
|
||
|
def get_all_buckets(self, headers=None):
|
||
|
response = self.make_request('GET', headers=headers)
|
||
|
body = response.read()
|
||
|
if response.status > 300:
|
||
|
raise self.provider.storage_response_error(
|
||
|
response.status, response.reason, body)
|
||
|
rs = ResultSet([('Bucket', self.bucket_class)])
|
||
|
h = handler.XmlHandler(rs, self)
|
||
|
if not isinstance(body, bytes):
|
||
|
body = body.encode('utf-8')
|
||
|
xml.sax.parseString(body, h)
|
||
|
return rs
|
||
|
|
||
|
def get_canonical_user_id(self, headers=None):
|
||
|
"""
|
||
|
Convenience method that returns the "CanonicalUserID" of the
|
||
|
user who's credentials are associated with the connection.
|
||
|
The only way to get this value is to do a GET request on the
|
||
|
service which returns all buckets associated with the account.
|
||
|
As part of that response, the canonical userid is returned.
|
||
|
This method simply does all of that and then returns just the
|
||
|
user id.
|
||
|
|
||
|
:rtype: string
|
||
|
:return: A string containing the canonical user id.
|
||
|
"""
|
||
|
rs = self.get_all_buckets(headers=headers)
|
||
|
return rs.owner.id
|
||
|
|
||
|
def get_bucket(self, bucket_name, validate=True, headers=None):
|
||
|
"""
|
||
|
Retrieves a bucket by name.
|
||
|
|
||
|
If the bucket does not exist, an ``S3ResponseError`` will be raised. If
|
||
|
you are unsure if the bucket exists or not, you can use the
|
||
|
``S3Connection.lookup`` method, which will either return a valid bucket
|
||
|
or ``None``.
|
||
|
|
||
|
If ``validate=False`` is passed, no request is made to the service (no
|
||
|
charge/communication delay). This is only safe to do if you are **sure**
|
||
|
the bucket exists.
|
||
|
|
||
|
If the default ``validate=True`` is passed, a request is made to the
|
||
|
service to ensure the bucket exists. Prior to Boto v2.25.0, this fetched
|
||
|
a list of keys (but with a max limit set to ``0``, always returning an empty
|
||
|
list) in the bucket (& included better error messages), at an
|
||
|
increased expense. As of Boto v2.25.0, this now performs a HEAD request
|
||
|
(less expensive but worse error messages).
|
||
|
|
||
|
If you were relying on parsing the error message before, you should call
|
||
|
something like::
|
||
|
|
||
|
bucket = conn.get_bucket('<bucket_name>', validate=False)
|
||
|
bucket.get_all_keys(maxkeys=0)
|
||
|
|
||
|
:type bucket_name: string
|
||
|
:param bucket_name: The name of the bucket
|
||
|
|
||
|
:type headers: dict
|
||
|
:param headers: Additional headers to pass along with the request to
|
||
|
AWS.
|
||
|
|
||
|
:type validate: boolean
|
||
|
:param validate: If ``True``, it will try to verify the bucket exists
|
||
|
on the service-side. (Default: ``True``)
|
||
|
"""
|
||
|
if validate:
|
||
|
return self.head_bucket(bucket_name, headers=headers)
|
||
|
else:
|
||
|
return self.bucket_class(self, bucket_name)
|
||
|
|
||
|
def head_bucket(self, bucket_name, headers=None):
|
||
|
"""
|
||
|
Determines if a bucket exists by name.
|
||
|
|
||
|
If the bucket does not exist, an ``S3ResponseError`` will be raised.
|
||
|
|
||
|
:type bucket_name: string
|
||
|
:param bucket_name: The name of the bucket
|
||
|
|
||
|
:type headers: dict
|
||
|
:param headers: Additional headers to pass along with the request to
|
||
|
AWS.
|
||
|
|
||
|
:returns: A <Bucket> object
|
||
|
"""
|
||
|
response = self.make_request('HEAD', bucket_name, headers=headers)
|
||
|
body = response.read()
|
||
|
if response.status == 200:
|
||
|
return self.bucket_class(self, bucket_name)
|
||
|
elif response.status == 403:
|
||
|
# For backward-compatibility, we'll populate part of the exception
|
||
|
# with the most-common default.
|
||
|
err = self.provider.storage_response_error(
|
||
|
response.status,
|
||
|
response.reason,
|
||
|
body
|
||
|
)
|
||
|
err.error_code = 'AccessDenied'
|
||
|
err.error_message = 'Access Denied'
|
||
|
raise err
|
||
|
elif response.status == 404:
|
||
|
# For backward-compatibility, we'll populate part of the exception
|
||
|
# with the most-common default.
|
||
|
err = self.provider.storage_response_error(
|
||
|
response.status,
|
||
|
response.reason,
|
||
|
body
|
||
|
)
|
||
|
err.error_code = 'NoSuchBucket'
|
||
|
err.error_message = 'The specified bucket does not exist'
|
||
|
raise err
|
||
|
else:
|
||
|
raise self.provider.storage_response_error(
|
||
|
response.status, response.reason, body)
|
||
|
|
||
|
def lookup(self, bucket_name, validate=True, headers=None):
|
||
|
"""
|
||
|
Attempts to get a bucket from S3.
|
||
|
|
||
|
Works identically to ``S3Connection.get_bucket``, save for that it
|
||
|
will return ``None`` if the bucket does not exist instead of throwing
|
||
|
an exception.
|
||
|
|
||
|
:type bucket_name: string
|
||
|
:param bucket_name: The name of the bucket
|
||
|
|
||
|
:type headers: dict
|
||
|
:param headers: Additional headers to pass along with the request to
|
||
|
AWS.
|
||
|
|
||
|
:type validate: boolean
|
||
|
:param validate: If ``True``, it will try to fetch all keys within the
|
||
|
given bucket. (Default: ``True``)
|
||
|
"""
|
||
|
try:
|
||
|
bucket = self.get_bucket(bucket_name, validate, headers=headers)
|
||
|
except:
|
||
|
bucket = None
|
||
|
return bucket
|
||
|
|
||
|
def create_bucket(self, bucket_name, headers=None,
|
||
|
location=Location.DEFAULT, policy=None):
|
||
|
"""
|
||
|
Creates a new located bucket. By default it's in the USA. You can pass
|
||
|
Location.EU to create a European bucket (S3) or European Union bucket
|
||
|
(GCS).
|
||
|
|
||
|
:type bucket_name: string
|
||
|
:param bucket_name: The name of the new bucket
|
||
|
|
||
|
:type headers: dict
|
||
|
:param headers: Additional headers to pass along with the request to AWS.
|
||
|
|
||
|
:type location: str
|
||
|
:param location: The location of the new bucket. You can use one of the
|
||
|
constants in :class:`boto.s3.connection.Location` (e.g. Location.EU,
|
||
|
Location.USWest, etc.).
|
||
|
|
||
|
:type policy: :class:`boto.s3.acl.CannedACLStrings`
|
||
|
:param policy: A canned ACL policy that will be applied to the
|
||
|
new key in S3.
|
||
|
|
||
|
"""
|
||
|
check_lowercase_bucketname(bucket_name)
|
||
|
|
||
|
if policy:
|
||
|
if headers:
|
||
|
headers[self.provider.acl_header] = policy
|
||
|
else:
|
||
|
headers = {self.provider.acl_header: policy}
|
||
|
if location == Location.DEFAULT:
|
||
|
data = ''
|
||
|
else:
|
||
|
data = '<CreateBucketConfiguration><LocationConstraint>' + \
|
||
|
location + '</LocationConstraint></CreateBucketConfiguration>'
|
||
|
response = self.make_request('PUT', bucket_name, headers=headers,
|
||
|
data=data)
|
||
|
body = response.read()
|
||
|
if response.status == 409:
|
||
|
raise self.provider.storage_create_error(
|
||
|
response.status, response.reason, body)
|
||
|
if response.status == 200:
|
||
|
return self.bucket_class(self, bucket_name)
|
||
|
else:
|
||
|
raise self.provider.storage_response_error(
|
||
|
response.status, response.reason, body)
|
||
|
|
||
|
def delete_bucket(self, bucket, headers=None):
|
||
|
"""
|
||
|
Removes an S3 bucket.
|
||
|
|
||
|
In order to remove the bucket, it must first be empty. If the bucket is
|
||
|
not empty, an ``S3ResponseError`` will be raised.
|
||
|
|
||
|
:type bucket_name: string
|
||
|
:param bucket_name: The name of the bucket
|
||
|
|
||
|
:type headers: dict
|
||
|
:param headers: Additional headers to pass along with the request to
|
||
|
AWS.
|
||
|
"""
|
||
|
response = self.make_request('DELETE', bucket, headers=headers)
|
||
|
body = response.read()
|
||
|
if response.status != 204:
|
||
|
raise self.provider.storage_response_error(
|
||
|
response.status, response.reason, body)
|
||
|
|
||
|
def make_request(self, method, bucket='', key='', headers=None, data='',
|
||
|
query_args=None, sender=None, override_num_retries=None,
|
||
|
retry_handler=None):
|
||
|
if isinstance(bucket, self.bucket_class):
|
||
|
bucket = bucket.name
|
||
|
if isinstance(key, Key):
|
||
|
key = key.name
|
||
|
path = self.calling_format.build_path_base(bucket, key)
|
||
|
boto.log.debug('path=%s' % path)
|
||
|
auth_path = self.calling_format.build_auth_path(bucket, key)
|
||
|
boto.log.debug('auth_path=%s' % auth_path)
|
||
|
host = self.calling_format.build_host(self.server_name(), bucket)
|
||
|
if query_args:
|
||
|
path += '?' + query_args
|
||
|
boto.log.debug('path=%s' % path)
|
||
|
auth_path += '?' + query_args
|
||
|
boto.log.debug('auth_path=%s' % auth_path)
|
||
|
return super(S3Connection, self).make_request(
|
||
|
method, path, headers,
|
||
|
data, host, auth_path, sender,
|
||
|
override_num_retries=override_num_retries,
|
||
|
retry_handler=retry_handler
|
||
|
)
|