From: David-Sarah Hopwood Date: Tue, 29 Jan 2013 23:17:31 +0000 (+0000) Subject: Add stub OpenStack container impl. as a copy of S3 container impl. X-Git-Url: https://git.rkrishnan.org/%5B/frontends/%22news.html/htmlfontify-example.html?a=commitdiff_plain;h=2dc48bc8c5cab570da6725d01f174722fae254a8;p=tahoe-lafs%2Ftahoe-lafs.git Add stub OpenStack container impl. as a copy of S3 container impl. Generalize the container instantiation to work for either. Signed-off-by: David-Sarah Hopwood --- diff --git a/src/allmydata/storage/backends/cloud/cloud_backend.py b/src/allmydata/storage/backends/cloud/cloud_backend.py index 4fcb0dc5..3d92525f 100644 --- a/src/allmydata/storage/backends/cloud/cloud_backend.py +++ b/src/allmydata/storage/backends/cloud/cloud_backend.py @@ -1,4 +1,6 @@ +import sys + from twisted.internet import defer from zope.interface import implements @@ -18,6 +20,9 @@ from allmydata.storage.backends.cloud.cloud_common import get_share_key, delete_ from allmydata.mutable.layout import MUTABLE_MAGIC +CLOUD_INTERFACES = ("cloud.s3", "cloud.openstack") + + def get_cloud_share(container, storage_index, shnum, total_size): key = get_share_key(storage_index, shnum) d = container.get_object(key) @@ -32,9 +37,6 @@ def get_cloud_share(container, storage_index, shnum, total_size): def configure_cloud_backend(storedir, config): - # REMIND: when multiple container implementations are supported, only import the container we're going to use. - from allmydata.storage.backends.cloud.s3.s3_container import configure_s3_container - if config.get_config("storage", "readonly", False, boolean=True): raise InvalidValueError("[storage]readonly is not supported by the cloud backend; " "make the container read-only instead.") @@ -43,15 +45,13 @@ def configure_cloud_backend(storedir, config): if backendtype == "s3": backendtype = "cloud.s3" - container_configurators = { - 'cloud.s3': configure_s3_container, - } - - if backendtype not in container_configurators: + if backendtype not in CLOUD_INTERFACES: raise InvalidValueError("%s is not supported by the cloud backend; it must be one of %s" - % (quote_output("[storage]backend = " + backendtype), container_configurators.keys()) ) + % (quote_output("[storage]backend = " + backendtype), CLOUD_INTERFACES) ) - container = container_configurators[backendtype](storedir, config) + pkgname = "allmydata.storage.backends." + backendtype + __import__(pkgname) + container = sys.modules[pkgname].configure_container(storedir, config) return CloudBackend(container) diff --git a/src/allmydata/storage/backends/cloud/openstack/__init__.py b/src/allmydata/storage/backends/cloud/openstack/__init__.py new file mode 100644 index 00000000..86695ed0 --- /dev/null +++ b/src/allmydata/storage/backends/cloud/openstack/__init__.py @@ -0,0 +1,4 @@ + +from allmydata.storage.backends.cloud.openstack.openstack_container import configure_openstack_container + +configure_container = configure_openstack_container diff --git a/src/allmydata/storage/backends/cloud/openstack/openstack_container.py b/src/allmydata/storage/backends/cloud/openstack/openstack_container.py new file mode 100644 index 00000000..ca5d889a --- /dev/null +++ b/src/allmydata/storage/backends/cloud/openstack/openstack_container.py @@ -0,0 +1,99 @@ + +from zope.interface import implements + +from allmydata.node import InvalidValueError +from allmydata.storage.backends.cloud.cloud_common import IContainer, \ + ContainerRetryMixin, ContainerListMixin + + +def configure_openstack_container(storedir, config): + accesskeyid = config.get_config("storage", "s3.access_key_id") + secretkey = config.get_or_create_private_config("s3secret") + usertoken = config.get_optional_private_config("s3usertoken") + producttoken = config.get_optional_private_config("s3producttoken") + if producttoken and not usertoken: + raise InvalidValueError("If private/s3producttoken is present, private/s3usertoken must also be present.") + url = config.get_config("storage", "s3.url", "http://s3.amazonaws.com") + container_name = config.get_config("storage", "s3.bucket") + + return S3Container(accesskeyid, secretkey, url, container_name, usertoken, producttoken) + + +class S3Container(ContainerRetryMixin, ContainerListMixin): + implements(IContainer) + """ + I represent a real S3 container (bucket), accessed using the txaws library. + """ + + def __init__(self, access_key, secret_key, url, container_name, usertoken=None, producttoken=None): + # We only depend on txaws when this class is actually instantiated. + from txaws.credentials import AWSCredentials + from txaws.service import AWSServiceEndpoint + from txaws.s3.client import S3Client, Query + from txaws.s3.exception import S3Error + + creds = AWSCredentials(access_key=access_key, secret_key=secret_key) + endpoint = AWSServiceEndpoint(uri=url) + + query_factory = None + if usertoken is not None: + def make_query(*args, **kwargs): + amz_headers = kwargs.get("amz_headers", {}) + if producttoken is not None: + amz_headers["security-token"] = (usertoken, producttoken) + else: + amz_headers["security-token"] = usertoken + kwargs["amz_headers"] = amz_headers + + return Query(*args, **kwargs) + query_factory = make_query + + self.client = S3Client(creds=creds, endpoint=endpoint, query_factory=query_factory) + self.container_name = container_name + self.ServiceError = S3Error + + def __repr__(self): + return ("<%s %r>" % (self.__class__.__name__, self.container_name,)) + + def create(self): + return self._do_request('create bucket', self.client.create, self.container_name) + + def delete(self): + return self._do_request('delete bucket', self.client.delete, self.container_name) + + def list_some_objects(self, **kwargs): + return self._do_request('list objects', self.client.get_bucket, self.container_name, **kwargs) + + def put_object(self, object_name, data, content_type='application/octet-stream', metadata={}): + return self._do_request('PUT object', self.client.put_object, self.container_name, + object_name, data, content_type, metadata) + + def get_object(self, object_name): + return self._do_request('GET object', self.client.get_object, self.container_name, object_name) + + def head_object(self, object_name): + return self._do_request('HEAD object', self.client.head_object, self.container_name, object_name) + + def delete_object(self, object_name): + return self._do_request('DELETE object', self.client.delete_object, self.container_name, object_name) + + def put_policy(self, policy): + """ + Set access control policy on a bucket. + """ + query = self.client.query_factory( + action='PUT', creds=self.client.creds, endpoint=self.client.endpoint, + bucket=self.container_name, object_name='?policy', data=policy) + return self._do_request('PUT policy', query.submit) + + def get_policy(self): + query = self.client.query_factory( + action='GET', creds=self.client.creds, endpoint=self.client.endpoint, + bucket=self.container_name, object_name='?policy') + return self._do_request('GET policy', query.submit) + + def delete_policy(self): + query = self.client.query_factory( + action='DELETE', creds=self.client.creds, endpoint=self.client.endpoint, + bucket=self.container_name, object_name='?policy') + return self._do_request('DELETE policy', query.submit) diff --git a/src/allmydata/storage/backends/cloud/s3/__init__.py b/src/allmydata/storage/backends/cloud/s3/__init__.py index e69de29b..1dc7b546 100644 --- a/src/allmydata/storage/backends/cloud/s3/__init__.py +++ b/src/allmydata/storage/backends/cloud/s3/__init__.py @@ -0,0 +1,4 @@ + +from allmydata.storage.backends.cloud.s3.s3_container import configure_s3_container + +configure_container = configure_s3_container diff --git a/src/allmydata/storage/backends/cloud/s3/s3_container.py b/src/allmydata/storage/backends/cloud/s3/s3_container.py index d1d6712a..0a9b8af6 100644 --- a/src/allmydata/storage/backends/cloud/s3/s3_container.py +++ b/src/allmydata/storage/backends/cloud/s3/s3_container.py @@ -7,8 +7,6 @@ from allmydata.storage.backends.cloud.cloud_common import IContainer, \ def configure_s3_container(storedir, config): - from allmydata.storage.backends.cloud.s3.s3_container import S3Container - accesskeyid = config.get_config("storage", "s3.access_key_id") secretkey = config.get_or_create_private_config("s3secret") usertoken = config.get_optional_private_config("s3usertoken")