]> git.rkrishnan.org Git - tahoe-lafs/tahoe-lafs.git/commitdiff
Add stub OpenStack container impl. as a copy of S3 container impl.
authorDaira Hopwood <daira@jacaranda.org>
Tue, 29 Jan 2013 23:17:31 +0000 (23:17 +0000)
committerDaira Hopwood <daira@jacaranda.org>
Fri, 17 Apr 2015 21:31:37 +0000 (22:31 +0100)
Generalize the container instantiation to work for either.

Signed-off-by: David-Sarah Hopwood <david-sarah@jacaranda.org>
src/allmydata/storage/backends/cloud/cloud_backend.py
src/allmydata/storage/backends/cloud/openstack/__init__.py [new file with mode: 0644]
src/allmydata/storage/backends/cloud/openstack/openstack_container.py [new file with mode: 0644]
src/allmydata/storage/backends/cloud/s3/__init__.py
src/allmydata/storage/backends/cloud/s3/s3_container.py

index 4fcb0dc544bf5e31508a279a2d8eaa48676df85f..3d92525fbebea2376c03ab92c1e0eaaa7a7af5b3 100644 (file)
@@ -1,4 +1,6 @@
 
+import sys
+
 from twisted.internet import defer
 
 from zope.interface import implements
@@ -18,6 +20,9 @@ from allmydata.storage.backends.cloud.cloud_common import get_share_key, delete_
 from allmydata.mutable.layout import MUTABLE_MAGIC
 
 
+CLOUD_INTERFACES = ("cloud.s3", "cloud.openstack")
+
+
 def get_cloud_share(container, storage_index, shnum, total_size):
     key = get_share_key(storage_index, shnum)
     d = container.get_object(key)
@@ -32,9 +37,6 @@ def get_cloud_share(container, storage_index, shnum, total_size):
 
 
 def configure_cloud_backend(storedir, config):
-    # REMIND: when multiple container implementations are supported, only import the container we're going to use.
-    from allmydata.storage.backends.cloud.s3.s3_container import configure_s3_container
-
     if config.get_config("storage", "readonly", False, boolean=True):
         raise InvalidValueError("[storage]readonly is not supported by the cloud backend; "
                                 "make the container read-only instead.")
@@ -43,15 +45,13 @@ def configure_cloud_backend(storedir, config):
     if backendtype == "s3":
         backendtype = "cloud.s3"
 
-    container_configurators = {
-        'cloud.s3': configure_s3_container,
-    }
-
-    if backendtype not in container_configurators:
+    if backendtype not in CLOUD_INTERFACES:
         raise InvalidValueError("%s is not supported by the cloud backend; it must be one of %s"
-                                % (quote_output("[storage]backend = " + backendtype), container_configurators.keys()) )
+                                % (quote_output("[storage]backend = " + backendtype), CLOUD_INTERFACES) )
 
-    container = container_configurators[backendtype](storedir, config)
+    pkgname = "allmydata.storage.backends." + backendtype
+    __import__(pkgname)
+    container = sys.modules[pkgname].configure_container(storedir, config)
     return CloudBackend(container)
 
 
diff --git a/src/allmydata/storage/backends/cloud/openstack/__init__.py b/src/allmydata/storage/backends/cloud/openstack/__init__.py
new file mode 100644 (file)
index 0000000..86695ed
--- /dev/null
@@ -0,0 +1,4 @@
+
+from allmydata.storage.backends.cloud.openstack.openstack_container import configure_openstack_container
+
+configure_container = configure_openstack_container
diff --git a/src/allmydata/storage/backends/cloud/openstack/openstack_container.py b/src/allmydata/storage/backends/cloud/openstack/openstack_container.py
new file mode 100644 (file)
index 0000000..ca5d889
--- /dev/null
@@ -0,0 +1,99 @@
+
+from zope.interface import implements
+
+from allmydata.node import InvalidValueError
+from allmydata.storage.backends.cloud.cloud_common import IContainer, \
+     ContainerRetryMixin, ContainerListMixin
+
+
+def configure_openstack_container(storedir, config):
+    accesskeyid = config.get_config("storage", "s3.access_key_id")
+    secretkey = config.get_or_create_private_config("s3secret")
+    usertoken = config.get_optional_private_config("s3usertoken")
+    producttoken = config.get_optional_private_config("s3producttoken")
+    if producttoken and not usertoken:
+        raise InvalidValueError("If private/s3producttoken is present, private/s3usertoken must also be present.")
+    url = config.get_config("storage", "s3.url", "http://s3.amazonaws.com")
+    container_name = config.get_config("storage", "s3.bucket")
+
+    return S3Container(accesskeyid, secretkey, url, container_name, usertoken, producttoken)
+
+
+class S3Container(ContainerRetryMixin, ContainerListMixin):
+    implements(IContainer)
+    """
+    I represent a real S3 container (bucket), accessed using the txaws library.
+    """
+
+    def __init__(self, access_key, secret_key, url, container_name, usertoken=None, producttoken=None):
+        # We only depend on txaws when this class is actually instantiated.
+        from txaws.credentials import AWSCredentials
+        from txaws.service import AWSServiceEndpoint
+        from txaws.s3.client import S3Client, Query
+        from txaws.s3.exception import S3Error
+
+        creds = AWSCredentials(access_key=access_key, secret_key=secret_key)
+        endpoint = AWSServiceEndpoint(uri=url)
+
+        query_factory = None
+        if usertoken is not None:
+            def make_query(*args, **kwargs):
+                amz_headers = kwargs.get("amz_headers", {})
+                if producttoken is not None:
+                    amz_headers["security-token"] = (usertoken, producttoken)
+                else:
+                    amz_headers["security-token"] = usertoken
+                kwargs["amz_headers"] = amz_headers
+
+                return Query(*args, **kwargs)
+            query_factory = make_query
+
+        self.client = S3Client(creds=creds, endpoint=endpoint, query_factory=query_factory)
+        self.container_name = container_name
+        self.ServiceError = S3Error
+
+    def __repr__(self):
+        return ("<%s %r>" % (self.__class__.__name__, self.container_name,))
+
+    def create(self):
+        return self._do_request('create bucket', self.client.create, self.container_name)
+
+    def delete(self):
+        return self._do_request('delete bucket', self.client.delete, self.container_name)
+
+    def list_some_objects(self, **kwargs):
+        return self._do_request('list objects', self.client.get_bucket, self.container_name, **kwargs)
+
+    def put_object(self, object_name, data, content_type='application/octet-stream', metadata={}):
+        return self._do_request('PUT object', self.client.put_object, self.container_name,
+                                object_name, data, content_type, metadata)
+
+    def get_object(self, object_name):
+        return self._do_request('GET object', self.client.get_object, self.container_name, object_name)
+
+    def head_object(self, object_name):
+        return self._do_request('HEAD object', self.client.head_object, self.container_name, object_name)
+
+    def delete_object(self, object_name):
+        return self._do_request('DELETE object', self.client.delete_object, self.container_name, object_name)
+
+    def put_policy(self, policy):
+        """
+        Set access control policy on a bucket.
+        """
+        query = self.client.query_factory(
+            action='PUT', creds=self.client.creds, endpoint=self.client.endpoint,
+            bucket=self.container_name, object_name='?policy', data=policy)
+        return self._do_request('PUT policy', query.submit)
+
+    def get_policy(self):
+        query = self.client.query_factory(
+            action='GET', creds=self.client.creds, endpoint=self.client.endpoint,
+            bucket=self.container_name, object_name='?policy')
+        return self._do_request('GET policy', query.submit)
+
+    def delete_policy(self):
+        query = self.client.query_factory(
+            action='DELETE', creds=self.client.creds, endpoint=self.client.endpoint,
+            bucket=self.container_name, object_name='?policy')
+        return self._do_request('DELETE policy', query.submit)
index e69de29bb2d1d6434b8b29ae775ad8c2e48c5391..1dc7b5465e3789ec14ec30570f1d06cbb4b88d96 100644 (file)
@@ -0,0 +1,4 @@
+
+from allmydata.storage.backends.cloud.s3.s3_container import configure_s3_container
+
+configure_container = configure_s3_container
index d1d6712a453a76c7348d784995612a8213cf3c9e..0a9b8af673bb1af387d274408bfea60a6e789740 100644 (file)
@@ -7,8 +7,6 @@ from allmydata.storage.backends.cloud.cloud_common import IContainer, \
 
 
 def configure_s3_container(storedir, config):
-    from allmydata.storage.backends.cloud.s3.s3_container import S3Container
-
     accesskeyid = config.get_config("storage", "s3.access_key_id")
     secretkey = config.get_or_create_private_config("s3secret")
     usertoken = config.get_optional_private_config("s3usertoken")