From: Brian Warner Date: Mon, 16 Feb 2009 21:58:44 +0000 (-0700) Subject: test/no_network: new test harness, like system-test but doesn't use the network so... X-Git-Tag: allmydata-tahoe-1.4.0~215 X-Git-Url: https://git.rkrishnan.org/components/com_hotproperty/%22doc.html/architecture.txt?a=commitdiff_plain;h=ebe0f2ac081be021724dd1e19bdbe83278fab016;p=tahoe-lafs%2Ftahoe-lafs.git test/no_network: new test harness, like system-test but doesn't use the network so it's faster --- diff --git a/src/allmydata/client.py b/src/allmydata/client.py index b8577978..97e16bbc 100644 --- a/src/allmydata/client.py +++ b/src/allmydata/client.py @@ -195,6 +195,9 @@ class Client(node.Node, pollmixin.PollMixin): self.download_cache = cachedir.CacheDirectoryManager(download_cachedir) self.download_cache.setServiceParent(self) self.add_service(Downloader(self.stats_provider)) + self.init_stub_client() + + def init_stub_client(self): def _publish(res): # we publish an empty object so that the introducer can count how # many clients are connected and see what versions they're diff --git a/src/allmydata/test/no_network.py b/src/allmydata/test/no_network.py new file mode 100644 index 00000000..55e678e0 --- /dev/null +++ b/src/allmydata/test/no_network.py @@ -0,0 +1,157 @@ + +# This contains a test harness that creates a full Tahoe grid in a single +# process (actually in a single MultiService) which does not use the network. +# It does not use an Introducer, and there are no foolscap Tubs. Each storage +# server puts real shares on disk, but is accessed through loopback +# RemoteReferences instead of over serialized SSL. It is not as complete as +# the common.SystemTestMixin framework (which does use the network), but +# should be considerably faster. It should be useful for tests which want to +# examine and/or manipulate the uploaded shares, checker/verifier/repairer +# tests, etc. The clients have no Tubs, so it is not useful for tests that +# involve a Helper, a KeyGenerator, or the control.furl . + +import os.path +import sha +from twisted.application import service +from foolscap import Referenceable +from foolscap.eventual import fireEventually +from base64 import b32encode +from allmydata.client import Client +from allmydata.storage import StorageServer +from allmydata.util import fileutil, idlib, hashutil +from allmydata.introducer.client import RemoteServiceConnector + +class IntentionalError(Exception): + pass + +class Marker: + pass + +class LocalWrapper: + def __init__(self, original): + self.original = original + self.broken = False + self.post_call_notifier = None + self.disconnectors = {} + + def callRemote(self, methname, *args, **kwargs): + # this is ideally a Membrane, but that's too hard + def wrap(a): + if isinstance(a, Referenceable): + return LocalWrapper(a) + else: + return a + args = tuple([wrap(a) for a in args]) + kwargs = dict([(k,wrap(kwargs[k])) for k in kwargs]) + def _call(): + if self.broken: + raise IntentionalError("I was asked to break") + meth = getattr(self.original, "remote_" + methname) + return meth(*args, **kwargs) + d = fireEventually() + d.addCallback(lambda res: _call()) + def _return_membrane(res): + # rather than complete the difficult task of building a + # fully-general Membrane (which would locate all Referenceable + # objects that cross the simulated wire and replace them with + # wrappers), we special-case certain methods that we happen to + # know will return Referenceables. + if methname == "allocate_buckets": + (alreadygot, allocated) = res + for shnum in allocated: + allocated[shnum] = LocalWrapper(allocated[shnum]) + if methname == "get_buckets": + for shnum in res: + res[shnum] = LocalWrapper(res[shnum]) + return res + d.addCallback(_return_membrane) + if self.post_call_notifier: + d.addCallback(self.post_call_notifier, methname) + return d + + def notifyOnDisconnect(self, f, *args, **kwargs): + m = Marker() + self.disconnectors[m] = (f, args, kwargs) + return m + def dontNotifyOnDisconnect(self, marker): + del self.disconnectors[marker] + +class VersionedLocalWrapper(LocalWrapper): + def __init__(self, original, service_name): + LocalWrapper.__init__(self, original) + try: + version = original.remote_get_version() + except AttributeError: + version = RemoteServiceConnector.VERSION_DEFAULTS[service_name] + self.version = version + +class NoNetworkClient(Client): + + def create_tub(self): + pass + def init_introducer_client(self): + pass + def setup_logging(self): + pass + def startService(self): + service.MultiService.startService(self) + def stopService(self): + service.MultiService.stopService(self) + def when_tub_ready(self): + raise RuntimeError("NoNetworkClient has no Tub") + def init_control(self): + pass + def init_helper(self): + pass + def init_key_gen(self): + pass + def init_storage(self): + pass + def init_stub_client(self): + pass + + def get_servers(self, service_name): + return self._servers + + def get_permuted_peers(self, service_name, key): + return sorted(self._servers, key=lambda x: sha.new(key+x[0]).digest()) + + +class NoNetworkGrid(service.MultiService): + def __init__(self, basedir, num_clients=1, num_servers=10): + service.MultiService.__init__(self) + self.basedir = basedir + fileutil.make_dirs(basedir) + + self.servers = {} + self.all_servers = [] + for i in range(num_servers): + serverid = hashutil.tagged_hash("serverid", str(i))[:20] + serverdir = os.path.join(basedir, "servers", + idlib.shortnodeid_b2a(serverid)) + fileutil.make_dirs(serverdir) + ss = StorageServer(serverdir) + self.add_server(serverid, ss) + + self.clients = [] + for i in range(num_clients): + clientid = hashutil.tagged_hash("clientid", str(i))[:20] + clientdir = os.path.join(basedir, "clients", + idlib.shortnodeid_b2a(clientid)) + fileutil.make_dirs(clientdir) + c = NoNetworkClient(clientdir) + c.nodeid = clientid + c.short_nodeid = b32encode(clientid).lower()[:8] + c._servers = self.all_servers # can be updated later + c.setServiceParent(self) + self.clients.append(c) + + def add_server(self, serverid, ss): + # TODO: ss.setServiceParent(self), but first remove the goofy + # self.parent.nodeid from Storage.startService . At the moment, + # Storage doesn't really need to be startServiced, but it will in + # the future. + ss.setNodeID(serverid) + lw = VersionedLocalWrapper(ss, "storage") + self.servers[serverid] = lw + self.all_servers.append( (serverid, lw) ) diff --git a/src/allmydata/test/test_no_network.py b/src/allmydata/test/test_no_network.py new file mode 100644 index 00000000..911dc86f --- /dev/null +++ b/src/allmydata/test/test_no_network.py @@ -0,0 +1,42 @@ + +# Test the NoNetworkGrid test harness + +from twisted.trial import unittest +from twisted.application import service +from allmydata.test.no_network import NoNetworkGrid +from allmydata.immutable.upload import Data + + +class Harness(unittest.TestCase): + def setUp(self): + self.s = service.MultiService() + self.s.startService() + + def tearDown(self): + return self.s.stopService() + + def test_create(self): + basedir = "no_network/Harness/create" + g = NoNetworkGrid(basedir) + g.startService() + return g.stopService() + + def test_upload(self): + basedir = "no_network/Harness/upload" + g = NoNetworkGrid(basedir) + g.setServiceParent(self.s) + + c0 = g.clients[0] + DATA = "Data to upload" * 100 + data = Data(DATA, "") + d = c0.upload(data) + def _uploaded(res): + n = c0.create_node_from_uri(res.uri) + return n.download_to_data() + d.addCallback(_uploaded) + def _check(res): + self.failUnlessEqual(res, DATA) + d.addCallback(_check) + + return d +