]> git.rkrishnan.org Git - tahoe-lafs/tahoe-lafs.git/commitdiff
hush pyflakes-0.4.0 warnings: remove trivial unused variables. For #900.
authorBrian Warner <warner@lothar.com>
Thu, 14 Jan 2010 22:15:29 +0000 (14:15 -0800)
committerBrian Warner <warner@lothar.com>
Thu, 14 Jan 2010 22:15:29 +0000 (14:15 -0800)
43 files changed:
src/allmydata/dirnode.py
src/allmydata/immutable/download.py
src/allmydata/immutable/upload.py
src/allmydata/mutable/common.py
src/allmydata/mutable/layout.py
src/allmydata/mutable/publish.py
src/allmydata/mutable/servermap.py
src/allmydata/provisioning.py
src/allmydata/reliability.py
src/allmydata/scripts/debug.py
src/allmydata/scripts/stats_gatherer.py
src/allmydata/scripts/tahoe_backup.py
src/allmydata/scripts/tahoe_cp.py
src/allmydata/scripts/tahoe_manifest.py
src/allmydata/scripts/tahoe_rm.py
src/allmydata/storage/crawler.py
src/allmydata/storage/server.py
src/allmydata/test/check_speed.py
src/allmydata/test/common.py
src/allmydata/test/test_client.py
src/allmydata/test/test_crawler.py
src/allmydata/test/test_deepcheck.py
src/allmydata/test/test_dirnode.py
src/allmydata/test/test_encode.py
src/allmydata/test/test_filenode.py
src/allmydata/test/test_hashtree.py
src/allmydata/test/test_immutable.py
src/allmydata/test/test_introducer.py
src/allmydata/test/test_mutable.py
src/allmydata/test/test_storage.py
src/allmydata/test/test_system.py
src/allmydata/test/test_uri.py
src/allmydata/test/test_util.py
src/allmydata/test/test_web.py
src/allmydata/uri.py
src/allmydata/util/base32.py
src/allmydata/util/mathutil.py
src/allmydata/web/check_results.py
src/allmydata/web/directory.py
src/allmydata/web/filenode.py
src/allmydata/web/info.py
src/allmydata/web/reliability.py
src/allmydata/web/storage.py

index ae8516f6e8359f4e80378cc967bb7f3528a8d69f..2951d5a9fad86c78c40960c2987a9054725fdaeb 100644 (file)
@@ -413,7 +413,6 @@ class DirectoryNode:
     def set_children(self, entries, overwrite=True):
         # this takes URIs
         a = Adder(self, overwrite=overwrite)
     def set_children(self, entries, overwrite=True):
         # this takes URIs
         a = Adder(self, overwrite=overwrite)
-        node_entries = []
         for (name, e) in entries.iteritems():
             assert isinstance(name, unicode)
             if len(e) == 2:
         for (name, e) in entries.iteritems():
             assert isinstance(name, unicode)
             if len(e) == 2:
index 261b65d666326ef3ec46098324ce1008e9caf584..e087b6181e36471bec4413a4e04d61efe6d99516 100644 (file)
@@ -82,9 +82,9 @@ class ValidatedThingObtainer:
         self._log_id = log_id
 
     def _bad(self, f, validatedthingproxy):
         self._log_id = log_id
 
     def _bad(self, f, validatedthingproxy):
-        failtype = f.trap(RemoteException, DeadReferenceError,
-                          IntegrityCheckReject, layout.LayoutInvalid,
-                          layout.ShareVersionIncompatible)
+        f.trap(RemoteException, DeadReferenceError,
+               IntegrityCheckReject, layout.LayoutInvalid,
+               layout.ShareVersionIncompatible)
         level = log.WEIRD
         if f.check(DeadReferenceError):
             level = log.UNUSUAL
         level = log.WEIRD
         if f.check(DeadReferenceError):
             level = log.UNUSUAL
@@ -598,9 +598,9 @@ class BlockDownloader(log.PrefixingLogMixin):
         self.parent.hold_block(self.blocknum, data)
 
     def _got_block_error(self, f):
         self.parent.hold_block(self.blocknum, data)
 
     def _got_block_error(self, f):
-        failtype = f.trap(RemoteException, DeadReferenceError,
-                          IntegrityCheckReject,
-                          layout.LayoutInvalid, layout.ShareVersionIncompatible)
+        f.trap(RemoteException, DeadReferenceError,
+               IntegrityCheckReject, layout.LayoutInvalid,
+               layout.ShareVersionIncompatible)
         if f.check(RemoteException, DeadReferenceError):
             level = log.UNUSUAL
         else:
         if f.check(RemoteException, DeadReferenceError):
             level = log.UNUSUAL
         else:
index 0b3bd7d9437bc051710fc0ea75b1f032e3e5b445..72826994588dbd89e61a28ac81e94f93463584a8 100644 (file)
@@ -777,7 +777,6 @@ class CHKUploader:
         for shnum in self._encoder.get_shares_placed():
             peer_tracker = self._peer_trackers[shnum]
             peerid = peer_tracker.peerid
         for shnum in self._encoder.get_shares_placed():
             peer_tracker = self._peer_trackers[shnum]
             peerid = peer_tracker.peerid
-            peerid_s = idlib.shortnodeid_b2a(peerid)
             r.sharemap.add(shnum, peerid)
             r.servermap.add(peerid, shnum)
         r.pushed_shares = len(self._encoder.get_shares_placed())
             r.sharemap.add(shnum, peerid)
             r.servermap.add(peerid, shnum)
         r.pushed_shares = len(self._encoder.get_shares_placed())
index 49a04ff2462363210485677d91cdf374a50e262e..fc217dfd33853141328134557942e07ef44b2a17 100644 (file)
@@ -121,7 +121,6 @@ class ResponseCache:
         # we have a fragment that contains the whole request
 
         index = (verinfo, shnum)
         # we have a fragment that contains the whole request
 
         index = (verinfo, shnum)
-        end = offset+length
         for entry in self.cache.get(index, set()):
             (e_start, e_data, e_timestamp) = entry
             if self._inside(offset, length, e_start, len(e_data)):
         for entry in self.cache.get(index, set()):
             (e_start, e_data, e_timestamp) = entry
             if self._inside(offset, length, e_start, len(e_data)):
index ee89128cc4d30208cc274158f995b3c2b1693a40..40c7174a9ad0a0ef3364ef807638c775678efe0e 100644 (file)
@@ -166,7 +166,7 @@ def pack_offsets(verification_key_length, signature_length,
     o3 = offsets['block_hash_tree'] = o2 + share_hash_chain_length
     o4 = offsets['share_data'] = o3 + block_hash_tree_length
     o5 = offsets['enc_privkey'] = o4 + share_data_length
     o3 = offsets['block_hash_tree'] = o2 + share_hash_chain_length
     o4 = offsets['share_data'] = o3 + block_hash_tree_length
     o5 = offsets['enc_privkey'] = o4 + share_data_length
-    o6 = offsets['EOF'] = o5 + encprivkey_length
+    offsets['EOF'] = o5 + encprivkey_length
 
     return struct.pack(">LLLLQQ",
                        offsets['signature'],
 
     return struct.pack(">LLLLQQ",
                        offsets['signature'],
index 39ed649b91065ea98ca4361603af7f97fd87fa63..7ff4350dae05e2aa0f91f56660a991daf7943e17 100644 (file)
@@ -374,7 +374,6 @@ class Publish:
                                         str(self._first_write_error),
                                         self._first_write_error)
 
                                         str(self._first_write_error),
                                         self._first_write_error)
 
-        new_assignments = []
         # we then index this peerlist with an integer, because we may have to
         # wrap. We update the goal as we go.
         i = 0
         # we then index this peerlist with an integer, because we may have to
         # wrap. We update the goal as we go.
         i = 0
@@ -465,7 +464,7 @@ class Publish:
             # build the block hash tree. SDMF has only one leaf.
             leaves = [hashutil.block_hash(share_data)]
             t = hashtree.HashTree(leaves)
             # build the block hash tree. SDMF has only one leaf.
             leaves = [hashutil.block_hash(share_data)]
             t = hashtree.HashTree(leaves)
-            block_hash_trees[shnum] = block_hash_tree = list(t)
+            block_hash_trees[shnum] = list(t)
             share_hash_leaves[shnum] = t[0]
         for leaf in share_hash_leaves:
             assert leaf is not None
             share_hash_leaves[shnum] = t[0]
         for leaf in share_hash_leaves:
             assert leaf is not None
index 5798a448fb32a052edb891a834e942e8c83628f7..f47deaef3838e5f5b361dbdf183e15a14ab7dff7 100644 (file)
@@ -500,7 +500,6 @@ class ServermapUpdater:
         self._status.set_status("Sending %d initial queries" % len(peerlist))
         self._queries_outstanding = set()
         self._sharemap = DictOfSets() # shnum -> [(peerid, seqnum, R)..]
         self._status.set_status("Sending %d initial queries" % len(peerlist))
         self._queries_outstanding = set()
         self._sharemap = DictOfSets() # shnum -> [(peerid, seqnum, R)..]
-        dl = []
         for (peerid, ss) in peerlist.items():
             self._queries_outstanding.add(peerid)
             self._do_query(ss, peerid, self._storage_index, self._read_size)
         for (peerid, ss) in peerlist.items():
             self._queries_outstanding.add(peerid)
             self._do_query(ss, peerid, self._storage_index, self._read_size)
index 2990689a6d9f453cca10058d72b1a01ab7fa6fa1..303bf1d0e74e326f3cc6406ed6ffaa51974adcba 100644 (file)
@@ -57,7 +57,6 @@ class ProvisioningTool(rend.Page):
             i_select = T.select(name=name)
             for (count, description) in options:
                 count = astype(count)
             i_select = T.select(name=name)
             for (count, description) in options:
                 count = astype(count)
-                selected = False
                 if ((current_value is not None and count == current_value) or
                     (current_value is None and count == default)):
                     o = T.option(value=str(count), selected="true")[description]
                 if ((current_value is not None and count == current_value) or
                     (current_value is None and count == default)):
                     o = T.option(value=str(count), selected="true")[description]
@@ -340,7 +339,6 @@ class ProvisioningTool(rend.Page):
             add_output("Users",
                        T.div["Average file size: ", number(file_size)])
             total_files = num_users * files_per_user / sharing_ratio
             add_output("Users",
                        T.div["Average file size: ", number(file_size)])
             total_files = num_users * files_per_user / sharing_ratio
-            user_file_check_interval = file_check_interval / files_per_user
 
             add_output("Grid",
                        T.div["Total number of files in grid: ",
 
             add_output("Grid",
                        T.div["Total number of files in grid: ",
@@ -711,6 +709,7 @@ class ProvisioningTool(rend.Page):
             from allmydata import reliability
             # we import this just to test to see if the page is available
             _hush_pyflakes = reliability
             from allmydata import reliability
             # we import this just to test to see if the page is available
             _hush_pyflakes = reliability
+            del _hush_pyflakes
             f = [T.div[T.a(href="../reliability")["Reliability Math"]], f]
         except ImportError:
             pass
             f = [T.div[T.a(href="../reliability")["Reliability Math"]], f]
         except ImportError:
             pass
index 76d930f356a67721f464edfba0a00a3e61596c95..a0d60769bdec9ca6d0020868854c647aabf44891 100644 (file)
@@ -79,7 +79,6 @@ class ReliabilityModel:
         #print "DIFF:", (old_post_repair - decay * repair)
 
         START = array([0]*N + [1])
         #print "DIFF:", (old_post_repair - decay * repair)
 
         START = array([0]*N + [1])
-        ALIVE = array([0]*k + [1]*(1+N-k))
         DEAD = array([1]*k + [0]*(1+N-k))
         REPAIRp = array([0]*k + [1]*(R-k) + [0]*(1+N-R))
         REPAIR_newshares = array([0]*k +
         DEAD = array([1]*k + [0]*(1+N-k))
         REPAIRp = array([0]*k + [1]*(R-k) + [0]*(1+N-R))
         REPAIR_newshares = array([0]*k +
@@ -87,7 +86,6 @@ class ReliabilityModel:
                                  [0]*(1+N-R))
         assert REPAIR_newshares.shape[0] == N+1
         #print "START", START
                                  [0]*(1+N-R))
         assert REPAIR_newshares.shape[0] == N+1
         #print "START", START
-        #print "ALIVE", ALIVE
         #print "REPAIRp", REPAIRp
         #print "REPAIR_newshares", REPAIR_newshares
 
         #print "REPAIRp", REPAIRp
         #print "REPAIR_newshares", REPAIR_newshares
 
index c85bfca305652cebb0cf4f745d1d2db96bd15ab1..c455866bd30e3c5b303c9b77800b45dcb1c93a82 100644 (file)
@@ -553,10 +553,7 @@ def describe_share(abs_sharefile, si_s, shnum_s, now, out):
         # mutable share
         m = MutableShareFile(abs_sharefile)
         WE, nodeid = m._read_write_enabler_and_nodeid(f)
         # mutable share
         m = MutableShareFile(abs_sharefile)
         WE, nodeid = m._read_write_enabler_and_nodeid(f)
-        num_extra_leases = m._read_num_extra_leases(f)
         data_length = m._read_data_length(f)
         data_length = m._read_data_length(f)
-        extra_lease_offset = m._read_extra_lease_offset(f)
-        container_size = extra_lease_offset - m.DATA_OFFSET
         expiration_time = min( [lease.expiration_time
                                 for (i,lease) in m._enumerate_leases(f)] )
         expiration = max(0, expiration_time - now)
         expiration_time = min( [lease.expiration_time
                                 for (i,lease) in m._enumerate_leases(f)] )
         expiration = max(0, expiration_time - now)
index 4d44e67f032812bc4b1653038e10eb813b94b27b..e799d4e8aac96cea7a534671e63c1150b23bb8c1 100644 (file)
@@ -27,7 +27,6 @@ g.setServiceParent(application)
 """
 
 def create_stats_gatherer(config):
 """
 
 def create_stats_gatherer(config):
-    out = config.stdout
     err = config.stderr
     basedir = config['basedir']
     if not basedir:
     err = config.stderr
     basedir = config['basedir']
     if not basedir:
index 4f7c70dda10202824d30d271d710d22adf4fe572..122018c982c404b69dd24235149311f16bf9ccbf 100644 (file)
@@ -72,14 +72,11 @@ class BackerUpper:
     def run(self):
         options = self.options
         nodeurl = options['node-url']
     def run(self):
         options = self.options
         nodeurl = options['node-url']
-        from_dir = options.from_dir
-        to_dir = options.to_dir
         self.verbosity = 1
         if options['quiet']:
             self.verbosity = 0
         if options['verbose']:
             self.verbosity = 2
         self.verbosity = 1
         if options['quiet']:
             self.verbosity = 0
         if options['verbose']:
             self.verbosity = 2
-        stdin = options.stdin
         stdout = options.stdout
         stderr = options.stderr
 
         stdout = options.stdout
         stderr = options.stderr
 
@@ -101,7 +98,6 @@ class BackerUpper:
             to_url += "/"
 
         archives_url = to_url + "Archives/"
             to_url += "/"
 
         archives_url = to_url + "Archives/"
-        latest_url = to_url + "Latest"
 
         # first step: make sure the target directory exists, as well as the
         # Archives/ subdirectory.
 
         # first step: make sure the target directory exists, as well as the
         # Archives/ subdirectory.
@@ -112,11 +108,6 @@ class BackerUpper:
                 print >>stderr, "Unable to create target directory: %s %s %s" % \
                       (resp.status, resp.reason, resp.read())
                 return 1
                 print >>stderr, "Unable to create target directory: %s %s %s" % \
                       (resp.status, resp.reason, resp.read())
                 return 1
-            archives_dir = {}
-        else:
-            jdata = simplejson.load(resp)
-            (otype, attrs) = jdata
-            archives_dir = attrs["children"]
 
         # second step: process the tree
         new_backup_dircap = self.process(options.from_dir)
 
         # second step: process the tree
         new_backup_dircap = self.process(options.from_dir)
index 8cf543da91bc6d8f0a4b1261f2c4c4dd9b1f78c0..02052641f5517296db45e3fd5891e115b137018c 100644 (file)
@@ -530,7 +530,6 @@ class Copier:
             url = self.nodeurl + "uri/%s" % urllib.quote(rootcap)
             if path:
                 url += "/" + escape_path(path)
             url = self.nodeurl + "uri/%s" % urllib.quote(rootcap)
             if path:
                 url += "/" + escape_path(path)
-                last_slash = path.rfind("/")
 
             resp = do_http("GET", url + "?t=json")
             if resp.status == 404:
 
             resp = do_http("GET", url + "?t=json")
             if resp.status == 404:
index 180bf7cb153011fd4342a9e1c8ee4f1e5fa9c5e5..5131aead90ac3fb3cbbb60bc5dcec45fefd8e7e0 100644 (file)
@@ -93,7 +93,6 @@ class StatsGrabber(SlowOperationRunner):
 
     def write_results(self, data):
         stdout = self.options.stdout
 
     def write_results(self, data):
         stdout = self.options.stdout
-        stderr = self.options.stderr
         keys = ("count-immutable-files",
                 "count-mutable-files",
                 "count-literal-files",
         keys = ("count-immutable-files",
                 "count-mutable-files",
                 "count-literal-files",
index a0ffe719617beb2930aba2d1e7527405a14545a0..3e854e97b7ebe836a63398a6d2d9b8daf8d2a694 100644 (file)
@@ -5,17 +5,11 @@ from allmydata.scripts.common import get_alias, DEFAULT_ALIAS, escape_path
 
 def rm(options):
     """
 
 def rm(options):
     """
-    @param verbosity: 0, 1, or 2, meaning quiet, verbose, or very verbose
-
     @return: a Deferred which eventually fires with the exit code
     """
     nodeurl = options['node-url']
     aliases = options.aliases
     where = options.where
     @return: a Deferred which eventually fires with the exit code
     """
     nodeurl = options['node-url']
     aliases = options.aliases
     where = options.where
-    if options['quiet']:
-        verbosity = 0
-    else:
-        verbosity = 2
     stdout = options.stdout
     stderr = options.stderr
 
     stdout = options.stdout
     stderr = options.stderr
 
index e3a50e42022a1fdedf8865026b98dd80fd7bc1c4..3a917b8d4db5ed9d9630a81a0fd392df5041b8ca 100644 (file)
@@ -258,7 +258,6 @@ class ShareCrawler(service.MultiService):
         self.current_sleep_time = None
         self.next_wake_time = None
         try:
         self.current_sleep_time = None
         self.next_wake_time = None
         try:
-            s = self.last_complete_prefix_index
             self.start_current_prefix(start_slice)
             finished_cycle = True
         except TimeSliceExceeded:
             self.start_current_prefix(start_slice)
             finished_cycle = True
         except TimeSliceExceeded:
index 2f71c71ec74112263985139128f90785c700cc86..22ec1621ca6d582a1b135d8bf00b435eb7d33da2 100644 (file)
@@ -76,7 +76,7 @@ class StorageServer(service.MultiService, Referenceable):
         self._clean_incomplete()
         fileutil.make_dirs(self.incomingdir)
         self._active_writers = weakref.WeakKeyDictionary()
         self._clean_incomplete()
         fileutil.make_dirs(self.incomingdir)
         self._active_writers = weakref.WeakKeyDictionary()
-        lp = log.msg("StorageServer created", facility="tahoe.storage")
+        log.msg("StorageServer created", facility="tahoe.storage")
 
         if reserved_space:
             if self.get_available_space() is None:
 
         if reserved_space:
             if self.get_available_space() is None:
@@ -479,7 +479,7 @@ class StorageServer(service.MultiService, Referenceable):
         start = time.time()
         self.count("writev")
         si_s = si_b2a(storage_index)
         start = time.time()
         self.count("writev")
         si_s = si_b2a(storage_index)
-        lp = log.msg("storage: slot_writev %s" % si_s)
+        log.msg("storage: slot_writev %s" % si_s)
         si_dir = storage_index_to_dir(storage_index)
         (write_enabler, renew_secret, cancel_secret) = secrets
         # shares exist if there is a file for them
         si_dir = storage_index_to_dir(storage_index)
         (write_enabler, renew_secret, cancel_secret) = secrets
         # shares exist if there is a file for them
index 84c21d47a1c5dbf8ef0f21d5ced851460e4617f4..8be4ca07f32228e4d4b1979b1cbba85475e594c8 100644 (file)
@@ -95,7 +95,6 @@ class SpeedTest:
 
     def do_test(self):
         print "doing test"
 
     def do_test(self):
         print "doing test"
-        rr = self.client_rref
         d = defer.succeed(None)
         d.addCallback(self.one_test, "startup", 1, 1000, False) #ignore this one
         d.addCallback(self.measure_rtt)
         d = defer.succeed(None)
         d.addCallback(self.one_test, "startup", 1, 1000, False) #ignore this one
         d.addCallback(self.measure_rtt)
index 45e6fcd3dfb4bf5df757432175ea8c67cf59851f..538f6627a72f81662c6dacb78df0c40eca596e4f 100644 (file)
@@ -1335,18 +1335,11 @@ def _corrupt_crypttext_hash_tree(data, debug=False):
     return corrupt_field(data, 0x0c+crypttexthashtreeoffset, blockhashesoffset-crypttexthashtreeoffset, debug=debug)
 
 def _corrupt_crypttext_hash_tree_byte_x221(data, debug=False):
     return corrupt_field(data, 0x0c+crypttexthashtreeoffset, blockhashesoffset-crypttexthashtreeoffset, debug=debug)
 
 def _corrupt_crypttext_hash_tree_byte_x221(data, debug=False):
-    """Scramble the file data -- the field containing the crypttext hash tree
-    will have the 7th bit of the 9th byte flipped.
+    """Scramble the file data -- the byte at offset 0x221 will have its 7th
+    (b1) bit flipped.
     """
     sharevernum = struct.unpack(">L", data[0x0c:0x0c+4])[0]
     assert sharevernum in (1, 2), "This test is designed to corrupt immutable shares of v1 or v2 in specific ways."
     """
     sharevernum = struct.unpack(">L", data[0x0c:0x0c+4])[0]
     assert sharevernum in (1, 2), "This test is designed to corrupt immutable shares of v1 or v2 in specific ways."
-    if sharevernum == 1:
-        crypttexthashtreeoffset = struct.unpack(">L", data[0x0c+0x14:0x0c+0x14+4])[0]
-        blockhashesoffset = struct.unpack(">L", data[0x0c+0x18:0x0c+0x18+4])[0]
-    else:
-        crypttexthashtreeoffset = struct.unpack(">Q", data[0x0c+0x24:0x0c+0x24+8])[0]
-        blockhashesoffset = struct.unpack(">Q", data[0x0c+0x2c:0x0c+0x2c+8])[0]
-
     if debug:
         log.msg("original data: %r" % (data,))
     return data[:0x0c+0x221] + chr(ord(data[0x0c+0x221])^0x02) + data[0x0c+0x2210+1:]
     if debug:
         log.msg("original data: %r" % (data,))
     return data[:0x0c+0x221] + chr(ord(data[0x0c+0x221])^0x02) + data[0x0c+0x2210+1:]
index 9d8f4756124dc923e4011a30c2ded40a8d2e8efa..beb6abe895d87a25f28121648ed3aa75c11232b4 100644 (file)
@@ -32,7 +32,7 @@ class Basic(unittest.TestCase):
         basedir = "test_client.Basic.test_loadable"
         os.mkdir(basedir)
         open(os.path.join(basedir, "introducer.furl"), "w").write("")
         basedir = "test_client.Basic.test_loadable"
         os.mkdir(basedir)
         open(os.path.join(basedir, "introducer.furl"), "w").write("")
-        c = client.Client(basedir)
+        client.Client(basedir)
 
     def test_loadable_old_config_bits(self):
         basedir = "test_client.Basic.test_loadable_old_config_bits"
 
     def test_loadable_old_config_bits(self):
         basedir = "test_client.Basic.test_loadable_old_config_bits"
@@ -190,7 +190,7 @@ class Run(unittest.TestCase, testutil.StallMixin):
         dummy = "pb://wl74cyahejagspqgy4x5ukrvfnevlknt@127.0.0.1:58889/bogus"
         open(os.path.join(basedir, "introducer.furl"), "w").write(dummy)
         open(os.path.join(basedir, "suicide_prevention_hotline"), "w")
         dummy = "pb://wl74cyahejagspqgy4x5ukrvfnevlknt@127.0.0.1:58889/bogus"
         open(os.path.join(basedir, "introducer.furl"), "w").write(dummy)
         open(os.path.join(basedir, "suicide_prevention_hotline"), "w")
-        c = client.Client(basedir)
+        client.Client(basedir)
 
     def test_reloadable(self):
         basedir = "test_client.Run.test_reloadable"
 
     def test_reloadable(self):
         basedir = "test_client.Run.test_reloadable"
index 4caa29064b06c430f68bc49c25de010ede5cc659..1c57e134a0be8b3b7cbe85f3121393f3f6c50e4e 100644 (file)
@@ -341,7 +341,8 @@ class Basic(unittest.TestCase, StallMixin, pollmixin.PollMixin):
         ss = StorageServer(self.basedir, serverid)
         ss.setServiceParent(self.s)
 
         ss = StorageServer(self.basedir, serverid)
         ss.setServiceParent(self.s)
 
-        sis = [self.write(i, ss, serverid) for i in range(10)]
+        for i in range(10):
+            self.write(i, ss, serverid)
 
         statefile = os.path.join(self.basedir, "statefile")
         c = ConsumingCrawler(ss, statefile)
 
         statefile = os.path.join(self.basedir, "statefile")
         c = ConsumingCrawler(ss, statefile)
@@ -385,7 +386,8 @@ class Basic(unittest.TestCase, StallMixin, pollmixin.PollMixin):
         ss = StorageServer(self.basedir, serverid)
         ss.setServiceParent(self.s)
 
         ss = StorageServer(self.basedir, serverid)
         ss.setServiceParent(self.s)
 
-        sis = [self.write(i, ss, serverid) for i in range(10)]
+        for i in range(10):
+            self.write(i, ss, serverid)
 
         statefile = os.path.join(self.basedir, "statefile")
         c = ShareCrawler(ss, statefile)
 
         statefile = os.path.join(self.basedir, "statefile")
         c = ShareCrawler(ss, statefile)
@@ -412,7 +414,8 @@ class Basic(unittest.TestCase, StallMixin, pollmixin.PollMixin):
         ss = StorageServer(self.basedir, serverid)
         ss.setServiceParent(self.s)
 
         ss = StorageServer(self.basedir, serverid)
         ss.setServiceParent(self.s)
 
-        sis = [self.write(i, ss, serverid) for i in range(30)]
+        for i in range(30):
+            self.write(i, ss, serverid)
 
         statefile = os.path.join(self.basedir, "statefile")
         c = OneShotCrawler(ss, statefile)
 
         statefile = os.path.join(self.basedir, "statefile")
         c = OneShotCrawler(ss, statefile)
index 165013c586f8a4f2249fb0a476d52deb49a08fb1..55e85680d07fe3f50bc46e5433795931aca25914 100644 (file)
@@ -38,7 +38,6 @@ class MutableChecker(GridTestMixin, unittest.TestCase, ErrorMixin):
         def _created(node):
             self.node = node
             self.fileurl = "uri/" + urllib.quote(node.get_uri())
         def _created(node):
             self.node = node
             self.fileurl = "uri/" + urllib.quote(node.get_uri())
-            si = self.node.get_storage_index()
         d.addCallback(_created)
         # now make sure the webapi verifier sees no problems
         d.addCallback(lambda ign: self.GET(self.fileurl+"?t=check&verify=true",
         d.addCallback(_created)
         # now make sure the webapi verifier sees no problems
         d.addCallback(lambda ign: self.GET(self.fileurl+"?t=check&verify=true",
@@ -395,13 +394,14 @@ class DeepCheckWebGood(DeepCheckBase, unittest.TestCase):
         self.check_stats_good(stats)
 
     def do_web_stream_check(self, ignored):
         self.check_stats_good(stats)
 
     def do_web_stream_check(self, ignored):
+        # TODO
         return
         d = self.web(self.root, t="stream-deep-check")
         def _check(res):
             units = list(self.parse_streamed_json(res))
         return
         d = self.web(self.root, t="stream-deep-check")
         def _check(res):
             units = list(self.parse_streamed_json(res))
-            files = [u for u in units if u["type"] in ("file", "directory")]
+            #files = [u for u in units if u["type"] in ("file", "directory")]
             assert units[-1]["type"] == "stats"
             assert units[-1]["type"] == "stats"
-            stats = units[-1]["stats"]
+            #stats = units[-1]["stats"]
             # ...
         d.addCallback(_check)
         return d
             # ...
         d.addCallback(_check)
         return d
index fd591bedae5d48b5a402393191f6df121a124f1a..a45b37f30f333a8ce51af909c70d91119a74d898 100644 (file)
@@ -294,7 +294,6 @@ class Dirnode(GridTestMixin, unittest.TestCase,
         return d
 
     def _mark_file_bad(self, rootnode):
         return d
 
     def _mark_file_bad(self, rootnode):
-        si = rootnode.get_storage_index()
         self.delete_shares_numbered(rootnode.get_uri(), [0])
         return rootnode
 
         self.delete_shares_numbered(rootnode.get_uri(), [0])
         return rootnode
 
@@ -976,25 +975,25 @@ class Packing(unittest.TestCase):
         self.failUnlessIn("lit", packed)
 
         kids = self._make_kids(nm, ["imm", "lit", "write"])
         self.failUnlessIn("lit", packed)
 
         kids = self._make_kids(nm, ["imm", "lit", "write"])
-        e = self.failUnlessRaises(dirnode.MustBeDeepImmutable,
-                                  dirnode.pack_children,
-                                  fn, kids, deep_immutable=True)
+        self.failUnlessRaises(dirnode.MustBeDeepImmutable,
+                              dirnode.pack_children,
+                              fn, kids, deep_immutable=True)
 
         # read-only is not enough: all children must be immutable
         kids = self._make_kids(nm, ["imm", "lit", "read"])
 
         # read-only is not enough: all children must be immutable
         kids = self._make_kids(nm, ["imm", "lit", "read"])
-        e = self.failUnlessRaises(dirnode.MustBeDeepImmutable,
-                                  dirnode.pack_children,
-                                  fn, kids, deep_immutable=True)
+        self.failUnlessRaises(dirnode.MustBeDeepImmutable,
+                              dirnode.pack_children,
+                              fn, kids, deep_immutable=True)
 
         kids = self._make_kids(nm, ["imm", "lit", "dirwrite"])
 
         kids = self._make_kids(nm, ["imm", "lit", "dirwrite"])
-        e = self.failUnlessRaises(dirnode.MustBeDeepImmutable,
-                                  dirnode.pack_children,
-                                  fn, kids, deep_immutable=True)
+        self.failUnlessRaises(dirnode.MustBeDeepImmutable,
+                              dirnode.pack_children,
+                              fn, kids, deep_immutable=True)
 
         kids = self._make_kids(nm, ["imm", "lit", "dirread"])
 
         kids = self._make_kids(nm, ["imm", "lit", "dirread"])
-        e = self.failUnlessRaises(dirnode.MustBeDeepImmutable,
-                                  dirnode.pack_children,
-                                  fn, kids, deep_immutable=True)
+        self.failUnlessRaises(dirnode.MustBeDeepImmutable,
+                              dirnode.pack_children,
+                              fn, kids, deep_immutable=True)
 
 class FakeMutableFile:
     implements(IMutableFileNode)
 
 class FakeMutableFile:
     implements(IMutableFileNode)
index 2d91dbc581b9aace89a0e00b4b6f8a78b072a37e..82e9056b94113f671d9bd645da61077d6ffe1723 100644 (file)
@@ -457,7 +457,6 @@ class Roundtrip(unittest.TestCase, testutil.ShouldFailMixin):
         def _ready(res):
             k,happy,n = e.get_param("share_counts")
             assert n == NUM_SHARES # else we'll be completely confused
         def _ready(res):
             k,happy,n = e.get_param("share_counts")
             assert n == NUM_SHARES # else we'll be completely confused
-            all_peers = []
             for shnum in range(NUM_SHARES):
                 mode = bucket_modes.get(shnum, "good")
                 peer = FakeBucketReaderWriterProxy(mode)
             for shnum in range(NUM_SHARES):
                 mode = bucket_modes.get(shnum, "good")
                 peer = FakeBucketReaderWriterProxy(mode)
index a8de4207bf90c27e63515b4f7456bb133f2541c8..a8ea9969effa9cced13f465062081cb9cb451087 100644 (file)
@@ -30,7 +30,6 @@ class Node(unittest.TestCase):
                            needed_shares=3,
                            total_shares=10,
                            size=1000)
                            needed_shares=3,
                            total_shares=10,
                            size=1000)
-        c = FakeClient()
         cf = cachedir.CacheFile("none")
         fn1 = ImmutableFileNode(u, None, None, None, None, cf)
         fn2 = ImmutableFileNode(u, None, None, None, None, cf)
         cf = cachedir.CacheFile("none")
         fn1 = ImmutableFileNode(u, None, None, None, None, cf)
         fn2 = ImmutableFileNode(u, None, None, None, None, cf)
@@ -55,7 +54,6 @@ class Node(unittest.TestCase):
     def test_literal_filenode(self):
         DATA = "I am a short file."
         u = uri.LiteralFileURI(data=DATA)
     def test_literal_filenode(self):
         DATA = "I am a short file."
         u = uri.LiteralFileURI(data=DATA)
-        c = None
         fn1 = LiteralFileNode(u)
         fn2 = LiteralFileNode(u)
         self.failUnlessEqual(fn1, fn2)
         fn1 = LiteralFileNode(u)
         fn2 = LiteralFileNode(u)
         self.failUnlessEqual(fn1, fn2)
@@ -91,7 +89,6 @@ class Node(unittest.TestCase):
     def test_mutable_filenode(self):
         client = FakeClient()
         wk = "\x00"*16
     def test_mutable_filenode(self):
         client = FakeClient()
         wk = "\x00"*16
-        fp = "\x00"*32
         rk = hashutil.ssk_readkey_hash(wk)
         si = hashutil.ssk_storage_index_hash(rk)
 
         rk = hashutil.ssk_readkey_hash(wk)
         si = hashutil.ssk_storage_index_hash(rk)
 
index 30d99194b5b8b8d0dbcbfc3a27818b3e415cd666..abcf4eb1b4e73a19f1af9eb8871981a7cba0a318 100644 (file)
@@ -81,7 +81,7 @@ class Incomplete(unittest.TestCase):
         self.failUnlessEqual(ht.needed_hashes(5, True), set([12, 11, 6, 1]))
 
     def test_depth_of(self):
         self.failUnlessEqual(ht.needed_hashes(5, True), set([12, 11, 6, 1]))
 
     def test_depth_of(self):
-        ht = hashtree.IncompleteHashTree(8)
+        hashtree.IncompleteHashTree(8)
         self.failUnlessEqual(hashtree.depth_of(0), 0)
         for i in [1,2]:
             self.failUnlessEqual(hashtree.depth_of(i), 1, "i=%d"%i)
         self.failUnlessEqual(hashtree.depth_of(0), 0)
         for i in [1,2]:
             self.failUnlessEqual(hashtree.depth_of(i), 1, "i=%d"%i)
index c9dfc064f97fdd27230f88ac0d913594a648ecd0..418956c44bd0499a26a1a29ffbb535fa000707b8 100644 (file)
@@ -78,7 +78,7 @@ class Test(common.ShareManglingMixin, unittest.TestCase):
             random.shuffle(shnums)
             for i in shnums[:7]:
                 self._corrupt_a_share(None, common._corrupt_offset_of_block_hashes_to_truncate_crypttext_hashes, i)
             random.shuffle(shnums)
             for i in shnums[:7]:
                 self._corrupt_a_share(None, common._corrupt_offset_of_block_hashes_to_truncate_crypttext_hashes, i)
-        before_download_reads = self._count_reads()
+        #before_download_reads = self._count_reads()
         d.addCallback(_then_corrupt_7)
         d.addCallback(self._download_and_check_plaintext)
         return d
         d.addCallback(_then_corrupt_7)
         d.addCallback(self._download_and_check_plaintext)
         return d
index 7958158df24ef785a985fd02e76561446b58f6be..9da6dccef45531c8d33c78581ce976fc65e1fde7 100644 (file)
@@ -129,7 +129,7 @@ class SystemTest(SystemTestMixin, unittest.TestCase):
             log.msg("creating client %d: %s" % (i, tub.getShortTubID()))
             c = IntroducerClient(tub, self.introducer_furl, u"nickname-%d" % i,
                                  "version", "oldest")
             log.msg("creating client %d: %s" % (i, tub.getShortTubID()))
             c = IntroducerClient(tub, self.introducer_furl, u"nickname-%d" % i,
                                  "version", "oldest")
-            received_announcements[c] = ra = {}
+            received_announcements[c] = {}
             def got(serverid, ann_d, announcements):
                 announcements[serverid] = ann_d
             c.subscribe_to("storage", got, received_announcements[c])
             def got(serverid, ann_d, announcements):
                 announcements[serverid] = ann_d
             c.subscribe_to("storage", got, received_announcements[c])
index a79197afdf10b289110cc06281a40a62d9ed5bb3..8afec13fde27b45eb7cb875650fb5ff5af8cf8f8 100644 (file)
@@ -6,7 +6,6 @@ from twisted.internet import defer, reactor
 from allmydata import uri, client
 from allmydata.nodemaker import NodeMaker
 from allmydata.util import base32
 from allmydata import uri, client
 from allmydata.nodemaker import NodeMaker
 from allmydata.util import base32
-from allmydata.util.idlib import shortnodeid_b2a
 from allmydata.util.hashutil import tagged_hash, ssk_writekey_hash, \
      ssk_pubkey_fingerprint_hash
 from allmydata.interfaces import IRepairResults, ICheckAndRepairResults, \
 from allmydata.util.hashutil import tagged_hash, ssk_writekey_hash, \
      ssk_pubkey_fingerprint_hash
 from allmydata.interfaces import IRepairResults, ICheckAndRepairResults, \
@@ -71,7 +70,6 @@ class FakeStorage:
         self._pending_timer = None
         pending = self._pending
         self._pending = {}
         self._pending_timer = None
         pending = self._pending
         self._pending = {}
-        extra = []
         for peerid in self._sequence:
             if peerid in pending:
                 d, shares = pending.pop(peerid)
         for peerid in self._sequence:
             if peerid in pending:
                 d, shares = pending.pop(peerid)
@@ -602,7 +600,6 @@ class PublishMixin:
         # publish a file and create shares, which can then be manipulated
         # later.
         self.CONTENTS = "New contents go here" * 1000
         # publish a file and create shares, which can then be manipulated
         # later.
         self.CONTENTS = "New contents go here" * 1000
-        num_peers = 20
         self._storage = FakeStorage()
         self._nodemaker = make_nodemaker(self._storage)
         self._storage_broker = self._nodemaker.storage_broker
         self._storage = FakeStorage()
         self._nodemaker = make_nodemaker(self._storage)
         self._storage_broker = self._nodemaker.storage_broker
@@ -620,7 +617,6 @@ class PublishMixin:
                          "Contents 3a",
                          "Contents 3b"]
         self._copied_shares = {}
                          "Contents 3a",
                          "Contents 3b"]
         self._copied_shares = {}
-        num_peers = 20
         self._storage = FakeStorage()
         self._nodemaker = make_nodemaker(self._storage)
         d = self._nodemaker.create_mutable_file(self.CONTENTS[0]) # seqnum=1
         self._storage = FakeStorage()
         self._nodemaker = make_nodemaker(self._storage)
         d = self._nodemaker.create_mutable_file(self.CONTENTS[0]) # seqnum=1
@@ -757,7 +753,6 @@ class Servermap(unittest.TestCase, PublishMixin):
     def test_mark_bad(self):
         d = defer.succeed(None)
         ms = self.make_servermap
     def test_mark_bad(self):
         d = defer.succeed(None)
         ms = self.make_servermap
-        us = self.update_servermap
 
         d.addCallback(lambda res: ms(mode=MODE_READ))
         d.addCallback(lambda sm: self.failUnlessOneRecoverable(sm, 6))
 
         d.addCallback(lambda res: ms(mode=MODE_READ))
         d.addCallback(lambda sm: self.failUnlessOneRecoverable(sm, 6))
@@ -1435,7 +1430,6 @@ class Repair(unittest.TestCase, PublishMixin, ShouldFailMixin):
             self.failUnlessEqual(len(smap.recoverable_versions()), 1)
             self.failIf(smap.unrecoverable_versions())
             # now, which should have won?
             self.failUnlessEqual(len(smap.recoverable_versions()), 1)
             self.failIf(smap.unrecoverable_versions())
             # now, which should have won?
-            roothash_s4a = self.get_roothash_for(3)
             expected_contents = self.CONTENTS[3]
             new_versionid = smap.best_recoverable_version()
             self.failUnlessEqual(new_versionid[0], 5) # seqnum 5
             expected_contents = self.CONTENTS[3]
             new_versionid = smap.best_recoverable_version()
             self.failUnlessEqual(new_versionid[0], 5) # seqnum 5
@@ -1586,7 +1580,6 @@ class MultipleEncodings(unittest.TestCase):
             sb = self._storage_broker
 
             for peerid in sorted(sb.get_all_serverids()):
             sb = self._storage_broker
 
             for peerid in sorted(sb.get_all_serverids()):
-                peerid_s = shortnodeid_b2a(peerid)
                 for shnum in self._shares1.get(peerid, {}):
                     if shnum < len(places):
                         which = places[shnum]
                 for shnum in self._shares1.get(peerid, {}):
                     if shnum < len(places):
                         which = places[shnum]
@@ -1596,7 +1589,6 @@ class MultipleEncodings(unittest.TestCase):
                     in_1 = shnum in self._shares1[peerid]
                     in_2 = shnum in self._shares2.get(peerid, {})
                     in_3 = shnum in self._shares3.get(peerid, {})
                     in_1 = shnum in self._shares1[peerid]
                     in_2 = shnum in self._shares2.get(peerid, {})
                     in_3 = shnum in self._shares3.get(peerid, {})
-                    #print peerid_s, shnum, which, in_1, in_2, in_3
                     if which == 1:
                         if in_1:
                             peers[shnum] = self._shares1[peerid][shnum]
                     if which == 1:
                         if in_1:
                             peers[shnum] = self._shares1[peerid][shnum]
index 7538b3908f56c7f2b80400aee04a1ec6041b425a..a1a9f7ab62af4a2a8ea2c0fad9b4f685ab0e0ecd 100644 (file)
@@ -253,7 +253,7 @@ class Server(unittest.TestCase):
         return ss
 
     def test_create(self):
         return ss
 
     def test_create(self):
-        ss = self.create("test_create")
+        self.create("test_create")
 
     def allocate(self, ss, storage_index, sharenums, size, canary=None):
         renew_secret = hashutil.tagged_hash("blah", "%d" % self._lease_secret.next())
 
     def allocate(self, ss, storage_index, sharenums, size, canary=None):
         renew_secret = hashutil.tagged_hash("blah", "%d" % self._lease_secret.next())
@@ -326,7 +326,6 @@ class Server(unittest.TestCase):
 
         self.failUnlessEqual(ss.remote_get_buckets("allocate"), {})
 
 
         self.failUnlessEqual(ss.remote_get_buckets("allocate"), {})
 
-        canary = FakeCanary()
         already,writers = self.allocate(ss, "allocate", [0,1,2], 75)
         self.failUnlessEqual(already, set())
         self.failUnlessEqual(set(writers.keys()), set([0,1,2]))
         already,writers = self.allocate(ss, "allocate", [0,1,2], 75)
         self.failUnlessEqual(already, set())
         self.failUnlessEqual(set(writers.keys()), set([0,1,2]))
@@ -387,7 +386,7 @@ class Server(unittest.TestCase):
         f.write(struct.pack(">L", 0)) # this is invalid: minimum used is v1
         f.close()
 
         f.write(struct.pack(">L", 0)) # this is invalid: minimum used is v1
         f.close()
 
-        b = ss.remote_get_buckets("allocate")
+        ss.remote_get_buckets("allocate")
 
         e = self.failUnlessRaises(UnknownImmutableContainerVersionError,
                                   ss.remote_get_buckets, "si1")
 
         e = self.failUnlessRaises(UnknownImmutableContainerVersionError,
                                   ss.remote_get_buckets, "si1")
@@ -654,7 +653,6 @@ class Server(unittest.TestCase):
         ss = StorageServer(workdir, "\x00" * 20, discard_storage=True)
         ss.setServiceParent(self.sparent)
 
         ss = StorageServer(workdir, "\x00" * 20, discard_storage=True)
         ss.setServiceParent(self.sparent)
 
-        canary = FakeCanary()
         already,writers = self.allocate(ss, "vid", [0,1,2], 75)
         self.failUnlessEqual(already, set())
         self.failUnlessEqual(set(writers.keys()), set([0,1,2]))
         already,writers = self.allocate(ss, "vid", [0,1,2], 75)
         self.failUnlessEqual(already, set())
         self.failUnlessEqual(set(writers.keys()), set([0,1,2]))
@@ -733,7 +731,7 @@ class MutableServer(unittest.TestCase):
         return ss
 
     def test_create(self):
         return ss
 
     def test_create(self):
-        ss = self.create("test_create")
+        self.create("test_create")
 
     def write_enabler(self, we_tag):
         return hashutil.tagged_hash("we_blah", we_tag)
 
     def write_enabler(self, we_tag):
         return hashutil.tagged_hash("we_blah", we_tag)
index fc2d09128ea37a87df307ad521c1bd97108a07f0..6b439a083bb36936b578806337539185036f61e1 100644 (file)
@@ -465,7 +465,6 @@ class SystemTest(SystemTestMixin, unittest.TestCase):
             def _done(res):
                 log.msg("DONE: %s" % (res,))
                 self._mutable_node_1 = res
             def _done(res):
                 log.msg("DONE: %s" % (res,))
                 self._mutable_node_1 = res
-                uri = res.get_uri()
             d1.addCallback(_done)
             return d1
         d.addCallback(_create_mutable)
             d1.addCallback(_done)
             return d1
         d.addCallback(_create_mutable)
@@ -1364,13 +1363,11 @@ class SystemTest(SystemTestMixin, unittest.TestCase):
         # network calls)
 
         private_uri = self._private_node.get_uri()
         # network calls)
 
         private_uri = self._private_node.get_uri()
-        some_uri = self._root_directory_uri
         client0_basedir = self.getdir("client0")
 
         nodeargs = [
             "--node-directory", client0_basedir,
             ]
         client0_basedir = self.getdir("client0")
 
         nodeargs = [
             "--node-directory", client0_basedir,
             ]
-        TESTDATA = "I will not write the same thing over and over.\n" * 100
 
         d = defer.succeed(None)
 
 
         d = defer.succeed(None)
 
index 6b4c3e533cbe6821bc6b146955f0d8c5f2599029..41d5927f93c94f4a90e6ce59c395574c4f0e3802 100644 (file)
@@ -380,7 +380,6 @@ class Dirnode(unittest.TestCase):
         u0 = uri.LiteralFileURI("data")
         u1 = uri.LiteralDirectoryURI(u0)
         self.failUnless(str(u1))
         u0 = uri.LiteralFileURI("data")
         u1 = uri.LiteralDirectoryURI(u0)
         self.failUnless(str(u1))
-        u1s = u1.to_string()
         self.failUnlessEqual(u1.to_string(), "URI:DIR2-LIT:mrqxiyi")
         self.failUnless(u1.is_readonly())
         self.failIf(u1.is_mutable())
         self.failUnlessEqual(u1.to_string(), "URI:DIR2-LIT:mrqxiyi")
         self.failUnless(u1.is_readonly())
         self.failIf(u1.is_mutable())
index 6607d458a2a1fd54f4644f17db669b41e9008abd..f66d55f637d2bc898e8c8ef367b1f89b7132e92c 100644 (file)
@@ -176,7 +176,7 @@ class Statistics(unittest.TestCase):
         try:
             func(*args, **kwargs)
             self.fail(msg)
         try:
             func(*args, **kwargs)
             self.fail(msg)
-        except AssertionError, e:
+        except AssertionError:
             pass
 
     def failUnlessListEqual(self, a, b, msg = None):
             pass
 
     def failUnlessListEqual(self, a, b, msg = None):
@@ -320,9 +320,6 @@ class Asserts(unittest.TestCase):
         self.fail("assert was not caught")
 
     def should_not_assert(self, func, *args, **kwargs):
         self.fail("assert was not caught")
 
     def should_not_assert(self, func, *args, **kwargs):
-        if "re" in kwargs:
-            regexp = kwargs["re"]
-            del kwargs["re"]
         try:
             func(*args, **kwargs)
         except AssertionError, e:
         try:
             func(*args, **kwargs)
         except AssertionError, e:
index 529afe76f948ae50421b4c5661914094057bf0d0..32765df68cb272b85a91ca09d7af23b46067f4ef 100644 (file)
@@ -438,6 +438,7 @@ class Web(WebMixin, WebErrorMixin, testutil.StallMixin, unittest.TestCase):
         try:
             from allmydata import reliability
             _hush_pyflakes = reliability
         try:
             from allmydata import reliability
             _hush_pyflakes = reliability
+            del _hush_pyflakes
         except:
             raise unittest.SkipTest("reliability tool requires NumPy")
 
         except:
             raise unittest.SkipTest("reliability tool requires NumPy")
 
@@ -483,7 +484,7 @@ class Web(WebMixin, WebErrorMixin, testutil.StallMixin, unittest.TestCase):
         def _check_json(res):
             data = simplejson.loads(res)
             self.failUnless(isinstance(data, dict))
         def _check_json(res):
             data = simplejson.loads(res)
             self.failUnless(isinstance(data, dict))
-            active = data["active"]
+            #active = data["active"]
             # TODO: test more. We need a way to fake an active operation
             # here.
         d.addCallback(_check_json)
             # TODO: test more. We need a way to fake an active operation
             # here.
         d.addCallback(_check_json)
index 7c0bc10f5c026d79691f8babe55555a16d3999ed..ec8977297f94702ae4aea70885f29feb80adcba3 100644 (file)
@@ -529,7 +529,7 @@ def from_string(s):
 
 def is_uri(s):
     try:
 
 def is_uri(s):
     try:
-        uri = from_string(s)
+        from_string(s)
         return True
     except (TypeError, AssertionError):
         return False
         return True
     except (TypeError, AssertionError):
         return False
index a943cacc6c2620dd642cfca572158fb3a3da0ca5..00424af52f548c803c94234327321c90215df46d 100644 (file)
@@ -247,7 +247,6 @@ def a2b_l(cs, lengthinbits):
     octets = []
     pos = 2048
     num = qs[0] * pos
     octets = []
     pos = 2048
     num = qs[0] * pos
-    readybits = 5
     i = 1
     while len(octets) < numoctets:
         while pos > 256:
     i = 1
     while len(octets) < numoctets:
         while pos > 256:
index ab7ebab852ef9a5f40169764e23c2be6212ad28e..bbdb40b08492a76269ea07f7ebe924c9e1cc63f6 100644 (file)
@@ -33,7 +33,6 @@ def next_power_of_k(n, k):
         x = 0
     else:
         x = int(math.log(n, k) + 0.5)
         x = 0
     else:
         x = int(math.log(n, k) + 0.5)
-    r = k**x
     if k**x < n:
         return k**(x+1)
     else:
     if k**x < n:
         return k**(x+1)
     else:
index 0f12f66e44356f7a94b547c0e9f454c3041a678e..96ce17bf4490dad7144633410a7aa5ac2cefe04b 100644 (file)
@@ -177,7 +177,6 @@ class ResultsBase:
 
     def _render_si_link(self, ctx, storage_index):
         si_s = base32.b2a(storage_index)
 
     def _render_si_link(self, ctx, storage_index):
         si_s = base32.b2a(storage_index)
-        root = get_root(ctx)
         req = inevow.IRequest(ctx)
         ophandle = req.prepath[-1]
         target = "%s/operations/%s/%s" % (get_root(ctx), ophandle, si_s)
         req = inevow.IRequest(ctx)
         ophandle = req.prepath[-1]
         target = "%s/operations/%s/%s" % (get_root(ctx), ophandle, si_s)
index d38c6457e85e41134ace8e041a414e5d80625b13..6b4faa0305ec72a997fd8caba2aa5fbf76519c6b 100644 (file)
@@ -60,7 +60,6 @@ class DirectoryNodeHandler(RenderMixin, rend.Page, ReplaceMeMixin):
         self.name = name
 
     def childFactory(self, ctx, name):
         self.name = name
 
     def childFactory(self, ctx, name):
-        req = IRequest(ctx)
         name = name.decode("utf-8")
         if not name:
             raise EmptyPathnameComponentError()
         name = name.decode("utf-8")
         if not name:
             raise EmptyPathnameComponentError()
index 9fd44022a61404a2ac843140a2a0bb313676623b..a2a1a6a926b3e6b8f583b5cbb79821aee9f9ed8c 100644 (file)
@@ -355,7 +355,6 @@ class FileDownloader(rend.Page):
         if encoding:
             req.setHeader("content-encoding", encoding)
 
         if encoding:
             req.setHeader("content-encoding", encoding)
 
-        save_to_filename = None
         if boolean_of_arg(get_arg(req, "save", "False")):
             # tell the browser to save the file rather display it we don't
             # try to encode the filename, instead we echo back the exact same
         if boolean_of_arg(get_arg(req, "save", "False")):
             # tell the browser to save the file rather display it we don't
             # try to encode the filename, instead we echo back the exact same
index 6e95452117bc019e2b2032500c54703d3285a276..c00465f39cfe686e614fbba04f4e80d20554b30d 100644 (file)
@@ -53,7 +53,6 @@ class MoreInfo(rend.Page):
 
     def render_size(self, ctx, data):
         node = self.original
 
     def render_size(self, ctx, data):
         node = self.original
-        si = node.get_storage_index()
         d = node.get_current_size()
         def _no_size(size):
             if size is None:
         d = node.get_current_size()
         def _no_size(size):
             if size is None:
index 5c8683ce0903c563458be2d721389c7da522b194..d5d340615a4c33ebf5f86a586cef220fc924c443 100644 (file)
@@ -1,5 +1,5 @@
 
 
-from nevow import rend, inevow, tags as T
+from nevow import rend, tags as T
 reliability = None # might not be usable
 try:
     from allmydata import reliability # requires NumPy
 reliability = None # might not be usable
 try:
     from allmydata import reliability # requires NumPy
@@ -58,7 +58,6 @@ class ReliabilityTool(rend.Page):
         return "%d" % s
 
     def get_parameters(self, ctx):
         return "%d" % s
 
     def get_parameters(self, ctx):
-        req = inevow.IRequest(ctx)
         parameters = {}
         for (name,default,argtype,description) in self.DEFAULT_PARAMETERS:
             v = get_arg(ctx, name, default)
         parameters = {}
         for (name,default,argtype,description) in self.DEFAULT_PARAMETERS:
             v = get_arg(ctx, name, default)
index 29f2c3a9630a484b27b213a65da26ce1ce749ead..1943d00f28c5f6a75f6f33747e9a3a8746f63a71 100644 (file)
@@ -184,7 +184,6 @@ class StorageStatus(rend.Page):
         ecr = ec["space-recovered"]
 
         p = T.ul()
         ecr = ec["space-recovered"]
 
         p = T.ul()
-        pieces = []
         def add(*pieces):
             p[T.li[pieces]]
 
         def add(*pieces):
             p[T.li[pieces]]
 
@@ -243,7 +242,6 @@ class StorageStatus(rend.Page):
                 ]
 
         p = T.ul()
                 ]
 
         p = T.ul()
-        pieces = []
         def add(*pieces):
             p[T.li[pieces]]
 
         def add(*pieces):
             p[T.li[pieces]]