SDMFSlotWriteProxy
KiB = 1024
-DEFAULT_MAX_SEGMENT_SIZE = 64
+DEFAULT_MAX_SEGMENT_SIZE = 128 * KiB
PUSHING_BLOCKS_STATE = 0
PUSHING_EVERYTHING_ELSE_STATE = 1
DONE_STATE = 2
hashed = sharedata
block_hash = hashutil.block_hash(hashed)
self.blockhashes[shareid][segnum] = block_hash
- log.msg("yyy 0 shareid: %s, segnum: %s, blockhash: %s, sharedata: %s, salt: %s" % (shareid, segnum, base32.b2a(block_hash), base32.b2a(sharedata), base32.b2a(salt),))
# find the writer for this share
writer = self.writers[shareid]
writer.put_block(sharedata, segnum, salt)
sharehashes[1].keys())
bht = self._block_hash_trees[reader.shnum]
- for bhk, bhv in blockhashes.iteritems():
- log.msg("xxx 0 blockhash: %s %s" % (bhk, base32.b2a(bhv),))
-
if bht.needed_hashes(segnum, include_leaf=True):
try:
bht.set_hashes(blockhashes)
self.set_up_grid()
self.c = self.g.clients[0]
self.nm = self.c.nodemaker
- self.data = "test data" * 32 # about 900 KiB; MDMF
+ self.data = "test data" * 100000 # about 900 KiB; MDMF
self.small_data = "test data" * 10 # about 90 B; SDMF
self.set_up_grid()
self.c = self.g.clients[0]
self.nm = self.c.nodemaker
- self.data = "testdata " * 30 # about 900 KiB; MDMF
+ self.data = "testdata " * 100000 # about 900 KiB; MDMF
self.small_data = "test data" * 10 # about 90 B; SDMF