From 38f09c64ffe20088ca1fa567f27b9b2daa7e1a10 Mon Sep 17 00:00:00 2001 From: Michael Vrable Date: Wed, 16 Jun 2010 13:50:29 -0700 Subject: [PATCH] Testing S3 with more object sizes. --- cloudbench/paralleltest.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/cloudbench/paralleltest.py b/cloudbench/paralleltest.py index 8af4cb4..3f74b2f 100755 --- a/cloudbench/paralleltest.py +++ b/cloudbench/paralleltest.py @@ -95,11 +95,11 @@ def run_test(size, threads, num, logfile=sys.stdout, delay=1.0): connections = [S3TestConnection() for _ in range(128)] logfile = open('multifetch-simulation.data', 'a') -for s in SIZES: +for s in [(1 << s) for s in range(16, 27)]: print "Priming objects: %d-byte objects" % (s,) run_test(s, 1, 100, open('/dev/null', 'w'), 0.0) - for blocksize in [1 << 20, 2 << 20, 4 << 20, 8 << 20]: + for blocksize in [x << 20 for x in (4, 8, 16, 32, 64, 128)]: if s > blocksize: continue for t in [4, 2, 1]: for rep in range(10): -- 2.20.1