projects
/
bluesky.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
Create a simple Python script for sending pipelined GET requests to S3.
[bluesky.git]
/
microbench
/
run.sh
diff --git
a/microbench/run.sh
b/microbench/run.sh
index
8a0dd74
..
28571ef
100755
(executable)
--- a/
microbench/run.sh
+++ b/
microbench/run.sh
@@
-18,7
+18,7
@@
run_setup() {
$NFSD &
NFSPID=$!
sleep 0.5
$NFSD &
NFSPID=$!
sleep 0.5
- sudo mount -t nfs -o tcp,hard,intr localhost:/bluesky /mnt/bluesky || die "Mounting NFS"
+ sudo mount -t nfs -o tcp,hard,intr
,vers=3
localhost:/bluesky /mnt/bluesky || die "Mounting NFS"
}
run_cleanup() {
}
run_cleanup() {
@@
-32,7
+32,7
@@
benchmark_setup() {
echo "Checking for sudo access..." 1>&2
sudo whoami
echo "Preparing for benchmark run..." 1>&2
echo "Checking for sudo access..." 1>&2
sudo whoami
echo "Preparing for benchmark run..." 1>&2
-
$HOME/docs/2009/fall/s3test/s3-cleanup.py
mvrable-bluesky
+
s3cmd -r --force del s3://
mvrable-bluesky
run_setup
for i in {1..64}; do
dd if=/dev/urandom of=/mnt/bluesky/file-$i bs=32k count=1
run_setup
for i in {1..64}; do
dd if=/dev/urandom of=/mnt/bluesky/file-$i bs=32k count=1
@@
-69,6
+69,7
@@
for s in $SIZES; do
run_cleanup
done
run_cleanup
done
+if false; then
export BLUESKY_OPT_SYNC_FRONTENDS=0
for t in 1 2 4 6 8 10 12 14 16 20 24 28 32 40 48 56 64; do
run_thread_test $t
export BLUESKY_OPT_SYNC_FRONTENDS=0
for t in 1 2 4 6 8 10 12 14 16 20 24 28 32 40 48 56 64; do
run_thread_test $t
@@
-78,5
+79,6
@@
export BLUESKY_OPT_SYNC_FRONTENDS=1
for t in 1 2 4 6 8 10 12 14 16 20 24 28 32 40 48 56 64; do
run_thread_test $t
done
for t in 1 2 4 6 8 10 12 14 16 20 24 28 32 40 48 56 64; do
run_thread_test $t
done
+fi
benchmark_cleanup
benchmark_cleanup