#!/usr/bin/env python # # A simple cleaner for the BlueSky cloud file system. At the moment this is an # offline cleaner--the file system must not be in use at the time that the # cleaning is performed. Later, it will be extended to be an online/concurrent # cleaner, where cleaning is performed even while the file system is mounted. # # Copyright (C) 2010 The Regents of the University of California # Written by Michael Vrable # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions # are met: # 1. Redistributions of source code must retain the above copyright # notice, this list of conditions and the following disclaimer. # 2. Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in the # documentation and/or other materials provided with the distribution. # 3. Neither the name of the University nor the names of its contributors # may be used to endorse or promote products derived from this software # without specific prior written permission. # # THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND # ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE # IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE # ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE # FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL # DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS # OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) # HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT # LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY # OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF # SUCH DAMAGE. import base64, os, re, struct, sys, time import boto from boto.s3.key import Key # The BlueSky 'struct cloudlog_header' data type. HEADER_FORMAT = '<4s48sb16sQIII' HEADER_CRYPTBYTES = 48 HEADER_MAGIC1 = 'AgI-' # Unencrypted data HEADER_MAGIC2 = 'AgI=' # Encrypted data HEADER_SIZE = struct.calcsize(HEADER_FORMAT) CHECKPOINT_MAGIC = struct.pack(' 0: fp.seek(offset) if length is None: return fp.read() else: return fp.read(length) def write(self, filename, data): fp = open(os.path.join(self.path, filename), 'wb') fp.write(data) fp.close() def delete(self, filename): os.unlink(os.path.join(self.path, filename)) def retry_wrap(method): def wrapped(self, *args, **kwargs): for retries in range(3): try: return method(self, *args, **kwargs) except: print >>sys.stderr, "S3 operation failed, retrying..." print >>sys.stderr, " %s %s %s" % (method, args, kwargs) self.connect() time.sleep(1.0) return method(self, *args, **kwargs) return wrapped class S3Backend(Backend): """An interface to BlueSky where the log segments are on in Amazon S3.""" def __init__(self, bucket, path='', cachedir="."): self.bucket_name = bucket self.path = path self.cachedir = cachedir self.cache = {} for f in os.listdir(cachedir): self.cache[f] = True #print "Initial cache contents:", list(self.cache.keys()) self.connect() self.stats_get = [0, 0] self.stats_put = [0, 0] def connect(self): self.conn = boto.connect_s3(is_secure=False) self.bucket = self.conn.get_bucket(self.bucket_name) def list(self, directory=0): files = [] prefix = "log-%08d-" % (directory,) for k in self.bucket.list(self.path + prefix): files.append((k.key, k.size)) return files @retry_wrap def read(self, filename, offset=0, length=None): if filename in self.cache: fp = open(os.path.join(self.cachedir, filename), 'rb') if offset > 0: fp.seek(offset) if length is None: return fp.read() else: return fp.read(length) else: k = Key(self.bucket) k.key = self.path + filename data = k.get_contents_as_string() fp = open(os.path.join(self.cachedir, filename), 'wb') fp.write(data) fp.close() self.cache[filename] = True self.stats_get[0] += 1 self.stats_get[1] += len(data) if offset > 0: data = data[offset:] if length is not None: data = data[0:length] return data @retry_wrap def write(self, filename, data): k = Key(self.bucket) k.key = self.path + filename k.set_contents_from_string(data) self.stats_put[0] += 1 self.stats_put[1] += len(data) if filename in self.cache: del self.cache[filename] @retry_wrap def delete(self, filename): k = Key(self.bucket) k.key = self.path + filename k.delete() if filename in self.cache: del self.cache[filename] def dump_stats(self): print "S3 statistics:" print "GET: %d ops / %d bytes" % tuple(self.stats_get) print "PUT: %d ops / %d bytes" % tuple(self.stats_put) benchlog_write("s3_get: %d", self.stats_get[1]) benchlog_write("s3_put: %d", self.stats_put[1]) class SimpleBackend(Backend): """An interface to the simple BlueSky test network server.""" def __init__(self, server=('localhost', 12345), cachedir="."): self.bucket_name = bucket self.server_address = server self.cachedir = cachedir self.cache = {} def _get_socket(self): return socket.create_connection(self.server_address).makefile() def list(self, directory=0): files = [] prefix = "log-%08d-" % (directory,) for k in self.bucket.list(self.path + prefix): files.append((k.key, k.size)) return files def read(self, filename, offset=0, length=None): if filename in self.cache: fp = open(os.path.join(self.cachedir, filename), 'rb') if offset > 0: fp.seek(offset) if length is None: return fp.read() else: return fp.read(length) else: f = self._get_socket() f.write("GET %s %d %d\n" % (filename, 0, 0)) f.flush() datalen = int(f.readline()) if datalen < 0: raise RuntimeError data = f.read(datalen) fp = open(os.path.join(self.cachedir, filename), 'wb') fp.write(data) fp.close() self.cache[filename] = True if offset > 0: data = data[offset:] if length is not None: data = data[0:length] return data def write(self, filename, data): f = self._get_socket() f.write("PUT %s %d %d\n" % (filename, len(data))) f.write(data) f.flush() result = int(f.readline()) if filename in self.cache: del self.cache[filename] def delete(self, filename): pass class LogItem: """In-memory representation of a single item stored in a log file.""" def __init__(self): self.cryptkeys = '\0' * HEADER_CRYPTBYTES self.encrypted = False def __str__(self): return "" % (self.encrypted and '$' or '', self.type, self.location, self.size, base64.b16encode(self.id).lower()[0:8]) @staticmethod def random_id(): return open('/dev/urandom').read(16) def serialize(self): link_ids = [] link_locs = [] for (i, l) in self.links: link_ids.append(i) if i != '\0' * 16: link_locs.append(struct.pack('= LogDirectory.TARGET_SIZE: seg.close() del self.groups[segment_group] def close_all(self): for k in list(self.groups.keys()): self.groups[k].close() del self.groups[k] class UtilizationTracker: """A simple object that tracks what fraction of each segment is used. This data can be used to guide segment cleaning decisions.""" def __init__(self, backend): self.segments = {} for (segment, size) in backend.list(0) + backend.list(1): self.segments[segment] = [size, 0] def add_item(self, item): if isinstance(item, LogItem): item = item.location if item is None: return (dir, seq, offset, size) = item filename = "log-%08d-%08d" % (dir, seq) self.segments[filename][1] += size def parse_item(data): if len(data) < HEADER_SIZE: return header = struct.unpack_from(HEADER_FORMAT, data, 0) size = HEADER_SIZE + sum(header[5:8]) if header[0] not in (HEADER_MAGIC1, HEADER_MAGIC2): print "Bad header magic!" return if len(data) != size: print "Item size does not match: %d != %d" % (size, len(data)) return item = LogItem() if header[0] == HEADER_MAGIC2: item.encrypted = True item.cryptkeys = header[1] item.id = header[3] item.inum = header[4] item.location = None item.type = chr(header[2]) item.size = size item.data = data[HEADER_SIZE : HEADER_SIZE + header[5]] links = [] link_ids = data[HEADER_SIZE + header[5] : HEADER_SIZE + header[5] + header[6]] link_locs = data[HEADER_SIZE + header[5] + header[6] : HEADER_SIZE + sum(header[5:8])] for i in range(len(link_ids) // 16): id = link_ids[16*i : 16*i + 16] if id == '\0' * 16: loc = None else: loc = struct.unpack('= HEADER_SIZE: header = struct.unpack_from(HEADER_FORMAT, data, offset) size = HEADER_SIZE + sum(header[5:8]) if header[0] not in (HEADER_MAGIC1, HEADER_MAGIC2): print "Bad header magic!" break if size + offset > len(data): print "Short data record at end of log: %s < %s" % (len(data) - offset, size) break item = parse_item(data[offset : offset + size]) if location is not None: item.location = location + (offset, size) if item is not None: yield item offset += size def load_checkpoint_record(backend, directory=0): for (log, size) in reversed(backend.list(directory)): for item in reversed(list(parse_log(backend.read(log), log))): print item if item.type == ITEM_TYPE.CHECKPOINT: return item class InodeMap: def __init__(self): pass def build(self, backend, checkpoint_record): """Reconstruct the inode map from the checkpoint record given. This will also build up information about segment utilization.""" self.version_vector = {} self.checkpoint_record = checkpoint_record util = UtilizationTracker(backend) util.add_item(checkpoint_record) inodes = {} self.obsolete_segments = set() data = checkpoint_record.data if not data.startswith(CHECKPOINT_MAGIC): raise ValueError, "Invalid checkpoint record!" data = data[len(CHECKPOINT_MAGIC):] (vvlen,) = struct.unpack_from(" end: old_location = self.checkpoint_record.links[i][1][0:2] if old_location not in self.obsolete_segments: new_checkpoint.links.append(self.checkpoint_record.links[i]) continue # Case 2: Some inodes have been updated. Create a new inode map # block, write it out, and point the new checkpoint at it. inodes = [k for k in self.inodes if k >= start and k <= end] inodes.sort() block = LogItem() block.id = LogItem.random_id() block.inum = 0 block.type = ITEM_TYPE.INODE_MAP block.links = [] block.data = "" for j in inodes: block.data += struct.pack(" 0 and updated_inodes[-1] <= end: updated_inodes.pop() log.write(new_checkpoint, 2) self.checkpoint_record = new_checkpoint def rewrite_inode(backend, inode_map, inum, log, copy_data=True): inode = inode_map.inodes[inum] if copy_data: newlinks = [] for l in inode.links: if l[1] is not None: data = load_item(backend, l[1]) log.write(data, 0) newlinks.append((data.id, data.location)) else: newlinks.append(l) inode.links = newlinks log.write(inode, 1) inode_map.mark_updated(inum) def run_cleaner(backend, inode_map, log, repack_inodes=False): # Determine which segments are poorly utilized and should be cleaned. We # need better heuristics here. for (s, u) in sorted(inode_map.util.segments.items()): if (float(u[1]) / u[0] < 0.6) and u[1] > 0: print "Should clean segment", s loc = backend.name_to_loc(s) if s: inode_map.obsolete_segments.add(loc) # TODO: We probably also want heuristics that will find inodes with # badly-fragmented data and rewrite that to achieve better locality. # Given that list of segments to clean, scan through those segments to find # data which is still live and mark relevant inodes as needing to be # rewritten. if repack_inodes: dirty_inodes = set(inode_map.inodes) else: dirty_inodes = set() dirty_inode_data = set() for s in inode_map.obsolete_segments: filename = backend.loc_to_name(s) #print "Scanning", filename, "for live data" for item in parse_log(backend.read(filename), filename): if item.type in (ITEM_TYPE.DATA, ITEM_TYPE.INODE): if item.inum != 0: inode = inode_map.inodes[item.inum] if s == inode.location[0:2]: dirty_inodes.add(item.inum) if item.inum not in dirty_inode_data: for b in inode.links: if b[1] is not None and s == b[1][0:2]: dirty_inode_data.add(item.inum) break #print "Inodes to rewrite:", dirty_inodes #print "Inodes with data to rewrite:", dirty_inode_data for i in sorted(dirty_inodes.union(dirty_inode_data)): rewrite_inode(backend, inode_map, i, log, i in dirty_inode_data) if __name__ == '__main__': benchlog = open('cleaner.log', 'a') benchlog_write("*** START CLEANER RUN ***") start_time = time.time() backend = S3Backend("mvrable-bluesky-west", cachedir="/tmp/bluesky-cache") #backend = FileBackend(".") chkpt = load_checkpoint_record(backend) #print backend.list() log_dir = LogDirectory(backend, 1) imap = InodeMap() imap.build(backend, chkpt) print chkpt print "Version vector:", imap.version_vector print "Last cleaner log file:", log_dir.seq_num - 1 if imap.version_vector.get(1, -1) != log_dir.seq_num - 1: print "Proxy hasn't updated to latest cleaner segment yet!" benchlog_write("waiting for proxy...") sys.exit(0) run_cleaner(backend, imap, log_dir) print "Version vector:", imap.version_vector imap.write(backend, log_dir) log_dir.close_all() end_time = time.time() backend.dump_stats() benchlog_write("running_time: %s", end_time - start_time) benchlog_write("")