3 # Utility for managing LBS archives.
5 import getpass, os, stat, sys, time
6 from optparse import OptionParser
9 # We support up to "LBS Snapshot v0.6" formats, but are also limited by the lbs
11 FORMAT_VERSION = min(lbs.FORMAT_VERSION, (0, 6))
13 def check_version(format):
14 ver = lbs.parse_metadata_version(format)
15 if ver > FORMAT_VERSION:
16 raise RuntimeError("Unsupported LBS format: " + format)
18 parser = OptionParser(usage="%prog [option]... command [arg]...")
19 parser.add_option("-v", action="store_true", dest="verbose", default=False,
20 help="increase verbosity")
21 parser.add_option("--store", dest="store",
22 help="specify path to backup data store")
23 parser.add_option("--localdb", dest="localdb",
24 help="specify path to local database")
25 parser.add_option("--intent", dest="intent", default=1.0,
26 help="give expected next snapshot type when cleaning")
27 (options, args) = parser.parse_args(sys.argv[1:])
29 # Read a passphrase from the user and store it in the LBS_GPG_PASSPHRASE
30 # environment variable.
32 ENV_KEY = 'LBS_GPG_PASSPHRASE'
33 if not os.environ.has_key(ENV_KEY):
34 os.environ[ENV_KEY] = getpass.getpass()
36 # Delete old snapshots from the local database, though do not actually schedule
37 # any segment cleaning.
38 # Syntax: $0 --localdb=LOCALDB prune-db
40 db = lbs.LocalDatabase(options.localdb)
42 # Delete old snapshots from the local database.
46 # Run the segment cleaner.
47 # Syntax: $0 --localdb=LOCALDB clean
48 def cmd_clean(clean_threshold=7.0):
49 db = lbs.LocalDatabase(options.localdb)
51 # Delete old snapshots from the local database.
52 intent = float(options.intent)
53 for s in db.list_schemes():
54 db.garbage_collect(s, intent)
56 # Expire segments which are poorly-utilized.
57 for s in db.get_segment_cleaning_list():
58 if s.cleaning_benefit > clean_threshold:
59 print "Cleaning segment %d (benefit %.2f)" % (s.id,
61 db.mark_segment_expired(s)
64 db.balance_expired_objects()
67 # List snapshots stored.
68 # Syntax: $0 --data=DATADIR list-snapshots
69 def cmd_list_snapshots():
70 store = lbs.LowlevelDataStore(options.store)
71 for s in sorted(store.list_snapshots()):
74 # List size of data needed for each snapshot.
75 # Syntax: $0 --data=DATADIR list-snapshot-sizes
76 def cmd_list_snapshot_sizes():
77 lowlevel = lbs.LowlevelDataStore(options.store)
78 store = lbs.ObjectStore(lowlevel)
80 for s in sorted(lowlevel.list_snapshots()):
81 d = lbs.parse_full(store.load_snapshot(s))
82 check_version(d['Format'])
83 segments = d['Segments'].split()
84 (size, added, removed) = (0, 0, 0)
86 segsize = lowlevel.lowlevel_stat(seg + ".tar.gpg")['size']
88 if seg not in previous: added += segsize
90 if seg not in segments:
91 removed += lowlevel.lowlevel_stat(seg + ".tar.gpg")['size']
92 previous = set(segments)
93 print "%s: %.3f +%.3f -%.3f" % (s, size / 1024.0**2, added / 1024.0**2, removed / 1024.0**2)
95 # Build checksum list for objects in the given segments, or all segments if
97 def cmd_object_checksums(segments):
99 lowlevel = lbs.LowlevelDataStore(options.store)
100 store = lbs.ObjectStore(lowlevel)
101 if len(segments) == 0:
102 segments = sorted(lowlevel.list_segments())
104 for (o, data) in store.load_segment(s):
105 csum = lbs.ChecksumCreator().update(data).compute()
106 print "%s/%s:%d:%s" % (s, o, len(data), csum)
109 # Read a snapshot file
110 def cmd_read_snapshots(snapshots):
112 lowlevel = lbs.LowlevelDataStore(options.store)
113 store = lbs.ObjectStore(lowlevel)
115 d = lbs.parse_full(store.load_snapshot(s))
116 check_version(d['Format'])
118 print d['Segments'].split()
121 # Produce a flattened metadata dump from a snapshot
122 def cmd_read_metadata(snapshot):
124 lowlevel = lbs.LowlevelDataStore(options.store)
125 store = lbs.ObjectStore(lowlevel)
126 d = lbs.parse_full(store.load_snapshot(snapshot))
127 check_version(d['Format'])
128 metadata = lbs.read_metadata(store, d['Root'])
133 # Verify snapshot integrity
134 def cmd_verify_snapshots(snapshots):
136 lowlevel = lbs.LowlevelDataStore(options.store)
137 store = lbs.ObjectStore(lowlevel)
139 lbs.accessed_segments.clear()
140 print "#### Snapshot", s
141 d = lbs.parse_full(store.load_snapshot(s))
142 check_version(d['Format'])
143 print "## Root:", d['Root']
144 metadata = lbs.iterate_metadata(store, d['Root'])
146 if m.fields['type'] not in ('-', 'f'): continue
147 print "%s [%d bytes]" % (m.fields['name'], int(m.fields['size']))
148 verifier = lbs.ChecksumVerifier(m.fields['checksum'])
150 for block in m.data():
151 data = store.get(block)
152 verifier.update(data)
154 if int(m.fields['size']) != size:
155 raise ValueError("File size does not match!")
156 if not verifier.valid():
157 raise ValueError("Bad checksum found")
159 # Verify that the list of segments included with the snapshot was
160 # actually accurate: covered all segments that were really read, and
161 # doesn't contain duplicates.
162 listed_segments = set(d['Segments'].split())
163 if lbs.accessed_segments - listed_segments:
164 print "Error: Some segments not listed in descriptor!"
165 print sorted(list(lbs.accessed_segments - listed_segments))
166 if listed_segments - lbs.accessed_segments :
167 print "Warning: Extra unused segments listed in descriptor!"
168 print sorted(list(listed_segments - lbs.accessed_segments))
171 # Restore a snapshot, or some subset of files from it
172 def cmd_restore_snapshot(args):
174 lowlevel = lbs.LowlevelDataStore(options.store)
175 store = lbs.ObjectStore(lowlevel)
176 snapshot = lbs.parse_full(store.load_snapshot(args[0]))
177 check_version(snapshot['Format'])
182 "Return true if the specified path should be included in the restore."
184 # No specification of what to restore => restore everything
185 if len(paths) == 0: return True
188 if path == p: return True
189 if path.startswith(p + "/"): return True
193 print "Warning: %s: %s" % (m.items.name, msg)
195 # Phase 1: Read the complete metadata log and create directory structure.
198 metadata_segments = {}
199 for m in lbs.iterate_metadata(store, snapshot['Root']):
200 pathname = os.path.normpath(m.items.name)
201 while os.path.isabs(pathname):
202 pathname = pathname[1:]
203 if not matchpath(pathname): continue
205 destpath = os.path.join(destdir, pathname)
206 if m.items.type == 'd':
209 (path, filename) = os.path.split(destpath)
211 metadata_items.append((pathname, m))
212 if m.items.type in ('-', 'f'):
213 metadata_paths[pathname] = m
214 for block in m.data():
215 (segment, object, checksum, slice) \
216 = lbs.ObjectStore.parse_ref(block)
217 if segment not in metadata_segments:
218 metadata_segments[segment] = set()
219 metadata_segments[segment].add(pathname)
222 if not os.path.isdir(path):
226 warn(m, "Error creating directory structure: %s" % (e,))
229 # Phase 2: Restore files, ordered by how data is stored in segments.
230 def restore_file(pathname, m):
231 assert m.items.type in ('-', 'f')
232 print "extract:", pathname
233 destpath = os.path.join(destdir, pathname)
235 file = open(destpath, 'wb')
236 verifier = lbs.ChecksumVerifier(m.items.checksum)
238 for block in m.data():
239 data = store.get(block)
240 verifier.update(data)
244 if int(m.fields['size']) != size:
245 raise ValueError("File size does not match!")
246 if not verifier.valid():
247 raise ValueError("Bad checksum found")
249 while metadata_segments:
250 (segment, items) = metadata_segments.popitem()
251 print "+ Segment", segment
252 for pathname in sorted(items):
253 if pathname in metadata_paths:
254 restore_file(pathname, metadata_paths[pathname])
255 del metadata_paths[pathname]
257 print "+ Remaining files"
258 while metadata_paths:
259 (pathname, m) = metadata_paths.popitem()
260 restore_file(pathname, m)
262 # Phase 3: Restore special files (symlinks, devices).
263 # Phase 4: Restore directory permissions and modification times.
264 for (pathname, m) in reversed(metadata_items):
265 print "permissions:", pathname
266 destpath = os.path.join(destdir, pathname)
267 (path, filename) = os.path.split(destpath)
269 # TODO: Check for ../../../paths that might attempt to write outside
270 # the destination directory. Maybe also check attempts to follow
271 # symlinks pointing outside?
274 if m.items.type in ('-', 'f', 'd'):
276 elif m.items.type == 'l':
278 target = m.items.target
280 # Old (v0.2 format) name for 'target'
281 target = m.items.contents
282 os.symlink(target, destpath)
283 elif m.items.type == 'p':
285 elif m.items.type in ('c', 'b'):
286 if m.items.type == 'c':
287 mode = 0600 | stat.S_IFCHR
289 mode = 0600 | stat.S_IFBLK
290 os.mknod(destpath, mode, os.makedev(*m.items.device))
291 elif m.items.type == 's':
292 pass # TODO: Implement
294 warn(m, "Unknown type code: " + m.items.type)
298 warn(m, "Error restoring: %s" % (e,))
302 uid = m.items.user[0]
303 gid = m.items.group[0]
304 os.lchown(destpath, uid, gid)
306 warn(m, "Error restoring file ownership: %s" % (e,))
308 if m.items.type == 'l':
312 os.chmod(destpath, m.items.mode)
314 warn(m, "Error restoring file permissions: %s" % (e,))
317 os.utime(destpath, (time.time(), m.items.mtime))
319 warn(m, "Error restoring file timestamps: %s" % (e,))
330 elif cmd == 'prune-db':
332 elif cmd == 'list-snapshots':
334 elif cmd == 'object-sums':
335 cmd_object_checksums(args)
336 elif cmd == 'read-snapshots':
337 cmd_read_snapshots(args)
338 elif cmd == 'read-metadata':
339 cmd_read_metadata(args[0])
340 elif cmd == 'list-snapshot-sizes':
341 cmd_list_snapshot_sizes()
342 elif cmd == 'verify-snapshots':
343 cmd_verify_snapshots(args)
344 elif cmd == 'restore-snapshot':
345 cmd_restore_snapshot(args)
347 print "Unknown command:", cmd