-/* Cumulus: Smart Filesystem Backup to Dumb Servers
- *
- * Copyright (C) 2008 The Regents of the University of California
- * Written by Michael Vrable <mvrable@cs.ucsd.edu>
+/* Cumulus: Efficient Filesystem Backup to the Cloud
+ * Copyright (C) 2008-2009 The Cumulus Developers
+ * See the AUTHORS file for a list of contributors.
*
* This program is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License as published by
#include <string>
#include <iostream>
+#include "hash.h"
+#include "localdb.h"
#include "store.h"
#include "ref.h"
+#include "util.h"
using std::max;
using std::list;
const char *filter_program = "bzip2 -c";
const char *filter_extension = ".bz2";
-static void cloexec(int fd)
-{
- long flags = fcntl(fd, F_GETFD);
-
- if (flags < 0)
- return;
-
- fcntl(fd, F_SETFD, flags | FD_CLOEXEC);
-}
-
Tarfile::Tarfile(RemoteFile *file, const string &segment)
: size(0),
segment_name(segment)
tar_write(buf, TAR_BLOCK_SIZE);
if (close(filter_fd) != 0)
- throw IOException("Error closing Tarfile");
+ fatal("Error closing Tarfile");
/* ...and wait for filter process to finish. */
int status;
waitpid(filter_pid, &status, 0);
if (!WIFEXITED(status) || WEXITSTATUS(status) != 0) {
- throw IOException("Filter process error");
+ fatal("Filter process error");
}
close(real_fd);
/* Create a pipe for communicating with the filter process. */
if (pipe(fds) < 0) {
- throw IOException("Unable to create pipe for filter");
+ fatal("Unable to create pipe for filter");
}
/* Create a child process which can exec() the filter program. */
pid = fork();
if (pid < 0)
- throw IOException("Unable to fork filter process");
+ fatal("Unable to fork filter process");
if (pid > 0) {
/* Parent process */
if (errno == EINTR)
continue;
fprintf(stderr, "Write error: %m\n");
- throw IOException("Write error");
+ fatal("Write error");
}
len -= res;
sprintf(header.mode, "%07o", 0600);
sprintf(header.uid, "%07o", 0);
sprintf(header.gid, "%07o", 0);
- sprintf(header.size, "%011o", len);
+ sprintf(header.size, "%011o", (int)len);
sprintf(header.mtime, "%011o", (int)time(NULL));
header.typeflag = '0';
strcpy(header.magic, "ustar ");
static map<string, pair<int64_t, int64_t> > group_sizes;
ObjectReference TarSegmentStore::write_object(const char *data, size_t len,
- const std::string &group)
+ const std::string &group,
+ const std::string &checksum,
+ double age)
{
struct segment_info *segment;
segment->basename = segment->name + ".tar";
segment->basename += filter_extension;
segment->count = 0;
- segment->size = 0;
- segment->rf = remote->alloc_file(segment->basename, "segments");
+ segment->data_size = 0;
+ segment->rf = remote->alloc_file(segment->basename,
+ group == "metadata" ? "segments0"
+ : "segments1");
segment->file = new Tarfile(segment->rf, segment->name);
segments[group] = segment;
segment->file->write_object(id, data, len);
segment->count++;
- segment->size += len;
+ segment->data_size += len;
group_sizes[group].first += len;
ObjectReference ref(segment->name, id_buf);
+ ref.set_range(0, len, true);
+ if (checksum.size() > 0)
+ ref.set_checksum(checksum);
+ if (db != NULL)
+ db->StoreObject(ref, age);
// If this segment meets or exceeds the size target, close it so that
// future objects will go into a new segment.
for (map<string, pair<int64_t, int64_t> >::iterator i = group_sizes.begin();
i != group_sizes.end(); ++i) {
printf(" %s: %lld (%lld compressed)\n", i->first.c_str(),
- i->second.first, i->second.second);
+ (long long)i->second.first, (long long)i->second.second);
}
}
delete segment->file;
if (db != NULL) {
+ struct stat stat_buf;
+ int disk_size = 0;
+ if (stat(segment->rf->get_local_path().c_str(), &stat_buf) == 0) {
+ disk_size = stat_buf.st_size;
+ group_sizes[segment->group].second += disk_size;
+ }
+
SHA1Checksum segment_checksum;
+ string checksum;
if (segment_checksum.process_file(segment->rf->get_local_path().c_str())) {
- string checksum = segment_checksum.checksum_str();
- db->SetSegmentChecksum(segment->name, segment->basename, checksum,
- segment->size);
+ checksum = segment_checksum.checksum_str();
}
- struct stat stat_buf;
- if (stat(segment->rf->get_local_path().c_str(), &stat_buf) == 0) {
- group_sizes[segment->group].second += stat_buf.st_size;
- }
+ db->SetSegmentMetadata(segment->name, segment->basename, checksum,
+ group, segment->data_size, disk_size);
}
segment->rf->send();
}
LbsObject::LbsObject()
- : group(""), data(NULL), data_len(0), written(false)
+ : group(""), age(0.0), data(NULL), data_len(0), written(false)
{
}
{
}
-void LbsObject::write(TarSegmentStore *store)
+void LbsObject::set_data(const char *d, size_t len, const char *checksum)
{
- assert(data != NULL);
- assert(!written);
+ data = d;
+ data_len = len;
- ref = store->write_object(data, data_len, group);
- written = true;
+ if (checksum != NULL) {
+ this->checksum = checksum;
+ } else {
+ Hash *hash = Hash::New();
+ hash->update(data, data_len);
+ this->checksum = hash->digest_str();
+ delete hash;
+ }
}
-void LbsObject::checksum()
+void LbsObject::write(TarSegmentStore *store)
{
- assert(written);
+ assert(data != NULL);
+ assert(!written);
- SHA1Checksum hash;
- hash.process(data, data_len);
- ref.set_checksum(hash.checksum_str());
+ ref = store->write_object(data, data_len, group, checksum, age);
+ written = true;
}