projects
/
cumulus.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
Support for spreading objects across segments.
[cumulus.git]
/
scandir.cc
diff --git
a/scandir.cc
b/scandir.cc
index
3e004fd
..
043d35e
100644
(file)
--- a/
scandir.cc
+++ b/
scandir.cc
@@
-15,12
+15,16
@@
#include <vector>
#include "store.h"
#include <vector>
#include "store.h"
+#include "sha1.h"
using std::string;
using std::vector;
using std::string;
using std::vector;
+static SegmentStore *segment_store;
static OutputStream *info_dump = NULL;
static OutputStream *info_dump = NULL;
+static SegmentPartitioner *index_segment, *data_segment;
+
void scandir(const string& path);
/* Converts time to microseconds since the epoch. */
void scandir(const string& path);
/* Converts time to microseconds since the epoch. */
@@
-29,7
+33,7
@@
int64_t encode_time(time_t time)
return (int64_t)time * 1000000;
}
return (int64_t)time * 1000000;
}
-void dumpfile(int fd)
+void dumpfile(int fd
, dictionary &file_info
)
{
struct stat stat_buf;
fstat(fd, &stat_buf);
{
struct stat stat_buf;
fstat(fd, &stat_buf);
@@
-42,6
+46,7
@@
void dumpfile(int fd)
return;
}
return;
}
+ SHA1Checksum hash;
while (true) {
ssize_t res = read(fd, buf, sizeof(buf));
if (res < 0) {
while (true) {
ssize_t res = read(fd, buf, sizeof(buf));
if (res < 0) {
@@
-52,11
+57,15
@@
void dumpfile(int fd)
} else if (res == 0) {
break;
} else {
} else if (res == 0) {
break;
} else {
+ hash.process(buf, res);
+ OutputStream *block = data_segment->new_object();
+ block->write(buf, res);
size += res;
}
}
size += res;
}
}
- printf(" bytes=%Ld\n", size);
+ file_info["sha1"] = string((const char *)hash.checksum(),
+ hash.checksum_size());
}
void scanfile(const string& path)
}
void scanfile(const string& path)
@@
-67,6
+76,9
@@
void scanfile(const string& path)
char *buf;
ssize_t len;
char *buf;
ssize_t len;
+ // Set to true if the item is a directory and we should recursively scan
+ bool recurse = false;
+
dictionary file_info;
lstat(path.c_str(), &stat_buf);
dictionary file_info;
lstat(path.c_str(), &stat_buf);
@@
-137,14
+149,14
@@
void scanfile(const string& path)
flags = fcntl(fd, F_GETFL);
fcntl(fd, F_SETFL, flags & ~O_NONBLOCK);
flags = fcntl(fd, F_GETFL);
fcntl(fd, F_SETFL, flags & ~O_NONBLOCK);
- //dumpfile(fd);
file_info["size"] = encode_u64(stat_buf.st_size);
file_info["size"] = encode_u64(stat_buf.st_size);
+ dumpfile(fd, file_info);
close(fd);
break;
case S_IFDIR:
inode_type = 'd';
close(fd);
break;
case S_IFDIR:
inode_type = 'd';
-
scandir(path)
;
+
recurse = true
;
break;
default:
break;
default:
@@
-156,6
+168,11
@@
void scanfile(const string& path)
info_dump->write_string(path);
info_dump->write_dictionary(file_info);
info_dump->write_string(path);
info_dump->write_dictionary(file_info);
+
+ // If we hit a directory, now that we've written the directory itself,
+ // recursively scan the directory.
+ if (recurse)
+ scandir(path);
}
void scandir(const string& path)
}
void scandir(const string& path)
@@
-189,14
+206,15
@@
void scandir(const string& path)
int main(int argc, char *argv[])
{
int main(int argc, char *argv[])
{
- FILE *dump = fopen("fileinfo", "w");
- if (dump == NULL) {
- fprintf(stderr, "Cannot open fileinfo: %m\n");
- return 1;
- }
+ segment_store = new SegmentStore(".");
+ SegmentWriter *sw = segment_store->new_segment();
+ info_dump = sw->new_object();
- FileOutputStream os(dump);
- info_dump = &os;
+ index_segment = new SegmentPartitioner(segment_store);
+ data_segment = new SegmentPartitioner(segment_store);
+
+ string uuid = SegmentWriter::format_uuid(sw->get_uuid());
+ printf("Backup UUID: %s\n", uuid.c_str());
try {
scanfile(".");
try {
scanfile(".");
@@
-204,5
+222,9
@@
int main(int argc, char *argv[])
fprintf(stderr, "IOException: %s\n", e.getError().c_str());
}
fprintf(stderr, "IOException: %s\n", e.getError().c_str());
}
+ delete index_segment;
+ delete data_segment;
+ delete sw;
+
return 0;
}
return 0;
}