* - Writing batches of data to the cloud.
*/
+#if 0
static void writeback_complete(gpointer a, gpointer i)
{
BlueSkyInode *inode = (BlueSkyInode *)i;
g_mutex_unlock(inode->lock);
}
+#endif
static void flushd_dirty_inode(BlueSkyInode *inode)
{
if (inode->change_count == inode->change_commit)
return;
- /* Inode writeback is in progress; put back on the dirty list. */
- if (inode->change_pending) {
- /* Waiting for an earlier writeback to finish, so don't start a new
- * writeback yet. */
- g_mutex_lock(fs->lock);
- inode->change_time = bluesky_get_current_time();
- bluesky_list_unlink(&fs->dirty_list, inode->dirty_list);
- inode->dirty_list = bluesky_list_prepend(&fs->dirty_list, inode);
- g_mutex_unlock(fs->lock);
- return;
- }
-
if (bluesky_verbose) {
g_log("bluesky/flushd", G_LOG_LEVEL_DEBUG,
"Starting flush of inode %"PRIu64, inode->inum);
}
- inode->change_pending = inode->change_count;
-
- /* Create a store barrier. All operations part of the writeback will be
- * added to this barrier, so when the barrier completes we know that the
- * writeback is finished. */
- BlueSkyStoreAsync *barrier = bluesky_store_async_new(fs->store);
- barrier->op = STORE_OP_BARRIER;
-
- bluesky_inode_start_sync(inode, barrier);
- bluesky_store_async_add_notifier(barrier, writeback_complete, inode);
- bluesky_store_async_submit(barrier);
- bluesky_store_async_unref(barrier);
+ bluesky_inode_start_sync(inode);
}
/* Try to flush dirty data to disk, either due to memory pressure or due to
bluesky_file_drop_cached(inode);
}
-/* Drop clean data fromt the cache if needed due to memory pressure. */
+/* Drop clean data from the cache if needed due to memory pressure. */
static void flushd_clean(BlueSkyFS *fs)
{
g_mutex_lock(fs->lock);
inode_count = 1;
while (inode_count-- > 0) {
+#if 0
if (g_atomic_int_get(&fs->cache_total) < bluesky_watermark_medium_total)
break;
+#endif
BlueSkyInode *inode;
if (fs->accessed_list.prev == NULL)