+ g_mutex_lock(inode->lock);
+
+ if (inode->change_count == inode->change_commit) {
+ uint64_t delay = bluesky_get_current_time() - inode->access_time;
+ if (delay >= CACHE_CLEAN_DELAY) {
+ drop_caches(inode);
+
+ /* If the only references are the one we hold and the one in the
+ * filesystem inum->inode hash table... First check the refcount
+ * without the lock for speed, but if the check looks good verify
+ * it after taking the filesystem lock. */
+ if (inode->refcount == 2) {
+ g_mutex_lock(fs->lock);
+ if (inode->refcount == 2) {
+ g_log("bluesky/flushd", G_LOG_LEVEL_DEBUG,
+ "Trying to drop inode %"PRIu64" from cache",
+ inode->inum);
+ if (g_hash_table_remove(fs->inodes, &inode->inum))
+ bluesky_inode_unref(inode);
+ }
+ bluesky_list_unlink(&inode->fs->accessed_list,
+ inode->accessed_list);
+ inode->accessed_list = NULL;
+ bluesky_list_unlink(&inode->fs->dirty_list,
+ inode->dirty_list);
+ inode->dirty_list = NULL;
+ g_mutex_unlock(fs->lock);
+ }
+ }
+
+ g_mutex_unlock(inode->lock);
+ bluesky_inode_unref(inode);
+ return;
+ }
+
+ if (inode->change_pending) {
+ /* Waiting for an earlier writeback to finish, so don't start a new
+ * writeback yet. */
+ g_mutex_unlock(inode->lock);
+ bluesky_inode_unref(inode);
+ return;
+ }
+
+ uint64_t elapsed = bluesky_get_current_time() - inode->change_time;
+ if (elapsed < WRITEBACK_DELAY) {
+ /* Give a bit more time before starting writeback. */
+ g_mutex_unlock(inode->lock);
+ bluesky_inode_unref(inode);
+ return;
+ }
+
+ inode->change_pending = inode->change_count;
+
+ g_log("bluesky/flushd", G_LOG_LEVEL_DEBUG,
+ "Starting flush of inode %"PRIu64, inode->inum);
+
+ /* Create a store barrier. All operations part of the writeback will be
+ * added to this barrier, so when the barrier completes we know that the
+ * writeback is finished. */
+ BlueSkyStoreAsync *barrier = bluesky_store_async_new(fs->store);
+ barrier->op = STORE_OP_BARRIER;
+
+ bluesky_inode_start_sync(inode, barrier);
+
+ bluesky_store_async_add_notifier(barrier, writeback_complete, inode);
+ bluesky_store_async_submit(barrier);
+ bluesky_store_async_unref(barrier);
+
+ g_mutex_unlock(inode->lock);
+ bluesky_inode_unref(inode);
+}
+#endif
+
+static void flushd_dirty_inode(BlueSkyInode *inode)
+{
+ BlueSkyFS *fs = inode->fs;
+
+ /* Inode is clean; nothing to do. */