/*
* simple init on struct btrfs_workers
*/
-void btrfs_init_workers(struct btrfs_workers *workers, int max)
+void btrfs_init_workers(struct btrfs_workers *workers, char *name, int max)
{
workers->num_workers = 0;
INIT_LIST_HEAD(&workers->worker_list);
spin_lock_init(&workers->lock);
workers->max_workers = max;
workers->idle_thresh = 32;
+ workers->name = name;
}
/*
INIT_LIST_HEAD(&worker->worker_list);
spin_lock_init(&worker->lock);
atomic_set(&worker->num_pending, 0);
- worker->task = kthread_run(worker_loop, worker, "btrfs");
+ worker->task = kthread_run(worker_loop, worker,
+ "btrfs-%s-%d", workers->name,
+ workers->num_workers + i);
worker->workers = workers;
if (IS_ERR(worker->task)) {
kfree(worker);
/* lock for finding the next worker thread to queue on */
spinlock_t lock;
+
+ /* extra name for this worker */
+ char *name;
};
int btrfs_queue_worker(struct btrfs_workers *workers, struct btrfs_work *work);
int btrfs_start_workers(struct btrfs_workers *workers, int num_workers);
int btrfs_stop_workers(struct btrfs_workers *workers);
-void btrfs_init_workers(struct btrfs_workers *workers, int max);
+void btrfs_init_workers(struct btrfs_workers *workers, char *name, int max);
int btrfs_requeue_work(struct btrfs_work *work);
#endif
/*
* when we're called for a write, we're already in the async
- * submission context. Just jump ingo btrfs_map_bio
+ * submission context. Just jump into btrfs_map_bio
*/
if (rw & (1 << BIO_RW)) {
return btrfs_map_bio(BTRFS_I(inode)->root, rw, bio,
- mirror_num, 0);
+ mirror_num, 1);
}
/*
{
struct extent_io_tree *tree;
tree = &BTRFS_I(page->mapping->host)->io_tree;
+
+ if (current->flags & PF_MEMALLOC) {
+ redirty_page_for_writepage(wbc, page);
+ unlock_page(page);
+ return 0;
+ }
return extent_write_full_page(tree, page, btree_get_extent, wbc);
}
* queue work function gets called at interrupt time, and so it
* cannot dynamically grow.
*/
- btrfs_init_workers(&fs_info->workers, fs_info->thread_pool_size);
- btrfs_init_workers(&fs_info->submit_workers,
+ btrfs_init_workers(&fs_info->workers, "worker",
+ fs_info->thread_pool_size);
+ btrfs_init_workers(&fs_info->submit_workers, "submit",
min_t(u64, fs_devices->num_devices,
fs_info->thread_pool_size));
*/
fs_info->submit_workers.idle_thresh = 64;
- btrfs_init_workers(&fs_info->fixup_workers, 1);
- btrfs_init_workers(&fs_info->endio_workers, fs_info->thread_pool_size);
- btrfs_init_workers(&fs_info->endio_write_workers,
+ btrfs_init_workers(&fs_info->fixup_workers, "fixup", 1);
+ btrfs_init_workers(&fs_info->endio_workers, "endio",
+ fs_info->thread_pool_size);
+ btrfs_init_workers(&fs_info->endio_write_workers, "endio-write",
fs_info->thread_pool_size);
/*