io context sharing: preliminary support
Detach task state from ioc, instead keep track of how many processes are accessing the ioc. Signed-off-by: Jens Axboe <jens.axboe@oracle.com>
This commit is contained in:
parent
fd0928df98
commit
d38ecf935f
5 changed files with 49 additions and 23 deletions
|
@ -3854,12 +3854,13 @@ int __init blk_dev_init(void)
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* IO Context helper functions
|
* IO Context helper functions. put_io_context() returns 1 if there are no
|
||||||
|
* more users of this io context, 0 otherwise.
|
||||||
*/
|
*/
|
||||||
void put_io_context(struct io_context *ioc)
|
int put_io_context(struct io_context *ioc)
|
||||||
{
|
{
|
||||||
if (ioc == NULL)
|
if (ioc == NULL)
|
||||||
return;
|
return 1;
|
||||||
|
|
||||||
BUG_ON(atomic_read(&ioc->refcount) == 0);
|
BUG_ON(atomic_read(&ioc->refcount) == 0);
|
||||||
|
|
||||||
|
@ -3878,7 +3879,9 @@ void put_io_context(struct io_context *ioc)
|
||||||
rcu_read_unlock();
|
rcu_read_unlock();
|
||||||
|
|
||||||
kmem_cache_free(iocontext_cachep, ioc);
|
kmem_cache_free(iocontext_cachep, ioc);
|
||||||
|
return 1;
|
||||||
}
|
}
|
||||||
|
return 0;
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(put_io_context);
|
EXPORT_SYMBOL(put_io_context);
|
||||||
|
|
||||||
|
@ -3893,15 +3896,17 @@ void exit_io_context(void)
|
||||||
current->io_context = NULL;
|
current->io_context = NULL;
|
||||||
task_unlock(current);
|
task_unlock(current);
|
||||||
|
|
||||||
ioc->task = NULL;
|
if (atomic_dec_and_test(&ioc->nr_tasks)) {
|
||||||
if (ioc->aic && ioc->aic->exit)
|
if (ioc->aic && ioc->aic->exit)
|
||||||
ioc->aic->exit(ioc->aic);
|
ioc->aic->exit(ioc->aic);
|
||||||
if (ioc->cic_root.rb_node != NULL) {
|
if (ioc->cic_root.rb_node != NULL) {
|
||||||
cic = rb_entry(rb_first(&ioc->cic_root), struct cfq_io_context, rb_node);
|
cic = rb_entry(rb_first(&ioc->cic_root),
|
||||||
cic->exit(ioc);
|
struct cfq_io_context, rb_node);
|
||||||
}
|
cic->exit(ioc);
|
||||||
|
}
|
||||||
|
|
||||||
put_io_context(ioc);
|
put_io_context(ioc);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
struct io_context *alloc_io_context(gfp_t gfp_flags, int node)
|
struct io_context *alloc_io_context(gfp_t gfp_flags, int node)
|
||||||
|
@ -3911,7 +3916,8 @@ struct io_context *alloc_io_context(gfp_t gfp_flags, int node)
|
||||||
ret = kmem_cache_alloc_node(iocontext_cachep, gfp_flags, node);
|
ret = kmem_cache_alloc_node(iocontext_cachep, gfp_flags, node);
|
||||||
if (ret) {
|
if (ret) {
|
||||||
atomic_set(&ret->refcount, 1);
|
atomic_set(&ret->refcount, 1);
|
||||||
ret->task = current;
|
atomic_set(&ret->nr_tasks, 1);
|
||||||
|
spin_lock_init(&ret->lock);
|
||||||
ret->ioprio_changed = 0;
|
ret->ioprio_changed = 0;
|
||||||
ret->ioprio = 0;
|
ret->ioprio = 0;
|
||||||
ret->last_waited = jiffies; /* doesn't matter... */
|
ret->last_waited = jiffies; /* doesn't matter... */
|
||||||
|
@ -3959,10 +3965,18 @@ static struct io_context *current_io_context(gfp_t gfp_flags, int node)
|
||||||
*/
|
*/
|
||||||
struct io_context *get_io_context(gfp_t gfp_flags, int node)
|
struct io_context *get_io_context(gfp_t gfp_flags, int node)
|
||||||
{
|
{
|
||||||
struct io_context *ret;
|
struct io_context *ret = NULL;
|
||||||
ret = current_io_context(gfp_flags, node);
|
|
||||||
if (likely(ret))
|
/*
|
||||||
atomic_inc(&ret->refcount);
|
* Check for unlikely race with exiting task. ioc ref count is
|
||||||
|
* zero when ioc is being detached.
|
||||||
|
*/
|
||||||
|
do {
|
||||||
|
ret = current_io_context(gfp_flags, node);
|
||||||
|
if (unlikely(!ret))
|
||||||
|
break;
|
||||||
|
} while (!atomic_inc_not_zero(&ret->refcount));
|
||||||
|
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(get_io_context);
|
EXPORT_SYMBOL(get_io_context);
|
||||||
|
|
|
@ -54,7 +54,6 @@ static int set_task_ioprio(struct task_struct *task, int ioprio)
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
task->io_context = ioc;
|
task->io_context = ioc;
|
||||||
ioc->task = task;
|
|
||||||
} while (1);
|
} while (1);
|
||||||
|
|
||||||
if (!err) {
|
if (!err) {
|
||||||
|
|
|
@ -34,7 +34,7 @@ struct sg_io_hdr;
|
||||||
#define BLKDEV_MIN_RQ 4
|
#define BLKDEV_MIN_RQ 4
|
||||||
#define BLKDEV_MAX_RQ 128 /* Default maximum */
|
#define BLKDEV_MAX_RQ 128 /* Default maximum */
|
||||||
|
|
||||||
void put_io_context(struct io_context *ioc);
|
int put_io_context(struct io_context *ioc);
|
||||||
void exit_io_context(void);
|
void exit_io_context(void);
|
||||||
struct io_context *get_io_context(gfp_t gfp_flags, int node);
|
struct io_context *get_io_context(gfp_t gfp_flags, int node);
|
||||||
struct io_context *alloc_io_context(gfp_t gfp_flags, int node);
|
struct io_context *alloc_io_context(gfp_t gfp_flags, int node);
|
||||||
|
|
|
@ -54,13 +54,15 @@ struct cfq_io_context {
|
||||||
};
|
};
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* This is the per-process I/O subsystem state. It is refcounted and
|
* I/O subsystem state of the associated processes. It is refcounted
|
||||||
* kmalloc'ed. Currently all fields are modified in process io context
|
* and kmalloc'ed. These could be shared between processes.
|
||||||
* (apart from the atomic refcount), so require no locking.
|
|
||||||
*/
|
*/
|
||||||
struct io_context {
|
struct io_context {
|
||||||
atomic_t refcount;
|
atomic_t refcount;
|
||||||
struct task_struct *task;
|
atomic_t nr_tasks;
|
||||||
|
|
||||||
|
/* all the fields below are protected by this lock */
|
||||||
|
spinlock_t lock;
|
||||||
|
|
||||||
unsigned short ioprio;
|
unsigned short ioprio;
|
||||||
unsigned short ioprio_changed;
|
unsigned short ioprio_changed;
|
||||||
|
@ -76,4 +78,16 @@ struct io_context {
|
||||||
void *ioc_data;
|
void *ioc_data;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
static inline struct io_context *ioc_task_link(struct io_context *ioc)
|
||||||
|
{
|
||||||
|
/*
|
||||||
|
* if ref count is zero, don't allow sharing (ioc is going away, it's
|
||||||
|
* a race).
|
||||||
|
*/
|
||||||
|
if (ioc && atomic_inc_not_zero(&ioc->refcount))
|
||||||
|
return ioc;
|
||||||
|
|
||||||
|
return NULL;
|
||||||
|
}
|
||||||
|
|
||||||
#endif
|
#endif
|
||||||
|
|
|
@ -805,7 +805,6 @@ static int copy_io(struct task_struct *tsk)
|
||||||
if (unlikely(!tsk->io_context))
|
if (unlikely(!tsk->io_context))
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
tsk->io_context->task = tsk;
|
|
||||||
tsk->io_context->ioprio = ioc->ioprio;
|
tsk->io_context->ioprio = ioc->ioprio;
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
Loading…
Reference in a new issue