projects
/
firefly-linux-kernel-4.4.55.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
Merge remote-tracking branch 'lsk/v3.10/topic/coresight' into linux-linaro-lsk
[firefly-linux-kernel-4.4.55.git]
/
fs
/
aio.c
diff --git
a/fs/aio.c
b/fs/aio.c
index 2bbcacf74d0c64f8814cb68d8c4eff423dfe98d4..ded94c4fa30d31a6997625d007548d119c2aa75d 100644
(file)
--- a/
fs/aio.c
+++ b/
fs/aio.c
@@
-310,7
+310,6
@@
static void free_ioctx(struct kioctx *ctx)
avail = (head <= ctx->tail ? ctx->tail : ctx->nr_events) - head;
avail = (head <= ctx->tail ? ctx->tail : ctx->nr_events) - head;
- atomic_sub(avail, &ctx->reqs_active);
head += avail;
head %= ctx->nr_events;
}
head += avail;
head %= ctx->nr_events;
}
@@
-423,10
+422,12
@@
static void kill_ioctx_rcu(struct rcu_head *head)
* when the processes owning a context have all exited to encourage
* the rapid destruction of the kioctx.
*/
* when the processes owning a context have all exited to encourage
* the rapid destruction of the kioctx.
*/
-static void kill_ioctx(struct kioctx *ctx)
+static void kill_ioctx(struct
mm_struct *mm, struct
kioctx *ctx)
{
if (!atomic_xchg(&ctx->dead, 1)) {
{
if (!atomic_xchg(&ctx->dead, 1)) {
+ spin_lock(&mm->ioctx_lock);
hlist_del_rcu(&ctx->list);
hlist_del_rcu(&ctx->list);
+ spin_unlock(&mm->ioctx_lock);
/*
* It'd be more correct to do this in free_ioctx(), after all
/*
* It'd be more correct to do this in free_ioctx(), after all
@@
-494,7
+495,7
@@
void exit_aio(struct mm_struct *mm)
*/
ctx->mmap_size = 0;
*/
ctx->mmap_size = 0;
- kill_ioctx(ctx);
+ kill_ioctx(
mm,
ctx);
}
}
}
}
@@
-676,6
+677,7
@@
void aio_complete(struct kiocb *iocb, long res, long res2)
put_rq:
/* everything turned out well, dispose of the aiocb. */
aio_put_req(iocb);
put_rq:
/* everything turned out well, dispose of the aiocb. */
aio_put_req(iocb);
+ atomic_dec(&ctx->reqs_active);
/*
* We have to order our ring_info tail store above and test
/*
* We have to order our ring_info tail store above and test
@@
-715,6
+717,8
@@
static long aio_read_events_ring(struct kioctx *ctx,
if (head == ctx->tail)
goto out;
if (head == ctx->tail)
goto out;
+ head %= ctx->nr_events;
+
while (ret < nr) {
long avail;
struct io_event *ev;
while (ret < nr) {
long avail;
struct io_event *ev;
@@
-753,8
+757,6
@@
static long aio_read_events_ring(struct kioctx *ctx,
flush_dcache_page(ctx->ring_pages[0]);
pr_debug("%li h%u t%u\n", ret, head, ctx->tail);
flush_dcache_page(ctx->ring_pages[0]);
pr_debug("%li h%u t%u\n", ret, head, ctx->tail);
-
- atomic_sub(ret, &ctx->reqs_active);
out:
mutex_unlock(&ctx->ring_lock);
out:
mutex_unlock(&ctx->ring_lock);
@@
-852,7
+854,7
@@
SYSCALL_DEFINE2(io_setup, unsigned, nr_events, aio_context_t __user *, ctxp)
if (!IS_ERR(ioctx)) {
ret = put_user(ioctx->user_id, ctxp);
if (ret)
if (!IS_ERR(ioctx)) {
ret = put_user(ioctx->user_id, ctxp);
if (ret)
- kill_ioctx(ioctx);
+ kill_ioctx(
current->mm,
ioctx);
put_ioctx(ioctx);
}
put_ioctx(ioctx);
}
@@
-870,7
+872,7
@@
SYSCALL_DEFINE1(io_destroy, aio_context_t, ctx)
{
struct kioctx *ioctx = lookup_ioctx(ctx);
if (likely(NULL != ioctx)) {
{
struct kioctx *ioctx = lookup_ioctx(ctx);
if (likely(NULL != ioctx)) {
- kill_ioctx(ioctx);
+ kill_ioctx(
current->mm,
ioctx);
put_ioctx(ioctx);
return 0;
}
put_ioctx(ioctx);
return 0;
}