Message ID | 20180321074032.14211-7-hch@lst.de |
---|---|
State | Not Applicable, archived |
Delegated to: | David Miller |
Headers | show |
Series | [01/28] fs: unexport poll_schedule_timeout | expand |
On Wed, Mar 21, 2018 at 08:40:10AM +0100, Christoph Hellwig wrote: > Simple one-shot poll through the io_submit() interface. To poll for > a file descriptor the application should submit an iocb of type > IOCB_CMD_POLL. It will poll the fd for the events specified in the > the first 32 bits of the aio_buf field of the iocb. > > Unlike poll or epoll without EPOLLONESHOT this interface always works > in one shot mode, that is once the iocb is completed, it will have to be > resubmitted. > > Signed-off-by: Christoph Hellwig <hch@lst.de> > Acked-by: Jeff Moyer <jmoyer@redhat.com> Reviewed-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
On Wed, Mar 21, 2018 at 08:40:10AM +0100, Christoph Hellwig wrote: > Simple one-shot poll through the io_submit() interface. To poll for > a file descriptor the application should submit an iocb of type > IOCB_CMD_POLL. It will poll the fd for the events specified in the > the first 32 bits of the aio_buf field of the iocb. > > Unlike poll or epoll without EPOLLONESHOT this interface always works > in one shot mode, that is once the iocb is completed, it will have to be > resubmitted. > > Signed-off-by: Christoph Hellwig <hch@lst.de> > Acked-by: Jeff Moyer <jmoyer@redhat.com> Looks ok, Reviewed-by: Darrick J. Wong <darrick.wong@oracle.com> After this point in the series my familiarity with network code and sockets drops way off, so please don't be too surprised if I don't get any further. --D > --- > fs/aio.c | 102 ++++++++++++++++++++++++++++++++++++++++++- > include/uapi/linux/aio_abi.h | 6 +-- > 2 files changed, 103 insertions(+), 5 deletions(-) > > diff --git a/fs/aio.c b/fs/aio.c > index 79d3eb3d2dd9..38b408129697 100644 > --- a/fs/aio.c > +++ b/fs/aio.c > @@ -5,6 +5,7 @@ > * Implements an efficient asynchronous io interface. > * > * Copyright 2000, 2001, 2002 Red Hat, Inc. All Rights Reserved. > + * Copyright 2018 Christoph Hellwig. > * > * See ../COPYING for licensing terms. > */ > @@ -162,10 +163,18 @@ struct fsync_iocb { > bool datasync; > }; > > +struct poll_iocb { > + struct file *file; > + __poll_t events; > + struct wait_queue_head *head; > + struct wait_queue_entry wait; > +}; > + > struct aio_kiocb { > union { > struct kiocb rw; > struct fsync_iocb fsync; > + struct poll_iocb poll; > }; > > struct kioctx *ki_ctx; > @@ -1590,7 +1599,6 @@ static int aio_fsync(struct fsync_iocb *req, struct iocb *iocb, bool datasync) > return -EINVAL; > if (iocb->aio_offset || iocb->aio_nbytes || iocb->aio_rw_flags) > return -EINVAL; > - > req->file = fget(iocb->aio_fildes); > if (unlikely(!req->file)) > return -EBADF; > @@ -1609,6 +1617,96 @@ static int aio_fsync(struct fsync_iocb *req, struct iocb *iocb, bool datasync) > return ret; > } > > +static void __aio_complete_poll(struct poll_iocb *req, __poll_t mask) > +{ > + fput(req->file); > + aio_complete(container_of(req, struct aio_kiocb, poll), > + mangle_poll(mask), 0); > +} > + > +static void aio_complete_poll(struct poll_iocb *req, __poll_t mask) > +{ > + struct aio_kiocb *iocb = container_of(req, struct aio_kiocb, poll); > + > + if (!(iocb->flags & AIO_IOCB_CANCELLED)) > + __aio_complete_poll(req, mask); > +} > + > +static int aio_poll_cancel(struct kiocb *rw) > +{ > + struct aio_kiocb *iocb = container_of(rw, struct aio_kiocb, rw); > + > + remove_wait_queue(iocb->poll.head, &iocb->poll.wait); > + __aio_complete_poll(&iocb->poll, 0); /* no events to report */ > + return 0; > +} > + > +static int aio_poll_wake(struct wait_queue_entry *wait, unsigned mode, int sync, > + void *key) > +{ > + struct poll_iocb *req = container_of(wait, struct poll_iocb, wait); > + struct file *file = req->file; > + __poll_t mask = key_to_poll(key); > + > + assert_spin_locked(&req->head->lock); > + > + /* for instances that support it check for an event match first: */ > + if (mask && !(mask & req->events)) > + return 0; > + > + mask = vfs_poll_mask(file, req->events); > + if (!mask) > + return 0; > + > + __remove_wait_queue(req->head, &req->wait); > + aio_complete_poll(req, mask); > + return 1; > +} > + > +static ssize_t aio_poll(struct aio_kiocb *aiocb, struct iocb *iocb) > +{ > + struct poll_iocb *req = &aiocb->poll; > + unsigned long flags; > + __poll_t mask; > + > + /* reject any unknown events outside the normal event mask. */ > + if ((u16)iocb->aio_buf != iocb->aio_buf) > + return -EINVAL; > + /* reject fields that are not defined for poll */ > + if (iocb->aio_offset || iocb->aio_nbytes || iocb->aio_rw_flags) > + return -EINVAL; > + > + req->events = demangle_poll(iocb->aio_buf) | POLLERR | POLLHUP; > + req->file = fget(iocb->aio_fildes); > + if (unlikely(!req->file)) > + return -EBADF; > + > + req->head = vfs_get_poll_head(req->file, req->events); > + if (!req->head) { > + fput(req->file); > + return -EINVAL; /* same as no support for IOCB_CMD_POLL */ > + } > + if (IS_ERR(req->head)) { > + mask = PTR_TO_POLL(req->head); > + goto done; > + } > + > + init_waitqueue_func_entry(&req->wait, aio_poll_wake); > + > + spin_lock_irqsave(&req->head->lock, flags); > + mask = vfs_poll_mask(req->file, req->events); > + if (!mask) { > + __kiocb_set_cancel_fn(aiocb, aio_poll_cancel, > + AIO_IOCB_DELAYED_CANCEL); > + __add_wait_queue(req->head, &req->wait); > + } > + spin_unlock_irqrestore(&req->head->lock, flags); > +done: > + if (mask) > + aio_complete_poll(req, mask); > + return -EIOCBQUEUED; > +} > + > static int io_submit_one(struct kioctx *ctx, struct iocb __user *user_iocb, > struct iocb *iocb, bool compat) > { > @@ -1677,6 +1775,8 @@ static int io_submit_one(struct kioctx *ctx, struct iocb __user *user_iocb, > break; > case IOCB_CMD_FDSYNC: > ret = aio_fsync(&req->fsync, iocb, true); > + case IOCB_CMD_POLL: > + ret = aio_poll(req, iocb); > break; > default: > pr_debug("invalid aio operation %d\n", iocb->aio_lio_opcode); > diff --git a/include/uapi/linux/aio_abi.h b/include/uapi/linux/aio_abi.h > index 2c0a3415beee..ed0185945bb2 100644 > --- a/include/uapi/linux/aio_abi.h > +++ b/include/uapi/linux/aio_abi.h > @@ -39,10 +39,8 @@ enum { > IOCB_CMD_PWRITE = 1, > IOCB_CMD_FSYNC = 2, > IOCB_CMD_FDSYNC = 3, > - /* These two are experimental. > - * IOCB_CMD_PREADX = 4, > - * IOCB_CMD_POLL = 5, > - */ > + /* 4 was the experimental IOCB_CMD_PREADX */ > + IOCB_CMD_POLL = 5, > IOCB_CMD_NOOP = 6, > IOCB_CMD_PREADV = 7, > IOCB_CMD_PWRITEV = 8, > -- > 2.14.2 >
On Wed, Mar 21, 2018 at 08:40:10AM +0100, Christoph Hellwig wrote: > Simple one-shot poll through the io_submit() interface. To poll for > a file descriptor the application should submit an iocb of type > IOCB_CMD_POLL. It will poll the fd for the events specified in the > the first 32 bits of the aio_buf field of the iocb. > > Unlike poll or epoll without EPOLLONESHOT this interface always works > in one shot mode, that is once the iocb is completed, it will have to be > resubmitted. AFAICS, your wakeup can race with io_cancel(), leading to double fput(). You are checking the "somebody had committed itself to cancelling that thing" bit outside of ->ctx_lock on the wakeup side, and I don't see anything to prevent both getting to __aio_poll_complete() on the same iocb, with obvious results. I might be missing something subtle in there, but then it would be nice to have it covered in commit message...
On Thu, Mar 22, 2018 at 04:52:55PM +0000, Al Viro wrote: > On Wed, Mar 21, 2018 at 08:40:10AM +0100, Christoph Hellwig wrote: > > Simple one-shot poll through the io_submit() interface. To poll for > > a file descriptor the application should submit an iocb of type > > IOCB_CMD_POLL. It will poll the fd for the events specified in the > > the first 32 bits of the aio_buf field of the iocb. > > > > Unlike poll or epoll without EPOLLONESHOT this interface always works > > in one shot mode, that is once the iocb is completed, it will have to be > > resubmitted. > > AFAICS, your wakeup can race with io_cancel(), leading to double fput(). > You are checking the "somebody had committed itself to cancelling that > thing" bit outside of ->ctx_lock on the wakeup side, and I don't see > anything to prevent both getting to __aio_poll_complete() on the same > iocb, with obvious results. True. Probably wants something like this to fix, although for this is entirely untested: diff --git a/fs/aio.c b/fs/aio.c index 38b408129697..66d5cc272617 100644 --- a/fs/aio.c +++ b/fs/aio.c @@ -187,8 +187,9 @@ struct aio_kiocb { * for cancellation */ unsigned int flags; /* protected by ctx->ctx_lock */ -#define AIO_IOCB_DELAYED_CANCEL (1 << 0) -#define AIO_IOCB_CANCELLED (1 << 1) +#define AIO_IOCB_CAN_CANCEL (1 << 0) +#define AIO_IOCB_DELAYED_CANCEL (1 << 1) +#define AIO_IOCB_CANCELLED (1 << 2) /* * If the aio_resfd field of the userspace iocb is not zero, @@ -568,7 +569,7 @@ static void __kiocb_set_cancel_fn(struct aio_kiocb *req, spin_lock_irqsave(&ctx->ctx_lock, flags); list_add_tail(&req->ki_list, &ctx->active_reqs); req->ki_cancel = cancel; - req->flags |= iocb_flags; + req->flags |= (AIO_IOCB_CAN_CANCEL | iocb_flags); spin_unlock_irqrestore(&ctx->ctx_lock, flags); } @@ -1086,22 +1087,30 @@ static struct kioctx *lookup_ioctx(unsigned long ctx_id) return ret; } +#define AIO_COMPLETE_CANCEL (1 << 0) + /* aio_complete * Called when the io request on the given iocb is complete. */ -static void aio_complete(struct aio_kiocb *iocb, long res, long res2) +static bool aio_complete(struct aio_kiocb *iocb, long res, long res2, + unsigned complete_flags) { struct kioctx *ctx = iocb->ki_ctx; struct aio_ring *ring; struct io_event *ev_page, *event; unsigned tail, pos, head; - unsigned long flags; - - if (!list_empty_careful(iocb->ki_list.next)) { - unsigned long flags; + unsigned long flags; + if (iocb->flags & AIO_IOCB_CAN_CANCEL) { spin_lock_irqsave(&ctx->ctx_lock, flags); - list_del(&iocb->ki_list); + if (!(complete_flags & AIO_COMPLETE_CANCEL) && + (iocb->flags & AIO_IOCB_CANCELLED)) { + spin_unlock_irqrestore(&ctx->ctx_lock, flags); + return false; + } + + if (!list_empty(&iocb->ki_list)) + list_del(&iocb->ki_list); spin_unlock_irqrestore(&ctx->ctx_lock, flags); } @@ -1177,6 +1186,7 @@ static void aio_complete(struct aio_kiocb *iocb, long res, long res2) wake_up(&ctx->wait); percpu_ref_put(&ctx->reqs); + return true; } /* aio_read_events_ring @@ -1425,6 +1435,7 @@ SYSCALL_DEFINE1(io_destroy, aio_context_t, ctx) static void aio_complete_rw(struct kiocb *kiocb, long res, long res2) { struct aio_kiocb *iocb = container_of(kiocb, struct aio_kiocb, rw); + struct file *file = kiocb->ki_filp; WARN_ON_ONCE(is_sync_kiocb(kiocb)); @@ -1440,8 +1451,8 @@ static void aio_complete_rw(struct kiocb *kiocb, long res, long res2) file_end_write(kiocb->ki_filp); } - fput(kiocb->ki_filp); - aio_complete(iocb, res, res2); + if (aio_complete(iocb, res, res2, 0)) + fput(file); } static int aio_prep_rw(struct kiocb *req, struct iocb *iocb) @@ -1584,11 +1595,13 @@ static ssize_t aio_write(struct kiocb *req, struct iocb *iocb, bool vectored, static void aio_fsync_work(struct work_struct *work) { struct fsync_iocb *req = container_of(work, struct fsync_iocb, work); + struct aio_kiocb *iocb = container_of(req, struct aio_kiocb, fsync); + struct file *file = req->file; int ret; ret = vfs_fsync(req->file, req->datasync); - fput(req->file); - aio_complete(container_of(req, struct aio_kiocb, fsync), ret, 0); + if (aio_complete(iocb, ret, 0, 0)) + fput(file); } static int aio_fsync(struct fsync_iocb *req, struct iocb *iocb, bool datasync) @@ -1617,27 +1630,23 @@ static int aio_fsync(struct fsync_iocb *req, struct iocb *iocb, bool datasync) return ret; } -static void __aio_complete_poll(struct poll_iocb *req, __poll_t mask) -{ - fput(req->file); - aio_complete(container_of(req, struct aio_kiocb, poll), - mangle_poll(mask), 0); -} - static void aio_complete_poll(struct poll_iocb *req, __poll_t mask) { struct aio_kiocb *iocb = container_of(req, struct aio_kiocb, poll); + struct file *file = req->file; - if (!(iocb->flags & AIO_IOCB_CANCELLED)) - __aio_complete_poll(req, mask); + if (aio_complete(iocb, mangle_poll(mask), 0, 0)) + fput(file); } static int aio_poll_cancel(struct kiocb *rw) { struct aio_kiocb *iocb = container_of(rw, struct aio_kiocb, rw); + struct file *file = iocb->poll.file; remove_wait_queue(iocb->poll.head, &iocb->poll.wait); - __aio_complete_poll(&iocb->poll, 0); /* no events to report */ + if (aio_complete(iocb, 0, 0, AIO_COMPLETE_CANCEL)) + fput(file); return 0; }
On Thu, Mar 22, 2018 at 06:24:10PM +0100, Christoph Hellwig wrote: > -static void aio_complete(struct aio_kiocb *iocb, long res, long res2) > +static bool aio_complete(struct aio_kiocb *iocb, long res, long res2, > + unsigned complete_flags) Looks like all callers are following that with "if returned true, fput(something)". Does it really make any sense to keep that struct file * in different fields? Wait a sec... What ordering do we want for * call(s) of ->ki_complete * call (if any) of ->ki_cancel * dropping reference to struct file and what are the expected call chains for all of those?
On Thu, Mar 22, 2018 at 06:16:53PM +0000, Al Viro wrote: > On Thu, Mar 22, 2018 at 06:24:10PM +0100, Christoph Hellwig wrote: > > > -static void aio_complete(struct aio_kiocb *iocb, long res, long res2) > > +static bool aio_complete(struct aio_kiocb *iocb, long res, long res2, > > + unsigned complete_flags) > > Looks like all callers are following that with "if returned true, > fput(something)". Does it really make any sense to keep that struct > file * in different fields? struct kiocb is used not just for aio, but for our normal read/write_iter APIs, and it is not suitable for poll or fsync. So I can't really find a good way to keep it common except for duplicating it in struct kiocb and strut aio_iocb. But maybe we could pass a struct file argument to aio_complete(). > Wait a sec... What ordering do we want for > * call(s) of ->ki_complete > * call (if any) of ->ki_cancel > * dropping reference to struct file > and what are the expected call chains for all of those? fput must be done exactly once from inside ->ki_complete OR ->ki_cancel in case it did manage to do the actual completion. Reference to struct file isn't needed in aio_complete, but if aio_complete decided who won the race we'll have to put after it (or inside it if we want to make it common)
diff --git a/fs/aio.c b/fs/aio.c index 79d3eb3d2dd9..38b408129697 100644 --- a/fs/aio.c +++ b/fs/aio.c @@ -5,6 +5,7 @@ * Implements an efficient asynchronous io interface. * * Copyright 2000, 2001, 2002 Red Hat, Inc. All Rights Reserved. + * Copyright 2018 Christoph Hellwig. * * See ../COPYING for licensing terms. */ @@ -162,10 +163,18 @@ struct fsync_iocb { bool datasync; }; +struct poll_iocb { + struct file *file; + __poll_t events; + struct wait_queue_head *head; + struct wait_queue_entry wait; +}; + struct aio_kiocb { union { struct kiocb rw; struct fsync_iocb fsync; + struct poll_iocb poll; }; struct kioctx *ki_ctx; @@ -1590,7 +1599,6 @@ static int aio_fsync(struct fsync_iocb *req, struct iocb *iocb, bool datasync) return -EINVAL; if (iocb->aio_offset || iocb->aio_nbytes || iocb->aio_rw_flags) return -EINVAL; - req->file = fget(iocb->aio_fildes); if (unlikely(!req->file)) return -EBADF; @@ -1609,6 +1617,96 @@ static int aio_fsync(struct fsync_iocb *req, struct iocb *iocb, bool datasync) return ret; } +static void __aio_complete_poll(struct poll_iocb *req, __poll_t mask) +{ + fput(req->file); + aio_complete(container_of(req, struct aio_kiocb, poll), + mangle_poll(mask), 0); +} + +static void aio_complete_poll(struct poll_iocb *req, __poll_t mask) +{ + struct aio_kiocb *iocb = container_of(req, struct aio_kiocb, poll); + + if (!(iocb->flags & AIO_IOCB_CANCELLED)) + __aio_complete_poll(req, mask); +} + +static int aio_poll_cancel(struct kiocb *rw) +{ + struct aio_kiocb *iocb = container_of(rw, struct aio_kiocb, rw); + + remove_wait_queue(iocb->poll.head, &iocb->poll.wait); + __aio_complete_poll(&iocb->poll, 0); /* no events to report */ + return 0; +} + +static int aio_poll_wake(struct wait_queue_entry *wait, unsigned mode, int sync, + void *key) +{ + struct poll_iocb *req = container_of(wait, struct poll_iocb, wait); + struct file *file = req->file; + __poll_t mask = key_to_poll(key); + + assert_spin_locked(&req->head->lock); + + /* for instances that support it check for an event match first: */ + if (mask && !(mask & req->events)) + return 0; + + mask = vfs_poll_mask(file, req->events); + if (!mask) + return 0; + + __remove_wait_queue(req->head, &req->wait); + aio_complete_poll(req, mask); + return 1; +} + +static ssize_t aio_poll(struct aio_kiocb *aiocb, struct iocb *iocb) +{ + struct poll_iocb *req = &aiocb->poll; + unsigned long flags; + __poll_t mask; + + /* reject any unknown events outside the normal event mask. */ + if ((u16)iocb->aio_buf != iocb->aio_buf) + return -EINVAL; + /* reject fields that are not defined for poll */ + if (iocb->aio_offset || iocb->aio_nbytes || iocb->aio_rw_flags) + return -EINVAL; + + req->events = demangle_poll(iocb->aio_buf) | POLLERR | POLLHUP; + req->file = fget(iocb->aio_fildes); + if (unlikely(!req->file)) + return -EBADF; + + req->head = vfs_get_poll_head(req->file, req->events); + if (!req->head) { + fput(req->file); + return -EINVAL; /* same as no support for IOCB_CMD_POLL */ + } + if (IS_ERR(req->head)) { + mask = PTR_TO_POLL(req->head); + goto done; + } + + init_waitqueue_func_entry(&req->wait, aio_poll_wake); + + spin_lock_irqsave(&req->head->lock, flags); + mask = vfs_poll_mask(req->file, req->events); + if (!mask) { + __kiocb_set_cancel_fn(aiocb, aio_poll_cancel, + AIO_IOCB_DELAYED_CANCEL); + __add_wait_queue(req->head, &req->wait); + } + spin_unlock_irqrestore(&req->head->lock, flags); +done: + if (mask) + aio_complete_poll(req, mask); + return -EIOCBQUEUED; +} + static int io_submit_one(struct kioctx *ctx, struct iocb __user *user_iocb, struct iocb *iocb, bool compat) { @@ -1677,6 +1775,8 @@ static int io_submit_one(struct kioctx *ctx, struct iocb __user *user_iocb, break; case IOCB_CMD_FDSYNC: ret = aio_fsync(&req->fsync, iocb, true); + case IOCB_CMD_POLL: + ret = aio_poll(req, iocb); break; default: pr_debug("invalid aio operation %d\n", iocb->aio_lio_opcode); diff --git a/include/uapi/linux/aio_abi.h b/include/uapi/linux/aio_abi.h index 2c0a3415beee..ed0185945bb2 100644 --- a/include/uapi/linux/aio_abi.h +++ b/include/uapi/linux/aio_abi.h @@ -39,10 +39,8 @@ enum { IOCB_CMD_PWRITE = 1, IOCB_CMD_FSYNC = 2, IOCB_CMD_FDSYNC = 3, - /* These two are experimental. - * IOCB_CMD_PREADX = 4, - * IOCB_CMD_POLL = 5, - */ + /* 4 was the experimental IOCB_CMD_PREADX */ + IOCB_CMD_POLL = 5, IOCB_CMD_NOOP = 6, IOCB_CMD_PREADV = 7, IOCB_CMD_PWRITEV = 8,