diff options
author | alc <alc@FreeBSD.org> | 2001-03-10 22:47:57 +0000 |
---|---|---|
committer | alc <alc@FreeBSD.org> | 2001-03-10 22:47:57 +0000 |
commit | d25198ddf60b0b6a8d44e75417de128e29382130 (patch) | |
tree | 09bc7d7119aa29db59690f0236e0d21a77c6c7bb /sys/kern | |
parent | a8d215e3c1112edf88fab0b035dc71f12a2f11d8 (diff) | |
download | FreeBSD-src-d25198ddf60b0b6a8d44e75417de128e29382130.zip FreeBSD-src-d25198ddf60b0b6a8d44e75417de128e29382130.tar.gz |
When aio_read/write() is used on a raw device, physical buffers are
used for up to "vfs.aio.max_buf_aio" of the requests. If a request
size is MAXPHYS, but the request base isn't page aligned, vmapbuf()
will map the end of the user space buffer into the start of the kva
allocated for the next physical buffer. Don't use a physical buffer
in this case. (This change addresses problem report 25617.)
When an aio_read/write() on a raw device has completed, timeout() is
used to schedule a signal to the process. Thus, the reporting is
delayed up to 10 ms (assuming hz is 100). The process might have
terminated in the meantime, causing a trap 12 when attempting to
deliver the signal. Thus, the timeout must be cancelled when removing
the job.
aio jobs in state JOBST_JOBQGLOBAL should be removed from the
kaio_jobqueue list during process rundown.
During process rundown, some aio jobs might move from one list to a
different list that has already been "emptied", causing the rundown to
be incomplete. Retry the rundown.
A call to BUF_KERNPROC() is needed after obtaining a physical buffer
to disassociate the lock from the running process since it can return
to userland without releasing that lock.
PR: 25617
Submitted by: tegge
Diffstat (limited to 'sys/kern')
-rw-r--r-- | sys/kern/vfs_aio.c | 33 |
1 files changed, 26 insertions, 7 deletions
diff --git a/sys/kern/vfs_aio.c b/sys/kern/vfs_aio.c index 8ee7d6d..7e18d89 100644 --- a/sys/kern/vfs_aio.c +++ b/sys/kern/vfs_aio.c @@ -215,6 +215,7 @@ static void aio_physwakeup(struct buf *bp); static int aio_fphysio(struct proc *p, struct aiocblist *aiocbe); static int aio_qphysio(struct proc *p, struct aiocblist *iocb); static void aio_daemon(void *uproc); +static void process_signal(void *aioj); SYSINIT(aio, SI_SUB_VFS, SI_ORDER_ANY, aio_onceonly, NULL); @@ -298,11 +299,11 @@ aio_free_entry(struct aiocblist *aiocbe) if (ki == NULL) panic("aio_free_entry: missing p->p_aioinfo"); - if (aiocbe->jobstate == JOBST_JOBRUNNING) { + while (aiocbe->jobstate == JOBST_JOBRUNNING) { if (aiocbe->jobflags & AIOCBLIST_ASYNCFREE) return 0; aiocbe->jobflags |= AIOCBLIST_RUNDOWN; - tsleep(aiocbe, PRIBIO|PCATCH, "jobwai", 0); + tsleep(aiocbe, PRIBIO, "jobwai", 0); } aiocbe->jobflags &= ~AIOCBLIST_ASYNCFREE; @@ -353,9 +354,10 @@ aio_free_entry(struct aiocblist *aiocbe) } else if (aiocbe->jobstate == JOBST_JOBQPROC) { aiop = aiocbe->jobaioproc; TAILQ_REMOVE(&aiop->jobtorun, aiocbe, list); - } else if (aiocbe->jobstate == JOBST_JOBQGLOBAL) + } else if (aiocbe->jobstate == JOBST_JOBQGLOBAL) { TAILQ_REMOVE(&aio_jobs, aiocbe, list); - else if (aiocbe->jobstate == JOBST_JOBFINISHED) + TAILQ_REMOVE(&ki->kaio_jobqueue, aiocbe, plist); + } else if (aiocbe->jobstate == JOBST_JOBFINISHED) TAILQ_REMOVE(&ki->kaio_jobdone, aiocbe, plist); else if (aiocbe->jobstate == JOBST_JOBBFINISHED) { s = splbio(); @@ -372,6 +374,7 @@ aio_free_entry(struct aiocblist *aiocbe) zfree(aiolio_zone, lj); } aiocbe->jobstate = JOBST_NULL; + untimeout(process_signal, aiocbe, aiocbe->timeouthandle); zfree(aiocb_zone, aiocbe); return 0; } @@ -481,6 +484,16 @@ restart4: } splx(s); + /* + * If we've slept, jobs might have moved from one queue to another. + * Retry rundown if we didn't manage to empty the queues. + */ + if (TAILQ_FIRST(&ki->kaio_jobdone) != NULL || + TAILQ_FIRST(&ki->kaio_jobqueue) != NULL || + TAILQ_FIRST(&ki->kaio_bufqueue) != NULL || + TAILQ_FIRST(&ki->kaio_bufdone) != NULL) + goto restart1; + for (lj = TAILQ_FIRST(&ki->kaio_liojoblist); lj; lj = ljn) { ljn = TAILQ_NEXT(lj, lioj_list); if ((lj->lioj_buffer_count == 0) && (lj->lioj_queue_count == @@ -990,7 +1003,8 @@ aio_qphysio(struct proc *p, struct aiocblist *aiocbe) if (cb->aio_nbytes % vp->v_rdev->si_bsize_phys) return (-1); - if (cb->aio_nbytes > MAXPHYS) + if (cb->aio_nbytes > + MAXPHYS - (((vm_offset_t) cb->aio_buf) & PAGE_MASK)) return (-1); ki = p->p_aioinfo; @@ -1005,6 +1019,7 @@ aio_qphysio(struct proc *p, struct aiocblist *aiocbe) /* Create and build a buffer header for a transfer. */ bp = (struct buf *)getpbuf(NULL); + BUF_KERNPROC(bp); /* * Get a copy of the kva from the physical buffer. @@ -1209,6 +1224,7 @@ _aio_aqueue(struct proc *p, struct aiocb *job, struct aio_liojob *lj, int type) aiocbe = zalloc(aiocb_zone); aiocbe->inputcharge = 0; aiocbe->outputcharge = 0; + callout_handle_init(&aiocbe->timeouthandle); SLIST_INIT(&aiocbe->klist); suword(&job->_aiocb_private.status, -1); @@ -2116,7 +2132,9 @@ aio_physwakeup(struct buf *bp) (LIOJ_SIGNAL|LIOJ_SIGNAL_POSTED)) == LIOJ_SIGNAL) { lj->lioj_flags |= LIOJ_SIGNAL_POSTED; - timeout(process_signal, aiocbe, 0); + aiocbe->timeouthandle = + timeout(process_signal, + aiocbe, 0); } } } @@ -2137,7 +2155,8 @@ aio_physwakeup(struct buf *bp) } if (aiocbe->uaiocb.aio_sigevent.sigev_notify == SIGEV_SIGNAL) - timeout(process_signal, aiocbe, 0); + aiocbe->timeouthandle = + timeout(process_signal, aiocbe, 0); } } #endif /* VFS_AIO */ |