1 1.50 riastrad /* $NetBSD: sys_aio.c,v 1.50 2024/12/07 02:38:51 riastradh Exp $ */ 2 1.1 rmind 3 1.1 rmind /* 4 1.31 rmind * Copyright (c) 2007 Mindaugas Rasiukevicius <rmind at NetBSD org> 5 1.10 rmind * All rights reserved. 6 1.49 riastrad * 7 1.1 rmind * Redistribution and use in source and binary forms, with or without 8 1.1 rmind * modification, are permitted provided that the following conditions 9 1.1 rmind * are met: 10 1.1 rmind * 1. Redistributions of source code must retain the above copyright 11 1.1 rmind * notice, this list of conditions and the following disclaimer. 12 1.1 rmind * 2. Redistributions in binary form must reproduce the above copyright 13 1.1 rmind * notice, this list of conditions and the following disclaimer in the 14 1.1 rmind * documentation and/or other materials provided with the distribution. 15 1.1 rmind * 16 1.19 rmind * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 17 1.19 rmind * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 18 1.19 rmind * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 19 1.19 rmind * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE 20 1.19 rmind * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 21 1.19 rmind * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 22 1.19 rmind * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 23 1.19 rmind * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 24 1.19 rmind * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 25 1.19 rmind * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 26 1.19 rmind * SUCH DAMAGE. 27 1.1 rmind */ 28 1.1 rmind 29 1.1 rmind /* 30 1.19 rmind * Implementation of POSIX asynchronous I/O. 31 1.19 rmind * Defined in the Base Definitions volume of IEEE Std 1003.1-2001. 32 1.1 rmind */ 33 1.1 rmind 34 1.1 rmind #include <sys/cdefs.h> 35 1.50 riastrad __KERNEL_RCSID(0, "$NetBSD: sys_aio.c,v 1.50 2024/12/07 02:38:51 riastradh Exp $"); 36 1.4 rmind 37 1.20 ad #ifdef _KERNEL_OPT 38 1.4 rmind #include "opt_ddb.h" 39 1.20 ad #endif 40 1.1 rmind 41 1.1 rmind #include <sys/param.h> 42 1.49 riastrad #include <sys/types.h> 43 1.49 riastrad 44 1.49 riastrad #include <sys/atomic.h> 45 1.49 riastrad #include <sys/buf.h> 46 1.1 rmind #include <sys/condvar.h> 47 1.1 rmind #include <sys/file.h> 48 1.1 rmind #include <sys/filedesc.h> 49 1.1 rmind #include <sys/kernel.h> 50 1.1 rmind #include <sys/kmem.h> 51 1.1 rmind #include <sys/lwp.h> 52 1.49 riastrad #include <sys/module.h> 53 1.1 rmind #include <sys/mutex.h> 54 1.1 rmind #include <sys/pool.h> 55 1.1 rmind #include <sys/proc.h> 56 1.1 rmind #include <sys/queue.h> 57 1.50 riastrad #include <sys/sdt.h> 58 1.1 rmind #include <sys/signal.h> 59 1.1 rmind #include <sys/signalvar.h> 60 1.20 ad #include <sys/syscall.h> 61 1.1 rmind #include <sys/syscallargs.h> 62 1.20 ad #include <sys/syscallvar.h> 63 1.1 rmind #include <sys/sysctl.h> 64 1.1 rmind #include <sys/systm.h> 65 1.1 rmind #include <sys/types.h> 66 1.1 rmind #include <sys/vnode.h> 67 1.1 rmind 68 1.1 rmind #include <uvm/uvm_extern.h> 69 1.1 rmind 70 1.20 ad MODULE(MODULE_CLASS_MISC, aio, NULL); 71 1.20 ad 72 1.1 rmind /* 73 1.1 rmind * System-wide limits and counter of AIO operations. 74 1.1 rmind */ 75 1.31 rmind u_int aio_listio_max = AIO_LISTIO_MAX; 76 1.31 rmind static u_int aio_max = AIO_MAX; 77 1.31 rmind static u_int aio_jobs_count; 78 1.31 rmind 79 1.31 rmind static struct pool aio_job_pool; 80 1.31 rmind static struct pool aio_lio_pool; 81 1.31 rmind static void * aio_ehook; 82 1.31 rmind 83 1.35 jruoho static void aio_worker(void *); 84 1.35 jruoho static void aio_process(struct aio_job *); 85 1.35 jruoho static void aio_sendsig(struct proc *, struct sigevent *); 86 1.35 jruoho static int aio_enqueue_job(int, void *, struct lio_req *); 87 1.35 jruoho static void aio_exit(proc_t *, void *); 88 1.35 jruoho 89 1.35 jruoho static int sysctl_aio_listio_max(SYSCTLFN_PROTO); 90 1.35 jruoho static int sysctl_aio_max(SYSCTLFN_PROTO); 91 1.20 ad 92 1.20 ad static const struct syscall_package aio_syscalls[] = { 93 1.20 ad { SYS_aio_cancel, 0, (sy_call_t *)sys_aio_cancel }, 94 1.20 ad { SYS_aio_error, 0, (sy_call_t *)sys_aio_error }, 95 1.20 ad { SYS_aio_fsync, 0, (sy_call_t *)sys_aio_fsync }, 96 1.20 ad { SYS_aio_read, 0, (sy_call_t *)sys_aio_read }, 97 1.20 ad { SYS_aio_return, 0, (sy_call_t *)sys_aio_return }, 98 1.22 christos { SYS___aio_suspend50, 0, (sy_call_t *)sys___aio_suspend50 }, 99 1.20 ad { SYS_aio_write, 0, (sy_call_t *)sys_aio_write }, 100 1.20 ad { SYS_lio_listio, 0, (sy_call_t *)sys_lio_listio }, 101 1.20 ad { 0, 0, NULL }, 102 1.20 ad }; 103 1.1 rmind 104 1.1 rmind /* 105 1.20 ad * Tear down all AIO state. 106 1.4 rmind */ 107 1.20 ad static int 108 1.20 ad aio_fini(bool interface) 109 1.20 ad { 110 1.20 ad int error; 111 1.20 ad proc_t *p; 112 1.20 ad 113 1.20 ad if (interface) { 114 1.20 ad /* Stop syscall activity. */ 115 1.20 ad error = syscall_disestablish(NULL, aio_syscalls); 116 1.20 ad if (error != 0) 117 1.20 ad return error; 118 1.20 ad /* Abort if any processes are using AIO. */ 119 1.48 ad mutex_enter(&proc_lock); 120 1.20 ad PROCLIST_FOREACH(p, &allproc) { 121 1.20 ad if (p->p_aio != NULL) 122 1.20 ad break; 123 1.20 ad } 124 1.48 ad mutex_exit(&proc_lock); 125 1.20 ad if (p != NULL) { 126 1.20 ad error = syscall_establish(NULL, aio_syscalls); 127 1.20 ad KASSERT(error == 0); 128 1.50 riastrad return SET_ERROR(EBUSY); 129 1.20 ad } 130 1.20 ad } 131 1.35 jruoho 132 1.20 ad KASSERT(aio_jobs_count == 0); 133 1.20 ad exithook_disestablish(aio_ehook); 134 1.20 ad pool_destroy(&aio_job_pool); 135 1.20 ad pool_destroy(&aio_lio_pool); 136 1.20 ad return 0; 137 1.20 ad } 138 1.20 ad 139 1.20 ad /* 140 1.20 ad * Initialize global AIO state. 141 1.20 ad */ 142 1.20 ad static int 143 1.20 ad aio_init(void) 144 1.4 rmind { 145 1.20 ad int error; 146 1.4 rmind 147 1.4 rmind pool_init(&aio_job_pool, sizeof(struct aio_job), 0, 0, 0, 148 1.4 rmind "aio_jobs_pool", &pool_allocator_nointr, IPL_NONE); 149 1.4 rmind pool_init(&aio_lio_pool, sizeof(struct lio_req), 0, 0, 0, 150 1.4 rmind "aio_lio_pool", &pool_allocator_nointr, IPL_NONE); 151 1.20 ad aio_ehook = exithook_establish(aio_exit, NULL); 152 1.35 jruoho 153 1.20 ad error = syscall_establish(NULL, aio_syscalls); 154 1.20 ad if (error != 0) 155 1.35 jruoho (void)aio_fini(false); 156 1.20 ad return error; 157 1.20 ad } 158 1.20 ad 159 1.20 ad /* 160 1.20 ad * Module interface. 161 1.20 ad */ 162 1.20 ad static int 163 1.20 ad aio_modcmd(modcmd_t cmd, void *arg) 164 1.20 ad { 165 1.20 ad 166 1.20 ad switch (cmd) { 167 1.20 ad case MODULE_CMD_INIT: 168 1.20 ad return aio_init(); 169 1.20 ad case MODULE_CMD_FINI: 170 1.20 ad return aio_fini(true); 171 1.20 ad default: 172 1.50 riastrad return SET_ERROR(ENOTTY); 173 1.20 ad } 174 1.4 rmind } 175 1.4 rmind 176 1.4 rmind /* 177 1.1 rmind * Initialize Asynchronous I/O data structures for the process. 178 1.1 rmind */ 179 1.20 ad static int 180 1.20 ad aio_procinit(struct proc *p) 181 1.1 rmind { 182 1.1 rmind struct aioproc *aio; 183 1.1 rmind struct lwp *l; 184 1.8 ad int error; 185 1.1 rmind vaddr_t uaddr; 186 1.1 rmind 187 1.1 rmind /* Allocate and initialize AIO structure */ 188 1.15 ad aio = kmem_zalloc(sizeof(struct aioproc), KM_SLEEP); 189 1.1 rmind 190 1.4 rmind /* Initialize queue and their synchronization structures */ 191 1.1 rmind mutex_init(&aio->aio_mtx, MUTEX_DEFAULT, IPL_NONE); 192 1.1 rmind cv_init(&aio->aio_worker_cv, "aiowork"); 193 1.1 rmind cv_init(&aio->done_cv, "aiodone"); 194 1.1 rmind TAILQ_INIT(&aio->jobs_queue); 195 1.1 rmind 196 1.1 rmind /* 197 1.1 rmind * Create an AIO worker thread. 198 1.1 rmind * XXX: Currently, AIO thread is not protected against user's actions. 199 1.1 rmind */ 200 1.29 rmind uaddr = uvm_uarea_alloc(); 201 1.1 rmind if (uaddr == 0) { 202 1.5 rmind aio_exit(p, aio); 203 1.50 riastrad return SET_ERROR(EAGAIN); 204 1.1 rmind } 205 1.29 rmind error = lwp_create(curlwp, p, uaddr, 0, NULL, 0, aio_worker, 206 1.42 christos NULL, &l, curlwp->l_class, &curlwp->l_sigmask, &curlwp->l_sigstk); 207 1.8 ad if (error != 0) { 208 1.29 rmind uvm_uarea_free(uaddr); 209 1.5 rmind aio_exit(p, aio); 210 1.8 ad return error; 211 1.1 rmind } 212 1.1 rmind 213 1.5 rmind /* Recheck if we are really first */ 214 1.18 ad mutex_enter(p->p_lock); 215 1.5 rmind if (p->p_aio) { 216 1.18 ad mutex_exit(p->p_lock); 217 1.5 rmind aio_exit(p, aio); 218 1.5 rmind lwp_exit(l); 219 1.5 rmind return 0; 220 1.5 rmind } 221 1.5 rmind p->p_aio = aio; 222 1.5 rmind 223 1.1 rmind /* Complete the initialization of thread, and run it */ 224 1.1 rmind aio->aio_worker = l; 225 1.1 rmind lwp_lock(l); 226 1.45 ad lwp_changepri(l, MAXPRI_USER); 227 1.45 ad setrunnable(l); 228 1.45 ad /* LWP now unlocked */ 229 1.18 ad mutex_exit(p->p_lock); 230 1.1 rmind 231 1.1 rmind return 0; 232 1.1 rmind } 233 1.1 rmind 234 1.1 rmind /* 235 1.1 rmind * Exit of Asynchronous I/O subsystem of process. 236 1.1 rmind */ 237 1.20 ad static void 238 1.20 ad aio_exit(struct proc *p, void *cookie) 239 1.1 rmind { 240 1.1 rmind struct aio_job *a_job; 241 1.20 ad struct aioproc *aio; 242 1.1 rmind 243 1.20 ad if (cookie != NULL) 244 1.20 ad aio = cookie; 245 1.20 ad else if ((aio = p->p_aio) == NULL) 246 1.1 rmind return; 247 1.1 rmind 248 1.1 rmind /* Free AIO queue */ 249 1.1 rmind while (!TAILQ_EMPTY(&aio->jobs_queue)) { 250 1.1 rmind a_job = TAILQ_FIRST(&aio->jobs_queue); 251 1.1 rmind TAILQ_REMOVE(&aio->jobs_queue, a_job, list); 252 1.4 rmind pool_put(&aio_job_pool, a_job); 253 1.11 ad atomic_dec_uint(&aio_jobs_count); 254 1.1 rmind } 255 1.1 rmind 256 1.1 rmind /* Destroy and free the entire AIO data structure */ 257 1.1 rmind cv_destroy(&aio->aio_worker_cv); 258 1.1 rmind cv_destroy(&aio->done_cv); 259 1.1 rmind mutex_destroy(&aio->aio_mtx); 260 1.1 rmind kmem_free(aio, sizeof(struct aioproc)); 261 1.1 rmind } 262 1.1 rmind 263 1.1 rmind /* 264 1.1 rmind * AIO worker thread and processor. 265 1.1 rmind */ 266 1.26 yamt static void 267 1.1 rmind aio_worker(void *arg) 268 1.1 rmind { 269 1.1 rmind struct proc *p = curlwp->l_proc; 270 1.1 rmind struct aioproc *aio = p->p_aio; 271 1.1 rmind struct aio_job *a_job; 272 1.1 rmind struct lio_req *lio; 273 1.1 rmind sigset_t oss, nss; 274 1.38 martin int error __diagused, refcnt; 275 1.1 rmind 276 1.1 rmind /* 277 1.1 rmind * Make an empty signal mask, so it 278 1.1 rmind * handles only SIGKILL and SIGSTOP. 279 1.1 rmind */ 280 1.1 rmind sigfillset(&nss); 281 1.18 ad mutex_enter(p->p_lock); 282 1.1 rmind error = sigprocmask1(curlwp, SIG_SETMASK, &nss, &oss); 283 1.18 ad mutex_exit(p->p_lock); 284 1.1 rmind KASSERT(error == 0); 285 1.1 rmind 286 1.1 rmind for (;;) { 287 1.1 rmind /* 288 1.1 rmind * Loop for each job in the queue. If there 289 1.4 rmind * are no jobs then sleep. 290 1.1 rmind */ 291 1.1 rmind mutex_enter(&aio->aio_mtx); 292 1.1 rmind while ((a_job = TAILQ_FIRST(&aio->jobs_queue)) == NULL) { 293 1.1 rmind if (cv_wait_sig(&aio->aio_worker_cv, &aio->aio_mtx)) { 294 1.1 rmind /* 295 1.4 rmind * Thread was interrupted - check for 296 1.4 rmind * pending exit or suspend. 297 1.1 rmind */ 298 1.4 rmind mutex_exit(&aio->aio_mtx); 299 1.4 rmind lwp_userret(curlwp); 300 1.4 rmind mutex_enter(&aio->aio_mtx); 301 1.1 rmind } 302 1.1 rmind } 303 1.1 rmind 304 1.1 rmind /* Take the job from the queue */ 305 1.1 rmind aio->curjob = a_job; 306 1.1 rmind TAILQ_REMOVE(&aio->jobs_queue, a_job, list); 307 1.1 rmind 308 1.11 ad atomic_dec_uint(&aio_jobs_count); 309 1.1 rmind aio->jobs_count--; 310 1.1 rmind 311 1.1 rmind mutex_exit(&aio->aio_mtx); 312 1.1 rmind 313 1.1 rmind /* Process an AIO operation */ 314 1.1 rmind aio_process(a_job); 315 1.1 rmind 316 1.1 rmind /* Copy data structure back to the user-space */ 317 1.1 rmind (void)copyout(&a_job->aiocbp, a_job->aiocb_uptr, 318 1.1 rmind sizeof(struct aiocb)); 319 1.1 rmind 320 1.1 rmind mutex_enter(&aio->aio_mtx); 321 1.36 yamt KASSERT(aio->curjob == a_job); 322 1.1 rmind aio->curjob = NULL; 323 1.4 rmind 324 1.1 rmind /* Decrease a reference counter, if there is a LIO structure */ 325 1.1 rmind lio = a_job->lio; 326 1.4 rmind refcnt = (lio != NULL ? --lio->refcnt : -1); 327 1.4 rmind 328 1.1 rmind /* Notify all suspenders */ 329 1.1 rmind cv_broadcast(&aio->done_cv); 330 1.1 rmind mutex_exit(&aio->aio_mtx); 331 1.1 rmind 332 1.1 rmind /* Send a signal, if any */ 333 1.1 rmind aio_sendsig(p, &a_job->aiocbp.aio_sigevent); 334 1.1 rmind 335 1.1 rmind /* Destroy the LIO structure */ 336 1.4 rmind if (refcnt == 0) { 337 1.1 rmind aio_sendsig(p, &lio->sig); 338 1.4 rmind pool_put(&aio_lio_pool, lio); 339 1.1 rmind } 340 1.1 rmind 341 1.30 mbalmer /* Destroy the job */ 342 1.4 rmind pool_put(&aio_job_pool, a_job); 343 1.1 rmind } 344 1.1 rmind 345 1.4 rmind /* NOTREACHED */ 346 1.1 rmind } 347 1.1 rmind 348 1.1 rmind static void 349 1.1 rmind aio_process(struct aio_job *a_job) 350 1.1 rmind { 351 1.1 rmind struct proc *p = curlwp->l_proc; 352 1.1 rmind struct aiocb *aiocbp = &a_job->aiocbp; 353 1.1 rmind struct file *fp; 354 1.1 rmind int fd = aiocbp->aio_fildes; 355 1.1 rmind int error = 0; 356 1.1 rmind 357 1.1 rmind KASSERT(a_job->aio_op != 0); 358 1.1 rmind 359 1.4 rmind if ((a_job->aio_op & (AIO_READ | AIO_WRITE)) != 0) { 360 1.1 rmind struct iovec aiov; 361 1.1 rmind struct uio auio; 362 1.1 rmind 363 1.1 rmind if (aiocbp->aio_nbytes > SSIZE_MAX) { 364 1.50 riastrad error = SET_ERROR(EINVAL); 365 1.1 rmind goto done; 366 1.1 rmind } 367 1.1 rmind 368 1.16 ad fp = fd_getfile(fd); 369 1.1 rmind if (fp == NULL) { 370 1.50 riastrad error = SET_ERROR(EBADF); 371 1.1 rmind goto done; 372 1.1 rmind } 373 1.1 rmind 374 1.1 rmind aiov.iov_base = (void *)(uintptr_t)aiocbp->aio_buf; 375 1.1 rmind aiov.iov_len = aiocbp->aio_nbytes; 376 1.1 rmind auio.uio_iov = &aiov; 377 1.1 rmind auio.uio_iovcnt = 1; 378 1.1 rmind auio.uio_resid = aiocbp->aio_nbytes; 379 1.1 rmind auio.uio_vmspace = p->p_vmspace; 380 1.1 rmind 381 1.1 rmind if (a_job->aio_op & AIO_READ) { 382 1.1 rmind /* 383 1.1 rmind * Perform a Read operation 384 1.1 rmind */ 385 1.1 rmind KASSERT((a_job->aio_op & AIO_WRITE) == 0); 386 1.1 rmind 387 1.1 rmind if ((fp->f_flag & FREAD) == 0) { 388 1.16 ad fd_putfile(fd); 389 1.50 riastrad error = SET_ERROR(EBADF); 390 1.1 rmind goto done; 391 1.1 rmind } 392 1.1 rmind auio.uio_rw = UIO_READ; 393 1.1 rmind error = (*fp->f_ops->fo_read)(fp, &aiocbp->aio_offset, 394 1.1 rmind &auio, fp->f_cred, FOF_UPDATE_OFFSET); 395 1.1 rmind } else { 396 1.1 rmind /* 397 1.1 rmind * Perform a Write operation 398 1.1 rmind */ 399 1.1 rmind KASSERT(a_job->aio_op & AIO_WRITE); 400 1.1 rmind 401 1.1 rmind if ((fp->f_flag & FWRITE) == 0) { 402 1.16 ad fd_putfile(fd); 403 1.50 riastrad error = SET_ERROR(EBADF); 404 1.1 rmind goto done; 405 1.1 rmind } 406 1.1 rmind auio.uio_rw = UIO_WRITE; 407 1.1 rmind error = (*fp->f_ops->fo_write)(fp, &aiocbp->aio_offset, 408 1.1 rmind &auio, fp->f_cred, FOF_UPDATE_OFFSET); 409 1.1 rmind } 410 1.16 ad fd_putfile(fd); 411 1.1 rmind 412 1.1 rmind /* Store the result value */ 413 1.1 rmind a_job->aiocbp.aio_nbytes -= auio.uio_resid; 414 1.1 rmind a_job->aiocbp._retval = (error == 0) ? 415 1.1 rmind a_job->aiocbp.aio_nbytes : -1; 416 1.1 rmind 417 1.4 rmind } else if ((a_job->aio_op & (AIO_SYNC | AIO_DSYNC)) != 0) { 418 1.1 rmind /* 419 1.1 rmind * Perform a file Sync operation 420 1.1 rmind */ 421 1.1 rmind struct vnode *vp; 422 1.1 rmind 423 1.16 ad if ((error = fd_getvnode(fd, &fp)) != 0) 424 1.41 msaitoh goto done; 425 1.1 rmind 426 1.1 rmind if ((fp->f_flag & FWRITE) == 0) { 427 1.16 ad fd_putfile(fd); 428 1.50 riastrad error = SET_ERROR(EBADF); 429 1.1 rmind goto done; 430 1.1 rmind } 431 1.1 rmind 432 1.40 matt vp = fp->f_vnode; 433 1.1 rmind vn_lock(vp, LK_EXCLUSIVE | LK_RETRY); 434 1.1 rmind if (a_job->aio_op & AIO_DSYNC) { 435 1.1 rmind error = VOP_FSYNC(vp, fp->f_cred, 436 1.9 pooka FSYNC_WAIT | FSYNC_DATAONLY, 0, 0); 437 1.1 rmind } else if (a_job->aio_op & AIO_SYNC) { 438 1.1 rmind error = VOP_FSYNC(vp, fp->f_cred, 439 1.9 pooka FSYNC_WAIT, 0, 0); 440 1.1 rmind } 441 1.34 hannken VOP_UNLOCK(vp); 442 1.16 ad fd_putfile(fd); 443 1.1 rmind 444 1.1 rmind /* Store the result value */ 445 1.1 rmind a_job->aiocbp._retval = (error == 0) ? 0 : -1; 446 1.1 rmind 447 1.1 rmind } else 448 1.1 rmind panic("aio_process: invalid operation code\n"); 449 1.1 rmind 450 1.1 rmind done: 451 1.1 rmind /* Job is done, set the error, if any */ 452 1.1 rmind a_job->aiocbp._errno = error; 453 1.1 rmind a_job->aiocbp._state = JOB_DONE; 454 1.1 rmind } 455 1.1 rmind 456 1.1 rmind /* 457 1.1 rmind * Send AIO signal. 458 1.1 rmind */ 459 1.1 rmind static void 460 1.1 rmind aio_sendsig(struct proc *p, struct sigevent *sig) 461 1.1 rmind { 462 1.1 rmind ksiginfo_t ksi; 463 1.1 rmind 464 1.1 rmind if (sig->sigev_signo == 0 || sig->sigev_notify == SIGEV_NONE) 465 1.1 rmind return; 466 1.1 rmind 467 1.1 rmind KSI_INIT(&ksi); 468 1.1 rmind ksi.ksi_signo = sig->sigev_signo; 469 1.1 rmind ksi.ksi_code = SI_ASYNCIO; 470 1.3 christos ksi.ksi_value = sig->sigev_value; 471 1.48 ad mutex_enter(&proc_lock); 472 1.1 rmind kpsignal(p, &ksi, NULL); 473 1.48 ad mutex_exit(&proc_lock); 474 1.1 rmind } 475 1.1 rmind 476 1.1 rmind /* 477 1.1 rmind * Enqueue the job. 478 1.1 rmind */ 479 1.1 rmind static int 480 1.1 rmind aio_enqueue_job(int op, void *aiocb_uptr, struct lio_req *lio) 481 1.1 rmind { 482 1.1 rmind struct proc *p = curlwp->l_proc; 483 1.1 rmind struct aioproc *aio; 484 1.1 rmind struct aio_job *a_job; 485 1.1 rmind struct aiocb aiocbp; 486 1.1 rmind struct sigevent *sig; 487 1.1 rmind int error; 488 1.1 rmind 489 1.12 rmind /* Non-accurate check for the limit */ 490 1.12 rmind if (aio_jobs_count + 1 > aio_max) 491 1.50 riastrad return SET_ERROR(EAGAIN); 492 1.1 rmind 493 1.1 rmind /* Get the data structure from user-space */ 494 1.1 rmind error = copyin(aiocb_uptr, &aiocbp, sizeof(struct aiocb)); 495 1.1 rmind if (error) 496 1.1 rmind return error; 497 1.1 rmind 498 1.1 rmind /* Check if signal is set, and validate it */ 499 1.1 rmind sig = &aiocbp.aio_sigevent; 500 1.1 rmind if (sig->sigev_signo < 0 || sig->sigev_signo >= NSIG || 501 1.1 rmind sig->sigev_notify < SIGEV_NONE || sig->sigev_notify > SIGEV_SA) 502 1.50 riastrad return SET_ERROR(EINVAL); 503 1.1 rmind 504 1.1 rmind /* Buffer and byte count */ 505 1.1 rmind if (((AIO_SYNC | AIO_DSYNC) & op) == 0) 506 1.1 rmind if (aiocbp.aio_buf == NULL || aiocbp.aio_nbytes > SSIZE_MAX) 507 1.50 riastrad return SET_ERROR(EINVAL); 508 1.1 rmind 509 1.1 rmind /* Check the opcode, if LIO_NOP - simply ignore */ 510 1.1 rmind if (op == AIO_LIO) { 511 1.1 rmind KASSERT(lio != NULL); 512 1.1 rmind if (aiocbp.aio_lio_opcode == LIO_WRITE) 513 1.1 rmind op = AIO_WRITE; 514 1.1 rmind else if (aiocbp.aio_lio_opcode == LIO_READ) 515 1.1 rmind op = AIO_READ; 516 1.1 rmind else 517 1.50 riastrad return (aiocbp.aio_lio_opcode == LIO_NOP) ? 0 : 518 1.50 riastrad SET_ERROR(EINVAL); 519 1.1 rmind } else { 520 1.1 rmind KASSERT(lio == NULL); 521 1.1 rmind } 522 1.1 rmind 523 1.1 rmind /* 524 1.1 rmind * Look for already existing job. If found - the job is in-progress. 525 1.1 rmind * According to POSIX this is invalid, so return the error. 526 1.1 rmind */ 527 1.1 rmind aio = p->p_aio; 528 1.1 rmind if (aio) { 529 1.1 rmind mutex_enter(&aio->aio_mtx); 530 1.1 rmind TAILQ_FOREACH(a_job, &aio->jobs_queue, list) { 531 1.1 rmind if (a_job->aiocb_uptr != aiocb_uptr) 532 1.1 rmind continue; 533 1.1 rmind mutex_exit(&aio->aio_mtx); 534 1.50 riastrad return SET_ERROR(EINVAL); 535 1.1 rmind } 536 1.1 rmind mutex_exit(&aio->aio_mtx); 537 1.1 rmind } 538 1.1 rmind 539 1.1 rmind /* 540 1.1 rmind * Check if AIO structure is initialized, if not - initialize it. 541 1.1 rmind * In LIO case, we did that already. We will recheck this with 542 1.20 ad * the lock in aio_procinit(). 543 1.1 rmind */ 544 1.1 rmind if (lio == NULL && p->p_aio == NULL) 545 1.20 ad if (aio_procinit(p)) 546 1.50 riastrad return SET_ERROR(EAGAIN); 547 1.1 rmind aio = p->p_aio; 548 1.1 rmind 549 1.1 rmind /* 550 1.1 rmind * Set the state with errno, and copy data 551 1.1 rmind * structure back to the user-space. 552 1.1 rmind */ 553 1.1 rmind aiocbp._state = JOB_WIP; 554 1.50 riastrad aiocbp._errno = SET_ERROR(EINPROGRESS); 555 1.1 rmind aiocbp._retval = -1; 556 1.1 rmind error = copyout(&aiocbp, aiocb_uptr, sizeof(struct aiocb)); 557 1.1 rmind if (error) 558 1.1 rmind return error; 559 1.1 rmind 560 1.1 rmind /* Allocate and initialize a new AIO job */ 561 1.44 christos a_job = pool_get(&aio_job_pool, PR_WAITOK | PR_ZERO); 562 1.1 rmind 563 1.1 rmind /* 564 1.1 rmind * Set the data. 565 1.1 rmind * Store the user-space pointer for searching. Since we 566 1.1 rmind * are storing only per proc pointers - it is safe. 567 1.1 rmind */ 568 1.1 rmind memcpy(&a_job->aiocbp, &aiocbp, sizeof(struct aiocb)); 569 1.1 rmind a_job->aiocb_uptr = aiocb_uptr; 570 1.1 rmind a_job->aio_op |= op; 571 1.1 rmind a_job->lio = lio; 572 1.1 rmind 573 1.1 rmind /* 574 1.1 rmind * Add the job to the queue, update the counters, and 575 1.1 rmind * notify the AIO worker thread to handle the job. 576 1.1 rmind */ 577 1.1 rmind mutex_enter(&aio->aio_mtx); 578 1.1 rmind 579 1.1 rmind /* Fail, if the limit was reached */ 580 1.13 rmind if (atomic_inc_uint_nv(&aio_jobs_count) > aio_max || 581 1.13 rmind aio->jobs_count >= aio_listio_max) { 582 1.12 rmind atomic_dec_uint(&aio_jobs_count); 583 1.1 rmind mutex_exit(&aio->aio_mtx); 584 1.4 rmind pool_put(&aio_job_pool, a_job); 585 1.50 riastrad return SET_ERROR(EAGAIN); 586 1.1 rmind } 587 1.1 rmind 588 1.1 rmind TAILQ_INSERT_TAIL(&aio->jobs_queue, a_job, list); 589 1.1 rmind aio->jobs_count++; 590 1.1 rmind if (lio) 591 1.1 rmind lio->refcnt++; 592 1.1 rmind cv_signal(&aio->aio_worker_cv); 593 1.1 rmind 594 1.1 rmind mutex_exit(&aio->aio_mtx); 595 1.1 rmind 596 1.1 rmind /* 597 1.1 rmind * One would handle the errors only with aio_error() function. 598 1.1 rmind * This way is appropriate according to POSIX. 599 1.1 rmind */ 600 1.1 rmind return 0; 601 1.1 rmind } 602 1.1 rmind 603 1.1 rmind /* 604 1.1 rmind * Syscall functions. 605 1.1 rmind */ 606 1.1 rmind 607 1.1 rmind int 608 1.27 yamt sys_aio_cancel(struct lwp *l, const struct sys_aio_cancel_args *uap, 609 1.27 yamt register_t *retval) 610 1.1 rmind { 611 1.14 dsl /* { 612 1.1 rmind syscallarg(int) fildes; 613 1.1 rmind syscallarg(struct aiocb *) aiocbp; 614 1.14 dsl } */ 615 1.1 rmind struct proc *p = l->l_proc; 616 1.1 rmind struct aioproc *aio; 617 1.1 rmind struct aio_job *a_job; 618 1.1 rmind struct aiocb *aiocbp_ptr; 619 1.1 rmind struct lio_req *lio; 620 1.1 rmind struct filedesc *fdp = p->p_fd; 621 1.1 rmind unsigned int cn, errcnt, fildes; 622 1.24 ad fdtab_t *dt; 623 1.1 rmind 624 1.1 rmind TAILQ_HEAD(, aio_job) tmp_jobs_list; 625 1.1 rmind 626 1.1 rmind /* Check for invalid file descriptor */ 627 1.1 rmind fildes = (unsigned int)SCARG(uap, fildes); 628 1.46 riastrad dt = atomic_load_consume(&fdp->fd_dt); 629 1.24 ad if (fildes >= dt->dt_nfiles) 630 1.50 riastrad return SET_ERROR(EBADF); 631 1.24 ad if (dt->dt_ff[fildes] == NULL || dt->dt_ff[fildes]->ff_file == NULL) 632 1.50 riastrad return SET_ERROR(EBADF); 633 1.1 rmind 634 1.1 rmind /* Check if AIO structure is initialized */ 635 1.1 rmind if (p->p_aio == NULL) { 636 1.1 rmind *retval = AIO_NOTCANCELED; 637 1.1 rmind return 0; 638 1.1 rmind } 639 1.1 rmind 640 1.1 rmind aio = p->p_aio; 641 1.1 rmind aiocbp_ptr = (struct aiocb *)SCARG(uap, aiocbp); 642 1.1 rmind 643 1.1 rmind mutex_enter(&aio->aio_mtx); 644 1.1 rmind 645 1.1 rmind /* Cancel the jobs, and remove them from the queue */ 646 1.1 rmind cn = 0; 647 1.1 rmind TAILQ_INIT(&tmp_jobs_list); 648 1.1 rmind TAILQ_FOREACH(a_job, &aio->jobs_queue, list) { 649 1.1 rmind if (aiocbp_ptr) { 650 1.1 rmind if (aiocbp_ptr != a_job->aiocb_uptr) 651 1.1 rmind continue; 652 1.1 rmind if (fildes != a_job->aiocbp.aio_fildes) { 653 1.1 rmind mutex_exit(&aio->aio_mtx); 654 1.50 riastrad return SET_ERROR(EBADF); 655 1.1 rmind } 656 1.1 rmind } else if (a_job->aiocbp.aio_fildes != fildes) 657 1.1 rmind continue; 658 1.1 rmind 659 1.1 rmind TAILQ_REMOVE(&aio->jobs_queue, a_job, list); 660 1.1 rmind TAILQ_INSERT_TAIL(&tmp_jobs_list, a_job, list); 661 1.1 rmind 662 1.1 rmind /* Decrease the counters */ 663 1.11 ad atomic_dec_uint(&aio_jobs_count); 664 1.1 rmind aio->jobs_count--; 665 1.1 rmind lio = a_job->lio; 666 1.4 rmind if (lio != NULL && --lio->refcnt != 0) 667 1.4 rmind a_job->lio = NULL; 668 1.1 rmind 669 1.1 rmind cn++; 670 1.1 rmind if (aiocbp_ptr) 671 1.1 rmind break; 672 1.1 rmind } 673 1.1 rmind 674 1.1 rmind /* There are canceled jobs */ 675 1.1 rmind if (cn) 676 1.1 rmind *retval = AIO_CANCELED; 677 1.1 rmind 678 1.1 rmind /* We cannot cancel current job */ 679 1.1 rmind a_job = aio->curjob; 680 1.1 rmind if (a_job && ((a_job->aiocbp.aio_fildes == fildes) || 681 1.1 rmind (a_job->aiocb_uptr == aiocbp_ptr))) 682 1.1 rmind *retval = AIO_NOTCANCELED; 683 1.1 rmind 684 1.1 rmind mutex_exit(&aio->aio_mtx); 685 1.1 rmind 686 1.1 rmind /* Free the jobs after the lock */ 687 1.1 rmind errcnt = 0; 688 1.1 rmind while (!TAILQ_EMPTY(&tmp_jobs_list)) { 689 1.1 rmind a_job = TAILQ_FIRST(&tmp_jobs_list); 690 1.1 rmind TAILQ_REMOVE(&tmp_jobs_list, a_job, list); 691 1.1 rmind /* Set the errno and copy structures back to the user-space */ 692 1.50 riastrad a_job->aiocbp._errno = SET_ERROR(ECANCELED); 693 1.1 rmind a_job->aiocbp._state = JOB_DONE; 694 1.1 rmind if (copyout(&a_job->aiocbp, a_job->aiocb_uptr, 695 1.1 rmind sizeof(struct aiocb))) 696 1.1 rmind errcnt++; 697 1.1 rmind /* Send a signal if any */ 698 1.1 rmind aio_sendsig(p, &a_job->aiocbp.aio_sigevent); 699 1.6 rmind if (a_job->lio) { 700 1.6 rmind lio = a_job->lio; 701 1.6 rmind aio_sendsig(p, &lio->sig); 702 1.6 rmind pool_put(&aio_lio_pool, lio); 703 1.6 rmind } 704 1.4 rmind pool_put(&aio_job_pool, a_job); 705 1.1 rmind } 706 1.1 rmind 707 1.1 rmind if (errcnt) 708 1.50 riastrad return SET_ERROR(EFAULT); 709 1.1 rmind 710 1.1 rmind /* Set a correct return value */ 711 1.1 rmind if (*retval == 0) 712 1.1 rmind *retval = AIO_ALLDONE; 713 1.1 rmind 714 1.1 rmind return 0; 715 1.1 rmind } 716 1.1 rmind 717 1.1 rmind int 718 1.27 yamt sys_aio_error(struct lwp *l, const struct sys_aio_error_args *uap, 719 1.27 yamt register_t *retval) 720 1.1 rmind { 721 1.14 dsl /* { 722 1.1 rmind syscallarg(const struct aiocb *) aiocbp; 723 1.14 dsl } */ 724 1.1 rmind struct proc *p = l->l_proc; 725 1.1 rmind struct aioproc *aio = p->p_aio; 726 1.1 rmind struct aiocb aiocbp; 727 1.1 rmind int error; 728 1.1 rmind 729 1.1 rmind if (aio == NULL) 730 1.50 riastrad return SET_ERROR(EINVAL); 731 1.1 rmind 732 1.1 rmind error = copyin(SCARG(uap, aiocbp), &aiocbp, sizeof(struct aiocb)); 733 1.1 rmind if (error) 734 1.1 rmind return error; 735 1.1 rmind 736 1.1 rmind if (aiocbp._state == JOB_NONE) 737 1.50 riastrad return SET_ERROR(EINVAL); 738 1.1 rmind 739 1.1 rmind *retval = aiocbp._errno; 740 1.1 rmind 741 1.1 rmind return 0; 742 1.1 rmind } 743 1.1 rmind 744 1.1 rmind int 745 1.27 yamt sys_aio_fsync(struct lwp *l, const struct sys_aio_fsync_args *uap, 746 1.27 yamt register_t *retval) 747 1.1 rmind { 748 1.14 dsl /* { 749 1.1 rmind syscallarg(int) op; 750 1.1 rmind syscallarg(struct aiocb *) aiocbp; 751 1.14 dsl } */ 752 1.1 rmind int op = SCARG(uap, op); 753 1.1 rmind 754 1.1 rmind if ((op != O_DSYNC) && (op != O_SYNC)) 755 1.50 riastrad return SET_ERROR(EINVAL); 756 1.1 rmind 757 1.1 rmind op = O_DSYNC ? AIO_DSYNC : AIO_SYNC; 758 1.1 rmind 759 1.1 rmind return aio_enqueue_job(op, SCARG(uap, aiocbp), NULL); 760 1.1 rmind } 761 1.1 rmind 762 1.1 rmind int 763 1.27 yamt sys_aio_read(struct lwp *l, const struct sys_aio_read_args *uap, 764 1.27 yamt register_t *retval) 765 1.1 rmind { 766 1.14 dsl /* { 767 1.1 rmind syscallarg(struct aiocb *) aiocbp; 768 1.14 dsl } */ 769 1.1 rmind 770 1.1 rmind return aio_enqueue_job(AIO_READ, SCARG(uap, aiocbp), NULL); 771 1.1 rmind } 772 1.1 rmind 773 1.1 rmind int 774 1.27 yamt sys_aio_return(struct lwp *l, const struct sys_aio_return_args *uap, 775 1.27 yamt register_t *retval) 776 1.1 rmind { 777 1.14 dsl /* { 778 1.1 rmind syscallarg(struct aiocb *) aiocbp; 779 1.14 dsl } */ 780 1.1 rmind struct proc *p = l->l_proc; 781 1.1 rmind struct aioproc *aio = p->p_aio; 782 1.1 rmind struct aiocb aiocbp; 783 1.1 rmind int error; 784 1.1 rmind 785 1.1 rmind if (aio == NULL) 786 1.50 riastrad return SET_ERROR(EINVAL); 787 1.1 rmind 788 1.1 rmind error = copyin(SCARG(uap, aiocbp), &aiocbp, sizeof(struct aiocb)); 789 1.1 rmind if (error) 790 1.1 rmind return error; 791 1.1 rmind 792 1.1 rmind if (aiocbp._errno == EINPROGRESS || aiocbp._state != JOB_DONE) 793 1.50 riastrad return SET_ERROR(EINVAL); 794 1.1 rmind 795 1.1 rmind *retval = aiocbp._retval; 796 1.1 rmind 797 1.1 rmind /* Reset the internal variables */ 798 1.1 rmind aiocbp._errno = 0; 799 1.1 rmind aiocbp._retval = -1; 800 1.1 rmind aiocbp._state = JOB_NONE; 801 1.1 rmind error = copyout(&aiocbp, SCARG(uap, aiocbp), sizeof(struct aiocb)); 802 1.1 rmind 803 1.1 rmind return error; 804 1.1 rmind } 805 1.1 rmind 806 1.1 rmind int 807 1.22 christos sys___aio_suspend50(struct lwp *l, const struct sys___aio_suspend50_args *uap, 808 1.22 christos register_t *retval) 809 1.1 rmind { 810 1.14 dsl /* { 811 1.1 rmind syscallarg(const struct aiocb *const[]) list; 812 1.1 rmind syscallarg(int) nent; 813 1.1 rmind syscallarg(const struct timespec *) timeout; 814 1.14 dsl } */ 815 1.22 christos struct aiocb **list; 816 1.1 rmind struct timespec ts; 817 1.22 christos int error, nent; 818 1.1 rmind 819 1.1 rmind nent = SCARG(uap, nent); 820 1.1 rmind if (nent <= 0 || nent > aio_listio_max) 821 1.50 riastrad return SET_ERROR(EAGAIN); 822 1.1 rmind 823 1.1 rmind if (SCARG(uap, timeout)) { 824 1.1 rmind /* Convert timespec to ticks */ 825 1.1 rmind error = copyin(SCARG(uap, timeout), &ts, 826 1.1 rmind sizeof(struct timespec)); 827 1.1 rmind if (error) 828 1.1 rmind return error; 829 1.22 christos } 830 1.31 rmind 831 1.28 yamt list = kmem_alloc(nent * sizeof(*list), KM_SLEEP); 832 1.28 yamt error = copyin(SCARG(uap, list), list, nent * sizeof(*list)); 833 1.22 christos if (error) 834 1.22 christos goto out; 835 1.22 christos error = aio_suspend1(l, list, nent, SCARG(uap, timeout) ? &ts : NULL); 836 1.22 christos out: 837 1.28 yamt kmem_free(list, nent * sizeof(*list)); 838 1.22 christos return error; 839 1.22 christos } 840 1.22 christos 841 1.22 christos int 842 1.22 christos aio_suspend1(struct lwp *l, struct aiocb **aiocbp_list, int nent, 843 1.22 christos struct timespec *ts) 844 1.22 christos { 845 1.22 christos struct proc *p = l->l_proc; 846 1.22 christos struct aioproc *aio; 847 1.22 christos struct aio_job *a_job; 848 1.22 christos int i, error, timo; 849 1.22 christos 850 1.22 christos if (p->p_aio == NULL) 851 1.50 riastrad return SET_ERROR(EAGAIN); 852 1.22 christos aio = p->p_aio; 853 1.22 christos 854 1.22 christos if (ts) { 855 1.22 christos timo = mstohz((ts->tv_sec * 1000) + (ts->tv_nsec / 1000000)); 856 1.22 christos if (timo == 0 && ts->tv_sec == 0 && ts->tv_nsec > 0) 857 1.1 rmind timo = 1; 858 1.1 rmind if (timo <= 0) 859 1.50 riastrad return SET_ERROR(EAGAIN); 860 1.1 rmind } else 861 1.1 rmind timo = 0; 862 1.1 rmind 863 1.1 rmind mutex_enter(&aio->aio_mtx); 864 1.1 rmind for (;;) { 865 1.1 rmind for (i = 0; i < nent; i++) { 866 1.1 rmind 867 1.1 rmind /* Skip NULL entries */ 868 1.1 rmind if (aiocbp_list[i] == NULL) 869 1.1 rmind continue; 870 1.1 rmind 871 1.1 rmind /* Skip current job */ 872 1.1 rmind if (aio->curjob) { 873 1.1 rmind a_job = aio->curjob; 874 1.1 rmind if (a_job->aiocb_uptr == aiocbp_list[i]) 875 1.1 rmind continue; 876 1.1 rmind } 877 1.1 rmind 878 1.1 rmind /* Look for a job in the queue */ 879 1.1 rmind TAILQ_FOREACH(a_job, &aio->jobs_queue, list) 880 1.1 rmind if (a_job->aiocb_uptr == aiocbp_list[i]) 881 1.1 rmind break; 882 1.1 rmind 883 1.1 rmind if (a_job == NULL) { 884 1.1 rmind struct aiocb aiocbp; 885 1.1 rmind 886 1.1 rmind mutex_exit(&aio->aio_mtx); 887 1.1 rmind 888 1.31 rmind /* Check if the job is done. */ 889 1.1 rmind error = copyin(aiocbp_list[i], &aiocbp, 890 1.1 rmind sizeof(struct aiocb)); 891 1.1 rmind if (error == 0 && aiocbp._state != JOB_DONE) { 892 1.1 rmind mutex_enter(&aio->aio_mtx); 893 1.1 rmind continue; 894 1.1 rmind } 895 1.1 rmind return error; 896 1.1 rmind } 897 1.1 rmind } 898 1.1 rmind 899 1.1 rmind /* Wait for a signal or when timeout occurs */ 900 1.1 rmind error = cv_timedwait_sig(&aio->done_cv, &aio->aio_mtx, timo); 901 1.1 rmind if (error) { 902 1.1 rmind if (error == EWOULDBLOCK) 903 1.50 riastrad error = SET_ERROR(EAGAIN); 904 1.1 rmind break; 905 1.1 rmind } 906 1.1 rmind } 907 1.1 rmind mutex_exit(&aio->aio_mtx); 908 1.1 rmind return error; 909 1.1 rmind } 910 1.1 rmind 911 1.1 rmind int 912 1.27 yamt sys_aio_write(struct lwp *l, const struct sys_aio_write_args *uap, 913 1.27 yamt register_t *retval) 914 1.1 rmind { 915 1.14 dsl /* { 916 1.1 rmind syscallarg(struct aiocb *) aiocbp; 917 1.14 dsl } */ 918 1.1 rmind 919 1.1 rmind return aio_enqueue_job(AIO_WRITE, SCARG(uap, aiocbp), NULL); 920 1.1 rmind } 921 1.1 rmind 922 1.1 rmind int 923 1.27 yamt sys_lio_listio(struct lwp *l, const struct sys_lio_listio_args *uap, 924 1.27 yamt register_t *retval) 925 1.1 rmind { 926 1.14 dsl /* { 927 1.1 rmind syscallarg(int) mode; 928 1.1 rmind syscallarg(struct aiocb *const[]) list; 929 1.1 rmind syscallarg(int) nent; 930 1.1 rmind syscallarg(struct sigevent *) sig; 931 1.14 dsl } */ 932 1.1 rmind struct proc *p = l->l_proc; 933 1.1 rmind struct aioproc *aio; 934 1.1 rmind struct aiocb **aiocbp_list; 935 1.1 rmind struct lio_req *lio; 936 1.1 rmind int i, error, errcnt, mode, nent; 937 1.1 rmind 938 1.1 rmind mode = SCARG(uap, mode); 939 1.1 rmind nent = SCARG(uap, nent); 940 1.1 rmind 941 1.12 rmind /* Non-accurate checks for the limit and invalid values */ 942 1.1 rmind if (nent < 1 || nent > aio_listio_max) 943 1.50 riastrad return SET_ERROR(EINVAL); 944 1.12 rmind if (aio_jobs_count + nent > aio_max) 945 1.50 riastrad return SET_ERROR(EAGAIN); 946 1.1 rmind 947 1.1 rmind /* Check if AIO structure is initialized, if not - initialize it */ 948 1.1 rmind if (p->p_aio == NULL) 949 1.20 ad if (aio_procinit(p)) 950 1.50 riastrad return SET_ERROR(EAGAIN); 951 1.1 rmind aio = p->p_aio; 952 1.1 rmind 953 1.1 rmind /* Create a LIO structure */ 954 1.4 rmind lio = pool_get(&aio_lio_pool, PR_WAITOK); 955 1.4 rmind lio->refcnt = 1; 956 1.4 rmind error = 0; 957 1.4 rmind 958 1.4 rmind switch (mode) { 959 1.4 rmind case LIO_WAIT: 960 1.1 rmind memset(&lio->sig, 0, sizeof(struct sigevent)); 961 1.4 rmind break; 962 1.4 rmind case LIO_NOWAIT: 963 1.4 rmind /* Check for signal, validate it */ 964 1.4 rmind if (SCARG(uap, sig)) { 965 1.4 rmind struct sigevent *sig = &lio->sig; 966 1.4 rmind 967 1.4 rmind error = copyin(SCARG(uap, sig), &lio->sig, 968 1.4 rmind sizeof(struct sigevent)); 969 1.4 rmind if (error == 0 && 970 1.4 rmind (sig->sigev_signo < 0 || 971 1.4 rmind sig->sigev_signo >= NSIG || 972 1.4 rmind sig->sigev_notify < SIGEV_NONE || 973 1.4 rmind sig->sigev_notify > SIGEV_SA)) 974 1.50 riastrad error = SET_ERROR(EINVAL); 975 1.4 rmind } else 976 1.4 rmind memset(&lio->sig, 0, sizeof(struct sigevent)); 977 1.4 rmind break; 978 1.4 rmind default: 979 1.50 riastrad error = SET_ERROR(EINVAL); 980 1.4 rmind break; 981 1.4 rmind } 982 1.4 rmind 983 1.4 rmind if (error != 0) { 984 1.4 rmind pool_put(&aio_lio_pool, lio); 985 1.4 rmind return error; 986 1.4 rmind } 987 1.1 rmind 988 1.1 rmind /* Get the list from user-space */ 989 1.28 yamt aiocbp_list = kmem_alloc(nent * sizeof(*aiocbp_list), KM_SLEEP); 990 1.1 rmind error = copyin(SCARG(uap, list), aiocbp_list, 991 1.28 yamt nent * sizeof(*aiocbp_list)); 992 1.4 rmind if (error) { 993 1.4 rmind mutex_enter(&aio->aio_mtx); 994 1.1 rmind goto err; 995 1.4 rmind } 996 1.1 rmind 997 1.1 rmind /* Enqueue all jobs */ 998 1.1 rmind errcnt = 0; 999 1.1 rmind for (i = 0; i < nent; i++) { 1000 1.1 rmind error = aio_enqueue_job(AIO_LIO, aiocbp_list[i], lio); 1001 1.1 rmind /* 1002 1.1 rmind * According to POSIX, in such error case it may 1003 1.1 rmind * fail with other I/O operations initiated. 1004 1.1 rmind */ 1005 1.1 rmind if (error) 1006 1.1 rmind errcnt++; 1007 1.1 rmind } 1008 1.1 rmind 1009 1.4 rmind mutex_enter(&aio->aio_mtx); 1010 1.4 rmind 1011 1.1 rmind /* Return an error, if any */ 1012 1.1 rmind if (errcnt) { 1013 1.50 riastrad error = SET_ERROR(EIO); 1014 1.1 rmind goto err; 1015 1.1 rmind } 1016 1.1 rmind 1017 1.1 rmind if (mode == LIO_WAIT) { 1018 1.1 rmind /* 1019 1.1 rmind * Wait for AIO completion. In such case, 1020 1.1 rmind * the LIO structure will be freed here. 1021 1.1 rmind */ 1022 1.4 rmind while (lio->refcnt > 1 && error == 0) 1023 1.1 rmind error = cv_wait_sig(&aio->done_cv, &aio->aio_mtx); 1024 1.1 rmind if (error) 1025 1.50 riastrad error = SET_ERROR(EINTR); 1026 1.1 rmind } 1027 1.1 rmind 1028 1.1 rmind err: 1029 1.4 rmind if (--lio->refcnt != 0) 1030 1.4 rmind lio = NULL; 1031 1.4 rmind mutex_exit(&aio->aio_mtx); 1032 1.4 rmind if (lio != NULL) { 1033 1.4 rmind aio_sendsig(p, &lio->sig); 1034 1.4 rmind pool_put(&aio_lio_pool, lio); 1035 1.4 rmind } 1036 1.28 yamt kmem_free(aiocbp_list, nent * sizeof(*aiocbp_list)); 1037 1.1 rmind return error; 1038 1.1 rmind } 1039 1.1 rmind 1040 1.1 rmind /* 1041 1.1 rmind * SysCtl 1042 1.1 rmind */ 1043 1.1 rmind 1044 1.1 rmind static int 1045 1.1 rmind sysctl_aio_listio_max(SYSCTLFN_ARGS) 1046 1.1 rmind { 1047 1.1 rmind struct sysctlnode node; 1048 1.1 rmind int error, newsize; 1049 1.1 rmind 1050 1.1 rmind node = *rnode; 1051 1.1 rmind node.sysctl_data = &newsize; 1052 1.1 rmind 1053 1.1 rmind newsize = aio_listio_max; 1054 1.1 rmind error = sysctl_lookup(SYSCTLFN_CALL(&node)); 1055 1.1 rmind if (error || newp == NULL) 1056 1.1 rmind return error; 1057 1.1 rmind 1058 1.1 rmind if (newsize < 1 || newsize > aio_max) 1059 1.50 riastrad return SET_ERROR(EINVAL); 1060 1.1 rmind aio_listio_max = newsize; 1061 1.1 rmind 1062 1.1 rmind return 0; 1063 1.1 rmind } 1064 1.1 rmind 1065 1.1 rmind static int 1066 1.1 rmind sysctl_aio_max(SYSCTLFN_ARGS) 1067 1.1 rmind { 1068 1.1 rmind struct sysctlnode node; 1069 1.1 rmind int error, newsize; 1070 1.1 rmind 1071 1.1 rmind node = *rnode; 1072 1.1 rmind node.sysctl_data = &newsize; 1073 1.1 rmind 1074 1.1 rmind newsize = aio_max; 1075 1.1 rmind error = sysctl_lookup(SYSCTLFN_CALL(&node)); 1076 1.1 rmind if (error || newp == NULL) 1077 1.1 rmind return error; 1078 1.1 rmind 1079 1.1 rmind if (newsize < 1 || newsize < aio_listio_max) 1080 1.50 riastrad return SET_ERROR(EINVAL); 1081 1.1 rmind aio_max = newsize; 1082 1.1 rmind 1083 1.1 rmind return 0; 1084 1.1 rmind } 1085 1.1 rmind 1086 1.47 pgoyette SYSCTL_SETUP(sysctl_aio_init, "aio sysctl") 1087 1.1 rmind { 1088 1.35 jruoho int rv; 1089 1.35 jruoho 1090 1.47 pgoyette rv = sysctl_createv(clog, 0, NULL, NULL, 1091 1.49 riastrad CTLFLAG_PERMANENT | CTLFLAG_IMMEDIATE, 1092 1.49 riastrad CTLTYPE_INT, "posix_aio", 1093 1.49 riastrad SYSCTL_DESCR("Version of IEEE Std 1003.1 and its " 1094 1.49 riastrad "Asynchronous I/O option to which the " 1095 1.49 riastrad "system attempts to conform"), 1096 1.49 riastrad NULL, _POSIX_ASYNCHRONOUS_IO, NULL, 0, 1097 1.49 riastrad CTL_KERN, CTL_CREATE, CTL_EOL); 1098 1.35 jruoho 1099 1.35 jruoho if (rv != 0) 1100 1.47 pgoyette return; 1101 1.35 jruoho 1102 1.47 pgoyette rv = sysctl_createv(clog, 0, NULL, NULL, 1103 1.49 riastrad CTLFLAG_PERMANENT | CTLFLAG_READWRITE, 1104 1.49 riastrad CTLTYPE_INT, "aio_listio_max", 1105 1.49 riastrad SYSCTL_DESCR("Maximum number of asynchronous I/O " 1106 1.49 riastrad "operations in a single list I/O call"), 1107 1.49 riastrad sysctl_aio_listio_max, 0, &aio_listio_max, 0, 1108 1.49 riastrad CTL_KERN, CTL_CREATE, CTL_EOL); 1109 1.35 jruoho 1110 1.35 jruoho if (rv != 0) 1111 1.47 pgoyette return; 1112 1.35 jruoho 1113 1.47 pgoyette rv = sysctl_createv(clog, 0, NULL, NULL, 1114 1.49 riastrad CTLFLAG_PERMANENT | CTLFLAG_READWRITE, 1115 1.49 riastrad CTLTYPE_INT, "aio_max", 1116 1.49 riastrad SYSCTL_DESCR("Maximum number of asynchronous I/O " 1117 1.49 riastrad "operations"), 1118 1.49 riastrad sysctl_aio_max, 0, &aio_max, 0, 1119 1.49 riastrad CTL_KERN, CTL_CREATE, CTL_EOL); 1120 1.35 jruoho 1121 1.47 pgoyette return; 1122 1.1 rmind } 1123 1.1 rmind 1124 1.1 rmind /* 1125 1.1 rmind * Debugging 1126 1.1 rmind */ 1127 1.1 rmind #if defined(DDB) 1128 1.1 rmind void 1129 1.1 rmind aio_print_jobs(void (*pr)(const char *, ...)) 1130 1.1 rmind { 1131 1.37 matt struct proc *p = curlwp->l_proc; 1132 1.1 rmind struct aioproc *aio; 1133 1.1 rmind struct aio_job *a_job; 1134 1.1 rmind struct aiocb *aiocbp; 1135 1.1 rmind 1136 1.1 rmind if (p == NULL) { 1137 1.1 rmind (*pr)("AIO: We are not in the processes right now.\n"); 1138 1.1 rmind return; 1139 1.1 rmind } 1140 1.1 rmind 1141 1.1 rmind aio = p->p_aio; 1142 1.1 rmind if (aio == NULL) { 1143 1.1 rmind (*pr)("AIO data is not initialized (PID = %d).\n", p->p_pid); 1144 1.1 rmind return; 1145 1.1 rmind } 1146 1.1 rmind 1147 1.1 rmind (*pr)("AIO: PID = %d\n", p->p_pid); 1148 1.1 rmind (*pr)("AIO: Global count of the jobs = %u\n", aio_jobs_count); 1149 1.1 rmind (*pr)("AIO: Count of the jobs = %u\n", aio->jobs_count); 1150 1.1 rmind 1151 1.1 rmind if (aio->curjob) { 1152 1.1 rmind a_job = aio->curjob; 1153 1.1 rmind (*pr)("\nAIO current job:\n"); 1154 1.1 rmind (*pr)(" opcode = %d, errno = %d, state = %d, aiocb_ptr = %p\n", 1155 1.1 rmind a_job->aio_op, a_job->aiocbp._errno, 1156 1.1 rmind a_job->aiocbp._state, a_job->aiocb_uptr); 1157 1.1 rmind aiocbp = &a_job->aiocbp; 1158 1.1 rmind (*pr)(" fd = %d, offset = %u, buf = %p, nbytes = %u\n", 1159 1.1 rmind aiocbp->aio_fildes, aiocbp->aio_offset, 1160 1.1 rmind aiocbp->aio_buf, aiocbp->aio_nbytes); 1161 1.1 rmind } 1162 1.1 rmind 1163 1.1 rmind (*pr)("\nAIO queue:\n"); 1164 1.1 rmind TAILQ_FOREACH(a_job, &aio->jobs_queue, list) { 1165 1.1 rmind (*pr)(" opcode = %d, errno = %d, state = %d, aiocb_ptr = %p\n", 1166 1.1 rmind a_job->aio_op, a_job->aiocbp._errno, 1167 1.1 rmind a_job->aiocbp._state, a_job->aiocb_uptr); 1168 1.1 rmind aiocbp = &a_job->aiocbp; 1169 1.1 rmind (*pr)(" fd = %d, offset = %u, buf = %p, nbytes = %u\n", 1170 1.1 rmind aiocbp->aio_fildes, aiocbp->aio_offset, 1171 1.1 rmind aiocbp->aio_buf, aiocbp->aio_nbytes); 1172 1.1 rmind } 1173 1.1 rmind } 1174 1.1 rmind #endif /* defined(DDB) */ 1175