sys_generic.c revision 1.83.2.8 1 1.83.2.8 yamt /* $NetBSD: sys_generic.c,v 1.83.2.8 2008/03/17 09:15:34 yamt Exp $ */
2 1.83.2.4 yamt
3 1.83.2.4 yamt /*-
4 1.83.2.8 yamt * Copyright (c) 2007, 2008 The NetBSD Foundation, Inc.
5 1.83.2.4 yamt * All rights reserved.
6 1.83.2.4 yamt *
7 1.83.2.4 yamt * This code is derived from software contributed to The NetBSD Foundation
8 1.83.2.4 yamt * by Andrew Doran.
9 1.83.2.4 yamt *
10 1.83.2.4 yamt * Redistribution and use in source and binary forms, with or without
11 1.83.2.4 yamt * modification, are permitted provided that the following conditions
12 1.83.2.4 yamt * are met:
13 1.83.2.4 yamt * 1. Redistributions of source code must retain the above copyright
14 1.83.2.4 yamt * notice, this list of conditions and the following disclaimer.
15 1.83.2.4 yamt * 2. Redistributions in binary form must reproduce the above copyright
16 1.83.2.4 yamt * notice, this list of conditions and the following disclaimer in the
17 1.83.2.4 yamt * documentation and/or other materials provided with the distribution.
18 1.83.2.4 yamt * 3. All advertising materials mentioning features or use of this software
19 1.83.2.4 yamt * must display the following acknowledgement:
20 1.83.2.4 yamt * This product includes software developed by the NetBSD
21 1.83.2.4 yamt * Foundation, Inc. and its contributors.
22 1.83.2.4 yamt * 4. Neither the name of The NetBSD Foundation nor the names of its
23 1.83.2.4 yamt * contributors may be used to endorse or promote products derived
24 1.83.2.4 yamt * from this software without specific prior written permission.
25 1.83.2.4 yamt *
26 1.83.2.4 yamt * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
27 1.83.2.4 yamt * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
28 1.83.2.4 yamt * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
29 1.83.2.4 yamt * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
30 1.83.2.4 yamt * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
31 1.83.2.4 yamt * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
32 1.83.2.4 yamt * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
33 1.83.2.4 yamt * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
34 1.83.2.4 yamt * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
35 1.83.2.4 yamt * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
36 1.83.2.4 yamt * POSSIBILITY OF SUCH DAMAGE.
37 1.83.2.4 yamt */
38 1.15 cgd
39 1.15 cgd /*
40 1.15 cgd * Copyright (c) 1982, 1986, 1989, 1993
41 1.15 cgd * The Regents of the University of California. All rights reserved.
42 1.15 cgd * (c) UNIX System Laboratories, Inc.
43 1.15 cgd * All or some portions of this file are derived from material licensed
44 1.15 cgd * to the University of California by American Telephone and Telegraph
45 1.15 cgd * Co. or Unix System Laboratories, Inc. and are reproduced herein with
46 1.15 cgd * the permission of UNIX System Laboratories, Inc.
47 1.15 cgd *
48 1.15 cgd * Redistribution and use in source and binary forms, with or without
49 1.15 cgd * modification, are permitted provided that the following conditions
50 1.15 cgd * are met:
51 1.15 cgd * 1. Redistributions of source code must retain the above copyright
52 1.15 cgd * notice, this list of conditions and the following disclaimer.
53 1.15 cgd * 2. Redistributions in binary form must reproduce the above copyright
54 1.15 cgd * notice, this list of conditions and the following disclaimer in the
55 1.15 cgd * documentation and/or other materials provided with the distribution.
56 1.77 agc * 3. Neither the name of the University nor the names of its contributors
57 1.15 cgd * may be used to endorse or promote products derived from this software
58 1.15 cgd * without specific prior written permission.
59 1.15 cgd *
60 1.15 cgd * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
61 1.15 cgd * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
62 1.15 cgd * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
63 1.15 cgd * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
64 1.15 cgd * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
65 1.15 cgd * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
66 1.15 cgd * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
67 1.15 cgd * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
68 1.15 cgd * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
69 1.15 cgd * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
70 1.15 cgd * SUCH DAMAGE.
71 1.15 cgd *
72 1.36 fvdl * @(#)sys_generic.c 8.9 (Berkeley) 2/14/95
73 1.15 cgd */
74 1.59 lukem
75 1.83.2.4 yamt /*
76 1.83.2.4 yamt * System calls relating to files.
77 1.83.2.4 yamt */
78 1.37 thorpej
79 1.83.2.4 yamt #include <sys/cdefs.h>
80 1.83.2.8 yamt __KERNEL_RCSID(0, "$NetBSD: sys_generic.c,v 1.83.2.8 2008/03/17 09:15:34 yamt Exp $");
81 1.15 cgd
82 1.15 cgd #include <sys/param.h>
83 1.15 cgd #include <sys/systm.h>
84 1.15 cgd #include <sys/filedesc.h>
85 1.15 cgd #include <sys/ioctl.h>
86 1.15 cgd #include <sys/file.h>
87 1.15 cgd #include <sys/proc.h>
88 1.15 cgd #include <sys/socketvar.h>
89 1.22 christos #include <sys/signalvar.h>
90 1.15 cgd #include <sys/uio.h>
91 1.15 cgd #include <sys/kernel.h>
92 1.15 cgd #include <sys/stat.h>
93 1.83.2.4 yamt #include <sys/kmem.h>
94 1.28 mycroft #include <sys/poll.h>
95 1.83.2.4 yamt #include <sys/vnode.h>
96 1.16 cgd #include <sys/mount.h>
97 1.16 cgd #include <sys/syscallargs.h>
98 1.83.2.4 yamt #include <sys/ktrace.h>
99 1.22 christos
100 1.83.2.1 yamt #include <uvm/uvm_extern.h>
101 1.83.2.1 yamt
102 1.83.2.4 yamt /* Flags for lwp::l_selflag. */
103 1.83.2.4 yamt #define SEL_RESET 0 /* awoken, interrupted, or not yet polling */
104 1.83.2.4 yamt #define SEL_SCANNING 1 /* polling descriptors */
105 1.83.2.4 yamt #define SEL_BLOCKING 2 /* about to block on select_cv */
106 1.83.2.4 yamt
107 1.83.2.4 yamt static int selscan(lwp_t *, fd_mask *, fd_mask *, int, register_t *);
108 1.83.2.4 yamt static int pollscan(lwp_t *, struct pollfd *, int, register_t *);
109 1.83.2.4 yamt
110 1.83.2.4 yamt /* Global state for select()/poll(). */
111 1.83.2.4 yamt kmutex_t select_lock;
112 1.83.2.4 yamt kcondvar_t select_cv;
113 1.83.2.4 yamt int nselcoll;
114 1.82 matt
115 1.15 cgd /*
116 1.15 cgd * Read system call.
117 1.15 cgd */
118 1.15 cgd /* ARGSUSED */
119 1.22 christos int
120 1.83.2.7 yamt sys_read(struct lwp *l, const struct sys_read_args *uap, register_t *retval)
121 1.20 thorpej {
122 1.83.2.7 yamt /* {
123 1.53 lukem syscallarg(int) fd;
124 1.53 lukem syscallarg(void *) buf;
125 1.53 lukem syscallarg(size_t) nbyte;
126 1.83.2.7 yamt } */
127 1.53 lukem int fd;
128 1.53 lukem struct file *fp;
129 1.83.2.4 yamt proc_t *p;
130 1.53 lukem struct filedesc *fdp;
131 1.39 thorpej
132 1.53 lukem fd = SCARG(uap, fd);
133 1.69 thorpej p = l->l_proc;
134 1.53 lukem fdp = p->p_fd;
135 1.56 thorpej
136 1.56 thorpej if ((fp = fd_getfile(fdp, fd)) == NULL)
137 1.56 thorpej return (EBADF);
138 1.56 thorpej
139 1.70 pk if ((fp->f_flag & FREAD) == 0) {
140 1.83.2.7 yamt FILE_UNLOCK(fp);
141 1.39 thorpej return (EBADF);
142 1.70 pk }
143 1.39 thorpej
144 1.45 thorpej FILE_USE(fp);
145 1.45 thorpej
146 1.45 thorpej /* dofileread() will unuse the descriptor for us */
147 1.83.2.5 yamt return (dofileread(fd, fp, SCARG(uap, buf), SCARG(uap, nbyte),
148 1.39 thorpej &fp->f_offset, FOF_UPDATE_OFFSET, retval));
149 1.39 thorpej }
150 1.39 thorpej
151 1.39 thorpej int
152 1.83.2.5 yamt dofileread(int fd, struct file *fp, void *buf, size_t nbyte,
153 1.53 lukem off_t *offset, int flags, register_t *retval)
154 1.53 lukem {
155 1.83.2.1 yamt struct iovec aiov;
156 1.83.2.1 yamt struct uio auio;
157 1.83.2.1 yamt size_t cnt;
158 1.83.2.1 yamt int error;
159 1.83.2.5 yamt lwp_t *l;
160 1.83.2.1 yamt
161 1.83.2.5 yamt l = curlwp;
162 1.15 cgd
163 1.83.2.4 yamt aiov.iov_base = (void *)buf;
164 1.39 thorpej aiov.iov_len = nbyte;
165 1.15 cgd auio.uio_iov = &aiov;
166 1.15 cgd auio.uio_iovcnt = 1;
167 1.39 thorpej auio.uio_resid = nbyte;
168 1.15 cgd auio.uio_rw = UIO_READ;
169 1.83.2.5 yamt auio.uio_vmspace = l->l_proc->p_vmspace;
170 1.40 thorpej
171 1.40 thorpej /*
172 1.40 thorpej * Reads return ssize_t because -1 is returned on error. Therefore
173 1.40 thorpej * we must restrict the length to SSIZE_MAX to avoid garbage return
174 1.40 thorpej * values.
175 1.40 thorpej */
176 1.45 thorpej if (auio.uio_resid > SSIZE_MAX) {
177 1.45 thorpej error = EINVAL;
178 1.45 thorpej goto out;
179 1.45 thorpej }
180 1.40 thorpej
181 1.38 thorpej cnt = auio.uio_resid;
182 1.39 thorpej error = (*fp->f_ops->fo_read)(fp, offset, &auio, fp->f_cred, flags);
183 1.22 christos if (error)
184 1.15 cgd if (auio.uio_resid != cnt && (error == ERESTART ||
185 1.15 cgd error == EINTR || error == EWOULDBLOCK))
186 1.15 cgd error = 0;
187 1.15 cgd cnt -= auio.uio_resid;
188 1.83.2.4 yamt ktrgenio(fd, UIO_READ, buf, cnt, error);
189 1.15 cgd *retval = cnt;
190 1.45 thorpej out:
191 1.83.2.1 yamt FILE_UNUSE(fp, l);
192 1.15 cgd return (error);
193 1.15 cgd }
194 1.15 cgd
195 1.15 cgd /*
196 1.15 cgd * Scatter read system call.
197 1.15 cgd */
198 1.22 christos int
199 1.83.2.7 yamt sys_readv(struct lwp *l, const struct sys_readv_args *uap, register_t *retval)
200 1.20 thorpej {
201 1.83.2.7 yamt /* {
202 1.53 lukem syscallarg(int) fd;
203 1.53 lukem syscallarg(const struct iovec *) iovp;
204 1.53 lukem syscallarg(int) iovcnt;
205 1.83.2.7 yamt } */
206 1.83.2.4 yamt
207 1.83.2.5 yamt return do_filereadv(SCARG(uap, fd), SCARG(uap, iovp),
208 1.83.2.4 yamt SCARG(uap, iovcnt), NULL, FOF_UPDATE_OFFSET, retval);
209 1.83.2.4 yamt }
210 1.83.2.4 yamt
211 1.83.2.4 yamt int
212 1.83.2.5 yamt do_filereadv(int fd, const struct iovec *iovp, int iovcnt,
213 1.83.2.4 yamt off_t *offset, int flags, register_t *retval)
214 1.83.2.4 yamt {
215 1.83.2.4 yamt struct uio auio;
216 1.83.2.4 yamt struct iovec *iov, *needfree = NULL, aiov[UIO_SMALLIOV];
217 1.83.2.4 yamt int i, error;
218 1.83.2.4 yamt size_t cnt;
219 1.83.2.4 yamt u_int iovlen;
220 1.83.2.4 yamt struct file *fp;
221 1.83.2.4 yamt struct iovec *ktriov = NULL;
222 1.83.2.5 yamt lwp_t *l;
223 1.83.2.4 yamt
224 1.83.2.4 yamt if (iovcnt == 0)
225 1.83.2.4 yamt return EINVAL;
226 1.39 thorpej
227 1.83.2.5 yamt l = curlwp;
228 1.56 thorpej
229 1.83.2.5 yamt if ((fp = fd_getfile(l->l_proc->p_fd, fd)) == NULL)
230 1.83.2.4 yamt return EBADF;
231 1.56 thorpej
232 1.70 pk if ((fp->f_flag & FREAD) == 0) {
233 1.83.2.7 yamt FILE_UNLOCK(fp);
234 1.83.2.4 yamt return EBADF;
235 1.70 pk }
236 1.39 thorpej
237 1.45 thorpej FILE_USE(fp);
238 1.45 thorpej
239 1.83.2.4 yamt if (offset == NULL)
240 1.83.2.4 yamt offset = &fp->f_offset;
241 1.83.2.4 yamt else {
242 1.83.2.4 yamt struct vnode *vp = fp->f_data;
243 1.83.2.4 yamt if (fp->f_type != DTYPE_VNODE || vp->v_type == VFIFO) {
244 1.83.2.4 yamt error = ESPIPE;
245 1.83.2.4 yamt goto out;
246 1.83.2.4 yamt }
247 1.83.2.4 yamt /*
248 1.83.2.4 yamt * Test that the device is seekable ?
249 1.83.2.4 yamt * XXX This works because no file systems actually
250 1.83.2.4 yamt * XXX take any action on the seek operation.
251 1.83.2.4 yamt */
252 1.83.2.4 yamt error = VOP_SEEK(vp, fp->f_offset, *offset, fp->f_cred);
253 1.83.2.4 yamt if (error != 0)
254 1.83.2.4 yamt goto out;
255 1.83.2.4 yamt }
256 1.15 cgd
257 1.42 perry iovlen = iovcnt * sizeof(struct iovec);
258 1.83.2.4 yamt if (flags & FOF_IOV_SYSSPACE)
259 1.83.2.4 yamt iov = __UNCONST(iovp);
260 1.83.2.4 yamt else {
261 1.15 cgd iov = aiov;
262 1.83.2.4 yamt if ((u_int)iovcnt > UIO_SMALLIOV) {
263 1.83.2.4 yamt if ((u_int)iovcnt > IOV_MAX) {
264 1.83.2.4 yamt error = EINVAL;
265 1.83.2.4 yamt goto out;
266 1.83.2.4 yamt }
267 1.83.2.4 yamt iov = kmem_alloc(iovlen, KM_SLEEP);
268 1.83.2.4 yamt if (iov == NULL) {
269 1.83.2.4 yamt error = ENOMEM;
270 1.83.2.4 yamt goto out;
271 1.83.2.4 yamt }
272 1.83.2.4 yamt needfree = iov;
273 1.83.2.4 yamt }
274 1.83.2.4 yamt error = copyin(iovp, iov, iovlen);
275 1.83.2.4 yamt if (error)
276 1.83.2.4 yamt goto done;
277 1.45 thorpej }
278 1.41 kleink
279 1.15 cgd auio.uio_iov = iov;
280 1.34 mycroft auio.uio_iovcnt = iovcnt;
281 1.15 cgd auio.uio_rw = UIO_READ;
282 1.83.2.5 yamt auio.uio_vmspace = l->l_proc->p_vmspace;
283 1.83.2.4 yamt
284 1.15 cgd auio.uio_resid = 0;
285 1.83.2.4 yamt for (i = 0; i < iovcnt; i++, iov++) {
286 1.15 cgd auio.uio_resid += iov->iov_len;
287 1.40 thorpej /*
288 1.40 thorpej * Reads return ssize_t because -1 is returned on error.
289 1.40 thorpej * Therefore we must restrict the length to SSIZE_MAX to
290 1.40 thorpej * avoid garbage return values.
291 1.40 thorpej */
292 1.40 thorpej if (iov->iov_len > SSIZE_MAX || auio.uio_resid > SSIZE_MAX) {
293 1.15 cgd error = EINVAL;
294 1.15 cgd goto done;
295 1.15 cgd }
296 1.15 cgd }
297 1.83.2.4 yamt
298 1.15 cgd /*
299 1.15 cgd * if tracing, save a copy of iovec
300 1.15 cgd */
301 1.83.2.4 yamt if (ktrpoint(KTR_GENIO)) {
302 1.83.2.4 yamt ktriov = kmem_alloc(iovlen, KM_SLEEP);
303 1.83.2.4 yamt if (ktriov != NULL)
304 1.83.2.4 yamt memcpy(ktriov, auio.uio_iov, iovlen);
305 1.15 cgd }
306 1.83.2.4 yamt
307 1.15 cgd cnt = auio.uio_resid;
308 1.39 thorpej error = (*fp->f_ops->fo_read)(fp, offset, &auio, fp->f_cred, flags);
309 1.22 christos if (error)
310 1.15 cgd if (auio.uio_resid != cnt && (error == ERESTART ||
311 1.15 cgd error == EINTR || error == EWOULDBLOCK))
312 1.15 cgd error = 0;
313 1.15 cgd cnt -= auio.uio_resid;
314 1.83.2.4 yamt *retval = cnt;
315 1.83.2.4 yamt
316 1.58 itohy if (ktriov != NULL) {
317 1.83.2.4 yamt ktrgeniov(fd, UIO_READ, ktriov, cnt, error);
318 1.83.2.4 yamt kmem_free(ktriov, iovlen);
319 1.15 cgd }
320 1.83.2.4 yamt
321 1.45 thorpej done:
322 1.15 cgd if (needfree)
323 1.83.2.4 yamt kmem_free(needfree, iovlen);
324 1.45 thorpej out:
325 1.83.2.1 yamt FILE_UNUSE(fp, l);
326 1.15 cgd return (error);
327 1.15 cgd }
328 1.15 cgd
329 1.15 cgd /*
330 1.15 cgd * Write system call
331 1.15 cgd */
332 1.22 christos int
333 1.83.2.7 yamt sys_write(struct lwp *l, const struct sys_write_args *uap, register_t *retval)
334 1.20 thorpej {
335 1.83.2.7 yamt /* {
336 1.53 lukem syscallarg(int) fd;
337 1.53 lukem syscallarg(const void *) buf;
338 1.53 lukem syscallarg(size_t) nbyte;
339 1.83.2.7 yamt } */
340 1.53 lukem int fd;
341 1.53 lukem struct file *fp;
342 1.39 thorpej
343 1.53 lukem fd = SCARG(uap, fd);
344 1.56 thorpej
345 1.83.2.5 yamt if ((fp = fd_getfile(curproc->p_fd, fd)) == NULL)
346 1.56 thorpej return (EBADF);
347 1.56 thorpej
348 1.70 pk if ((fp->f_flag & FWRITE) == 0) {
349 1.83.2.7 yamt FILE_UNLOCK(fp);
350 1.39 thorpej return (EBADF);
351 1.70 pk }
352 1.39 thorpej
353 1.45 thorpej FILE_USE(fp);
354 1.45 thorpej
355 1.45 thorpej /* dofilewrite() will unuse the descriptor for us */
356 1.83.2.5 yamt return (dofilewrite(fd, fp, SCARG(uap, buf), SCARG(uap, nbyte),
357 1.39 thorpej &fp->f_offset, FOF_UPDATE_OFFSET, retval));
358 1.39 thorpej }
359 1.39 thorpej
360 1.39 thorpej int
361 1.83.2.5 yamt dofilewrite(int fd, struct file *fp, const void *buf,
362 1.53 lukem size_t nbyte, off_t *offset, int flags, register_t *retval)
363 1.53 lukem {
364 1.83.2.1 yamt struct iovec aiov;
365 1.83.2.1 yamt struct uio auio;
366 1.83.2.1 yamt size_t cnt;
367 1.83.2.1 yamt int error;
368 1.83.2.5 yamt lwp_t *l;
369 1.83.2.5 yamt
370 1.83.2.5 yamt l = curlwp;
371 1.15 cgd
372 1.83 christos aiov.iov_base = __UNCONST(buf); /* XXXUNCONST kills const */
373 1.39 thorpej aiov.iov_len = nbyte;
374 1.15 cgd auio.uio_iov = &aiov;
375 1.15 cgd auio.uio_iovcnt = 1;
376 1.39 thorpej auio.uio_resid = nbyte;
377 1.15 cgd auio.uio_rw = UIO_WRITE;
378 1.83.2.5 yamt auio.uio_vmspace = l->l_proc->p_vmspace;
379 1.40 thorpej
380 1.40 thorpej /*
381 1.40 thorpej * Writes return ssize_t because -1 is returned on error. Therefore
382 1.40 thorpej * we must restrict the length to SSIZE_MAX to avoid garbage return
383 1.40 thorpej * values.
384 1.40 thorpej */
385 1.45 thorpej if (auio.uio_resid > SSIZE_MAX) {
386 1.45 thorpej error = EINVAL;
387 1.45 thorpej goto out;
388 1.45 thorpej }
389 1.40 thorpej
390 1.38 thorpej cnt = auio.uio_resid;
391 1.39 thorpej error = (*fp->f_ops->fo_write)(fp, offset, &auio, fp->f_cred, flags);
392 1.22 christos if (error) {
393 1.15 cgd if (auio.uio_resid != cnt && (error == ERESTART ||
394 1.15 cgd error == EINTR || error == EWOULDBLOCK))
395 1.15 cgd error = 0;
396 1.83.2.3 yamt if (error == EPIPE) {
397 1.83.2.3 yamt mutex_enter(&proclist_mutex);
398 1.83.2.5 yamt psignal(l->l_proc, SIGPIPE);
399 1.83.2.3 yamt mutex_exit(&proclist_mutex);
400 1.83.2.3 yamt }
401 1.15 cgd }
402 1.15 cgd cnt -= auio.uio_resid;
403 1.83.2.4 yamt ktrgenio(fd, UIO_WRITE, buf, cnt, error);
404 1.15 cgd *retval = cnt;
405 1.45 thorpej out:
406 1.83.2.1 yamt FILE_UNUSE(fp, l);
407 1.15 cgd return (error);
408 1.15 cgd }
409 1.15 cgd
410 1.15 cgd /*
411 1.15 cgd * Gather write system call
412 1.15 cgd */
413 1.22 christos int
414 1.83.2.7 yamt sys_writev(struct lwp *l, const struct sys_writev_args *uap, register_t *retval)
415 1.20 thorpej {
416 1.83.2.7 yamt /* {
417 1.53 lukem syscallarg(int) fd;
418 1.53 lukem syscallarg(const struct iovec *) iovp;
419 1.53 lukem syscallarg(int) iovcnt;
420 1.83.2.7 yamt } */
421 1.83.2.4 yamt
422 1.83.2.5 yamt return do_filewritev(SCARG(uap, fd), SCARG(uap, iovp),
423 1.83.2.4 yamt SCARG(uap, iovcnt), NULL, FOF_UPDATE_OFFSET, retval);
424 1.83.2.4 yamt }
425 1.83.2.4 yamt
426 1.83.2.4 yamt int
427 1.83.2.5 yamt do_filewritev(int fd, const struct iovec *iovp, int iovcnt,
428 1.83.2.4 yamt off_t *offset, int flags, register_t *retval)
429 1.83.2.4 yamt {
430 1.83.2.4 yamt struct uio auio;
431 1.83.2.4 yamt struct iovec *iov, *needfree = NULL, aiov[UIO_SMALLIOV];
432 1.83.2.4 yamt int i, error;
433 1.83.2.4 yamt size_t cnt;
434 1.83.2.4 yamt u_int iovlen;
435 1.83.2.4 yamt struct file *fp;
436 1.83.2.4 yamt struct iovec *ktriov = NULL;
437 1.83.2.5 yamt lwp_t *l;
438 1.83.2.5 yamt
439 1.83.2.5 yamt l = curlwp;
440 1.83.2.4 yamt
441 1.83.2.4 yamt if (iovcnt == 0)
442 1.83.2.4 yamt return EINVAL;
443 1.39 thorpej
444 1.83.2.5 yamt if ((fp = fd_getfile(l->l_proc->p_fd, fd)) == NULL)
445 1.83.2.4 yamt return EBADF;
446 1.56 thorpej
447 1.70 pk if ((fp->f_flag & FWRITE) == 0) {
448 1.83.2.7 yamt FILE_UNLOCK(fp);
449 1.83.2.4 yamt return EBADF;
450 1.70 pk }
451 1.39 thorpej
452 1.45 thorpej FILE_USE(fp);
453 1.45 thorpej
454 1.83.2.4 yamt if (offset == NULL)
455 1.83.2.4 yamt offset = &fp->f_offset;
456 1.83.2.4 yamt else {
457 1.83.2.4 yamt struct vnode *vp = fp->f_data;
458 1.83.2.4 yamt if (fp->f_type != DTYPE_VNODE || vp->v_type == VFIFO) {
459 1.83.2.4 yamt error = ESPIPE;
460 1.83.2.4 yamt goto out;
461 1.83.2.4 yamt }
462 1.83.2.4 yamt /*
463 1.83.2.4 yamt * Test that the device is seekable ?
464 1.83.2.4 yamt * XXX This works because no file systems actually
465 1.83.2.4 yamt * XXX take any action on the seek operation.
466 1.83.2.4 yamt */
467 1.83.2.4 yamt error = VOP_SEEK(vp, fp->f_offset, *offset, fp->f_cred);
468 1.83.2.4 yamt if (error != 0)
469 1.83.2.4 yamt goto out;
470 1.83.2.4 yamt }
471 1.15 cgd
472 1.42 perry iovlen = iovcnt * sizeof(struct iovec);
473 1.83.2.4 yamt if (flags & FOF_IOV_SYSSPACE)
474 1.83.2.4 yamt iov = __UNCONST(iovp);
475 1.83.2.4 yamt else {
476 1.15 cgd iov = aiov;
477 1.83.2.4 yamt if ((u_int)iovcnt > UIO_SMALLIOV) {
478 1.83.2.4 yamt if ((u_int)iovcnt > IOV_MAX) {
479 1.83.2.4 yamt error = EINVAL;
480 1.83.2.4 yamt goto out;
481 1.83.2.4 yamt }
482 1.83.2.4 yamt iov = kmem_alloc(iovlen, KM_SLEEP);
483 1.83.2.4 yamt if (iov == NULL) {
484 1.83.2.4 yamt error = ENOMEM;
485 1.83.2.4 yamt goto out;
486 1.83.2.4 yamt }
487 1.83.2.4 yamt needfree = iov;
488 1.83.2.4 yamt }
489 1.83.2.4 yamt error = copyin(iovp, iov, iovlen);
490 1.83.2.4 yamt if (error)
491 1.83.2.4 yamt goto done;
492 1.45 thorpej }
493 1.41 kleink
494 1.15 cgd auio.uio_iov = iov;
495 1.34 mycroft auio.uio_iovcnt = iovcnt;
496 1.15 cgd auio.uio_rw = UIO_WRITE;
497 1.83.2.5 yamt auio.uio_vmspace = curproc->p_vmspace;
498 1.83.2.4 yamt
499 1.15 cgd auio.uio_resid = 0;
500 1.83.2.4 yamt for (i = 0; i < iovcnt; i++, iov++) {
501 1.15 cgd auio.uio_resid += iov->iov_len;
502 1.40 thorpej /*
503 1.40 thorpej * Writes return ssize_t because -1 is returned on error.
504 1.40 thorpej * Therefore we must restrict the length to SSIZE_MAX to
505 1.40 thorpej * avoid garbage return values.
506 1.40 thorpej */
507 1.40 thorpej if (iov->iov_len > SSIZE_MAX || auio.uio_resid > SSIZE_MAX) {
508 1.15 cgd error = EINVAL;
509 1.15 cgd goto done;
510 1.15 cgd }
511 1.15 cgd }
512 1.83.2.4 yamt
513 1.15 cgd /*
514 1.15 cgd * if tracing, save a copy of iovec
515 1.15 cgd */
516 1.83.2.4 yamt if (ktrpoint(KTR_GENIO)) {
517 1.83.2.4 yamt ktriov = kmem_alloc(iovlen, KM_SLEEP);
518 1.83.2.4 yamt if (ktriov != NULL)
519 1.83.2.4 yamt memcpy(ktriov, auio.uio_iov, iovlen);
520 1.15 cgd }
521 1.83.2.4 yamt
522 1.15 cgd cnt = auio.uio_resid;
523 1.39 thorpej error = (*fp->f_ops->fo_write)(fp, offset, &auio, fp->f_cred, flags);
524 1.22 christos if (error) {
525 1.15 cgd if (auio.uio_resid != cnt && (error == ERESTART ||
526 1.15 cgd error == EINTR || error == EWOULDBLOCK))
527 1.15 cgd error = 0;
528 1.83.2.3 yamt if (error == EPIPE) {
529 1.83.2.3 yamt mutex_enter(&proclist_mutex);
530 1.83.2.5 yamt psignal(l->l_proc, SIGPIPE);
531 1.83.2.3 yamt mutex_exit(&proclist_mutex);
532 1.83.2.3 yamt }
533 1.15 cgd }
534 1.15 cgd cnt -= auio.uio_resid;
535 1.83.2.4 yamt *retval = cnt;
536 1.83.2.4 yamt
537 1.78 drochner if (ktriov != NULL) {
538 1.83.2.4 yamt ktrgeniov(fd, UIO_WRITE, ktriov, cnt, error);
539 1.83.2.4 yamt kmem_free(ktriov, iovlen);
540 1.15 cgd }
541 1.83.2.4 yamt
542 1.45 thorpej done:
543 1.15 cgd if (needfree)
544 1.83.2.4 yamt kmem_free(needfree, iovlen);
545 1.45 thorpej out:
546 1.83.2.1 yamt FILE_UNUSE(fp, l);
547 1.15 cgd return (error);
548 1.15 cgd }
549 1.15 cgd
550 1.15 cgd /*
551 1.15 cgd * Ioctl system call
552 1.15 cgd */
553 1.15 cgd /* ARGSUSED */
554 1.22 christos int
555 1.83.2.7 yamt sys_ioctl(struct lwp *l, const struct sys_ioctl_args *uap, register_t *retval)
556 1.20 thorpej {
557 1.83.2.7 yamt /* {
558 1.53 lukem syscallarg(int) fd;
559 1.53 lukem syscallarg(u_long) com;
560 1.83.2.4 yamt syscallarg(void *) data;
561 1.83.2.7 yamt } */
562 1.53 lukem struct file *fp;
563 1.83.2.4 yamt proc_t *p;
564 1.53 lukem struct filedesc *fdp;
565 1.53 lukem u_long com;
566 1.53 lukem int error;
567 1.53 lukem u_int size;
568 1.83.2.4 yamt void *data, *memp;
569 1.53 lukem #define STK_PARAMS 128
570 1.53 lukem u_long stkbuf[STK_PARAMS/sizeof(u_long)];
571 1.15 cgd
572 1.53 lukem error = 0;
573 1.69 thorpej p = l->l_proc;
574 1.15 cgd fdp = p->p_fd;
575 1.56 thorpej
576 1.56 thorpej if ((fp = fd_getfile(fdp, SCARG(uap, fd))) == NULL)
577 1.15 cgd return (EBADF);
578 1.15 cgd
579 1.45 thorpej FILE_USE(fp);
580 1.45 thorpej
581 1.45 thorpej if ((fp->f_flag & (FREAD | FWRITE)) == 0) {
582 1.45 thorpej error = EBADF;
583 1.65 scw com = 0;
584 1.45 thorpej goto out;
585 1.45 thorpej }
586 1.15 cgd
587 1.16 cgd switch (com = SCARG(uap, com)) {
588 1.15 cgd case FIONCLEX:
589 1.83.2.5 yamt rw_enter(&fdp->fd_lock, RW_WRITER);
590 1.16 cgd fdp->fd_ofileflags[SCARG(uap, fd)] &= ~UF_EXCLOSE;
591 1.83.2.5 yamt rw_exit(&fdp->fd_lock);
592 1.45 thorpej goto out;
593 1.45 thorpej
594 1.15 cgd case FIOCLEX:
595 1.83.2.5 yamt rw_enter(&fdp->fd_lock, RW_WRITER);
596 1.16 cgd fdp->fd_ofileflags[SCARG(uap, fd)] |= UF_EXCLOSE;
597 1.83.2.5 yamt rw_exit(&fdp->fd_lock);
598 1.45 thorpej goto out;
599 1.15 cgd }
600 1.15 cgd
601 1.15 cgd /*
602 1.15 cgd * Interpret high order word to find amount of data to be
603 1.15 cgd * copied to/from the user's address space.
604 1.15 cgd */
605 1.15 cgd size = IOCPARM_LEN(com);
606 1.45 thorpej if (size > IOCPARM_MAX) {
607 1.45 thorpej error = ENOTTY;
608 1.45 thorpej goto out;
609 1.45 thorpej }
610 1.15 cgd memp = NULL;
611 1.42 perry if (size > sizeof(stkbuf)) {
612 1.83.2.4 yamt memp = kmem_alloc(size, KM_SLEEP);
613 1.15 cgd data = memp;
614 1.15 cgd } else
615 1.83.2.4 yamt data = (void *)stkbuf;
616 1.15 cgd if (com&IOC_IN) {
617 1.15 cgd if (size) {
618 1.31 cgd error = copyin(SCARG(uap, data), data, size);
619 1.15 cgd if (error) {
620 1.15 cgd if (memp)
621 1.83.2.4 yamt kmem_free(memp, size);
622 1.45 thorpej goto out;
623 1.15 cgd }
624 1.83.2.4 yamt ktrgenio(SCARG(uap, fd), UIO_WRITE, SCARG(uap, data),
625 1.83.2.4 yamt size, 0);
626 1.15 cgd } else
627 1.83.2.4 yamt *(void **)data = SCARG(uap, data);
628 1.15 cgd } else if ((com&IOC_OUT) && size)
629 1.15 cgd /*
630 1.15 cgd * Zero the buffer so the user always
631 1.15 cgd * gets back something deterministic.
632 1.15 cgd */
633 1.44 perry memset(data, 0, size);
634 1.15 cgd else if (com&IOC_VOID)
635 1.83.2.4 yamt *(void **)data = SCARG(uap, data);
636 1.15 cgd
637 1.15 cgd switch (com) {
638 1.15 cgd
639 1.15 cgd case FIONBIO:
640 1.83.2.7 yamt FILE_LOCK(fp);
641 1.79 jdolecek if (*(int *)data != 0)
642 1.15 cgd fp->f_flag |= FNONBLOCK;
643 1.15 cgd else
644 1.15 cgd fp->f_flag &= ~FNONBLOCK;
645 1.83.2.7 yamt FILE_UNLOCK(fp);
646 1.83.2.1 yamt error = (*fp->f_ops->fo_ioctl)(fp, FIONBIO, data, l);
647 1.15 cgd break;
648 1.15 cgd
649 1.15 cgd case FIOASYNC:
650 1.83.2.7 yamt FILE_LOCK(fp);
651 1.79 jdolecek if (*(int *)data != 0)
652 1.15 cgd fp->f_flag |= FASYNC;
653 1.15 cgd else
654 1.15 cgd fp->f_flag &= ~FASYNC;
655 1.83.2.7 yamt FILE_UNLOCK(fp);
656 1.83.2.1 yamt error = (*fp->f_ops->fo_ioctl)(fp, FIOASYNC, data, l);
657 1.15 cgd break;
658 1.15 cgd
659 1.15 cgd default:
660 1.83.2.1 yamt error = (*fp->f_ops->fo_ioctl)(fp, com, data, l);
661 1.15 cgd /*
662 1.15 cgd * Copy any data to user, size was
663 1.15 cgd * already set and checked above.
664 1.15 cgd */
665 1.73 dsl if (error == 0 && (com&IOC_OUT) && size) {
666 1.31 cgd error = copyout(data, SCARG(uap, data), size);
667 1.83.2.4 yamt ktrgenio(SCARG(uap, fd), UIO_READ, SCARG(uap, data),
668 1.83.2.4 yamt size, error);
669 1.73 dsl }
670 1.15 cgd break;
671 1.15 cgd }
672 1.15 cgd if (memp)
673 1.83.2.4 yamt kmem_free(memp, size);
674 1.45 thorpej out:
675 1.83.2.1 yamt FILE_UNUSE(fp, l);
676 1.61 atatat switch (error) {
677 1.61 atatat case -1:
678 1.61 atatat printf("sys_ioctl: _IO%s%s('%c', %lu, %lu) returned -1: "
679 1.61 atatat "pid=%d comm=%s\n",
680 1.61 atatat (com & IOC_IN) ? "W" : "", (com & IOC_OUT) ? "R" : "",
681 1.61 atatat (char)IOCGROUP(com), (com & 0xff), IOCPARM_LEN(com),
682 1.61 atatat p->p_pid, p->p_comm);
683 1.61 atatat /* FALLTHROUGH */
684 1.61 atatat case EPASSTHROUGH:
685 1.61 atatat error = ENOTTY;
686 1.61 atatat /* FALLTHROUGH */
687 1.61 atatat default:
688 1.61 atatat return (error);
689 1.61 atatat }
690 1.15 cgd }
691 1.15 cgd
692 1.15 cgd /*
693 1.15 cgd * Select system call.
694 1.15 cgd */
695 1.22 christos int
696 1.83.2.7 yamt sys_pselect(struct lwp *l, const struct sys_pselect_args *uap, register_t *retval)
697 1.82 matt {
698 1.83.2.7 yamt /* {
699 1.82 matt syscallarg(int) nd;
700 1.82 matt syscallarg(fd_set *) in;
701 1.82 matt syscallarg(fd_set *) ou;
702 1.82 matt syscallarg(fd_set *) ex;
703 1.82 matt syscallarg(const struct timespec *) ts;
704 1.82 matt syscallarg(sigset_t *) mask;
705 1.83.2.7 yamt } */
706 1.82 matt struct timespec ats;
707 1.82 matt struct timeval atv, *tv = NULL;
708 1.82 matt sigset_t amask, *mask = NULL;
709 1.82 matt int error;
710 1.82 matt
711 1.82 matt if (SCARG(uap, ts)) {
712 1.82 matt error = copyin(SCARG(uap, ts), &ats, sizeof(ats));
713 1.82 matt if (error)
714 1.82 matt return error;
715 1.82 matt atv.tv_sec = ats.tv_sec;
716 1.82 matt atv.tv_usec = ats.tv_nsec / 1000;
717 1.82 matt tv = &atv;
718 1.82 matt }
719 1.82 matt if (SCARG(uap, mask) != NULL) {
720 1.82 matt error = copyin(SCARG(uap, mask), &amask, sizeof(amask));
721 1.82 matt if (error)
722 1.82 matt return error;
723 1.82 matt mask = &amask;
724 1.82 matt }
725 1.82 matt
726 1.82 matt return selcommon(l, retval, SCARG(uap, nd), SCARG(uap, in),
727 1.82 matt SCARG(uap, ou), SCARG(uap, ex), tv, mask);
728 1.82 matt }
729 1.82 matt
730 1.82 matt int
731 1.83.2.2 yamt inittimeleft(struct timeval *tv, struct timeval *sleeptv)
732 1.83.2.2 yamt {
733 1.83.2.2 yamt if (itimerfix(tv))
734 1.83.2.2 yamt return -1;
735 1.83.2.2 yamt getmicrouptime(sleeptv);
736 1.83.2.2 yamt return 0;
737 1.83.2.2 yamt }
738 1.83.2.2 yamt
739 1.83.2.2 yamt int
740 1.83.2.2 yamt gettimeleft(struct timeval *tv, struct timeval *sleeptv)
741 1.83.2.2 yamt {
742 1.83.2.2 yamt /*
743 1.83.2.2 yamt * We have to recalculate the timeout on every retry.
744 1.83.2.2 yamt */
745 1.83.2.2 yamt struct timeval slepttv;
746 1.83.2.2 yamt /*
747 1.83.2.2 yamt * reduce tv by elapsed time
748 1.83.2.2 yamt * based on monotonic time scale
749 1.83.2.2 yamt */
750 1.83.2.2 yamt getmicrouptime(&slepttv);
751 1.83.2.2 yamt timeradd(tv, sleeptv, tv);
752 1.83.2.2 yamt timersub(tv, &slepttv, tv);
753 1.83.2.2 yamt *sleeptv = slepttv;
754 1.83.2.2 yamt return tvtohz(tv);
755 1.83.2.2 yamt }
756 1.83.2.2 yamt
757 1.83.2.2 yamt int
758 1.83.2.7 yamt sys_select(struct lwp *l, const struct sys_select_args *uap, register_t *retval)
759 1.20 thorpej {
760 1.83.2.7 yamt /* {
761 1.53 lukem syscallarg(int) nd;
762 1.53 lukem syscallarg(fd_set *) in;
763 1.53 lukem syscallarg(fd_set *) ou;
764 1.53 lukem syscallarg(fd_set *) ex;
765 1.53 lukem syscallarg(struct timeval *) tv;
766 1.83.2.7 yamt } */
767 1.82 matt struct timeval atv, *tv = NULL;
768 1.82 matt int error;
769 1.82 matt
770 1.82 matt if (SCARG(uap, tv)) {
771 1.83.2.4 yamt error = copyin(SCARG(uap, tv), (void *)&atv,
772 1.82 matt sizeof(atv));
773 1.82 matt if (error)
774 1.82 matt return error;
775 1.82 matt tv = &atv;
776 1.82 matt }
777 1.82 matt
778 1.82 matt return selcommon(l, retval, SCARG(uap, nd), SCARG(uap, in),
779 1.82 matt SCARG(uap, ou), SCARG(uap, ex), tv, NULL);
780 1.82 matt }
781 1.82 matt
782 1.82 matt int
783 1.83.2.4 yamt selcommon(lwp_t *l, register_t *retval, int nd, fd_set *u_in,
784 1.83.2.4 yamt fd_set *u_ou, fd_set *u_ex, struct timeval *tv, sigset_t *mask)
785 1.82 matt {
786 1.53 lukem char smallbits[howmany(FD_SETSIZE, NFDBITS) *
787 1.53 lukem sizeof(fd_mask) * 6];
788 1.83.2.4 yamt proc_t * const p = l->l_proc;
789 1.83.2.4 yamt char *bits;
790 1.83.2.4 yamt int ncoll, error, timo;
791 1.53 lukem size_t ni;
792 1.82 matt sigset_t oldmask;
793 1.83.2.2 yamt struct timeval sleeptv;
794 1.15 cgd
795 1.53 lukem error = 0;
796 1.82 matt if (nd < 0)
797 1.35 thorpej return (EINVAL);
798 1.82 matt if (nd > p->p_fd->fd_nfiles) {
799 1.16 cgd /* forgiving; slightly wrong */
800 1.82 matt nd = p->p_fd->fd_nfiles;
801 1.16 cgd }
802 1.82 matt ni = howmany(nd, NFDBITS) * sizeof(fd_mask);
803 1.27 mycroft if (ni * 6 > sizeof(smallbits))
804 1.83.2.4 yamt bits = kmem_alloc(ni * 6, KM_SLEEP);
805 1.25 mycroft else
806 1.26 cgd bits = smallbits;
807 1.15 cgd
808 1.53 lukem #define getbits(name, x) \
809 1.82 matt if (u_ ## name) { \
810 1.82 matt error = copyin(u_ ## name, bits + ni * x, ni); \
811 1.53 lukem if (error) \
812 1.53 lukem goto done; \
813 1.53 lukem } else \
814 1.44 perry memset(bits + ni * x, 0, ni);
815 1.15 cgd getbits(in, 0);
816 1.15 cgd getbits(ou, 1);
817 1.15 cgd getbits(ex, 2);
818 1.15 cgd #undef getbits
819 1.15 cgd
820 1.65 scw timo = 0;
821 1.83.2.2 yamt if (tv && inittimeleft(tv, &sleeptv) == -1) {
822 1.83.2.1 yamt error = EINVAL;
823 1.83.2.1 yamt goto done;
824 1.65 scw }
825 1.83.2.2 yamt
826 1.83.2.3 yamt if (mask) {
827 1.83.2.3 yamt sigminusset(&sigcantmask, mask);
828 1.83.2.3 yamt mutex_enter(&p->p_smutex);
829 1.83.2.3 yamt oldmask = l->l_sigmask;
830 1.83.2.3 yamt l->l_sigmask = *mask;
831 1.83.2.3 yamt mutex_exit(&p->p_smutex);
832 1.83.2.3 yamt } else
833 1.83.2.3 yamt oldmask = l->l_sigmask; /* XXXgcc */
834 1.65 scw
835 1.83.2.4 yamt mutex_enter(&select_lock);
836 1.83.2.4 yamt SLIST_INIT(&l->l_selwait);
837 1.83.2.4 yamt for (;;) {
838 1.83.2.4 yamt l->l_selflag = SEL_SCANNING;
839 1.83.2.4 yamt ncoll = nselcoll;
840 1.83.2.4 yamt mutex_exit(&select_lock);
841 1.83.2.4 yamt
842 1.83.2.4 yamt error = selscan(l, (fd_mask *)(bits + ni * 0),
843 1.83.2.4 yamt (fd_mask *)(bits + ni * 3), nd, retval);
844 1.83.2.4 yamt
845 1.83.2.4 yamt mutex_enter(&select_lock);
846 1.83.2.4 yamt if (error || *retval)
847 1.83.2.4 yamt break;
848 1.83.2.4 yamt if (tv && (timo = gettimeleft(tv, &sleeptv)) <= 0)
849 1.83.2.4 yamt break;
850 1.83.2.4 yamt if (l->l_selflag != SEL_SCANNING || ncoll != nselcoll)
851 1.83.2.4 yamt continue;
852 1.83.2.4 yamt l->l_selflag = SEL_BLOCKING;
853 1.83.2.4 yamt error = cv_timedwait_sig(&select_cv, &select_lock, timo);
854 1.83.2.4 yamt if (error != 0)
855 1.83.2.4 yamt break;
856 1.83.2.4 yamt }
857 1.83.2.4 yamt selclear();
858 1.83.2.4 yamt mutex_exit(&select_lock);
859 1.83.2.4 yamt
860 1.83.2.3 yamt if (mask) {
861 1.83.2.3 yamt mutex_enter(&p->p_smutex);
862 1.83.2.3 yamt l->l_sigmask = oldmask;
863 1.83.2.3 yamt mutex_exit(&p->p_smutex);
864 1.83.2.3 yamt }
865 1.83.2.4 yamt
866 1.83.2.2 yamt done:
867 1.15 cgd /* select is not restarted after signals... */
868 1.15 cgd if (error == ERESTART)
869 1.15 cgd error = EINTR;
870 1.15 cgd if (error == EWOULDBLOCK)
871 1.15 cgd error = 0;
872 1.83.2.4 yamt if (error == 0 && u_in != NULL)
873 1.83.2.4 yamt error = copyout(bits + ni * 3, u_in, ni);
874 1.83.2.4 yamt if (error == 0 && u_ou != NULL)
875 1.83.2.4 yamt error = copyout(bits + ni * 4, u_ou, ni);
876 1.83.2.4 yamt if (error == 0 && u_ex != NULL)
877 1.83.2.4 yamt error = copyout(bits + ni * 5, u_ex, ni);
878 1.83.2.4 yamt if (bits != smallbits)
879 1.83.2.4 yamt kmem_free(bits, ni * 6);
880 1.15 cgd return (error);
881 1.15 cgd }
882 1.15 cgd
883 1.22 christos int
884 1.83.2.4 yamt selscan(lwp_t *l, fd_mask *ibitp, fd_mask *obitp, int nfd,
885 1.53 lukem register_t *retval)
886 1.53 lukem {
887 1.63 jdolecek static const int flag[3] = { POLLRDNORM | POLLHUP | POLLERR,
888 1.28 mycroft POLLWRNORM | POLLHUP | POLLERR,
889 1.28 mycroft POLLRDBAND };
890 1.83.2.4 yamt proc_t *p = l->l_proc;
891 1.83.2.1 yamt struct filedesc *fdp;
892 1.83.2.1 yamt int msk, i, j, fd, n;
893 1.83.2.1 yamt fd_mask ibits, obits;
894 1.83.2.1 yamt struct file *fp;
895 1.15 cgd
896 1.53 lukem fdp = p->p_fd;
897 1.53 lukem n = 0;
898 1.15 cgd for (msk = 0; msk < 3; msk++) {
899 1.15 cgd for (i = 0; i < nfd; i += NFDBITS) {
900 1.25 mycroft ibits = *ibitp++;
901 1.25 mycroft obits = 0;
902 1.25 mycroft while ((j = ffs(ibits)) && (fd = i + --j) < nfd) {
903 1.25 mycroft ibits &= ~(1 << j);
904 1.56 thorpej if ((fp = fd_getfile(fdp, fd)) == NULL)
905 1.15 cgd return (EBADF);
906 1.45 thorpej FILE_USE(fp);
907 1.83.2.1 yamt if ((*fp->f_ops->fo_poll)(fp, flag[msk], l)) {
908 1.25 mycroft obits |= (1 << j);
909 1.15 cgd n++;
910 1.15 cgd }
911 1.83.2.1 yamt FILE_UNUSE(fp, l);
912 1.15 cgd }
913 1.25 mycroft *obitp++ = obits;
914 1.15 cgd }
915 1.15 cgd }
916 1.15 cgd *retval = n;
917 1.15 cgd return (0);
918 1.15 cgd }
919 1.15 cgd
920 1.28 mycroft /*
921 1.28 mycroft * Poll system call.
922 1.28 mycroft */
923 1.28 mycroft int
924 1.83.2.7 yamt sys_poll(struct lwp *l, const struct sys_poll_args *uap, register_t *retval)
925 1.28 mycroft {
926 1.83.2.7 yamt /* {
927 1.53 lukem syscallarg(struct pollfd *) fds;
928 1.53 lukem syscallarg(u_int) nfds;
929 1.53 lukem syscallarg(int) timeout;
930 1.83.2.7 yamt } */
931 1.82 matt struct timeval atv, *tv = NULL;
932 1.82 matt
933 1.82 matt if (SCARG(uap, timeout) != INFTIM) {
934 1.82 matt atv.tv_sec = SCARG(uap, timeout) / 1000;
935 1.82 matt atv.tv_usec = (SCARG(uap, timeout) % 1000) * 1000;
936 1.82 matt tv = &atv;
937 1.82 matt }
938 1.82 matt
939 1.82 matt return pollcommon(l, retval, SCARG(uap, fds), SCARG(uap, nfds),
940 1.82 matt tv, NULL);
941 1.82 matt }
942 1.82 matt
943 1.82 matt /*
944 1.82 matt * Poll system call.
945 1.82 matt */
946 1.82 matt int
947 1.83.2.7 yamt sys_pollts(struct lwp *l, const struct sys_pollts_args *uap, register_t *retval)
948 1.82 matt {
949 1.83.2.7 yamt /* {
950 1.82 matt syscallarg(struct pollfd *) fds;
951 1.82 matt syscallarg(u_int) nfds;
952 1.82 matt syscallarg(const struct timespec *) ts;
953 1.82 matt syscallarg(const sigset_t *) mask;
954 1.83.2.7 yamt } */
955 1.82 matt struct timespec ats;
956 1.82 matt struct timeval atv, *tv = NULL;
957 1.82 matt sigset_t amask, *mask = NULL;
958 1.82 matt int error;
959 1.82 matt
960 1.82 matt if (SCARG(uap, ts)) {
961 1.82 matt error = copyin(SCARG(uap, ts), &ats, sizeof(ats));
962 1.82 matt if (error)
963 1.82 matt return error;
964 1.82 matt atv.tv_sec = ats.tv_sec;
965 1.82 matt atv.tv_usec = ats.tv_nsec / 1000;
966 1.82 matt tv = &atv;
967 1.82 matt }
968 1.82 matt if (SCARG(uap, mask)) {
969 1.82 matt error = copyin(SCARG(uap, mask), &amask, sizeof(amask));
970 1.82 matt if (error)
971 1.82 matt return error;
972 1.82 matt mask = &amask;
973 1.82 matt }
974 1.82 matt
975 1.82 matt return pollcommon(l, retval, SCARG(uap, fds), SCARG(uap, nfds),
976 1.82 matt tv, mask);
977 1.82 matt }
978 1.82 matt
979 1.82 matt int
980 1.83.2.4 yamt pollcommon(lwp_t *l, register_t *retval,
981 1.82 matt struct pollfd *u_fds, u_int nfds,
982 1.82 matt struct timeval *tv, sigset_t *mask)
983 1.82 matt {
984 1.83.2.1 yamt char smallbits[32 * sizeof(struct pollfd)];
985 1.83.2.4 yamt proc_t * const p = l->l_proc;
986 1.83.2.4 yamt void * bits;
987 1.82 matt sigset_t oldmask;
988 1.83.2.4 yamt int ncoll, error, timo;
989 1.53 lukem size_t ni;
990 1.83.2.2 yamt struct timeval sleeptv;
991 1.28 mycroft
992 1.82 matt if (nfds > p->p_fd->fd_nfiles) {
993 1.28 mycroft /* forgiving; slightly wrong */
994 1.82 matt nfds = p->p_fd->fd_nfiles;
995 1.28 mycroft }
996 1.82 matt ni = nfds * sizeof(struct pollfd);
997 1.28 mycroft if (ni > sizeof(smallbits))
998 1.83.2.4 yamt bits = kmem_alloc(ni, KM_SLEEP);
999 1.28 mycroft else
1000 1.28 mycroft bits = smallbits;
1001 1.28 mycroft
1002 1.82 matt error = copyin(u_fds, bits, ni);
1003 1.28 mycroft if (error)
1004 1.28 mycroft goto done;
1005 1.28 mycroft
1006 1.65 scw timo = 0;
1007 1.83.2.2 yamt if (tv && inittimeleft(tv, &sleeptv) == -1) {
1008 1.83.2.1 yamt error = EINVAL;
1009 1.83.2.1 yamt goto done;
1010 1.65 scw }
1011 1.83.2.2 yamt
1012 1.83.2.3 yamt if (mask) {
1013 1.83.2.3 yamt sigminusset(&sigcantmask, mask);
1014 1.83.2.3 yamt mutex_enter(&p->p_smutex);
1015 1.83.2.3 yamt oldmask = l->l_sigmask;
1016 1.83.2.3 yamt l->l_sigmask = *mask;
1017 1.83.2.3 yamt mutex_exit(&p->p_smutex);
1018 1.83.2.3 yamt } else
1019 1.83.2.3 yamt oldmask = l->l_sigmask; /* XXXgcc */
1020 1.65 scw
1021 1.83.2.4 yamt mutex_enter(&select_lock);
1022 1.83.2.4 yamt SLIST_INIT(&l->l_selwait);
1023 1.83.2.4 yamt for (;;) {
1024 1.83.2.4 yamt ncoll = nselcoll;
1025 1.83.2.4 yamt l->l_selflag = SEL_SCANNING;
1026 1.83.2.4 yamt mutex_exit(&select_lock);
1027 1.83.2.4 yamt
1028 1.83.2.4 yamt error = pollscan(l, (struct pollfd *)bits, nfds, retval);
1029 1.83.2.4 yamt
1030 1.83.2.4 yamt mutex_enter(&select_lock);
1031 1.83.2.4 yamt if (error || *retval)
1032 1.83.2.4 yamt break;
1033 1.83.2.4 yamt if (tv && (timo = gettimeleft(tv, &sleeptv)) <= 0)
1034 1.83.2.4 yamt break;
1035 1.83.2.4 yamt if (l->l_selflag != SEL_SCANNING || nselcoll != ncoll)
1036 1.83.2.4 yamt continue;
1037 1.83.2.4 yamt l->l_selflag = SEL_BLOCKING;
1038 1.83.2.4 yamt error = cv_timedwait_sig(&select_cv, &select_lock, timo);
1039 1.83.2.4 yamt if (error != 0)
1040 1.83.2.4 yamt break;
1041 1.83.2.4 yamt }
1042 1.83.2.4 yamt selclear();
1043 1.83.2.4 yamt mutex_exit(&select_lock);
1044 1.83.2.4 yamt
1045 1.83.2.3 yamt if (mask) {
1046 1.83.2.3 yamt mutex_enter(&p->p_smutex);
1047 1.83.2.3 yamt l->l_sigmask = oldmask;
1048 1.83.2.3 yamt mutex_exit(&p->p_smutex);
1049 1.83.2.3 yamt }
1050 1.83.2.2 yamt done:
1051 1.28 mycroft /* poll is not restarted after signals... */
1052 1.28 mycroft if (error == ERESTART)
1053 1.28 mycroft error = EINTR;
1054 1.28 mycroft if (error == EWOULDBLOCK)
1055 1.28 mycroft error = 0;
1056 1.83.2.4 yamt if (error == 0)
1057 1.82 matt error = copyout(bits, u_fds, ni);
1058 1.83.2.4 yamt if (bits != smallbits)
1059 1.83.2.4 yamt kmem_free(bits, ni);
1060 1.28 mycroft return (error);
1061 1.28 mycroft }
1062 1.28 mycroft
1063 1.28 mycroft int
1064 1.83.2.4 yamt pollscan(lwp_t *l, struct pollfd *fds, int nfd, register_t *retval)
1065 1.53 lukem {
1066 1.83.2.4 yamt proc_t *p = l->l_proc;
1067 1.53 lukem struct filedesc *fdp;
1068 1.53 lukem int i, n;
1069 1.53 lukem struct file *fp;
1070 1.28 mycroft
1071 1.53 lukem fdp = p->p_fd;
1072 1.54 lukem n = 0;
1073 1.28 mycroft for (i = 0; i < nfd; i++, fds++) {
1074 1.60 christos if (fds->fd >= fdp->fd_nfiles) {
1075 1.28 mycroft fds->revents = POLLNVAL;
1076 1.28 mycroft n++;
1077 1.60 christos } else if (fds->fd < 0) {
1078 1.60 christos fds->revents = 0;
1079 1.28 mycroft } else {
1080 1.56 thorpej if ((fp = fd_getfile(fdp, fds->fd)) == NULL) {
1081 1.32 mrg fds->revents = POLLNVAL;
1082 1.28 mycroft n++;
1083 1.32 mrg } else {
1084 1.45 thorpej FILE_USE(fp);
1085 1.32 mrg fds->revents = (*fp->f_ops->fo_poll)(fp,
1086 1.83.2.1 yamt fds->events | POLLERR | POLLHUP, l);
1087 1.32 mrg if (fds->revents != 0)
1088 1.32 mrg n++;
1089 1.83.2.1 yamt FILE_UNUSE(fp, l);
1090 1.32 mrg }
1091 1.28 mycroft }
1092 1.28 mycroft }
1093 1.28 mycroft *retval = n;
1094 1.28 mycroft return (0);
1095 1.28 mycroft }
1096 1.28 mycroft
1097 1.15 cgd /*ARGSUSED*/
1098 1.22 christos int
1099 1.83.2.4 yamt seltrue(dev_t dev, int events, lwp_t *l)
1100 1.15 cgd {
1101 1.15 cgd
1102 1.28 mycroft return (events & (POLLIN | POLLOUT | POLLRDNORM | POLLWRNORM));
1103 1.15 cgd }
1104 1.15 cgd
1105 1.15 cgd /*
1106 1.15 cgd * Record a select request.
1107 1.15 cgd */
1108 1.15 cgd void
1109 1.83.2.4 yamt selrecord(lwp_t *selector, struct selinfo *sip)
1110 1.15 cgd {
1111 1.15 cgd
1112 1.83.2.4 yamt mutex_enter(&select_lock);
1113 1.83.2.4 yamt if (sip->sel_lwp == NULL) {
1114 1.83.2.4 yamt /* First named waiter, although there may be more. */
1115 1.83.2.4 yamt sip->sel_lwp = selector;
1116 1.83.2.4 yamt SLIST_INSERT_HEAD(&selector->l_selwait, sip, sel_chain);
1117 1.83.2.4 yamt } else if (sip->sel_lwp != selector) {
1118 1.83.2.4 yamt /* Multiple waiters. */
1119 1.83.2.4 yamt sip->sel_collision = true;
1120 1.69 thorpej }
1121 1.83.2.4 yamt mutex_exit(&select_lock);
1122 1.15 cgd }
1123 1.15 cgd
1124 1.15 cgd /*
1125 1.15 cgd * Do a wakeup when a selectable event occurs.
1126 1.15 cgd */
1127 1.15 cgd void
1128 1.83.2.8 yamt selnotify(struct selinfo *sip, int events, long knhint)
1129 1.15 cgd {
1130 1.83.2.4 yamt lwp_t *l;
1131 1.15 cgd
1132 1.83.2.4 yamt mutex_enter(&select_lock);
1133 1.73 dsl if (sip->sel_collision) {
1134 1.83.2.4 yamt /* Multiple waiters - just notify everybody. */
1135 1.15 cgd nselcoll++;
1136 1.83.2.4 yamt sip->sel_collision = false;
1137 1.83.2.4 yamt cv_broadcast(&select_cv);
1138 1.83.2.4 yamt } else if (sip->sel_lwp != NULL) {
1139 1.83.2.4 yamt /* Only one LWP waiting. */
1140 1.83.2.4 yamt l = sip->sel_lwp;
1141 1.83.2.4 yamt if (l->l_selflag == SEL_BLOCKING) {
1142 1.83.2.4 yamt /*
1143 1.83.2.4 yamt * If it's sleeping, wake it up. If not, it's
1144 1.83.2.4 yamt * already awake but hasn't yet removed itself
1145 1.83.2.4 yamt * from the selector. We reset the state below
1146 1.83.2.4 yamt * so that we only attempt to do this once.
1147 1.83.2.4 yamt */
1148 1.83.2.4 yamt lwp_lock(l);
1149 1.83.2.4 yamt if (l->l_wchan == &select_cv) {
1150 1.83.2.4 yamt /* lwp_unsleep() releases the LWP lock. */
1151 1.83.2.4 yamt lwp_unsleep(l);
1152 1.83.2.4 yamt } else
1153 1.83.2.4 yamt lwp_unlock(l);
1154 1.83.2.4 yamt } else {
1155 1.83.2.4 yamt /*
1156 1.83.2.4 yamt * Not yet asleep. Reset its state below so that
1157 1.83.2.4 yamt * it will go around again.
1158 1.83.2.4 yamt */
1159 1.83.2.4 yamt }
1160 1.83.2.4 yamt l->l_selflag = SEL_RESET;
1161 1.15 cgd }
1162 1.83.2.4 yamt mutex_exit(&select_lock);
1163 1.83.2.3 yamt
1164 1.83.2.4 yamt KNOTE(&sip->sel_klist, knhint);
1165 1.83.2.4 yamt }
1166 1.83.2.4 yamt
1167 1.83.2.4 yamt /*
1168 1.83.2.4 yamt * Remove an LWP from all objects that it is waiting for.
1169 1.83.2.4 yamt */
1170 1.83.2.8 yamt void
1171 1.83.2.4 yamt selclear(void)
1172 1.83.2.4 yamt {
1173 1.83.2.4 yamt struct selinfo *sip;
1174 1.83.2.4 yamt lwp_t *l = curlwp;
1175 1.83.2.4 yamt
1176 1.83.2.4 yamt KASSERT(mutex_owned(&select_lock));
1177 1.83.2.4 yamt
1178 1.83.2.4 yamt SLIST_FOREACH(sip, &l->l_selwait, sel_chain) {
1179 1.83.2.4 yamt KASSERT(sip->sel_lwp == l);
1180 1.83.2.4 yamt sip->sel_lwp = NULL;
1181 1.15 cgd }
1182 1.83.2.4 yamt }
1183 1.83.2.4 yamt
1184 1.83.2.4 yamt /*
1185 1.83.2.4 yamt * Initialize the select/poll system calls.
1186 1.83.2.4 yamt */
1187 1.83.2.4 yamt void
1188 1.83.2.4 yamt selsysinit(void)
1189 1.83.2.4 yamt {
1190 1.83.2.4 yamt
1191 1.83.2.6 yamt mutex_init(&select_lock, MUTEX_DEFAULT, IPL_VM);
1192 1.83.2.4 yamt cv_init(&select_cv, "select");
1193 1.15 cgd }
1194 1.83.2.5 yamt
1195 1.83.2.5 yamt /*
1196 1.83.2.5 yamt * Initialize a selector.
1197 1.83.2.5 yamt */
1198 1.83.2.5 yamt void
1199 1.83.2.5 yamt selinit(struct selinfo *sip)
1200 1.83.2.5 yamt {
1201 1.83.2.5 yamt
1202 1.83.2.5 yamt memset(sip, 0, sizeof(*sip));
1203 1.83.2.5 yamt }
1204 1.83.2.5 yamt
1205 1.83.2.5 yamt /*
1206 1.83.2.5 yamt * Destroy a selector. The owning object must not gain new
1207 1.83.2.5 yamt * references while this is in progress: all activity on the
1208 1.83.2.5 yamt * selector must be stopped.
1209 1.83.2.5 yamt */
1210 1.83.2.5 yamt void
1211 1.83.2.5 yamt seldestroy(struct selinfo *sip)
1212 1.83.2.5 yamt {
1213 1.83.2.5 yamt lwp_t *l;
1214 1.83.2.5 yamt
1215 1.83.2.5 yamt if (sip->sel_lwp == NULL)
1216 1.83.2.5 yamt return;
1217 1.83.2.5 yamt
1218 1.83.2.5 yamt mutex_enter(&select_lock);
1219 1.83.2.5 yamt if ((l = sip->sel_lwp) != NULL) {
1220 1.83.2.5 yamt /* This should rarely happen, so SLIST_REMOVE() is OK. */
1221 1.83.2.5 yamt SLIST_REMOVE(&l->l_selwait, sip, selinfo, sel_chain);
1222 1.83.2.5 yamt sip->sel_lwp = NULL;
1223 1.83.2.5 yamt }
1224 1.83.2.5 yamt mutex_exit(&select_lock);
1225 1.83.2.5 yamt }
1226