Home | History | Annotate | Line # | Download | only in uvm
uvm_unix.c revision 1.16
      1 /*	$NetBSD: uvm_unix.c,v 1.16 2000/08/24 06:09:25 chs Exp $	*/
      2 
      3 /*
      4  * Copyright (c) 1997 Charles D. Cranor and Washington University.
      5  * Copyright (c) 1991, 1993 The Regents of the University of California.
      6  * Copyright (c) 1988 University of Utah.
      7  *
      8  * All rights reserved.
      9  *
     10  * This code is derived from software contributed to Berkeley by
     11  * the Systems Programming Group of the University of Utah Computer
     12  * Science Department.
     13  *
     14  * Redistribution and use in source and binary forms, with or without
     15  * modification, are permitted provided that the following conditions
     16  * are met:
     17  * 1. Redistributions of source code must retain the above copyright
     18  *    notice, this list of conditions and the following disclaimer.
     19  * 2. Redistributions in binary form must reproduce the above copyright
     20  *    notice, this list of conditions and the following disclaimer in the
     21  *    documentation and/or other materials provided with the distribution.
     22  * 3. All advertising materials mentioning features or use of this software
     23  *    must display the following acknowledgement:
     24  *      This product includes software developed by Charles D. Cranor,
     25  *	Washington University, the University of California, Berkeley and
     26  *	its contributors.
     27  * 4. Neither the name of the University nor the names of its contributors
     28  *    may be used to endorse or promote products derived from this software
     29  *    without specific prior written permission.
     30  *
     31  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
     32  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
     33  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
     34  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
     35  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
     36  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
     37  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
     38  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
     39  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
     40  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
     41  * SUCH DAMAGE.
     42  *
     43  * from: Utah $Hdr: vm_unix.c 1.1 89/11/07$
     44  *      @(#)vm_unix.c   8.1 (Berkeley) 6/11/93
     45  * from: Id: uvm_unix.c,v 1.1.2.2 1997/08/25 18:52:30 chuck Exp
     46  */
     47 
     48 /*
     49  * uvm_unix.c: traditional sbrk/grow interface to vm.
     50  */
     51 #include "opt_compat_netbsd32.h"
     52 
     53 #include <sys/param.h>
     54 #include <sys/systm.h>
     55 #include <sys/proc.h>
     56 #include <sys/resourcevar.h>
     57 #include <sys/vnode.h>
     58 #include <sys/core.h>
     59 
     60 #include <sys/mount.h>
     61 #include <sys/syscallargs.h>
     62 
     63 #include <uvm/uvm.h>
     64 
     65 /*
     66  * sys_obreak: set break
     67  */
     68 
     69 int
     70 sys_obreak(p, v, retval)
     71 	struct proc *p;
     72 	void *v;
     73 	register_t *retval;
     74 {
     75 	struct sys_obreak_args /* {
     76 		syscallarg(char *) nsize;
     77 	} */ *uap = v;
     78 	struct vmspace *vm = p->p_vmspace;
     79 	vaddr_t new, old;
     80 	ssize_t diff;
     81 	int rv;
     82 
     83 	old = (vaddr_t)vm->vm_daddr;
     84 	new = round_page((vaddr_t)SCARG(uap, nsize));
     85 	if ((new - old) > p->p_rlimit[RLIMIT_DATA].rlim_cur)
     86 		return (ENOMEM);
     87 
     88 	old = round_page(old + ptoa(vm->vm_dsize));
     89 	diff = new - old;
     90 
     91 	if (diff == 0)
     92 		return (0);
     93 
     94 	/*
     95 	 * grow or shrink?
     96 	 */
     97 	if (diff > 0) {
     98 		rv = uvm_map(&vm->vm_map, &old, diff, NULL, UVM_UNKNOWN_OFFSET,
     99 		    UVM_MAPFLAG(UVM_PROT_ALL, UVM_PROT_ALL, UVM_INH_COPY,
    100 		    UVM_ADV_NORMAL, UVM_FLAG_AMAPPAD|UVM_FLAG_FIXED|
    101 		    UVM_FLAG_OVERLAY|UVM_FLAG_COPYONW));
    102 		if (rv == KERN_SUCCESS) {
    103 			vm->vm_dsize += atop(diff);
    104 			return (0);
    105 		}
    106 	} else {
    107 		rv = uvm_deallocate(&vm->vm_map, new, -diff);
    108 		if (rv == KERN_SUCCESS) {
    109 			vm->vm_dsize -= atop(-diff);
    110 			return (0);
    111 		}
    112 	}
    113 
    114 	uprintf("sbrk: %s %ld failed, return = %d\n",
    115 	    diff > 0 ? "grow" : "shrink",
    116 	    (long)(diff > 0 ? diff : -diff), rv);
    117 	return (ENOMEM);
    118 }
    119 
    120 /*
    121  * uvm_grow: enlarge the "stack segment" to include sp.
    122  */
    123 
    124 int
    125 uvm_grow(p, sp)
    126 	struct proc *p;
    127 	vaddr_t sp;
    128 {
    129 	struct vmspace *vm = p->p_vmspace;
    130 	int si;
    131 
    132 	/*
    133 	 * For user defined stacks (from sendsig).
    134 	 */
    135 	if (sp < (vaddr_t)vm->vm_maxsaddr)
    136 		return (0);
    137 
    138 	/*
    139 	 * For common case of already allocated (from trap).
    140 	 */
    141 	if (sp >= USRSTACK - ctob(vm->vm_ssize))
    142 		return (1);
    143 
    144 	/*
    145 	 * Really need to check vs limit and increment stack size if ok.
    146 	 */
    147 	si = btoc(USRSTACK-sp) - vm->vm_ssize;
    148 	if (vm->vm_ssize + si > btoc(p->p_rlimit[RLIMIT_STACK].rlim_cur))
    149 		return (0);
    150 	vm->vm_ssize += si;
    151 	return (1);
    152 }
    153 
    154 /*
    155  * sys_oadvise: old advice system call
    156  */
    157 
    158 /* ARGSUSED */
    159 int
    160 sys_ovadvise(p, v, retval)
    161 	struct proc *p;
    162 	void *v;
    163 	register_t *retval;
    164 {
    165 #if 0
    166 	struct sys_ovadvise_args /* {
    167 		syscallarg(int) anom;
    168 	} */ *uap = v;
    169 #endif
    170 
    171 	return (EINVAL);
    172 }
    173 
    174 /*
    175  * uvm_coredump: dump core!
    176  */
    177 
    178 int
    179 uvm_coredump(p, vp, cred, chdr)
    180 	struct proc *p;
    181 	struct vnode *vp;
    182 	struct ucred *cred;
    183 	struct core *chdr;
    184 {
    185 	struct vmspace *vm = p->p_vmspace;
    186 	vm_map_t map = &vm->vm_map;
    187 	vm_map_entry_t entry;
    188 	vaddr_t start, end, maxstack;
    189 	struct coreseg cseg;
    190 	off_t offset;
    191 	int flag, error = 0;
    192 
    193 	offset = chdr->c_hdrsize + chdr->c_seghdrsize + chdr->c_cpusize;
    194 	maxstack = trunc_page(USRSTACK - ctob(vm->vm_ssize));
    195 
    196 	for (entry = map->header.next; entry != &map->header;
    197 	    entry = entry->next) {
    198 
    199 		/* should never happen for a user process */
    200 		if (UVM_ET_ISSUBMAP(entry)) {
    201 			panic("uvm_coredump: user process with submap?");
    202 		}
    203 
    204 		if (!(entry->protection & VM_PROT_WRITE))
    205 			continue;
    206 
    207 		start = entry->start;
    208 		end = entry->end;
    209 
    210 		if (start >= VM_MAXUSER_ADDRESS)
    211 			continue;
    212 
    213 		if (end > VM_MAXUSER_ADDRESS)
    214 			end = VM_MAXUSER_ADDRESS;
    215 
    216 		if (start >= (vaddr_t)vm->vm_maxsaddr) {
    217 			if (end <= maxstack)
    218 				continue;
    219 			if (start < maxstack)
    220 				start = maxstack;
    221 			flag = CORE_STACK;
    222 		} else
    223 			flag = CORE_DATA;
    224 
    225 		/*
    226 		 * Set up a new core file segment.
    227 		 */
    228 		CORE_SETMAGIC(cseg, CORESEGMAGIC, CORE_GETMID(*chdr), flag);
    229 		cseg.c_addr = start;
    230 		cseg.c_size = end - start;
    231 
    232 		error = vn_rdwr(UIO_WRITE, vp,
    233 		    (caddr_t)&cseg, chdr->c_seghdrsize,
    234 		    offset, UIO_SYSSPACE,
    235 		    IO_NODELOCKED|IO_UNIT, cred, NULL, p);
    236 		if (error)
    237 			break;
    238 
    239 		offset += chdr->c_seghdrsize;
    240 		error = vn_rdwr(UIO_WRITE, vp,
    241 		    (caddr_t)cseg.c_addr, (int)cseg.c_size,
    242 		    offset, UIO_USERSPACE,
    243 		    IO_NODELOCKED|IO_UNIT, cred, NULL, p);
    244 		if (error)
    245 			break;
    246 
    247 		offset += cseg.c_size;
    248 		chdr->c_nseg++;
    249 	}
    250 
    251 	return (error);
    252 }
    253 
    254 #if COMPAT_NETBSD32
    255 /*
    256  * uvm_coredump32: dump 32-bit core!
    257  */
    258 
    259 int
    260 uvm_coredump32(p, vp, cred, chdr)
    261 	struct proc *p;
    262 	struct vnode *vp;
    263 	struct ucred *cred;
    264 	struct core32 *chdr;
    265 {
    266 	struct vmspace *vm = p->p_vmspace;
    267 	vm_map_t map = &vm->vm_map;
    268 	vm_map_entry_t entry;
    269 	vaddr_t start, end;
    270 	struct coreseg32 cseg;
    271 	off_t offset;
    272 	int flag, error = 0;
    273 
    274 	offset = chdr->c_hdrsize + chdr->c_seghdrsize + chdr->c_cpusize;
    275 
    276 	for (entry = map->header.next; entry != &map->header;
    277 	    entry = entry->next) {
    278 
    279 		/* should never happen for a user process */
    280 		if (UVM_ET_ISSUBMAP(entry)) {
    281 			panic("uvm_coredump: user process with submap?");
    282 		}
    283 
    284 		if (!(entry->protection & VM_PROT_WRITE))
    285 			continue;
    286 
    287 		start = entry->start;
    288 		end = entry->end;
    289 
    290 		if (start >= VM_MAXUSER_ADDRESS)
    291 			continue;
    292 
    293 		if (end > VM_MAXUSER_ADDRESS)
    294 			end = VM_MAXUSER_ADDRESS;
    295 
    296 		if (start >= (vaddr_t)vm->vm_maxsaddr) {
    297 			flag = CORE_STACK;
    298 			start = trunc_page(USRSTACK - ctob(vm->vm_ssize));
    299 			if (start >= end)
    300 				continue;
    301 		} else
    302 			flag = CORE_DATA;
    303 
    304 		/*
    305 		 * Set up a new core file segment.
    306 		 */
    307 		CORE_SETMAGIC(cseg, CORESEGMAGIC, CORE_GETMID(*chdr), flag);
    308 		cseg.c_addr = start;
    309 		cseg.c_size = end - start;
    310 
    311 		error = vn_rdwr(UIO_WRITE, vp,
    312 		    (caddr_t)&cseg, chdr->c_seghdrsize,
    313 		    offset, UIO_SYSSPACE,
    314 		    IO_NODELOCKED|IO_UNIT, cred, NULL, p);
    315 		if (error)
    316 			break;
    317 
    318 		offset += chdr->c_seghdrsize;
    319 		error = vn_rdwr(UIO_WRITE, vp,
    320 		    (caddr_t)(u_long)cseg.c_addr, (int)cseg.c_size,
    321 		    offset, UIO_USERSPACE,
    322 		    IO_NODELOCKED|IO_UNIT, cred, NULL, p);
    323 		if (error)
    324 			break;
    325 
    326 		offset += cseg.c_size;
    327 		chdr->c_nseg++;
    328 	}
    329 
    330 	return (error);
    331 }
    332 
    333 #endif
    334