Home | History | Annotate | Line # | Download | only in uvm
uvm_fault_i.h revision 1.10
      1  1.10      chs /*	$NetBSD: uvm_fault_i.h,v 1.10 2000/01/11 06:57:50 chs Exp $	*/
      2   1.1      mrg 
      3   1.1      mrg /*
      4   1.1      mrg  *
      5   1.1      mrg  * Copyright (c) 1997 Charles D. Cranor and Washington University.
      6   1.1      mrg  * All rights reserved.
      7   1.1      mrg  *
      8   1.1      mrg  * Redistribution and use in source and binary forms, with or without
      9   1.1      mrg  * modification, are permitted provided that the following conditions
     10   1.1      mrg  * are met:
     11   1.1      mrg  * 1. Redistributions of source code must retain the above copyright
     12   1.1      mrg  *    notice, this list of conditions and the following disclaimer.
     13   1.1      mrg  * 2. Redistributions in binary form must reproduce the above copyright
     14   1.1      mrg  *    notice, this list of conditions and the following disclaimer in the
     15   1.1      mrg  *    documentation and/or other materials provided with the distribution.
     16   1.1      mrg  * 3. All advertising materials mentioning features or use of this software
     17   1.1      mrg  *    must display the following acknowledgement:
     18   1.1      mrg  *      This product includes software developed by Charles D. Cranor and
     19   1.1      mrg  *      Washington University.
     20   1.1      mrg  * 4. The name of the author may not be used to endorse or promote products
     21   1.1      mrg  *    derived from this software without specific prior written permission.
     22   1.1      mrg  *
     23   1.1      mrg  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
     24   1.1      mrg  * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
     25   1.1      mrg  * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
     26   1.1      mrg  * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
     27   1.1      mrg  * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
     28   1.1      mrg  * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
     29   1.1      mrg  * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
     30   1.1      mrg  * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
     31   1.1      mrg  * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
     32   1.1      mrg  * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
     33   1.3      mrg  *
     34   1.3      mrg  * from: Id: uvm_fault_i.h,v 1.1.6.1 1997/12/08 16:07:12 chuck Exp
     35   1.1      mrg  */
     36   1.1      mrg 
     37   1.4    perry #ifndef _UVM_UVM_FAULT_I_H_
     38   1.4    perry #define _UVM_UVM_FAULT_I_H_
     39   1.4    perry 
     40   1.1      mrg /*
     41   1.1      mrg  * uvm_fault_i.h: fault inline functions
     42   1.1      mrg  */
     43   1.1      mrg 
     44   1.1      mrg /*
     45   1.1      mrg  * uvmfault_unlockmaps: unlock the maps
     46   1.1      mrg  */
     47   1.1      mrg 
     48   1.5      mrg static __inline void
     49   1.5      mrg uvmfault_unlockmaps(ufi, write_locked)
     50   1.5      mrg 	struct uvm_faultinfo *ufi;
     51   1.5      mrg 	boolean_t write_locked;
     52   1.5      mrg {
     53  1.10      chs 	/*
     54  1.10      chs 	 * ufi can be NULL when this isn't really a fault,
     55  1.10      chs 	 * but merely paging in anon data.
     56  1.10      chs 	 */
     57  1.10      chs 
     58  1.10      chs 	if (ufi == NULL) {
     59  1.10      chs 		return;
     60  1.10      chs 	}
     61   1.1      mrg 
     62   1.5      mrg 	if (write_locked) {
     63   1.5      mrg 		vm_map_unlock(ufi->map);
     64   1.5      mrg 	} else {
     65   1.5      mrg 		vm_map_unlock_read(ufi->map);
     66   1.5      mrg 	}
     67   1.1      mrg }
     68   1.1      mrg 
     69   1.1      mrg /*
     70   1.1      mrg  * uvmfault_unlockall: unlock everything passed in.
     71   1.1      mrg  *
     72   1.1      mrg  * => maps must be read-locked (not write-locked).
     73   1.1      mrg  */
     74   1.1      mrg 
     75   1.5      mrg static __inline void
     76   1.5      mrg uvmfault_unlockall(ufi, amap, uobj, anon)
     77   1.5      mrg 	struct uvm_faultinfo *ufi;
     78   1.5      mrg 	struct vm_amap *amap;
     79   1.5      mrg 	struct uvm_object *uobj;
     80   1.5      mrg 	struct vm_anon *anon;
     81   1.5      mrg {
     82   1.1      mrg 
     83   1.5      mrg 	if (anon)
     84   1.5      mrg 		simple_unlock(&anon->an_lock);
     85   1.5      mrg 	if (uobj)
     86   1.5      mrg 		simple_unlock(&uobj->vmobjlock);
     87   1.5      mrg 	if (amap)
     88   1.7    chuck 		amap_unlock(amap);
     89   1.5      mrg 	uvmfault_unlockmaps(ufi, FALSE);
     90   1.9  thorpej }
     91   1.9  thorpej 
     92   1.9  thorpej /*
     93   1.9  thorpej  * uvmfault_check_intrsafe: check for a virtual address managed by
     94   1.9  thorpej  * an interrupt-safe map.
     95   1.9  thorpej  *
     96   1.9  thorpej  * => caller must provide a uvm_faultinfo structure with the IN
     97   1.9  thorpej  *	params properly filled in
     98   1.9  thorpej  * => if we find an intersafe VA, we fill in ufi->map, and return TRUE
     99   1.9  thorpej  */
    100   1.9  thorpej 
    101   1.9  thorpej static __inline boolean_t
    102   1.9  thorpej uvmfault_check_intrsafe(ufi)
    103   1.9  thorpej 	struct uvm_faultinfo *ufi;
    104   1.9  thorpej {
    105   1.9  thorpej 	struct vm_map_intrsafe *vmi;
    106   1.9  thorpej 	int s;
    107   1.9  thorpej 
    108   1.9  thorpej 	s = vmi_list_lock();
    109   1.9  thorpej 	for (vmi = LIST_FIRST(&vmi_list); vmi != NULL;
    110   1.9  thorpej 	     vmi = LIST_NEXT(vmi, vmi_list)) {
    111   1.9  thorpej 		if (ufi->orig_rvaddr >= vm_map_min(&vmi->vmi_map) &&
    112   1.9  thorpej 		    ufi->orig_rvaddr < vm_map_max(&vmi->vmi_map))
    113   1.9  thorpej 			break;
    114   1.9  thorpej 	}
    115   1.9  thorpej 	vmi_list_unlock(s);
    116   1.9  thorpej 
    117   1.9  thorpej 	if (vmi != NULL) {
    118   1.9  thorpej 		ufi->map = &vmi->vmi_map;
    119   1.9  thorpej 		return (TRUE);
    120   1.9  thorpej 	}
    121   1.9  thorpej 
    122   1.9  thorpej 	return (FALSE);
    123   1.1      mrg }
    124   1.1      mrg 
    125   1.1      mrg /*
    126   1.1      mrg  * uvmfault_lookup: lookup a virtual address in a map
    127   1.1      mrg  *
    128   1.6    chuck  * => caller must provide a uvm_faultinfo structure with the IN
    129   1.1      mrg  *	params properly filled in
    130   1.6    chuck  * => we will lookup the map entry (handling submaps) as we go
    131   1.1      mrg  * => if the lookup is a success we will return with the maps locked
    132   1.1      mrg  * => if "write_lock" is TRUE, we write_lock the map, otherwise we only
    133   1.1      mrg  *	get a read lock.
    134   1.6    chuck  * => note that submaps can only appear in the kernel and they are
    135   1.6    chuck  *	required to use the same virtual addresses as the map they
    136   1.6    chuck  *	are referenced by (thus address translation between the main
    137   1.6    chuck  *	map and the submap is unnecessary).
    138   1.1      mrg  */
    139   1.1      mrg 
    140   1.5      mrg static __inline boolean_t
    141   1.5      mrg uvmfault_lookup(ufi, write_lock)
    142   1.5      mrg 	struct uvm_faultinfo *ufi;
    143   1.5      mrg 	boolean_t write_lock;
    144   1.1      mrg {
    145   1.5      mrg 	vm_map_t tmpmap;
    146   1.1      mrg 
    147   1.5      mrg 	/*
    148   1.5      mrg 	 * init ufi values for lookup.
    149   1.5      mrg 	 */
    150   1.5      mrg 
    151   1.5      mrg 	ufi->map = ufi->orig_map;
    152   1.5      mrg 	ufi->size = ufi->orig_size;
    153   1.5      mrg 
    154   1.5      mrg 	/*
    155   1.5      mrg 	 * keep going down levels until we are done.   note that there can
    156   1.5      mrg 	 * only be two levels so we won't loop very long.
    157   1.5      mrg 	 */
    158   1.5      mrg 
    159   1.5      mrg 	while (1) {
    160   1.5      mrg 
    161   1.5      mrg 		/*
    162   1.5      mrg 		 * lock map
    163   1.5      mrg 		 */
    164   1.5      mrg 		if (write_lock) {
    165   1.5      mrg 			vm_map_lock(ufi->map);
    166   1.5      mrg 		} else {
    167   1.5      mrg 			vm_map_lock_read(ufi->map);
    168   1.5      mrg 		}
    169   1.5      mrg 
    170   1.5      mrg 		/*
    171   1.5      mrg 		 * lookup
    172   1.5      mrg 		 */
    173   1.6    chuck 		if (!uvm_map_lookup_entry(ufi->map, ufi->orig_rvaddr,
    174   1.6    chuck 								&ufi->entry)) {
    175   1.5      mrg 			uvmfault_unlockmaps(ufi, write_lock);
    176   1.5      mrg 			return(FALSE);
    177   1.5      mrg 		}
    178   1.5      mrg 
    179   1.5      mrg 		/*
    180   1.5      mrg 		 * reduce size if necessary
    181   1.5      mrg 		 */
    182   1.6    chuck 		if (ufi->entry->end - ufi->orig_rvaddr < ufi->size)
    183   1.6    chuck 			ufi->size = ufi->entry->end - ufi->orig_rvaddr;
    184   1.5      mrg 
    185   1.5      mrg 		/*
    186   1.5      mrg 		 * submap?    replace map with the submap and lookup again.
    187   1.5      mrg 		 * note: VAs in submaps must match VAs in main map.
    188   1.5      mrg 		 */
    189   1.5      mrg 		if (UVM_ET_ISSUBMAP(ufi->entry)) {
    190   1.5      mrg 			tmpmap = ufi->entry->object.sub_map;
    191   1.5      mrg 			if (write_lock) {
    192   1.5      mrg 				vm_map_unlock(ufi->map);
    193   1.5      mrg 			} else {
    194   1.5      mrg 				vm_map_unlock_read(ufi->map);
    195   1.5      mrg 			}
    196   1.5      mrg 			ufi->map = tmpmap;
    197   1.5      mrg 			continue;
    198   1.5      mrg 		}
    199   1.5      mrg 
    200   1.5      mrg 		/*
    201   1.5      mrg 		 * got it!
    202   1.5      mrg 		 */
    203   1.1      mrg 
    204   1.5      mrg 		ufi->mapv = ufi->map->timestamp;
    205   1.5      mrg 		return(TRUE);
    206   1.1      mrg 
    207   1.5      mrg 	}	/* while loop */
    208   1.1      mrg 
    209   1.5      mrg 	/*NOTREACHED*/
    210   1.1      mrg }
    211   1.1      mrg 
    212   1.1      mrg /*
    213   1.1      mrg  * uvmfault_relock: attempt to relock the same version of the map
    214   1.1      mrg  *
    215   1.1      mrg  * => fault data structures should be unlocked before calling.
    216   1.1      mrg  * => if a success (TRUE) maps will be locked after call.
    217   1.1      mrg  */
    218   1.1      mrg 
    219   1.5      mrg static __inline boolean_t
    220   1.5      mrg uvmfault_relock(ufi)
    221   1.5      mrg 	struct uvm_faultinfo *ufi;
    222   1.5      mrg {
    223  1.10      chs 	/*
    224  1.10      chs 	 * ufi can be NULL when this isn't really a fault,
    225  1.10      chs 	 * but merely paging in anon data.
    226  1.10      chs 	 */
    227  1.10      chs 
    228  1.10      chs 	if (ufi == NULL) {
    229  1.10      chs 		return TRUE;
    230  1.10      chs 	}
    231   1.1      mrg 
    232   1.5      mrg 	uvmexp.fltrelck++;
    233  1.10      chs 
    234   1.5      mrg 	/*
    235   1.6    chuck 	 * relock map.   fail if version mismatch (in which case nothing
    236   1.6    chuck 	 * gets locked).
    237   1.5      mrg 	 */
    238   1.5      mrg 
    239   1.5      mrg 	vm_map_lock_read(ufi->map);
    240   1.5      mrg 	if (ufi->mapv != ufi->map->timestamp) {
    241   1.5      mrg 		vm_map_unlock_read(ufi->map);
    242   1.5      mrg 		return(FALSE);
    243   1.5      mrg 	}
    244   1.1      mrg 
    245   1.5      mrg 	uvmexp.fltrelckok++;
    246   1.5      mrg 	return(TRUE);		/* got it! */
    247   1.1      mrg }
    248   1.4    perry 
    249   1.4    perry #endif /* _UVM_UVM_FAULT_I_H_ */
    250