Home | History | Annotate | Line # | Download | only in include
xenpmap.h revision 1.16
      1 /*	$NetBSD: xenpmap.h,v 1.16 2007/11/22 16:16:58 bouyer Exp $	*/
      2 
      3 /*
      4  *
      5  * Copyright (c) 2004 Christian Limpach.
      6  * All rights reserved.
      7  *
      8  * Redistribution and use in source and binary forms, with or without
      9  * modification, are permitted provided that the following conditions
     10  * are met:
     11  * 1. Redistributions of source code must retain the above copyright
     12  *    notice, this list of conditions and the following disclaimer.
     13  * 2. Redistributions in binary form must reproduce the above copyright
     14  *    notice, this list of conditions and the following disclaimer in the
     15  *    documentation and/or other materials provided with the distribution.
     16  * 3. All advertising materials mentioning features or use of this software
     17  *    must display the following acknowledgement:
     18  *      This product includes software developed by Christian Limpach.
     19  * 4. The name of the author may not be used to endorse or promote products
     20  *    derived from this software without specific prior written permission.
     21  *
     22  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
     23  * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
     24  * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
     25  * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
     26  * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
     27  * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
     28  * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
     29  * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
     30  * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
     31  * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
     32  */
     33 
     34 
     35 #ifndef _XEN_XENPMAP_H_
     36 #define _XEN_XENPMAP_H_
     37 
     38 #define	INVALID_P2M_ENTRY	(~0UL)
     39 
     40 void xpq_queue_machphys_update(paddr_t, paddr_t);
     41 void xpq_queue_invlpg(vaddr_t);
     42 void xpq_queue_pde_update(pd_entry_t *, pd_entry_t);
     43 void xpq_queue_pte_update(pt_entry_t *, pt_entry_t);
     44 void xpq_queue_pt_switch(paddr_t);
     45 void xpq_flush_queue(void);
     46 void xpq_queue_set_ldt(vaddr_t, uint32_t);
     47 void xpq_queue_tlb_flush(void);
     48 void xpq_queue_pin_table(paddr_t);
     49 void xpq_queue_unpin_table(paddr_t);
     50 int  xpq_update_foreign(pt_entry_t *, pt_entry_t, int);
     51 
     52 extern paddr_t *xpmap_phys_to_machine_mapping;
     53 
     54 #ifndef XEN
     55 #define	PDE_GET(_pdp)						\
     56 	*(_pdp)
     57 #define PDE_SET(_pdp,_mapdp,_npde)				\
     58 	*(_mapdp) = (_npde)
     59 #define PDE_CLEAR(_pdp,_mapdp)					\
     60 	*(_mapdp) = 0
     61 #define PTE_SET(_ptp,_maptp,_npte)				\
     62 	*(_maptp) = (_npte)
     63 #define PTE_CLEAR(_ptp,_maptp)					\
     64 	*(_maptp) = 0
     65 #define PTE_ATOMIC_SET(_ptp,_maptp,_npte,_opte)			\
     66 	(_opte) = x86_atomic_testset_ul((_maptp), (_npte))
     67 #define PTE_ATOMIC_CLEAR(_ptp,_maptp,_opte)			\
     68 	(_opte) = x86_atomic_testset_ul((_maptp), 0)
     69 #define PDE_CLEARBITS(_pdp,_mapdp,_bits)			\
     70 	*(_mapdp) &= ~(_bits)
     71 #define PTE_ATOMIC_CLEARBITS(_ptp,_maptp,_bits)			\
     72 	x86_atomic_clearbits_l((_maptp), (_bits))
     73 #define PTE_SETBITS(_ptp,_maptp,_bits)				\
     74 	*(_maptp) |= (_bits)
     75 #define PTE_ATOMIC_SETBITS(_ptp,_maptp,_bits)			\
     76 	x86_atomic_setbits_l((_maptp), (_bits))
     77 #else
     78 paddr_t *xpmap_phys_to_machine_mapping;
     79 
     80 #define	PDE_GET(_pdp)						\
     81 	(pmap_valid_entry(*(_pdp)) ? xpmap_mtop(*(_pdp)) : *(_pdp))
     82 #define PDE_SET(_pdp,_mapdp,_npde) do {				\
     83 	int _s = splvm();					\
     84 	xpq_queue_pde_update((_mapdp), xpmap_ptom((_npde)));	\
     85 	xpq_flush_queue();					\
     86 	splx(_s);						\
     87 } while (/*CONSTCOND*/0)
     88 #define PDE_CLEAR(_pdp,_mapdp) do {				\
     89 	int _s = splvm();					\
     90 	xpq_queue_pde_update((_mapdp), 0);			\
     91 	xpq_flush_queue();					\
     92 	splx(_s);						\
     93 } while (/*CONSTCOND*/0)
     94 #define	PTE_GET(_ptp)						\
     95 	(pmap_valid_entry(*(_ptp)) ? xpmap_mtop(*(_ptp)) : *(_ptp))
     96 #define	PTE_GET_MA(_ptp)					\
     97 	*(_ptp)
     98 #define PTE_SET(_ptp,_maptp,_npte) do {				\
     99 	int _s = splvm();					\
    100 	xpq_queue_pte_update((_maptp), xpmap_ptom((_npte)));	\
    101 	xpq_flush_queue();					\
    102 	splx(_s);						\
    103 } while (/*CONSTCOND*/0)
    104 #define PTE_SET_MA(_ptp,_maptp,_npte) do {			\
    105 	int _s = splvm();					\
    106 	xpq_queue_pte_update((_maptp), (_npte));		\
    107 	xpq_flush_queue();					\
    108 	splx(_s);						\
    109 } while (/*CONSTCOND*/0)
    110 #define PTE_CLEAR(_ptp,_maptp) do {				\
    111 	int _s = splvm();					\
    112 	xpq_queue_pte_update((_maptp), 0);			\
    113 	xpq_flush_queue();					\
    114 	splx(_s);						\
    115 } while (/*CONSTCOND*/0)
    116 #define PTE_ATOMIC_SET(_ptp,_maptp,_npte,_opte) do {		\
    117 	int _s;							\
    118 	(_opte) = PTE_GET(_ptp);				\
    119 	_s = splvm();						\
    120 	xpq_queue_pte_update((_maptp), xpmap_ptom((_npte)));	\
    121 	xpq_flush_queue();					\
    122 	splx(_s);						\
    123 } while (/*CONSTCOND*/0)
    124 #define PTE_ATOMIC_SET_MA(_ptp,_maptp,_npte,_opte) do {		\
    125 	int _s;							\
    126 	(_opte) = *(_ptp);					\
    127 	_s = splvm();						\
    128 	xpq_queue_pte_update((_maptp), (_npte));		\
    129 	xpq_flush_queue();					\
    130 	splx(_s);						\
    131 } while (/*CONSTCOND*/0)
    132 #define PTE_ATOMIC_CLEAR(_ptp,_maptp,_opte) do {		\
    133 	int _s;							\
    134 	(_opte) = PTE_GET(_ptp);				\
    135 	_s = splvm();						\
    136 	xpq_queue_pte_update((_maptp), 0);			\
    137 	xpq_flush_queue();					\
    138 	splx(_s);						\
    139 } while (/*CONSTCOND*/0)
    140 #define PTE_ATOMIC_CLEAR_MA(_ptp,_maptp,_opte) do {		\
    141 	int _s;							\
    142 	(_opte) = *(_ptp);					\
    143 	_s = splvm();						\
    144 	xpq_queue_pte_update((_maptp), 0);			\
    145 	xpq_flush_queue();					\
    146 	splx(_s);						\
    147 } while (/*CONSTCOND*/0)
    148 #define PDE_CLEARBITS(_pdp,_mapdp,_bits) do {			\
    149 	int _s = splvm();					\
    150 	xpq_queue_pte_update((_mapdp), *(_pdp) & ~((_bits) & ~PG_FRAME)); \
    151 	xpq_flush_queue();					\
    152 	splx(_s);						\
    153 } while (/*CONSTCOND*/0)
    154 #define PTE_CLEARBITS(_ptp,_maptp,_bits) do {			\
    155 	int _s = splvm();					\
    156 	xpq_queue_pte_update((_maptp), *(_ptp) & ~((_bits) & ~PG_FRAME)); \
    157 	xpq_flush_queue();					\
    158 	splx(_s);						\
    159 } while (/*CONSTCOND*/0)
    160 #define PDE_ATOMIC_CLEARBITS(_pdp,_mapdp,_bits) do {		\
    161 	int _s = splvm();					\
    162 	xpq_queue_pde_update((_mapdp), *(_pdp) & ~((_bits) & ~PG_FRAME)); \
    163 	xpq_flush_queue();					\
    164 	splx(_s);						\
    165 } while (/*CONSTCOND*/0)
    166 #define PTE_ATOMIC_CLEARBITS(_ptp,_maptp,_bits) do {		\
    167 	int _s = splvm();					\
    168 	xpq_queue_pte_update((_maptp), *(_ptp) & ~((_bits) & ~PG_FRAME)); \
    169 	xpq_flush_queue();					\
    170 	splx(_s);						\
    171 } while (/*CONSTCOND*/0)
    172 #define PTE_SETBITS(_ptp,_maptp,_bits) do {			\
    173 	int _s = splvm();					\
    174 	xpq_queue_pte_update((_maptp), *(_ptp) | ((_bits) & ~PG_FRAME)); \
    175 	xpq_flush_queue();					\
    176 	splx(_s);						\
    177 } while (/*CONSTCOND*/0)
    178 #define PDE_ATOMIC_SETBITS(_pdp,_mapdp,_bits) do {		\
    179 	int _s = splvm();					\
    180 	xpq_queue_pde_update((_mapdp), *(_pdp) | ((_bits) & ~PG_FRAME)); \
    181 	xpq_flush_queue();					\
    182 	splx(_s);						\
    183 } while (/*CONSTCOND*/0)
    184 #define PTE_ATOMIC_SETBITS(_ptp,_maptp,_bits) do {		\
    185 	int _s = splvm();					\
    186 	xpq_queue_pte_update((_maptp), *(_ptp) | ((_bits) & ~PG_FRAME)); \
    187 	xpq_flush_queue();					\
    188 	splx(_s);						\
    189 } while (/*CONSTCOND*/0)
    190 #define PDE_COPY(_dpdp,_madpdp,_spdp) do {			\
    191 	int _s = splvm();					\
    192 	xpq_queue_pde_update((_madpdp), *(_spdp));		\
    193 	xpq_flush_queue();					\
    194 	splx(_s);						\
    195 } while (/*CONSTCOND*/0)
    196 #define	PTE_UPDATES_FLUSH() do {				\
    197 	int _s = splvm();					\
    198 	xpq_flush_queue();					\
    199 	splx(_s);						\
    200 } while (/*CONSTCOND*/0)
    201 
    202 #endif
    203 
    204 /*
    205  * On Xen-2, the start of the day virual memory starts at KERNTEXTOFF
    206  * (0xc0100000). On Xen-3 for domain0 it starts at KERNBASE (0xc0000000).
    207  * So the offset between physical and virtual address is different on
    208  * Xen-2 and Xen-3 for domain0.
    209  * starting with xen-3.0.2, we can add notes so that virual memory starts
    210  * at KERNBASE for domU as well.
    211  */
    212 #if defined(XEN3) && (defined(DOM0OPS) || !defined(XEN_COMPAT_030001))
    213 #define XPMAP_OFFSET	0
    214 #else
    215 #define	XPMAP_OFFSET	(KERNTEXTOFF - KERNBASE)
    216 #endif
    217 
    218 static __inline paddr_t
    219 xpmap_mtop(paddr_t mpa)
    220 {
    221 	return ((machine_to_phys_mapping[mpa >> PAGE_SHIFT] << PAGE_SHIFT) +
    222 	    XPMAP_OFFSET) | (mpa & ~PG_FRAME);
    223 }
    224 
    225 static __inline paddr_t
    226 xpmap_mtop_masked(paddr_t mpa)
    227 {
    228 	return ((machine_to_phys_mapping[mpa >> PAGE_SHIFT] << PAGE_SHIFT) +
    229 	    XPMAP_OFFSET);
    230 }
    231 
    232 static __inline paddr_t
    233 xpmap_ptom(paddr_t ppa)
    234 {
    235 	return (xpmap_phys_to_machine_mapping[(ppa -
    236 	    XPMAP_OFFSET) >> PAGE_SHIFT] << PAGE_SHIFT)
    237 		| (ppa & ~PG_FRAME);
    238 }
    239 
    240 static __inline paddr_t
    241 xpmap_ptom_masked(paddr_t ppa)
    242 {
    243 	return (xpmap_phys_to_machine_mapping[(ppa -
    244 	    XPMAP_OFFSET) >> PAGE_SHIFT] << PAGE_SHIFT);
    245 }
    246 
    247 #ifdef XEN3
    248 static inline void
    249 MULTI_update_va_mapping(
    250 	multicall_entry_t *mcl, vaddr_t va,
    251 	paddr_t new_val, unsigned long flags)
    252 {
    253 	mcl->op = __HYPERVISOR_update_va_mapping;
    254 	mcl->args[0] = va;
    255 #if defined(__x86_64__)
    256 	mcl->args[1] = new_val;
    257 	mcl->args[2] = flags;
    258 #else
    259 	mcl->args[1] = new_val;
    260 	mcl->args[2] = 0;
    261 	mcl->args[3] = flags;
    262 #endif
    263 }
    264 
    265 static inline void
    266 MULTI_update_va_mapping_otherdomain(
    267 	multicall_entry_t *mcl, vaddr_t va,
    268 	paddr_t new_val, unsigned long flags, domid_t domid)
    269 {
    270 	mcl->op = __HYPERVISOR_update_va_mapping_otherdomain;
    271 	mcl->args[0] = va;
    272 #if defined(__x86_64__)
    273 	mcl->args[1] = new_val;
    274 	mcl->args[2] = flags;
    275 	mcl->args[3] = domid;
    276 #else
    277 	mcl->args[1] = new_val;
    278 	mcl->args[2] = 0;
    279 	mcl->args[3] = flags;
    280 	mcl->args[4] = domid;
    281 #endif
    282 }
    283 #if defined(__x86_64__)
    284 #define MULTI_UVMFLAGS_INDEX 2
    285 #define MULTI_UVMDOMID_INDEX 3
    286 #else
    287 #define MULTI_UVMFLAGS_INDEX 3
    288 #define MULTI_UVMDOMID_INDEX 4
    289 #endif
    290 
    291 #if defined(__x86_64__)
    292 void xen_set_user_pgd(paddr_t);
    293 #endif
    294 
    295 #endif /* XEN3 */
    296 
    297 #endif /* _XEN_XENPMAP_H_ */
    298