xenpmap.h revision 1.3 1 /* $NetBSD: xenpmap.h,v 1.3 2004/04/26 22:05:05 cl Exp $ */
2
3 /*
4 *
5 * Copyright (c) 2004 Christian Limpach.
6 * All rights reserved.
7 *
8 * Redistribution and use in source and binary forms, with or without
9 * modification, are permitted provided that the following conditions
10 * are met:
11 * 1. Redistributions of source code must retain the above copyright
12 * notice, this list of conditions and the following disclaimer.
13 * 2. Redistributions in binary form must reproduce the above copyright
14 * notice, this list of conditions and the following disclaimer in the
15 * documentation and/or other materials provided with the distribution.
16 * 3. All advertising materials mentioning features or use of this software
17 * must display the following acknowledgement:
18 * This product includes software developed by Christian Limpach.
19 * 4. The name of the author may not be used to endorse or promote products
20 * derived from this software without specific prior written permission.
21 *
22 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
23 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
24 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
25 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
26 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
27 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
28 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
29 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
30 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
31 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
32 */
33
34
35 #ifndef _XEN_XENPMAP_H_
36 #define _XEN_XENPMAP_H_
37
38 void xpq_queue_invlpg(vaddr_t);
39 void xpq_queue_pde_update(pd_entry_t *, pd_entry_t);
40 void xpq_queue_pte_update(pt_entry_t *, pt_entry_t);
41 void xpq_queue_unchecked_pte_update(pt_entry_t *, pt_entry_t);
42 void xpq_queue_pt_switch(paddr_t);
43 void xpq_flush_queue(void);
44 void xpq_queue_set_ldt(vaddr_t, uint32_t);
45 void xpq_queue_tlb_flush(void);
46 void xpq_queue_pin_table(paddr_t, int);
47 void xpq_queue_unpin_table(paddr_t);
48
49 extern paddr_t *xpmap_phys_to_machine_mapping;
50
51 #define XPQ_PIN_L1_TABLE 1
52 #define XPQ_PIN_L2_TABLE 2
53
54 #ifndef XEN
55 #define PDE_GET(_pdp) \
56 *(_pdp)
57 #define PDE_SET(_pdp,_npde) \
58 *(_pdp) = (_npde)
59 #define PDE_CLEAR(_pdp) \
60 *(_pdp) = 0
61 #define PTE_SET(_ptp,_npte) \
62 *(_ptp) = (_npte)
63 #define PTE_CLEAR(_ptp) \
64 *(_ptp) = 0
65 #define PTE_ATOMIC_SET(_ptp,_npte,_opte) \
66 (_opte) = x86_atomic_testset_ul((_ptp), (_npte))
67 #define PTE_ATOMIC_CLEAR(_ptp,_opte) \
68 (_opte) = x86_atomic_testset_ul((_ptp), 0)
69 #define PDE_CLEARBITS(_pdp,_bits) \
70 *(_pdp) &= ~(_bits)
71 #define PTE_ATOMIC_CLEARBITS(_ptp,_bits) \
72 x86_atomic_clearbits_l((_ptp), (_bits))
73 #define PTE_SETBITS(_ptp,_bits) \
74 *(_ptp) |= (_bits)
75 #define PTE_ATOMIC_SETBITS(_ptp,_bits) \
76 x86_atomic_setbits_l((_ptp), (_bits))
77 #else
78 paddr_t *xpmap_phys_to_machine_mapping;
79
80 #define PDE_GET(_pdp) \
81 (pmap_valid_entry(*(_pdp)) ? xpmap_mtop(*(_pdp)) : *(_pdp))
82 #define PDE_SET(_pdp,_npde) do { \
83 xpq_queue_pde_update((_pdp), xpmap_ptom((_npde))); \
84 xpq_flush_queue(); \
85 } while (/*CONSTCOND*/0)
86 #define PDE_CLEAR(_pdp) do { \
87 xpq_queue_pde_update((_pdp), 0); \
88 xpq_flush_queue(); \
89 } while (/*CONSTCOND*/0)
90 #define PTE_GET(_ptp) \
91 (pmap_valid_entry(*(_ptp)) ? xpmap_mtop(*(_ptp)) : *(_ptp))
92 #define PTE_GET_MA(_ptp) \
93 *(_ptp)
94 #define PTE_SET(_ptp,_npte) do { \
95 xpq_queue_pte_update((_ptp), xpmap_ptom((_npte))); \
96 xpq_flush_queue(); \
97 } while (/*CONSTCOND*/0)
98 #define PTE_SET_MA(_ptp,_npte) do { \
99 xpq_queue_pte_update((_ptp), (_npte)); \
100 xpq_flush_queue(); \
101 } while (/*CONSTCOND*/0)
102 #define PTE_SET_MA_UNCHECKED(_ptp,_npte) do { \
103 xpq_queue_unchecked_pte_update((_ptp), (_npte)); \
104 xpq_flush_queue(); \
105 } while (/*CONSTCOND*/0)
106 #define PTE_CLEAR(_ptp) do { \
107 xpq_queue_pte_update((_ptp), 0); \
108 xpq_flush_queue(); \
109 } while (/*CONSTCOND*/0)
110 #define PTE_ATOMIC_SET(_ptp,_npte,_opte) do { \
111 (_opte) = PTE_GET(_ptp); \
112 xpq_queue_pte_update((_ptp), xpmap_ptom((_npte))); \
113 xpq_flush_queue(); \
114 } while (/*CONSTCOND*/0)
115 #define PTE_ATOMIC_SET_MA(_ptp,_npte,_opte) do { \
116 (_opte) = *(_ptp); \
117 xpq_queue_pte_update((_ptp), (_npte)); \
118 xpq_flush_queue(); \
119 } while (/*CONSTCOND*/0)
120 #define PTE_ATOMIC_CLEAR(_ptp,_opte) do { \
121 (_opte) = PTE_GET(_ptp); \
122 xpq_queue_pte_update((_ptp), 0); \
123 xpq_flush_queue(); \
124 } while (/*CONSTCOND*/0)
125 #define PTE_ATOMIC_CLEAR_MA(_ptp,_opte) do { \
126 (_opte) = *(_ptp); \
127 xpq_queue_pte_update((_ptp), 0); \
128 xpq_flush_queue(); \
129 } while (/*CONSTCOND*/0)
130 #define PDE_CLEARBITS(_pdp,_bits) do { \
131 xpq_queue_pte_update((_pdp), *(_pdp) & ~((_bits) & ~PG_FRAME)); \
132 xpq_flush_queue(); \
133 } while (/*CONSTCOND*/0)
134 #define PTE_CLEARBITS(_ptp,_bits) do { \
135 xpq_queue_pte_update((_ptp), *(_ptp) & ~((_bits) & ~PG_FRAME)); \
136 xpq_flush_queue(); \
137 } while (/*CONSTCOND*/0)
138 #define PDE_ATOMIC_CLEARBITS(_pdp,_bits) do { \
139 xpq_queue_pde_update((_pdp), *(_pdp) & ~((_bits) & ~PG_FRAME)); \
140 xpq_flush_queue(); \
141 } while (/*CONSTCOND*/0)
142 #define PTE_ATOMIC_CLEARBITS(_ptp,_bits) do { \
143 xpq_queue_pte_update((_ptp), *(_ptp) & ~((_bits) & ~PG_FRAME)); \
144 xpq_flush_queue(); \
145 } while (/*CONSTCOND*/0)
146 #define PTE_SETBITS(_ptp,_bits) do { \
147 xpq_queue_pte_update((_ptp), *(_ptp) | ((_bits) & ~PG_FRAME)); \
148 xpq_flush_queue(); \
149 } while (/*CONSTCOND*/0)
150 #define PDE_ATOMIC_SETBITS(_pdp,_bits) do { \
151 xpq_queue_pde_update((_pdp), *(_pdp) | ((_bits) & ~PG_FRAME)); \
152 xpq_flush_queue(); \
153 } while (/*CONSTCOND*/0)
154 #define PTE_ATOMIC_SETBITS(_ptp,_bits) do { \
155 xpq_queue_pde_update((_ptp), *(_ptp) | ((_bits) & ~PG_FRAME)); \
156 xpq_flush_queue(); \
157 } while (/*CONSTCOND*/0)
158 #define PDE_COPY(_dpdp,_spdp) do { \
159 xpq_queue_pde_update((_dpdp), *(_spdp)); \
160 xpq_flush_queue(); \
161 } while (/*CONSTCOND*/0)
162 #define PTE_UPDATES_FLUSH() do { \
163 xpq_flush_queue(); \
164 } while (/*CONSTCOND*/0)
165
166 #endif
167
168 static __inline paddr_t
169 xpmap_mtop(paddr_t mpa)
170 {
171 return ((machine_to_phys_mapping[mpa >> PAGE_SHIFT] << PAGE_SHIFT) +
172 (KERNTEXTOFF - KERNBASE_LOCORE)) | (mpa & ~PG_FRAME);
173 }
174
175 static __inline paddr_t
176 xpmap_ptom(paddr_t ppa)
177 {
178 return (xpmap_phys_to_machine_mapping[(ppa -
179 (KERNTEXTOFF - KERNBASE_LOCORE)) >> PAGE_SHIFT] << PAGE_SHIFT)
180 | (ppa & ~PG_FRAME);
181 }
182
183 static __inline paddr_t
184 xpmap_ptom_masked(paddr_t ppa)
185 {
186 return (xpmap_phys_to_machine_mapping[(ppa -
187 (KERNTEXTOFF - KERNBASE_LOCORE)) >> PAGE_SHIFT] << PAGE_SHIFT);
188 }
189
190 #endif /* _XEN_XENPMAP_H_ */
191