cia_dma.c revision 1.32 1 /* $NetBSD: cia_dma.c,v 1.32 2021/05/05 02:15:18 thorpej Exp $ */
2
3 /*-
4 * Copyright (c) 1997, 1998 The NetBSD Foundation, Inc.
5 * All rights reserved.
6 *
7 * This code is derived from software contributed to The NetBSD Foundation
8 * by Jason R. Thorpe of the Numerical Aerospace Simulation Facility,
9 * NASA Ames Research Center.
10 *
11 * Redistribution and use in source and binary forms, with or without
12 * modification, are permitted provided that the following conditions
13 * are met:
14 * 1. Redistributions of source code must retain the above copyright
15 * notice, this list of conditions and the following disclaimer.
16 * 2. Redistributions in binary form must reproduce the above copyright
17 * notice, this list of conditions and the following disclaimer in the
18 * documentation and/or other materials provided with the distribution.
19 *
20 * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
21 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
22 * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
23 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
24 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
25 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
26 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
27 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
28 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
29 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
30 * POSSIBILITY OF SUCH DAMAGE.
31 */
32
33 #include <sys/cdefs.h> /* RCS ID & Copyright macro defns */
34
35 __KERNEL_RCSID(0, "$NetBSD: cia_dma.c,v 1.32 2021/05/05 02:15:18 thorpej Exp $");
36
37 #include <sys/param.h>
38 #include <sys/systm.h>
39 #include <sys/kernel.h>
40 #include <sys/device.h>
41 #include <sys/malloc.h>
42
43 #define _ALPHA_BUS_DMA_PRIVATE
44 #include <sys/bus.h>
45
46 #include <dev/pci/pcireg.h>
47 #include <dev/pci/pcivar.h>
48 #include <alpha/pci/ciareg.h>
49 #include <alpha/pci/ciavar.h>
50
51 static bus_dma_tag_t cia_dma_get_tag(bus_dma_tag_t, alpha_bus_t);
52
53 static int cia_bus_dmamap_create_direct(bus_dma_tag_t, bus_size_t, int,
54 bus_size_t, bus_size_t, int, bus_dmamap_t *);
55
56 static int cia_bus_dmamap_load_sgmap(bus_dma_tag_t, bus_dmamap_t, void *,
57 bus_size_t, struct proc *, int);
58
59 static int cia_bus_dmamap_load_mbuf_sgmap(bus_dma_tag_t, bus_dmamap_t,
60 struct mbuf *, int);
61
62 static int cia_bus_dmamap_load_uio_sgmap(bus_dma_tag_t, bus_dmamap_t,
63 struct uio *, int);
64
65 static int cia_bus_dmamap_load_raw_sgmap(bus_dma_tag_t, bus_dmamap_t,
66 bus_dma_segment_t *, int, bus_size_t, int);
67
68 static void cia_bus_dmamap_unload_sgmap(bus_dma_tag_t, bus_dmamap_t);
69
70 /*
71 * Direct-mapped window: 1G at 1G
72 */
73 #define CIA_DIRECT_MAPPED_BASE (1*1024*1024*1024)
74 #define CIA_DIRECT_MAPPED_SIZE (1*1024*1024*1024)
75
76 /*
77 * SGMAP window: 8M at 8M
78 */
79 #define CIA_SGMAP_MAPPED_BASE (8*1024*1024)
80 #define CIA_SGMAP_MAPPED_SIZE (8*1024*1024)
81
82 /* ALCOR/ALGOR2/PYXIS have a 256-byte out-bound DMA prefetch threshold. */
83 #define CIA_SGMAP_PFTHRESH 256
84
85 static void cia_tlb_invalidate(void);
86 static void cia_broken_pyxis_tlb_invalidate(void);
87
88 static void (*cia_tlb_invalidate_fn)(void);
89
90 #define CIA_TLB_INVALIDATE() (*cia_tlb_invalidate_fn)()
91
92 struct alpha_sgmap cia_pyxis_bug_sgmap;
93 #define CIA_PYXIS_BUG_BASE (128*1024*1024)
94 #define CIA_PYXIS_BUG_SIZE (2*1024*1024)
95
96 void
97 cia_dma_init(struct cia_config *ccp)
98 {
99 bus_addr_t tbase;
100 bus_dma_tag_t t;
101
102 /*
103 * Initialize the DMA tag used for direct-mapped DMA.
104 */
105 t = &ccp->cc_dmat_direct;
106 t->_cookie = ccp;
107 t->_wbase = CIA_DIRECT_MAPPED_BASE;
108 t->_wsize = CIA_DIRECT_MAPPED_SIZE;
109 t->_next_window = &ccp->cc_dmat_sgmap;
110 t->_boundary = 0;
111 t->_sgmap = NULL;
112 t->_get_tag = cia_dma_get_tag;
113 t->_dmamap_create = cia_bus_dmamap_create_direct;
114 t->_dmamap_destroy = _bus_dmamap_destroy;
115 t->_dmamap_load = _bus_dmamap_load_direct;
116 t->_dmamap_load_mbuf = _bus_dmamap_load_mbuf_direct;
117 t->_dmamap_load_uio = _bus_dmamap_load_uio_direct;
118 t->_dmamap_load_raw = _bus_dmamap_load_raw_direct;
119 t->_dmamap_unload = _bus_dmamap_unload;
120 t->_dmamap_sync = _bus_dmamap_sync;
121
122 t->_dmamem_alloc = _bus_dmamem_alloc;
123 t->_dmamem_free = _bus_dmamem_free;
124 t->_dmamem_map = _bus_dmamem_map;
125 t->_dmamem_unmap = _bus_dmamem_unmap;
126 t->_dmamem_mmap = _bus_dmamem_mmap;
127
128 /*
129 * Initialize the DMA tag used for sgmap-mapped DMA.
130 */
131 t = &ccp->cc_dmat_sgmap;
132 t->_cookie = ccp;
133 t->_wbase = CIA_SGMAP_MAPPED_BASE;
134 t->_wsize = CIA_SGMAP_MAPPED_SIZE;
135 t->_next_window = NULL;
136 t->_boundary = 0;
137 t->_sgmap = &ccp->cc_sgmap;
138 t->_pfthresh = CIA_SGMAP_PFTHRESH;
139 t->_get_tag = cia_dma_get_tag;
140 t->_dmamap_create = alpha_sgmap_dmamap_create;
141 t->_dmamap_destroy = alpha_sgmap_dmamap_destroy;
142 t->_dmamap_load = cia_bus_dmamap_load_sgmap;
143 t->_dmamap_load_mbuf = cia_bus_dmamap_load_mbuf_sgmap;
144 t->_dmamap_load_uio = cia_bus_dmamap_load_uio_sgmap;
145 t->_dmamap_load_raw = cia_bus_dmamap_load_raw_sgmap;
146 t->_dmamap_unload = cia_bus_dmamap_unload_sgmap;
147 t->_dmamap_sync = _bus_dmamap_sync;
148
149 t->_dmamem_alloc = _bus_dmamem_alloc;
150 t->_dmamem_free = _bus_dmamem_free;
151 t->_dmamem_map = _bus_dmamem_map;
152 t->_dmamem_unmap = _bus_dmamem_unmap;
153 t->_dmamem_mmap = _bus_dmamem_mmap;
154
155 /*
156 * The firmware has set up window 1 as a 1G direct-mapped DMA
157 * window beginning at 1G. We leave it alone. Leave window
158 * 0 alone until we reconfigure it for SGMAP-mapped DMA.
159 * Windows 2 and 3 are already disabled.
160 */
161
162 /*
163 * Initialize the SGMAP. Must align page table to 32k
164 * (hardware bug?).
165 */
166 alpha_sgmap_init(t, &ccp->cc_sgmap, "cia_sgmap",
167 CIA_SGMAP_MAPPED_BASE, 0, CIA_SGMAP_MAPPED_SIZE,
168 sizeof(uint64_t), NULL, (32*1024));
169
170 /*
171 * Set up window 0 as an 8MB SGMAP-mapped window
172 * starting at 8MB.
173 */
174 REGVAL(CIA_PCI_W0BASE) = CIA_SGMAP_MAPPED_BASE |
175 CIA_PCI_WnBASE_SG_EN | CIA_PCI_WnBASE_W_EN;
176 alpha_mb();
177
178 REGVAL(CIA_PCI_W0MASK) = CIA_PCI_WnMASK_8M;
179 alpha_mb();
180
181 tbase = ccp->cc_sgmap.aps_ptpa >> CIA_PCI_TnBASE_SHIFT;
182 if ((tbase & CIA_PCI_TnBASE_MASK) != tbase)
183 panic("cia_dma_init: bad page table address");
184 REGVAL(CIA_PCI_T0BASE) = tbase;
185 alpha_mb();
186
187 /*
188 * Pass 1 and 2 (i.e. revision <= 1) of the Pyxis have a
189 * broken scatter/gather TLB; it cannot be invalidated. To
190 * work around this problem, we configure window 2 as an SG
191 * 2M window at 128M, which we use in DMA loopback mode to
192 * read a spill page. This works by causing TLB misses,
193 * causing the old entries to be purged to make room for
194 * the new entries coming in for the spill page.
195 */
196 if ((ccp->cc_flags & CCF_ISPYXIS) != 0 && ccp->cc_rev <= 1) {
197 uint64_t *page_table;
198 int i;
199
200 cia_tlb_invalidate_fn =
201 cia_broken_pyxis_tlb_invalidate;
202
203 alpha_sgmap_init(t, &cia_pyxis_bug_sgmap,
204 "pyxis_bug_sgmap", CIA_PYXIS_BUG_BASE, 0,
205 CIA_PYXIS_BUG_SIZE, sizeof(uint64_t), NULL,
206 (32*1024));
207
208 REGVAL(CIA_PCI_W2BASE) = CIA_PYXIS_BUG_BASE |
209 CIA_PCI_WnBASE_SG_EN | CIA_PCI_WnBASE_W_EN;
210 alpha_mb();
211
212 REGVAL(CIA_PCI_W2MASK) = CIA_PCI_WnMASK_2M;
213 alpha_mb();
214
215 tbase = cia_pyxis_bug_sgmap.aps_ptpa >>
216 CIA_PCI_TnBASE_SHIFT;
217 if ((tbase & CIA_PCI_TnBASE_MASK) != tbase)
218 panic("cia_dma_init: bad page table address");
219 REGVAL(CIA_PCI_T2BASE) = tbase;
220 alpha_mb();
221
222 /*
223 * Initialize the page table to point at the spill
224 * page. Leave the last entry invalid.
225 */
226 pci_sgmap_pte64_init_spill_page_pte();
227 for (i = 0, page_table = cia_pyxis_bug_sgmap.aps_pt;
228 i < (CIA_PYXIS_BUG_SIZE / PAGE_SIZE) - 1; i++) {
229 page_table[i] =
230 pci_sgmap_pte64_prefetch_spill_page_pte;
231 }
232 alpha_mb();
233 } else
234 cia_tlb_invalidate_fn = cia_tlb_invalidate;
235
236 CIA_TLB_INVALIDATE();
237 }
238
239 /*
240 * Return the bus dma tag to be used for the specified bus type.
241 * INTERNAL USE ONLY!
242 */
243 static bus_dma_tag_t
244 cia_dma_get_tag(bus_dma_tag_t t, alpha_bus_t bustype)
245 {
246 struct cia_config *ccp = t->_cookie;
247
248 switch (bustype) {
249 case ALPHA_BUS_PCI:
250 case ALPHA_BUS_EISA:
251 /*
252 * Systems with a CIA can only support 1G
253 * of memory, so we use the direct-mapped window
254 * on busses that have 32-bit DMA.
255 *
256 * Ahem: I have a PWS 500au with 1.5G of memory, and it
257 * had problems doing DMA because it was not falling back
258 * to using SGMAPs. I've fixed that and my PWS now works with
259 * 1.5G. There have been other reports about failures with
260 * more than 1.0G of memory. Michael Hitch
261 */
262 return (&ccp->cc_dmat_direct);
263
264 case ALPHA_BUS_ISA:
265 /*
266 * ISA doesn't have enough address bits to use
267 * the direct-mapped DMA window, so we must use
268 * SGMAPs.
269 */
270 return (&ccp->cc_dmat_sgmap);
271
272 default:
273 panic("cia_dma_get_tag: shouldn't be here, really...");
274 }
275 }
276
277 /*
278 * Create a CIA direct-mapped DMA map.
279 */
280 static int
281 cia_bus_dmamap_create_direct(
282 bus_dma_tag_t t,
283 bus_size_t size,
284 int nsegments,
285 bus_size_t maxsegsz,
286 bus_size_t boundary,
287 int flags,
288 bus_dmamap_t *dmamp)
289 {
290 struct cia_config *ccp = t->_cookie;
291 bus_dmamap_t map;
292 int error;
293
294 error = _bus_dmamap_create(t, size, nsegments, maxsegsz,
295 boundary, flags, dmamp);
296 if (error)
297 return (error);
298
299 map = *dmamp;
300
301 if ((ccp->cc_flags & CCF_PYXISBUG) != 0 &&
302 map->_dm_segcnt > 1) {
303 /*
304 * We have a Pyxis with the DMA page crossing bug, make
305 * sure we don't coalesce adjacent DMA segments.
306 *
307 * NOTE: We can only do this if the max segment count
308 * is greater than 1. This is because many network
309 * drivers allocate large contiguous blocks of memory
310 * for control data structures, even though they won't
311 * do any single DMA that crosses a page coundary.
312 * -- thorpej (at) NetBSD.org, 2/5/2000
313 */
314 map->_dm_flags |= DMAMAP_NO_COALESCE;
315 }
316
317 return (0);
318 }
319
320 /*
321 * Load a CIA SGMAP-mapped DMA map with a linear buffer.
322 */
323 static int
324 cia_bus_dmamap_load_sgmap(bus_dma_tag_t t, bus_dmamap_t map, void *buf, bus_size_t buflen, struct proc *p, int flags)
325 {
326 int error;
327
328 error = pci_sgmap_pte64_load(t, map, buf, buflen, p, flags,
329 t->_sgmap);
330 if (error == 0)
331 CIA_TLB_INVALIDATE();
332
333 return (error);
334 }
335
336 /*
337 * Load a CIA SGMAP-mapped DMA map with an mbuf chain.
338 */
339 static int
340 cia_bus_dmamap_load_mbuf_sgmap(bus_dma_tag_t t, bus_dmamap_t map, struct mbuf *m, int flags)
341 {
342 int error;
343
344 error = pci_sgmap_pte64_load_mbuf(t, map, m, flags, t->_sgmap);
345 if (error == 0)
346 CIA_TLB_INVALIDATE();
347
348 return (error);
349 }
350
351 /*
352 * Load a CIA SGMAP-mapped DMA map with a uio.
353 */
354 static int
355 cia_bus_dmamap_load_uio_sgmap(bus_dma_tag_t t, bus_dmamap_t map, struct uio *uio, int flags)
356 {
357 int error;
358
359 error = pci_sgmap_pte64_load_uio(t, map, uio, flags, t->_sgmap);
360 if (error == 0)
361 CIA_TLB_INVALIDATE();
362
363 return (error);
364 }
365
366 /*
367 * Load a CIA SGMAP-mapped DMA map with raw memory.
368 */
369 static int
370 cia_bus_dmamap_load_raw_sgmap(bus_dma_tag_t t, bus_dmamap_t map, bus_dma_segment_t *segs, int nsegs, bus_size_t size, int flags)
371 {
372 int error;
373
374 error = pci_sgmap_pte64_load_raw(t, map, segs, nsegs, size, flags,
375 t->_sgmap);
376 if (error == 0)
377 CIA_TLB_INVALIDATE();
378
379 return (error);
380 }
381
382 /*
383 * Unload a CIA DMA map.
384 */
385 static void
386 cia_bus_dmamap_unload_sgmap(bus_dma_tag_t t, bus_dmamap_t map)
387 {
388
389 /*
390 * Invalidate any SGMAP page table entries used by this
391 * mapping.
392 */
393 pci_sgmap_pte64_unload(t, map, t->_sgmap);
394 CIA_TLB_INVALIDATE();
395
396 /*
397 * Do the generic bits of the unload.
398 */
399 _bus_dmamap_unload_common(t, map);
400 }
401
402 /*
403 * Flush the CIA scatter/gather TLB.
404 */
405 static void
406 cia_tlb_invalidate(void)
407 {
408
409 alpha_mb();
410 REGVAL(CIA_PCI_TBIA) = CIA_PCI_TBIA_ALL;
411 alpha_mb();
412 }
413
414 /*
415 * Flush the scatter/gather TLB on broken Pyxis chips.
416 */
417 static void
418 cia_broken_pyxis_tlb_invalidate(void)
419 {
420 uint32_t ctrl;
421 int i, s;
422
423 s = splhigh();
424
425 /*
426 * Put the Pyxis into PCI loopback mode.
427 */
428 alpha_mb();
429 ctrl = REGVAL(CIA_CSR_CTRL);
430 REGVAL(CIA_CSR_CTRL) = ctrl | CTRL_PCI_LOOP_EN;
431 alpha_mb();
432
433 /*
434 * Now, read from PCI dense memory space at offset 128M (our
435 * target window base), skipping 64k on each read. This forces
436 * S/G TLB misses.
437 *
438 * XXX Looks like the TLB entries are `not quite LRU'. We need
439 * XXX to read more times than there are actual tags!
440 */
441 for (i = 0; i < CIA_TLB_NTAGS + 4; i++) {
442 volatile uint64_t dummy;
443 dummy = *((volatile uint64_t *)
444 ALPHA_PHYS_TO_K0SEG(CIA_PCI_DENSE + CIA_PYXIS_BUG_BASE +
445 (i * 65536)));
446 __USE(dummy);
447 }
448
449 /*
450 * Restore normal PCI operation.
451 */
452 alpha_mb();
453 REGVAL(CIA_CSR_CTRL) = ctrl;
454 alpha_mb();
455
456 splx(s);
457 }
458