rk_drm.c revision 1.17 1 /* $NetBSD: rk_drm.c,v 1.17 2021/12/19 12:45:04 riastradh Exp $ */
2
3 /*-
4 * Copyright (c) 2019 Jared D. McNeill <jmcneill (at) invisible.ca>
5 * All rights reserved.
6 *
7 * Redistribution and use in source and binary forms, with or without
8 * modification, are permitted provided that the following conditions
9 * are met:
10 * 1. Redistributions of source code must retain the above copyright
11 * notice, this list of conditions and the following disclaimer.
12 * 2. Redistributions in binary form must reproduce the above copyright
13 * notice, this list of conditions and the following disclaimer in the
14 * documentation and/or other materials provided with the distribution.
15 *
16 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
17 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
18 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
19 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
20 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
21 * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
22 * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED
23 * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
24 * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
25 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
26 * SUCH DAMAGE.
27 */
28
29 #include <sys/cdefs.h>
30 __KERNEL_RCSID(0, "$NetBSD: rk_drm.c,v 1.17 2021/12/19 12:45:04 riastradh Exp $");
31
32 #include <sys/param.h>
33 #include <sys/bus.h>
34 #include <sys/conf.h>
35 #include <sys/device.h>
36 #include <sys/intr.h>
37 #include <sys/kernel.h>
38 #include <sys/systm.h>
39
40 #include <uvm/uvm_device.h>
41 #include <uvm/uvm_extern.h>
42 #include <uvm/uvm_object.h>
43
44 #include <dev/fdt/fdt_port.h>
45 #include <dev/fdt/fdtvar.h>
46
47 #include <arm/rockchip/rk_drm.h>
48
49 #include <drm/drm_atomic_helper.h>
50 #include <drm/drm_auth.h>
51 #include <drm/drm_crtc_helper.h>
52 #include <drm/drm_damage_helper.h>
53 #include <drm/drm_drv.h>
54 #include <drm/drm_fb_helper.h>
55 #include <drm/drm_fourcc.h>
56 #include <drm/drm_vblank.h>
57
58 #define RK_DRM_MAX_WIDTH 3840
59 #define RK_DRM_MAX_HEIGHT 2160
60
61 static TAILQ_HEAD(, rk_drm_ports) rk_drm_ports =
62 TAILQ_HEAD_INITIALIZER(rk_drm_ports);
63
64 static const struct device_compatible_entry compat_data[] = {
65 { .compat = "rockchip,display-subsystem" },
66 DEVICE_COMPAT_EOL
67 };
68
69 static const char * fb_compatible[] = {
70 "simple-framebuffer",
71 NULL
72 };
73
74 static int rk_drm_match(device_t, cfdata_t, void *);
75 static void rk_drm_attach(device_t, device_t, void *);
76
77 static void rk_drm_init(device_t);
78 static vmem_t *rk_drm_alloc_cma_pool(struct drm_device *, size_t);
79
80 static uint32_t rk_drm_get_vblank_counter(struct drm_device *, unsigned int);
81 static int rk_drm_enable_vblank(struct drm_device *, unsigned int);
82 static void rk_drm_disable_vblank(struct drm_device *, unsigned int);
83
84 static int rk_drm_load(struct drm_device *, unsigned long);
85 static void rk_drm_unload(struct drm_device *);
86
87 static void rk_drm_task_work(struct work *, void *);
88
89 static struct drm_driver rk_drm_driver = {
90 .driver_features = DRIVER_MODESET | DRIVER_ATOMIC | DRIVER_GEM,
91 .dev_priv_size = 0,
92 .load = rk_drm_load,
93 .unload = rk_drm_unload,
94
95 .gem_free_object = drm_gem_cma_free_object,
96 .mmap_object = drm_gem_or_legacy_mmap_object,
97 .gem_uvm_ops = &drm_gem_cma_uvm_ops,
98
99 .dumb_create = drm_gem_cma_dumb_create,
100 .dumb_destroy = drm_gem_dumb_destroy,
101
102 .get_vblank_counter = rk_drm_get_vblank_counter,
103 .enable_vblank = rk_drm_enable_vblank,
104 .disable_vblank = rk_drm_disable_vblank,
105
106 .name = DRIVER_NAME,
107 .desc = DRIVER_DESC,
108 .date = DRIVER_DATE,
109 .major = DRIVER_MAJOR,
110 .minor = DRIVER_MINOR,
111 .patchlevel = DRIVER_PATCHLEVEL,
112 };
113
114 CFATTACH_DECL_NEW(rk_drm, sizeof(struct rk_drm_softc),
115 rk_drm_match, rk_drm_attach, NULL, NULL);
116
117 static int
118 rk_drm_match(device_t parent, cfdata_t cf, void *aux)
119 {
120 struct fdt_attach_args * const faa = aux;
121
122 return of_compatible_match(faa->faa_phandle, compat_data);
123 }
124
125 static void
126 rk_drm_attach(device_t parent, device_t self, void *aux)
127 {
128 struct rk_drm_softc * const sc = device_private(self);
129 struct fdt_attach_args * const faa = aux;
130 struct drm_driver * const driver = &rk_drm_driver;
131 prop_dictionary_t dict = device_properties(self);
132 bool is_disabled;
133
134 aprint_naive("\n");
135
136 if (prop_dictionary_get_bool(dict, "disabled", &is_disabled) &&
137 is_disabled) {
138 aprint_normal(": (disabled)\n");
139 return;
140 }
141
142 aprint_normal("\n");
143
144 sc->sc_dev = self;
145 sc->sc_dmat = faa->faa_dmat;
146 sc->sc_bst = faa->faa_bst;
147 sc->sc_phandle = faa->faa_phandle;
148 sc->sc_task_thread = NULL;
149 SIMPLEQ_INIT(&sc->sc_tasks);
150 if (workqueue_create(&sc->sc_task_wq, "rkdrm",
151 &rk_drm_task_work, NULL, PRI_NONE, IPL_NONE, WQ_MPSAFE)) {
152 aprint_error_dev(self, "unable to create workqueue\n");
153 sc->sc_task_wq = NULL;
154 return;
155 }
156
157 sc->sc_ddev = drm_dev_alloc(driver, sc->sc_dev);
158 if (IS_ERR(sc->sc_ddev)) {
159 aprint_error_dev(self, "couldn't allocate DRM device\n");
160 return;
161 }
162 sc->sc_ddev->dev_private = sc;
163 sc->sc_ddev->bst = sc->sc_bst;
164 sc->sc_ddev->bus_dmat = sc->sc_dmat;
165 sc->sc_ddev->dmat = sc->sc_ddev->bus_dmat;
166 sc->sc_ddev->dmat_subregion_p = false;
167
168 fdt_remove_bycompat(fb_compatible);
169
170 /*
171 * Wait until rk_vop is attached as a sibling to this device --
172 * we need that to actually display our framebuffer.
173 */
174 config_defer(self, rk_drm_init);
175 }
176
177 static void
178 rk_drm_init(device_t dev)
179 {
180 struct rk_drm_softc * const sc = device_private(dev);
181 struct drm_driver * const driver = &rk_drm_driver;
182 int error;
183
184 /*
185 * Cause any tasks issued synchronously during attach to be
186 * processed at the end of this function.
187 */
188 sc->sc_task_thread = curlwp;
189
190 error = -drm_dev_register(sc->sc_ddev, 0);
191 if (error) {
192 aprint_error_dev(dev, "couldn't register DRM device: %d\n",
193 error);
194 goto out;
195 }
196 sc->sc_dev_registered = true;
197
198 aprint_normal_dev(dev, "initialized %s %d.%d.%d %s on minor %d\n",
199 driver->name, driver->major, driver->minor, driver->patchlevel,
200 driver->date, sc->sc_ddev->primary->index);
201
202 /*
203 * Process asynchronous tasks queued synchronously during
204 * attach. This will be for display detection to attach a
205 * framebuffer, so we have the opportunity for a console device
206 * to attach before autoconf has completed, in time for init(8)
207 * to find that console without panicking.
208 */
209 while (!SIMPLEQ_EMPTY(&sc->sc_tasks)) {
210 struct rk_drm_task *const task = SIMPLEQ_FIRST(&sc->sc_tasks);
211
212 SIMPLEQ_REMOVE_HEAD(&sc->sc_tasks, rdt_u.queue);
213 (*task->rdt_fn)(task);
214 }
215
216 out: /* Cause any subesquent tasks to be processed by the workqueue. */
217 atomic_store_relaxed(&sc->sc_task_thread, NULL);
218 }
219
220 static vmem_t *
221 rk_drm_alloc_cma_pool(struct drm_device *ddev, size_t cma_size)
222 {
223 struct rk_drm_softc * const sc = rk_drm_private(ddev);
224 bus_dma_segment_t segs[1];
225 int nsegs;
226 int error;
227
228 error = bus_dmamem_alloc(sc->sc_dmat, cma_size, PAGE_SIZE, 0,
229 segs, 1, &nsegs, BUS_DMA_NOWAIT);
230 if (error) {
231 aprint_error_dev(sc->sc_dev, "couldn't allocate CMA pool\n");
232 return NULL;
233 }
234
235 return vmem_create("rkdrm", segs[0].ds_addr, segs[0].ds_len,
236 PAGE_SIZE, NULL, NULL, NULL, 0, VM_SLEEP, IPL_NONE);
237 }
238
239 static int
240 rk_drm_fb_create_handle(struct drm_framebuffer *fb,
241 struct drm_file *file, unsigned int *handle)
242 {
243 struct rk_drm_framebuffer *sfb = to_rk_drm_framebuffer(fb);
244
245 return drm_gem_handle_create(file, &sfb->obj->base, handle);
246 }
247
248 static void
249 rk_drm_fb_destroy(struct drm_framebuffer *fb)
250 {
251 struct rk_drm_framebuffer *sfb = to_rk_drm_framebuffer(fb);
252
253 drm_framebuffer_cleanup(fb);
254 drm_gem_object_put_unlocked(&sfb->obj->base);
255 kmem_free(sfb, sizeof(*sfb));
256 }
257
258 static const struct drm_framebuffer_funcs rk_drm_framebuffer_funcs = {
259 .create_handle = rk_drm_fb_create_handle,
260 .destroy = rk_drm_fb_destroy,
261 .dirty = drm_atomic_helper_dirtyfb,
262 };
263
264 static struct drm_framebuffer *
265 rk_drm_fb_create(struct drm_device *ddev, struct drm_file *file,
266 const struct drm_mode_fb_cmd2 *cmd)
267 {
268 struct rk_drm_framebuffer *fb;
269 struct drm_gem_object *gem_obj;
270 int error;
271
272 if (cmd->flags)
273 return NULL;
274
275 gem_obj = drm_gem_object_lookup(file, cmd->handles[0]);
276 if (gem_obj == NULL)
277 return NULL;
278
279 fb = kmem_zalloc(sizeof(*fb), KM_SLEEP);
280 drm_helper_mode_fill_fb_struct(ddev, &fb->base, cmd);
281 fb->obj = to_drm_gem_cma_obj(gem_obj);
282
283 error = drm_framebuffer_init(ddev, &fb->base, &rk_drm_framebuffer_funcs);
284 if (error != 0)
285 goto dealloc;
286
287 return &fb->base;
288
289 dealloc:
290 drm_framebuffer_cleanup(&fb->base);
291 kmem_free(fb, sizeof(*fb));
292 drm_gem_object_put_unlocked(gem_obj);
293
294 return NULL;
295 }
296
297 static struct drm_mode_config_funcs rk_drm_mode_config_funcs = {
298 .fb_create = rk_drm_fb_create,
299 .atomic_check = drm_atomic_helper_check,
300 .atomic_commit = drm_atomic_helper_commit,
301 };
302
303 static struct drm_mode_config_helper_funcs rk_drm_mode_config_helper_funcs = {
304 .atomic_commit_tail = drm_atomic_helper_commit_tail_rpm,
305 };
306
307 static int
308 rk_drm_fb_probe(struct drm_fb_helper *helper, struct drm_fb_helper_surface_size *sizes)
309 {
310 struct rk_drm_softc * const sc = rk_drm_private(helper->dev);
311 struct drm_device *ddev = helper->dev;
312 struct rk_drm_framebuffer *sfb = to_rk_drm_framebuffer(helper->fb);
313 struct drm_framebuffer *fb = helper->fb;
314 struct rk_drmfb_attach_args sfa;
315 size_t cma_size;
316 int error;
317
318 const u_int width = sizes->surface_width;
319 const u_int height = sizes->surface_height;
320 const u_int pitch = width * (32 / 8);
321
322 const size_t size = roundup(height * pitch, PAGE_SIZE);
323
324 /* Reserve enough memory for the FB console plus a 4K plane, rounded to 1MB */
325 cma_size = size;
326 cma_size += (RK_DRM_MAX_WIDTH * RK_DRM_MAX_HEIGHT * 4);
327 cma_size = roundup(cma_size, 1024 * 1024);
328 sc->sc_ddev->cma_pool = rk_drm_alloc_cma_pool(sc->sc_ddev, cma_size);
329 if (sc->sc_ddev->cma_pool != NULL)
330 aprint_normal_dev(sc->sc_dev, "reserved %u MB DRAM for CMA\n",
331 (u_int)(cma_size / (1024 * 1024)));
332
333 sfb->obj = drm_gem_cma_create(ddev, size);
334 if (sfb->obj == NULL) {
335 DRM_ERROR("failed to allocate memory for framebuffer\n");
336 return -ENOMEM;
337 }
338
339 /* similar to drm_helper_mode_fill_fb_struct(), but we have no cmd */
340 fb->pitches[0] = pitch;
341 fb->offsets[0] = 0;
342 fb->width = width;
343 fb->height = height;
344 fb->modifier = 0;
345 fb->flags = 0;
346 #ifdef __ARM_BIG_ENDIAN
347 fb->format = drm_format_info(DRM_FORMAT_BGRX8888);
348 #else
349 fb->format = drm_format_info(DRM_FORMAT_XRGB8888);
350 #endif
351 fb->dev = ddev;
352
353 error = drm_framebuffer_init(ddev, fb, &rk_drm_framebuffer_funcs);
354 if (error != 0) {
355 DRM_ERROR("failed to initialize framebuffer\n");
356 return error;
357 }
358
359 memset(&sfa, 0, sizeof(sfa));
360 sfa.sfa_drm_dev = ddev;
361 sfa.sfa_fb_helper = helper;
362 sfa.sfa_fb_sizes = *sizes;
363 sfa.sfa_fb_bst = sc->sc_bst;
364 sfa.sfa_fb_dmat = sc->sc_dmat;
365 sfa.sfa_fb_linebytes = helper->fb->pitches[0];
366
367 helper->fbdev = config_found(ddev->dev, &sfa, NULL,
368 CFARGS(.iattr = "rkfbbus"));
369 if (helper->fbdev == NULL) {
370 DRM_ERROR("unable to attach framebuffer\n");
371 return -ENXIO;
372 }
373
374 return 0;
375 }
376
377 static struct drm_fb_helper_funcs rk_drm_fb_helper_funcs = {
378 .fb_probe = rk_drm_fb_probe,
379 };
380
381 static int
382 rk_drm_load(struct drm_device *ddev, unsigned long flags)
383 {
384 struct rk_drm_softc * const sc = rk_drm_private(ddev);
385 struct rk_drm_ports *sport;
386 struct rk_drm_fbdev *fbdev;
387 struct fdt_endpoint *ep;
388 const u_int *data;
389 int datalen, error, num_crtc, ep_index;
390
391 drm_mode_config_init(ddev);
392 ddev->mode_config.min_width = 0;
393 ddev->mode_config.min_height = 0;
394 ddev->mode_config.max_width = RK_DRM_MAX_WIDTH;
395 ddev->mode_config.max_height = RK_DRM_MAX_HEIGHT;
396 ddev->mode_config.funcs = &rk_drm_mode_config_funcs;
397 ddev->mode_config.helper_private = &rk_drm_mode_config_helper_funcs;
398
399 num_crtc = 0;
400 data = fdtbus_get_prop(sc->sc_phandle, "ports", &datalen);
401 while (datalen >= 4) {
402 const int crtc_phandle = fdtbus_get_phandle_from_native(be32dec(data));
403
404 TAILQ_FOREACH(sport, &rk_drm_ports, entries)
405 if (sport->phandle == crtc_phandle && sport->ddev == NULL) {
406 sport->ddev = ddev;
407 for (ep_index = 0; (ep = fdt_endpoint_get_from_index(sport->port, 0, ep_index)) != NULL; ep_index++) {
408 error = fdt_endpoint_activate_direct(ep, true);
409 if (error != 0)
410 aprint_debug_dev(sc->sc_dev,
411 "failed to activate endpoint %d: %d\n",
412 ep_index, error);
413 }
414 num_crtc++;
415 }
416
417 datalen -= 4;
418 data++;
419 }
420
421 if (num_crtc == 0) {
422 aprint_error_dev(sc->sc_dev, "no display interface ports configured\n");
423 error = ENXIO;
424 goto drmerr;
425 }
426
427 drm_mode_config_reset(ddev);
428
429 fbdev = kmem_zalloc(sizeof(*fbdev), KM_SLEEP);
430
431 drm_fb_helper_prepare(ddev, &fbdev->helper, &rk_drm_fb_helper_funcs);
432
433 error = drm_fb_helper_init(ddev, &fbdev->helper, num_crtc);
434 if (error)
435 goto allocerr;
436
437 fbdev->helper.fb = kmem_zalloc(sizeof(struct rk_drm_framebuffer), KM_SLEEP);
438
439 drm_fb_helper_single_add_all_connectors(&fbdev->helper);
440
441 drm_fb_helper_initial_config(&fbdev->helper, 32);
442
443 /* XXX */
444 ddev->irq_enabled = true;
445 drm_vblank_init(ddev, num_crtc);
446
447 return 0;
448
449 allocerr:
450 kmem_free(fbdev, sizeof(*fbdev));
451 drmerr:
452 drm_mode_config_cleanup(ddev);
453
454 return error;
455 }
456
457 static uint32_t
458 rk_drm_get_vblank_counter(struct drm_device *ddev, unsigned int crtc)
459 {
460 struct rk_drm_softc * const sc = rk_drm_private(ddev);
461
462 if (crtc >= __arraycount(sc->sc_vbl))
463 return 0;
464
465 if (sc->sc_vbl[crtc].get_vblank_counter == NULL)
466 return 0;
467
468 return sc->sc_vbl[crtc].get_vblank_counter(sc->sc_vbl[crtc].priv);
469 }
470
471 static int
472 rk_drm_enable_vblank(struct drm_device *ddev, unsigned int crtc)
473 {
474 struct rk_drm_softc * const sc = rk_drm_private(ddev);
475
476 if (crtc >= __arraycount(sc->sc_vbl))
477 return 0;
478
479 if (sc->sc_vbl[crtc].enable_vblank == NULL)
480 return 0;
481
482 sc->sc_vbl[crtc].enable_vblank(sc->sc_vbl[crtc].priv);
483
484 return 0;
485 }
486
487 static void
488 rk_drm_disable_vblank(struct drm_device *ddev, unsigned int crtc)
489 {
490 struct rk_drm_softc * const sc = rk_drm_private(ddev);
491
492 if (crtc >= __arraycount(sc->sc_vbl))
493 return;
494
495 if (sc->sc_vbl[crtc].disable_vblank == NULL)
496 return;
497
498 sc->sc_vbl[crtc].disable_vblank(sc->sc_vbl[crtc].priv);
499 }
500
501 static void
502 rk_drm_unload(struct drm_device *ddev)
503 {
504 drm_mode_config_cleanup(ddev);
505 }
506
507 int
508 rk_drm_register_port(int phandle, struct fdt_device_ports *port)
509 {
510 struct rk_drm_ports *sport;
511
512 sport = kmem_zalloc(sizeof(*sport), KM_SLEEP);
513 sport->phandle = phandle;
514 sport->port = port;
515 sport->ddev = NULL;
516 TAILQ_INSERT_TAIL(&rk_drm_ports, sport, entries);
517
518 return 0;
519 }
520
521 struct drm_device *
522 rk_drm_port_device(struct fdt_device_ports *port)
523 {
524 struct rk_drm_ports *sport;
525
526 TAILQ_FOREACH(sport, &rk_drm_ports, entries)
527 if (sport->port == port)
528 return sport->ddev;
529
530 return NULL;
531 }
532
533 static void
534 rk_drm_task_work(struct work *work, void *cookie)
535 {
536 struct rk_drm_task *task = container_of(work, struct rk_drm_task,
537 rdt_u.work);
538
539 (*task->rdt_fn)(task);
540 }
541
542 void
543 rk_task_init(struct rk_drm_task *task,
544 void (*fn)(struct rk_drm_task *))
545 {
546
547 task->rdt_fn = fn;
548 }
549
550 void
551 rk_task_schedule(device_t self, struct rk_drm_task *task)
552 {
553 struct rk_drm_softc *sc = device_private(self);
554
555 if (atomic_load_relaxed(&sc->sc_task_thread) == curlwp)
556 SIMPLEQ_INSERT_TAIL(&sc->sc_tasks, task, rdt_u.queue);
557 else
558 workqueue_enqueue(sc->sc_task_wq, &task->rdt_u.work, NULL);
559 }
560