omap_drm.c revision 7cdc0497
1/* -*- mode: C; c-file-style: "k&r"; tab-width 4; indent-tabs-mode: t; -*- */
2
3/*
4 * Copyright (C) 2011 Texas Instruments, Inc
5 *
6 * Permission is hereby granted, free of charge, to any person obtaining a
7 * copy of this software and associated documentation files (the "Software"),
8 * to deal in the Software without restriction, including without limitation
9 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
10 * and/or sell copies of the Software, and to permit persons to whom the
11 * Software is furnished to do so, subject to the following conditions:
12 *
13 * The above copyright notice and this permission notice (including the next
14 * paragraph) shall be included in all copies or substantial portions of the
15 * Software.
16 *
17 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
18 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
19 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
20 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
21 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
22 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
23 * SOFTWARE.
24 *
25 * Authors:
26 *    Rob Clark <rob@ti.com>
27 */
28
29#include <stdlib.h>
30#include <linux/stddef.h>
31#include <linux/types.h>
32#include <errno.h>
33#include <sys/mman.h>
34#include <fcntl.h>
35#include <unistd.h>
36#include <pthread.h>
37
38#include <libdrm_macros.h>
39#include <xf86drm.h>
40#include <xf86atomic.h>
41
42#include "omap_drm.h"
43#include "omap_drmif.h"
44
45#define __round_mask(x, y) ((__typeof__(x))((y)-1))
46#define round_up(x, y) ((((x)-1) | __round_mask(x, y))+1)
47#define PAGE_SIZE 4096
48
49static pthread_mutex_t table_lock = PTHREAD_MUTEX_INITIALIZER;
50static void * dev_table;
51
52struct omap_device {
53	int fd;
54	atomic_t refcnt;
55
56	/* The handle_table is used to track GEM bo handles associated w/
57	 * this fd.  This is needed, in particular, when importing
58	 * dmabuf's because we don't want multiple 'struct omap_bo's
59	 * floating around with the same handle.  Otherwise, when the
60	 * first one is omap_bo_del()'d the handle becomes no longer
61	 * valid, and the remaining 'struct omap_bo's are left pointing
62	 * to an invalid handle (and possible a GEM bo that is already
63	 * free'd).
64	 */
65	void *handle_table;
66};
67
68/* a GEM buffer object allocated from the DRM device */
69struct omap_bo {
70	struct omap_device	*dev;
71	void		*map;		/* userspace mmap'ing (if there is one) */
72	uint32_t	size;
73	uint32_t	handle;
74	uint32_t	name;		/* flink global handle (DRI2 name) */
75	uint64_t	offset;		/* offset to mmap() */
76	int		fd;		/* dmabuf handle */
77	atomic_t	refcnt;
78};
79
80static struct omap_device * omap_device_new_impl(int fd)
81{
82	struct omap_device *dev = calloc(sizeof(*dev), 1);
83	if (!dev)
84		return NULL;
85	dev->fd = fd;
86	atomic_set(&dev->refcnt, 1);
87	dev->handle_table = drmHashCreate();
88	return dev;
89}
90
91drm_public struct omap_device * omap_device_new(int fd)
92{
93	struct omap_device *dev = NULL;
94
95	pthread_mutex_lock(&table_lock);
96
97	if (!dev_table)
98		dev_table = drmHashCreate();
99
100	if (drmHashLookup(dev_table, fd, (void **)&dev)) {
101		/* not found, create new device */
102		dev = omap_device_new_impl(fd);
103		drmHashInsert(dev_table, fd, dev);
104	} else {
105		/* found, just incr refcnt */
106		dev = omap_device_ref(dev);
107	}
108
109	pthread_mutex_unlock(&table_lock);
110
111	return dev;
112}
113
114drm_public struct omap_device * omap_device_ref(struct omap_device *dev)
115{
116	atomic_inc(&dev->refcnt);
117	return dev;
118}
119
120drm_public void omap_device_del(struct omap_device *dev)
121{
122	if (!atomic_dec_and_test(&dev->refcnt))
123		return;
124	pthread_mutex_lock(&table_lock);
125	drmHashDestroy(dev->handle_table);
126	drmHashDelete(dev_table, dev->fd);
127	pthread_mutex_unlock(&table_lock);
128	free(dev);
129}
130
131drm_public int
132omap_get_param(struct omap_device *dev, uint64_t param, uint64_t *value)
133{
134	struct drm_omap_param req = {
135			.param = param,
136	};
137	int ret;
138
139	ret = drmCommandWriteRead(dev->fd, DRM_OMAP_GET_PARAM, &req, sizeof(req));
140	if (ret) {
141		return ret;
142	}
143
144	*value = req.value;
145
146	return 0;
147}
148
149drm_public int
150omap_set_param(struct omap_device *dev, uint64_t param, uint64_t value)
151{
152	struct drm_omap_param req = {
153			.param = param,
154			.value = value,
155	};
156	return drmCommandWrite(dev->fd, DRM_OMAP_SET_PARAM, &req, sizeof(req));
157}
158
159/* lookup a buffer from it's handle, call w/ table_lock held: */
160static struct omap_bo * lookup_bo(struct omap_device *dev,
161		uint32_t handle)
162{
163	struct omap_bo *bo = NULL;
164	if (!drmHashLookup(dev->handle_table, handle, (void **)&bo)) {
165		/* found, incr refcnt and return: */
166		bo = omap_bo_ref(bo);
167	}
168	return bo;
169}
170
171/* allocate a new buffer object, call w/ table_lock held */
172static struct omap_bo * bo_from_handle(struct omap_device *dev,
173		uint32_t handle)
174{
175	struct omap_bo *bo = calloc(sizeof(*bo), 1);
176	if (!bo) {
177		struct drm_gem_close req = {
178				.handle = handle,
179		};
180		drmIoctl(dev->fd, DRM_IOCTL_GEM_CLOSE, &req);
181		return NULL;
182	}
183	bo->dev = omap_device_ref(dev);
184	bo->handle = handle;
185	bo->fd = -1;
186	atomic_set(&bo->refcnt, 1);
187	/* add ourselves to the handle table: */
188	drmHashInsert(dev->handle_table, handle, bo);
189	return bo;
190}
191
192/* allocate a new buffer object */
193static struct omap_bo * omap_bo_new_impl(struct omap_device *dev,
194		union omap_gem_size size, uint32_t flags)
195{
196	struct omap_bo *bo = NULL;
197	struct drm_omap_gem_new req = {
198			.size = size,
199			.flags = flags,
200	};
201
202	if (size.bytes == 0) {
203		goto fail;
204	}
205
206	if (drmCommandWriteRead(dev->fd, DRM_OMAP_GEM_NEW, &req, sizeof(req))) {
207		goto fail;
208	}
209
210	pthread_mutex_lock(&table_lock);
211	bo = bo_from_handle(dev, req.handle);
212	pthread_mutex_unlock(&table_lock);
213
214	if (flags & OMAP_BO_TILED) {
215		bo->size = round_up(size.tiled.width, PAGE_SIZE) * size.tiled.height;
216	} else {
217		bo->size = size.bytes;
218	}
219
220	return bo;
221
222fail:
223	free(bo);
224	return NULL;
225}
226
227
228/* allocate a new (un-tiled) buffer object */
229drm_public struct omap_bo *
230omap_bo_new(struct omap_device *dev, uint32_t size, uint32_t flags)
231{
232	union omap_gem_size gsize = {
233			.bytes = size,
234	};
235	if (flags & OMAP_BO_TILED) {
236		return NULL;
237	}
238	return omap_bo_new_impl(dev, gsize, flags);
239}
240
241/* allocate a new buffer object */
242drm_public struct omap_bo *
243omap_bo_new_tiled(struct omap_device *dev, uint32_t width,
244		  uint32_t height, uint32_t flags)
245{
246	union omap_gem_size gsize = {
247			.tiled = {
248				.width = width,
249				.height = height,
250			},
251	};
252	if (!(flags & OMAP_BO_TILED)) {
253		return NULL;
254	}
255	return omap_bo_new_impl(dev, gsize, flags);
256}
257
258drm_public struct omap_bo *omap_bo_ref(struct omap_bo *bo)
259{
260	atomic_inc(&bo->refcnt);
261	return bo;
262}
263
264/* get buffer info */
265static int get_buffer_info(struct omap_bo *bo)
266{
267	struct drm_omap_gem_info req = {
268			.handle = bo->handle,
269	};
270	int ret = drmCommandWriteRead(bo->dev->fd, DRM_OMAP_GEM_INFO,
271			&req, sizeof(req));
272	if (ret) {
273		return ret;
274	}
275
276	/* really all we need for now is mmap offset */
277	bo->offset = req.offset;
278	bo->size = req.size;
279
280	return 0;
281}
282
283/* import a buffer object from DRI2 name */
284drm_public struct omap_bo *
285omap_bo_from_name(struct omap_device *dev, uint32_t name)
286{
287	struct omap_bo *bo = NULL;
288	struct drm_gem_open req = {
289			.name = name,
290	};
291
292	pthread_mutex_lock(&table_lock);
293
294	if (drmIoctl(dev->fd, DRM_IOCTL_GEM_OPEN, &req)) {
295		goto fail;
296	}
297
298	bo = lookup_bo(dev, req.handle);
299	if (!bo) {
300		bo = bo_from_handle(dev, req.handle);
301		bo->name = name;
302	}
303
304	pthread_mutex_unlock(&table_lock);
305
306	return bo;
307
308fail:
309	pthread_mutex_unlock(&table_lock);
310	free(bo);
311	return NULL;
312}
313
314/* import a buffer from dmabuf fd, does not take ownership of the
315 * fd so caller should close() the fd when it is otherwise done
316 * with it (even if it is still using the 'struct omap_bo *')
317 */
318drm_public struct omap_bo *
319omap_bo_from_dmabuf(struct omap_device *dev, int fd)
320{
321	struct omap_bo *bo = NULL;
322	struct drm_prime_handle req = {
323			.fd = fd,
324	};
325	int ret;
326
327	pthread_mutex_lock(&table_lock);
328
329	ret = drmIoctl(dev->fd, DRM_IOCTL_PRIME_FD_TO_HANDLE, &req);
330	if (ret) {
331		goto fail;
332	}
333
334	bo = lookup_bo(dev, req.handle);
335	if (!bo) {
336		bo = bo_from_handle(dev, req.handle);
337	}
338
339	pthread_mutex_unlock(&table_lock);
340
341	return bo;
342
343fail:
344	pthread_mutex_unlock(&table_lock);
345	free(bo);
346	return NULL;
347}
348
349/* destroy a buffer object */
350drm_public void omap_bo_del(struct omap_bo *bo)
351{
352	if (!bo) {
353		return;
354	}
355
356	if (!atomic_dec_and_test(&bo->refcnt))
357		return;
358
359	if (bo->map) {
360		munmap(bo->map, bo->size);
361	}
362
363	if (bo->fd >= 0) {
364		close(bo->fd);
365	}
366
367	if (bo->handle) {
368		struct drm_gem_close req = {
369				.handle = bo->handle,
370		};
371		pthread_mutex_lock(&table_lock);
372		drmHashDelete(bo->dev->handle_table, bo->handle);
373		drmIoctl(bo->dev->fd, DRM_IOCTL_GEM_CLOSE, &req);
374		pthread_mutex_unlock(&table_lock);
375	}
376
377	omap_device_del(bo->dev);
378
379	free(bo);
380}
381
382/* get the global flink/DRI2 buffer name */
383drm_public int omap_bo_get_name(struct omap_bo *bo, uint32_t *name)
384{
385	if (!bo->name) {
386		struct drm_gem_flink req = {
387				.handle = bo->handle,
388		};
389		int ret;
390
391		ret = drmIoctl(bo->dev->fd, DRM_IOCTL_GEM_FLINK, &req);
392		if (ret) {
393			return ret;
394		}
395
396		bo->name = req.name;
397	}
398
399	*name = bo->name;
400
401	return 0;
402}
403
404drm_public uint32_t omap_bo_handle(struct omap_bo *bo)
405{
406	return bo->handle;
407}
408
409/* caller owns the dmabuf fd that is returned and is responsible
410 * to close() it when done
411 */
412drm_public int omap_bo_dmabuf(struct omap_bo *bo)
413{
414	if (bo->fd < 0) {
415		struct drm_prime_handle req = {
416				.handle = bo->handle,
417				.flags = DRM_CLOEXEC,
418		};
419		int ret;
420
421		ret = drmIoctl(bo->dev->fd, DRM_IOCTL_PRIME_HANDLE_TO_FD, &req);
422		if (ret) {
423			return ret;
424		}
425
426		bo->fd = req.fd;
427	}
428	return dup(bo->fd);
429}
430
431drm_public uint32_t omap_bo_size(struct omap_bo *bo)
432{
433	if (!bo->size) {
434		get_buffer_info(bo);
435	}
436	return bo->size;
437}
438
439drm_public void *omap_bo_map(struct omap_bo *bo)
440{
441	if (!bo->map) {
442		if (!bo->offset) {
443			get_buffer_info(bo);
444		}
445
446		bo->map = mmap(0, bo->size, PROT_READ | PROT_WRITE,
447				MAP_SHARED, bo->dev->fd, bo->offset);
448		if (bo->map == MAP_FAILED) {
449			bo->map = NULL;
450		}
451	}
452	return bo->map;
453}
454
455drm_public int omap_bo_cpu_prep(struct omap_bo *bo, enum omap_gem_op op)
456{
457	struct drm_omap_gem_cpu_prep req = {
458			.handle = bo->handle,
459			.op = op,
460	};
461	return drmCommandWrite(bo->dev->fd,
462			DRM_OMAP_GEM_CPU_PREP, &req, sizeof(req));
463}
464
465drm_public int omap_bo_cpu_fini(struct omap_bo *bo, enum omap_gem_op op)
466{
467	struct drm_omap_gem_cpu_fini req = {
468			.handle = bo->handle,
469			.op = op,
470			.nregions = 0,
471	};
472	return drmCommandWrite(bo->dev->fd,
473			DRM_OMAP_GEM_CPU_FINI, &req, sizeof(req));
474}
475