Lines Matching defs:rdev
55 int radeon_gem_object_create(struct radeon_device *rdev, unsigned long size,
73 max_size = rdev->mc.gtt_size - rdev->gart_pin_size;
81 r = radeon_bo_create(rdev, size, alignment, kernel, initial_domain,
99 mutex_lock(&rdev->gem.mutex);
100 list_add_tail(&robj->list, &rdev->gem.objects);
101 mutex_unlock(&rdev->gem.mutex);
143 int radeon_gem_init(struct radeon_device *rdev)
145 INIT_LIST_HEAD(&rdev->gem.objects);
149 void radeon_gem_fini(struct radeon_device *rdev)
151 radeon_bo_force_delete(rdev);
161 struct radeon_device *rdev = rbo->rdev;
167 if ((rdev->family < CHIP_CAYMAN) ||
168 (!rdev->accel_working)) {
179 bo_va = radeon_vm_bo_add(rdev, vm, rbo);
192 struct radeon_device *rdev = rbo->rdev;
198 if ((rdev->family < CHIP_CAYMAN) ||
199 (!rdev->accel_working)) {
205 dev_err(rdev->dev, "leaking bo va because "
212 radeon_vm_bo_rmv(rdev, bo_va);
218 static int radeon_gem_handle_lockup(struct radeon_device *rdev, int r)
221 r = radeon_gpu_reset(rdev);
234 struct radeon_device *rdev = dev->dev_private;
238 man = &rdev->mman.bdev.man[TTM_PL_VRAM];
241 args->vram_visible = rdev->mc.visible_vram_size;
242 args->vram_visible -= rdev->vram_pin_size;
243 args->gart_size = rdev->mc.gtt_size;
244 args->gart_size -= rdev->gart_pin_size;
268 struct radeon_device *rdev = dev->dev_private;
274 down_read(&rdev->exclusive_lock);
277 r = radeon_gem_object_create(rdev, args->size, args->alignment,
281 up_read(&rdev->exclusive_lock);
282 r = radeon_gem_handle_lockup(rdev, r);
289 up_read(&rdev->exclusive_lock);
290 r = radeon_gem_handle_lockup(rdev, r);
294 up_read(&rdev->exclusive_lock);
302 struct radeon_device *rdev = dev->dev_private;
322 if (rdev->family < CHIP_R600)
333 down_read(&rdev->exclusive_lock);
336 r = radeon_gem_object_create(rdev, args->size, 0,
388 up_read(&rdev->exclusive_lock);
395 up_read(&rdev->exclusive_lock);
396 r = radeon_gem_handle_lockup(rdev, r);
406 struct radeon_device *rdev = dev->dev_private;
414 down_read(&rdev->exclusive_lock);
419 up_read(&rdev->exclusive_lock);
427 up_read(&rdev->exclusive_lock);
428 r = radeon_gem_handle_lockup(robj->rdev, r);
491 struct radeon_device *rdev = dev->dev_private;
513 if (rdev->asic->mmio_hdp_flush &&
515 robj->rdev->asic->mmio_hdp_flush(rdev);
517 r = radeon_gem_handle_lockup(rdev, r);
565 * @rdev: radeon_device pointer
571 static void radeon_gem_va_update_vm(struct radeon_device *rdev,
587 vm_bos = radeon_vm_get_bos(rdev, bo_va->vm, &list);
604 r = radeon_vm_clear_freed(rdev, bo_va->vm);
609 r = radeon_vm_bo_update(rdev, bo_va, &bo_va->bo->tbo.mem);
629 struct radeon_device *rdev = dev->dev_private;
636 if (!rdev->vm_manager.enabled) {
711 r = radeon_vm_bo_set_addr(rdev, bo_va, args->offset, args->flags);
714 r = radeon_vm_bo_set_addr(rdev, bo_va, 0, 0);
720 radeon_gem_va_update_vm(rdev, bo_va);
775 struct radeon_device *rdev = dev->dev_private;
780 args->pitch = radeon_align_pitch(rdev, args->width,
785 r = radeon_gem_object_create(rdev, args->size, 0,
806 struct radeon_device *rdev = dev->dev_private;
810 mutex_lock(&rdev->gem.mutex);
811 list_for_each_entry(rbo, &rdev->gem.objects, list) {
833 mutex_unlock(&rdev->gem.mutex);
842 int radeon_gem_debugfs_init(struct radeon_device *rdev)
845 return radeon_debugfs_add_files(rdev, radeon_debugfs_gem_list, 1);