Lines Matching defs:rdev
138 (i <= 0 || pci_find_capability(p->rdev->ddev->pdev,
140 p->rdev->family == CHIP_RS780 ||
141 p->rdev->family == CHIP_RS880)) {
201 p->vm_bos = radeon_vm_get_bos(p->rdev, p->ib.vm,
211 r = radeon_bo_list_validate(p->rdev, &p->ticket, &p->validated, p->ring);
236 if (p->rdev->family >= CHIP_TAHITI) {
245 if (p->rdev->family >= CHIP_CAYMAN) {
250 } else if (p->rdev->family >= CHIP_RV770) {
276 r = radeon_sync_resv(p->rdev, &p->ib.sync, resv,
362 if (!p->rdev || !(p->rdev->flags & RADEON_IS_AGP))
384 if (p->rdev) {
386 !p->rdev->vm_manager.enabled) {
396 if (p->rdev->asic->ring[p->ring]->cs_parse == NULL) {
401 if (p->rdev->asic->ring[p->ring]->ib_parse == NULL) {
471 radeon_ib_free(parser->rdev, &parser->ib);
472 radeon_ib_free(parser->rdev, &parser->const_ib);
475 static int radeon_cs_ib_chunk(struct radeon_device *rdev,
486 r = radeon_cs_parse(rdev, parser->ring, parser);
500 radeon_uvd_note_usage(rdev);
503 radeon_vce_note_usage(rdev);
505 r = radeon_ib_schedule(rdev, &parser->ib, NULL, true);
515 struct radeon_device *rdev = p->rdev;
519 r = radeon_vm_update_page_directory(rdev, vm);
523 r = radeon_vm_clear_freed(rdev, vm);
532 r = radeon_vm_bo_update(rdev, vm->ib_bo_va,
533 &rdev->ring_tmp_bo.bo->tbo.mem);
543 dev_err(rdev->dev, "bo %p not in vm %p\n", bo, vm);
547 r = radeon_vm_bo_update(rdev, bo_va, &bo->tbo.mem);
554 return radeon_vm_clear_invalids(rdev, vm);
557 static int radeon_cs_ib_vm_chunk(struct radeon_device *rdev,
570 r = radeon_ring_ib_parse(rdev, parser->ring, &parser->const_ib);
576 r = radeon_ring_ib_parse(rdev, parser->ring, &parser->ib);
582 radeon_uvd_note_usage(rdev);
597 if ((rdev->family >= CHIP_TAHITI) &&
599 r = radeon_ib_schedule(rdev
601 r = radeon_ib_schedule(rdev, &parser->ib, NULL, true);
609 static int radeon_cs_handle_lockup(struct radeon_device *rdev, int r)
612 r = radeon_gpu_reset(rdev);
619 static int radeon_cs_ib_fill(struct radeon_device *rdev, struct radeon_cs_parser *parser)
632 if ((rdev->family >= CHIP_TAHITI) &&
639 r = radeon_ib_get(rdev, parser->ring, &parser->const_ib,
661 r = radeon_ib_get(rdev, parser->ring, &parser->ib,
677 struct radeon_device *rdev = dev->dev_private;
681 down_read(&rdev->exclusive_lock);
682 if (!rdev->accel_working) {
683 up_read(&rdev->exclusive_lock);
686 if (rdev->in_reset) {
687 up_read(&rdev->exclusive_lock);
688 r = radeon_gpu_reset(rdev);
696 parser.rdev = rdev;
697 parser.dev = rdev->dev;
698 parser.family = rdev->family;
703 up_read(&rdev->exclusive_lock);
704 r = radeon_cs_handle_lockup(rdev, r);
708 r = radeon_cs_ib_fill(rdev, &parser);
717 up_read(&rdev->exclusive_lock);
718 r = radeon_cs_handle_lockup(rdev, r);
724 r = radeon_cs_ib_chunk(rdev, &parser);
728 r = radeon_cs_ib_vm_chunk(rdev, &parser);
734 up_read(&rdev->exclusive_lock);
735 r = radeon_cs_handle_lockup(rdev, r);
752 struct radeon_device *rdev = p->rdev;
768 if (rdev->family < CHIP_R600) {