Lines Matching refs:inactive
900 static struct list *inactive(struct kgem *kgem, int num_pages)
904 return &kgem->inactive[cache_bucket(num_pages)];
1522 for (i = 0; i < ARRAY_SIZE(kgem->inactive); i++)
1523 list_init(&kgem->inactive[i]);
1529 for (j = 0; j < ARRAY_SIZE(kgem->vma[i].inactive); j++)
1530 list_init(&kgem->vma[i].inactive[j]);
2141 DBG(("%s: moving handle=%d to inactive\n", __FUNCTION__, bo->handle));
2169 list_move(&bo->list, &kgem->inactive[bucket(bo)]);
2175 list_add(&bo->vma, &kgem->vma[0].inactive[bucket(bo)]);
2179 list_add(&bo->vma, &kgem->vma[1].inactive[bucket(bo)]);
2220 DBG(("%s: removing handle=%d from inactive\n", __FUNCTION__, bo->handle));
2376 DBG(("%s: inactive and cache empty\n", __FUNCTION__));
2992 for (i = 0; i < ARRAY_SIZE(kgem->inactive); i++)
2993 kgem_close_list(kgem, &kgem->inactive[i]);
3799 for (i = 0; i < ARRAY_SIZE(kgem->inactive); i++) {
3800 list_for_each_entry_safe(bo, next, &kgem->inactive[i], list) {
3926 for (i = 0; i < ARRAY_SIZE(kgem->inactive); i++) {
3927 idle &= list_is_empty(&kgem->inactive[i]);
3928 list_for_each_entry(bo, &kgem->inactive[i], list) {
3944 for (i = 0; i < ARRAY_SIZE(kgem->inactive); i++) {
3948 while (!list_is_empty(&kgem->inactive[i])) {
3949 bo = list_last_entry(&kgem->inactive[i],
3969 preserve.prev->next = kgem->inactive[i].next;
3970 kgem->inactive[i].next->prev = preserve.prev;
3971 kgem->inactive[i].next = preserve.next;
3972 preserve.next->prev = &kgem->inactive[i];
3980 for (i = 0; i < ARRAY_SIZE(kgem->inactive); i++)
3981 list_for_each_entry(bo, &kgem->inactive[i], list)
3983 DBG(("%s: still allocated %d bo, %ld bytes, in inactive cache\n",
4030 for (i = 0; i < ARRAY_SIZE(kgem->inactive); i++) {
4031 while (!list_is_empty(&kgem->inactive[i]))
4033 list_last_entry(&kgem->inactive[i],
4123 if (!use_active && list_is_empty(inactive(kgem, num_pages))) {
4124 DBG(("%s: inactive and cache bucket empty\n",
4142 if (list_is_empty(inactive(kgem, num_pages))) {
4151 DBG(("%s: searching for inactive %s map\n",
4153 cache = &kgem->vma[for_cpu].inactive[cache_bucket(num_pages)];
4163 DBG(("inactive too small: %d < %d\n",
4201 cache = use_active ? active(kgem, num_pages, I915_TILING_NONE) : inactive(kgem, num_pages);
4277 use_active ? "active" : "inactive"));
4299 use_active ? "active" : "inactive"));
4830 DBG(("%s(%dx%d, bpp=%d, tiling=%d, exact=%d, inactive=%d, cpu-mapping=%d, gtt-mapping=%d, scanout?=%d, prime?=%d, temp?=%d)\n", __FUNCTION__,
5045 DBG((" 1:from large inactive: pitch=%d, tiling=%d, handle=%d, id=%d\n",
5067 cache = &kgem->vma[for_cpu].inactive[bucket];
5081 DBG(("inactive too small: %d < %d\n",
5094 DBG(("inactive GTT vma with wrong tiling: %d < %d\n",
5116 DBG((" from inactive vma: pitch=%d, tiling=%d: handle=%d, id=%d\n",
5131 list_is_empty(&kgem->inactive[bucket]))
5313 cache = &kgem->inactive[bucket];
5322 DBG(("inactive too small: %d < %d\n",
5349 DBG((" from inactive: pitch=%d, tiling=%d: handle=%d, id=%d\n",
6168 bo == NULL && j < ARRAY_SIZE(kgem->vma[type].inactive);
6170 struct list *head = &kgem->vma[type].inactive[i++%ARRAY_SIZE(kgem->vma[type].inactive)];
6177 DBG(("%s: discarding inactive %s vma cache for %d\n",
7032 DBG(("%s: searching for an inactive GTT map for upload\n",