/src/games/backgammon/common_source/ |
allow.c | 44 movallow(struct move *mm) 48 if (mm->d0) 49 mswap(mm); 50 m = (mm->D0 == mm->D1 ? 4 : 2); 52 mm->p[i] = bar; 57 mm->h[i] = 0; 60 mm->g[i] = bar + cturn * mm->D1; 62 mm->g[i] = bar + cturn * mm->D0 [all...] |
one.c | 47 makmove(struct move *mm, int i) 52 d = mm->d0; 53 n = abs(mm->g[i] - mm->p[i]); 55 if (board[mm->p[i]] * cturn <= 0) 56 return (checkd(mm, d) + 2); 57 if (mm->g[i] != home && board[mm->g[i]] * cturn < -1) 58 return (checkd(mm, d) + 3); 59 if (i || mm->D0 == mm->D1) [all...] |
init.c | 97 move_init(struct move *mm) 99 mm->D0 = 0; 100 mm->D1 = 0; 101 mm->mvlim = 0; 102 mm->p[0] = mm->p[1] = mm->p[2] = mm->p[3] = mm->p[4] = 0; 103 mm->g[0] = mm->g[1] = mm->g[2] = mm->g[3] = mm->g[4] = 0 [all...] |
check.c | 44 getmove(struct move *mm) 50 i = checkmove(mm, c); 54 if (movokay(mm, mm->mvlim)) { 59 for (i = 0; i < mm->mvlim; i++) 60 if (mm->h[i]) 61 wrhit(mm->g[i]); 85 writec(mm->mvlim + '0'); 87 if (mm->mvlim > 1) 94 if (quit(mm)) [all...] |
table.c | 86 checkmove(struct move *mm, int ist) 101 mm->p[j] = mm->g[j] = -1; 108 save(mm, 1); 114 proll(mm); 129 n = rsetbrd(mm); 154 proll(mm); 159 n = dotable(mm, c, ist); 171 if (n == -1 && mvl >= mm->mvlim) 173 if (n == -1 && mvl < mm->mvlim - 1 [all...] |
/src/games/backgammon/teachgammon/ |
tutor.c | 51 tutor(struct move *mm) 75 movback(mm, mm->mvlim); 81 proll(mm); 85 getmove(mm); 97 mm->D0 = test[i].roll1; 98 mm->D1 = test[i].roll2; 99 mm->d0 = 0; 100 mm->mvlim = 0; 104 mm->p[j] = test[i].mp[j] [all...] |
/src/sys/external/bsd/drm/dist/bsd-core/ |
drm_mm.c | 52 unsigned long drm_mm_tail_space(struct drm_mm *mm) 57 tail_node = mm->ml_entry.prev; 65 int drm_mm_remove_space_from_tail(struct drm_mm *mm, unsigned long size) 70 tail_node = mm->ml_entry.prev; 82 static struct drm_mm_node *drm_mm_kmalloc(struct drm_mm *mm, int atomic) 92 mutex_enter(&mm->unused_lock); 93 if (list_empty(&mm->unused_nodes)) 97 list_entry(mm->unused_nodes.next, 100 --mm->num_unused; 102 mutex_exit(&mm->unused_lock) 231 struct drm_mm *mm = cur->mm; local in function:drm_mm_put_block [all...] |
drm_mm.h | 45 struct drm_mm *mm; member in struct:drm_mm_node 77 extern struct drm_mm_node *drm_mm_search_free(const struct drm_mm *mm, 81 extern int drm_mm_init(struct drm_mm *mm, unsigned long start, 83 extern void drm_mm_takedown(struct drm_mm *mm); 84 extern int drm_mm_clean(struct drm_mm *mm); 85 extern unsigned long drm_mm_tail_space(struct drm_mm *mm); 86 extern int drm_mm_remove_space_from_tail(struct drm_mm *mm, 88 extern int drm_mm_add_space_to_tail(struct drm_mm *mm, 90 extern int drm_mm_pre_get(struct drm_mm *mm); 94 return block->mm; [all...] |
/src/sys/external/bsd/drm2/dist/drm/i915/gem/ |
i915_gem_userptr.c | 16 #include <linux/sched/mm.h> 29 struct vmspace *mm; member in struct:i915_mm_struct 31 struct mm_struct *mm; 48 struct i915_mm_struct *mm; member in struct:i915_mmu_notifier 171 i915_mmu_notifier_create(struct i915_mm_struct *mm) 182 mn->mm = mm; 203 i915_mmu_notifier_find(struct i915_mm_struct *mm) 208 mn = mm->mn; 212 mn = i915_mmu_notifier_create(mm); 336 struct i915_mm_struct *mm; local in function:__i915_mm_struct_find 350 struct i915_mm_struct *mm; local in function:i915_gem_userptr_init__mm_struct 403 struct i915_mm_struct *mm = container_of(work, typeof(*mm), work); local in function:__i915_mm_struct_free__worker 412 struct i915_mm_struct *mm = container_of(kref, typeof(*mm), kref); local in function:__i915_mm_struct_free 499 struct vmspace *mm = obj->userptr.mm->mm; local in function:__i915_gem_userptr_get_pages_worker 605 struct vmspace *mm = obj->userptr.mm->mm; local in function:i915_gem_userptr_get_pages 607 struct mm_struct *mm = obj->userptr.mm->mm; local in function:i915_gem_userptr_get_pages [all...] |
i915_gem_shrinker.c | 13 #include <linux/sched/mm.h> 44 if (atomic_read(&obj->mm.pages_pin_count) > atomic_read(&obj->bind_count)) 52 return swap_available() || obj->mm.madv == I915_MADV_DONTNEED; 73 switch (obj->mm.madv) { 98 * (like cpu mmaps) or the mm core has reused them before we could grab them. 119 { &i915->mm.purge_list, ~0u }, 121 &i915->mm.shrink_list, 188 spin_lock_irqsave(&i915->mm.obj_lock, flags); 192 mm.link))) { 193 list_move_tail(&obj->mm.link, &still_in_list) [all...] |
i915_gem_pm.c | 46 mm.link); 53 &i915->mm.shrink_list, 54 &i915->mm.purge_list, 81 spin_lock_irqsave(&i915->mm.obj_lock, flags); 86 list_move_tail(&obj->mm.link, &keep); 92 spin_unlock_irqrestore(&i915->mm.obj_lock, flags); 99 spin_lock_irqsave(&i915->mm.obj_lock, flags); 104 spin_unlock_irqrestore(&i915->mm.obj_lock, flags);
|
i915_gem_internal.c | 52 obj->mm.u.internal.rsegs = obj->mm.u.internal.nsegs = 0; 54 KASSERT(obj->mm.u.internal.segs == NULL); 57 nsegs > SIZE_MAX/sizeof(obj->mm.u.internal.segs[0])) { 61 obj->mm.u.internal.segs = kmem_alloc( 62 nsegs * sizeof(obj->mm.u.internal.segs[0]), 64 if (obj->mm.u.internal.segs == NULL) { 68 obj->mm.u.internal.nsegs = nsegs; 72 obj->mm.u.internal.segs, nsegs, &obj->mm.u.internal.rsegs [all...] |
/src/sys/external/bsd/drm2/dist/drm/i915/ |
i915_buddy.c | 85 static void mark_free(struct i915_buddy_mm *mm, 92 &mm->free_list[i915_buddy_block_order(block)]); 103 int i915_buddy_init(struct i915_buddy_mm *mm, u64 size, u64 chunk_size) 119 mm->size = size; 120 mm->chunk_size = chunk_size; 121 mm->max_order = ilog2(size) - ilog2(chunk_size); 123 GEM_BUG_ON(mm->max_order > I915_BUDDY_MAX_ORDER); 125 mm->free_list = kmalloc_array(mm->max_order + 1, 128 if (!mm->free_list [all...] |
i915_buddy.h | 33 * i915_buddy_free* ownership is given back to the mm. 62 * the mm. Everything above is considered do-not-touch. 109 i915_buddy_block_size(struct i915_buddy_mm *mm, 112 return mm->chunk_size << i915_buddy_block_order(block); 115 int i915_buddy_init(struct i915_buddy_mm *mm, u64 size, u64 chunk_size); 117 void i915_buddy_fini(struct i915_buddy_mm *mm); 120 i915_buddy_alloc(struct i915_buddy_mm *mm, unsigned int order); 122 int i915_buddy_alloc_range(struct i915_buddy_mm *mm, 126 void i915_buddy_free(struct i915_buddy_mm *mm, struct i915_buddy_block *block); 128 void i915_buddy_free_list(struct i915_buddy_mm *mm, struct list_head *objects) [all...] |
/src/games/hack/ |
hack.mkmaze.c | 85 coord mm; local in function:makemaz 113 mm = mazexy(); 114 zx = mm.x; 115 zy = mm.y; 134 mm = mazexy(); 135 (void) mkobj_at(rn2(2) ? GEM_SYM : 0, mm.x, mm.y); 138 mm = mazexy(); 139 (void) mkobj_at(ROCK_SYM, mm.x, mm.y) 213 coord mm; local in function:mazexy [all...] |
/src/sys/external/bsd/drm2/dist/drm/amd/amdkfd/ |
kfd_mqd_manager.h | 74 void (*init_mqd)(struct mqd_manager *mm, void **mqd, 78 int (*load_mqd)(struct mqd_manager *mm, void *mqd, 83 void (*update_mqd)(struct mqd_manager *mm, void *mqd, 86 int (*destroy_mqd)(struct mqd_manager *mm, void *mqd, 91 void (*free_mqd)(struct mqd_manager *mm, void *mqd, 94 bool (*is_occupied)(struct mqd_manager *mm, void *mqd, 98 int (*get_wave_state)(struct mqd_manager *mm, void *mqd, 117 void free_mqd_hiq_sdma(struct mqd_manager *mm, void *mqd, 120 void mqd_symmetrically_map_cu_mask(struct mqd_manager *mm,
|
kfd_mqd_manager_vi.c | 52 static void update_cu_mask(struct mqd_manager *mm, void *mqd, 61 mqd_symmetrically_map_cu_mask(mm, 95 static void init_mqd(struct mqd_manager *mm, void **mqd, 142 if (mm->dev->cwsr_enabled && q->ctx_save_restore_area_address) { 158 mm->update_mqd(mm, m, q); 161 static int load_mqd(struct mqd_manager *mm, void *mqd, 169 return mm->dev->kfd2kgd->hqd_load(mm->dev->kgd, mqd, pipe_id, queue_id, 174 static void __update_mqd(struct mqd_manager *mm, void *mqd [all...] |
/src/games/backgammon/backgammon/ |
move.c | 103 move(struct move *mm, int okay) 118 roll(mm); 137 writec(mm->D0 + '0'); 139 writec(mm->D1 + '0'); 146 mm->mvlim = movallow(mm); 147 if (mm->mvlim == 0) { 158 trymove(mm, 0, 0); 159 pickmove(mm); 163 for (i = 0; i < mm->mvlim; i++) [all...] |
main.c | 96 struct move mmstore, *mm; local in function:main 121 mm = &mmstore; 122 move_init(mm); 124 getarg(mm, &argv); 241 roll(mm); 247 while (mm->D0 == mm->D1) /* no doubles */ 248 roll(mm); 252 writec(mm->D0 + '0'); 254 writec(mm->D1 + '0') [all...] |
/src/sys/external/bsd/drm2/dist/drm/i915/selftests/ |
i915_buddy.c | 18 static void __igt_dump_block(struct i915_buddy_mm *mm, 27 i915_buddy_block_size(mm, block), 32 static void igt_dump_block(struct i915_buddy_mm *mm, 37 __igt_dump_block(mm, block, false); 41 __igt_dump_block(mm, buddy, true); 44 static int igt_check_block(struct i915_buddy_mm *mm, 62 block_size = i915_buddy_block_size(mm, block); 65 if (block_size < mm->chunk_size) { 75 if (!IS_ALIGNED(block_size, mm->chunk_size)) { 80 if (!IS_ALIGNED(offset, mm->chunk_size)) 305 struct i915_buddy_mm mm; local in function:igt_buddy_alloc_smoke 399 struct i915_buddy_mm mm; local in function:igt_buddy_alloc_pessimistic 491 struct i915_buddy_mm mm; local in function:igt_buddy_alloc_optimistic 541 struct i915_buddy_mm mm; local in function:igt_buddy_alloc_pathological 622 struct i915_buddy_mm mm; local in function:igt_buddy_alloc_range [all...] |
/src/sys/external/bsd/drm2/dist/drm/nouveau/nvkm/core/ |
nouveau_nvkm_core_mm.c | 29 #include <core/mm.h> 31 #define node(root, dir) ((root)->nl_entry.dir == &mm->nodes) ? NULL : \ 35 nvkm_mm_dump(struct nvkm_mm *mm, const char *header) 41 list_for_each_entry(node, &mm->nodes, nl_entry) { 46 list_for_each_entry(node, &mm->free, fl_entry) { 53 nvkm_mm_free(struct nvkm_mm *mm, struct nvkm_mm_node **pthis) 77 list_for_each_entry(prev, &mm->free, fl_entry) { 91 region_head(struct nvkm_mm *mm, struct nvkm_mm_node *a, u32 size) 116 nvkm_mm_head(struct nvkm_mm *mm, u8 heap, u8 type, u32 size_max, u32 size_min, 126 list_for_each_entry(this, &mm->free, fl_entry) [all...] |
/src/sys/external/bsd/drm2/dist/drm/amd/amdgpu/ |
amdgpu_amdkfd_fence.c | 34 #include <linux/sched/mm.h> 68 struct mm_struct *mm) 77 mmgrab(mm); 78 fence->mm = mm; 130 if (!kgd2kfd_schedule_evict_and_restore_process(fence->mm, f)) 154 mmdrop(fence->mm); 159 * amdkfd_fence_check_mm - Check if @mm is same as that of the fence @f 163 * @mm: [IN] mm that needs to be verifie [all...] |
/src/sys/external/bsd/drm2/dist/include/drm/ |
drm_mm.h | 165 struct drm_mm *mm; member in struct:drm_mm_node 232 struct drm_mm *mm; member in struct:drm_mm_scan 268 * @mm: drm_mm to check 277 * True if the @mm is initialized. 279 static inline bool drm_mm_initialized(const struct drm_mm *mm) 281 return mm->hole_stack.next; 346 * @mm: the struct drm_mm range manger 356 #define drm_mm_nodes(mm) (&(mm)->head_node.node_list) 361 * @mm: &drm_mm allocator to wal [all...] |
/src/sys/external/bsd/drm2/i915drm/ |
i915_gem_userptr.c | 51 dev_priv->mm.userptr_wq = 55 if (!dev_priv->mm.userptr_wq) 64 destroy_workqueue(dev_priv->mm.userptr_wq);
|
/src/sys/external/bsd/drm2/dist/drm/ |
drm_mm.c | 123 static void show_leaks(struct drm_mm *mm) 134 list_for_each_entry(node, drm_mm_nodes(mm), node_list) { 154 static void show_leaks(struct drm_mm *mm) { } 169 struct drm_mm *mm = __UNCONST(mm_const); local in function:INTERVAL_TREE_DEFINE 172 list_for_each_entry(node, &mm->head_node.node_list, node_list) { 176 return &mm->head_node; 178 return drm_mm_interval_tree_iter_first((struct rb_root_cached *)&mm->interval_tree, 179 start, last) ?: (struct drm_mm_node *)&mm->head_node; 188 struct drm_mm *mm = hole_node->mm; local in function:drm_mm_interval_tree_add_node 363 struct drm_mm *mm = node->mm; local in function:add_hole 717 struct drm_mm *mm = node->mm; local in function:drm_mm_remove_node 754 struct drm_mm *mm = old->mm; local in function:drm_mm_replace_node 877 struct drm_mm *mm = scan->mm; local in function:drm_mm_scan_add_block 1008 struct drm_mm *mm = scan->mm; local in function:drm_mm_scan_color_evict [all...] |