1 /* $NetBSD: nouveau_nvkm_subdev_mmu_uvmm.c,v 1.3 2021/12/19 10:51:58 riastradh Exp $ */ 2 3 /* 4 * Copyright 2017 Red Hat Inc. 5 * 6 * Permission is hereby granted, free of charge, to any person obtaining a 7 * copy of this software and associated documentation files (the "Software"), 8 * to deal in the Software without restriction, including without limitation 9 * the rights to use, copy, modify, merge, publish, distribute, sublicense, 10 * and/or sell copies of the Software, and to permit persons to whom the 11 * Software is furnished to do so, subject to the following conditions: 12 * 13 * The above copyright notice and this permission notice shall be included in 14 * all copies or substantial portions of the Software. 15 * 16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 19 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR 20 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, 21 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR 22 * OTHER DEALINGS IN THE SOFTWARE. 23 */ 24 #include <sys/cdefs.h> 25 __KERNEL_RCSID(0, "$NetBSD: nouveau_nvkm_subdev_mmu_uvmm.c,v 1.3 2021/12/19 10:51:58 riastradh Exp $"); 26 27 #include "uvmm.h" 28 #include "umem.h" 29 #include "ummu.h" 30 31 #include <core/client.h> 32 #include <core/memory.h> 33 34 #include <nvif/if000c.h> 35 #include <nvif/unpack.h> 36 37 static const struct nvkm_object_func nvkm_uvmm; 38 struct nvkm_vmm * 39 nvkm_uvmm_search(struct nvkm_client *client, u64 handle) 40 { 41 struct nvkm_object *object; 42 43 object = nvkm_object_search(client, handle, &nvkm_uvmm); 44 if (IS_ERR(object)) 45 return (void *)object; 46 47 return nvkm_uvmm(object)->vmm; 48 } 49 50 static int 51 nvkm_uvmm_mthd_pfnclr(struct nvkm_uvmm *uvmm, void *argv, u32 argc) 52 { 53 struct nvkm_client *client = uvmm->object.client; 54 union { 55 struct nvif_vmm_pfnclr_v0 v0; 56 } *args = argv; 57 struct nvkm_vmm *vmm = uvmm->vmm; 58 int ret = -ENOSYS; 59 u64 addr, size; 60 61 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) { 62 addr = args->v0.addr; 63 size = args->v0.size; 64 } else 65 return ret; 66 67 if (!client->super) 68 return -ENOENT; 69 70 if (size) { 71 mutex_lock(&vmm->mutex); 72 ret = nvkm_vmm_pfn_unmap(vmm, addr, size); 73 mutex_unlock(&vmm->mutex); 74 } 75 76 return ret; 77 } 78 79 static int 80 nvkm_uvmm_mthd_pfnmap(struct nvkm_uvmm *uvmm, void *argv, u32 argc) 81 { 82 struct nvkm_client *client = uvmm->object.client; 83 union { 84 struct nvif_vmm_pfnmap_v0 v0; 85 } *args = argv; 86 struct nvkm_vmm *vmm = uvmm->vmm; 87 int ret = -ENOSYS; 88 u64 addr, size, *phys; 89 u8 page; 90 91 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, true))) { 92 page = args->v0.page; 93 addr = args->v0.addr; 94 size = args->v0.size; 95 phys = args->v0.phys; 96 if (argc != (size >> page) * sizeof(args->v0.phys[0])) 97 return -EINVAL; 98 } else 99 return ret; 100 101 if (!client->super) 102 return -ENOENT; 103 104 if (size) { 105 mutex_lock(&vmm->mutex); 106 ret = nvkm_vmm_pfn_map(vmm, page, addr, size, phys); 107 mutex_unlock(&vmm->mutex); 108 } 109 110 return ret; 111 } 112 113 static int 114 nvkm_uvmm_mthd_unmap(struct nvkm_uvmm *uvmm, void *argv, u32 argc) 115 { 116 struct nvkm_client *client = uvmm->object.client; 117 union { 118 struct nvif_vmm_unmap_v0 v0; 119 } *args = argv; 120 struct nvkm_vmm *vmm = uvmm->vmm; 121 struct nvkm_vma *vma; 122 int ret = -ENOSYS; 123 u64 addr; 124 125 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) { 126 addr = args->v0.addr; 127 } else 128 return ret; 129 130 mutex_lock(&vmm->mutex); 131 vma = nvkm_vmm_node_search(vmm, addr); 132 if (ret = -ENOENT, !vma || vma->addr != addr) { 133 VMM_DEBUG(vmm, "lookup %016"PRIx64": %016"PRIx64"", 134 addr, vma ? vma->addr : ~(u64)0); 135 goto done; 136 } 137 138 if (ret = -ENOENT, (!vma->user && !client->super) || vma->busy) { 139 VMM_DEBUG(vmm, "denied %016"PRIx64": %d %d %d", addr, 140 vma->user, !client->super, vma->busy); 141 goto done; 142 } 143 144 if (ret = -EINVAL, !vma->memory) { 145 VMM_DEBUG(vmm, "unmapped"); 146 goto done; 147 } 148 149 nvkm_vmm_unmap_locked(vmm, vma, false); 150 ret = 0; 151 done: 152 mutex_unlock(&vmm->mutex); 153 return ret; 154 } 155 156 static int 157 nvkm_uvmm_mthd_map(struct nvkm_uvmm *uvmm, void *argv, u32 argc) 158 { 159 struct nvkm_client *client = uvmm->object.client; 160 union { 161 struct nvif_vmm_map_v0 v0; 162 } *args = argv; 163 u64 addr, size, handle, offset; 164 struct nvkm_vmm *vmm = uvmm->vmm; 165 struct nvkm_vma *vma; 166 struct nvkm_memory *memory; 167 int ret = -ENOSYS; 168 169 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, true))) { 170 addr = args->v0.addr; 171 size = args->v0.size; 172 handle = args->v0.memory; 173 offset = args->v0.offset; 174 } else 175 return ret; 176 177 memory = nvkm_umem_search(client, handle); 178 if (IS_ERR(memory)) { 179 VMM_DEBUG(vmm, "memory %016"PRIx64" %ld\n", handle, PTR_ERR(memory)); 180 return PTR_ERR(memory); 181 } 182 183 mutex_lock(&vmm->mutex); 184 if (ret = -ENOENT, !(vma = nvkm_vmm_node_search(vmm, addr))) { 185 VMM_DEBUG(vmm, "lookup %016"PRIx64"", addr); 186 goto fail; 187 } 188 189 if (ret = -ENOENT, (!vma->user && !client->super) || vma->busy) { 190 VMM_DEBUG(vmm, "denied %016"PRIx64": %d %d %d", addr, 191 vma->user, !client->super, vma->busy); 192 goto fail; 193 } 194 195 if (ret = -EINVAL, vma->mapped && !vma->memory) { 196 VMM_DEBUG(vmm, "pfnmap %016"PRIx64"", addr); 197 goto fail; 198 } 199 200 if (ret = -EINVAL, vma->addr != addr || vma->size != size) { 201 if (addr + size > vma->addr + vma->size || vma->memory || 202 (vma->refd == NVKM_VMA_PAGE_NONE && !vma->mapref)) { 203 VMM_DEBUG(vmm, "split %d %d %d " 204 "%016"PRIx64" %016"PRIx64" %016"PRIx64" %016"PRIx64"", 205 !!vma->memory, vma->refd, vma->mapref, 206 addr, size, vma->addr, (u64)vma->size); 207 goto fail; 208 } 209 210 vma = nvkm_vmm_node_split(vmm, vma, addr, size); 211 if (!vma) { 212 ret = -ENOMEM; 213 goto fail; 214 } 215 } 216 vma->busy = true; 217 mutex_unlock(&vmm->mutex); 218 219 ret = nvkm_memory_map(memory, offset, vmm, vma, argv, argc); 220 if (ret == 0) { 221 /* Successful map will clear vma->busy. */ 222 nvkm_memory_unref(&memory); 223 return 0; 224 } 225 226 mutex_lock(&vmm->mutex); 227 vma->busy = false; 228 nvkm_vmm_unmap_region(vmm, vma); 229 fail: 230 mutex_unlock(&vmm->mutex); 231 nvkm_memory_unref(&memory); 232 return ret; 233 } 234 235 static int 236 nvkm_uvmm_mthd_put(struct nvkm_uvmm *uvmm, void *argv, u32 argc) 237 { 238 struct nvkm_client *client = uvmm->object.client; 239 union { 240 struct nvif_vmm_put_v0 v0; 241 } *args = argv; 242 struct nvkm_vmm *vmm = uvmm->vmm; 243 struct nvkm_vma *vma; 244 int ret = -ENOSYS; 245 u64 addr; 246 247 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) { 248 addr = args->v0.addr; 249 } else 250 return ret; 251 252 mutex_lock(&vmm->mutex); 253 vma = nvkm_vmm_node_search(vmm, args->v0.addr); 254 if (ret = -ENOENT, !vma || vma->addr != addr || vma->part) { 255 VMM_DEBUG(vmm, "lookup %016"PRIx64": %016"PRIx64" %d", addr, 256 vma ? vma->addr : ~(u64)0, vma ? vma->part : 0); 257 goto done; 258 } 259 260 if (ret = -ENOENT, (!vma->user && !client->super) || vma->busy) { 261 VMM_DEBUG(vmm, "denied %016"PRIx64": %d %d %d", addr, 262 vma->user, !client->super, vma->busy); 263 goto done; 264 } 265 266 nvkm_vmm_put_locked(vmm, vma); 267 ret = 0; 268 done: 269 mutex_unlock(&vmm->mutex); 270 return ret; 271 } 272 273 static int 274 nvkm_uvmm_mthd_get(struct nvkm_uvmm *uvmm, void *argv, u32 argc) 275 { 276 struct nvkm_client *client = uvmm->object.client; 277 union { 278 struct nvif_vmm_get_v0 v0; 279 } *args = argv; 280 struct nvkm_vmm *vmm = uvmm->vmm; 281 struct nvkm_vma *vma; 282 int ret = -ENOSYS; 283 bool getref, mapref, sparse; 284 u8 page, align; 285 u64 size; 286 287 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) { 288 getref = args->v0.type == NVIF_VMM_GET_V0_PTES; 289 mapref = args->v0.type == NVIF_VMM_GET_V0_ADDR; 290 sparse = args->v0.sparse; 291 page = args->v0.page; 292 align = args->v0.align; 293 size = args->v0.size; 294 } else 295 return ret; 296 297 mutex_lock(&vmm->mutex); 298 ret = nvkm_vmm_get_locked(vmm, getref, mapref, sparse, 299 page, align, size, &vma); 300 mutex_unlock(&vmm->mutex); 301 if (ret) 302 return ret; 303 304 args->v0.addr = vma->addr; 305 vma->user = !client->super; 306 return ret; 307 } 308 309 static int 310 nvkm_uvmm_mthd_page(struct nvkm_uvmm *uvmm, void *argv, u32 argc) 311 { 312 union { 313 struct nvif_vmm_page_v0 v0; 314 } *args = argv; 315 const struct nvkm_vmm_page *page; 316 int ret = -ENOSYS; 317 u8 type, index, nr; 318 319 page = uvmm->vmm->func->page; 320 for (nr = 0; page[nr].shift; nr++); 321 322 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) { 323 if ((index = args->v0.index) >= nr) 324 return -EINVAL; 325 type = page[index].type; 326 args->v0.shift = page[index].shift; 327 args->v0.sparse = !!(type & NVKM_VMM_PAGE_SPARSE); 328 args->v0.vram = !!(type & NVKM_VMM_PAGE_VRAM); 329 args->v0.host = !!(type & NVKM_VMM_PAGE_HOST); 330 args->v0.comp = !!(type & NVKM_VMM_PAGE_COMP); 331 } else 332 return -ENOSYS; 333 334 return 0; 335 } 336 337 static int 338 nvkm_uvmm_mthd(struct nvkm_object *object, u32 mthd, void *argv, u32 argc) 339 { 340 struct nvkm_uvmm *uvmm = nvkm_uvmm(object); 341 switch (mthd) { 342 case NVIF_VMM_V0_PAGE : return nvkm_uvmm_mthd_page (uvmm, argv, argc); 343 case NVIF_VMM_V0_GET : return nvkm_uvmm_mthd_get (uvmm, argv, argc); 344 case NVIF_VMM_V0_PUT : return nvkm_uvmm_mthd_put (uvmm, argv, argc); 345 case NVIF_VMM_V0_MAP : return nvkm_uvmm_mthd_map (uvmm, argv, argc); 346 case NVIF_VMM_V0_UNMAP : return nvkm_uvmm_mthd_unmap (uvmm, argv, argc); 347 case NVIF_VMM_V0_PFNMAP: return nvkm_uvmm_mthd_pfnmap(uvmm, argv, argc); 348 case NVIF_VMM_V0_PFNCLR: return nvkm_uvmm_mthd_pfnclr(uvmm, argv, argc); 349 case NVIF_VMM_V0_MTHD(0x00) ... NVIF_VMM_V0_MTHD(0x7f): 350 if (uvmm->vmm->func->mthd) { 351 return uvmm->vmm->func->mthd(uvmm->vmm, 352 uvmm->object.client, 353 mthd, argv, argc); 354 } 355 break; 356 default: 357 break; 358 } 359 return -EINVAL; 360 } 361 362 static void * 363 nvkm_uvmm_dtor(struct nvkm_object *object) 364 { 365 struct nvkm_uvmm *uvmm = nvkm_uvmm(object); 366 nvkm_vmm_unref(&uvmm->vmm); 367 return uvmm; 368 } 369 370 static const struct nvkm_object_func 371 nvkm_uvmm = { 372 .dtor = nvkm_uvmm_dtor, 373 .mthd = nvkm_uvmm_mthd, 374 }; 375 376 int 377 nvkm_uvmm_new(const struct nvkm_oclass *oclass, void *argv, u32 argc, 378 struct nvkm_object **pobject) 379 { 380 struct nvkm_mmu *mmu = nvkm_ummu(oclass->parent)->mmu; 381 const bool more = oclass->base.maxver >= 0; 382 union { 383 struct nvif_vmm_v0 v0; 384 } *args = argv; 385 const struct nvkm_vmm_page *page; 386 struct nvkm_uvmm *uvmm; 387 int ret = -ENOSYS; 388 u64 addr, size; 389 bool managed; 390 391 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, more))) { 392 managed = args->v0.managed != 0; 393 addr = args->v0.addr; 394 size = args->v0.size; 395 } else 396 return ret; 397 398 if (!(uvmm = kzalloc(sizeof(*uvmm), GFP_KERNEL))) 399 return -ENOMEM; 400 nvkm_object_ctor(&nvkm_uvmm, oclass, &uvmm->object); 401 *pobject = &uvmm->object; 402 403 if (!mmu->vmm) { 404 ret = mmu->func->vmm.ctor(mmu, managed, addr, size, argv, argc, 405 NULL, "user", &uvmm->vmm); 406 if (ret) 407 return ret; 408 409 uvmm->vmm->debug = max(uvmm->vmm->debug, oclass->client->debug); 410 } else { 411 if (size) 412 return -EINVAL; 413 414 uvmm->vmm = nvkm_vmm_ref(mmu->vmm); 415 } 416 417 page = uvmm->vmm->func->page; 418 args->v0.page_nr = 0; 419 while (page && (page++)->shift) 420 args->v0.page_nr++; 421 args->v0.addr = uvmm->vmm->start; 422 args->v0.size = uvmm->vmm->limit; 423 return 0; 424 } 425