1 1.22 riastrad /* $NetBSD: linux_dma_resv.c,v 1.22 2022/02/15 22:51:03 riastradh Exp $ */ 2 1.1 riastrad 3 1.1 riastrad /*- 4 1.1 riastrad * Copyright (c) 2018 The NetBSD Foundation, Inc. 5 1.1 riastrad * All rights reserved. 6 1.1 riastrad * 7 1.1 riastrad * This code is derived from software contributed to The NetBSD Foundation 8 1.1 riastrad * by Taylor R. Campbell. 9 1.1 riastrad * 10 1.1 riastrad * Redistribution and use in source and binary forms, with or without 11 1.1 riastrad * modification, are permitted provided that the following conditions 12 1.1 riastrad * are met: 13 1.1 riastrad * 1. Redistributions of source code must retain the above copyright 14 1.1 riastrad * notice, this list of conditions and the following disclaimer. 15 1.1 riastrad * 2. Redistributions in binary form must reproduce the above copyright 16 1.1 riastrad * notice, this list of conditions and the following disclaimer in the 17 1.1 riastrad * documentation and/or other materials provided with the distribution. 18 1.1 riastrad * 19 1.1 riastrad * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS 20 1.1 riastrad * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED 21 1.1 riastrad * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR 22 1.1 riastrad * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS 23 1.1 riastrad * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 24 1.1 riastrad * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 25 1.1 riastrad * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 26 1.1 riastrad * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 27 1.1 riastrad * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 28 1.1 riastrad * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 29 1.1 riastrad * POSSIBILITY OF SUCH DAMAGE. 30 1.1 riastrad */ 31 1.1 riastrad 32 1.1 riastrad #include <sys/cdefs.h> 33 1.22 riastrad __KERNEL_RCSID(0, "$NetBSD: linux_dma_resv.c,v 1.22 2022/02/15 22:51:03 riastradh Exp $"); 34 1.1 riastrad 35 1.1 riastrad #include <sys/param.h> 36 1.1 riastrad #include <sys/poll.h> 37 1.1 riastrad #include <sys/select.h> 38 1.1 riastrad 39 1.1 riastrad #include <linux/dma-fence.h> 40 1.1 riastrad #include <linux/dma-resv.h> 41 1.1 riastrad #include <linux/seqlock.h> 42 1.1 riastrad #include <linux/ww_mutex.h> 43 1.1 riastrad 44 1.1 riastrad DEFINE_WW_CLASS(reservation_ww_class __cacheline_aligned); 45 1.1 riastrad 46 1.1 riastrad static struct dma_resv_list * 47 1.1 riastrad objlist_tryalloc(uint32_t n) 48 1.1 riastrad { 49 1.1 riastrad struct dma_resv_list *list; 50 1.1 riastrad 51 1.1 riastrad list = kmem_alloc(offsetof(typeof(*list), shared[n]), KM_NOSLEEP); 52 1.1 riastrad if (list == NULL) 53 1.1 riastrad return NULL; 54 1.1 riastrad list->shared_max = n; 55 1.1 riastrad 56 1.1 riastrad return list; 57 1.1 riastrad } 58 1.1 riastrad 59 1.22 riastrad static struct dma_resv_list * 60 1.22 riastrad objlist_alloc(uint32_t n) 61 1.22 riastrad { 62 1.22 riastrad struct dma_resv_list *list; 63 1.22 riastrad 64 1.22 riastrad list = kmem_alloc(offsetof(typeof(*list), shared[n]), KM_SLEEP); 65 1.22 riastrad list->shared_max = n; 66 1.22 riastrad 67 1.22 riastrad return list; 68 1.22 riastrad } 69 1.22 riastrad 70 1.1 riastrad static void 71 1.1 riastrad objlist_free(struct dma_resv_list *list) 72 1.1 riastrad { 73 1.1 riastrad uint32_t n = list->shared_max; 74 1.1 riastrad 75 1.1 riastrad kmem_free(list, offsetof(typeof(*list), shared[n])); 76 1.1 riastrad } 77 1.1 riastrad 78 1.1 riastrad static void 79 1.1 riastrad objlist_free_cb(struct rcu_head *rcu) 80 1.1 riastrad { 81 1.1 riastrad struct dma_resv_list *list = container_of(rcu, 82 1.1 riastrad struct dma_resv_list, rol_rcu); 83 1.1 riastrad 84 1.1 riastrad objlist_free(list); 85 1.1 riastrad } 86 1.1 riastrad 87 1.1 riastrad static void 88 1.1 riastrad objlist_defer_free(struct dma_resv_list *list) 89 1.1 riastrad { 90 1.1 riastrad 91 1.1 riastrad call_rcu(&list->rol_rcu, objlist_free_cb); 92 1.1 riastrad } 93 1.1 riastrad 94 1.1 riastrad /* 95 1.1 riastrad * dma_resv_init(robj) 96 1.1 riastrad * 97 1.1 riastrad * Initialize a reservation object. Caller must later destroy it 98 1.1 riastrad * with dma_resv_fini. 99 1.1 riastrad */ 100 1.1 riastrad void 101 1.1 riastrad dma_resv_init(struct dma_resv *robj) 102 1.1 riastrad { 103 1.1 riastrad 104 1.1 riastrad ww_mutex_init(&robj->lock, &reservation_ww_class); 105 1.1 riastrad seqcount_init(&robj->seq); 106 1.1 riastrad robj->fence_excl = NULL; 107 1.1 riastrad robj->fence = NULL; 108 1.1 riastrad robj->robj_prealloc = NULL; 109 1.1 riastrad } 110 1.1 riastrad 111 1.1 riastrad /* 112 1.1 riastrad * dma_resv_fini(robj) 113 1.1 riastrad * 114 1.1 riastrad * Destroy a reservation object, freeing any memory that had been 115 1.1 riastrad * allocated for it. Caller must have exclusive access to it. 116 1.1 riastrad */ 117 1.1 riastrad void 118 1.1 riastrad dma_resv_fini(struct dma_resv *robj) 119 1.1 riastrad { 120 1.1 riastrad unsigned i; 121 1.1 riastrad 122 1.10 riastrad if (robj->robj_prealloc) { 123 1.1 riastrad objlist_free(robj->robj_prealloc); 124 1.10 riastrad robj->robj_prealloc = NULL; /* paranoia */ 125 1.10 riastrad } 126 1.1 riastrad if (robj->fence) { 127 1.10 riastrad for (i = 0; i < robj->fence->shared_count; i++) { 128 1.1 riastrad dma_fence_put(robj->fence->shared[i]); 129 1.10 riastrad robj->fence->shared[i] = NULL; /* paranoia */ 130 1.10 riastrad } 131 1.1 riastrad objlist_free(robj->fence); 132 1.10 riastrad robj->fence = NULL; /* paranoia */ 133 1.1 riastrad } 134 1.10 riastrad if (robj->fence_excl) { 135 1.1 riastrad dma_fence_put(robj->fence_excl); 136 1.10 riastrad robj->fence_excl = NULL; /* paranoia */ 137 1.10 riastrad } 138 1.1 riastrad ww_mutex_destroy(&robj->lock); 139 1.1 riastrad } 140 1.1 riastrad 141 1.1 riastrad /* 142 1.1 riastrad * dma_resv_lock(robj, ctx) 143 1.1 riastrad * 144 1.1 riastrad * Acquire a reservation object's lock. Return 0 on success, 145 1.1 riastrad * -EALREADY if caller already holds it, -EDEADLK if a 146 1.1 riastrad * higher-priority owner holds it and the caller must back out and 147 1.1 riastrad * retry. 148 1.1 riastrad */ 149 1.1 riastrad int 150 1.1 riastrad dma_resv_lock(struct dma_resv *robj, 151 1.1 riastrad struct ww_acquire_ctx *ctx) 152 1.1 riastrad { 153 1.1 riastrad 154 1.1 riastrad return ww_mutex_lock(&robj->lock, ctx); 155 1.1 riastrad } 156 1.1 riastrad 157 1.1 riastrad /* 158 1.2 riastrad * dma_resv_lock_slow(robj, ctx) 159 1.2 riastrad * 160 1.2 riastrad * Acquire a reservation object's lock. Caller must not hold 161 1.2 riastrad * this lock or any others -- this is to be used in slow paths 162 1.2 riastrad * after dma_resv_lock or dma_resv_lock_interruptible has failed 163 1.2 riastrad * and the caller has backed out all other locks. 164 1.2 riastrad */ 165 1.2 riastrad void 166 1.2 riastrad dma_resv_lock_slow(struct dma_resv *robj, 167 1.2 riastrad struct ww_acquire_ctx *ctx) 168 1.2 riastrad { 169 1.2 riastrad 170 1.2 riastrad ww_mutex_lock_slow(&robj->lock, ctx); 171 1.2 riastrad } 172 1.2 riastrad 173 1.2 riastrad /* 174 1.1 riastrad * dma_resv_lock_interruptible(robj, ctx) 175 1.1 riastrad * 176 1.1 riastrad * Acquire a reservation object's lock. Return 0 on success, 177 1.1 riastrad * -EALREADY if caller already holds it, -EDEADLK if a 178 1.1 riastrad * higher-priority owner holds it and the caller must back out and 179 1.21 riastrad * retry, -EINTR if interrupted. 180 1.1 riastrad */ 181 1.1 riastrad int 182 1.1 riastrad dma_resv_lock_interruptible(struct dma_resv *robj, 183 1.1 riastrad struct ww_acquire_ctx *ctx) 184 1.1 riastrad { 185 1.1 riastrad 186 1.1 riastrad return ww_mutex_lock_interruptible(&robj->lock, ctx); 187 1.1 riastrad } 188 1.1 riastrad 189 1.1 riastrad /* 190 1.2 riastrad * dma_resv_lock_slow_interruptible(robj, ctx) 191 1.2 riastrad * 192 1.2 riastrad * Acquire a reservation object's lock. Caller must not hold 193 1.2 riastrad * this lock or any others -- this is to be used in slow paths 194 1.2 riastrad * after dma_resv_lock or dma_resv_lock_interruptible has failed 195 1.2 riastrad * and the caller has backed out all other locks. Return 0 on 196 1.21 riastrad * success, -EINTR if interrupted. 197 1.2 riastrad */ 198 1.2 riastrad int 199 1.2 riastrad dma_resv_lock_slow_interruptible(struct dma_resv *robj, 200 1.2 riastrad struct ww_acquire_ctx *ctx) 201 1.2 riastrad { 202 1.2 riastrad 203 1.2 riastrad return ww_mutex_lock_slow_interruptible(&robj->lock, ctx); 204 1.2 riastrad } 205 1.2 riastrad 206 1.2 riastrad /* 207 1.1 riastrad * dma_resv_trylock(robj) 208 1.1 riastrad * 209 1.1 riastrad * Try to acquire a reservation object's lock without blocking. 210 1.1 riastrad * Return true on success, false on failure. 211 1.1 riastrad */ 212 1.1 riastrad bool 213 1.1 riastrad dma_resv_trylock(struct dma_resv *robj) 214 1.1 riastrad { 215 1.1 riastrad 216 1.1 riastrad return ww_mutex_trylock(&robj->lock); 217 1.1 riastrad } 218 1.1 riastrad 219 1.1 riastrad /* 220 1.4 riastrad * dma_resv_locking_ctx(robj) 221 1.4 riastrad * 222 1.4 riastrad * Return a pointer to the ww_acquire_ctx used by the owner of 223 1.4 riastrad * the reservation object's lock, or NULL if it is either not 224 1.4 riastrad * owned or if it is locked without context. 225 1.4 riastrad */ 226 1.4 riastrad struct ww_acquire_ctx * 227 1.4 riastrad dma_resv_locking_ctx(struct dma_resv *robj) 228 1.4 riastrad { 229 1.4 riastrad 230 1.4 riastrad return ww_mutex_locking_ctx(&robj->lock); 231 1.4 riastrad } 232 1.4 riastrad 233 1.4 riastrad /* 234 1.1 riastrad * dma_resv_unlock(robj) 235 1.1 riastrad * 236 1.1 riastrad * Release a reservation object's lock. 237 1.1 riastrad */ 238 1.1 riastrad void 239 1.1 riastrad dma_resv_unlock(struct dma_resv *robj) 240 1.1 riastrad { 241 1.1 riastrad 242 1.1 riastrad return ww_mutex_unlock(&robj->lock); 243 1.1 riastrad } 244 1.1 riastrad 245 1.1 riastrad /* 246 1.11 riastrad * dma_resv_is_locked(robj) 247 1.11 riastrad * 248 1.11 riastrad * True if robj is locked. 249 1.11 riastrad */ 250 1.11 riastrad bool 251 1.11 riastrad dma_resv_is_locked(struct dma_resv *robj) 252 1.11 riastrad { 253 1.11 riastrad 254 1.11 riastrad return ww_mutex_is_locked(&robj->lock); 255 1.11 riastrad } 256 1.11 riastrad 257 1.11 riastrad /* 258 1.1 riastrad * dma_resv_held(robj) 259 1.1 riastrad * 260 1.1 riastrad * True if robj is locked. 261 1.1 riastrad */ 262 1.1 riastrad bool 263 1.1 riastrad dma_resv_held(struct dma_resv *robj) 264 1.1 riastrad { 265 1.1 riastrad 266 1.1 riastrad return ww_mutex_is_locked(&robj->lock); 267 1.1 riastrad } 268 1.1 riastrad 269 1.1 riastrad /* 270 1.1 riastrad * dma_resv_assert_held(robj) 271 1.1 riastrad * 272 1.1 riastrad * Panic if robj is not held, in DIAGNOSTIC builds. 273 1.1 riastrad */ 274 1.1 riastrad void 275 1.1 riastrad dma_resv_assert_held(struct dma_resv *robj) 276 1.1 riastrad { 277 1.1 riastrad 278 1.1 riastrad KASSERT(dma_resv_held(robj)); 279 1.1 riastrad } 280 1.1 riastrad 281 1.1 riastrad /* 282 1.1 riastrad * dma_resv_get_excl(robj) 283 1.1 riastrad * 284 1.1 riastrad * Return a pointer to the exclusive fence of the reservation 285 1.1 riastrad * object robj. 286 1.1 riastrad * 287 1.1 riastrad * Caller must have robj locked. 288 1.1 riastrad */ 289 1.1 riastrad struct dma_fence * 290 1.1 riastrad dma_resv_get_excl(struct dma_resv *robj) 291 1.1 riastrad { 292 1.1 riastrad 293 1.1 riastrad KASSERT(dma_resv_held(robj)); 294 1.1 riastrad return robj->fence_excl; 295 1.1 riastrad } 296 1.1 riastrad 297 1.1 riastrad /* 298 1.1 riastrad * dma_resv_get_list(robj) 299 1.1 riastrad * 300 1.1 riastrad * Return a pointer to the shared fence list of the reservation 301 1.1 riastrad * object robj. 302 1.1 riastrad * 303 1.1 riastrad * Caller must have robj locked. 304 1.1 riastrad */ 305 1.1 riastrad struct dma_resv_list * 306 1.1 riastrad dma_resv_get_list(struct dma_resv *robj) 307 1.1 riastrad { 308 1.1 riastrad 309 1.1 riastrad KASSERT(dma_resv_held(robj)); 310 1.1 riastrad return robj->fence; 311 1.1 riastrad } 312 1.1 riastrad 313 1.1 riastrad /* 314 1.18 riastrad * dma_resv_reserve_shared(robj, num_fences) 315 1.1 riastrad * 316 1.18 riastrad * Reserve space in robj to add num_fences shared fences. To be 317 1.18 riastrad * used only once before calling dma_resv_add_shared_fence. 318 1.1 riastrad * 319 1.1 riastrad * Caller must have robj locked. 320 1.1 riastrad * 321 1.1 riastrad * Internally, we start with room for four entries and double if 322 1.1 riastrad * we don't have enough. This is not guaranteed. 323 1.1 riastrad */ 324 1.1 riastrad int 325 1.3 riastrad dma_resv_reserve_shared(struct dma_resv *robj, unsigned int num_fences) 326 1.1 riastrad { 327 1.1 riastrad struct dma_resv_list *list, *prealloc; 328 1.1 riastrad uint32_t n, nalloc; 329 1.1 riastrad 330 1.1 riastrad KASSERT(dma_resv_held(robj)); 331 1.1 riastrad 332 1.1 riastrad list = robj->fence; 333 1.1 riastrad prealloc = robj->robj_prealloc; 334 1.1 riastrad 335 1.1 riastrad /* If there's an existing list, check it for space. */ 336 1.1 riastrad if (list) { 337 1.1 riastrad /* If there's too many already, give up. */ 338 1.18 riastrad if (list->shared_count > UINT32_MAX - num_fences) 339 1.1 riastrad return -ENOMEM; 340 1.1 riastrad 341 1.18 riastrad /* Add some more. */ 342 1.18 riastrad n = list->shared_count + num_fences; 343 1.1 riastrad 344 1.1 riastrad /* If there's enough for one more, we're done. */ 345 1.1 riastrad if (n <= list->shared_max) 346 1.1 riastrad return 0; 347 1.1 riastrad } else { 348 1.18 riastrad /* No list already. We need space for num_fences. */ 349 1.18 riastrad n = num_fences; 350 1.1 riastrad } 351 1.1 riastrad 352 1.1 riastrad /* If not, maybe there's a preallocated list ready. */ 353 1.1 riastrad if (prealloc != NULL) { 354 1.1 riastrad /* If there's enough room in it, stop here. */ 355 1.1 riastrad if (n <= prealloc->shared_max) 356 1.1 riastrad return 0; 357 1.1 riastrad 358 1.1 riastrad /* Try to double its capacity. */ 359 1.1 riastrad nalloc = n > UINT32_MAX/2 ? UINT32_MAX : 2*n; 360 1.22 riastrad prealloc = objlist_alloc(nalloc); 361 1.1 riastrad 362 1.1 riastrad /* Swap the new preallocated list and free the old one. */ 363 1.1 riastrad objlist_free(robj->robj_prealloc); 364 1.1 riastrad robj->robj_prealloc = prealloc; 365 1.1 riastrad } else { 366 1.1 riastrad /* Start with some spare. */ 367 1.1 riastrad nalloc = n > UINT32_MAX/2 ? UINT32_MAX : MAX(2*n, 4); 368 1.22 riastrad prealloc = objlist_alloc(nalloc); 369 1.22 riastrad 370 1.1 riastrad /* Save the new preallocated list. */ 371 1.1 riastrad robj->robj_prealloc = prealloc; 372 1.1 riastrad } 373 1.1 riastrad 374 1.1 riastrad /* Success! */ 375 1.1 riastrad return 0; 376 1.1 riastrad } 377 1.1 riastrad 378 1.1 riastrad struct dma_resv_write_ticket { 379 1.1 riastrad }; 380 1.1 riastrad 381 1.1 riastrad /* 382 1.1 riastrad * dma_resv_write_begin(robj, ticket) 383 1.1 riastrad * 384 1.1 riastrad * Begin an atomic batch of writes to robj, and initialize opaque 385 1.1 riastrad * ticket for it. The ticket must be passed to 386 1.1 riastrad * dma_resv_write_commit to commit the writes. 387 1.1 riastrad * 388 1.1 riastrad * Caller must have robj locked. 389 1.1 riastrad * 390 1.1 riastrad * Implies membar_producer, i.e. store-before-store barrier. Does 391 1.1 riastrad * NOT serve as an acquire operation, however. 392 1.1 riastrad */ 393 1.1 riastrad static void 394 1.1 riastrad dma_resv_write_begin(struct dma_resv *robj, 395 1.1 riastrad struct dma_resv_write_ticket *ticket) 396 1.1 riastrad { 397 1.1 riastrad 398 1.1 riastrad KASSERT(dma_resv_held(robj)); 399 1.1 riastrad 400 1.1 riastrad write_seqcount_begin(&robj->seq); 401 1.1 riastrad } 402 1.1 riastrad 403 1.1 riastrad /* 404 1.1 riastrad * dma_resv_write_commit(robj, ticket) 405 1.1 riastrad * 406 1.1 riastrad * Commit an atomic batch of writes to robj begun with the call to 407 1.1 riastrad * dma_resv_write_begin that returned ticket. 408 1.1 riastrad * 409 1.1 riastrad * Caller must have robj locked. 410 1.1 riastrad * 411 1.1 riastrad * Implies membar_producer, i.e. store-before-store barrier. Does 412 1.1 riastrad * NOT serve as a release operation, however. 413 1.1 riastrad */ 414 1.1 riastrad static void 415 1.1 riastrad dma_resv_write_commit(struct dma_resv *robj, 416 1.1 riastrad struct dma_resv_write_ticket *ticket) 417 1.1 riastrad { 418 1.1 riastrad 419 1.1 riastrad KASSERT(dma_resv_held(robj)); 420 1.1 riastrad 421 1.1 riastrad write_seqcount_end(&robj->seq); 422 1.1 riastrad } 423 1.1 riastrad 424 1.1 riastrad struct dma_resv_read_ticket { 425 1.1 riastrad unsigned version; 426 1.1 riastrad }; 427 1.1 riastrad 428 1.1 riastrad /* 429 1.1 riastrad * dma_resv_read_begin(robj, ticket) 430 1.1 riastrad * 431 1.1 riastrad * Begin a read section, and initialize opaque ticket for it. The 432 1.1 riastrad * ticket must be passed to dma_resv_read_exit, and the 433 1.1 riastrad * caller must be prepared to retry reading if it fails. 434 1.1 riastrad */ 435 1.1 riastrad static void 436 1.1 riastrad dma_resv_read_begin(const struct dma_resv *robj, 437 1.1 riastrad struct dma_resv_read_ticket *ticket) 438 1.1 riastrad { 439 1.1 riastrad 440 1.1 riastrad ticket->version = read_seqcount_begin(&robj->seq); 441 1.1 riastrad } 442 1.1 riastrad 443 1.1 riastrad /* 444 1.1 riastrad * dma_resv_read_valid(robj, ticket) 445 1.1 riastrad * 446 1.1 riastrad * Test whether the read sections are valid. Return true on 447 1.1 riastrad * success, or false on failure if the read ticket has been 448 1.1 riastrad * invalidated. 449 1.1 riastrad */ 450 1.1 riastrad static bool 451 1.1 riastrad dma_resv_read_valid(const struct dma_resv *robj, 452 1.1 riastrad struct dma_resv_read_ticket *ticket) 453 1.1 riastrad { 454 1.1 riastrad 455 1.1 riastrad return !read_seqcount_retry(&robj->seq, ticket->version); 456 1.1 riastrad } 457 1.1 riastrad 458 1.1 riastrad /* 459 1.12 riastrad * dma_resv_get_shared_reader(robj, listp, shared_countp, ticket) 460 1.12 riastrad * 461 1.12 riastrad * Set *listp and *shared_countp to a snapshot of the pointer to 462 1.12 riastrad * and length of the shared fence list of robj and return true, or 463 1.12 riastrad * set them to NULL/0 and return false if a writer intervened so 464 1.12 riastrad * the caller must start over. 465 1.12 riastrad * 466 1.12 riastrad * Both *listp and *shared_countp are unconditionally initialized 467 1.12 riastrad * on return. They may be NULL/0 even on success, if there is no 468 1.12 riastrad * shared list at the moment. Does not take any fence references. 469 1.12 riastrad */ 470 1.12 riastrad static bool 471 1.12 riastrad dma_resv_get_shared_reader(const struct dma_resv *robj, 472 1.12 riastrad const struct dma_resv_list **listp, unsigned *shared_countp, 473 1.12 riastrad struct dma_resv_read_ticket *ticket) 474 1.12 riastrad { 475 1.12 riastrad struct dma_resv_list *list; 476 1.12 riastrad unsigned shared_count = 0; 477 1.12 riastrad 478 1.12 riastrad /* 479 1.12 riastrad * Get the list and, if it is present, its length. If the list 480 1.12 riastrad * is present, it has a valid length. The atomic_load_consume 481 1.12 riastrad * pairs with the membar_producer in dma_resv_write_begin. 482 1.12 riastrad */ 483 1.12 riastrad list = atomic_load_consume(&robj->fence); 484 1.12 riastrad shared_count = list ? atomic_load_relaxed(&list->shared_count) : 0; 485 1.12 riastrad 486 1.12 riastrad /* 487 1.12 riastrad * We are done reading from robj and list. Validate our 488 1.12 riastrad * parking ticket. If it's invalid, do not pass go and do not 489 1.12 riastrad * collect $200. 490 1.12 riastrad */ 491 1.12 riastrad if (!dma_resv_read_valid(robj, ticket)) 492 1.12 riastrad goto fail; 493 1.12 riastrad 494 1.12 riastrad /* Success! */ 495 1.12 riastrad *listp = list; 496 1.12 riastrad *shared_countp = shared_count; 497 1.12 riastrad return true; 498 1.12 riastrad 499 1.12 riastrad fail: *listp = NULL; 500 1.12 riastrad *shared_countp = 0; 501 1.12 riastrad return false; 502 1.12 riastrad } 503 1.12 riastrad 504 1.12 riastrad /* 505 1.12 riastrad * dma_resv_get_excl_reader(robj, fencep, ticket) 506 1.12 riastrad * 507 1.12 riastrad * Set *fencep to the exclusive fence of robj and return true, or 508 1.12 riastrad * set it to NULL and return false if either 509 1.12 riastrad * (a) a writer intervened, or 510 1.12 riastrad * (b) the fence is scheduled to be destroyed after this RCU grace 511 1.12 riastrad * period, 512 1.12 riastrad * in either case meaning the caller must restart. 513 1.12 riastrad * 514 1.12 riastrad * The value of *fencep is unconditionally initialized on return. 515 1.12 riastrad * It may be NULL, if there is no exclusive fence at the moment. 516 1.12 riastrad * If nonnull, *fencep is referenced; caller must dma_fence_put. 517 1.12 riastrad */ 518 1.12 riastrad static bool 519 1.12 riastrad dma_resv_get_excl_reader(const struct dma_resv *robj, 520 1.12 riastrad struct dma_fence **fencep, 521 1.12 riastrad struct dma_resv_read_ticket *ticket) 522 1.12 riastrad { 523 1.12 riastrad struct dma_fence *fence; 524 1.12 riastrad 525 1.12 riastrad /* 526 1.12 riastrad * Get the candidate fence pointer. The atomic_load_consume 527 1.12 riastrad * pairs with the membar_consumer in dma_resv_write_begin. 528 1.12 riastrad */ 529 1.12 riastrad fence = atomic_load_consume(&robj->fence_excl); 530 1.12 riastrad 531 1.12 riastrad /* 532 1.12 riastrad * The load of robj->fence_excl is atomic, but the caller may 533 1.12 riastrad * have previously loaded the shared fence list and should 534 1.12 riastrad * restart if its view of the entire dma_resv object is not a 535 1.12 riastrad * consistent snapshot. 536 1.12 riastrad */ 537 1.12 riastrad if (!dma_resv_read_valid(robj, ticket)) 538 1.12 riastrad goto fail; 539 1.12 riastrad 540 1.12 riastrad /* 541 1.12 riastrad * If the fence is already scheduled to away after this RCU 542 1.12 riastrad * read section, give up. Otherwise, take a reference so it 543 1.12 riastrad * won't go away until after dma_fence_put. 544 1.12 riastrad */ 545 1.12 riastrad if (fence != NULL && 546 1.12 riastrad (fence = dma_fence_get_rcu(fence)) == NULL) 547 1.12 riastrad goto fail; 548 1.12 riastrad 549 1.12 riastrad /* Success! */ 550 1.12 riastrad *fencep = fence; 551 1.12 riastrad return true; 552 1.12 riastrad 553 1.12 riastrad fail: *fencep = NULL; 554 1.12 riastrad return false; 555 1.12 riastrad } 556 1.12 riastrad 557 1.12 riastrad /* 558 1.1 riastrad * dma_resv_add_excl_fence(robj, fence) 559 1.1 riastrad * 560 1.1 riastrad * Empty and release all of robj's shared fences, and clear and 561 1.1 riastrad * release its exclusive fence. If fence is nonnull, acquire a 562 1.1 riastrad * reference to it and save it as robj's exclusive fence. 563 1.1 riastrad * 564 1.1 riastrad * Caller must have robj locked. 565 1.1 riastrad */ 566 1.1 riastrad void 567 1.1 riastrad dma_resv_add_excl_fence(struct dma_resv *robj, 568 1.1 riastrad struct dma_fence *fence) 569 1.1 riastrad { 570 1.1 riastrad struct dma_fence *old_fence = robj->fence_excl; 571 1.1 riastrad struct dma_resv_list *old_list = robj->fence; 572 1.1 riastrad uint32_t old_shared_count; 573 1.1 riastrad struct dma_resv_write_ticket ticket; 574 1.1 riastrad 575 1.1 riastrad KASSERT(dma_resv_held(robj)); 576 1.1 riastrad 577 1.1 riastrad /* 578 1.1 riastrad * If we are setting rather than just removing a fence, acquire 579 1.1 riastrad * a reference for ourselves. 580 1.1 riastrad */ 581 1.1 riastrad if (fence) 582 1.1 riastrad (void)dma_fence_get(fence); 583 1.1 riastrad 584 1.1 riastrad /* If there are any shared fences, remember how many. */ 585 1.1 riastrad if (old_list) 586 1.1 riastrad old_shared_count = old_list->shared_count; 587 1.1 riastrad 588 1.7 riastrad /* Begin an update. Implies membar_producer for fence. */ 589 1.1 riastrad dma_resv_write_begin(robj, &ticket); 590 1.1 riastrad 591 1.1 riastrad /* Replace the fence and zero the shared count. */ 592 1.7 riastrad atomic_store_relaxed(&robj->fence_excl, fence); 593 1.1 riastrad if (old_list) 594 1.1 riastrad old_list->shared_count = 0; 595 1.1 riastrad 596 1.1 riastrad /* Commit the update. */ 597 1.1 riastrad dma_resv_write_commit(robj, &ticket); 598 1.1 riastrad 599 1.1 riastrad /* Release the old exclusive fence, if any. */ 600 1.10 riastrad if (old_fence) { 601 1.1 riastrad dma_fence_put(old_fence); 602 1.10 riastrad old_fence = NULL; /* paranoia */ 603 1.10 riastrad } 604 1.1 riastrad 605 1.1 riastrad /* Release any old shared fences. */ 606 1.1 riastrad if (old_list) { 607 1.10 riastrad while (old_shared_count--) { 608 1.1 riastrad dma_fence_put(old_list->shared[old_shared_count]); 609 1.10 riastrad /* paranoia */ 610 1.10 riastrad old_list->shared[old_shared_count] = NULL; 611 1.10 riastrad } 612 1.1 riastrad } 613 1.1 riastrad } 614 1.1 riastrad 615 1.1 riastrad /* 616 1.1 riastrad * dma_resv_add_shared_fence(robj, fence) 617 1.1 riastrad * 618 1.1 riastrad * Acquire a reference to fence and add it to robj's shared list. 619 1.1 riastrad * If any fence was already added with the same context number, 620 1.1 riastrad * release it and replace it by this one. 621 1.1 riastrad * 622 1.1 riastrad * Caller must have robj locked, and must have preceded with a 623 1.1 riastrad * call to dma_resv_reserve_shared for each shared fence 624 1.1 riastrad * added. 625 1.1 riastrad */ 626 1.1 riastrad void 627 1.1 riastrad dma_resv_add_shared_fence(struct dma_resv *robj, 628 1.1 riastrad struct dma_fence *fence) 629 1.1 riastrad { 630 1.1 riastrad struct dma_resv_list *list = robj->fence; 631 1.1 riastrad struct dma_resv_list *prealloc = robj->robj_prealloc; 632 1.1 riastrad struct dma_resv_write_ticket ticket; 633 1.1 riastrad struct dma_fence *replace = NULL; 634 1.1 riastrad uint32_t i; 635 1.1 riastrad 636 1.1 riastrad KASSERT(dma_resv_held(robj)); 637 1.1 riastrad 638 1.1 riastrad /* Acquire a reference to the fence. */ 639 1.1 riastrad KASSERT(fence != NULL); 640 1.1 riastrad (void)dma_fence_get(fence); 641 1.1 riastrad 642 1.1 riastrad /* Check for a preallocated replacement list. */ 643 1.1 riastrad if (prealloc == NULL) { 644 1.1 riastrad /* 645 1.1 riastrad * If there is no preallocated replacement list, then 646 1.1 riastrad * there must be room in the current list. 647 1.1 riastrad */ 648 1.1 riastrad KASSERT(list != NULL); 649 1.1 riastrad KASSERT(list->shared_count < list->shared_max); 650 1.1 riastrad 651 1.1 riastrad /* Begin an update. Implies membar_producer for fence. */ 652 1.1 riastrad dma_resv_write_begin(robj, &ticket); 653 1.1 riastrad 654 1.1 riastrad /* Find a fence with the same context number. */ 655 1.1 riastrad for (i = 0; i < list->shared_count; i++) { 656 1.1 riastrad if (list->shared[i]->context == fence->context) { 657 1.1 riastrad replace = list->shared[i]; 658 1.7 riastrad atomic_store_relaxed(&list->shared[i], fence); 659 1.1 riastrad break; 660 1.1 riastrad } 661 1.1 riastrad } 662 1.1 riastrad 663 1.1 riastrad /* If we didn't find one, add it at the end. */ 664 1.7 riastrad if (i == list->shared_count) { 665 1.7 riastrad atomic_store_relaxed(&list->shared[list->shared_count], 666 1.7 riastrad fence); 667 1.7 riastrad atomic_store_relaxed(&list->shared_count, 668 1.7 riastrad list->shared_count + 1); 669 1.7 riastrad } 670 1.1 riastrad 671 1.1 riastrad /* Commit the update. */ 672 1.1 riastrad dma_resv_write_commit(robj, &ticket); 673 1.1 riastrad } else { 674 1.1 riastrad /* 675 1.1 riastrad * There is a preallocated replacement list. There may 676 1.1 riastrad * not be a current list. If not, treat it as a zero- 677 1.1 riastrad * length list. 678 1.1 riastrad */ 679 1.1 riastrad uint32_t shared_count = (list == NULL? 0 : list->shared_count); 680 1.1 riastrad 681 1.1 riastrad /* There had better be room in the preallocated list. */ 682 1.1 riastrad KASSERT(shared_count < prealloc->shared_max); 683 1.1 riastrad 684 1.1 riastrad /* 685 1.1 riastrad * Copy the fences over, but replace if we find one 686 1.1 riastrad * with the same context number. 687 1.1 riastrad */ 688 1.1 riastrad for (i = 0; i < shared_count; i++) { 689 1.1 riastrad if (replace == NULL && 690 1.1 riastrad list->shared[i]->context == fence->context) { 691 1.1 riastrad replace = list->shared[i]; 692 1.1 riastrad prealloc->shared[i] = fence; 693 1.1 riastrad } else { 694 1.1 riastrad prealloc->shared[i] = list->shared[i]; 695 1.1 riastrad } 696 1.1 riastrad } 697 1.1 riastrad prealloc->shared_count = shared_count; 698 1.1 riastrad 699 1.1 riastrad /* If we didn't find one, add it at the end. */ 700 1.22 riastrad if (replace == NULL) { 701 1.22 riastrad KASSERT(prealloc->shared_count < prealloc->shared_max); 702 1.1 riastrad prealloc->shared[prealloc->shared_count++] = fence; 703 1.22 riastrad } 704 1.1 riastrad 705 1.1 riastrad /* 706 1.1 riastrad * Now ready to replace the list. Begin an update. 707 1.1 riastrad * Implies membar_producer for fence and prealloc. 708 1.1 riastrad */ 709 1.1 riastrad dma_resv_write_begin(robj, &ticket); 710 1.1 riastrad 711 1.1 riastrad /* Replace the list. */ 712 1.7 riastrad atomic_store_relaxed(&robj->fence, prealloc); 713 1.1 riastrad robj->robj_prealloc = NULL; 714 1.1 riastrad 715 1.1 riastrad /* Commit the update. */ 716 1.1 riastrad dma_resv_write_commit(robj, &ticket); 717 1.1 riastrad 718 1.1 riastrad /* 719 1.1 riastrad * If there is an old list, free it when convenient. 720 1.1 riastrad * (We are not in a position at this point to sleep 721 1.1 riastrad * waiting for activity on all CPUs.) 722 1.1 riastrad */ 723 1.1 riastrad if (list) 724 1.1 riastrad objlist_defer_free(list); 725 1.1 riastrad } 726 1.1 riastrad 727 1.1 riastrad /* Release a fence if we replaced it. */ 728 1.10 riastrad if (replace) { 729 1.1 riastrad dma_fence_put(replace); 730 1.10 riastrad replace = NULL; /* paranoia */ 731 1.10 riastrad } 732 1.1 riastrad } 733 1.1 riastrad 734 1.1 riastrad /* 735 1.1 riastrad * dma_resv_get_excl_rcu(robj) 736 1.1 riastrad * 737 1.1 riastrad * Note: Caller need not call this from an RCU read section. 738 1.1 riastrad */ 739 1.1 riastrad struct dma_fence * 740 1.1 riastrad dma_resv_get_excl_rcu(const struct dma_resv *robj) 741 1.1 riastrad { 742 1.1 riastrad struct dma_fence *fence; 743 1.1 riastrad 744 1.1 riastrad rcu_read_lock(); 745 1.1 riastrad fence = dma_fence_get_rcu_safe(&robj->fence_excl); 746 1.1 riastrad rcu_read_unlock(); 747 1.1 riastrad 748 1.1 riastrad return fence; 749 1.1 riastrad } 750 1.1 riastrad 751 1.1 riastrad /* 752 1.1 riastrad * dma_resv_get_fences_rcu(robj, fencep, nsharedp, sharedp) 753 1.13 riastrad * 754 1.13 riastrad * Get a snapshot of the exclusive and shared fences of robj. The 755 1.13 riastrad * shared fences are returned as a pointer *sharedp to an array, 756 1.13 riastrad * to be freed by the caller with kfree, of *nsharedp elements. 757 1.17 riastrad * If fencep is null, then add the exclusive fence, if any, at the 758 1.17 riastrad * end of the array instead. 759 1.13 riastrad * 760 1.13 riastrad * Returns zero on success, negative (Linux-style) error code on 761 1.13 riastrad * failure. On failure, *fencep, *nsharedp, and *sharedp are 762 1.13 riastrad * untouched. 763 1.1 riastrad */ 764 1.1 riastrad int 765 1.1 riastrad dma_resv_get_fences_rcu(const struct dma_resv *robj, 766 1.1 riastrad struct dma_fence **fencep, unsigned *nsharedp, struct dma_fence ***sharedp) 767 1.1 riastrad { 768 1.10 riastrad const struct dma_resv_list *list = NULL; 769 1.10 riastrad struct dma_fence *fence = NULL; 770 1.1 riastrad struct dma_fence **shared = NULL; 771 1.20 riastrad unsigned shared_alloc = 0, shared_count, i; 772 1.1 riastrad struct dma_resv_read_ticket ticket; 773 1.1 riastrad 774 1.10 riastrad top: KASSERT(fence == NULL); 775 1.10 riastrad 776 1.1 riastrad /* Enter an RCU read section and get a read ticket. */ 777 1.1 riastrad rcu_read_lock(); 778 1.1 riastrad dma_resv_read_begin(robj, &ticket); 779 1.1 riastrad 780 1.12 riastrad /* If there is a shared list, grab it. */ 781 1.12 riastrad if (!dma_resv_get_shared_reader(robj, &list, &shared_count, &ticket)) 782 1.12 riastrad goto restart; 783 1.12 riastrad if (list != NULL) { 784 1.1 riastrad 785 1.17 riastrad /* 786 1.17 riastrad * Avoid arithmetic overflow with `+ 1' below. 787 1.17 riastrad * Strictly speaking we don't need this if the caller 788 1.17 riastrad * specified fencep or if there is no exclusive fence, 789 1.17 riastrad * but it is simpler to not have to consider those 790 1.17 riastrad * cases. 791 1.17 riastrad */ 792 1.17 riastrad KASSERT(shared_count <= list->shared_max); 793 1.17 riastrad if (list->shared_max == UINT_MAX) 794 1.17 riastrad return -ENOMEM; 795 1.17 riastrad 796 1.1 riastrad /* Check whether we have a buffer. */ 797 1.1 riastrad if (shared == NULL) { 798 1.1 riastrad /* 799 1.1 riastrad * We don't have a buffer yet. Try to allocate 800 1.1 riastrad * one without waiting. 801 1.1 riastrad */ 802 1.17 riastrad shared_alloc = list->shared_max + 1; 803 1.1 riastrad shared = kcalloc(shared_alloc, sizeof(shared[0]), 804 1.1 riastrad GFP_NOWAIT); 805 1.1 riastrad if (shared == NULL) { 806 1.1 riastrad /* 807 1.1 riastrad * Couldn't do it immediately. Back 808 1.1 riastrad * out of RCU and allocate one with 809 1.1 riastrad * waiting. 810 1.1 riastrad */ 811 1.1 riastrad rcu_read_unlock(); 812 1.1 riastrad shared = kcalloc(shared_alloc, 813 1.1 riastrad sizeof(shared[0]), GFP_KERNEL); 814 1.1 riastrad if (shared == NULL) 815 1.1 riastrad return -ENOMEM; 816 1.1 riastrad goto top; 817 1.1 riastrad } 818 1.17 riastrad } else if (shared_alloc < list->shared_max + 1) { 819 1.1 riastrad /* 820 1.1 riastrad * We have a buffer but it's too small. We're 821 1.1 riastrad * already racing in this case, so just back 822 1.1 riastrad * out and wait to allocate a bigger one. 823 1.1 riastrad */ 824 1.17 riastrad shared_alloc = list->shared_max + 1; 825 1.1 riastrad rcu_read_unlock(); 826 1.1 riastrad kfree(shared); 827 1.1 riastrad shared = kcalloc(shared_alloc, sizeof(shared[0]), 828 1.1 riastrad GFP_KERNEL); 829 1.1 riastrad if (shared == NULL) 830 1.1 riastrad return -ENOMEM; 831 1.19 riastrad goto top; 832 1.1 riastrad } 833 1.1 riastrad 834 1.1 riastrad /* 835 1.1 riastrad * We got a buffer large enough. Copy into the buffer 836 1.7 riastrad * and record the number of elements. Could safely use 837 1.7 riastrad * memcpy here, because even if we race with a writer 838 1.7 riastrad * it'll invalidate the read ticket and we'll start 839 1.14 riastrad * over, but atomic_load in a loop will pacify kcsan. 840 1.7 riastrad */ 841 1.7 riastrad for (i = 0; i < shared_count; i++) 842 1.7 riastrad shared[i] = atomic_load_relaxed(&list->shared[i]); 843 1.14 riastrad 844 1.14 riastrad /* If anything changed while we were copying, restart. */ 845 1.14 riastrad if (!dma_resv_read_valid(robj, &ticket)) 846 1.14 riastrad goto restart; 847 1.1 riastrad } 848 1.1 riastrad 849 1.1 riastrad /* If there is an exclusive fence, grab it. */ 850 1.10 riastrad KASSERT(fence == NULL); 851 1.12 riastrad if (!dma_resv_get_excl_reader(robj, &fence, &ticket)) 852 1.1 riastrad goto restart; 853 1.1 riastrad 854 1.1 riastrad /* 855 1.1 riastrad * Try to get a reference to all of the shared fences. 856 1.1 riastrad */ 857 1.1 riastrad for (i = 0; i < shared_count; i++) { 858 1.7 riastrad if (dma_fence_get_rcu(atomic_load_relaxed(&shared[i])) == NULL) 859 1.1 riastrad goto put_restart; 860 1.1 riastrad } 861 1.1 riastrad 862 1.1 riastrad /* Success! */ 863 1.1 riastrad rcu_read_unlock(); 864 1.20 riastrad KASSERT(shared_count <= shared_alloc); 865 1.20 riastrad KASSERT(shared_alloc == 0 || shared_count < shared_alloc); 866 1.20 riastrad KASSERT(shared_alloc <= UINT_MAX); 867 1.17 riastrad if (fencep) { 868 1.17 riastrad *fencep = fence; 869 1.17 riastrad } else if (fence) { 870 1.20 riastrad if (shared_count) { 871 1.20 riastrad shared[shared_count++] = fence; 872 1.20 riastrad } else { 873 1.20 riastrad shared = kmalloc(sizeof(shared[0]), GFP_KERNEL); 874 1.20 riastrad shared[0] = fence; 875 1.20 riastrad shared_count = 1; 876 1.20 riastrad } 877 1.17 riastrad } 878 1.1 riastrad *nsharedp = shared_count; 879 1.1 riastrad *sharedp = shared; 880 1.1 riastrad return 0; 881 1.1 riastrad 882 1.1 riastrad put_restart: 883 1.1 riastrad /* Back out. */ 884 1.1 riastrad while (i --> 0) { 885 1.1 riastrad dma_fence_put(shared[i]); 886 1.1 riastrad shared[i] = NULL; /* paranoia */ 887 1.1 riastrad } 888 1.1 riastrad if (fence) { 889 1.1 riastrad dma_fence_put(fence); 890 1.10 riastrad fence = NULL; 891 1.1 riastrad } 892 1.1 riastrad 893 1.1 riastrad restart: 894 1.10 riastrad KASSERT(fence == NULL); 895 1.1 riastrad rcu_read_unlock(); 896 1.1 riastrad goto top; 897 1.1 riastrad } 898 1.1 riastrad 899 1.1 riastrad /* 900 1.1 riastrad * dma_resv_copy_fences(dst, src) 901 1.1 riastrad * 902 1.1 riastrad * Copy the exclusive fence and all the shared fences from src to 903 1.1 riastrad * dst. 904 1.1 riastrad * 905 1.1 riastrad * Caller must have dst locked. 906 1.1 riastrad */ 907 1.1 riastrad int 908 1.1 riastrad dma_resv_copy_fences(struct dma_resv *dst_robj, 909 1.1 riastrad const struct dma_resv *src_robj) 910 1.1 riastrad { 911 1.1 riastrad const struct dma_resv_list *src_list; 912 1.1 riastrad struct dma_resv_list *dst_list = NULL; 913 1.1 riastrad struct dma_resv_list *old_list; 914 1.1 riastrad struct dma_fence *fence = NULL; 915 1.1 riastrad struct dma_fence *old_fence; 916 1.1 riastrad uint32_t shared_count, i; 917 1.1 riastrad struct dma_resv_read_ticket read_ticket; 918 1.1 riastrad struct dma_resv_write_ticket write_ticket; 919 1.1 riastrad 920 1.1 riastrad KASSERT(dma_resv_held(dst_robj)); 921 1.1 riastrad 922 1.10 riastrad top: KASSERT(fence == NULL); 923 1.10 riastrad 924 1.1 riastrad /* Enter an RCU read section and get a read ticket. */ 925 1.1 riastrad rcu_read_lock(); 926 1.1 riastrad dma_resv_read_begin(src_robj, &read_ticket); 927 1.1 riastrad 928 1.1 riastrad /* Get the shared list. */ 929 1.12 riastrad if (!dma_resv_get_shared_reader(src_robj, &src_list, &shared_count, 930 1.12 riastrad &read_ticket)) 931 1.12 riastrad goto restart; 932 1.22 riastrad if (src_list) { 933 1.22 riastrad /* Allocate a new list, if necessary. */ 934 1.1 riastrad if (dst_list == NULL) 935 1.22 riastrad dst_list = objlist_tryalloc(shared_count); 936 1.22 riastrad if (dst_list == NULL || dst_list->shared_max < shared_count) { 937 1.22 riastrad rcu_read_unlock(); 938 1.22 riastrad if (dst_list) { 939 1.22 riastrad objlist_free(dst_list); 940 1.22 riastrad dst_list = NULL; 941 1.22 riastrad } 942 1.22 riastrad dst_list = objlist_alloc(shared_count); 943 1.22 riastrad dst_list->shared_count = 0; /* paranoia */ 944 1.22 riastrad goto top; 945 1.22 riastrad } 946 1.1 riastrad 947 1.1 riastrad /* Copy over all fences that are not yet signalled. */ 948 1.1 riastrad dst_list->shared_count = 0; 949 1.1 riastrad for (i = 0; i < shared_count; i++) { 950 1.10 riastrad KASSERT(fence == NULL); 951 1.7 riastrad fence = atomic_load_relaxed(&src_list->shared[i]); 952 1.9 riastrad if ((fence = dma_fence_get_rcu(fence)) == NULL) 953 1.1 riastrad goto restart; 954 1.1 riastrad if (dma_fence_is_signaled(fence)) { 955 1.1 riastrad dma_fence_put(fence); 956 1.1 riastrad fence = NULL; 957 1.1 riastrad continue; 958 1.1 riastrad } 959 1.1 riastrad dst_list->shared[dst_list->shared_count++] = fence; 960 1.1 riastrad fence = NULL; 961 1.1 riastrad } 962 1.14 riastrad 963 1.14 riastrad /* If anything changed while we were copying, restart. */ 964 1.14 riastrad if (!dma_resv_read_valid(src_robj, &read_ticket)) 965 1.14 riastrad goto restart; 966 1.1 riastrad } 967 1.1 riastrad 968 1.1 riastrad /* Get the exclusive fence. */ 969 1.10 riastrad KASSERT(fence == NULL); 970 1.12 riastrad if (!dma_resv_get_excl_reader(src_robj, &fence, &read_ticket)) 971 1.12 riastrad goto restart; 972 1.1 riastrad 973 1.1 riastrad /* All done with src; exit the RCU read section. */ 974 1.1 riastrad rcu_read_unlock(); 975 1.1 riastrad 976 1.1 riastrad /* 977 1.1 riastrad * We now have a snapshot of the shared and exclusive fences of 978 1.1 riastrad * src_robj and we have acquired references to them so they 979 1.1 riastrad * won't go away. Transfer them over to dst_robj, releasing 980 1.1 riastrad * references to any that were there. 981 1.1 riastrad */ 982 1.1 riastrad 983 1.1 riastrad /* Get the old shared and exclusive fences, if any. */ 984 1.1 riastrad old_list = dst_robj->fence; 985 1.1 riastrad old_fence = dst_robj->fence_excl; 986 1.1 riastrad 987 1.7 riastrad /* 988 1.7 riastrad * Begin an update. Implies membar_producer for dst_list and 989 1.7 riastrad * fence. 990 1.7 riastrad */ 991 1.1 riastrad dma_resv_write_begin(dst_robj, &write_ticket); 992 1.1 riastrad 993 1.1 riastrad /* Replace the fences. */ 994 1.6 riastrad atomic_store_relaxed(&dst_robj->fence, dst_list); 995 1.6 riastrad atomic_store_relaxed(&dst_robj->fence_excl, fence); 996 1.1 riastrad 997 1.1 riastrad /* Commit the update. */ 998 1.1 riastrad dma_resv_write_commit(dst_robj, &write_ticket); 999 1.1 riastrad 1000 1.1 riastrad /* Release the old exclusive fence, if any. */ 1001 1.10 riastrad if (old_fence) { 1002 1.1 riastrad dma_fence_put(old_fence); 1003 1.10 riastrad old_fence = NULL; /* paranoia */ 1004 1.10 riastrad } 1005 1.1 riastrad 1006 1.1 riastrad /* Release any old shared fences. */ 1007 1.1 riastrad if (old_list) { 1008 1.10 riastrad for (i = old_list->shared_count; i --> 0;) { 1009 1.1 riastrad dma_fence_put(old_list->shared[i]); 1010 1.10 riastrad old_list->shared[i] = NULL; /* paranoia */ 1011 1.10 riastrad } 1012 1.10 riastrad objlist_free(old_list); 1013 1.10 riastrad old_list = NULL; /* paranoia */ 1014 1.1 riastrad } 1015 1.1 riastrad 1016 1.1 riastrad /* Success! */ 1017 1.1 riastrad return 0; 1018 1.1 riastrad 1019 1.1 riastrad restart: 1020 1.10 riastrad KASSERT(fence == NULL); 1021 1.1 riastrad rcu_read_unlock(); 1022 1.1 riastrad if (dst_list) { 1023 1.1 riastrad for (i = dst_list->shared_count; i --> 0;) { 1024 1.1 riastrad dma_fence_put(dst_list->shared[i]); 1025 1.10 riastrad dst_list->shared[i] = NULL; /* paranoia */ 1026 1.1 riastrad } 1027 1.22 riastrad /* reuse dst_list allocation for the next attempt */ 1028 1.1 riastrad } 1029 1.1 riastrad goto top; 1030 1.1 riastrad } 1031 1.1 riastrad 1032 1.1 riastrad /* 1033 1.1 riastrad * dma_resv_test_signaled_rcu(robj, shared) 1034 1.1 riastrad * 1035 1.1 riastrad * If shared is true, test whether all of the shared fences are 1036 1.1 riastrad * signalled, or if there are none, test whether the exclusive 1037 1.1 riastrad * fence is signalled. If shared is false, test only whether the 1038 1.1 riastrad * exclusive fence is signalled. 1039 1.1 riastrad * 1040 1.1 riastrad * XXX Why does this _not_ test the exclusive fence if shared is 1041 1.1 riastrad * true only if there are no shared fences? This makes no sense. 1042 1.1 riastrad */ 1043 1.1 riastrad bool 1044 1.1 riastrad dma_resv_test_signaled_rcu(const struct dma_resv *robj, 1045 1.1 riastrad bool shared) 1046 1.1 riastrad { 1047 1.1 riastrad struct dma_resv_read_ticket ticket; 1048 1.12 riastrad const struct dma_resv_list *list; 1049 1.10 riastrad struct dma_fence *fence = NULL; 1050 1.1 riastrad uint32_t i, shared_count; 1051 1.1 riastrad bool signaled = true; 1052 1.1 riastrad 1053 1.10 riastrad top: KASSERT(fence == NULL); 1054 1.10 riastrad 1055 1.1 riastrad /* Enter an RCU read section and get a read ticket. */ 1056 1.1 riastrad rcu_read_lock(); 1057 1.1 riastrad dma_resv_read_begin(robj, &ticket); 1058 1.1 riastrad 1059 1.1 riastrad /* If shared is requested and there is a shared list, test it. */ 1060 1.12 riastrad if (shared) { 1061 1.12 riastrad if (!dma_resv_get_shared_reader(robj, &list, &shared_count, 1062 1.12 riastrad &ticket)) 1063 1.1 riastrad goto restart; 1064 1.12 riastrad } else { 1065 1.12 riastrad list = NULL; 1066 1.12 riastrad shared_count = 0; 1067 1.12 riastrad } 1068 1.12 riastrad if (list != NULL) { 1069 1.1 riastrad /* 1070 1.1 riastrad * For each fence, if it is going away, restart. 1071 1.1 riastrad * Otherwise, acquire a reference to it to test whether 1072 1.1 riastrad * it is signalled. Stop if we find any that is not 1073 1.1 riastrad * signalled. 1074 1.1 riastrad */ 1075 1.1 riastrad for (i = 0; i < shared_count; i++) { 1076 1.10 riastrad KASSERT(fence == NULL); 1077 1.7 riastrad fence = atomic_load_relaxed(&list->shared[i]); 1078 1.10 riastrad if ((fence = dma_fence_get_rcu(fence)) == NULL) 1079 1.1 riastrad goto restart; 1080 1.1 riastrad signaled &= dma_fence_is_signaled(fence); 1081 1.1 riastrad dma_fence_put(fence); 1082 1.10 riastrad fence = NULL; 1083 1.1 riastrad if (!signaled) 1084 1.1 riastrad goto out; 1085 1.1 riastrad } 1086 1.14 riastrad 1087 1.14 riastrad /* If anything changed while we were testing, restart. */ 1088 1.14 riastrad if (!dma_resv_read_valid(robj, &ticket)) 1089 1.14 riastrad goto restart; 1090 1.1 riastrad } 1091 1.15 riastrad if (shared_count) 1092 1.15 riastrad goto out; 1093 1.1 riastrad 1094 1.1 riastrad /* If there is an exclusive fence, test it. */ 1095 1.10 riastrad KASSERT(fence == NULL); 1096 1.12 riastrad if (!dma_resv_get_excl_reader(robj, &fence, &ticket)) 1097 1.12 riastrad goto restart; 1098 1.12 riastrad if (fence != NULL) { 1099 1.12 riastrad /* Test whether it is signalled. If no, stop. */ 1100 1.1 riastrad signaled &= dma_fence_is_signaled(fence); 1101 1.1 riastrad dma_fence_put(fence); 1102 1.10 riastrad fence = NULL; 1103 1.1 riastrad if (!signaled) 1104 1.1 riastrad goto out; 1105 1.1 riastrad } 1106 1.1 riastrad 1107 1.10 riastrad out: KASSERT(fence == NULL); 1108 1.10 riastrad rcu_read_unlock(); 1109 1.1 riastrad return signaled; 1110 1.1 riastrad 1111 1.1 riastrad restart: 1112 1.10 riastrad KASSERT(fence == NULL); 1113 1.1 riastrad rcu_read_unlock(); 1114 1.1 riastrad goto top; 1115 1.1 riastrad } 1116 1.1 riastrad 1117 1.1 riastrad /* 1118 1.1 riastrad * dma_resv_wait_timeout_rcu(robj, shared, intr, timeout) 1119 1.1 riastrad * 1120 1.1 riastrad * If shared is true, wait for all of the shared fences to be 1121 1.1 riastrad * signalled, or if there are none, wait for the exclusive fence 1122 1.1 riastrad * to be signalled. If shared is false, wait only for the 1123 1.1 riastrad * exclusive fence to be signalled. If timeout is zero, don't 1124 1.1 riastrad * wait, only test. 1125 1.1 riastrad * 1126 1.1 riastrad * XXX Why does this _not_ wait for the exclusive fence if shared 1127 1.1 riastrad * is true only if there are no shared fences? This makes no 1128 1.1 riastrad * sense. 1129 1.1 riastrad */ 1130 1.1 riastrad long 1131 1.1 riastrad dma_resv_wait_timeout_rcu(const struct dma_resv *robj, 1132 1.1 riastrad bool shared, bool intr, unsigned long timeout) 1133 1.1 riastrad { 1134 1.1 riastrad struct dma_resv_read_ticket ticket; 1135 1.12 riastrad const struct dma_resv_list *list; 1136 1.10 riastrad struct dma_fence *fence = NULL; 1137 1.1 riastrad uint32_t i, shared_count; 1138 1.1 riastrad long ret; 1139 1.1 riastrad 1140 1.1 riastrad if (timeout == 0) 1141 1.1 riastrad return dma_resv_test_signaled_rcu(robj, shared); 1142 1.1 riastrad 1143 1.10 riastrad top: KASSERT(fence == NULL); 1144 1.10 riastrad 1145 1.1 riastrad /* Enter an RCU read section and get a read ticket. */ 1146 1.1 riastrad rcu_read_lock(); 1147 1.1 riastrad dma_resv_read_begin(robj, &ticket); 1148 1.1 riastrad 1149 1.1 riastrad /* If shared is requested and there is a shared list, wait on it. */ 1150 1.12 riastrad if (shared) { 1151 1.12 riastrad if (!dma_resv_get_shared_reader(robj, &list, &shared_count, 1152 1.12 riastrad &ticket)) 1153 1.1 riastrad goto restart; 1154 1.12 riastrad } else { 1155 1.12 riastrad list = NULL; 1156 1.12 riastrad shared_count = 0; 1157 1.12 riastrad } 1158 1.12 riastrad if (list != NULL) { 1159 1.1 riastrad /* 1160 1.1 riastrad * For each fence, if it is going away, restart. 1161 1.1 riastrad * Otherwise, acquire a reference to it to test whether 1162 1.1 riastrad * it is signalled. Stop and wait if we find any that 1163 1.1 riastrad * is not signalled. 1164 1.1 riastrad */ 1165 1.1 riastrad for (i = 0; i < shared_count; i++) { 1166 1.10 riastrad KASSERT(fence == NULL); 1167 1.7 riastrad fence = atomic_load_relaxed(&list->shared[i]); 1168 1.10 riastrad if ((fence = dma_fence_get_rcu(fence)) == NULL) 1169 1.1 riastrad goto restart; 1170 1.1 riastrad if (!dma_fence_is_signaled(fence)) 1171 1.1 riastrad goto wait; 1172 1.1 riastrad dma_fence_put(fence); 1173 1.10 riastrad fence = NULL; 1174 1.1 riastrad } 1175 1.14 riastrad 1176 1.14 riastrad /* If anything changed while we were testing, restart. */ 1177 1.14 riastrad if (!dma_resv_read_valid(robj, &ticket)) 1178 1.14 riastrad goto restart; 1179 1.1 riastrad } 1180 1.15 riastrad if (shared_count) 1181 1.15 riastrad goto out; 1182 1.1 riastrad 1183 1.1 riastrad /* If there is an exclusive fence, test it. */ 1184 1.10 riastrad KASSERT(fence == NULL); 1185 1.12 riastrad if (!dma_resv_get_excl_reader(robj, &fence, &ticket)) 1186 1.12 riastrad goto restart; 1187 1.12 riastrad if (fence != NULL) { 1188 1.12 riastrad /* Test whether it is signalled. If no, wait. */ 1189 1.1 riastrad if (!dma_fence_is_signaled(fence)) 1190 1.1 riastrad goto wait; 1191 1.1 riastrad dma_fence_put(fence); 1192 1.10 riastrad fence = NULL; 1193 1.1 riastrad } 1194 1.1 riastrad 1195 1.15 riastrad out: /* Success! Return the number of ticks left. */ 1196 1.1 riastrad rcu_read_unlock(); 1197 1.10 riastrad KASSERT(fence == NULL); 1198 1.1 riastrad return timeout; 1199 1.1 riastrad 1200 1.1 riastrad restart: 1201 1.10 riastrad KASSERT(fence == NULL); 1202 1.1 riastrad rcu_read_unlock(); 1203 1.1 riastrad goto top; 1204 1.1 riastrad 1205 1.1 riastrad wait: 1206 1.1 riastrad /* 1207 1.5 riastrad * Exit the RCU read section, wait for it, and release the 1208 1.5 riastrad * fence when we're done. If we time out or fail, bail. 1209 1.5 riastrad * Otherwise, go back to the top. 1210 1.1 riastrad */ 1211 1.1 riastrad KASSERT(fence != NULL); 1212 1.1 riastrad rcu_read_unlock(); 1213 1.1 riastrad ret = dma_fence_wait_timeout(fence, intr, timeout); 1214 1.1 riastrad dma_fence_put(fence); 1215 1.10 riastrad fence = NULL; 1216 1.1 riastrad if (ret <= 0) 1217 1.1 riastrad return ret; 1218 1.1 riastrad KASSERT(ret <= timeout); 1219 1.1 riastrad timeout = ret; 1220 1.1 riastrad goto top; 1221 1.1 riastrad } 1222 1.1 riastrad 1223 1.1 riastrad /* 1224 1.1 riastrad * dma_resv_poll_init(rpoll, lock) 1225 1.1 riastrad * 1226 1.1 riastrad * Initialize reservation poll state. 1227 1.1 riastrad */ 1228 1.1 riastrad void 1229 1.1 riastrad dma_resv_poll_init(struct dma_resv_poll *rpoll) 1230 1.1 riastrad { 1231 1.1 riastrad 1232 1.1 riastrad mutex_init(&rpoll->rp_lock, MUTEX_DEFAULT, IPL_VM); 1233 1.1 riastrad selinit(&rpoll->rp_selq); 1234 1.1 riastrad rpoll->rp_claimed = 0; 1235 1.1 riastrad } 1236 1.1 riastrad 1237 1.1 riastrad /* 1238 1.1 riastrad * dma_resv_poll_fini(rpoll) 1239 1.1 riastrad * 1240 1.1 riastrad * Release any resource associated with reservation poll state. 1241 1.1 riastrad */ 1242 1.1 riastrad void 1243 1.1 riastrad dma_resv_poll_fini(struct dma_resv_poll *rpoll) 1244 1.1 riastrad { 1245 1.1 riastrad 1246 1.1 riastrad KASSERT(rpoll->rp_claimed == 0); 1247 1.1 riastrad seldestroy(&rpoll->rp_selq); 1248 1.1 riastrad mutex_destroy(&rpoll->rp_lock); 1249 1.1 riastrad } 1250 1.1 riastrad 1251 1.1 riastrad /* 1252 1.1 riastrad * dma_resv_poll_cb(fence, fcb) 1253 1.1 riastrad * 1254 1.1 riastrad * Callback to notify a reservation poll that a fence has 1255 1.1 riastrad * completed. Notify any waiters and allow the next poller to 1256 1.1 riastrad * claim the callback. 1257 1.1 riastrad * 1258 1.1 riastrad * If one thread is waiting for the exclusive fence only, and we 1259 1.1 riastrad * spuriously notify them about a shared fence, tough. 1260 1.1 riastrad */ 1261 1.1 riastrad static void 1262 1.1 riastrad dma_resv_poll_cb(struct dma_fence *fence, struct dma_fence_cb *fcb) 1263 1.1 riastrad { 1264 1.1 riastrad struct dma_resv_poll *rpoll = container_of(fcb, 1265 1.1 riastrad struct dma_resv_poll, rp_fcb); 1266 1.1 riastrad 1267 1.1 riastrad mutex_enter(&rpoll->rp_lock); 1268 1.1 riastrad selnotify(&rpoll->rp_selq, 0, NOTE_SUBMIT); 1269 1.1 riastrad rpoll->rp_claimed = 0; 1270 1.1 riastrad mutex_exit(&rpoll->rp_lock); 1271 1.1 riastrad } 1272 1.1 riastrad 1273 1.1 riastrad /* 1274 1.1 riastrad * dma_resv_do_poll(robj, events, rpoll) 1275 1.1 riastrad * 1276 1.1 riastrad * Poll for reservation object events using the reservation poll 1277 1.1 riastrad * state in rpoll: 1278 1.1 riastrad * 1279 1.1 riastrad * - POLLOUT wait for all fences shared and exclusive 1280 1.1 riastrad * - POLLIN wait for the exclusive fence 1281 1.1 riastrad * 1282 1.1 riastrad * Return the subset of events in events that are ready. If any 1283 1.1 riastrad * are requested but not ready, arrange to be notified with 1284 1.1 riastrad * selnotify when they are. 1285 1.1 riastrad */ 1286 1.1 riastrad int 1287 1.1 riastrad dma_resv_do_poll(const struct dma_resv *robj, int events, 1288 1.1 riastrad struct dma_resv_poll *rpoll) 1289 1.1 riastrad { 1290 1.1 riastrad struct dma_resv_read_ticket ticket; 1291 1.12 riastrad const struct dma_resv_list *list; 1292 1.10 riastrad struct dma_fence *fence = NULL; 1293 1.1 riastrad uint32_t i, shared_count; 1294 1.1 riastrad int revents; 1295 1.1 riastrad bool recorded = false; /* curlwp is on the selq */ 1296 1.1 riastrad bool claimed = false; /* we claimed the callback */ 1297 1.1 riastrad bool callback = false; /* we requested a callback */ 1298 1.1 riastrad 1299 1.1 riastrad /* 1300 1.1 riastrad * Start with the maximal set of events that could be ready. 1301 1.1 riastrad * We will eliminate the events that are definitely not ready 1302 1.1 riastrad * as we go at the same time as we add callbacks to notify us 1303 1.1 riastrad * that they may be ready. 1304 1.1 riastrad */ 1305 1.1 riastrad revents = events & (POLLIN|POLLOUT); 1306 1.1 riastrad if (revents == 0) 1307 1.1 riastrad return 0; 1308 1.1 riastrad 1309 1.10 riastrad top: KASSERT(fence == NULL); 1310 1.10 riastrad 1311 1.1 riastrad /* Enter an RCU read section and get a read ticket. */ 1312 1.1 riastrad rcu_read_lock(); 1313 1.1 riastrad dma_resv_read_begin(robj, &ticket); 1314 1.1 riastrad 1315 1.1 riastrad /* If we want to wait for all fences, get the shared list. */ 1316 1.12 riastrad if (events & POLLOUT) { 1317 1.12 riastrad if (!dma_resv_get_shared_reader(robj, &list, &shared_count, 1318 1.12 riastrad &ticket)) 1319 1.1 riastrad goto restart; 1320 1.12 riastrad } else { 1321 1.12 riastrad list = NULL; 1322 1.12 riastrad shared_count = 0; 1323 1.12 riastrad } 1324 1.12 riastrad if (list != NULL) do { 1325 1.1 riastrad /* 1326 1.1 riastrad * For each fence, if it is going away, restart. 1327 1.1 riastrad * Otherwise, acquire a reference to it to test whether 1328 1.1 riastrad * it is signalled. Stop and request a callback if we 1329 1.1 riastrad * find any that is not signalled. 1330 1.1 riastrad */ 1331 1.1 riastrad for (i = 0; i < shared_count; i++) { 1332 1.10 riastrad KASSERT(fence == NULL); 1333 1.7 riastrad fence = atomic_load_relaxed(&list->shared[i]); 1334 1.10 riastrad if ((fence = dma_fence_get_rcu(fence)) == NULL) 1335 1.1 riastrad goto restart; 1336 1.1 riastrad if (!dma_fence_is_signaled(fence)) { 1337 1.1 riastrad dma_fence_put(fence); 1338 1.10 riastrad fence = NULL; 1339 1.1 riastrad break; 1340 1.1 riastrad } 1341 1.1 riastrad dma_fence_put(fence); 1342 1.10 riastrad fence = NULL; 1343 1.1 riastrad } 1344 1.1 riastrad 1345 1.1 riastrad /* If all shared fences have been signalled, move on. */ 1346 1.1 riastrad if (i == shared_count) 1347 1.1 riastrad break; 1348 1.1 riastrad 1349 1.1 riastrad /* Put ourselves on the selq if we haven't already. */ 1350 1.1 riastrad if (!recorded) 1351 1.1 riastrad goto record; 1352 1.1 riastrad 1353 1.1 riastrad /* 1354 1.1 riastrad * If someone else claimed the callback, or we already 1355 1.1 riastrad * requested it, we're guaranteed to be notified, so 1356 1.1 riastrad * assume the event is not ready. 1357 1.1 riastrad */ 1358 1.1 riastrad if (!claimed || callback) { 1359 1.1 riastrad revents &= ~POLLOUT; 1360 1.1 riastrad break; 1361 1.1 riastrad } 1362 1.1 riastrad 1363 1.1 riastrad /* 1364 1.1 riastrad * Otherwise, find the first fence that is not 1365 1.1 riastrad * signalled, request the callback, and clear POLLOUT 1366 1.1 riastrad * from the possible ready events. If they are all 1367 1.1 riastrad * signalled, leave POLLOUT set; we will simulate the 1368 1.1 riastrad * callback later. 1369 1.1 riastrad */ 1370 1.1 riastrad for (i = 0; i < shared_count; i++) { 1371 1.10 riastrad KASSERT(fence == NULL); 1372 1.7 riastrad fence = atomic_load_relaxed(&list->shared[i]); 1373 1.10 riastrad if ((fence = dma_fence_get_rcu(fence)) == NULL) 1374 1.1 riastrad goto restart; 1375 1.1 riastrad if (!dma_fence_add_callback(fence, &rpoll->rp_fcb, 1376 1.1 riastrad dma_resv_poll_cb)) { 1377 1.1 riastrad dma_fence_put(fence); 1378 1.10 riastrad fence = NULL; 1379 1.1 riastrad revents &= ~POLLOUT; 1380 1.1 riastrad callback = true; 1381 1.1 riastrad break; 1382 1.1 riastrad } 1383 1.1 riastrad dma_fence_put(fence); 1384 1.10 riastrad fence = NULL; 1385 1.1 riastrad } 1386 1.1 riastrad } while (0); 1387 1.1 riastrad 1388 1.1 riastrad /* We always wait for at least the exclusive fence, so get it. */ 1389 1.10 riastrad KASSERT(fence == NULL); 1390 1.12 riastrad if (!dma_resv_get_excl_reader(robj, &fence, &ticket)) 1391 1.12 riastrad goto restart; 1392 1.12 riastrad if (fence != NULL) do { 1393 1.1 riastrad /* 1394 1.12 riastrad * Test whether it is signalled. If not, stop and 1395 1.12 riastrad * request a callback. 1396 1.1 riastrad */ 1397 1.16 riastrad if (dma_fence_is_signaled(fence)) 1398 1.1 riastrad break; 1399 1.1 riastrad 1400 1.1 riastrad /* Put ourselves on the selq if we haven't already. */ 1401 1.1 riastrad if (!recorded) { 1402 1.1 riastrad dma_fence_put(fence); 1403 1.10 riastrad fence = NULL; 1404 1.1 riastrad goto record; 1405 1.1 riastrad } 1406 1.1 riastrad 1407 1.1 riastrad /* 1408 1.1 riastrad * If someone else claimed the callback, or we already 1409 1.1 riastrad * requested it, we're guaranteed to be notified, so 1410 1.1 riastrad * assume the event is not ready. 1411 1.1 riastrad */ 1412 1.1 riastrad if (!claimed || callback) { 1413 1.1 riastrad revents = 0; 1414 1.1 riastrad break; 1415 1.1 riastrad } 1416 1.1 riastrad 1417 1.1 riastrad /* 1418 1.1 riastrad * Otherwise, try to request the callback, and clear 1419 1.1 riastrad * all possible ready events. If the fence has been 1420 1.1 riastrad * signalled in the interim, leave the events set; we 1421 1.1 riastrad * will simulate the callback later. 1422 1.1 riastrad */ 1423 1.1 riastrad if (!dma_fence_add_callback(fence, &rpoll->rp_fcb, 1424 1.1 riastrad dma_resv_poll_cb)) { 1425 1.1 riastrad revents = 0; 1426 1.1 riastrad callback = true; 1427 1.1 riastrad break; 1428 1.1 riastrad } 1429 1.16 riastrad } while (0); 1430 1.16 riastrad if (fence != NULL) { 1431 1.1 riastrad dma_fence_put(fence); 1432 1.10 riastrad fence = NULL; 1433 1.16 riastrad } 1434 1.1 riastrad 1435 1.1 riastrad /* All done reading the fences. */ 1436 1.1 riastrad rcu_read_unlock(); 1437 1.1 riastrad 1438 1.1 riastrad if (claimed && !callback) { 1439 1.1 riastrad /* 1440 1.1 riastrad * We claimed the callback but we didn't actually 1441 1.1 riastrad * request it because a fence was signalled while we 1442 1.1 riastrad * were claiming it. Call it ourselves now. The 1443 1.1 riastrad * callback doesn't use the fence nor rely on holding 1444 1.1 riastrad * any of the fence locks, so this is safe. 1445 1.1 riastrad */ 1446 1.1 riastrad dma_resv_poll_cb(NULL, &rpoll->rp_fcb); 1447 1.1 riastrad } 1448 1.1 riastrad return revents; 1449 1.1 riastrad 1450 1.1 riastrad restart: 1451 1.10 riastrad KASSERT(fence == NULL); 1452 1.1 riastrad rcu_read_unlock(); 1453 1.1 riastrad goto top; 1454 1.1 riastrad 1455 1.1 riastrad record: 1456 1.10 riastrad KASSERT(fence == NULL); 1457 1.1 riastrad rcu_read_unlock(); 1458 1.1 riastrad mutex_enter(&rpoll->rp_lock); 1459 1.1 riastrad selrecord(curlwp, &rpoll->rp_selq); 1460 1.1 riastrad if (!rpoll->rp_claimed) 1461 1.1 riastrad claimed = rpoll->rp_claimed = true; 1462 1.1 riastrad mutex_exit(&rpoll->rp_lock); 1463 1.1 riastrad recorded = true; 1464 1.1 riastrad goto top; 1465 1.1 riastrad } 1466 1.1 riastrad 1467 1.1 riastrad /* 1468 1.1 riastrad * dma_resv_kqfilter(robj, kn, rpoll) 1469 1.1 riastrad * 1470 1.1 riastrad * Kqueue filter for reservation objects. Currently not 1471 1.1 riastrad * implemented because the logic to implement it is nontrivial, 1472 1.1 riastrad * and userland will presumably never use it, so it would be 1473 1.1 riastrad * dangerous to add never-tested complex code paths to the kernel. 1474 1.1 riastrad */ 1475 1.1 riastrad int 1476 1.1 riastrad dma_resv_kqfilter(const struct dma_resv *robj, 1477 1.1 riastrad struct knote *kn, struct dma_resv_poll *rpoll) 1478 1.1 riastrad { 1479 1.1 riastrad 1480 1.1 riastrad return EINVAL; 1481 1.1 riastrad } 1482