| /src/sys/dev/pci/ |
| viornd.c | 88 struct virtqueue *vq = &sc->sc_vq; local 103 if (virtio_enqueue_prep(vsc, vq, &slot)) { 106 if (virtio_enqueue_reserve(vsc, vq, slot, 1)) { 109 virtio_enqueue(vsc, vq, slot, sc->sc_dmamap, 0); 110 virtio_enqueue_commit(vsc, vq, slot, 1); 221 viornd_vq_done(struct virtqueue *vq) 223 struct virtio_softc *vsc = vq->vq_owner; 229 if (virtio_dequeue(vsc, vq, &slot, &len) != 0) { 252 virtio_dequeue_commit(vsc, vq, slot);
|
| viomb.c | 279 struct virtqueue *vq = &sc->sc_vq[VQ_INFLATE]; local 305 if (virtio_enqueue_prep(vsc, vq, &slot) != 0) { 311 if (virtio_enqueue_reserve(vsc, vq, slot, 1)) { 319 virtio_enqueue(vsc, vq, slot, b->bl_dmamap, true); 320 virtio_enqueue_commit(vsc, vq, slot, true); 327 inflateq_done(struct virtqueue *vq) 329 struct virtio_softc *vsc = vq->vq_owner; 344 struct virtqueue *vq = &sc->sc_vq[VQ_INFLATE]; local 350 r = virtio_dequeue(vsc, vq, &slot, NULL); 356 virtio_dequeue_commit(vsc, vq, slot) 390 struct virtqueue *vq = &sc->sc_vq[VQ_DEFLATE]; local 461 struct virtqueue *vq = &sc->sc_vq[VQ_DEFLATE]; local [all...] |
| vio9p.c | 221 struct virtqueue *vq = &sc->sc_vq[0]; local 263 error = virtio_dequeue(vsc, vq, &slot, &len); 276 virtio_dequeue_commit(vsc, vq, slot); 317 struct virtqueue *vq = &sc->sc_vq[0]; local 356 error = virtio_enqueue_prep(vsc, vq, &slot); 363 error = virtio_enqueue_reserve(vsc, vq, slot, 374 virtio_enqueue(vsc, vq, slot, sc->sc_dmamap_tx, true); 378 virtio_enqueue(vsc, vq, slot, sc->sc_dmamap_rx, false); 379 virtio_enqueue_commit(vsc, vq, slot, true); 617 vio9p_request_done(struct virtqueue *vq) [all...] |
| vioscsi.c | 274 struct virtqueue *vq = &sc->sc_vqs[VIOSCSI_VQ_REQUEST]; local 381 virtio_enqueue_abort(vsc, vq, slot); 390 error = virtio_enqueue_reserve(vsc, vq, slot, nsegs); 413 virtio_enqueue_p(vsc, vq, slot, vr->vr_control, 417 virtio_enqueue(vsc, vq, slot, vr->vr_data, 1); 418 virtio_enqueue_p(vsc, vq, slot, vr->vr_control, 422 virtio_enqueue(vsc, vq, slot, vr->vr_data, 0); 424 virtio_enqueue_commit(vsc, vq, slot, 1); 452 struct vioscsi_req *vr, struct virtqueue *vq, int slot) 497 virtio_dequeue_commit(vsc, vq, slot) 556 struct virtqueue *vq = &sc->sc_vqs[VIOSCSI_VQ_REQUEST]; local [all...] |
| viogpu.c | 78 static int viogpu_vq_done(struct virtqueue *vq); 483 struct virtqueue *vq = &sc->sc_vqs[VQCTRL]; local 502 error = virtio_enqueue_prep(vsc, vq, &slot); 504 panic("%s: control vq busy", device_xname(sc->sc_dev)); 506 error = virtio_enqueue_reserve(vsc, vq, slot, 509 panic("%s: control vq busy", device_xname(sc->sc_dev)); 513 virtio_enqueue_p(vsc, vq, slot, sc->sc_dma_map, 0, cmd_size, true); 517 virtio_enqueue_p(vsc, vq, slot, sc->sc_dma_map, cmd_size, ret_size, 520 virtio_enqueue_commit(vsc, vq, slot, true); 530 viogpu_vq_done(struct virtqueue *vq) [all...] |
| /src/crypto/external/bsd/heimdal/dist/lib/hcrypto/ |
| rsa-gmp.c | 85 mpz_t vp, vq, u; local 86 mpz_init(vp); mpz_init(vq); mpz_init(u); 88 /* vq = c ^ (d mod (q - 1)) mod q */ 92 mpz_fdiv_r(vq, in, q); 93 mpz_powm(vq, vq, dmq1, q); 96 /* u = (vp - vq)C2 mod p. */ 97 mpz_sub(u, vp, vq); 105 /* c ^ d mod n = vq + u q */ 107 mpz_add(out, u, vq); [all...] |
| rsa-ltm.c | 109 mp_int vp, vq, u; local 111 mp_init_multi(&vp, &vq, &u, NULL); 113 /* vq = c ^ (d mod (q - 1)) mod q */ 118 mp_exptmod(&u, dmq1, q, &vq); 121 /* u = (vp - vq)C2 mod p. */ 122 mp_sub(&vp, &vq, &u); 128 /* c ^ d mod n = vq + u q */ 130 mp_add(&u, &vq, out); 132 mp_clear_multi(&vp, &vq, &u, NULL);
|
| rsa-tfm.c | 67 fp_int vp, vq, u; local 69 fp_init_multi(&vp, &vq, &u, NULL); 71 /* vq = c ^ (d mod (q - 1)) mod q */ 76 fp_exptmod(&u, dmq1, q, &vq); 79 /* u = (vp - vq)C2 mod p. */ 80 fp_sub(&vp, &vq, &u); 86 /* c ^ d mod n = vq + u q */ 88 fp_add(&u, &vq, out); 90 fp_zero_multi(&vp, &vq, &u, NULL);
|
| /src/external/bsd/am-utils/dist/fsinfo/ |
| fsi_util.c | 382 qelem *vq = (qelem *) v; local 389 ins_que(vq, q->q_back); 395 qelem *vq = (qelem *) v; local 402 ins_que(vq, q->q_back);
|
| /src/external/gpl3/gdb/dist/gdb/arch/ |
| aarch64.h | 29 /* A non zero VQ value indicates both the presence of SVE and the 33 The maximum value for VQ is 16 (5 bits). */ 34 uint64_t vq = 0; member in struct:aarch64_features 58 return lhs.vq == rhs.vq 75 h = features.vq; 179 VQ : Vector Quotient. 190 #define sve_vl_from_vq(vq) ((vq) * 0x10) 193 #define sve_vg_from_vq(vq) (sve_vg_from_vl (sve_vl_from_vq (vq)) [all...] |
| /src/external/gpl3/gdb/dist/gdb/nat/ |
| aarch64-scalable-linux-ptrace.c | 237 /* Ptrace gives the vector length in bytes. Convert it to VQ, the number of 238 128bit chunks in a Z register. We use VQ because 128 bits is the minimum 240 uint64_t vq = sve_vq_from_vl (header.vl); local 248 return vq; 254 aarch64_sve_set_vq (int tid, uint64_t vq) 272 header.vl = sve_vl_from_vq (vq); 299 uint64_t vq = aarch64_sve_get_vq (tid); local 302 if (vq == 0) 305 reg_vg = sve_vg_from_vq (vq); 322 uint64_t vq = sve_vq_from_vl (header.vl) 413 uint64_t vq = aarch64_sve_get_vq (tid); local 619 uint64_t vq = sve_vq_from_vl (header->vl); local 719 uint64_t vq = sve_vq_from_vl (header->vl); local [all...] |
| /src/external/gpl3/gdb.old/dist/gdb/arch/ |
| aarch64.h | 29 /* A non zero VQ value indicates both the presence of SVE and the 33 The maximum value for VQ is 16 (5 bits). */ 34 uint64_t vq = 0; member in struct:aarch64_features 58 return lhs.vq == rhs.vq 75 h = features.vq; 179 VQ : Vector Quotient. 190 #define sve_vl_from_vq(vq) ((vq) * 0x10) 193 #define sve_vg_from_vq(vq) (sve_vg_from_vl (sve_vl_from_vq (vq)) [all...] |
| /src/external/gpl3/gdb.old/dist/gdb/nat/ |
| aarch64-scalable-linux-ptrace.c | 237 /* Ptrace gives the vector length in bytes. Convert it to VQ, the number of 238 128bit chunks in a Z register. We use VQ because 128 bits is the minimum 240 uint64_t vq = sve_vq_from_vl (header.vl); local 248 return vq; 254 aarch64_sve_set_vq (int tid, uint64_t vq) 272 header.vl = sve_vl_from_vq (vq); 299 uint64_t vq = aarch64_sve_get_vq (tid); local 302 if (vq == 0) 305 reg_vg = sve_vg_from_vq (vq); 322 uint64_t vq = sve_vq_from_vl (header.vl) 413 uint64_t vq = aarch64_sve_get_vq (tid); local 619 uint64_t vq = sve_vq_from_vl (header->vl); local 719 uint64_t vq = sve_vq_from_vl (header->vl); local [all...] |
| /src/sys/dev/virtio/ |
| virtio_mmio.c | 211 struct virtqueue *vq; local 220 vq = &vsc->sc_vqs[idx]; 221 KASSERT(vq->vq_index == idx); 228 addr + vq->vq_availoffset); 230 addr + vq->vq_usedoffset);
|
| /src/external/gpl3/gdb/dist/gdb/ |
| aarch64-linux-nat.c | 881 /* SVE/SSVE check. Reading VQ may return either the regular vector length 884 features.vq = aarch64_sve_get_vq (tid); 994 /* Only return the inferior's gdbarch if both vq and svq match the ones in 998 uint64_t vq = aarch64_sve_get_vq (ptid.lwp ()); 1000 if (vq == tdep->vq && svq == tdep->sme_svq) 1005 new one) by using a target description that corresponds to the new vq/svq 1010 features.vq = vq; 996 uint64_t vq = aarch64_sve_get_vq (ptid.lwp ()); local
|
| aarch64-tdep.h | 103 /* The VQ value for SVE targets, or zero if SVE is not supported. */ 104 uint64_t vq = 0; member in struct:aarch64_gdbarch_tdep 109 return vq != 0;
|
| /src/external/gpl3/gdb.old/dist/gdb/ |
| aarch64-linux-nat.c | 883 /* SVE/SSVE check. Reading VQ may return either the regular vector length 886 features.vq = aarch64_sve_get_vq (tid); 996 /* Only return the inferior's gdbarch if both vq and svq match the ones in 1000 uint64_t vq = aarch64_sve_get_vq (ptid.lwp ()); 1002 if (vq == tdep->vq && svq == tdep->sme_svq) 1007 new one) by using a target description that corresponds to the new vq/svq 1012 features.vq = vq; 998 uint64_t vq = aarch64_sve_get_vq (ptid.lwp ()); local
|
| aarch64-tdep.h | 103 /* The VQ value for SVE targets, or zero if SVE is not supported. */ 104 uint64_t vq = 0; member in struct:aarch64_gdbarch_tdep 109 return vq != 0;
|
| /src/bin/ksh/ |
| c_ksh.c | 1309 struct tbl *vq, *voptarg; local 1388 vq = global(var); 1390 if (!setstr(vq, buf, KSH_RETURN_ERROR))
|
| expr.c | 592 struct tbl *vq; local 599 vq = tempvar(); 600 if (setint_v(vq, vp) == NULL) { 605 v_evaluate(vq, str_val(vp), KSH_UNWIND_ERROR); 609 return vq;
|
| var.c | 72 struct tbl *vp, **vpp = l->vars.tbls, *vq; local 78 if ((vq = global(vp->name))->flag & ISSET) 79 setspec(vq); 81 unsetspec(vq); 278 struct tbl *vq = (struct tbl *) 0; local 280 while ((ll = ll->next) && !(vq = mytsearch(&ll->vars, n, h))) 282 if (vq) { 283 vp->flag |= vq->flag & (EXPORT|INTEGER|RDONLY 286 if (vq->flag & INTEGER) 287 vp->type = vq->type 1238 struct tbl *vp, *vq; local [all...] |
| /src/sys/external/bsd/drm2/dist/drm/virtio/ |
| virtgpu_drv.h | 147 struct virtqueue *vq; member in struct:virtio_gpu_queue 330 void virtio_gpu_ctrl_ack(struct virtqueue *vq); 331 void virtio_gpu_cursor_ack(struct virtqueue *vq); 332 void virtio_gpu_fence_ack(struct virtqueue *vq);
|
| virtgpu_vq.c | 59 void virtio_gpu_ctrl_ack(struct virtqueue *vq) 61 struct drm_device *dev = vq->vdev->priv; 67 void virtio_gpu_cursor_ack(struct virtqueue *vq) 69 struct drm_device *dev = vq->vdev->priv; 178 static void reclaim_vbufs(struct virtqueue *vq, struct list_head *reclaim_list) 184 while ((vbuf = virtqueue_get_buf(vq, &len))) { 205 virtqueue_disable_cb(vgdev->ctrlq.vq); 206 reclaim_vbufs(vgdev->ctrlq.vq, &reclaim_list); 208 } while (!virtqueue_enable_cb(vgdev->ctrlq.vq)); 214 trace_virtio_gpu_cmd_response(vgdev->ctrlq.vq, resp) 321 struct virtqueue *vq = vgdev->ctrlq.vq; variable in typeref:struct:virtqueue 366 struct virtqueue *vq = vgdev->ctrlq.vq; local 449 struct virtqueue *vq = vgdev->cursorq.vq; local [all...] |
| /src/sys/fs/nfs/client/ |
| nfs_clvfsops.c | 1792 struct vfsquery vq; local 1795 bzero(&vq, sizeof(vq)); 1819 vq.vq_flags |= VQ_NOTRESP; 1824 vq.vq_flags |= VQ_NOTRESPLOCK; 1826 error = SYSCTL_OUT(req, &vq, sizeof(vq));
|
| /src/sys/miscfs/specfs/ |
| spec_vnops.c | 497 vnode_t *vp, *vq; local 500 for (i = 0, vq = NULL; i < SPECHSZ && vq == NULL; i++) { 504 vq = vp->v_specnode->sn_dev->sd_bdevvp; 505 if (vq != NULL && 506 vq->v_specnode->sn_dev->sd_mountpoint == mp) 508 vq = NULL; 511 if (vq == NULL) { 515 mutex_enter(vq->v_interlock); 517 error = vcache_vget(vq); [all...] |