HomeSort by: relevance | last modified time | path
    Searched defs:vq (Results 1 - 25 of 43) sorted by relevancy

1 2

  /src/sys/dev/pci/
viornd.c 88 struct virtqueue *vq = &sc->sc_vq; local
103 if (virtio_enqueue_prep(vsc, vq, &slot)) {
106 if (virtio_enqueue_reserve(vsc, vq, slot, 1)) {
109 virtio_enqueue(vsc, vq, slot, sc->sc_dmamap, 0);
110 virtio_enqueue_commit(vsc, vq, slot, 1);
221 viornd_vq_done(struct virtqueue *vq)
223 struct virtio_softc *vsc = vq->vq_owner;
229 if (virtio_dequeue(vsc, vq, &slot, &len) != 0) {
252 virtio_dequeue_commit(vsc, vq, slot);
viomb.c 279 struct virtqueue *vq = &sc->sc_vq[VQ_INFLATE]; local
305 if (virtio_enqueue_prep(vsc, vq, &slot) != 0) {
311 if (virtio_enqueue_reserve(vsc, vq, slot, 1)) {
319 virtio_enqueue(vsc, vq, slot, b->bl_dmamap, true);
320 virtio_enqueue_commit(vsc, vq, slot, true);
327 inflateq_done(struct virtqueue *vq)
329 struct virtio_softc *vsc = vq->vq_owner;
344 struct virtqueue *vq = &sc->sc_vq[VQ_INFLATE]; local
350 r = virtio_dequeue(vsc, vq, &slot, NULL);
356 virtio_dequeue_commit(vsc, vq, slot)
390 struct virtqueue *vq = &sc->sc_vq[VQ_DEFLATE]; local
461 struct virtqueue *vq = &sc->sc_vq[VQ_DEFLATE]; local
    [all...]
vio9p.c 221 struct virtqueue *vq = &sc->sc_vq[0]; local
263 error = virtio_dequeue(vsc, vq, &slot, &len);
276 virtio_dequeue_commit(vsc, vq, slot);
317 struct virtqueue *vq = &sc->sc_vq[0]; local
356 error = virtio_enqueue_prep(vsc, vq, &slot);
363 error = virtio_enqueue_reserve(vsc, vq, slot,
374 virtio_enqueue(vsc, vq, slot, sc->sc_dmamap_tx, true);
378 virtio_enqueue(vsc, vq, slot, sc->sc_dmamap_rx, false);
379 virtio_enqueue_commit(vsc, vq, slot, true);
617 vio9p_request_done(struct virtqueue *vq)
    [all...]
vioscsi.c 274 struct virtqueue *vq = &sc->sc_vqs[VIOSCSI_VQ_REQUEST]; local
381 virtio_enqueue_abort(vsc, vq, slot);
390 error = virtio_enqueue_reserve(vsc, vq, slot, nsegs);
413 virtio_enqueue_p(vsc, vq, slot, vr->vr_control,
417 virtio_enqueue(vsc, vq, slot, vr->vr_data, 1);
418 virtio_enqueue_p(vsc, vq, slot, vr->vr_control,
422 virtio_enqueue(vsc, vq, slot, vr->vr_data, 0);
424 virtio_enqueue_commit(vsc, vq, slot, 1);
452 struct vioscsi_req *vr, struct virtqueue *vq, int slot)
497 virtio_dequeue_commit(vsc, vq, slot)
556 struct virtqueue *vq = &sc->sc_vqs[VIOSCSI_VQ_REQUEST]; local
    [all...]
viogpu.c 78 static int viogpu_vq_done(struct virtqueue *vq);
483 struct virtqueue *vq = &sc->sc_vqs[VQCTRL]; local
502 error = virtio_enqueue_prep(vsc, vq, &slot);
504 panic("%s: control vq busy", device_xname(sc->sc_dev));
506 error = virtio_enqueue_reserve(vsc, vq, slot,
509 panic("%s: control vq busy", device_xname(sc->sc_dev));
513 virtio_enqueue_p(vsc, vq, slot, sc->sc_dma_map, 0, cmd_size, true);
517 virtio_enqueue_p(vsc, vq, slot, sc->sc_dma_map, cmd_size, ret_size,
520 virtio_enqueue_commit(vsc, vq, slot, true);
530 viogpu_vq_done(struct virtqueue *vq)
    [all...]
  /src/crypto/external/bsd/heimdal/dist/lib/hcrypto/
rsa-gmp.c 85 mpz_t vp, vq, u; local
86 mpz_init(vp); mpz_init(vq); mpz_init(u);
88 /* vq = c ^ (d mod (q - 1)) mod q */
92 mpz_fdiv_r(vq, in, q);
93 mpz_powm(vq, vq, dmq1, q);
96 /* u = (vp - vq)C2 mod p. */
97 mpz_sub(u, vp, vq);
105 /* c ^ d mod n = vq + u q */
107 mpz_add(out, u, vq);
    [all...]
rsa-ltm.c 109 mp_int vp, vq, u; local
111 mp_init_multi(&vp, &vq, &u, NULL);
113 /* vq = c ^ (d mod (q - 1)) mod q */
118 mp_exptmod(&u, dmq1, q, &vq);
121 /* u = (vp - vq)C2 mod p. */
122 mp_sub(&vp, &vq, &u);
128 /* c ^ d mod n = vq + u q */
130 mp_add(&u, &vq, out);
132 mp_clear_multi(&vp, &vq, &u, NULL);
rsa-tfm.c 67 fp_int vp, vq, u; local
69 fp_init_multi(&vp, &vq, &u, NULL);
71 /* vq = c ^ (d mod (q - 1)) mod q */
76 fp_exptmod(&u, dmq1, q, &vq);
79 /* u = (vp - vq)C2 mod p. */
80 fp_sub(&vp, &vq, &u);
86 /* c ^ d mod n = vq + u q */
88 fp_add(&u, &vq, out);
90 fp_zero_multi(&vp, &vq, &u, NULL);
  /src/external/bsd/am-utils/dist/fsinfo/
fsi_util.c 382 qelem *vq = (qelem *) v; local
389 ins_que(vq, q->q_back);
395 qelem *vq = (qelem *) v; local
402 ins_que(vq, q->q_back);
  /src/external/gpl3/gdb/dist/gdb/arch/
aarch64.h 29 /* A non zero VQ value indicates both the presence of SVE and the
33 The maximum value for VQ is 16 (5 bits). */
34 uint64_t vq = 0; member in struct:aarch64_features
58 return lhs.vq == rhs.vq
75 h = features.vq;
179 VQ : Vector Quotient.
190 #define sve_vl_from_vq(vq) ((vq) * 0x10)
193 #define sve_vg_from_vq(vq) (sve_vg_from_vl (sve_vl_from_vq (vq))
    [all...]
  /src/external/gpl3/gdb/dist/gdb/nat/
aarch64-scalable-linux-ptrace.c 237 /* Ptrace gives the vector length in bytes. Convert it to VQ, the number of
238 128bit chunks in a Z register. We use VQ because 128 bits is the minimum
240 uint64_t vq = sve_vq_from_vl (header.vl); local
248 return vq;
254 aarch64_sve_set_vq (int tid, uint64_t vq)
272 header.vl = sve_vl_from_vq (vq);
299 uint64_t vq = aarch64_sve_get_vq (tid); local
302 if (vq == 0)
305 reg_vg = sve_vg_from_vq (vq);
322 uint64_t vq = sve_vq_from_vl (header.vl)
413 uint64_t vq = aarch64_sve_get_vq (tid); local
619 uint64_t vq = sve_vq_from_vl (header->vl); local
719 uint64_t vq = sve_vq_from_vl (header->vl); local
    [all...]
  /src/external/gpl3/gdb.old/dist/gdb/arch/
aarch64.h 29 /* A non zero VQ value indicates both the presence of SVE and the
33 The maximum value for VQ is 16 (5 bits). */
34 uint64_t vq = 0; member in struct:aarch64_features
58 return lhs.vq == rhs.vq
75 h = features.vq;
179 VQ : Vector Quotient.
190 #define sve_vl_from_vq(vq) ((vq) * 0x10)
193 #define sve_vg_from_vq(vq) (sve_vg_from_vl (sve_vl_from_vq (vq))
    [all...]
  /src/external/gpl3/gdb.old/dist/gdb/nat/
aarch64-scalable-linux-ptrace.c 237 /* Ptrace gives the vector length in bytes. Convert it to VQ, the number of
238 128bit chunks in a Z register. We use VQ because 128 bits is the minimum
240 uint64_t vq = sve_vq_from_vl (header.vl); local
248 return vq;
254 aarch64_sve_set_vq (int tid, uint64_t vq)
272 header.vl = sve_vl_from_vq (vq);
299 uint64_t vq = aarch64_sve_get_vq (tid); local
302 if (vq == 0)
305 reg_vg = sve_vg_from_vq (vq);
322 uint64_t vq = sve_vq_from_vl (header.vl)
413 uint64_t vq = aarch64_sve_get_vq (tid); local
619 uint64_t vq = sve_vq_from_vl (header->vl); local
719 uint64_t vq = sve_vq_from_vl (header->vl); local
    [all...]
  /src/sys/dev/virtio/
virtio_mmio.c 211 struct virtqueue *vq; local
220 vq = &vsc->sc_vqs[idx];
221 KASSERT(vq->vq_index == idx);
228 addr + vq->vq_availoffset);
230 addr + vq->vq_usedoffset);
  /src/external/gpl3/gdb/dist/gdb/
aarch64-linux-nat.c 881 /* SVE/SSVE check. Reading VQ may return either the regular vector length
884 features.vq = aarch64_sve_get_vq (tid);
994 /* Only return the inferior's gdbarch if both vq and svq match the ones in
998 uint64_t vq = aarch64_sve_get_vq (ptid.lwp ());
1000 if (vq == tdep->vq && svq == tdep->sme_svq)
1005 new one) by using a target description that corresponds to the new vq/svq
1010 features.vq = vq;
996 uint64_t vq = aarch64_sve_get_vq (ptid.lwp ()); local
aarch64-tdep.h 103 /* The VQ value for SVE targets, or zero if SVE is not supported. */
104 uint64_t vq = 0; member in struct:aarch64_gdbarch_tdep
109 return vq != 0;
  /src/external/gpl3/gdb.old/dist/gdb/
aarch64-linux-nat.c 883 /* SVE/SSVE check. Reading VQ may return either the regular vector length
886 features.vq = aarch64_sve_get_vq (tid);
996 /* Only return the inferior's gdbarch if both vq and svq match the ones in
1000 uint64_t vq = aarch64_sve_get_vq (ptid.lwp ());
1002 if (vq == tdep->vq && svq == tdep->sme_svq)
1007 new one) by using a target description that corresponds to the new vq/svq
1012 features.vq = vq;
998 uint64_t vq = aarch64_sve_get_vq (ptid.lwp ()); local
aarch64-tdep.h 103 /* The VQ value for SVE targets, or zero if SVE is not supported. */
104 uint64_t vq = 0; member in struct:aarch64_gdbarch_tdep
109 return vq != 0;
  /src/bin/ksh/
c_ksh.c 1309 struct tbl *vq, *voptarg; local
1388 vq = global(var);
1390 if (!setstr(vq, buf, KSH_RETURN_ERROR))
expr.c 592 struct tbl *vq; local
599 vq = tempvar();
600 if (setint_v(vq, vp) == NULL) {
605 v_evaluate(vq, str_val(vp), KSH_UNWIND_ERROR);
609 return vq;
var.c 72 struct tbl *vp, **vpp = l->vars.tbls, *vq; local
78 if ((vq = global(vp->name))->flag & ISSET)
79 setspec(vq);
81 unsetspec(vq);
278 struct tbl *vq = (struct tbl *) 0; local
280 while ((ll = ll->next) && !(vq = mytsearch(&ll->vars, n, h)))
282 if (vq) {
283 vp->flag |= vq->flag & (EXPORT|INTEGER|RDONLY
286 if (vq->flag & INTEGER)
287 vp->type = vq->type
1238 struct tbl *vp, *vq; local
    [all...]
  /src/sys/external/bsd/drm2/dist/drm/virtio/
virtgpu_drv.h 147 struct virtqueue *vq; member in struct:virtio_gpu_queue
330 void virtio_gpu_ctrl_ack(struct virtqueue *vq);
331 void virtio_gpu_cursor_ack(struct virtqueue *vq);
332 void virtio_gpu_fence_ack(struct virtqueue *vq);
virtgpu_vq.c 59 void virtio_gpu_ctrl_ack(struct virtqueue *vq)
61 struct drm_device *dev = vq->vdev->priv;
67 void virtio_gpu_cursor_ack(struct virtqueue *vq)
69 struct drm_device *dev = vq->vdev->priv;
178 static void reclaim_vbufs(struct virtqueue *vq, struct list_head *reclaim_list)
184 while ((vbuf = virtqueue_get_buf(vq, &len))) {
205 virtqueue_disable_cb(vgdev->ctrlq.vq);
206 reclaim_vbufs(vgdev->ctrlq.vq, &reclaim_list);
208 } while (!virtqueue_enable_cb(vgdev->ctrlq.vq));
214 trace_virtio_gpu_cmd_response(vgdev->ctrlq.vq, resp)
321 struct virtqueue *vq = vgdev->ctrlq.vq; variable in typeref:struct:virtqueue
366 struct virtqueue *vq = vgdev->ctrlq.vq; local
449 struct virtqueue *vq = vgdev->cursorq.vq; local
    [all...]
  /src/sys/fs/nfs/client/
nfs_clvfsops.c 1792 struct vfsquery vq; local
1795 bzero(&vq, sizeof(vq));
1819 vq.vq_flags |= VQ_NOTRESP;
1824 vq.vq_flags |= VQ_NOTRESPLOCK;
1826 error = SYSCTL_OUT(req, &vq, sizeof(vq));
  /src/sys/miscfs/specfs/
spec_vnops.c 497 vnode_t *vp, *vq; local
500 for (i = 0, vq = NULL; i < SPECHSZ && vq == NULL; i++) {
504 vq = vp->v_specnode->sn_dev->sd_bdevvp;
505 if (vq != NULL &&
506 vq->v_specnode->sn_dev->sd_mountpoint == mp)
508 vq = NULL;
511 if (vq == NULL) {
515 mutex_enter(vq->v_interlock);
517 error = vcache_vget(vq);
    [all...]

Completed in 49 milliseconds

1 2