Home | History | Annotate | Line # | Download | only in qat
      1  1.1  hikaru /*	$NetBSD: qat_hw15.c,v 1.1 2019/11/20 09:37:46 hikaru Exp $	*/
      2  1.1  hikaru 
      3  1.1  hikaru /*
      4  1.1  hikaru  * Copyright (c) 2019 Internet Initiative Japan, Inc.
      5  1.1  hikaru  * All rights reserved.
      6  1.1  hikaru  *
      7  1.1  hikaru  * Redistribution and use in source and binary forms, with or without
      8  1.1  hikaru  * modification, are permitted provided that the following conditions
      9  1.1  hikaru  * are met:
     10  1.1  hikaru  * 1. Redistributions of source code must retain the above copyright
     11  1.1  hikaru  *    notice, this list of conditions and the following disclaimer.
     12  1.1  hikaru  * 2. Redistributions in binary form must reproduce the above copyright
     13  1.1  hikaru  *    notice, this list of conditions and the following disclaimer in the
     14  1.1  hikaru  *    documentation and/or other materials provided with the distribution.
     15  1.1  hikaru  *
     16  1.1  hikaru  * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
     17  1.1  hikaru  * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
     18  1.1  hikaru  * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
     19  1.1  hikaru  * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
     20  1.1  hikaru  * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
     21  1.1  hikaru  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
     22  1.1  hikaru  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
     23  1.1  hikaru  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
     24  1.1  hikaru  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
     25  1.1  hikaru  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
     26  1.1  hikaru  * POSSIBILITY OF SUCH DAMAGE.
     27  1.1  hikaru  */
     28  1.1  hikaru 
     29  1.1  hikaru /*
     30  1.1  hikaru  *   Copyright(c) 2007-2013 Intel Corporation. All rights reserved.
     31  1.1  hikaru  *
     32  1.1  hikaru  *   Redistribution and use in source and binary forms, with or without
     33  1.1  hikaru  *   modification, are permitted provided that the following conditions
     34  1.1  hikaru  *   are met:
     35  1.1  hikaru  *
     36  1.1  hikaru  *     * Redistributions of source code must retain the above copyright
     37  1.1  hikaru  *       notice, this list of conditions and the following disclaimer.
     38  1.1  hikaru  *     * Redistributions in binary form must reproduce the above copyright
     39  1.1  hikaru  *       notice, this list of conditions and the following disclaimer in
     40  1.1  hikaru  *       the documentation and/or other materials provided with the
     41  1.1  hikaru  *       distribution.
     42  1.1  hikaru  *     * Neither the name of Intel Corporation nor the names of its
     43  1.1  hikaru  *       contributors may be used to endorse or promote products derived
     44  1.1  hikaru  *       from this software without specific prior written permission.
     45  1.1  hikaru  *
     46  1.1  hikaru  *   THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
     47  1.1  hikaru  *   "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
     48  1.1  hikaru  *   LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
     49  1.1  hikaru  *   A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
     50  1.1  hikaru  *   OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
     51  1.1  hikaru  *   SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
     52  1.1  hikaru  *   LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
     53  1.1  hikaru  *   DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
     54  1.1  hikaru  *   THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
     55  1.1  hikaru  *   (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
     56  1.1  hikaru  *   OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
     57  1.1  hikaru  */
     58  1.1  hikaru 
     59  1.1  hikaru #include <sys/cdefs.h>
     60  1.1  hikaru __KERNEL_RCSID(0, "$NetBSD: qat_hw15.c,v 1.1 2019/11/20 09:37:46 hikaru Exp $");
     61  1.1  hikaru 
     62  1.1  hikaru #include <sys/param.h>
     63  1.1  hikaru #include <sys/systm.h>
     64  1.1  hikaru #include <sys/proc.h>
     65  1.1  hikaru 
     66  1.1  hikaru #include <opencrypto/xform.h>
     67  1.1  hikaru 
     68  1.1  hikaru /* XXX same as sys/arch/x86/x86/via_padlock.c */
     69  1.1  hikaru #include <opencrypto/cryptosoft_xform.c>
     70  1.1  hikaru 
     71  1.1  hikaru #include <dev/pci/pcireg.h>
     72  1.1  hikaru #include <dev/pci/pcivar.h>
     73  1.1  hikaru 
     74  1.1  hikaru #include "qatreg.h"
     75  1.1  hikaru #include "qat_hw15reg.h"
     76  1.1  hikaru #include "qatvar.h"
     77  1.1  hikaru #include "qat_hw15var.h"
     78  1.1  hikaru 
     79  1.1  hikaru int		qat_adm_ring_init_ring_table(struct qat_softc *);
     80  1.1  hikaru void		qat_adm_ring_build_slice_mask(uint16_t *, uint32_t, uint32_t);
     81  1.1  hikaru void		qat_adm_ring_build_shram_mask(uint64_t *, uint32_t, uint32_t);
     82  1.1  hikaru int		qat_adm_ring_build_ring_table(struct qat_softc *, uint32_t);
     83  1.1  hikaru int		qat_adm_ring_build_init_msg(struct qat_softc *,
     84  1.1  hikaru 		    struct fw_init_req *, enum fw_init_cmd_id, uint32_t,
     85  1.1  hikaru 		    struct qat_accel_init_cb *);
     86  1.1  hikaru int		qat_adm_ring_send_init_msg_sync(struct qat_softc *,
     87  1.1  hikaru 		    enum fw_init_cmd_id, uint32_t);
     88  1.1  hikaru int		qat_adm_ring_send_init_msg(struct qat_softc *,
     89  1.1  hikaru 		    enum fw_init_cmd_id);
     90  1.1  hikaru int		qat_adm_ring_intr(struct qat_softc *, void *, void *);
     91  1.1  hikaru 
     92  1.1  hikaru uint32_t	qat_crypto_setup_cipher_desc(struct qat_session *,
     93  1.1  hikaru 		    struct qat_crypto_desc *desc, struct cryptoini *,
     94  1.1  hikaru 		    struct fw_cipher_hdr *, uint8_t *, uint32_t, enum fw_slice);
     95  1.1  hikaru uint32_t	qat_crypto_setup_auth_desc(struct qat_session *,
     96  1.1  hikaru 		    struct qat_crypto_desc *, struct cryptoini *,
     97  1.1  hikaru 		    struct fw_auth_hdr *, uint8_t *, uint32_t, enum fw_slice);
     98  1.1  hikaru 
     99  1.1  hikaru void
    100  1.1  hikaru qat_msg_req_type_populate(struct arch_if_req_hdr *msg, enum arch_if_req type,
    101  1.1  hikaru     uint32_t rxring)
    102  1.1  hikaru {
    103  1.1  hikaru 
    104  1.1  hikaru 	memset(msg, 0, sizeof(struct arch_if_req_hdr));
    105  1.1  hikaru 	msg->flags = ARCH_IF_FLAGS_VALID_FLAG |
    106  1.1  hikaru 	    ARCH_IF_FLAGS_RESP_RING_TYPE_ET | ARCH_IF_FLAGS_RESP_TYPE_S;
    107  1.1  hikaru 	msg->req_type = type;
    108  1.1  hikaru 	msg->resp_pipe_id = rxring;
    109  1.1  hikaru }
    110  1.1  hikaru 
    111  1.1  hikaru void
    112  1.1  hikaru qat_msg_cmn_hdr_populate(struct fw_la_bulk_req *msg, bus_addr_t desc_paddr,
    113  1.1  hikaru     uint8_t hdrsz, uint8_t hwblksz, uint16_t comn_req_flags, uint32_t flow_id)
    114  1.1  hikaru {
    115  1.1  hikaru 	struct fw_comn_req_hdr *hdr = &msg->comn_hdr;
    116  1.1  hikaru 
    117  1.1  hikaru 	hdr->comn_req_flags = comn_req_flags;
    118  1.1  hikaru 	hdr->content_desc_params_sz = hwblksz;
    119  1.1  hikaru 	hdr->content_desc_hdr_sz = hdrsz;
    120  1.1  hikaru 	hdr->content_desc_addr = desc_paddr;
    121  1.1  hikaru 	msg->flow_id = flow_id;
    122  1.1  hikaru }
    123  1.1  hikaru 
    124  1.1  hikaru void
    125  1.1  hikaru qat_msg_service_cmd_populate(struct fw_la_bulk_req *msg, enum fw_la_cmd_id cmdid,
    126  1.1  hikaru     uint16_t cmd_flags)
    127  1.1  hikaru {
    128  1.1  hikaru 	msg->comn_la_req.la_cmd_id = cmdid;
    129  1.1  hikaru 	msg->comn_la_req.u.la_flags = cmd_flags;
    130  1.1  hikaru }
    131  1.1  hikaru 
    132  1.1  hikaru void
    133  1.1  hikaru qat_msg_cmn_mid_populate(struct fw_comn_req_mid *msg, void *cookie,
    134  1.1  hikaru     uint64_t src, uint64_t dst)
    135  1.1  hikaru {
    136  1.1  hikaru 
    137  1.1  hikaru 	msg->opaque_data = (uint64_t)(uintptr_t)cookie;
    138  1.1  hikaru 	msg->src_data_addr = src;
    139  1.1  hikaru 	if (dst == 0)
    140  1.1  hikaru 		msg->dest_data_addr = src;
    141  1.1  hikaru 	else
    142  1.1  hikaru 		msg->dest_data_addr = dst;
    143  1.1  hikaru }
    144  1.1  hikaru 
    145  1.1  hikaru void
    146  1.1  hikaru qat_msg_req_params_populate(struct fw_la_bulk_req *msg,
    147  1.1  hikaru     bus_addr_t req_params_paddr, uint8_t req_params_sz)
    148  1.1  hikaru {
    149  1.1  hikaru 	msg->req_params_addr = req_params_paddr;
    150  1.1  hikaru 	msg->comn_la_req.u1.req_params_blk_sz = req_params_sz / 8;
    151  1.1  hikaru }
    152  1.1  hikaru 
    153  1.1  hikaru 
    154  1.1  hikaru void
    155  1.1  hikaru qat_msg_cmn_footer_populate(union fw_comn_req_ftr *msg, uint64_t next_addr)
    156  1.1  hikaru {
    157  1.1  hikaru 	msg->next_request_addr = next_addr;
    158  1.1  hikaru }
    159  1.1  hikaru 
    160  1.1  hikaru void
    161  1.1  hikaru qat_msg_params_populate(struct fw_la_bulk_req *msg,
    162  1.1  hikaru     struct qat_crypto_desc *desc, uint8_t req_params_sz,
    163  1.1  hikaru     uint16_t service_cmd_flags, uint16_t comn_req_flags)
    164  1.1  hikaru {
    165  1.1  hikaru 	qat_msg_cmn_hdr_populate(msg, desc->qcd_desc_paddr,
    166  1.1  hikaru 	    desc->qcd_hdr_sz, desc->qcd_hw_blk_sz, comn_req_flags, 0);
    167  1.1  hikaru 	qat_msg_service_cmd_populate(msg, desc->qcd_cmd_id, service_cmd_flags);
    168  1.1  hikaru 	qat_msg_cmn_mid_populate(&msg->comn_mid, NULL, 0, 0);
    169  1.1  hikaru 	qat_msg_req_params_populate(msg, 0, req_params_sz);
    170  1.1  hikaru 	qat_msg_cmn_footer_populate(&msg->comn_ftr, 0);
    171  1.1  hikaru }
    172  1.1  hikaru 
    173  1.1  hikaru int
    174  1.1  hikaru qat_adm_ring_init_ring_table(struct qat_softc *sc)
    175  1.1  hikaru {
    176  1.1  hikaru 	struct qat_admin_rings *qadr = &sc->sc_admin_rings;
    177  1.1  hikaru 
    178  1.1  hikaru 	if (sc->sc_ae_num == 1) {
    179  1.1  hikaru 		qadr->qadr_cya_ring_tbl =
    180  1.1  hikaru 		    &qadr->qadr_master_ring_tbl[0];
    181  1.1  hikaru 		qadr->qadr_srv_mask[0] = QAT_SERVICE_CRYPTO_A;
    182  1.1  hikaru 	} else if (sc->sc_ae_num == 2 ||
    183  1.1  hikaru 	    sc->sc_ae_num == 4) {
    184  1.1  hikaru 		qadr->qadr_cya_ring_tbl =
    185  1.1  hikaru 		    &qadr->qadr_master_ring_tbl[0];
    186  1.1  hikaru 		qadr->qadr_srv_mask[0] = QAT_SERVICE_CRYPTO_A;
    187  1.1  hikaru 		qadr->qadr_cyb_ring_tbl =
    188  1.1  hikaru 		    &qadr->qadr_master_ring_tbl[1];
    189  1.1  hikaru 		qadr->qadr_srv_mask[1] = QAT_SERVICE_CRYPTO_B;
    190  1.1  hikaru 	}
    191  1.1  hikaru 
    192  1.1  hikaru 	return 0;
    193  1.1  hikaru }
    194  1.1  hikaru 
    195  1.1  hikaru int
    196  1.1  hikaru qat_adm_ring_init(struct qat_softc *sc)
    197  1.1  hikaru {
    198  1.1  hikaru 	struct qat_admin_rings *qadr = &sc->sc_admin_rings;
    199  1.1  hikaru 	int error, i, j;
    200  1.1  hikaru 
    201  1.1  hikaru 	error = qat_alloc_dmamem(sc, &qadr->qadr_dma,
    202  1.1  hikaru 	    PAGE_SIZE, PAGE_SIZE);
    203  1.1  hikaru 	if (error)
    204  1.1  hikaru 		return error;
    205  1.1  hikaru 
    206  1.1  hikaru 	qadr->qadr_master_ring_tbl = qadr->qadr_dma.qdm_dma_vaddr;
    207  1.1  hikaru 
    208  1.1  hikaru 	KASSERT(sc->sc_ae_num *
    209  1.1  hikaru 	    sizeof(struct fw_init_ring_table) <= PAGE_SIZE);
    210  1.1  hikaru 
    211  1.1  hikaru 	/* Initialize the Master Ring Table */
    212  1.1  hikaru 	for (i = 0; i < sc->sc_ae_num; i++) {
    213  1.1  hikaru 		struct fw_init_ring_table *firt =
    214  1.1  hikaru 		    &qadr->qadr_master_ring_tbl[i];
    215  1.1  hikaru 
    216  1.1  hikaru 		for (j = 0; j < INIT_RING_TABLE_SZ; j++) {
    217  1.1  hikaru 			struct fw_init_ring_params *firp =
    218  1.1  hikaru 			    &firt->firt_bulk_rings[j];
    219  1.1  hikaru 
    220  1.1  hikaru 			firp->firp_reserved = 0;
    221  1.1  hikaru 			firp->firp_curr_weight = QAT_DEFAULT_RING_WEIGHT;
    222  1.1  hikaru 			firp->firp_init_weight = QAT_DEFAULT_RING_WEIGHT;
    223  1.1  hikaru 			firp->firp_ring_pvl = QAT_DEFAULT_PVL;
    224  1.1  hikaru 		}
    225  1.1  hikaru 		memset(firt->firt_ring_mask, 0, sizeof(firt->firt_ring_mask));
    226  1.1  hikaru 	}
    227  1.1  hikaru 
    228  1.1  hikaru 	error = qat_etr_setup_ring(sc, 0, RING_NUM_ADMIN_TX,
    229  1.1  hikaru 	    ADMIN_RING_SIZE, sc->sc_hw.qhw_fw_req_size,
    230  1.1  hikaru 	    NULL, NULL, "admin_tx", &qadr->qadr_admin_tx);
    231  1.1  hikaru 	if (error)
    232  1.1  hikaru 		return error;
    233  1.1  hikaru 
    234  1.1  hikaru 	error = qat_etr_setup_ring(sc, 0, RING_NUM_ADMIN_RX,
    235  1.1  hikaru 	    ADMIN_RING_SIZE, sc->sc_hw.qhw_fw_resp_size,
    236  1.1  hikaru 	    qat_adm_ring_intr, qadr, "admin_rx", &qadr->qadr_admin_rx);
    237  1.1  hikaru 	if (error)
    238  1.1  hikaru 		return error;
    239  1.1  hikaru 
    240  1.1  hikaru 	/*
    241  1.1  hikaru 	 * Finally set up the service indices into the Master Ring Table
    242  1.1  hikaru 	 * and convenient ring table pointers for each service enabled.
    243  1.1  hikaru 	 * Only the Admin rings are initialized.
    244  1.1  hikaru 	 */
    245  1.1  hikaru 	error = qat_adm_ring_init_ring_table(sc);
    246  1.1  hikaru 	if (error)
    247  1.1  hikaru 		return error;
    248  1.1  hikaru 
    249  1.1  hikaru 	/*
    250  1.1  hikaru 	 * Calculate the number of active AEs per QAT
    251  1.1  hikaru 	 * needed for Shram partitioning.
    252  1.1  hikaru 	 */
    253  1.1  hikaru 	for (i = 0; i < sc->sc_ae_num; i++) {
    254  1.1  hikaru 		if (qadr->qadr_srv_mask[i])
    255  1.1  hikaru 			qadr->qadr_active_aes_per_accel++;
    256  1.1  hikaru 	}
    257  1.1  hikaru 
    258  1.1  hikaru 	return 0;
    259  1.1  hikaru }
    260  1.1  hikaru 
    261  1.1  hikaru void
    262  1.1  hikaru qat_adm_ring_build_slice_mask(uint16_t *slice_mask, uint32_t srv_mask,
    263  1.1  hikaru    uint32_t init_shram)
    264  1.1  hikaru {
    265  1.1  hikaru 	uint16_t shram = 0, comn_req = 0;
    266  1.1  hikaru 
    267  1.1  hikaru 	if (init_shram)
    268  1.1  hikaru 		shram = COMN_REQ_SHRAM_INIT_REQUIRED;
    269  1.1  hikaru 
    270  1.1  hikaru 	if (srv_mask & QAT_SERVICE_CRYPTO_A)
    271  1.1  hikaru 		comn_req |= COMN_REQ_CY0_ONLY(shram);
    272  1.1  hikaru 	if (srv_mask & QAT_SERVICE_CRYPTO_B)
    273  1.1  hikaru 		comn_req |= COMN_REQ_CY1_ONLY(shram);
    274  1.1  hikaru 
    275  1.1  hikaru 	*slice_mask = comn_req;
    276  1.1  hikaru }
    277  1.1  hikaru 
    278  1.1  hikaru void
    279  1.1  hikaru qat_adm_ring_build_shram_mask(uint64_t *shram_mask, uint32_t active_aes,
    280  1.1  hikaru     uint32_t ae)
    281  1.1  hikaru {
    282  1.1  hikaru 	*shram_mask = 0;
    283  1.1  hikaru 
    284  1.1  hikaru 	if (active_aes == 1) {
    285  1.1  hikaru 		*shram_mask = ~(*shram_mask);
    286  1.1  hikaru 	} else if (active_aes == 2) {
    287  1.1  hikaru 		if (ae == 1)
    288  1.1  hikaru 			*shram_mask = ((~(*shram_mask)) & 0xffffffff);
    289  1.1  hikaru 		else
    290  1.1  hikaru 			*shram_mask = ((~(*shram_mask)) & 0xffffffff00000000ull);
    291  1.1  hikaru 	} else if (active_aes == 3) {
    292  1.1  hikaru 		if (ae == 0)
    293  1.1  hikaru 			*shram_mask = ((~(*shram_mask)) & 0x7fffff);
    294  1.1  hikaru 		else if (ae == 1)
    295  1.1  hikaru 			*shram_mask = ((~(*shram_mask)) & 0x3fffff800000ull);
    296  1.1  hikaru 		else
    297  1.1  hikaru 			*shram_mask = ((~(*shram_mask)) & 0xffffc00000000000ull);
    298  1.1  hikaru 	} else {
    299  1.1  hikaru 		panic("Only three services are supported in current version");
    300  1.1  hikaru 	}
    301  1.1  hikaru }
    302  1.1  hikaru 
    303  1.1  hikaru int
    304  1.1  hikaru qat_adm_ring_build_ring_table(struct qat_softc *sc, uint32_t ae)
    305  1.1  hikaru {
    306  1.1  hikaru 	struct qat_admin_rings *qadr = &sc->sc_admin_rings;
    307  1.1  hikaru 	struct fw_init_ring_table *tbl;
    308  1.1  hikaru 	struct fw_init_ring_params *param;
    309  1.1  hikaru 	uint8_t srv_mask = sc->sc_admin_rings.qadr_srv_mask[ae];
    310  1.1  hikaru 
    311  1.1  hikaru 	if ((srv_mask & QAT_SERVICE_CRYPTO_A)) {
    312  1.1  hikaru 		tbl = qadr->qadr_cya_ring_tbl;
    313  1.1  hikaru 	} else if ((srv_mask & QAT_SERVICE_CRYPTO_B)) {
    314  1.1  hikaru 		tbl = qadr->qadr_cyb_ring_tbl;
    315  1.1  hikaru 	} else {
    316  1.1  hikaru 		aprint_error_dev(sc->sc_dev,
    317  1.1  hikaru 		    "Invalid execution engine %d\n", ae);
    318  1.1  hikaru 		return EINVAL;
    319  1.1  hikaru 	}
    320  1.1  hikaru 
    321  1.1  hikaru 	param = &tbl->firt_bulk_rings[sc->sc_hw.qhw_ring_sym_tx];
    322  1.1  hikaru 	param->firp_curr_weight = QAT_HI_PRIO_RING_WEIGHT;
    323  1.1  hikaru 	param->firp_init_weight = QAT_HI_PRIO_RING_WEIGHT;
    324  1.1  hikaru 	FW_INIT_RING_MASK_SET(tbl, sc->sc_hw.qhw_ring_sym_tx);
    325  1.1  hikaru 
    326  1.1  hikaru 	return 0;
    327  1.1  hikaru }
    328  1.1  hikaru 
    329  1.1  hikaru int
    330  1.1  hikaru qat_adm_ring_build_init_msg(struct qat_softc *sc,
    331  1.1  hikaru     struct fw_init_req *initmsg, enum fw_init_cmd_id cmd, uint32_t ae,
    332  1.1  hikaru     struct qat_accel_init_cb *cb)
    333  1.1  hikaru {
    334  1.1  hikaru 	struct fw_init_set_ae_info_hdr *aehdr;
    335  1.1  hikaru 	struct fw_init_set_ae_info *aeinfo;
    336  1.1  hikaru 	struct fw_init_set_ring_info_hdr *ringhdr;
    337  1.1  hikaru 	struct fw_init_set_ring_info *ringinfo;
    338  1.1  hikaru 	int init_shram = 0, tgt_id, cluster_id;
    339  1.1  hikaru 	uint32_t srv_mask;
    340  1.1  hikaru 
    341  1.1  hikaru 	srv_mask = sc->sc_admin_rings.qadr_srv_mask[
    342  1.1  hikaru 	    ae % sc->sc_ae_num];
    343  1.1  hikaru 
    344  1.1  hikaru 	memset(initmsg, 0, sizeof(struct fw_init_req));
    345  1.1  hikaru 
    346  1.1  hikaru 	qat_msg_req_type_populate(&initmsg->comn_hdr.arch_if,
    347  1.1  hikaru 	    ARCH_IF_REQ_QAT_FW_INIT,
    348  1.1  hikaru 	    sc->sc_admin_rings.qadr_admin_rx->qr_ring_id);
    349  1.1  hikaru 
    350  1.1  hikaru 	qat_msg_cmn_mid_populate(&initmsg->comn_mid, cb, 0, 0);
    351  1.1  hikaru 
    352  1.1  hikaru 	switch (cmd) {
    353  1.1  hikaru 	case FW_INIT_CMD_SET_AE_INFO:
    354  1.1  hikaru 		if (ae % sc->sc_ae_num == 0)
    355  1.1  hikaru 			init_shram = 1;
    356  1.1  hikaru 		if (ae >= sc->sc_ae_num) {
    357  1.1  hikaru 			tgt_id = 1;
    358  1.1  hikaru 			cluster_id = 1;
    359  1.1  hikaru 		} else {
    360  1.1  hikaru 			cluster_id = 0;
    361  1.1  hikaru 			if (sc->sc_ae_mask)
    362  1.1  hikaru 				tgt_id = 0;
    363  1.1  hikaru 			else
    364  1.1  hikaru 				tgt_id = 1;
    365  1.1  hikaru 		}
    366  1.1  hikaru 		aehdr = &initmsg->u.set_ae_info;
    367  1.1  hikaru 		aeinfo = &initmsg->u1.set_ae_info;
    368  1.1  hikaru 
    369  1.1  hikaru 		aehdr->init_cmd_id = cmd;
    370  1.1  hikaru 		/* XXX that does not support sparse ae_mask */
    371  1.1  hikaru 		aehdr->init_trgt_id = ae;
    372  1.1  hikaru 		aehdr->init_ring_cluster_id = cluster_id;
    373  1.1  hikaru 		aehdr->init_qat_id = tgt_id;
    374  1.1  hikaru 
    375  1.1  hikaru 		qat_adm_ring_build_slice_mask(&aehdr->init_slice_mask, srv_mask,
    376  1.1  hikaru 		    init_shram);
    377  1.1  hikaru 
    378  1.1  hikaru 		qat_adm_ring_build_shram_mask(&aeinfo->init_shram_mask,
    379  1.1  hikaru 		    sc->sc_admin_rings.qadr_active_aes_per_accel,
    380  1.1  hikaru 		    ae % sc->sc_ae_num);
    381  1.1  hikaru 
    382  1.1  hikaru 		break;
    383  1.1  hikaru 	case FW_INIT_CMD_SET_RING_INFO:
    384  1.1  hikaru 		ringhdr = &initmsg->u.set_ring_info;
    385  1.1  hikaru 		ringinfo = &initmsg->u1.set_ring_info;
    386  1.1  hikaru 
    387  1.1  hikaru 		ringhdr->init_cmd_id = cmd;
    388  1.1  hikaru 		/* XXX that does not support sparse ae_mask */
    389  1.1  hikaru 		ringhdr->init_trgt_id = ae;
    390  1.1  hikaru 
    391  1.1  hikaru 		/* XXX */
    392  1.1  hikaru 		qat_adm_ring_build_ring_table(sc,
    393  1.1  hikaru 		    ae % sc->sc_ae_num);
    394  1.1  hikaru 
    395  1.1  hikaru 		ringhdr->init_ring_tbl_sz = sizeof(struct fw_init_ring_table);
    396  1.1  hikaru 
    397  1.1  hikaru 		ringinfo->init_ring_table_ptr =
    398  1.1  hikaru 		    sc->sc_admin_rings.qadr_dma.qdm_dma_seg.ds_addr +
    399  1.1  hikaru 		    ((ae % sc->sc_ae_num) *
    400  1.1  hikaru 		    sizeof(struct fw_init_ring_table));
    401  1.1  hikaru 
    402  1.1  hikaru 		break;
    403  1.1  hikaru 	default:
    404  1.1  hikaru 		return ENOTSUP;
    405  1.1  hikaru 	}
    406  1.1  hikaru 
    407  1.1  hikaru 	return 0;
    408  1.1  hikaru }
    409  1.1  hikaru 
    410  1.1  hikaru int
    411  1.1  hikaru qat_adm_ring_send_init_msg_sync(struct qat_softc *sc,
    412  1.1  hikaru     enum fw_init_cmd_id cmd, uint32_t ae)
    413  1.1  hikaru {
    414  1.1  hikaru 	struct fw_init_req initmsg;
    415  1.1  hikaru 	struct qat_accel_init_cb cb;
    416  1.1  hikaru 	int error;
    417  1.1  hikaru 
    418  1.1  hikaru 	error = qat_adm_ring_build_init_msg(sc, &initmsg, cmd, ae, &cb);
    419  1.1  hikaru 	if (error)
    420  1.1  hikaru 		return error;
    421  1.1  hikaru 
    422  1.1  hikaru 	error = qat_etr_put_msg(sc, sc->sc_admin_rings.qadr_admin_tx,
    423  1.1  hikaru 	    (uint32_t *)&initmsg);
    424  1.1  hikaru 	if (error)
    425  1.1  hikaru 		return error;
    426  1.1  hikaru 
    427  1.1  hikaru 	error = tsleep(&cb, PZERO, "qat_init", hz * 3 / 2);
    428  1.1  hikaru 	if (error) {
    429  1.1  hikaru 		aprint_error_dev(sc->sc_dev,
    430  1.1  hikaru 		    "Timed out initialization firmware: %d\n", error);
    431  1.1  hikaru 		return error;
    432  1.1  hikaru 	}
    433  1.1  hikaru 	if (cb.qaic_status) {
    434  1.1  hikaru 		aprint_error_dev(sc->sc_dev, "Failed to initialize firmware\n");
    435  1.1  hikaru 		return EIO;
    436  1.1  hikaru 	}
    437  1.1  hikaru 
    438  1.1  hikaru 	return error;
    439  1.1  hikaru }
    440  1.1  hikaru 
    441  1.1  hikaru int
    442  1.1  hikaru qat_adm_ring_send_init_msg(struct qat_softc *sc,
    443  1.1  hikaru     enum fw_init_cmd_id cmd)
    444  1.1  hikaru {
    445  1.1  hikaru 	struct qat_admin_rings *qadr = &sc->sc_admin_rings;
    446  1.1  hikaru 	uint32_t error, ae;
    447  1.1  hikaru 
    448  1.1  hikaru 	for (ae = 0; ae < sc->sc_ae_num; ae++) {
    449  1.1  hikaru 		uint8_t srv_mask = qadr->qadr_srv_mask[ae];
    450  1.1  hikaru 		switch (cmd) {
    451  1.1  hikaru 		case FW_INIT_CMD_SET_AE_INFO:
    452  1.1  hikaru 		case FW_INIT_CMD_SET_RING_INFO:
    453  1.1  hikaru 			if (!srv_mask)
    454  1.1  hikaru 				continue;
    455  1.1  hikaru 			break;
    456  1.1  hikaru 		case FW_INIT_CMD_TRNG_ENABLE:
    457  1.1  hikaru 		case FW_INIT_CMD_TRNG_DISABLE:
    458  1.1  hikaru 			if (!(srv_mask & QAT_SERVICE_CRYPTO_A))
    459  1.1  hikaru 				continue;
    460  1.1  hikaru 			break;
    461  1.1  hikaru 		default:
    462  1.1  hikaru 			return ENOTSUP;
    463  1.1  hikaru 		}
    464  1.1  hikaru 
    465  1.1  hikaru 		error = qat_adm_ring_send_init_msg_sync(sc, cmd, ae);
    466  1.1  hikaru 		if (error)
    467  1.1  hikaru 			return error;
    468  1.1  hikaru 	}
    469  1.1  hikaru 
    470  1.1  hikaru 	return 0;
    471  1.1  hikaru }
    472  1.1  hikaru 
    473  1.1  hikaru int
    474  1.1  hikaru qat_adm_ring_send_init(struct qat_softc *sc)
    475  1.1  hikaru {
    476  1.1  hikaru 	int error;
    477  1.1  hikaru 
    478  1.1  hikaru 	error = qat_adm_ring_send_init_msg(sc, FW_INIT_CMD_SET_AE_INFO);
    479  1.1  hikaru 	if (error)
    480  1.1  hikaru 		return error;
    481  1.1  hikaru 
    482  1.1  hikaru 	error = qat_adm_ring_send_init_msg(sc, FW_INIT_CMD_SET_RING_INFO);
    483  1.1  hikaru 	if (error)
    484  1.1  hikaru 		return error;
    485  1.1  hikaru 
    486  1.1  hikaru 	aprint_verbose_dev(sc->sc_dev, "Initialization completed\n");
    487  1.1  hikaru 
    488  1.1  hikaru 	return 0;
    489  1.1  hikaru }
    490  1.1  hikaru 
    491  1.1  hikaru int
    492  1.1  hikaru qat_adm_ring_intr(struct qat_softc *sc, void *arg, void *msg)
    493  1.1  hikaru {
    494  1.1  hikaru 	struct arch_if_resp_hdr *resp;
    495  1.1  hikaru 	struct fw_init_resp *init_resp;
    496  1.1  hikaru 	struct qat_accel_init_cb *init_cb;
    497  1.1  hikaru 	int handled = 0;
    498  1.1  hikaru 
    499  1.1  hikaru 	resp = (struct arch_if_resp_hdr *)msg;
    500  1.1  hikaru 
    501  1.1  hikaru 	switch (resp->resp_type) {
    502  1.1  hikaru 	case ARCH_IF_REQ_QAT_FW_INIT:
    503  1.1  hikaru 		init_resp = (struct fw_init_resp *)msg;
    504  1.1  hikaru 		init_cb = (struct qat_accel_init_cb *)
    505  1.1  hikaru 		    (uintptr_t)init_resp->comn_resp.opaque_data;
    506  1.1  hikaru 		init_cb->qaic_status =
    507  1.1  hikaru 		    __SHIFTOUT(init_resp->comn_resp.comn_status,
    508  1.1  hikaru 		    COMN_RESP_INIT_ADMIN_STATUS);
    509  1.1  hikaru 		wakeup(init_cb);
    510  1.1  hikaru 		break;
    511  1.1  hikaru 	default:
    512  1.1  hikaru 		aprint_error_dev(sc->sc_dev,
    513  1.1  hikaru 		    "unknown resp type %d\n", resp->resp_type);
    514  1.1  hikaru 		break;
    515  1.1  hikaru 	}
    516  1.1  hikaru 
    517  1.1  hikaru 	return handled;
    518  1.1  hikaru }
    519  1.1  hikaru 
    520  1.1  hikaru static inline uint16_t
    521  1.1  hikaru qat_hw15_get_comn_req_flags(uint8_t ae)
    522  1.1  hikaru {
    523  1.1  hikaru 	if (ae == 0) {
    524  1.1  hikaru 		return COMN_REQ_ORD_STRICT | COMN_REQ_PTR_TYPE_SGL |
    525  1.1  hikaru 		    COMN_REQ_AUTH0_SLICE_REQUIRED |
    526  1.1  hikaru 		    COMN_REQ_CIPHER0_SLICE_REQUIRED;
    527  1.1  hikaru 	} else {
    528  1.1  hikaru 		return COMN_REQ_ORD_STRICT | COMN_REQ_PTR_TYPE_SGL |
    529  1.1  hikaru 		    COMN_REQ_AUTH1_SLICE_REQUIRED |
    530  1.1  hikaru 		    COMN_REQ_CIPHER1_SLICE_REQUIRED;
    531  1.1  hikaru 	}
    532  1.1  hikaru }
    533  1.1  hikaru 
    534  1.1  hikaru uint32_t
    535  1.1  hikaru qat_crypto_setup_cipher_desc(struct qat_session *qs,
    536  1.1  hikaru     struct qat_crypto_desc *desc, struct cryptoini *crie,
    537  1.1  hikaru     struct fw_cipher_hdr *cipher_hdr, uint8_t *hw_blk_ptr,
    538  1.1  hikaru     uint32_t hw_blk_offset, enum fw_slice next_slice)
    539  1.1  hikaru {
    540  1.1  hikaru 	struct hw_cipher_config *cipher_config = (struct hw_cipher_config *)
    541  1.1  hikaru 	    (hw_blk_ptr + hw_blk_offset);
    542  1.1  hikaru 	uint32_t hw_blk_size;
    543  1.1  hikaru 	uint8_t *cipher_key = (uint8_t *)(cipher_config + 1);
    544  1.1  hikaru 
    545  1.1  hikaru 	cipher_config->val = qat_crypto_load_cipher_cryptoini(desc, crie);
    546  1.1  hikaru 	cipher_config->reserved = 0;
    547  1.1  hikaru 
    548  1.1  hikaru 	cipher_hdr->state_padding_sz = 0;
    549  1.1  hikaru 	cipher_hdr->key_sz = crie->cri_klen / 64; /* bits to quad words */
    550  1.1  hikaru 
    551  1.1  hikaru 	cipher_hdr->state_sz = desc->qcd_cipher_blk_sz / 8;
    552  1.1  hikaru 
    553  1.1  hikaru 	cipher_hdr->next_id = next_slice;
    554  1.1  hikaru 	cipher_hdr->curr_id = FW_SLICE_CIPHER;
    555  1.1  hikaru 	cipher_hdr->offset = hw_blk_offset / 8;
    556  1.1  hikaru 	cipher_hdr->resrvd = 0;
    557  1.1  hikaru 
    558  1.1  hikaru 	hw_blk_size = sizeof(struct hw_cipher_config);
    559  1.1  hikaru 
    560  1.1  hikaru 	memcpy(cipher_key, crie->cri_key, crie->cri_klen / 8);
    561  1.1  hikaru 	hw_blk_size += crie->cri_klen / 8;
    562  1.1  hikaru 
    563  1.1  hikaru 	return hw_blk_size;
    564  1.1  hikaru }
    565  1.1  hikaru 
    566  1.1  hikaru uint32_t
    567  1.1  hikaru qat_crypto_setup_auth_desc(struct qat_session *qs, struct qat_crypto_desc *desc,
    568  1.1  hikaru     struct cryptoini *cria, struct fw_auth_hdr *auth_hdr, uint8_t *hw_blk_ptr,
    569  1.1  hikaru     uint32_t hw_blk_offset, enum fw_slice next_slice)
    570  1.1  hikaru {
    571  1.1  hikaru 	struct qat_sym_hash_def const *hash_def;
    572  1.1  hikaru 	const struct swcr_auth_hash *sah;
    573  1.1  hikaru 	struct hw_auth_setup *auth_setup;
    574  1.1  hikaru 	uint32_t hw_blk_size;
    575  1.1  hikaru 	uint8_t *state1, *state2;
    576  1.1  hikaru 	uint32_t state_size;
    577  1.1  hikaru 
    578  1.1  hikaru 	auth_setup = (struct hw_auth_setup *)(hw_blk_ptr + hw_blk_offset);
    579  1.1  hikaru 
    580  1.1  hikaru 	auth_setup->auth_config.config =
    581  1.1  hikaru 	    qat_crypto_load_auth_cryptoini(desc, cria, &hash_def);
    582  1.1  hikaru 	sah = hash_def->qshd_alg->qshai_sah;
    583  1.1  hikaru 	auth_setup->auth_config.reserved = 0;
    584  1.1  hikaru 
    585  1.1  hikaru 	/* for HMAC in mode 1 authCounter is the block size
    586  1.1  hikaru 	 * else the authCounter is 0. The firmware expects the counter to be
    587  1.1  hikaru 	 * big endian */
    588  1.1  hikaru 	auth_setup->auth_counter.counter =
    589  1.1  hikaru 	    htonl(hash_def->qshd_qat->qshqi_auth_counter);
    590  1.1  hikaru 	auth_setup->auth_counter.reserved = 0;
    591  1.1  hikaru 
    592  1.1  hikaru 	auth_hdr->next_id = next_slice;
    593  1.1  hikaru 	auth_hdr->curr_id = FW_SLICE_AUTH;
    594  1.1  hikaru 	auth_hdr->offset = hw_blk_offset / 8;
    595  1.1  hikaru 	auth_hdr->resrvd = 0;
    596  1.1  hikaru 
    597  1.1  hikaru 	auth_hdr->hash_flags = FW_AUTH_HDR_FLAG_NO_NESTED;
    598  1.1  hikaru 	auth_hdr->u.inner_prefix_sz = 0;
    599  1.1  hikaru 	auth_hdr->outer_prefix_sz = 0;
    600  1.1  hikaru 	auth_hdr->final_sz = sah->auth_hash->authsize;
    601  1.1  hikaru 	auth_hdr->inner_state1_sz =
    602  1.1  hikaru 	    roundup(hash_def->qshd_qat->qshqi_state1_len, 8);
    603  1.1  hikaru 	auth_hdr->inner_res_sz = hash_def->qshd_alg->qshai_digest_len;
    604  1.1  hikaru 	auth_hdr->inner_state2_sz =
    605  1.1  hikaru 	    roundup(hash_def->qshd_qat->qshqi_state2_len, 8);
    606  1.1  hikaru 	auth_hdr->inner_state2_off = auth_hdr->offset +
    607  1.1  hikaru 	    ((sizeof(struct hw_auth_setup) + auth_hdr->inner_state1_sz) / 8);
    608  1.1  hikaru 
    609  1.1  hikaru 	hw_blk_size = sizeof(struct hw_auth_setup) + auth_hdr->inner_state1_sz +
    610  1.1  hikaru 	    auth_hdr->inner_state2_sz;
    611  1.1  hikaru 
    612  1.1  hikaru 	auth_hdr->outer_config_off = 0;
    613  1.1  hikaru 	auth_hdr->outer_state1_sz = 0;
    614  1.1  hikaru 	auth_hdr->outer_res_sz = 0;
    615  1.1  hikaru 	auth_hdr->outer_prefix_off = 0;
    616  1.1  hikaru 
    617  1.1  hikaru 	state1 = (uint8_t *)(auth_setup + 1);
    618  1.1  hikaru 	state2 = state1 + auth_hdr->inner_state1_sz;
    619  1.1  hikaru 
    620  1.1  hikaru 	state_size = hash_def->qshd_alg->qshai_state_size;
    621  1.1  hikaru 	if (hash_def->qshd_qat->qshqi_algo_enc == HW_AUTH_ALGO_SHA1) {
    622  1.1  hikaru 		uint32_t state1_pad_len = auth_hdr->inner_state1_sz -
    623  1.1  hikaru 		    state_size;
    624  1.1  hikaru 		uint32_t state2_pad_len = auth_hdr->inner_state2_sz -
    625  1.1  hikaru 		    state_size;
    626  1.1  hikaru 		if (state1_pad_len > 0)
    627  1.1  hikaru 			memset(state1 + state_size, 0, state1_pad_len);
    628  1.1  hikaru 		if (state2_pad_len > 0)
    629  1.1  hikaru 			memset(state2 + state_size, 0, state2_pad_len);
    630  1.1  hikaru 	}
    631  1.1  hikaru 
    632  1.1  hikaru 	desc->qcd_state_storage_sz = (sizeof(struct hw_auth_counter) +
    633  1.1  hikaru 	    roundup(state_size, 8)) / 8;
    634  1.1  hikaru 
    635  1.1  hikaru 	qat_crypto_hmac_precompute(desc, cria, hash_def, state1, state2);
    636  1.1  hikaru 
    637  1.1  hikaru 	return hw_blk_size;
    638  1.1  hikaru }
    639  1.1  hikaru 
    640  1.1  hikaru 
    641  1.1  hikaru void
    642  1.1  hikaru qat_hw15_crypto_setup_desc(struct qat_crypto *qcy, struct qat_session *qs,
    643  1.1  hikaru     struct qat_crypto_desc *desc,
    644  1.1  hikaru     struct cryptoini *crie, struct cryptoini *cria)
    645  1.1  hikaru {
    646  1.1  hikaru 	struct fw_cipher_hdr *cipher_hdr;
    647  1.1  hikaru 	struct fw_auth_hdr *auth_hdr;
    648  1.1  hikaru 	struct fw_la_bulk_req *req_cache;
    649  1.1  hikaru 	uint32_t ctrl_blk_size = 0, ctrl_blk_offset = 0, hw_blk_offset = 0;
    650  1.1  hikaru 	int i;
    651  1.1  hikaru 	uint16_t la_cmd_flags = 0;
    652  1.1  hikaru 	uint8_t req_params_sz = 0;
    653  1.1  hikaru 	uint8_t *ctrl_blk_ptr;
    654  1.1  hikaru 	uint8_t *hw_blk_ptr;
    655  1.1  hikaru 
    656  1.1  hikaru 	if (crie != NULL)
    657  1.1  hikaru 		ctrl_blk_size += sizeof(struct fw_cipher_hdr);
    658  1.1  hikaru 	if (cria != NULL)
    659  1.1  hikaru 		ctrl_blk_size += sizeof(struct fw_auth_hdr);
    660  1.1  hikaru 
    661  1.1  hikaru 	ctrl_blk_ptr = desc->qcd_content_desc;
    662  1.1  hikaru 	hw_blk_ptr = ctrl_blk_ptr + ctrl_blk_size;
    663  1.1  hikaru 
    664  1.1  hikaru 	for (i = 0; i < MAX_FW_SLICE; i++) {
    665  1.1  hikaru 		switch (desc->qcd_slices[i]) {
    666  1.1  hikaru 		case FW_SLICE_CIPHER:
    667  1.1  hikaru 			cipher_hdr = (struct fw_cipher_hdr *)(ctrl_blk_ptr +
    668  1.1  hikaru 			    ctrl_blk_offset);
    669  1.1  hikaru 			ctrl_blk_offset += sizeof(struct fw_cipher_hdr);
    670  1.1  hikaru 			hw_blk_offset += qat_crypto_setup_cipher_desc(qs, desc,
    671  1.1  hikaru 			    crie, cipher_hdr, hw_blk_ptr, hw_blk_offset,
    672  1.1  hikaru 			    desc->qcd_slices[i + 1]);
    673  1.1  hikaru 			req_params_sz += sizeof(struct fw_la_cipher_req_params);
    674  1.1  hikaru 			break;
    675  1.1  hikaru 		case FW_SLICE_AUTH:
    676  1.1  hikaru 			auth_hdr = (struct fw_auth_hdr *)(ctrl_blk_ptr +
    677  1.1  hikaru 			    ctrl_blk_offset);
    678  1.1  hikaru 			ctrl_blk_offset += sizeof(struct fw_auth_hdr);
    679  1.1  hikaru 			hw_blk_offset += qat_crypto_setup_auth_desc(qs, desc,
    680  1.1  hikaru 			    cria, auth_hdr, hw_blk_ptr, hw_blk_offset,
    681  1.1  hikaru 			    desc->qcd_slices[i + 1]);
    682  1.1  hikaru 			req_params_sz += sizeof(struct fw_la_auth_req_params);
    683  1.1  hikaru 			la_cmd_flags |= LA_FLAGS_RET_AUTH_RES;
    684  1.1  hikaru 			/* no digest verify */
    685  1.1  hikaru 			break;
    686  1.1  hikaru 		case FW_SLICE_DRAM_WR:
    687  1.1  hikaru 			i = MAX_FW_SLICE; /* end of chain */
    688  1.1  hikaru 			break;
    689  1.1  hikaru 		default:
    690  1.1  hikaru 			KASSERT(0);
    691  1.1  hikaru 			break;
    692  1.1  hikaru 		}
    693  1.1  hikaru 	}
    694  1.1  hikaru 
    695  1.1  hikaru 	desc->qcd_hdr_sz = ctrl_blk_offset / 8;
    696  1.1  hikaru 	desc->qcd_hw_blk_sz = hw_blk_offset / 8;
    697  1.1  hikaru 
    698  1.1  hikaru 	req_cache = (struct fw_la_bulk_req *)desc->qcd_req_cache;
    699  1.1  hikaru 	qat_msg_req_type_populate(
    700  1.1  hikaru 	    &req_cache->comn_hdr.arch_if,
    701  1.1  hikaru 	    ARCH_IF_REQ_QAT_FW_LA, 0);
    702  1.1  hikaru 
    703  1.1  hikaru 	la_cmd_flags |= LA_FLAGS_PROTO_NO;
    704  1.1  hikaru 
    705  1.1  hikaru 	qat_msg_params_populate(req_cache,
    706  1.1  hikaru 	    desc, req_params_sz, la_cmd_flags, 0);
    707  1.1  hikaru 
    708  1.1  hikaru #ifdef QAT_DUMP
    709  1.1  hikaru 	qat_dump_raw(QAT_DUMP_DESC, "qcd_content_desc",
    710  1.1  hikaru 	    desc->qcd_content_desc, sizeof(desc->qcd_content_desc));
    711  1.1  hikaru 	qat_dump_raw(QAT_DUMP_DESC, "qcd_req_cache",
    712  1.1  hikaru 	    &desc->qcd_req_cache, sizeof(desc->qcd_req_cache));
    713  1.1  hikaru #endif
    714  1.1  hikaru 
    715  1.1  hikaru 	bus_dmamap_sync(qcy->qcy_sc->sc_dmat,
    716  1.1  hikaru 	    qcy->qcy_session_dmamems[qs->qs_lid].qdm_dma_map, 0,
    717  1.1  hikaru 	    sizeof(struct qat_session),
    718  1.1  hikaru 	    BUS_DMASYNC_PREWRITE);
    719  1.1  hikaru }
    720  1.1  hikaru 
    721  1.1  hikaru void
    722  1.1  hikaru qat_hw15_crypto_setup_req_params(struct qat_crypto_bank *qcb, struct qat_session *qs,
    723  1.1  hikaru     struct qat_crypto_desc const *desc, struct qat_sym_cookie *qsc,
    724  1.1  hikaru     struct cryptodesc *crde, struct cryptodesc *crda, bus_addr_t icv_paddr)
    725  1.1  hikaru {
    726  1.1  hikaru 	struct qat_sym_bulk_cookie *qsbc;
    727  1.1  hikaru 	struct fw_la_bulk_req *bulk_req;
    728  1.1  hikaru 	struct fw_la_cipher_req_params *cipher_req;
    729  1.1  hikaru 	struct fw_la_auth_req_params *auth_req;
    730  1.1  hikaru 	uint32_t req_params_offset = 0;
    731  1.1  hikaru 	uint8_t *req_params_ptr;
    732  1.1  hikaru 	enum fw_la_cmd_id cmd_id = desc->qcd_cmd_id;
    733  1.1  hikaru 	enum fw_slice next_slice;
    734  1.1  hikaru 
    735  1.1  hikaru 	qsbc = &qsc->u.qsc_bulk_cookie;
    736  1.1  hikaru 
    737  1.1  hikaru 	bulk_req = (struct fw_la_bulk_req *)qsbc->qsbc_msg;
    738  1.1  hikaru 	memcpy(bulk_req, &desc->qcd_req_cache, QAT_HW15_SESSION_REQ_CACHE_SIZE);
    739  1.1  hikaru 	bulk_req->comn_hdr.arch_if.resp_pipe_id = qcb->qcb_sym_rx->qr_ring_id;
    740  1.1  hikaru 	bulk_req->comn_hdr.comn_req_flags =
    741  1.1  hikaru 	    qat_hw15_get_comn_req_flags(qcb->qcb_bank % 2);
    742  1.1  hikaru 	bulk_req->comn_mid.src_data_addr = qsc->qsc_buffer_list_desc_paddr;
    743  1.1  hikaru 	bulk_req->comn_mid.dest_data_addr = qsc->qsc_buffer_list_desc_paddr;
    744  1.1  hikaru 	bulk_req->req_params_addr = qsc->qsc_bulk_req_params_buf_paddr;
    745  1.1  hikaru 	bulk_req->comn_ftr.next_request_addr = 0;
    746  1.1  hikaru 	bulk_req->comn_mid.opaque_data = (uint64_t)(uintptr_t)qsc;
    747  1.1  hikaru 
    748  1.1  hikaru 	if (icv_paddr != 0)
    749  1.1  hikaru 		bulk_req->comn_la_req.u.la_flags |= LA_FLAGS_DIGEST_IN_BUFFER;
    750  1.1  hikaru 
    751  1.1  hikaru 	req_params_ptr = qsbc->qsbc_req_params_buf;
    752  1.1  hikaru 
    753  1.1  hikaru 	if (cmd_id != FW_LA_CMD_AUTH) {
    754  1.1  hikaru 		cipher_req = (struct fw_la_cipher_req_params *)
    755  1.1  hikaru 		    (req_params_ptr + req_params_offset);
    756  1.1  hikaru 		req_params_offset += sizeof(struct fw_la_cipher_req_params);
    757  1.1  hikaru 
    758  1.1  hikaru 		if (cmd_id == FW_LA_CMD_CIPHER || cmd_id == FW_LA_CMD_HASH_CIPHER)
    759  1.1  hikaru 			next_slice = FW_SLICE_DRAM_WR;
    760  1.1  hikaru 		else
    761  1.1  hikaru 			next_slice = FW_SLICE_AUTH;
    762  1.1  hikaru 
    763  1.1  hikaru 		cipher_req->resrvd = 0;
    764  1.1  hikaru 
    765  1.1  hikaru 		cipher_req->cipher_state_sz = desc->qcd_cipher_blk_sz / 8;
    766  1.1  hikaru 
    767  1.1  hikaru 		cipher_req->curr_id = FW_SLICE_CIPHER;
    768  1.1  hikaru 		cipher_req->next_id = next_slice;
    769  1.1  hikaru 
    770  1.1  hikaru 		cipher_req->resrvd1 = 0;
    771  1.1  hikaru 
    772  1.1  hikaru 		cipher_req->cipher_off = crde->crd_skip;
    773  1.1  hikaru 		cipher_req->cipher_len = crde->crd_len;
    774  1.1  hikaru 		cipher_req->state_address = qsc->qsc_iv_buf_paddr;
    775  1.1  hikaru 
    776  1.1  hikaru 	}
    777  1.1  hikaru 	if (cmd_id != FW_LA_CMD_CIPHER) {
    778  1.1  hikaru 		auth_req = (struct fw_la_auth_req_params *)
    779  1.1  hikaru 		    (req_params_ptr + req_params_offset);
    780  1.1  hikaru 		req_params_offset += sizeof(struct fw_la_auth_req_params);
    781  1.1  hikaru 
    782  1.1  hikaru 		if (cmd_id == FW_LA_CMD_HASH_CIPHER)
    783  1.1  hikaru 			next_slice = FW_SLICE_CIPHER;
    784  1.1  hikaru 		else
    785  1.1  hikaru 			next_slice = FW_SLICE_DRAM_WR;
    786  1.1  hikaru 
    787  1.1  hikaru 		auth_req->next_id = next_slice;
    788  1.1  hikaru 		auth_req->curr_id = FW_SLICE_AUTH;
    789  1.1  hikaru 
    790  1.1  hikaru 		auth_req->auth_res_address = icv_paddr;
    791  1.1  hikaru 		auth_req->auth_res_sz = 0; /* no digest verify */
    792  1.1  hikaru 
    793  1.1  hikaru 		auth_req->auth_len = crda->crd_len;
    794  1.1  hikaru 		auth_req->auth_off = crda->crd_skip;
    795  1.1  hikaru 
    796  1.1  hikaru 		auth_req->hash_state_sz = 0;
    797  1.1  hikaru 		auth_req->u1.prefix_addr = desc->qcd_hash_state_paddr +
    798  1.1  hikaru 		    desc->qcd_state_storage_sz;
    799  1.1  hikaru 
    800  1.1  hikaru 		auth_req->u.resrvd = 0;
    801  1.1  hikaru 	}
    802  1.1  hikaru 
    803  1.1  hikaru #ifdef QAT_DUMP
    804  1.1  hikaru 	qat_dump_raw(QAT_DUMP_DESC, "req_params", req_params_ptr, req_params_offset);
    805  1.1  hikaru #endif
    806  1.1  hikaru }
    807  1.1  hikaru 
    808