Home | History | Annotate | Line # | Download | only in qat
      1  1.1  hikaru /*	$NetBSD: qat_hw17.c,v 1.1 2019/11/20 09:37:46 hikaru Exp $	*/
      2  1.1  hikaru 
      3  1.1  hikaru /*
      4  1.1  hikaru  * Copyright (c) 2019 Internet Initiative Japan, Inc.
      5  1.1  hikaru  * All rights reserved.
      6  1.1  hikaru  *
      7  1.1  hikaru  * Redistribution and use in source and binary forms, with or without
      8  1.1  hikaru  * modification, are permitted provided that the following conditions
      9  1.1  hikaru  * are met:
     10  1.1  hikaru  * 1. Redistributions of source code must retain the above copyright
     11  1.1  hikaru  *    notice, this list of conditions and the following disclaimer.
     12  1.1  hikaru  * 2. Redistributions in binary form must reproduce the above copyright
     13  1.1  hikaru  *    notice, this list of conditions and the following disclaimer in the
     14  1.1  hikaru  *    documentation and/or other materials provided with the distribution.
     15  1.1  hikaru  *
     16  1.1  hikaru  * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
     17  1.1  hikaru  * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
     18  1.1  hikaru  * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
     19  1.1  hikaru  * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
     20  1.1  hikaru  * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
     21  1.1  hikaru  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
     22  1.1  hikaru  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
     23  1.1  hikaru  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
     24  1.1  hikaru  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
     25  1.1  hikaru  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
     26  1.1  hikaru  * POSSIBILITY OF SUCH DAMAGE.
     27  1.1  hikaru  */
     28  1.1  hikaru 
     29  1.1  hikaru /*
     30  1.1  hikaru  *   Copyright(c) 2014 Intel Corporation.
     31  1.1  hikaru  *   Redistribution and use in source and binary forms, with or without
     32  1.1  hikaru  *   modification, are permitted provided that the following conditions
     33  1.1  hikaru  *   are met:
     34  1.1  hikaru  *
     35  1.1  hikaru  *     * Redistributions of source code must retain the above copyright
     36  1.1  hikaru  *       notice, this list of conditions and the following disclaimer.
     37  1.1  hikaru  *     * Redistributions in binary form must reproduce the above copyright
     38  1.1  hikaru  *       notice, this list of conditions and the following disclaimer in
     39  1.1  hikaru  *       the documentation and/or other materials provided with the
     40  1.1  hikaru  *       distribution.
     41  1.1  hikaru  *     * Neither the name of Intel Corporation nor the names of its
     42  1.1  hikaru  *       contributors may be used to endorse or promote products derived
     43  1.1  hikaru  *       from this software without specific prior written permission.
     44  1.1  hikaru  *
     45  1.1  hikaru  *   THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
     46  1.1  hikaru  *   "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
     47  1.1  hikaru  *   LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
     48  1.1  hikaru  *   A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
     49  1.1  hikaru  *   OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
     50  1.1  hikaru  *   SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
     51  1.1  hikaru  *   LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
     52  1.1  hikaru  *   DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
     53  1.1  hikaru  *   THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
     54  1.1  hikaru  *   (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
     55  1.1  hikaru  *   OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
     56  1.1  hikaru  */
     57  1.1  hikaru 
     58  1.1  hikaru #include <sys/cdefs.h>
     59  1.1  hikaru __KERNEL_RCSID(0, "$NetBSD: qat_hw17.c,v 1.1 2019/11/20 09:37:46 hikaru Exp $");
     60  1.1  hikaru 
     61  1.1  hikaru #include <sys/param.h>
     62  1.1  hikaru #include <sys/systm.h>
     63  1.1  hikaru #include <sys/proc.h>
     64  1.1  hikaru 
     65  1.1  hikaru #include <opencrypto/xform.h>
     66  1.1  hikaru 
     67  1.1  hikaru /* XXX same as sys/arch/x86/x86/via_padlock.c */
     68  1.1  hikaru #include <opencrypto/cryptosoft_xform.c>
     69  1.1  hikaru 
     70  1.1  hikaru #include <dev/pci/pcireg.h>
     71  1.1  hikaru #include <dev/pci/pcivar.h>
     72  1.1  hikaru 
     73  1.1  hikaru #include "qatreg.h"
     74  1.1  hikaru #include "qat_hw17reg.h"
     75  1.1  hikaru #include "qatvar.h"
     76  1.1  hikaru #include "qat_hw17var.h"
     77  1.1  hikaru 
     78  1.1  hikaru int		qat_adm_mailbox_put_msg_sync(struct qat_softc *, uint32_t,
     79  1.1  hikaru 		    void *, void *);
     80  1.1  hikaru int		qat_adm_mailbox_send(struct qat_softc *,
     81  1.1  hikaru 		    struct fw_init_admin_req *, struct fw_init_admin_resp *);
     82  1.1  hikaru int		qat_adm_mailbox_send_init_me(struct qat_softc *);
     83  1.1  hikaru int		qat_adm_mailbox_send_hb_timer(struct qat_softc *);
     84  1.1  hikaru int		qat_adm_mailbox_send_fw_status(struct qat_softc *);
     85  1.1  hikaru int		qat_adm_mailbox_send_constants(struct qat_softc *);
     86  1.1  hikaru 
     87  1.1  hikaru uint32_t	qat_hw17_crypto_setup_cipher_desc(struct qat_session *,
     88  1.1  hikaru 		    struct qat_crypto_desc *, struct cryptoini *,
     89  1.1  hikaru 		    union hw_cipher_algo_blk *, uint32_t, struct fw_la_bulk_req *,
     90  1.1  hikaru 		    enum fw_slice);
     91  1.1  hikaru uint32_t	qat_hw17_crypto_setup_auth_desc(struct qat_session *,
     92  1.1  hikaru 		    struct qat_crypto_desc *, struct cryptoini *,
     93  1.1  hikaru 		    union hw_auth_algo_blk *, uint32_t, struct fw_la_bulk_req *,
     94  1.1  hikaru 		    enum fw_slice);
     95  1.1  hikaru void		qat_hw17_init_comn_req_hdr(struct qat_crypto_desc *,
     96  1.1  hikaru 		    struct fw_la_bulk_req *);
     97  1.1  hikaru 
     98  1.1  hikaru int
     99  1.1  hikaru qat_adm_mailbox_init(struct qat_softc *sc)
    100  1.1  hikaru {
    101  1.1  hikaru 	uint64_t addr;
    102  1.1  hikaru 	int error;
    103  1.1  hikaru 	struct qat_dmamem *qdm;
    104  1.1  hikaru 
    105  1.1  hikaru 	error = qat_alloc_dmamem(sc, &sc->sc_admin_comms.qadc_dma,
    106  1.1  hikaru 	    PAGE_SIZE, PAGE_SIZE);
    107  1.1  hikaru 	if (error)
    108  1.1  hikaru 		return error;
    109  1.1  hikaru 
    110  1.1  hikaru 	qdm = &sc->sc_admin_comms.qadc_const_tbl_dma;
    111  1.1  hikaru 	error = qat_alloc_dmamem(sc, qdm, PAGE_SIZE, PAGE_SIZE);
    112  1.1  hikaru 	if (error)
    113  1.1  hikaru 		return error;
    114  1.1  hikaru 
    115  1.1  hikaru 	memcpy(qdm->qdm_dma_vaddr,
    116  1.1  hikaru 	    mailbox_const_tab, sizeof(mailbox_const_tab));
    117  1.1  hikaru 
    118  1.1  hikaru 	bus_dmamap_sync(sc->sc_dmat, qdm->qdm_dma_map, 0,
    119  1.1  hikaru 	    qdm->qdm_dma_map->dm_mapsize, BUS_DMASYNC_PREWRITE);
    120  1.1  hikaru 
    121  1.1  hikaru 	error = qat_alloc_dmamem(sc, &sc->sc_admin_comms.qadc_hb_dma,
    122  1.1  hikaru 	    PAGE_SIZE, PAGE_SIZE);
    123  1.1  hikaru 	if (error)
    124  1.1  hikaru 		return error;
    125  1.1  hikaru 
    126  1.1  hikaru 	addr = (uint64_t)sc->sc_admin_comms.qadc_dma.qdm_dma_seg.ds_addr;
    127  1.1  hikaru 	qat_misc_write_4(sc, ADMINMSGUR, addr >> 32);
    128  1.1  hikaru 	qat_misc_write_4(sc, ADMINMSGLR, addr);
    129  1.1  hikaru 
    130  1.1  hikaru 	return 0;
    131  1.1  hikaru }
    132  1.1  hikaru 
    133  1.1  hikaru int
    134  1.1  hikaru qat_adm_mailbox_put_msg_sync(struct qat_softc *sc, uint32_t ae,
    135  1.1  hikaru     void *in, void *out)
    136  1.1  hikaru {
    137  1.1  hikaru 	uint32_t mailbox;
    138  1.1  hikaru 	bus_size_t mb_offset = MAILBOX_BASE + (ae * MAILBOX_STRIDE);
    139  1.1  hikaru 	int offset = ae * ADMINMSG_LEN * 2;
    140  1.1  hikaru 	int times, received;
    141  1.1  hikaru 	uint8_t *buf = (uint8_t *)sc->sc_admin_comms.qadc_dma.qdm_dma_vaddr + offset;
    142  1.1  hikaru 
    143  1.1  hikaru 	mailbox = qat_misc_read_4(sc, mb_offset);
    144  1.1  hikaru 	if (mailbox == 1)
    145  1.1  hikaru 		return EAGAIN;
    146  1.1  hikaru 
    147  1.1  hikaru 	memcpy(buf, in, ADMINMSG_LEN);
    148  1.1  hikaru 	bus_dmamap_sync(sc->sc_dmat, sc->sc_admin_comms.qadc_dma.qdm_dma_map, 0,
    149  1.1  hikaru 	    sc->sc_admin_comms.qadc_dma.qdm_dma_map->dm_mapsize,
    150  1.1  hikaru 	    BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
    151  1.1  hikaru 	qat_misc_write_4(sc, mb_offset, 1);
    152  1.1  hikaru 
    153  1.1  hikaru 	received = 0;
    154  1.1  hikaru 	for (times = 0; times < 50; times++) {
    155  1.1  hikaru 		delay(20000);
    156  1.1  hikaru 		if (qat_misc_read_4(sc, mb_offset) == 0) {
    157  1.1  hikaru 			received = 1;
    158  1.1  hikaru 			break;
    159  1.1  hikaru 		}
    160  1.1  hikaru 	}
    161  1.1  hikaru 	if (received) {
    162  1.1  hikaru 		bus_dmamap_sync(sc->sc_dmat, sc->sc_admin_comms.qadc_dma.qdm_dma_map, 0,
    163  1.1  hikaru 		    sc->sc_admin_comms.qadc_dma.qdm_dma_map->dm_mapsize,
    164  1.1  hikaru 		    BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
    165  1.1  hikaru 		memcpy(out, buf + ADMINMSG_LEN, ADMINMSG_LEN);
    166  1.1  hikaru 	} else {
    167  1.1  hikaru 		aprint_error_dev(sc->sc_dev,
    168  1.1  hikaru 		    "Failed to send admin msg to accelerator\n");
    169  1.1  hikaru 	}
    170  1.1  hikaru 
    171  1.1  hikaru 	return received ? 0 : EFAULT;
    172  1.1  hikaru }
    173  1.1  hikaru 
    174  1.1  hikaru int
    175  1.1  hikaru qat_adm_mailbox_send(struct qat_softc *sc,
    176  1.1  hikaru     struct fw_init_admin_req *req, struct fw_init_admin_resp *resp)
    177  1.1  hikaru {
    178  1.1  hikaru 	int error;
    179  1.1  hikaru 	uint32_t mask;
    180  1.1  hikaru 	uint8_t ae;
    181  1.1  hikaru 
    182  1.1  hikaru 	for (ae = 0, mask = sc->sc_ae_mask; mask; ae++, mask >>= 1) {
    183  1.1  hikaru 		if (!(mask & 1))
    184  1.1  hikaru 			continue;
    185  1.1  hikaru 
    186  1.1  hikaru 		error = qat_adm_mailbox_put_msg_sync(sc, ae, req, resp);
    187  1.1  hikaru 		if (error)
    188  1.1  hikaru 			return error;
    189  1.1  hikaru 		if (resp->init_resp_hdr.status) {
    190  1.1  hikaru 			aprint_error_dev(sc->sc_dev,
    191  1.1  hikaru 			    "Failed to send admin msg: cmd %d\n",
    192  1.1  hikaru 			    req->init_admin_cmd_id);
    193  1.1  hikaru 			return EFAULT;
    194  1.1  hikaru 		}
    195  1.1  hikaru 	}
    196  1.1  hikaru 
    197  1.1  hikaru 	return 0;
    198  1.1  hikaru }
    199  1.1  hikaru 
    200  1.1  hikaru int
    201  1.1  hikaru qat_adm_mailbox_send_init_me(struct qat_softc *sc)
    202  1.1  hikaru {
    203  1.1  hikaru 	struct fw_init_admin_req req;
    204  1.1  hikaru 	struct fw_init_admin_resp resp;
    205  1.1  hikaru 
    206  1.1  hikaru 	memset(&req, 0, sizeof(req));
    207  1.1  hikaru 	req.init_admin_cmd_id = FW_INIT_ME;
    208  1.1  hikaru 
    209  1.1  hikaru 	return qat_adm_mailbox_send(sc, &req, &resp);
    210  1.1  hikaru }
    211  1.1  hikaru 
    212  1.1  hikaru int
    213  1.1  hikaru qat_adm_mailbox_send_hb_timer(struct qat_softc *sc)
    214  1.1  hikaru {
    215  1.1  hikaru 	struct fw_init_admin_req req;
    216  1.1  hikaru 	struct fw_init_admin_resp resp;
    217  1.1  hikaru 
    218  1.1  hikaru 	memset(&req, 0, sizeof(req));
    219  1.1  hikaru 	req.init_admin_cmd_id = FW_HEARTBEAT_TIMER_SET;
    220  1.1  hikaru 
    221  1.1  hikaru 	req.init_cfg_ptr = sc->sc_admin_comms.qadc_hb_dma.qdm_dma_seg.ds_addr;
    222  1.1  hikaru 	req.heartbeat_ticks =
    223  1.1  hikaru 	    sc->sc_hw.qhw_clock_per_sec / 1000 * QAT_HB_INTERVAL;
    224  1.1  hikaru 
    225  1.1  hikaru 	return qat_adm_mailbox_send(sc, &req, &resp);
    226  1.1  hikaru }
    227  1.1  hikaru 
    228  1.1  hikaru int
    229  1.1  hikaru qat_adm_mailbox_send_fw_status(struct qat_softc *sc)
    230  1.1  hikaru {
    231  1.1  hikaru 	int error;
    232  1.1  hikaru 	struct fw_init_admin_req req;
    233  1.1  hikaru 	struct fw_init_admin_resp resp;
    234  1.1  hikaru 
    235  1.1  hikaru 	memset(&req, 0, sizeof(req));
    236  1.1  hikaru 	req.init_admin_cmd_id = FW_STATUS_GET;
    237  1.1  hikaru 
    238  1.1  hikaru 	error = qat_adm_mailbox_send(sc, &req, &resp);
    239  1.1  hikaru 	if (error)
    240  1.1  hikaru 		return error;
    241  1.1  hikaru 
    242  1.1  hikaru 	aprint_normal_dev(sc->sc_dev,
    243  1.1  hikaru 	    "loaded firmware: version %d.%d.%d\n",
    244  1.1  hikaru 	    resp.u.s.version_major_num,
    245  1.1  hikaru 	    resp.u.s.version_minor_num,
    246  1.1  hikaru 	    resp.init_resp_pars.u.s1.version_patch_num);
    247  1.1  hikaru 
    248  1.1  hikaru 	return 0;
    249  1.1  hikaru }
    250  1.1  hikaru 
    251  1.1  hikaru int
    252  1.1  hikaru qat_adm_mailbox_send_constants(struct qat_softc *sc)
    253  1.1  hikaru {
    254  1.1  hikaru 	struct fw_init_admin_req req;
    255  1.1  hikaru 	struct fw_init_admin_resp resp;
    256  1.1  hikaru 
    257  1.1  hikaru 	memset(&req, 0, sizeof(req));
    258  1.1  hikaru 	req.init_admin_cmd_id = FW_CONSTANTS_CFG;
    259  1.1  hikaru 
    260  1.1  hikaru 	req.init_cfg_sz = 1024;
    261  1.1  hikaru 	req.init_cfg_ptr =
    262  1.1  hikaru 	    sc->sc_admin_comms.qadc_const_tbl_dma.qdm_dma_seg.ds_addr;
    263  1.1  hikaru 
    264  1.1  hikaru 	return qat_adm_mailbox_send(sc, &req, &resp);
    265  1.1  hikaru }
    266  1.1  hikaru 
    267  1.1  hikaru int
    268  1.1  hikaru qat_adm_mailbox_send_init(struct qat_softc *sc)
    269  1.1  hikaru {
    270  1.1  hikaru 	int error;
    271  1.1  hikaru 
    272  1.1  hikaru 	error = qat_adm_mailbox_send_init_me(sc);
    273  1.1  hikaru 	if (error)
    274  1.1  hikaru 		return error;
    275  1.1  hikaru 
    276  1.1  hikaru 	error = qat_adm_mailbox_send_hb_timer(sc);
    277  1.1  hikaru 	if (error)
    278  1.1  hikaru 		return error;
    279  1.1  hikaru 
    280  1.1  hikaru 	error = qat_adm_mailbox_send_fw_status(sc);
    281  1.1  hikaru 	if (error)
    282  1.1  hikaru 		return error;
    283  1.1  hikaru 
    284  1.1  hikaru 	return qat_adm_mailbox_send_constants(sc);
    285  1.1  hikaru }
    286  1.1  hikaru 
    287  1.1  hikaru int
    288  1.1  hikaru qat_arb_init(struct qat_softc *sc)
    289  1.1  hikaru {
    290  1.1  hikaru 	uint32_t arb_cfg = 0x1 << 31 | 0x4 << 4 | 0x1;
    291  1.1  hikaru 	uint32_t arb, i;
    292  1.1  hikaru 	const uint32_t *thd_2_arb_cfg;
    293  1.1  hikaru 
    294  1.1  hikaru 	/* Service arb configured for 32 bytes responses and
    295  1.1  hikaru 	 * ring flow control check enabled. */
    296  1.1  hikaru 	for (arb = 0; arb < MAX_ARB; arb++)
    297  1.1  hikaru 		qat_arb_sarconfig_write_4(sc, arb, arb_cfg);
    298  1.1  hikaru 
    299  1.1  hikaru 	/* Map worker threads to service arbiters */
    300  1.1  hikaru 	sc->sc_hw.qhw_get_arb_mapping(sc, &thd_2_arb_cfg);
    301  1.1  hikaru 
    302  1.1  hikaru 	if (!thd_2_arb_cfg)
    303  1.1  hikaru 		return EINVAL;
    304  1.1  hikaru 
    305  1.1  hikaru 	for (i = 0; i < sc->sc_hw.qhw_num_engines; i++)
    306  1.1  hikaru 		qat_arb_wrk_2_ser_map_write_4(sc, i, *(thd_2_arb_cfg + i));
    307  1.1  hikaru 
    308  1.1  hikaru 	return 0;
    309  1.1  hikaru }
    310  1.1  hikaru 
    311  1.1  hikaru int
    312  1.1  hikaru qat_set_ssm_wdtimer(struct qat_softc *sc)
    313  1.1  hikaru {
    314  1.1  hikaru 	uint32_t timer;
    315  1.1  hikaru 	u_int mask;
    316  1.1  hikaru 	int i;
    317  1.1  hikaru 
    318  1.1  hikaru 	timer = sc->sc_hw.qhw_clock_per_sec / 1000 * QAT_SSM_WDT;
    319  1.1  hikaru 	for (i = 0, mask = sc->sc_accel_mask; mask; i++, mask >>= 1) {
    320  1.1  hikaru 		if (!(mask & 1))
    321  1.1  hikaru 			continue;
    322  1.1  hikaru 		qat_misc_write_4(sc, SSMWDT(i), timer);
    323  1.1  hikaru 		qat_misc_write_4(sc, SSMWDTPKE(i), timer);
    324  1.1  hikaru 	}
    325  1.1  hikaru 
    326  1.1  hikaru 	return 0;
    327  1.1  hikaru }
    328  1.1  hikaru 
    329  1.1  hikaru int
    330  1.1  hikaru qat_check_slice_hang(struct qat_softc *sc)
    331  1.1  hikaru {
    332  1.1  hikaru 	int handled = 0;
    333  1.1  hikaru 
    334  1.1  hikaru 	return handled;
    335  1.1  hikaru }
    336  1.1  hikaru 
    337  1.1  hikaru uint32_t
    338  1.1  hikaru qat_hw17_crypto_setup_cipher_desc(struct qat_session *qs,
    339  1.1  hikaru     struct qat_crypto_desc *desc, struct cryptoini *crie,
    340  1.1  hikaru     union hw_cipher_algo_blk *cipher, uint32_t cd_blk_offset,
    341  1.1  hikaru     struct fw_la_bulk_req *req_tmpl, enum fw_slice next_slice)
    342  1.1  hikaru {
    343  1.1  hikaru 	struct fw_cipher_cd_ctrl_hdr *cipher_cd_ctrl =
    344  1.1  hikaru 	    (struct fw_cipher_cd_ctrl_hdr *)&req_tmpl->cd_ctrl;
    345  1.1  hikaru 	int keylen = crie->cri_klen / 8;
    346  1.1  hikaru 
    347  1.1  hikaru 	cipher->max.cipher_config.val =
    348  1.1  hikaru 	    qat_crypto_load_cipher_cryptoini(desc, crie);
    349  1.1  hikaru 	memcpy(cipher->max.key, crie->cri_key, keylen);
    350  1.1  hikaru 
    351  1.1  hikaru 	cipher_cd_ctrl->cipher_state_sz = desc->qcd_cipher_blk_sz >> 3;
    352  1.1  hikaru 	cipher_cd_ctrl->cipher_key_sz = keylen >> 3;
    353  1.1  hikaru 	cipher_cd_ctrl->cipher_cfg_offset = cd_blk_offset >> 3;
    354  1.1  hikaru 	FW_COMN_CURR_ID_SET(cipher_cd_ctrl, FW_SLICE_CIPHER);
    355  1.1  hikaru 	FW_COMN_NEXT_ID_SET(cipher_cd_ctrl, next_slice);
    356  1.1  hikaru 
    357  1.1  hikaru 	return roundup(sizeof(struct hw_cipher_config) + keylen, 8);
    358  1.1  hikaru }
    359  1.1  hikaru 
    360  1.1  hikaru uint32_t
    361  1.1  hikaru qat_hw17_crypto_setup_auth_desc(struct qat_session *qs,
    362  1.1  hikaru     struct qat_crypto_desc *desc, struct cryptoini *cria,
    363  1.1  hikaru     union hw_auth_algo_blk *auth, uint32_t cd_blk_offset,
    364  1.1  hikaru     struct fw_la_bulk_req *req_tmpl, enum fw_slice next_slice)
    365  1.1  hikaru {
    366  1.1  hikaru 	struct fw_auth_cd_ctrl_hdr *auth_cd_ctrl =
    367  1.1  hikaru 	    (struct fw_auth_cd_ctrl_hdr *)&req_tmpl->cd_ctrl;
    368  1.1  hikaru 	struct qat_sym_hash_def const *hash_def;
    369  1.1  hikaru 	uint8_t *state1, *state2;
    370  1.1  hikaru 
    371  1.1  hikaru 	auth->max.inner_setup.auth_config.config =
    372  1.1  hikaru 	    qat_crypto_load_auth_cryptoini(desc, cria, &hash_def);
    373  1.1  hikaru 	auth->max.inner_setup.auth_counter.counter =
    374  1.1  hikaru 	    htonl(hash_def->qshd_qat->qshqi_auth_counter);
    375  1.1  hikaru 
    376  1.1  hikaru 	auth_cd_ctrl->hash_cfg_offset = cd_blk_offset >> 3;
    377  1.1  hikaru 	auth_cd_ctrl->hash_flags = FW_AUTH_HDR_FLAG_NO_NESTED;
    378  1.1  hikaru 	auth_cd_ctrl->inner_res_sz = hash_def->qshd_alg->qshai_digest_len;
    379  1.1  hikaru 	auth_cd_ctrl->final_sz = desc->qcd_auth_sz;
    380  1.1  hikaru 
    381  1.1  hikaru 	auth_cd_ctrl->inner_state1_sz =
    382  1.1  hikaru 	    roundup(hash_def->qshd_qat->qshqi_state1_len, 8);
    383  1.1  hikaru 	auth_cd_ctrl->inner_state2_sz =
    384  1.1  hikaru 	    roundup(hash_def->qshd_qat->qshqi_state2_len, 8);
    385  1.1  hikaru 	auth_cd_ctrl->inner_state2_offset =
    386  1.1  hikaru 	    auth_cd_ctrl->hash_cfg_offset +
    387  1.1  hikaru 	    ((sizeof(struct hw_auth_setup) +
    388  1.1  hikaru 	    auth_cd_ctrl->inner_state1_sz) >> 3);
    389  1.1  hikaru 
    390  1.1  hikaru 	state1 = auth->max.state1;
    391  1.1  hikaru 	state2 = auth->max.state1 + auth_cd_ctrl->inner_state1_sz;
    392  1.1  hikaru 	qat_crypto_hmac_precompute(desc, cria, hash_def, state1, state2);
    393  1.1  hikaru 
    394  1.1  hikaru 	FW_COMN_CURR_ID_SET(auth_cd_ctrl, FW_SLICE_AUTH);
    395  1.1  hikaru 	FW_COMN_NEXT_ID_SET(auth_cd_ctrl, next_slice);
    396  1.1  hikaru 
    397  1.1  hikaru 	return roundup(auth_cd_ctrl->inner_state1_sz +
    398  1.1  hikaru 	    auth_cd_ctrl->inner_state2_sz +
    399  1.1  hikaru 	    sizeof(struct hw_auth_setup), 8);
    400  1.1  hikaru }
    401  1.1  hikaru 
    402  1.1  hikaru void
    403  1.1  hikaru qat_hw17_init_comn_req_hdr(struct qat_crypto_desc *desc,
    404  1.1  hikaru     struct fw_la_bulk_req *req)
    405  1.1  hikaru {
    406  1.1  hikaru 	union fw_comn_req_hdr_cd_pars *cd_pars = &req->cd_pars;
    407  1.1  hikaru 	struct fw_comn_req_hdr *req_hdr = &req->comn_hdr;
    408  1.1  hikaru 
    409  1.1  hikaru 	req_hdr->service_cmd_id = desc->qcd_cmd_id;
    410  1.1  hikaru 	req_hdr->hdr_flags = FW_COMN_VALID;
    411  1.1  hikaru 	req_hdr->service_type = FW_COMN_REQ_CPM_FW_LA;
    412  1.1  hikaru 	req_hdr->comn_req_flags = FW_COMN_FLAGS_BUILD(
    413  1.1  hikaru 	    COMN_CD_FLD_TYPE_64BIT_ADR, COMN_PTR_TYPE_SGL);
    414  1.1  hikaru 	req_hdr->serv_specif_flags = 0;
    415  1.1  hikaru 	cd_pars->s.content_desc_addr = desc->qcd_desc_paddr;
    416  1.1  hikaru }
    417  1.1  hikaru 
    418  1.1  hikaru void
    419  1.1  hikaru qat_hw17_crypto_setup_desc(struct qat_crypto *qcy, struct qat_session *qs,
    420  1.1  hikaru     struct qat_crypto_desc *desc,
    421  1.1  hikaru     struct cryptoini *crie, struct cryptoini *cria)
    422  1.1  hikaru {
    423  1.1  hikaru 	union hw_cipher_algo_blk *cipher;
    424  1.1  hikaru 	union hw_auth_algo_blk *auth;
    425  1.1  hikaru 	struct fw_la_bulk_req *req_tmpl;
    426  1.1  hikaru 	struct fw_comn_req_hdr *req_hdr;
    427  1.1  hikaru 	union fw_comn_req_hdr_cd_pars *cd_pars;
    428  1.1  hikaru 	uint32_t cd_blk_offset = 0;
    429  1.1  hikaru 	int i;
    430  1.1  hikaru 	uint8_t *cd_blk_ptr;
    431  1.1  hikaru 
    432  1.1  hikaru 	req_tmpl = (struct fw_la_bulk_req *)desc->qcd_req_cache;
    433  1.1  hikaru 	req_hdr = &req_tmpl->comn_hdr;
    434  1.1  hikaru 	cd_pars = &req_tmpl->cd_pars;
    435  1.1  hikaru 	cd_blk_ptr = desc->qcd_content_desc;
    436  1.1  hikaru 
    437  1.1  hikaru 	memset(req_tmpl, 0, sizeof(struct fw_la_bulk_req));
    438  1.1  hikaru 	qat_hw17_init_comn_req_hdr(desc, req_tmpl);
    439  1.1  hikaru 
    440  1.1  hikaru 	for (i = 0; i < MAX_FW_SLICE; i++) {
    441  1.1  hikaru 		switch (desc->qcd_slices[i]) {
    442  1.1  hikaru 		case FW_SLICE_CIPHER:
    443  1.1  hikaru 			cipher = (union hw_cipher_algo_blk *)(cd_blk_ptr +
    444  1.1  hikaru 			    cd_blk_offset);
    445  1.1  hikaru 			cd_blk_offset += qat_hw17_crypto_setup_cipher_desc(
    446  1.1  hikaru 			    qs, desc, crie, cipher, cd_blk_offset, req_tmpl,
    447  1.1  hikaru 			    desc->qcd_slices[i + 1]);
    448  1.1  hikaru 			break;
    449  1.1  hikaru 		case FW_SLICE_AUTH:
    450  1.1  hikaru 			auth = (union hw_auth_algo_blk *)(cd_blk_ptr +
    451  1.1  hikaru 			    cd_blk_offset);
    452  1.1  hikaru 			cd_blk_offset += qat_hw17_crypto_setup_auth_desc(
    453  1.1  hikaru 			    qs, desc, cria, auth, cd_blk_offset, req_tmpl,
    454  1.1  hikaru 			    desc->qcd_slices[i + 1]);
    455  1.1  hikaru 			req_hdr->serv_specif_flags |= FW_LA_RET_AUTH_RES;
    456  1.1  hikaru 			/* no digest verify */
    457  1.1  hikaru 			break;
    458  1.1  hikaru 		case FW_SLICE_DRAM_WR:
    459  1.1  hikaru 			i = MAX_FW_SLICE; /* end of chain */
    460  1.1  hikaru 			break;
    461  1.1  hikaru 		default:
    462  1.1  hikaru 			KASSERT(0);
    463  1.1  hikaru 			break;
    464  1.1  hikaru 		}
    465  1.1  hikaru 	}
    466  1.1  hikaru 
    467  1.1  hikaru 	cd_pars->s.content_desc_params_sz =
    468  1.1  hikaru 	    roundup(cd_blk_offset, QAT_OPTIMAL_ALIGN) >> 3;
    469  1.1  hikaru 
    470  1.1  hikaru #ifdef QAT_DUMP
    471  1.1  hikaru 	qat_dump_raw(QAT_DUMP_DESC, "qcd_content_desc",
    472  1.1  hikaru 	    desc->qcd_content_desc, cd_pars->s.content_desc_params_sz << 3);
    473  1.1  hikaru 	qat_dump_raw(QAT_DUMP_DESC, "qcd_req_cache",
    474  1.1  hikaru 	    &desc->qcd_req_cache, sizeof(desc->qcd_req_cache));
    475  1.1  hikaru #endif
    476  1.1  hikaru 
    477  1.1  hikaru 	bus_dmamap_sync(qcy->qcy_sc->sc_dmat,
    478  1.1  hikaru 	    qcy->qcy_session_dmamems[qs->qs_lid].qdm_dma_map, 0,
    479  1.1  hikaru 	    sizeof(struct qat_session),
    480  1.1  hikaru 	    BUS_DMASYNC_PREWRITE);
    481  1.1  hikaru }
    482  1.1  hikaru 
    483  1.1  hikaru void
    484  1.1  hikaru qat_hw17_crypto_setup_req_params(struct qat_crypto_bank *qcb, struct qat_session *qs,
    485  1.1  hikaru     struct qat_crypto_desc const *desc, struct qat_sym_cookie *qsc,
    486  1.1  hikaru     struct cryptodesc *crde, struct cryptodesc *crda, bus_addr_t icv_paddr)
    487  1.1  hikaru {
    488  1.1  hikaru 	struct qat_sym_bulk_cookie *qsbc;
    489  1.1  hikaru 	struct fw_la_bulk_req *bulk_req;
    490  1.1  hikaru 	struct fw_la_cipher_req_params *cipher_param;
    491  1.1  hikaru 	struct fw_la_auth_req_params *auth_param;
    492  1.1  hikaru 	uint32_t req_params_offset = 0;
    493  1.1  hikaru 	uint8_t *req_params_ptr;
    494  1.1  hikaru 	enum fw_la_cmd_id cmd_id = desc->qcd_cmd_id;
    495  1.1  hikaru 
    496  1.1  hikaru 	qsbc = &qsc->u.qsc_bulk_cookie;
    497  1.1  hikaru 	bulk_req = (struct fw_la_bulk_req *)qsbc->qsbc_msg;
    498  1.1  hikaru 
    499  1.1  hikaru 	memcpy(bulk_req, desc->qcd_req_cache, sizeof(struct fw_la_bulk_req));
    500  1.1  hikaru 	bulk_req->comn_mid.opaque_data = (uint64_t)(uintptr_t)qsc;
    501  1.1  hikaru 	bulk_req->comn_mid.src_data_addr = qsc->qsc_buffer_list_desc_paddr;
    502  1.1  hikaru 	bulk_req->comn_mid.dest_data_addr = qsc->qsc_buffer_list_desc_paddr;
    503  1.1  hikaru 
    504  1.1  hikaru 	if (icv_paddr != 0)
    505  1.1  hikaru 		bulk_req->comn_hdr.serv_specif_flags |= FW_LA_DIGEST_IN_BUFFER;
    506  1.1  hikaru 
    507  1.1  hikaru 	req_params_ptr = (uint8_t *)&bulk_req->serv_specif_rqpars;
    508  1.1  hikaru 
    509  1.1  hikaru 	if (cmd_id != FW_LA_CMD_AUTH) {
    510  1.1  hikaru 		cipher_param = (struct fw_la_cipher_req_params *)
    511  1.1  hikaru 		    (req_params_ptr + req_params_offset);
    512  1.1  hikaru 		req_params_offset += sizeof(struct fw_la_cipher_req_params);
    513  1.1  hikaru 
    514  1.1  hikaru 		cipher_param->u.s.cipher_IV_ptr = qsc->qsc_iv_buf_paddr;
    515  1.1  hikaru 		cipher_param->cipher_offset = crde->crd_skip;
    516  1.1  hikaru 		cipher_param->cipher_length = crde->crd_len;
    517  1.1  hikaru 	}
    518  1.1  hikaru 
    519  1.1  hikaru 	if (cmd_id != FW_LA_CMD_CIPHER) {
    520  1.1  hikaru 		auth_param = (struct fw_la_auth_req_params *)
    521  1.1  hikaru 		    (req_params_ptr + req_params_offset);
    522  1.1  hikaru 		req_params_offset += sizeof(struct fw_la_auth_req_params);
    523  1.1  hikaru 
    524  1.1  hikaru 		auth_param->auth_off = crda->crd_skip;
    525  1.1  hikaru 		auth_param->auth_len = crda->crd_len;
    526  1.1  hikaru 		auth_param->auth_res_addr = icv_paddr;
    527  1.1  hikaru 		auth_param->auth_res_sz = 0; /* XXX no digest verify */
    528  1.1  hikaru 		auth_param->hash_state_sz = 0;
    529  1.1  hikaru 		auth_param->u1.auth_partial_st_prefix = 0;
    530  1.1  hikaru 		auth_param->u2.aad_sz = 0;
    531  1.1  hikaru 	}
    532  1.1  hikaru 
    533  1.1  hikaru #ifdef QAT_DUMP
    534  1.1  hikaru 	qat_dump_raw(QAT_DUMP_DESC, "req_params", req_params_ptr, req_params_offset);
    535  1.1  hikaru #endif
    536  1.1  hikaru }
    537  1.1  hikaru 
    538