freebsd-dev/sys/dev/qat_c2xxx/qat_hw15.c
Mark Johnston f4f56ff43d qat: Rename to qat_c2xxx and remove support for modern chipsets
A replacement QAT driver will be imported, but this replacement does not
support Atom C2xxx hardware.  So, the existing driver will be kept
around to provide opencrypto offload support for those chipsets.

Reviewed by:	pauamma, emaste
Sponsored by:	The FreeBSD Foundation
Differential Revision:	https://reviews.freebsd.org/D35817
2022-07-27 11:10:52 -04:00

966 lines
28 KiB
C

/* SPDX-License-Identifier: BSD-2-Clause-NetBSD AND BSD-3-Clause */
/* $NetBSD: qat_hw15.c,v 1.1 2019/11/20 09:37:46 hikaru Exp $ */
/*
* Copyright (c) 2019 Internet Initiative Japan, Inc.
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
* ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
* TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
* BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
* POSSIBILITY OF SUCH DAMAGE.
*/
/*
* Copyright(c) 2007-2013 Intel Corporation. All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
*
* * Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* * Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in
* the documentation and/or other materials provided with the
* distribution.
* * Neither the name of Intel Corporation nor the names of its
* contributors may be used to endorse or promote products derived
* from this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#include <sys/cdefs.h>
__FBSDID("$FreeBSD$");
#if 0
__KERNEL_RCSID(0, "$NetBSD: qat_hw15.c,v 1.1 2019/11/20 09:37:46 hikaru Exp $");
#endif
#include <sys/param.h>
#include <sys/bus.h>
#include <sys/kernel.h>
#include <sys/proc.h>
#include <sys/systm.h>
#include <machine/bus.h>
#include <opencrypto/xform.h>
#include <dev/pci/pcireg.h>
#include <dev/pci/pcivar.h>
#include "qatreg.h"
#include "qat_hw15reg.h"
#include "qatvar.h"
#include "qat_hw15var.h"
static int qat_adm_ring_init_ring_table(struct qat_softc *);
static void qat_adm_ring_build_slice_mask(uint16_t *, uint32_t, uint32_t);
static void qat_adm_ring_build_shram_mask(uint64_t *, uint32_t, uint32_t);
static int qat_adm_ring_build_ring_table(struct qat_softc *, uint32_t);
static int qat_adm_ring_build_init_msg(struct qat_softc *,
struct fw_init_req *, enum fw_init_cmd_id, uint32_t,
struct qat_accel_init_cb *);
static int qat_adm_ring_send_init_msg_sync(struct qat_softc *,
enum fw_init_cmd_id, uint32_t);
static int qat_adm_ring_send_init_msg(struct qat_softc *,
enum fw_init_cmd_id);
static int qat_adm_ring_intr(struct qat_softc *, void *, void *);
void
qat_msg_req_type_populate(struct arch_if_req_hdr *msg, enum arch_if_req type,
uint32_t rxring)
{
memset(msg, 0, sizeof(struct arch_if_req_hdr));
msg->flags = ARCH_IF_FLAGS_VALID_FLAG |
ARCH_IF_FLAGS_RESP_RING_TYPE_ET | ARCH_IF_FLAGS_RESP_TYPE_S;
msg->req_type = type;
msg->resp_pipe_id = rxring;
}
void
qat_msg_cmn_hdr_populate(struct fw_la_bulk_req *msg, bus_addr_t desc_paddr,
uint8_t hdrsz, uint8_t hwblksz, uint16_t comn_req_flags, uint32_t flow_id)
{
struct fw_comn_req_hdr *hdr = &msg->comn_hdr;
hdr->comn_req_flags = comn_req_flags;
hdr->content_desc_params_sz = hwblksz;
hdr->content_desc_hdr_sz = hdrsz;
hdr->content_desc_addr = desc_paddr;
msg->flow_id = flow_id;
}
void
qat_msg_service_cmd_populate(struct fw_la_bulk_req *msg, enum fw_la_cmd_id cmdid,
uint16_t cmd_flags)
{
msg->comn_la_req.la_cmd_id = cmdid;
msg->comn_la_req.u.la_flags = cmd_flags;
}
void
qat_msg_cmn_mid_populate(struct fw_comn_req_mid *msg, void *cookie,
uint64_t src, uint64_t dst)
{
msg->opaque_data = (uint64_t)(uintptr_t)cookie;
msg->src_data_addr = src;
if (dst == 0)
msg->dest_data_addr = src;
else
msg->dest_data_addr = dst;
}
void
qat_msg_req_params_populate(struct fw_la_bulk_req *msg,
bus_addr_t req_params_paddr, uint8_t req_params_sz)
{
msg->req_params_addr = req_params_paddr;
msg->comn_la_req.u1.req_params_blk_sz = req_params_sz / 8;
}
void
qat_msg_cmn_footer_populate(union fw_comn_req_ftr *msg, uint64_t next_addr)
{
msg->next_request_addr = next_addr;
}
void
qat_msg_params_populate(struct fw_la_bulk_req *msg,
struct qat_crypto_desc *desc, uint8_t req_params_sz,
uint16_t service_cmd_flags, uint16_t comn_req_flags)
{
qat_msg_cmn_hdr_populate(msg, desc->qcd_desc_paddr,
desc->qcd_hdr_sz, desc->qcd_hw_blk_sz, comn_req_flags, 0);
qat_msg_service_cmd_populate(msg, desc->qcd_cmd_id, service_cmd_flags);
qat_msg_cmn_mid_populate(&msg->comn_mid, NULL, 0, 0);
qat_msg_req_params_populate(msg, 0, req_params_sz);
qat_msg_cmn_footer_populate(&msg->comn_ftr, 0);
}
static int
qat_adm_ring_init_ring_table(struct qat_softc *sc)
{
struct qat_admin_rings *qadr = &sc->sc_admin_rings;
if (sc->sc_ae_num == 1) {
qadr->qadr_cya_ring_tbl =
&qadr->qadr_master_ring_tbl[0];
qadr->qadr_srv_mask[0] = QAT_SERVICE_CRYPTO_A;
} else if (sc->sc_ae_num == 2 || sc->sc_ae_num == 4) {
qadr->qadr_cya_ring_tbl =
&qadr->qadr_master_ring_tbl[0];
qadr->qadr_srv_mask[0] = QAT_SERVICE_CRYPTO_A;
qadr->qadr_cyb_ring_tbl =
&qadr->qadr_master_ring_tbl[1];
qadr->qadr_srv_mask[1] = QAT_SERVICE_CRYPTO_B;
}
return 0;
}
int
qat_adm_ring_init(struct qat_softc *sc)
{
struct qat_admin_rings *qadr = &sc->sc_admin_rings;
int error, i, j;
error = qat_alloc_dmamem(sc, &qadr->qadr_dma, 1, PAGE_SIZE, PAGE_SIZE);
if (error)
return error;
qadr->qadr_master_ring_tbl = qadr->qadr_dma.qdm_dma_vaddr;
MPASS(sc->sc_ae_num *
sizeof(struct fw_init_ring_table) <= PAGE_SIZE);
/* Initialize the Master Ring Table */
for (i = 0; i < sc->sc_ae_num; i++) {
struct fw_init_ring_table *firt =
&qadr->qadr_master_ring_tbl[i];
for (j = 0; j < INIT_RING_TABLE_SZ; j++) {
struct fw_init_ring_params *firp =
&firt->firt_bulk_rings[j];
firp->firp_reserved = 0;
firp->firp_curr_weight = QAT_DEFAULT_RING_WEIGHT;
firp->firp_init_weight = QAT_DEFAULT_RING_WEIGHT;
firp->firp_ring_pvl = QAT_DEFAULT_PVL;
}
memset(firt->firt_ring_mask, 0, sizeof(firt->firt_ring_mask));
}
error = qat_etr_setup_ring(sc, 0, RING_NUM_ADMIN_TX,
ADMIN_RING_SIZE, sc->sc_hw.qhw_fw_req_size,
NULL, NULL, "admin_tx", &qadr->qadr_admin_tx);
if (error)
return error;
error = qat_etr_setup_ring(sc, 0, RING_NUM_ADMIN_RX,
ADMIN_RING_SIZE, sc->sc_hw.qhw_fw_resp_size,
qat_adm_ring_intr, qadr, "admin_rx", &qadr->qadr_admin_rx);
if (error)
return error;
/*
* Finally set up the service indices into the Master Ring Table
* and convenient ring table pointers for each service enabled.
* Only the Admin rings are initialized.
*/
error = qat_adm_ring_init_ring_table(sc);
if (error)
return error;
/*
* Calculate the number of active AEs per QAT
* needed for Shram partitioning.
*/
for (i = 0; i < sc->sc_ae_num; i++) {
if (qadr->qadr_srv_mask[i])
qadr->qadr_active_aes_per_accel++;
}
return 0;
}
static void
qat_adm_ring_build_slice_mask(uint16_t *slice_mask, uint32_t srv_mask,
uint32_t init_shram)
{
uint16_t shram = 0, comn_req = 0;
if (init_shram)
shram = COMN_REQ_SHRAM_INIT_REQUIRED;
if (srv_mask & QAT_SERVICE_CRYPTO_A)
comn_req |= COMN_REQ_CY0_ONLY(shram);
if (srv_mask & QAT_SERVICE_CRYPTO_B)
comn_req |= COMN_REQ_CY1_ONLY(shram);
*slice_mask = comn_req;
}
static void
qat_adm_ring_build_shram_mask(uint64_t *shram_mask, uint32_t active_aes,
uint32_t ae)
{
*shram_mask = 0;
if (active_aes == 1) {
*shram_mask = ~(*shram_mask);
} else if (active_aes == 2) {
if (ae == 1)
*shram_mask = ((~(*shram_mask)) & 0xffffffff);
else
*shram_mask = ((~(*shram_mask)) & 0xffffffff00000000ull);
} else if (active_aes == 3) {
if (ae == 0)
*shram_mask = ((~(*shram_mask)) & 0x7fffff);
else if (ae == 1)
*shram_mask = ((~(*shram_mask)) & 0x3fffff800000ull);
else
*shram_mask = ((~(*shram_mask)) & 0xffffc00000000000ull);
} else {
panic("Only three services are supported in current version");
}
}
static int
qat_adm_ring_build_ring_table(struct qat_softc *sc, uint32_t ae)
{
struct qat_admin_rings *qadr = &sc->sc_admin_rings;
struct fw_init_ring_table *tbl;
struct fw_init_ring_params *param;
uint8_t srv_mask = sc->sc_admin_rings.qadr_srv_mask[ae];
if ((srv_mask & QAT_SERVICE_CRYPTO_A)) {
tbl = qadr->qadr_cya_ring_tbl;
} else if ((srv_mask & QAT_SERVICE_CRYPTO_B)) {
tbl = qadr->qadr_cyb_ring_tbl;
} else {
device_printf(sc->sc_dev,
"Invalid execution engine %d\n", ae);
return EINVAL;
}
param = &tbl->firt_bulk_rings[sc->sc_hw.qhw_ring_sym_tx];
param->firp_curr_weight = QAT_HI_PRIO_RING_WEIGHT;
param->firp_init_weight = QAT_HI_PRIO_RING_WEIGHT;
FW_INIT_RING_MASK_SET(tbl, sc->sc_hw.qhw_ring_sym_tx);
return 0;
}
static int
qat_adm_ring_build_init_msg(struct qat_softc *sc,
struct fw_init_req *initmsg, enum fw_init_cmd_id cmd, uint32_t ae,
struct qat_accel_init_cb *cb)
{
struct fw_init_set_ae_info_hdr *aehdr;
struct fw_init_set_ae_info *aeinfo;
struct fw_init_set_ring_info_hdr *ringhdr;
struct fw_init_set_ring_info *ringinfo;
int init_shram = 0, tgt_id, cluster_id;
uint32_t srv_mask;
srv_mask = sc->sc_admin_rings.qadr_srv_mask[
ae % sc->sc_ae_num];
memset(initmsg, 0, sizeof(struct fw_init_req));
qat_msg_req_type_populate(&initmsg->comn_hdr.arch_if,
ARCH_IF_REQ_QAT_FW_INIT,
sc->sc_admin_rings.qadr_admin_rx->qr_ring_id);
qat_msg_cmn_mid_populate(&initmsg->comn_mid, cb, 0, 0);
switch (cmd) {
case FW_INIT_CMD_SET_AE_INFO:
if (ae % sc->sc_ae_num == 0)
init_shram = 1;
if (ae >= sc->sc_ae_num) {
tgt_id = 1;
cluster_id = 1;
} else {
cluster_id = 0;
if (sc->sc_ae_mask)
tgt_id = 0;
else
tgt_id = 1;
}
aehdr = &initmsg->u.set_ae_info;
aeinfo = &initmsg->u1.set_ae_info;
aehdr->init_cmd_id = cmd;
/* XXX that does not support sparse ae_mask */
aehdr->init_trgt_id = ae;
aehdr->init_ring_cluster_id = cluster_id;
aehdr->init_qat_id = tgt_id;
qat_adm_ring_build_slice_mask(&aehdr->init_slice_mask, srv_mask,
init_shram);
qat_adm_ring_build_shram_mask(&aeinfo->init_shram_mask,
sc->sc_admin_rings.qadr_active_aes_per_accel,
ae % sc->sc_ae_num);
break;
case FW_INIT_CMD_SET_RING_INFO:
ringhdr = &initmsg->u.set_ring_info;
ringinfo = &initmsg->u1.set_ring_info;
ringhdr->init_cmd_id = cmd;
/* XXX that does not support sparse ae_mask */
ringhdr->init_trgt_id = ae;
/* XXX */
qat_adm_ring_build_ring_table(sc,
ae % sc->sc_ae_num);
ringhdr->init_ring_tbl_sz = sizeof(struct fw_init_ring_table);
ringinfo->init_ring_table_ptr =
sc->sc_admin_rings.qadr_dma.qdm_dma_seg.ds_addr +
((ae % sc->sc_ae_num) *
sizeof(struct fw_init_ring_table));
break;
default:
return ENOTSUP;
}
return 0;
}
static int
qat_adm_ring_send_init_msg_sync(struct qat_softc *sc,
enum fw_init_cmd_id cmd, uint32_t ae)
{
struct fw_init_req initmsg;
struct qat_accel_init_cb cb;
int error;
error = qat_adm_ring_build_init_msg(sc, &initmsg, cmd, ae, &cb);
if (error)
return error;
error = qat_etr_put_msg(sc, sc->sc_admin_rings.qadr_admin_tx,
(uint32_t *)&initmsg);
if (error)
return error;
error = tsleep(&cb, PZERO, "qat_init", hz * 3 / 2);
if (error) {
device_printf(sc->sc_dev,
"Timed out initialization firmware: %d\n", error);
return error;
}
if (cb.qaic_status) {
device_printf(sc->sc_dev, "Failed to initialize firmware\n");
return EIO;
}
return error;
}
static int
qat_adm_ring_send_init_msg(struct qat_softc *sc,
enum fw_init_cmd_id cmd)
{
struct qat_admin_rings *qadr = &sc->sc_admin_rings;
uint32_t error, ae;
for (ae = 0; ae < sc->sc_ae_num; ae++) {
uint8_t srv_mask = qadr->qadr_srv_mask[ae];
switch (cmd) {
case FW_INIT_CMD_SET_AE_INFO:
case FW_INIT_CMD_SET_RING_INFO:
if (!srv_mask)
continue;
break;
case FW_INIT_CMD_TRNG_ENABLE:
case FW_INIT_CMD_TRNG_DISABLE:
if (!(srv_mask & QAT_SERVICE_CRYPTO_A))
continue;
break;
default:
return ENOTSUP;
}
error = qat_adm_ring_send_init_msg_sync(sc, cmd, ae);
if (error)
return error;
}
return 0;
}
int
qat_adm_ring_send_init(struct qat_softc *sc)
{
int error;
error = qat_adm_ring_send_init_msg(sc, FW_INIT_CMD_SET_AE_INFO);
if (error)
return error;
error = qat_adm_ring_send_init_msg(sc, FW_INIT_CMD_SET_RING_INFO);
if (error)
return error;
return 0;
}
static int
qat_adm_ring_intr(struct qat_softc *sc, void *arg, void *msg)
{
struct arch_if_resp_hdr *resp;
struct fw_init_resp *init_resp;
struct qat_accel_init_cb *init_cb;
int handled = 0;
resp = (struct arch_if_resp_hdr *)msg;
switch (resp->resp_type) {
case ARCH_IF_REQ_QAT_FW_INIT:
init_resp = (struct fw_init_resp *)msg;
init_cb = (struct qat_accel_init_cb *)
(uintptr_t)init_resp->comn_resp.opaque_data;
init_cb->qaic_status =
__SHIFTOUT(init_resp->comn_resp.comn_status,
COMN_RESP_INIT_ADMIN_STATUS);
wakeup(init_cb);
break;
default:
device_printf(sc->sc_dev,
"unknown resp type %d\n", resp->resp_type);
break;
}
return handled;
}
static inline uint16_t
qat_hw15_get_comn_req_flags(uint8_t ae)
{
if (ae == 0) {
return COMN_REQ_ORD_STRICT | COMN_REQ_PTR_TYPE_SGL |
COMN_REQ_AUTH0_SLICE_REQUIRED |
COMN_REQ_CIPHER0_SLICE_REQUIRED;
} else {
return COMN_REQ_ORD_STRICT | COMN_REQ_PTR_TYPE_SGL |
COMN_REQ_AUTH1_SLICE_REQUIRED |
COMN_REQ_CIPHER1_SLICE_REQUIRED;
}
}
static uint32_t
qat_hw15_crypto_setup_cipher_desc(struct qat_crypto_desc *desc,
struct qat_session *qs, struct fw_cipher_hdr *cipher_hdr,
uint32_t hw_blk_offset, enum fw_slice next_slice)
{
desc->qcd_cipher_blk_sz = HW_AES_BLK_SZ;
cipher_hdr->state_padding_sz = 0;
cipher_hdr->key_sz = qs->qs_cipher_klen / 8;
cipher_hdr->state_sz = desc->qcd_cipher_blk_sz / 8;
cipher_hdr->next_id = next_slice;
cipher_hdr->curr_id = FW_SLICE_CIPHER;
cipher_hdr->offset = hw_blk_offset / 8;
cipher_hdr->resrvd = 0;
return sizeof(struct hw_cipher_config) + qs->qs_cipher_klen;
}
static void
qat_hw15_crypto_setup_cipher_config(const struct qat_crypto_desc *desc,
const struct qat_session *qs, const struct cryptop *crp,
struct hw_cipher_config *cipher_config)
{
const uint8_t *key;
uint8_t *cipher_key;
cipher_config->val = qat_crypto_load_cipher_session(desc, qs);
cipher_config->reserved = 0;
cipher_key = (uint8_t *)(cipher_config + 1);
if (crp != NULL && crp->crp_cipher_key != NULL)
key = crp->crp_cipher_key;
else
key = qs->qs_cipher_key;
memcpy(cipher_key, key, qs->qs_cipher_klen);
}
static uint32_t
qat_hw15_crypto_setup_auth_desc(struct qat_crypto_desc *desc,
struct qat_session *qs, struct fw_auth_hdr *auth_hdr,
uint32_t ctrl_blk_offset, uint32_t hw_blk_offset,
enum fw_slice next_slice)
{
const struct qat_sym_hash_def *hash_def;
(void)qat_crypto_load_auth_session(desc, qs, &hash_def);
auth_hdr->next_id = next_slice;
auth_hdr->curr_id = FW_SLICE_AUTH;
auth_hdr->offset = hw_blk_offset / 8;
auth_hdr->resrvd = 0;
auth_hdr->hash_flags = FW_AUTH_HDR_FLAG_NO_NESTED;
auth_hdr->u.inner_prefix_sz = 0;
auth_hdr->outer_prefix_sz = 0;
auth_hdr->final_sz = hash_def->qshd_alg->qshai_digest_len;
auth_hdr->inner_state1_sz =
roundup(hash_def->qshd_qat->qshqi_state1_len, 8);
auth_hdr->inner_res_sz = hash_def->qshd_alg->qshai_digest_len;
auth_hdr->inner_state2_sz =
roundup(hash_def->qshd_qat->qshqi_state2_len, 8);
auth_hdr->inner_state2_off = auth_hdr->offset +
((sizeof(struct hw_auth_setup) + auth_hdr->inner_state1_sz) / 8);
auth_hdr->outer_config_off = 0;
auth_hdr->outer_state1_sz = 0;
auth_hdr->outer_res_sz = 0;
auth_hdr->outer_prefix_off = 0;
desc->qcd_auth_sz = hash_def->qshd_alg->qshai_sah->hashsize;
desc->qcd_state_storage_sz = (sizeof(struct hw_auth_counter) +
roundup(hash_def->qshd_alg->qshai_state_size, 8)) / 8;
desc->qcd_gcm_aad_sz_offset1 = desc->qcd_auth_offset +
sizeof(struct hw_auth_setup) + auth_hdr->inner_state1_sz +
AES_BLOCK_LEN;
desc->qcd_gcm_aad_sz_offset2 = ctrl_blk_offset +
offsetof(struct fw_auth_hdr, u.aad_sz);
return sizeof(struct hw_auth_setup) + auth_hdr->inner_state1_sz +
auth_hdr->inner_state2_sz;
}
static void
qat_hw15_crypto_setup_auth_setup(const struct qat_crypto_desc *desc,
const struct qat_session *qs, const struct cryptop *crp,
struct hw_auth_setup *auth_setup)
{
const struct qat_sym_hash_def *hash_def;
const uint8_t *key;
uint8_t *state1, *state2;
uint32_t state_sz, state1_sz, state2_sz, state1_pad_len, state2_pad_len;
auth_setup->auth_config.config = qat_crypto_load_auth_session(desc, qs,
&hash_def);
auth_setup->auth_config.reserved = 0;
auth_setup->auth_counter.counter =
htobe32(hash_def->qshd_qat->qshqi_auth_counter);
auth_setup->auth_counter.reserved = 0;
state1 = (uint8_t *)(auth_setup + 1);
state2 = state1 + roundup(hash_def->qshd_qat->qshqi_state1_len, 8);
switch (qs->qs_auth_algo) {
case HW_AUTH_ALGO_GALOIS_128:
qat_crypto_gmac_precompute(desc, qs->qs_cipher_key,
qs->qs_cipher_klen, hash_def, state2);
break;
case HW_AUTH_ALGO_SHA1:
state_sz = hash_def->qshd_alg->qshai_state_size;
state1_sz = roundup(hash_def->qshd_qat->qshqi_state1_len, 8);
state2_sz = roundup(hash_def->qshd_qat->qshqi_state2_len, 8);
if (qs->qs_auth_mode == HW_AUTH_MODE1) {
state1_pad_len = state1_sz - state_sz;
state2_pad_len = state2_sz - state_sz;
if (state1_pad_len > 0)
memset(state1 + state_sz, 0, state1_pad_len);
if (state2_pad_len > 0)
memset(state2 + state_sz, 0, state2_pad_len);
}
/* FALLTHROUGH */
case HW_AUTH_ALGO_SHA256:
case HW_AUTH_ALGO_SHA384:
case HW_AUTH_ALGO_SHA512:
switch (qs->qs_auth_mode) {
case HW_AUTH_MODE0:
memcpy(state1, hash_def->qshd_alg->qshai_init_state,
state1_sz);
/* Override for mode 0 hashes. */
auth_setup->auth_counter.counter = 0;
break;
case HW_AUTH_MODE1:
if (crp != NULL && crp->crp_auth_key != NULL)
key = crp->crp_auth_key;
else
key = qs->qs_auth_key;
if (key != NULL) {
qat_crypto_hmac_precompute(desc, key,
qs->qs_auth_klen, hash_def, state1, state2);
}
break;
default:
panic("%s: unhandled auth mode %d", __func__,
qs->qs_auth_mode);
}
break;
default:
panic("%s: unhandled auth algorithm %d", __func__,
qs->qs_auth_algo);
}
}
void
qat_hw15_crypto_setup_desc(struct qat_crypto *qcy, struct qat_session *qs,
struct qat_crypto_desc *desc)
{
struct fw_cipher_hdr *cipher_hdr;
struct fw_auth_hdr *auth_hdr;
struct fw_la_bulk_req *req_cache;
struct hw_auth_setup *auth_setup;
struct hw_cipher_config *cipher_config;
uint32_t ctrl_blk_sz, ctrl_blk_offset, hw_blk_offset;
int i;
uint16_t la_cmd_flags;
uint8_t req_params_sz;
uint8_t *ctrl_blk_ptr, *hw_blk_ptr;
ctrl_blk_sz = 0;
if (qs->qs_cipher_algo != HW_CIPHER_ALGO_NULL)
ctrl_blk_sz += sizeof(struct fw_cipher_hdr);
if (qs->qs_auth_algo != HW_AUTH_ALGO_NULL)
ctrl_blk_sz += sizeof(struct fw_auth_hdr);
ctrl_blk_ptr = desc->qcd_content_desc;
ctrl_blk_offset = 0;
hw_blk_ptr = ctrl_blk_ptr + ctrl_blk_sz;
hw_blk_offset = 0;
la_cmd_flags = 0;
req_params_sz = 0;
for (i = 0; i < MAX_FW_SLICE; i++) {
switch (desc->qcd_slices[i]) {
case FW_SLICE_CIPHER:
cipher_hdr = (struct fw_cipher_hdr *)(ctrl_blk_ptr +
ctrl_blk_offset);
cipher_config = (struct hw_cipher_config *)(hw_blk_ptr +
hw_blk_offset);
desc->qcd_cipher_offset = ctrl_blk_sz + hw_blk_offset;
hw_blk_offset += qat_hw15_crypto_setup_cipher_desc(desc,
qs, cipher_hdr, hw_blk_offset,
desc->qcd_slices[i + 1]);
qat_hw15_crypto_setup_cipher_config(desc, qs, NULL,
cipher_config);
ctrl_blk_offset += sizeof(struct fw_cipher_hdr);
req_params_sz += sizeof(struct fw_la_cipher_req_params);
break;
case FW_SLICE_AUTH:
auth_hdr = (struct fw_auth_hdr *)(ctrl_blk_ptr +
ctrl_blk_offset);
auth_setup = (struct hw_auth_setup *)(hw_blk_ptr +
hw_blk_offset);
desc->qcd_auth_offset = ctrl_blk_sz + hw_blk_offset;
hw_blk_offset += qat_hw15_crypto_setup_auth_desc(desc,
qs, auth_hdr, ctrl_blk_offset, hw_blk_offset,
desc->qcd_slices[i + 1]);
qat_hw15_crypto_setup_auth_setup(desc, qs, NULL,
auth_setup);
ctrl_blk_offset += sizeof(struct fw_auth_hdr);
req_params_sz += sizeof(struct fw_la_auth_req_params);
la_cmd_flags |= LA_FLAGS_RET_AUTH_RES;
/* no digest verify */
break;
case FW_SLICE_DRAM_WR:
i = MAX_FW_SLICE; /* end of chain */
break;
default:
MPASS(0);
break;
}
}
desc->qcd_hdr_sz = ctrl_blk_offset / 8;
desc->qcd_hw_blk_sz = hw_blk_offset / 8;
req_cache = (struct fw_la_bulk_req *)desc->qcd_req_cache;
qat_msg_req_type_populate(
&req_cache->comn_hdr.arch_if,
ARCH_IF_REQ_QAT_FW_LA, 0);
if (qs->qs_auth_algo == HW_AUTH_ALGO_GALOIS_128)
la_cmd_flags |= LA_FLAGS_PROTO_GCM | LA_FLAGS_GCM_IV_LEN_FLAG;
else
la_cmd_flags |= LA_FLAGS_PROTO_NO;
qat_msg_params_populate(req_cache, desc, req_params_sz,
la_cmd_flags, 0);
bus_dmamap_sync(qs->qs_desc_mem.qdm_dma_tag,
qs->qs_desc_mem.qdm_dma_map, BUS_DMASYNC_PREWRITE);
}
static void
qat_hw15_crypto_req_setkey(const struct qat_crypto_desc *desc,
const struct qat_session *qs, struct qat_sym_cookie *qsc,
struct fw_la_bulk_req *bulk_req, struct cryptop *crp)
{
struct hw_auth_setup *auth_setup;
struct hw_cipher_config *cipher_config;
uint8_t *cdesc;
int i;
cdesc = qsc->qsc_content_desc;
memcpy(cdesc, desc->qcd_content_desc, CONTENT_DESC_MAX_SIZE);
for (i = 0; i < MAX_FW_SLICE; i++) {
switch (desc->qcd_slices[i]) {
case FW_SLICE_CIPHER:
cipher_config = (struct hw_cipher_config *)
(cdesc + desc->qcd_cipher_offset);
qat_hw15_crypto_setup_cipher_config(desc, qs, crp,
cipher_config);
break;
case FW_SLICE_AUTH:
auth_setup = (struct hw_auth_setup *)
(cdesc + desc->qcd_auth_offset);
qat_hw15_crypto_setup_auth_setup(desc, qs, crp,
auth_setup);
break;
case FW_SLICE_DRAM_WR:
i = MAX_FW_SLICE; /* end of chain */
break;
default:
MPASS(0);
}
}
bulk_req->comn_hdr.content_desc_addr = qsc->qsc_content_desc_paddr;
}
void
qat_hw15_crypto_setup_req_params(struct qat_crypto_bank *qcb,
struct qat_session *qs, struct qat_crypto_desc const *desc,
struct qat_sym_cookie *qsc, struct cryptop *crp)
{
struct qat_sym_bulk_cookie *qsbc;
struct fw_la_bulk_req *bulk_req;
struct fw_la_cipher_req_params *cipher_req;
struct fw_la_auth_req_params *auth_req;
bus_addr_t digest_paddr;
uint8_t *aad_szp2, *req_params_ptr;
uint32_t aad_sz, *aad_szp1;
enum fw_la_cmd_id cmd_id = desc->qcd_cmd_id;
enum fw_slice next_slice;
qsbc = &qsc->qsc_bulk_cookie;
bulk_req = (struct fw_la_bulk_req *)qsbc->qsbc_msg;
memcpy(bulk_req, &desc->qcd_req_cache, QAT_HW15_SESSION_REQ_CACHE_SIZE);
bulk_req->comn_hdr.arch_if.resp_pipe_id = qcb->qcb_sym_rx->qr_ring_id;
bulk_req->comn_hdr.comn_req_flags =
qat_hw15_get_comn_req_flags(qcb->qcb_bank % 2);
bulk_req->comn_mid.src_data_addr = qsc->qsc_buffer_list_desc_paddr;
if (CRYPTO_HAS_OUTPUT_BUFFER(crp)) {
bulk_req->comn_mid.dest_data_addr =
qsc->qsc_obuffer_list_desc_paddr;
} else {
bulk_req->comn_mid.dest_data_addr =
qsc->qsc_buffer_list_desc_paddr;
}
bulk_req->req_params_addr = qsc->qsc_bulk_req_params_buf_paddr;
bulk_req->comn_ftr.next_request_addr = 0;
bulk_req->comn_mid.opaque_data = (uint64_t)(uintptr_t)qsc;
if (__predict_false(crp->crp_cipher_key != NULL ||
crp->crp_auth_key != NULL)) {
qat_hw15_crypto_req_setkey(desc, qs, qsc, bulk_req, crp);
}
digest_paddr = 0;
if (desc->qcd_auth_sz != 0)
digest_paddr = qsc->qsc_auth_res_paddr;
req_params_ptr = qsbc->qsbc_req_params_buf;
memset(req_params_ptr, 0, sizeof(qsbc->qsbc_req_params_buf));
/*
* The SG list layout is a bit different for GCM and GMAC, it's simpler
* to handle those cases separately.
*/
if (qs->qs_auth_algo == HW_AUTH_ALGO_GALOIS_128) {
cipher_req = (struct fw_la_cipher_req_params *)req_params_ptr;
auth_req = (struct fw_la_auth_req_params *)
(req_params_ptr + sizeof(struct fw_la_cipher_req_params));
cipher_req->cipher_state_sz = desc->qcd_cipher_blk_sz / 8;
cipher_req->curr_id = FW_SLICE_CIPHER;
if (cmd_id == FW_LA_CMD_HASH_CIPHER || cmd_id == FW_LA_CMD_AUTH)
cipher_req->next_id = FW_SLICE_DRAM_WR;
else
cipher_req->next_id = FW_SLICE_AUTH;
cipher_req->state_address = qsc->qsc_iv_buf_paddr;
if (cmd_id != FW_LA_CMD_AUTH) {
/*
* Don't fill out the cipher block if we're doing GMAC
* only.
*/
cipher_req->cipher_off = 0;
cipher_req->cipher_len = crp->crp_payload_length;
}
auth_req->curr_id = FW_SLICE_AUTH;
if (cmd_id == FW_LA_CMD_HASH_CIPHER || cmd_id == FW_LA_CMD_AUTH)
auth_req->next_id = FW_SLICE_CIPHER;
else
auth_req->next_id = FW_SLICE_DRAM_WR;
auth_req->auth_res_address = digest_paddr;
auth_req->auth_res_sz = desc->qcd_auth_sz;
auth_req->auth_off = 0;
auth_req->auth_len = crp->crp_payload_length;
auth_req->hash_state_sz =
roundup2(crp->crp_aad_length, QAT_AES_GCM_AAD_ALIGN) >> 3;
auth_req->u1.aad_addr = crp->crp_aad_length > 0 ?
qsc->qsc_gcm_aad_paddr : 0;
/*
* Update the hash state block if necessary. This only occurs
* when the AAD length changes between requests in a session and
* is synchronized by qat_process().
*/
aad_sz = htobe32(crp->crp_aad_length);
aad_szp1 = (uint32_t *)(
__DECONST(uint8_t *, desc->qcd_content_desc) +
desc->qcd_gcm_aad_sz_offset1);
aad_szp2 = __DECONST(uint8_t *, desc->qcd_content_desc) +
desc->qcd_gcm_aad_sz_offset2;
if (__predict_false(*aad_szp1 != aad_sz)) {
*aad_szp1 = aad_sz;
*aad_szp2 = (uint8_t)roundup2(crp->crp_aad_length,
QAT_AES_GCM_AAD_ALIGN);
bus_dmamap_sync(qs->qs_desc_mem.qdm_dma_tag,
qs->qs_desc_mem.qdm_dma_map,
BUS_DMASYNC_PREWRITE);
}
} else {
cipher_req = (struct fw_la_cipher_req_params *)req_params_ptr;
if (cmd_id != FW_LA_CMD_AUTH) {
if (cmd_id == FW_LA_CMD_CIPHER ||
cmd_id == FW_LA_CMD_HASH_CIPHER)
next_slice = FW_SLICE_DRAM_WR;
else
next_slice = FW_SLICE_AUTH;
cipher_req->cipher_state_sz =
desc->qcd_cipher_blk_sz / 8;
cipher_req->curr_id = FW_SLICE_CIPHER;
cipher_req->next_id = next_slice;
if (crp->crp_aad_length == 0) {
cipher_req->cipher_off = 0;
} else if (crp->crp_aad == NULL) {
cipher_req->cipher_off =
crp->crp_payload_start - crp->crp_aad_start;
} else {
cipher_req->cipher_off = crp->crp_aad_length;
}
cipher_req->cipher_len = crp->crp_payload_length;
cipher_req->state_address = qsc->qsc_iv_buf_paddr;
}
if (cmd_id != FW_LA_CMD_CIPHER) {
if (cmd_id == FW_LA_CMD_AUTH)
auth_req = (struct fw_la_auth_req_params *)
req_params_ptr;
else
auth_req = (struct fw_la_auth_req_params *)
(cipher_req + 1);
if (cmd_id == FW_LA_CMD_HASH_CIPHER)
next_slice = FW_SLICE_CIPHER;
else
next_slice = FW_SLICE_DRAM_WR;
auth_req->curr_id = FW_SLICE_AUTH;
auth_req->next_id = next_slice;
auth_req->auth_res_address = digest_paddr;
auth_req->auth_res_sz = desc->qcd_auth_sz;
auth_req->auth_len =
crp->crp_payload_length + crp->crp_aad_length;
auth_req->auth_off = 0;
auth_req->hash_state_sz = 0;
auth_req->u1.prefix_addr = desc->qcd_hash_state_paddr +
desc->qcd_state_storage_sz;
}
}
}