freebsd-nq/sys/mips/nlm/hal/nlmsaelib.h
Jayachandran C. 9b4d140639 Opencrypto driver for XLP Security and RSA/ECC blocks
Support for the Security and RSA blocks on XLP SoC. Even though
the XLP supports many more algorithms, only the ones supported
in OCF have been added.

Submitted by:	Venkatesh J. V. (venkatesh at netlogicmicro com)
2012-03-27 11:43:46 +00:00

608 lines
19 KiB
C

/*-
* Copyright (c) 2003-2012 Broadcom Corporation
* All Rights Reserved
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
*
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in
* the documentation and/or other materials provided with the
* distribution.
*
* THIS SOFTWARE IS PROVIDED BY BROADCOM ``AS IS'' AND ANY EXPRESS OR
* IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
* WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
* ARE DISCLAIMED. IN NO EVENT SHALL BROADCOM OR CONTRIBUTORS BE LIABLE
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR
* BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,
* WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE
* OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN
* IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*
* $FreeBSD$
*/
#ifndef _NLM_HAL_CRYPTO_H_
#define _NLM_HAL_CRYPTO_H_
#define SAE_CFG_REG 0x00
#define SAE_ENG_SEL_0 0x01
#define SAE_ENG_SEL_1 0x02
#define SAE_ENG_SEL_2 0x03
#define SAE_ENG_SEL_3 0x04
#define SAE_ENG_SEL_4 0x05
#define SAE_ENG_SEL_5 0x06
#define SAE_ENG_SEL_6 0x07
#define SAE_ENG_SEL_7 0x08
#define RSA_CFG_REG 0x00
#define RSA_ENG_SEL_0 0x01
#define RSA_ENG_SEL_1 0x02
#define RSA_ENG_SEL_2 0x03
#define nlm_read_sec_reg(b, r) nlm_read_reg(b, r)
#define nlm_write_sec_reg(b, r, v) nlm_write_reg(b, r, v)
#define nlm_get_sec_pcibase(node) nlm_pcicfg_base(XLP_IO_SEC_OFFSET(node))
#define nlm_get_sec_regbase(node) \
(nlm_get_sec_pcibase(node) + XLP_IO_PCI_HDRSZ)
#define nlm_read_rsa_reg(b, r) nlm_read_reg(b, r)
#define nlm_write_rsa_reg(b, r, v) nlm_write_reg(b, r, v)
#define nlm_get_rsa_pcibase(node) nlm_pcicfg_base(XLP_IO_RSA_OFFSET(node))
#define nlm_get_rsa_regbase(node) \
(nlm_get_rsa_pcibase(node) + XLP_IO_PCI_HDRSZ)
#define nlm_pcibase_sec(node) nlm_pcicfg_base(XLP_IO_SEC_OFFSET(node))
#define nlm_qidstart_sec(node) nlm_qidstart_kseg(nlm_pcibase_sec(node))
#define nlm_qnum_sec(node) nlm_qnum_kseg(nlm_pcibase_sec(node))
/*
* Since buffer allocation for crypto at kernel is done as malloc, each
* segment size is given as page size which is 4K by default
*/
#define NLM_CRYPTO_MAX_SEG_LEN PAGE_SIZE
#define MAX_KEY_LEN_IN_DW 20
#define left_shift64(x, bitshift, numofbits) \
((uint64_t)(x) << (bitshift))
#define left_shift64_mask(x, bitshift, numofbits) \
(((uint64_t)(x) & ((1ULL << (numofbits)) - 1)) << (bitshift))
/**
* @brief cipher algorithms
* @ingroup crypto
*/
enum nlm_cipher_algo {
NLM_CIPHER_BYPASS = 0,
NLM_CIPHER_DES = 1,
NLM_CIPHER_3DES = 2,
NLM_CIPHER_AES128 = 3,
NLM_CIPHER_AES192 = 4,
NLM_CIPHER_AES256 = 5,
NLM_CIPHER_ARC4 = 6,
NLM_CIPHER_KASUMI_F8 = 7,
NLM_CIPHER_SNOW3G_F8 = 8,
NLM_CIPHER_CAMELLIA128 = 9,
NLM_CIPHER_CAMELLIA192 = 0xA,
NLM_CIPHER_CAMELLIA256 = 0xB,
NLM_CIPHER_MAX = 0xC,
};
/**
* @brief cipher modes
* @ingroup crypto
*/
enum nlm_cipher_mode {
NLM_CIPHER_MODE_ECB = 0,
NLM_CIPHER_MODE_CBC = 1,
NLM_CIPHER_MODE_CFB = 2,
NLM_CIPHER_MODE_OFB = 3,
NLM_CIPHER_MODE_CTR = 4,
NLM_CIPHER_MODE_AES_F8 = 5,
NLM_CIPHER_MODE_GCM = 6,
NLM_CIPHER_MODE_CCM = 7,
NLM_CIPHER_MODE_UNDEFINED1 = 8,
NLM_CIPHER_MODE_UNDEFINED2 = 9,
NLM_CIPHER_MODE_LRW = 0xA,
NLM_CIPHER_MODE_XTS = 0xB,
NLM_CIPHER_MODE_MAX = 0xC,
};
/**
* @brief hash algorithms
* @ingroup crypto
*/
enum nlm_hash_algo {
NLM_HASH_BYPASS = 0,
NLM_HASH_MD5 = 1,
NLM_HASH_SHA = 2,
NLM_HASH_UNDEFINED = 3,
NLM_HASH_AES128 = 4,
NLM_HASH_AES192 = 5,
NLM_HASH_AES256 = 6,
NLM_HASH_KASUMI_F9 = 7,
NLM_HASH_SNOW3G_F9 = 8,
NLM_HASH_CAMELLIA128 = 9,
NLM_HASH_CAMELLIA192 = 0xA,
NLM_HASH_CAMELLIA256 = 0xB,
NLM_HASH_GHASH = 0xC,
NLM_HASH_MAX = 0xD
};
/**
* @brief hash modes
* @ingroup crypto
*/
enum nlm_hash_mode {
NLM_HASH_MODE_SHA1 = 0, /* Only SHA */
NLM_HASH_MODE_SHA224 = 1, /* Only SHA */
NLM_HASH_MODE_SHA256 = 2, /* Only SHA */
NLM_HASH_MODE_SHA384 = 3, /* Only SHA */
NLM_HASH_MODE_SHA512 = 4, /* Only SHA */
NLM_HASH_MODE_CMAC = 5, /* AES and Camellia */
NLM_HASH_MODE_XCBC = 6, /* AES and Camellia */
NLM_HASH_MODE_CBC_MAC = 7, /* AES and Camellia */
NLM_HASH_MODE_CCM = 8, /* AES */
NLM_HASH_MODE_GCM = 9, /* AES */
NLM_HASH_MODE_MAX = 0xA,
};
/**
* @brief crypto control descriptor, should be cache aligned
* @ingroup crypto
*/
struct nlm_crypto_pkt_ctrl {
uint64_t desc0;
/* combination of cipher and hash keys */
uint64_t key[MAX_KEY_LEN_IN_DW];
uint32_t cipherkeylen;
uint32_t hashkeylen;
uint32_t taglen;
};
/**
* @brief crypto packet descriptor, should be cache aligned
* @ingroup crypto
*/
struct nlm_crypto_pkt_param {
uint64_t desc0;
uint64_t desc1;
uint64_t desc2;
uint64_t desc3;
uint64_t segment[1][2];
};
static __inline__ uint64_t
nlm_crypto_form_rsa_ecc_fmn_entry0(unsigned int l3alloc, unsigned int type,
unsigned int func, uint64_t srcaddr)
{
return (left_shift64(l3alloc, 61, 1) |
left_shift64(type, 46, 7) |
left_shift64(func, 40, 6) |
left_shift64(srcaddr, 0, 40));
}
static __inline__ uint64_t
nlm_crypto_form_rsa_ecc_fmn_entry1(unsigned int dstclobber,
unsigned int l3alloc, unsigned int fbvc, uint64_t dstaddr)
{
return (left_shift64(dstclobber, 62, 1) |
left_shift64(l3alloc, 61, 1) |
left_shift64(fbvc, 40, 12) |
left_shift64(dstaddr, 0, 40));
}
/**
* @brief Generate cypto control descriptor
* @ingroup crypto
* hmac : 1 for hash with hmac
* hashalg, see hash_alg enums
* hashmode, see hash_mode enums
* cipherhalg, see cipher_alg enums
* ciphermode, see cipher_mode enums
* arc4_cipherkeylen : length of arc4 cipher key, 0 is interpreted as 32
* arc4_keyinit :
* cfbmask : cipher text for feedback,
* 0(1 bit), 1(2 bits), 2(4 bits), 3(8 bits), 4(16bits), 5(32 bits),
* 6(64 bits), 7(128 bits)
*/
static __inline__ uint64_t
nlm_crypto_form_pkt_ctrl_desc(unsigned int hmac, unsigned int hashalg,
unsigned int hashmode, unsigned int cipheralg, unsigned int ciphermode,
unsigned int arc4_cipherkeylen, unsigned int arc4_keyinit,
unsigned int cfbmask)
{
return (left_shift64(hmac, 61, 1) |
left_shift64(hashalg, 52, 8) |
left_shift64(hashmode, 43, 8) |
left_shift64(cipheralg, 34, 8) |
left_shift64(ciphermode, 25, 8) |
left_shift64(arc4_cipherkeylen, 18, 5) |
left_shift64(arc4_keyinit, 17, 1) |
left_shift64(cfbmask, 0, 3));
}
/**
* @brief Generate cypto packet descriptor 0
* @ingroup crypto
* tls : 1 (tls enabled) 0(tls disabled)
* hash_source : 1 (encrypted data is sent to the auth engine)
* 0 (plain data is sent to the auth engine)
* hashout_l3alloc : 1 (auth output is transited through l3 cache)
* encrypt : 1 (for encrypt) 0 (for decrypt)
* ivlen : iv length in bytes
* hashdst_addr : hash out physical address, byte aligned
*/
static __inline__ uint64_t
nlm_crypto_form_pkt_desc0(unsigned int tls, unsigned int hash_source,
unsigned int hashout_l3alloc, unsigned int encrypt, unsigned int ivlen,
uint64_t hashdst_addr)
{
return (left_shift64(tls, 63, 1) |
left_shift64(hash_source, 62, 1) |
left_shift64(hashout_l3alloc, 60, 1) |
left_shift64(encrypt, 59, 1) |
left_shift64_mask((ivlen - 1), 41, 16) |
left_shift64(hashdst_addr, 0, 40));
}
/**
* @brief Generate cypto packet descriptor 1
* @ingroup crypto
* cipherlen : cipher length in bytes
* hashlen : hash length in bytes
*/
static __inline__ uint64_t
nlm_crypto_form_pkt_desc1(unsigned int cipherlen, unsigned int hashlen)
{
return (left_shift64_mask((cipherlen - 1), 32, 32) |
left_shift64_mask((hashlen - 1), 0, 32));
}
/**
* @brief Generate cypto packet descriptor 2
* @ingroup crypto
* ivoff : iv offset, offset from start of src data addr
* ciperbit_cnt : number of valid bits in the last input byte to the cipher,
* 0 (8 bits), 1 (1 bit)..7 (7 bits)
* cipheroff : cipher offset, offset from start of src data addr
* hashbit_cnt : number of valid bits in the last input byte to the auth
* 0 (8 bits), 1 (1 bit)..7 (7 bits)
* hashclobber : 1 (hash output will be written as multiples of cachelines, no
* read modify write)
* hashoff : hash offset, offset from start of src data addr
*/
static __inline__ uint64_t
nlm_crypto_form_pkt_desc2(unsigned int ivoff, unsigned int cipherbit_cnt,
unsigned int cipheroff, unsigned int hashbit_cnt, unsigned int hashclobber,
unsigned int hashoff)
{
return (left_shift64(ivoff , 45, 16) |
left_shift64(cipherbit_cnt, 42, 3) |
left_shift64(cipheroff, 22, 16) |
left_shift64(hashbit_cnt, 19, 3) |
left_shift64(hashclobber, 18, 1) |
left_shift64(hashoff, 0, 16));
}
/**
* @brief Generate cypto packet descriptor 3
* @ingroup crypto
* designer_vc : designer freeback fmn destination id
* taglen : length in bits of the tag generated by the auth engine
* md5 (128 bits), sha1 (160), sha224 (224), sha384 (384),
* sha512 (512), Kasumi (32), snow3g (32), gcm (128)
* hmacpad : 1 if hmac padding is already done
*/
static __inline__ uint64_t
nlm_crypto_form_pkt_desc3(unsigned int designer_vc, unsigned int taglen,
unsigned int arc4_state_save_l3, unsigned int arc4_save_state,
unsigned int hmacpad)
{
return (left_shift64(designer_vc, 48, 16) |
left_shift64(taglen, 11, 16) |
left_shift64(arc4_state_save_l3, 8, 1) |
left_shift64(arc4_save_state, 6, 1) |
left_shift64(hmacpad, 5, 1));
}
/**
* @brief Generate cypto packet descriptor 4
* @ingroup crypto
* srcfraglen : length of the source fragment(header + data + tail) in bytes
* srcfragaddr : physical address of the srouce fragment
*/
static __inline__ uint64_t
nlm_crypto_form_pkt_desc4(uint64_t srcfraglen,
unsigned int srcfragaddr )
{
return (left_shift64_mask((srcfraglen - 1), 48, 16) |
left_shift64(srcfragaddr, 0, 40));
}
/**
* @brief Generate cypto packet descriptor 5
* @ingroup crypto
* dstfraglen : length of the dst fragment(header + data + tail) in bytes
* chipherout_l3alloc : 1(cipher output is transited through l3 cache)
* cipherclobber : 1 (cipher output will be written as multiples of cachelines,
* no read modify write)
* chiperdst_addr : physical address of the cipher destination address
*/
static __inline__ uint64_t
nlm_crypto_form_pkt_desc5(unsigned int dstfraglen,
unsigned int cipherout_l3alloc, unsigned int cipherclobber,
uint64_t cipherdst_addr)
{
return (left_shift64_mask((dstfraglen - 1), 48, 16) |
left_shift64(cipherout_l3alloc, 46, 1) |
left_shift64(cipherclobber, 41, 1) |
left_shift64(cipherdst_addr, 0, 40));
}
/**
* @brief Generate crypto packet fmn message entry 0
* @ingroup crypto
* freeback_vc: freeback response destination address
* designer_fblen : Designer freeback length, 1 - 4
* designerdesc_valid : designer desc valid or not
* cipher_keylen : cipher key length in bytes
* ctrldesc_addr : physicall address of the control descriptor
*/
static __inline__ uint64_t
nlm_crypto_form_pkt_fmn_entry0(unsigned int freeback_vc,
unsigned int designer_fblen, unsigned int designerdesc_valid,
unsigned int cipher_keylen, uint64_t cntldesc_addr)
{
return (left_shift64(freeback_vc, 48, 16) |
left_shift64_mask(designer_fblen - 1, 46, 2) |
left_shift64(designerdesc_valid, 45, 1) |
left_shift64_mask(((cipher_keylen + 7) >> 3), 40, 5) |
left_shift64(cntldesc_addr >> 6, 0, 34));
}
/**
* @brief Generate crypto packet fmn message entry 1
* @ingroup crypto
* arc4load_state : 1 if load state required 0 otherwise
* hash_keylen : hash key length in bytes
* pktdesc_size : packet descriptor size in bytes
* pktdesc_addr : physicall address of the packet descriptor
*/
static __inline__ uint64_t
nlm_crypto_form_pkt_fmn_entry1(unsigned int arc4load_state,
unsigned int hash_keylen, unsigned int pktdesc_size,
uint64_t pktdesc_addr)
{
return (left_shift64(arc4load_state, 63, 1) |
left_shift64_mask(((hash_keylen + 7) >> 3), 56, 5) |
left_shift64_mask(((pktdesc_size >> 4) - 1), 43, 12) |
left_shift64(pktdesc_addr >> 6, 0, 34));
}
static __inline__ int
nlm_crypto_get_hklen_taglen(enum nlm_hash_algo hashalg,
enum nlm_hash_mode hashmode, unsigned int *taglen, unsigned int *hklen)
{
if (hashalg == NLM_HASH_MD5) {
*taglen = 128;
*hklen = 64;
} else if (hashalg == NLM_HASH_SHA) {
switch (hashmode) {
case NLM_HASH_MODE_SHA1:
*taglen = 160;
*hklen = 64;
break;
case NLM_HASH_MODE_SHA224:
*taglen = 224;
*hklen = 64;
break;
case NLM_HASH_MODE_SHA256:
*taglen = 256;
*hklen = 64;
break;
case NLM_HASH_MODE_SHA384:
*taglen = 384;
*hklen = 128;
break;
case NLM_HASH_MODE_SHA512:
*taglen = 512;
*hklen = 128;
break;
default:
printf("Error : invalid shaid (%s)\n", __func__);
return (-1);
}
} else if (hashalg == NLM_HASH_KASUMI_F9) {
*taglen = 32;
*hklen = 0;
} else if (hashalg == NLM_HASH_SNOW3G_F9) {
*taglen = 32;
*hklen = 0;
} else if (hashmode == NLM_HASH_MODE_XCBC) {
*taglen = 128;
*hklen = 0;
} else if (hashmode == NLM_HASH_MODE_GCM) {
*taglen = 128;
*hklen = 0;
} else if (hashalg == NLM_HASH_BYPASS) {
*taglen = 0;
*hklen = 0;
} else {
printf("Error:Hash alg/mode not found\n");
return (-1);
}
/* TODO : Add remaining cases */
return (0);
}
/**
* @brief Generate fill cryto control info structure
* @ingroup crypto
* hmac : 1 for hash with hmac
* hashalg: see above, hash_alg enums
* hashmode: see above, hash_mode enums
* cipherhalg: see above, cipher_alg enums
* ciphermode: see above, cipher_mode enums
*
*/
static __inline__ int
nlm_crypto_fill_pkt_ctrl(struct nlm_crypto_pkt_ctrl *ctrl, unsigned int hmac,
enum nlm_hash_algo hashalg, enum nlm_hash_mode hashmode,
enum nlm_cipher_algo cipheralg, enum nlm_cipher_mode ciphermode,
unsigned char *cipherkey, unsigned int cipherkeylen,
unsigned char *hashkey, unsigned int hashkeylen)
{
unsigned int taglen = 0, hklen = 0;
ctrl->desc0 = nlm_crypto_form_pkt_ctrl_desc(hmac, hashalg, hashmode,
cipheralg, ciphermode, 0, 0, 0);
memset(ctrl->key, 0, sizeof(ctrl->key));
if (cipherkey)
memcpy(ctrl->key, cipherkey, cipherkeylen);
if (hashkey)
memcpy((unsigned char *)&ctrl->key[(cipherkeylen + 7) / 8],
hashkey, hashkeylen);
if (nlm_crypto_get_hklen_taglen(hashalg, hashmode, &taglen, &hklen)
< 0)
return (-1);
ctrl->cipherkeylen = cipherkeylen;
ctrl->hashkeylen = hklen;
ctrl->taglen = taglen;
/* TODO : add the invalid checks and return error */
return (0);
}
/**
* @brief Top level function for generation pkt desc 0 to 3 for cipher auth
* @ingroup crypto
* ctrl : pointer to control structure
* param : pointer to the param structure
* encrypt : 1(for encrypt) 0(for decrypt)
* hash_source : 1(encrypted data is sent to the auth engine) 0(plain data is
* sent to the auth engine)
* ivoff : iv offset from start of data
* ivlen : iv length in bytes
* hashoff : hash offset from start of data
* hashlen : hash length in bytes
* hmacpad : hmac padding required or not, 1 if already padded
* cipheroff : cipher offset from start of data
* cipherlen : cipher length in bytes
* hashdst_addr : hash destination physical address
*/
static __inline__ void
nlm_crypto_fill_cipher_auth_pkt_param(struct nlm_crypto_pkt_ctrl *ctrl,
struct nlm_crypto_pkt_param *param, unsigned int encrypt,
unsigned int hash_source, unsigned int ivoff, unsigned int ivlen,
unsigned int hashoff, unsigned int hashlen, unsigned int hmacpad,
unsigned int cipheroff, unsigned int cipherlen, unsigned char *hashdst_addr)
{
param->desc0 = nlm_crypto_form_pkt_desc0(0, hash_source, 1, encrypt,
ivlen, vtophys(hashdst_addr));
param->desc1 = nlm_crypto_form_pkt_desc1(cipherlen, hashlen);
param->desc2 = nlm_crypto_form_pkt_desc2(ivoff, 0, cipheroff, 0, 0,
hashoff);
param->desc3 = nlm_crypto_form_pkt_desc3(0, ctrl->taglen, 0, 0,
hmacpad);
}
/**
* @brief Top level function for generation pkt desc 0 to 3 for cipher operation
* @ingroup crypto
* ctrl : pointer to control structure
* param : pointer to the param structure
* encrypt : 1(for encrypt) 0(for decrypt)
* ivoff : iv offset from start of data
* ivlen : iv length in bytes
* cipheroff : cipher offset from start of data
* cipherlen : cipher length in bytes
*/
static __inline__ void
nlm_crypto_fill_cipher_pkt_param(struct nlm_crypto_pkt_ctrl *ctrl,
struct nlm_crypto_pkt_param *param, unsigned int encrypt,
unsigned int ivoff, unsigned int ivlen, unsigned int cipheroff,
unsigned int cipherlen)
{
param->desc0 = nlm_crypto_form_pkt_desc0(0, 0, 0, encrypt, ivlen, 0ULL);
param->desc1 = nlm_crypto_form_pkt_desc1(cipherlen, 1);
param->desc2 = nlm_crypto_form_pkt_desc2(ivoff, 0, cipheroff, 0, 0, 0);
param->desc3 = nlm_crypto_form_pkt_desc3(0, ctrl->taglen, 0, 0, 0);
}
/**
* @brief Top level function for generation pkt desc 0 to 3 for auth operation
* @ingroup crypto
* ctrl : pointer to control structure
* param : pointer to the param structure
* hashoff : hash offset from start of data
* hashlen : hash length in bytes
* hmacpad : hmac padding required or not, 1 if already padded
* hashdst_addr : hash destination physical address
*/
static __inline__ void
nlm_crypto_fill_auth_pkt_param(struct nlm_crypto_pkt_ctrl *ctrl,
struct nlm_crypto_pkt_param *param, unsigned int hashoff,
unsigned int hashlen, unsigned int hmacpad, unsigned char *hashdst_addr)
{
param->desc0 = nlm_crypto_form_pkt_desc0(0, 0, 1, 0, 1,
vtophys(hashdst_addr));
param->desc1 = nlm_crypto_form_pkt_desc1(1, hashlen);
param->desc2 = nlm_crypto_form_pkt_desc2(0, 0, 0, 0, 0, hashoff);
param->desc3 = nlm_crypto_form_pkt_desc3(0, ctrl->taglen, 0, 0,
hmacpad);
}
static __inline__ unsigned int
nlm_crypto_fill_src_seg(struct nlm_crypto_pkt_param *param, int seg,
unsigned char *input, unsigned int inlen)
{
unsigned off = 0, len = 0;
unsigned int remlen = inlen;
for (; remlen > 0;) {
len = remlen > NLM_CRYPTO_MAX_SEG_LEN ?
NLM_CRYPTO_MAX_SEG_LEN : remlen;
param->segment[seg][0] = nlm_crypto_form_pkt_desc4(len,
vtophys(input + off));
remlen -= len;
off += len;
seg++;
}
return (seg);
}
static __inline__ unsigned int
nlm_crypto_fill_dst_seg(struct nlm_crypto_pkt_param *param,
int seg, unsigned char *output, unsigned int outlen)
{
unsigned off = 0, len = 0;
unsigned int remlen = outlen;
for (; remlen > 0;) {
len = remlen > NLM_CRYPTO_MAX_SEG_LEN ?
NLM_CRYPTO_MAX_SEG_LEN : remlen;
param->segment[seg][1] = nlm_crypto_form_pkt_desc5(len, 1, 0,
vtophys(output + off));
remlen -= len;
off += len;
seg++;
}
return (seg);
}
#endif