2538 lines
71 KiB
C
2538 lines
71 KiB
C
/*-
|
|
* Copyright (c) 2009-2011 Spectra Logic Corporation
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions, and the following disclaimer,
|
|
* without modification.
|
|
* 2. Redistributions in binary form must reproduce at minimum a disclaimer
|
|
* substantially similar to the "NO WARRANTY" disclaimer below
|
|
* ("Disclaimer") and any redistribution must be conditioned upon
|
|
* including a substantially similar Disclaimer requirement for further
|
|
* binary redistribution.
|
|
*
|
|
* NO WARRANTY
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTIBILITY AND FITNESS FOR
|
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
* HOLDERS OR CONTRIBUTORS BE LIABLE FOR SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,
|
|
* STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING
|
|
* IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
* POSSIBILITY OF SUCH DAMAGES.
|
|
*
|
|
* Authors: Justin T. Gibbs (Spectra Logic Corporation)
|
|
* Alan Somers (Spectra Logic Corporation)
|
|
* John Suykerbuyk (Spectra Logic Corporation)
|
|
*/
|
|
|
|
#include <sys/cdefs.h>
|
|
__FBSDID("$FreeBSD$");
|
|
|
|
/**
|
|
* \file netback_unit_tests.c
|
|
*
|
|
* \brief Unit tests for the Xen netback driver.
|
|
*
|
|
* Due to the driver's use of static functions, these tests cannot be compiled
|
|
* standalone; they must be #include'd from the driver's .c file.
|
|
*/
|
|
|
|
|
|
/** Helper macro used to snprintf to a buffer and update the buffer pointer */
|
|
#define SNCATF(buffer, buflen, ...) do { \
|
|
size_t new_chars = snprintf(buffer, buflen, __VA_ARGS__); \
|
|
buffer += new_chars; \
|
|
/* be careful; snprintf's return value can be > buflen */ \
|
|
buflen -= MIN(buflen, new_chars); \
|
|
} while (0)
|
|
|
|
/* STRINGIFY and TOSTRING are used only to help turn __LINE__ into a string */
|
|
#define STRINGIFY(x) #x
|
|
#define TOSTRING(x) STRINGIFY(x)
|
|
|
|
/**
|
|
* Writes an error message to buffer if cond is false
|
|
* Note the implied parameters buffer and
|
|
* buflen
|
|
*/
|
|
#define XNB_ASSERT(cond) ({ \
|
|
int passed = (cond); \
|
|
char *_buffer = (buffer); \
|
|
size_t _buflen = (buflen); \
|
|
if (! passed) { \
|
|
strlcat(_buffer, __func__, _buflen); \
|
|
strlcat(_buffer, ":" TOSTRING(__LINE__) \
|
|
" Assertion Error: " #cond "\n", _buflen); \
|
|
} \
|
|
})
|
|
|
|
|
|
/**
|
|
* The signature used by all testcases. If the test writes anything
|
|
* to buffer, then it will be considered a failure
|
|
* \param buffer Return storage for error messages
|
|
* \param buflen The space available in the buffer
|
|
*/
|
|
typedef void testcase_t(char *buffer, size_t buflen);
|
|
|
|
/**
|
|
* Signature used by setup functions
|
|
* \return nonzero on error
|
|
*/
|
|
typedef int setup_t(void);
|
|
|
|
typedef void teardown_t(void);
|
|
|
|
/** A simple test fixture comprising setup, teardown, and test */
|
|
struct test_fixture {
|
|
/** Will be run before the test to allocate and initialize variables */
|
|
setup_t *setup;
|
|
|
|
/** Will be run if setup succeeds */
|
|
testcase_t *test;
|
|
|
|
/** Cleans up test data whether or not the setup suceeded*/
|
|
teardown_t *teardown;
|
|
};
|
|
|
|
typedef struct test_fixture test_fixture_t;
|
|
|
|
static int xnb_get1pkt(struct xnb_pkt *pkt, size_t size, uint16_t flags);
|
|
static int xnb_unit_test_runner(test_fixture_t const tests[], int ntests,
|
|
char *buffer, size_t buflen);
|
|
|
|
static int __unused
|
|
null_setup(void) { return 0; }
|
|
|
|
static void __unused
|
|
null_teardown(void) { }
|
|
|
|
static setup_t setup_pvt_data;
|
|
static teardown_t teardown_pvt_data;
|
|
static testcase_t xnb_ring2pkt_emptyring;
|
|
static testcase_t xnb_ring2pkt_1req;
|
|
static testcase_t xnb_ring2pkt_2req;
|
|
static testcase_t xnb_ring2pkt_3req;
|
|
static testcase_t xnb_ring2pkt_extra;
|
|
static testcase_t xnb_ring2pkt_partial;
|
|
static testcase_t xnb_ring2pkt_wraps;
|
|
static testcase_t xnb_txpkt2rsp_emptypkt;
|
|
static testcase_t xnb_txpkt2rsp_1req;
|
|
static testcase_t xnb_txpkt2rsp_extra;
|
|
static testcase_t xnb_txpkt2rsp_long;
|
|
static testcase_t xnb_txpkt2rsp_invalid;
|
|
static testcase_t xnb_txpkt2rsp_error;
|
|
static testcase_t xnb_txpkt2rsp_wraps;
|
|
static testcase_t xnb_pkt2mbufc_empty;
|
|
static testcase_t xnb_pkt2mbufc_short;
|
|
static testcase_t xnb_pkt2mbufc_csum;
|
|
static testcase_t xnb_pkt2mbufc_1cluster;
|
|
static testcase_t xnb_pkt2mbufc_largecluster;
|
|
static testcase_t xnb_pkt2mbufc_2cluster;
|
|
static testcase_t xnb_txpkt2gnttab_empty;
|
|
static testcase_t xnb_txpkt2gnttab_short;
|
|
static testcase_t xnb_txpkt2gnttab_2req;
|
|
static testcase_t xnb_txpkt2gnttab_2cluster;
|
|
static testcase_t xnb_update_mbufc_short;
|
|
static testcase_t xnb_update_mbufc_2req;
|
|
static testcase_t xnb_update_mbufc_2cluster;
|
|
static testcase_t xnb_mbufc2pkt_empty;
|
|
static testcase_t xnb_mbufc2pkt_short;
|
|
static testcase_t xnb_mbufc2pkt_1cluster;
|
|
static testcase_t xnb_mbufc2pkt_2short;
|
|
static testcase_t xnb_mbufc2pkt_long;
|
|
static testcase_t xnb_mbufc2pkt_extra;
|
|
static testcase_t xnb_mbufc2pkt_nospace;
|
|
static testcase_t xnb_rxpkt2gnttab_empty;
|
|
static testcase_t xnb_rxpkt2gnttab_short;
|
|
static testcase_t xnb_rxpkt2gnttab_2req;
|
|
static testcase_t xnb_rxpkt2rsp_empty;
|
|
static testcase_t xnb_rxpkt2rsp_short;
|
|
static testcase_t xnb_rxpkt2rsp_extra;
|
|
static testcase_t xnb_rxpkt2rsp_2short;
|
|
static testcase_t xnb_rxpkt2rsp_2slots;
|
|
static testcase_t xnb_rxpkt2rsp_copyerror;
|
|
static testcase_t xnb_sscanf_llu;
|
|
static testcase_t xnb_sscanf_lld;
|
|
static testcase_t xnb_sscanf_hhu;
|
|
static testcase_t xnb_sscanf_hhd;
|
|
static testcase_t xnb_sscanf_hhn;
|
|
|
|
#if defined(INET) || defined(INET6)
|
|
/* TODO: add test cases for xnb_add_mbuf_cksum for IPV6 tcp and udp */
|
|
static testcase_t xnb_add_mbuf_cksum_arp;
|
|
static testcase_t xnb_add_mbuf_cksum_tcp;
|
|
static testcase_t xnb_add_mbuf_cksum_udp;
|
|
static testcase_t xnb_add_mbuf_cksum_icmp;
|
|
static testcase_t xnb_add_mbuf_cksum_tcp_swcksum;
|
|
static void xnb_fill_eh_and_ip(struct mbuf *m, uint16_t ip_len,
|
|
uint16_t ip_id, uint16_t ip_p,
|
|
uint16_t ip_off, uint16_t ip_sum);
|
|
static void xnb_fill_tcp(struct mbuf *m);
|
|
#endif /* INET || INET6 */
|
|
|
|
/** Private data used by unit tests */
|
|
static struct {
|
|
gnttab_copy_table gnttab;
|
|
netif_rx_back_ring_t rxb;
|
|
netif_rx_front_ring_t rxf;
|
|
netif_tx_back_ring_t txb;
|
|
netif_tx_front_ring_t txf;
|
|
struct ifnet* ifp;
|
|
netif_rx_sring_t* rxs;
|
|
netif_tx_sring_t* txs;
|
|
} xnb_unit_pvt;
|
|
|
|
static inline void safe_m_freem(struct mbuf **ppMbuf) {
|
|
if (*ppMbuf != NULL) {
|
|
m_freem(*ppMbuf);
|
|
*ppMbuf = NULL;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* The unit test runner. It will run every supplied test and return an
|
|
* output message as a string
|
|
* \param tests An array of tests. Every test will be attempted.
|
|
* \param ntests The length of tests
|
|
* \param buffer Return storage for the result string
|
|
* \param buflen The length of buffer
|
|
* \return The number of tests that failed
|
|
*/
|
|
static int
|
|
xnb_unit_test_runner(test_fixture_t const tests[], int ntests, char *buffer,
|
|
size_t buflen)
|
|
{
|
|
int i;
|
|
int n_passes;
|
|
int n_failures = 0;
|
|
|
|
for (i = 0; i < ntests; i++) {
|
|
int error = tests[i].setup();
|
|
if (error != 0) {
|
|
SNCATF(buffer, buflen,
|
|
"Setup failed for test idx %d\n", i);
|
|
n_failures++;
|
|
} else {
|
|
size_t new_chars;
|
|
|
|
tests[i].test(buffer, buflen);
|
|
new_chars = strnlen(buffer, buflen);
|
|
buffer += new_chars;
|
|
buflen -= new_chars;
|
|
|
|
if (new_chars > 0) {
|
|
n_failures++;
|
|
}
|
|
}
|
|
tests[i].teardown();
|
|
}
|
|
|
|
n_passes = ntests - n_failures;
|
|
if (n_passes > 0) {
|
|
SNCATF(buffer, buflen, "%d Tests Passed\n", n_passes);
|
|
}
|
|
if (n_failures > 0) {
|
|
SNCATF(buffer, buflen, "%d Tests FAILED\n", n_failures);
|
|
}
|
|
|
|
return n_failures;
|
|
}
|
|
|
|
/** Number of unit tests. Must match the length of the tests array below */
|
|
#define TOTAL_TESTS (53)
|
|
/**
|
|
* Max memory available for returning results. 400 chars/test should give
|
|
* enough space for a five line error message for every test
|
|
*/
|
|
#define TOTAL_BUFLEN (400 * TOTAL_TESTS + 2)
|
|
|
|
/**
|
|
* Called from userspace by a sysctl. Runs all internal unit tests, and
|
|
* returns the results to userspace as a string
|
|
* \param oidp unused
|
|
* \param arg1 pointer to an xnb_softc for a specific xnb device
|
|
* \param arg2 unused
|
|
* \param req sysctl access structure
|
|
* \return a string via the special SYSCTL_OUT macro.
|
|
*/
|
|
|
|
static int
|
|
xnb_unit_test_main(SYSCTL_HANDLER_ARGS) {
|
|
test_fixture_t const tests[TOTAL_TESTS] = {
|
|
{setup_pvt_data, xnb_ring2pkt_emptyring, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_ring2pkt_1req, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_ring2pkt_2req, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_ring2pkt_3req, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_ring2pkt_extra, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_ring2pkt_partial, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_ring2pkt_wraps, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_txpkt2rsp_emptypkt, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_txpkt2rsp_1req, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_txpkt2rsp_extra, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_txpkt2rsp_long, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_txpkt2rsp_invalid, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_txpkt2rsp_error, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_txpkt2rsp_wraps, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_pkt2mbufc_empty, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_pkt2mbufc_short, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_pkt2mbufc_csum, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_pkt2mbufc_1cluster, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_pkt2mbufc_largecluster, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_pkt2mbufc_2cluster, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_txpkt2gnttab_empty, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_txpkt2gnttab_short, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_txpkt2gnttab_2req, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_txpkt2gnttab_2cluster, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_update_mbufc_short, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_update_mbufc_2req, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_update_mbufc_2cluster, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_mbufc2pkt_empty, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_mbufc2pkt_short, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_mbufc2pkt_1cluster, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_mbufc2pkt_2short, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_mbufc2pkt_long, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_mbufc2pkt_extra, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_mbufc2pkt_nospace, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_rxpkt2gnttab_empty, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_rxpkt2gnttab_short, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_rxpkt2gnttab_2req, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_rxpkt2rsp_empty, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_rxpkt2rsp_short, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_rxpkt2rsp_extra, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_rxpkt2rsp_2short, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_rxpkt2rsp_2slots, teardown_pvt_data},
|
|
{setup_pvt_data, xnb_rxpkt2rsp_copyerror, teardown_pvt_data},
|
|
#if defined(INET) || defined(INET6)
|
|
{null_setup, xnb_add_mbuf_cksum_arp, null_teardown},
|
|
{null_setup, xnb_add_mbuf_cksum_icmp, null_teardown},
|
|
{null_setup, xnb_add_mbuf_cksum_tcp, null_teardown},
|
|
{null_setup, xnb_add_mbuf_cksum_tcp_swcksum, null_teardown},
|
|
{null_setup, xnb_add_mbuf_cksum_udp, null_teardown},
|
|
#endif
|
|
{null_setup, xnb_sscanf_hhd, null_teardown},
|
|
{null_setup, xnb_sscanf_hhu, null_teardown},
|
|
{null_setup, xnb_sscanf_lld, null_teardown},
|
|
{null_setup, xnb_sscanf_llu, null_teardown},
|
|
{null_setup, xnb_sscanf_hhn, null_teardown},
|
|
};
|
|
/**
|
|
* results is static so that the data will persist after this function
|
|
* returns. The sysctl code expects us to return a constant string.
|
|
* \todo: the static variable is not thread safe. Put a mutex around
|
|
* it.
|
|
*/
|
|
static char results[TOTAL_BUFLEN];
|
|
|
|
/* empty the result strings */
|
|
results[0] = 0;
|
|
xnb_unit_test_runner(tests, TOTAL_TESTS, results, TOTAL_BUFLEN);
|
|
|
|
return (SYSCTL_OUT(req, results, strnlen(results, TOTAL_BUFLEN)));
|
|
}
|
|
|
|
static int
|
|
setup_pvt_data(void)
|
|
{
|
|
int error = 0;
|
|
|
|
bzero(xnb_unit_pvt.gnttab, sizeof(xnb_unit_pvt.gnttab));
|
|
|
|
xnb_unit_pvt.txs = malloc(PAGE_SIZE, M_XENNETBACK, M_WAITOK|M_ZERO);
|
|
if (xnb_unit_pvt.txs != NULL) {
|
|
SHARED_RING_INIT(xnb_unit_pvt.txs);
|
|
BACK_RING_INIT(&xnb_unit_pvt.txb, xnb_unit_pvt.txs, PAGE_SIZE);
|
|
FRONT_RING_INIT(&xnb_unit_pvt.txf, xnb_unit_pvt.txs, PAGE_SIZE);
|
|
} else {
|
|
error = 1;
|
|
}
|
|
|
|
xnb_unit_pvt.ifp = if_alloc(IFT_ETHER);
|
|
if (xnb_unit_pvt.ifp == NULL) {
|
|
error = 1;
|
|
}
|
|
|
|
xnb_unit_pvt.rxs = malloc(PAGE_SIZE, M_XENNETBACK, M_WAITOK|M_ZERO);
|
|
if (xnb_unit_pvt.rxs != NULL) {
|
|
SHARED_RING_INIT(xnb_unit_pvt.rxs);
|
|
BACK_RING_INIT(&xnb_unit_pvt.rxb, xnb_unit_pvt.rxs, PAGE_SIZE);
|
|
FRONT_RING_INIT(&xnb_unit_pvt.rxf, xnb_unit_pvt.rxs, PAGE_SIZE);
|
|
} else {
|
|
error = 1;
|
|
}
|
|
|
|
return error;
|
|
}
|
|
|
|
static void
|
|
teardown_pvt_data(void)
|
|
{
|
|
if (xnb_unit_pvt.txs != NULL) {
|
|
free(xnb_unit_pvt.txs, M_XENNETBACK);
|
|
}
|
|
if (xnb_unit_pvt.rxs != NULL) {
|
|
free(xnb_unit_pvt.rxs, M_XENNETBACK);
|
|
}
|
|
if (xnb_unit_pvt.ifp != NULL) {
|
|
if_free(xnb_unit_pvt.ifp);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Verify that xnb_ring2pkt will not consume any requests from an empty ring
|
|
*/
|
|
static void
|
|
xnb_ring2pkt_emptyring(char *buffer, size_t buflen)
|
|
{
|
|
struct xnb_pkt pkt;
|
|
int num_consumed;
|
|
|
|
num_consumed = xnb_ring2pkt(&pkt, &xnb_unit_pvt.txb,
|
|
xnb_unit_pvt.txb.req_cons);
|
|
XNB_ASSERT(num_consumed == 0);
|
|
}
|
|
|
|
/**
|
|
* Verify that xnb_ring2pkt can convert a single request packet correctly
|
|
*/
|
|
static void
|
|
xnb_ring2pkt_1req(char *buffer, size_t buflen)
|
|
{
|
|
struct xnb_pkt pkt;
|
|
int num_consumed;
|
|
struct netif_tx_request *req;
|
|
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
|
|
req->flags = 0;
|
|
req->size = 69; /* arbitrary number for test */
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
RING_PUSH_REQUESTS(&xnb_unit_pvt.txf);
|
|
|
|
num_consumed = xnb_ring2pkt(&pkt, &xnb_unit_pvt.txb,
|
|
xnb_unit_pvt.txb.req_cons);
|
|
XNB_ASSERT(num_consumed == 1);
|
|
XNB_ASSERT(pkt.size == 69);
|
|
XNB_ASSERT(pkt.car_size == 69);
|
|
XNB_ASSERT(pkt.flags == 0);
|
|
XNB_ASSERT(xnb_pkt_is_valid(&pkt));
|
|
XNB_ASSERT(pkt.list_len == 1);
|
|
XNB_ASSERT(pkt.car == 0);
|
|
}
|
|
|
|
/**
|
|
* Verify that xnb_ring2pkt can convert a two request packet correctly.
|
|
* This tests handling of the MORE_DATA flag and cdr
|
|
*/
|
|
static void
|
|
xnb_ring2pkt_2req(char *buffer, size_t buflen)
|
|
{
|
|
struct xnb_pkt pkt;
|
|
int num_consumed;
|
|
struct netif_tx_request *req;
|
|
RING_IDX start_idx = xnb_unit_pvt.txf.req_prod_pvt;
|
|
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->flags = NETTXF_more_data;
|
|
req->size = 100;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->flags = 0;
|
|
req->size = 40;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
RING_PUSH_REQUESTS(&xnb_unit_pvt.txf);
|
|
|
|
num_consumed = xnb_ring2pkt(&pkt, &xnb_unit_pvt.txb,
|
|
xnb_unit_pvt.txb.req_cons);
|
|
XNB_ASSERT(num_consumed == 2);
|
|
XNB_ASSERT(pkt.size == 100);
|
|
XNB_ASSERT(pkt.car_size == 60);
|
|
XNB_ASSERT(pkt.flags == 0);
|
|
XNB_ASSERT(xnb_pkt_is_valid(&pkt));
|
|
XNB_ASSERT(pkt.list_len == 2);
|
|
XNB_ASSERT(pkt.car == start_idx);
|
|
XNB_ASSERT(pkt.cdr == start_idx + 1);
|
|
}
|
|
|
|
/**
|
|
* Verify that xnb_ring2pkt can convert a three request packet correctly
|
|
*/
|
|
static void
|
|
xnb_ring2pkt_3req(char *buffer, size_t buflen)
|
|
{
|
|
struct xnb_pkt pkt;
|
|
int num_consumed;
|
|
struct netif_tx_request *req;
|
|
RING_IDX start_idx = xnb_unit_pvt.txf.req_prod_pvt;
|
|
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->flags = NETTXF_more_data;
|
|
req->size = 200;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->flags = NETTXF_more_data;
|
|
req->size = 40;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->flags = 0;
|
|
req->size = 50;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
RING_PUSH_REQUESTS(&xnb_unit_pvt.txf);
|
|
|
|
num_consumed = xnb_ring2pkt(&pkt, &xnb_unit_pvt.txb,
|
|
xnb_unit_pvt.txb.req_cons);
|
|
XNB_ASSERT(num_consumed == 3);
|
|
XNB_ASSERT(pkt.size == 200);
|
|
XNB_ASSERT(pkt.car_size == 110);
|
|
XNB_ASSERT(pkt.flags == 0);
|
|
XNB_ASSERT(xnb_pkt_is_valid(&pkt));
|
|
XNB_ASSERT(pkt.list_len == 3);
|
|
XNB_ASSERT(pkt.car == start_idx);
|
|
XNB_ASSERT(pkt.cdr == start_idx + 1);
|
|
XNB_ASSERT(RING_GET_REQUEST(&xnb_unit_pvt.txb, pkt.cdr + 1) == req);
|
|
}
|
|
|
|
/**
|
|
* Verify that xnb_ring2pkt can read extra inf
|
|
*/
|
|
static void
|
|
xnb_ring2pkt_extra(char *buffer, size_t buflen)
|
|
{
|
|
struct xnb_pkt pkt;
|
|
int num_consumed;
|
|
struct netif_tx_request *req;
|
|
struct netif_extra_info *ext;
|
|
RING_IDX start_idx = xnb_unit_pvt.txf.req_prod_pvt;
|
|
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->flags = NETTXF_extra_info | NETTXF_more_data;
|
|
req->size = 150;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
ext = (struct netif_extra_info*) RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
ext->flags = 0;
|
|
ext->type = XEN_NETIF_EXTRA_TYPE_GSO;
|
|
ext->u.gso.size = 250;
|
|
ext->u.gso.type = XEN_NETIF_GSO_TYPE_TCPV4;
|
|
ext->u.gso.features = 0;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->flags = 0;
|
|
req->size = 50;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
RING_PUSH_REQUESTS(&xnb_unit_pvt.txf);
|
|
|
|
num_consumed = xnb_ring2pkt(&pkt, &xnb_unit_pvt.txb,
|
|
xnb_unit_pvt.txb.req_cons);
|
|
XNB_ASSERT(num_consumed == 3);
|
|
XNB_ASSERT(pkt.extra.flags == 0);
|
|
XNB_ASSERT(pkt.extra.type == XEN_NETIF_EXTRA_TYPE_GSO);
|
|
XNB_ASSERT(pkt.extra.u.gso.size == 250);
|
|
XNB_ASSERT(pkt.extra.u.gso.type = XEN_NETIF_GSO_TYPE_TCPV4);
|
|
XNB_ASSERT(pkt.size == 150);
|
|
XNB_ASSERT(pkt.car_size == 100);
|
|
XNB_ASSERT(pkt.flags == NETTXF_extra_info);
|
|
XNB_ASSERT(xnb_pkt_is_valid(&pkt));
|
|
XNB_ASSERT(pkt.list_len == 2);
|
|
XNB_ASSERT(pkt.car == start_idx);
|
|
XNB_ASSERT(pkt.cdr == start_idx + 2);
|
|
XNB_ASSERT(RING_GET_REQUEST(&xnb_unit_pvt.txb, pkt.cdr) == req);
|
|
}
|
|
|
|
/**
|
|
* Verify that xnb_ring2pkt will consume no requests if the entire packet is
|
|
* not yet in the ring
|
|
*/
|
|
static void
|
|
xnb_ring2pkt_partial(char *buffer, size_t buflen)
|
|
{
|
|
struct xnb_pkt pkt;
|
|
int num_consumed;
|
|
struct netif_tx_request *req;
|
|
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->flags = NETTXF_more_data;
|
|
req->size = 150;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
RING_PUSH_REQUESTS(&xnb_unit_pvt.txf);
|
|
|
|
num_consumed = xnb_ring2pkt(&pkt, &xnb_unit_pvt.txb,
|
|
xnb_unit_pvt.txb.req_cons);
|
|
XNB_ASSERT(num_consumed == 0);
|
|
XNB_ASSERT(! xnb_pkt_is_valid(&pkt));
|
|
}
|
|
|
|
/**
|
|
* Verity that xnb_ring2pkt can read a packet whose requests wrap around
|
|
* the end of the ring
|
|
*/
|
|
static void
|
|
xnb_ring2pkt_wraps(char *buffer, size_t buflen)
|
|
{
|
|
struct xnb_pkt pkt;
|
|
int num_consumed;
|
|
struct netif_tx_request *req;
|
|
unsigned int rsize;
|
|
|
|
/*
|
|
* Manually tweak the ring indices to create a ring with no responses
|
|
* and the next request slot at position 2 from the end
|
|
*/
|
|
rsize = RING_SIZE(&xnb_unit_pvt.txf);
|
|
xnb_unit_pvt.txf.req_prod_pvt = rsize - 2;
|
|
xnb_unit_pvt.txf.rsp_cons = rsize - 2;
|
|
xnb_unit_pvt.txs->req_prod = rsize - 2;
|
|
xnb_unit_pvt.txs->req_event = rsize - 1;
|
|
xnb_unit_pvt.txs->rsp_prod = rsize - 2;
|
|
xnb_unit_pvt.txs->rsp_event = rsize - 1;
|
|
xnb_unit_pvt.txb.rsp_prod_pvt = rsize - 2;
|
|
xnb_unit_pvt.txb.req_cons = rsize - 2;
|
|
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->flags = NETTXF_more_data;
|
|
req->size = 550;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->flags = NETTXF_more_data;
|
|
req->size = 100;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->flags = 0;
|
|
req->size = 50;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
RING_PUSH_REQUESTS(&xnb_unit_pvt.txf);
|
|
|
|
num_consumed = xnb_ring2pkt(&pkt, &xnb_unit_pvt.txb,
|
|
xnb_unit_pvt.txb.req_cons);
|
|
XNB_ASSERT(num_consumed == 3);
|
|
XNB_ASSERT(xnb_pkt_is_valid(&pkt));
|
|
XNB_ASSERT(pkt.list_len == 3);
|
|
XNB_ASSERT(RING_GET_REQUEST(&xnb_unit_pvt.txb, pkt.cdr + 1) == req);
|
|
}
|
|
|
|
|
|
/**
|
|
* xnb_txpkt2rsp should do nothing for an empty packet
|
|
*/
|
|
static void
|
|
xnb_txpkt2rsp_emptypkt(char *buffer, size_t buflen)
|
|
{
|
|
int num_consumed;
|
|
struct xnb_pkt pkt;
|
|
netif_tx_back_ring_t txb_backup = xnb_unit_pvt.txb;
|
|
netif_tx_sring_t txs_backup = *xnb_unit_pvt.txs;
|
|
pkt.list_len = 0;
|
|
|
|
/* must call xnb_ring2pkt just to intialize pkt */
|
|
num_consumed = xnb_ring2pkt(&pkt, &xnb_unit_pvt.txb,
|
|
xnb_unit_pvt.txb.req_cons);
|
|
xnb_txpkt2rsp(&pkt, &xnb_unit_pvt.txb, 0);
|
|
XNB_ASSERT(
|
|
memcmp(&txb_backup, &xnb_unit_pvt.txb, sizeof(txb_backup)) == 0);
|
|
XNB_ASSERT(
|
|
memcmp(&txs_backup, xnb_unit_pvt.txs, sizeof(txs_backup)) == 0);
|
|
}
|
|
|
|
/**
|
|
* xnb_txpkt2rsp responding to one request
|
|
*/
|
|
static void
|
|
xnb_txpkt2rsp_1req(char *buffer, size_t buflen)
|
|
{
|
|
uint16_t num_consumed;
|
|
struct xnb_pkt pkt;
|
|
struct netif_tx_request *req;
|
|
struct netif_tx_response *rsp;
|
|
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->size = 1000;
|
|
req->flags = 0;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
RING_PUSH_REQUESTS(&xnb_unit_pvt.txf);
|
|
|
|
num_consumed = xnb_ring2pkt(&pkt, &xnb_unit_pvt.txb,
|
|
xnb_unit_pvt.txb.req_cons);
|
|
xnb_unit_pvt.txb.req_cons += num_consumed;
|
|
|
|
xnb_txpkt2rsp(&pkt, &xnb_unit_pvt.txb, 0);
|
|
rsp = RING_GET_RESPONSE(&xnb_unit_pvt.txb, xnb_unit_pvt.txf.rsp_cons);
|
|
|
|
XNB_ASSERT(
|
|
xnb_unit_pvt.txb.rsp_prod_pvt == xnb_unit_pvt.txs->req_prod);
|
|
XNB_ASSERT(rsp->id == req->id);
|
|
XNB_ASSERT(rsp->status == NETIF_RSP_OKAY);
|
|
};
|
|
|
|
/**
|
|
* xnb_txpkt2rsp responding to 1 data request and 1 extra info
|
|
*/
|
|
static void
|
|
xnb_txpkt2rsp_extra(char *buffer, size_t buflen)
|
|
{
|
|
uint16_t num_consumed;
|
|
struct xnb_pkt pkt;
|
|
struct netif_tx_request *req;
|
|
netif_extra_info_t *ext;
|
|
struct netif_tx_response *rsp;
|
|
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->size = 1000;
|
|
req->flags = NETTXF_extra_info;
|
|
req->id = 69;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
ext = (netif_extra_info_t*) RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
ext->type = XEN_NETIF_EXTRA_TYPE_GSO;
|
|
ext->flags = 0;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
RING_PUSH_REQUESTS(&xnb_unit_pvt.txf);
|
|
|
|
num_consumed = xnb_ring2pkt(&pkt, &xnb_unit_pvt.txb,
|
|
xnb_unit_pvt.txb.req_cons);
|
|
xnb_unit_pvt.txb.req_cons += num_consumed;
|
|
|
|
xnb_txpkt2rsp(&pkt, &xnb_unit_pvt.txb, 0);
|
|
|
|
XNB_ASSERT(
|
|
xnb_unit_pvt.txb.rsp_prod_pvt == xnb_unit_pvt.txs->req_prod);
|
|
|
|
rsp = RING_GET_RESPONSE(&xnb_unit_pvt.txb, xnb_unit_pvt.txf.rsp_cons);
|
|
XNB_ASSERT(rsp->id == req->id);
|
|
XNB_ASSERT(rsp->status == NETIF_RSP_OKAY);
|
|
|
|
rsp = RING_GET_RESPONSE(&xnb_unit_pvt.txb,
|
|
xnb_unit_pvt.txf.rsp_cons + 1);
|
|
XNB_ASSERT(rsp->status == NETIF_RSP_NULL);
|
|
};
|
|
|
|
/**
|
|
* xnb_pkg2rsp responding to 3 data requests and 1 extra info
|
|
*/
|
|
static void
|
|
xnb_txpkt2rsp_long(char *buffer, size_t buflen)
|
|
{
|
|
uint16_t num_consumed;
|
|
struct xnb_pkt pkt;
|
|
struct netif_tx_request *req;
|
|
netif_extra_info_t *ext;
|
|
struct netif_tx_response *rsp;
|
|
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->size = 1000;
|
|
req->flags = NETTXF_extra_info | NETTXF_more_data;
|
|
req->id = 254;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
ext = (netif_extra_info_t*) RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
ext->type = XEN_NETIF_EXTRA_TYPE_GSO;
|
|
ext->flags = 0;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->size = 300;
|
|
req->flags = NETTXF_more_data;
|
|
req->id = 1034;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->size = 400;
|
|
req->flags = 0;
|
|
req->id = 34;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
RING_PUSH_REQUESTS(&xnb_unit_pvt.txf);
|
|
|
|
num_consumed = xnb_ring2pkt(&pkt, &xnb_unit_pvt.txb,
|
|
xnb_unit_pvt.txb.req_cons);
|
|
xnb_unit_pvt.txb.req_cons += num_consumed;
|
|
|
|
xnb_txpkt2rsp(&pkt, &xnb_unit_pvt.txb, 0);
|
|
|
|
XNB_ASSERT(
|
|
xnb_unit_pvt.txb.rsp_prod_pvt == xnb_unit_pvt.txs->req_prod);
|
|
|
|
rsp = RING_GET_RESPONSE(&xnb_unit_pvt.txb, xnb_unit_pvt.txf.rsp_cons);
|
|
XNB_ASSERT(rsp->id ==
|
|
RING_GET_REQUEST(&xnb_unit_pvt.txf, 0)->id);
|
|
XNB_ASSERT(rsp->status == NETIF_RSP_OKAY);
|
|
|
|
rsp = RING_GET_RESPONSE(&xnb_unit_pvt.txb,
|
|
xnb_unit_pvt.txf.rsp_cons + 1);
|
|
XNB_ASSERT(rsp->status == NETIF_RSP_NULL);
|
|
|
|
rsp = RING_GET_RESPONSE(&xnb_unit_pvt.txb,
|
|
xnb_unit_pvt.txf.rsp_cons + 2);
|
|
XNB_ASSERT(rsp->id ==
|
|
RING_GET_REQUEST(&xnb_unit_pvt.txf, 2)->id);
|
|
XNB_ASSERT(rsp->status == NETIF_RSP_OKAY);
|
|
|
|
rsp = RING_GET_RESPONSE(&xnb_unit_pvt.txb,
|
|
xnb_unit_pvt.txf.rsp_cons + 3);
|
|
XNB_ASSERT(rsp->id ==
|
|
RING_GET_REQUEST(&xnb_unit_pvt.txf, 3)->id);
|
|
XNB_ASSERT(rsp->status == NETIF_RSP_OKAY);
|
|
}
|
|
|
|
/**
|
|
* xnb_txpkt2rsp responding to an invalid packet.
|
|
* Note: this test will result in an error message being printed to the console
|
|
* such as:
|
|
* xnb(xnb_ring2pkt:1306): Unknown extra info type 255. Discarding packet
|
|
*/
|
|
static void
|
|
xnb_txpkt2rsp_invalid(char *buffer, size_t buflen)
|
|
{
|
|
uint16_t num_consumed;
|
|
struct xnb_pkt pkt;
|
|
struct netif_tx_request *req;
|
|
netif_extra_info_t *ext;
|
|
struct netif_tx_response *rsp;
|
|
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->size = 1000;
|
|
req->flags = NETTXF_extra_info;
|
|
req->id = 69;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
ext = (netif_extra_info_t*) RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
ext->type = 0xFF; /* Invalid extra type */
|
|
ext->flags = 0;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
RING_PUSH_REQUESTS(&xnb_unit_pvt.txf);
|
|
|
|
num_consumed = xnb_ring2pkt(&pkt, &xnb_unit_pvt.txb,
|
|
xnb_unit_pvt.txb.req_cons);
|
|
xnb_unit_pvt.txb.req_cons += num_consumed;
|
|
XNB_ASSERT(! xnb_pkt_is_valid(&pkt));
|
|
|
|
xnb_txpkt2rsp(&pkt, &xnb_unit_pvt.txb, 0);
|
|
|
|
XNB_ASSERT(
|
|
xnb_unit_pvt.txb.rsp_prod_pvt == xnb_unit_pvt.txs->req_prod);
|
|
|
|
rsp = RING_GET_RESPONSE(&xnb_unit_pvt.txb, xnb_unit_pvt.txf.rsp_cons);
|
|
XNB_ASSERT(rsp->id == req->id);
|
|
XNB_ASSERT(rsp->status == NETIF_RSP_ERROR);
|
|
|
|
rsp = RING_GET_RESPONSE(&xnb_unit_pvt.txb,
|
|
xnb_unit_pvt.txf.rsp_cons + 1);
|
|
XNB_ASSERT(rsp->status == NETIF_RSP_NULL);
|
|
};
|
|
|
|
/**
|
|
* xnb_txpkt2rsp responding to one request which caused an error
|
|
*/
|
|
static void
|
|
xnb_txpkt2rsp_error(char *buffer, size_t buflen)
|
|
{
|
|
uint16_t num_consumed;
|
|
struct xnb_pkt pkt;
|
|
struct netif_tx_request *req;
|
|
struct netif_tx_response *rsp;
|
|
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->size = 1000;
|
|
req->flags = 0;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
RING_PUSH_REQUESTS(&xnb_unit_pvt.txf);
|
|
|
|
num_consumed = xnb_ring2pkt(&pkt, &xnb_unit_pvt.txb,
|
|
xnb_unit_pvt.txb.req_cons);
|
|
xnb_unit_pvt.txb.req_cons += num_consumed;
|
|
|
|
xnb_txpkt2rsp(&pkt, &xnb_unit_pvt.txb, 1);
|
|
rsp = RING_GET_RESPONSE(&xnb_unit_pvt.txb, xnb_unit_pvt.txf.rsp_cons);
|
|
|
|
XNB_ASSERT(
|
|
xnb_unit_pvt.txb.rsp_prod_pvt == xnb_unit_pvt.txs->req_prod);
|
|
XNB_ASSERT(rsp->id == req->id);
|
|
XNB_ASSERT(rsp->status == NETIF_RSP_ERROR);
|
|
};
|
|
|
|
/**
|
|
* xnb_txpkt2rsp's responses wrap around the end of the ring
|
|
*/
|
|
static void
|
|
xnb_txpkt2rsp_wraps(char *buffer, size_t buflen)
|
|
{
|
|
struct xnb_pkt pkt;
|
|
int num_consumed;
|
|
struct netif_tx_request *req;
|
|
struct netif_tx_response *rsp;
|
|
unsigned int rsize;
|
|
|
|
/*
|
|
* Manually tweak the ring indices to create a ring with no responses
|
|
* and the next request slot at position 2 from the end
|
|
*/
|
|
rsize = RING_SIZE(&xnb_unit_pvt.txf);
|
|
xnb_unit_pvt.txf.req_prod_pvt = rsize - 2;
|
|
xnb_unit_pvt.txf.rsp_cons = rsize - 2;
|
|
xnb_unit_pvt.txs->req_prod = rsize - 2;
|
|
xnb_unit_pvt.txs->req_event = rsize - 1;
|
|
xnb_unit_pvt.txs->rsp_prod = rsize - 2;
|
|
xnb_unit_pvt.txs->rsp_event = rsize - 1;
|
|
xnb_unit_pvt.txb.rsp_prod_pvt = rsize - 2;
|
|
xnb_unit_pvt.txb.req_cons = rsize - 2;
|
|
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->flags = NETTXF_more_data;
|
|
req->size = 550;
|
|
req->id = 1;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->flags = NETTXF_more_data;
|
|
req->size = 100;
|
|
req->id = 2;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->flags = 0;
|
|
req->size = 50;
|
|
req->id = 3;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
RING_PUSH_REQUESTS(&xnb_unit_pvt.txf);
|
|
|
|
num_consumed = xnb_ring2pkt(&pkt, &xnb_unit_pvt.txb,
|
|
xnb_unit_pvt.txb.req_cons);
|
|
|
|
xnb_txpkt2rsp(&pkt, &xnb_unit_pvt.txb, 0);
|
|
|
|
XNB_ASSERT(
|
|
xnb_unit_pvt.txb.rsp_prod_pvt == xnb_unit_pvt.txs->req_prod);
|
|
rsp = RING_GET_RESPONSE(&xnb_unit_pvt.txb,
|
|
xnb_unit_pvt.txf.rsp_cons + 2);
|
|
XNB_ASSERT(rsp->id == req->id);
|
|
XNB_ASSERT(rsp->status == NETIF_RSP_OKAY);
|
|
}
|
|
|
|
|
|
/**
|
|
* Helper function used to setup pkt2mbufc tests
|
|
* \param size size in bytes of the single request to push to the ring
|
|
* \param flags optional flags to put in the netif request
|
|
* \param[out] pkt the returned packet object
|
|
* \return number of requests consumed from the ring
|
|
*/
|
|
static int
|
|
xnb_get1pkt(struct xnb_pkt *pkt, size_t size, uint16_t flags)
|
|
{
|
|
struct netif_tx_request *req;
|
|
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->flags = flags;
|
|
req->size = size;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
RING_PUSH_REQUESTS(&xnb_unit_pvt.txf);
|
|
|
|
return xnb_ring2pkt(pkt, &xnb_unit_pvt.txb,
|
|
xnb_unit_pvt.txb.req_cons);
|
|
}
|
|
|
|
/**
|
|
* xnb_pkt2mbufc on an empty packet
|
|
*/
|
|
static void
|
|
xnb_pkt2mbufc_empty(char *buffer, size_t buflen)
|
|
{
|
|
int num_consumed;
|
|
struct xnb_pkt pkt;
|
|
struct mbuf *pMbuf;
|
|
pkt.list_len = 0;
|
|
|
|
/* must call xnb_ring2pkt just to intialize pkt */
|
|
num_consumed = xnb_ring2pkt(&pkt, &xnb_unit_pvt.txb,
|
|
xnb_unit_pvt.txb.req_cons);
|
|
pkt.size = 0;
|
|
pMbuf = xnb_pkt2mbufc(&pkt, xnb_unit_pvt.ifp);
|
|
safe_m_freem(&pMbuf);
|
|
}
|
|
|
|
/**
|
|
* xnb_pkt2mbufc on short packet that can fit in an mbuf internal buffer
|
|
*/
|
|
static void
|
|
xnb_pkt2mbufc_short(char *buffer, size_t buflen)
|
|
{
|
|
const size_t size = MINCLSIZE - 1;
|
|
struct xnb_pkt pkt;
|
|
struct mbuf *pMbuf;
|
|
|
|
xnb_get1pkt(&pkt, size, 0);
|
|
|
|
pMbuf = xnb_pkt2mbufc(&pkt, xnb_unit_pvt.ifp);
|
|
XNB_ASSERT(M_TRAILINGSPACE(pMbuf) >= size);
|
|
safe_m_freem(&pMbuf);
|
|
}
|
|
|
|
/**
|
|
* xnb_pkt2mbufc on short packet whose checksum was validated by the netfron
|
|
*/
|
|
static void
|
|
xnb_pkt2mbufc_csum(char *buffer, size_t buflen)
|
|
{
|
|
const size_t size = MINCLSIZE - 1;
|
|
struct xnb_pkt pkt;
|
|
struct mbuf *pMbuf;
|
|
|
|
xnb_get1pkt(&pkt, size, NETTXF_data_validated);
|
|
|
|
pMbuf = xnb_pkt2mbufc(&pkt, xnb_unit_pvt.ifp);
|
|
XNB_ASSERT(M_TRAILINGSPACE(pMbuf) >= size);
|
|
XNB_ASSERT(pMbuf->m_pkthdr.csum_flags & CSUM_IP_CHECKED);
|
|
XNB_ASSERT(pMbuf->m_pkthdr.csum_flags & CSUM_IP_VALID);
|
|
XNB_ASSERT(pMbuf->m_pkthdr.csum_flags & CSUM_DATA_VALID);
|
|
XNB_ASSERT(pMbuf->m_pkthdr.csum_flags & CSUM_PSEUDO_HDR);
|
|
safe_m_freem(&pMbuf);
|
|
}
|
|
|
|
/**
|
|
* xnb_pkt2mbufc on packet that can fit in one cluster
|
|
*/
|
|
static void
|
|
xnb_pkt2mbufc_1cluster(char *buffer, size_t buflen)
|
|
{
|
|
const size_t size = MINCLSIZE;
|
|
struct xnb_pkt pkt;
|
|
struct mbuf *pMbuf;
|
|
|
|
xnb_get1pkt(&pkt, size, 0);
|
|
|
|
pMbuf = xnb_pkt2mbufc(&pkt, xnb_unit_pvt.ifp);
|
|
XNB_ASSERT(M_TRAILINGSPACE(pMbuf) >= size);
|
|
safe_m_freem(&pMbuf);
|
|
}
|
|
|
|
/**
|
|
* xnb_pkt2mbufc on packet that cannot fit in one regular cluster
|
|
*/
|
|
static void
|
|
xnb_pkt2mbufc_largecluster(char *buffer, size_t buflen)
|
|
{
|
|
const size_t size = MCLBYTES + 1;
|
|
struct xnb_pkt pkt;
|
|
struct mbuf *pMbuf;
|
|
|
|
xnb_get1pkt(&pkt, size, 0);
|
|
|
|
pMbuf = xnb_pkt2mbufc(&pkt, xnb_unit_pvt.ifp);
|
|
XNB_ASSERT(M_TRAILINGSPACE(pMbuf) >= size);
|
|
safe_m_freem(&pMbuf);
|
|
}
|
|
|
|
/**
|
|
* xnb_pkt2mbufc on packet that cannot fit in one clusters
|
|
*/
|
|
static void
|
|
xnb_pkt2mbufc_2cluster(char *buffer, size_t buflen)
|
|
{
|
|
const size_t size = 2 * MCLBYTES + 1;
|
|
size_t space = 0;
|
|
struct xnb_pkt pkt;
|
|
struct mbuf *pMbuf;
|
|
struct mbuf *m;
|
|
|
|
xnb_get1pkt(&pkt, size, 0);
|
|
|
|
pMbuf = xnb_pkt2mbufc(&pkt, xnb_unit_pvt.ifp);
|
|
|
|
for (m = pMbuf; m != NULL; m = m->m_next) {
|
|
space += M_TRAILINGSPACE(m);
|
|
}
|
|
XNB_ASSERT(space >= size);
|
|
safe_m_freem(&pMbuf);
|
|
}
|
|
|
|
/**
|
|
* xnb_txpkt2gnttab on an empty packet. Should return empty gnttab
|
|
*/
|
|
static void
|
|
xnb_txpkt2gnttab_empty(char *buffer, size_t buflen)
|
|
{
|
|
int n_entries;
|
|
struct xnb_pkt pkt;
|
|
struct mbuf *pMbuf;
|
|
pkt.list_len = 0;
|
|
|
|
/* must call xnb_ring2pkt just to intialize pkt */
|
|
xnb_ring2pkt(&pkt, &xnb_unit_pvt.txb, xnb_unit_pvt.txb.req_cons);
|
|
pkt.size = 0;
|
|
pMbuf = xnb_pkt2mbufc(&pkt, xnb_unit_pvt.ifp);
|
|
n_entries = xnb_txpkt2gnttab(&pkt, pMbuf, xnb_unit_pvt.gnttab,
|
|
&xnb_unit_pvt.txb, DOMID_FIRST_RESERVED);
|
|
XNB_ASSERT(n_entries == 0);
|
|
safe_m_freem(&pMbuf);
|
|
}
|
|
|
|
/**
|
|
* xnb_txpkt2gnttab on a short packet, that can fit in one mbuf internal buffer
|
|
* and has one request
|
|
*/
|
|
static void
|
|
xnb_txpkt2gnttab_short(char *buffer, size_t buflen)
|
|
{
|
|
const size_t size = MINCLSIZE - 1;
|
|
int n_entries;
|
|
struct xnb_pkt pkt;
|
|
struct mbuf *pMbuf;
|
|
|
|
struct netif_tx_request *req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->flags = 0;
|
|
req->size = size;
|
|
req->gref = 7;
|
|
req->offset = 17;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
RING_PUSH_REQUESTS(&xnb_unit_pvt.txf);
|
|
|
|
xnb_ring2pkt(&pkt, &xnb_unit_pvt.txb, xnb_unit_pvt.txb.req_cons);
|
|
|
|
pMbuf = xnb_pkt2mbufc(&pkt, xnb_unit_pvt.ifp);
|
|
n_entries = xnb_txpkt2gnttab(&pkt, pMbuf, xnb_unit_pvt.gnttab,
|
|
&xnb_unit_pvt.txb, DOMID_FIRST_RESERVED);
|
|
XNB_ASSERT(n_entries == 1);
|
|
XNB_ASSERT(xnb_unit_pvt.gnttab[0].len == size);
|
|
/* flags should indicate gref's for source */
|
|
XNB_ASSERT(xnb_unit_pvt.gnttab[0].flags & GNTCOPY_source_gref);
|
|
XNB_ASSERT(xnb_unit_pvt.gnttab[0].source.offset == req->offset);
|
|
XNB_ASSERT(xnb_unit_pvt.gnttab[0].source.domid == DOMID_SELF);
|
|
XNB_ASSERT(xnb_unit_pvt.gnttab[0].dest.offset == virt_to_offset(
|
|
mtod(pMbuf, vm_offset_t)));
|
|
XNB_ASSERT(xnb_unit_pvt.gnttab[0].dest.u.gmfn ==
|
|
virt_to_mfn(mtod(pMbuf, vm_offset_t)));
|
|
XNB_ASSERT(xnb_unit_pvt.gnttab[0].dest.domid == DOMID_FIRST_RESERVED);
|
|
safe_m_freem(&pMbuf);
|
|
}
|
|
|
|
/**
|
|
* xnb_txpkt2gnttab on a packet with two requests, that can fit into a single
|
|
* mbuf cluster
|
|
*/
|
|
static void
|
|
xnb_txpkt2gnttab_2req(char *buffer, size_t buflen)
|
|
{
|
|
int n_entries;
|
|
struct xnb_pkt pkt;
|
|
struct mbuf *pMbuf;
|
|
|
|
struct netif_tx_request *req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->flags = NETTXF_more_data;
|
|
req->size = 1900;
|
|
req->gref = 7;
|
|
req->offset = 0;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->flags = 0;
|
|
req->size = 500;
|
|
req->gref = 8;
|
|
req->offset = 0;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
RING_PUSH_REQUESTS(&xnb_unit_pvt.txf);
|
|
|
|
xnb_ring2pkt(&pkt, &xnb_unit_pvt.txb, xnb_unit_pvt.txb.req_cons);
|
|
|
|
pMbuf = xnb_pkt2mbufc(&pkt, xnb_unit_pvt.ifp);
|
|
n_entries = xnb_txpkt2gnttab(&pkt, pMbuf, xnb_unit_pvt.gnttab,
|
|
&xnb_unit_pvt.txb, DOMID_FIRST_RESERVED);
|
|
|
|
XNB_ASSERT(n_entries == 2);
|
|
XNB_ASSERT(xnb_unit_pvt.gnttab[0].len == 1400);
|
|
XNB_ASSERT(xnb_unit_pvt.gnttab[0].dest.offset == virt_to_offset(
|
|
mtod(pMbuf, vm_offset_t)));
|
|
|
|
XNB_ASSERT(xnb_unit_pvt.gnttab[1].len == 500);
|
|
XNB_ASSERT(xnb_unit_pvt.gnttab[1].dest.offset == virt_to_offset(
|
|
mtod(pMbuf, vm_offset_t) + 1400));
|
|
safe_m_freem(&pMbuf);
|
|
}
|
|
|
|
/**
|
|
* xnb_txpkt2gnttab on a single request that spans two mbuf clusters
|
|
*/
|
|
static void
|
|
xnb_txpkt2gnttab_2cluster(char *buffer, size_t buflen)
|
|
{
|
|
int n_entries;
|
|
struct xnb_pkt pkt;
|
|
struct mbuf *pMbuf;
|
|
const uint16_t data_this_transaction = (MCLBYTES*2) + 1;
|
|
|
|
struct netif_tx_request *req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->flags = 0;
|
|
req->size = data_this_transaction;
|
|
req->gref = 8;
|
|
req->offset = 0;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
RING_PUSH_REQUESTS(&xnb_unit_pvt.txf);
|
|
xnb_ring2pkt(&pkt, &xnb_unit_pvt.txb, xnb_unit_pvt.txb.req_cons);
|
|
|
|
pMbuf = xnb_pkt2mbufc(&pkt, xnb_unit_pvt.ifp);
|
|
n_entries = xnb_txpkt2gnttab(&pkt, pMbuf, xnb_unit_pvt.gnttab,
|
|
&xnb_unit_pvt.txb, DOMID_FIRST_RESERVED);
|
|
|
|
if (M_TRAILINGSPACE(pMbuf) == MCLBYTES) {
|
|
/* there should be three mbufs and three gnttab entries */
|
|
XNB_ASSERT(n_entries == 3);
|
|
XNB_ASSERT(xnb_unit_pvt.gnttab[0].len == MCLBYTES);
|
|
XNB_ASSERT(
|
|
xnb_unit_pvt.gnttab[0].dest.offset == virt_to_offset(
|
|
mtod(pMbuf, vm_offset_t)));
|
|
XNB_ASSERT(xnb_unit_pvt.gnttab[0].source.offset == 0);
|
|
|
|
XNB_ASSERT(xnb_unit_pvt.gnttab[1].len == MCLBYTES);
|
|
XNB_ASSERT(
|
|
xnb_unit_pvt.gnttab[1].dest.offset == virt_to_offset(
|
|
mtod(pMbuf->m_next, vm_offset_t)));
|
|
XNB_ASSERT(xnb_unit_pvt.gnttab[1].source.offset == MCLBYTES);
|
|
|
|
XNB_ASSERT(xnb_unit_pvt.gnttab[2].len == 1);
|
|
XNB_ASSERT(
|
|
xnb_unit_pvt.gnttab[2].dest.offset == virt_to_offset(
|
|
mtod(pMbuf->m_next, vm_offset_t)));
|
|
XNB_ASSERT(xnb_unit_pvt.gnttab[2].source.offset == 2 *
|
|
MCLBYTES);
|
|
} else if (M_TRAILINGSPACE(pMbuf) == 2 * MCLBYTES) {
|
|
/* there should be two mbufs and two gnttab entries */
|
|
XNB_ASSERT(n_entries == 2);
|
|
XNB_ASSERT(xnb_unit_pvt.gnttab[0].len == 2 * MCLBYTES);
|
|
XNB_ASSERT(
|
|
xnb_unit_pvt.gnttab[0].dest.offset == virt_to_offset(
|
|
mtod(pMbuf, vm_offset_t)));
|
|
XNB_ASSERT(xnb_unit_pvt.gnttab[0].source.offset == 0);
|
|
|
|
XNB_ASSERT(xnb_unit_pvt.gnttab[1].len == 1);
|
|
XNB_ASSERT(
|
|
xnb_unit_pvt.gnttab[1].dest.offset == virt_to_offset(
|
|
mtod(pMbuf->m_next, vm_offset_t)));
|
|
XNB_ASSERT(
|
|
xnb_unit_pvt.gnttab[1].source.offset == 2 * MCLBYTES);
|
|
|
|
} else {
|
|
/* should never get here */
|
|
XNB_ASSERT(0);
|
|
}
|
|
if (pMbuf != NULL)
|
|
m_freem(pMbuf);
|
|
}
|
|
|
|
|
|
/**
|
|
* xnb_update_mbufc on a short packet that only has one gnttab entry
|
|
*/
|
|
static void
|
|
xnb_update_mbufc_short(char *buffer, size_t buflen)
|
|
{
|
|
const size_t size = MINCLSIZE - 1;
|
|
int n_entries;
|
|
struct xnb_pkt pkt;
|
|
struct mbuf *pMbuf;
|
|
|
|
struct netif_tx_request *req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->flags = 0;
|
|
req->size = size;
|
|
req->gref = 7;
|
|
req->offset = 17;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
RING_PUSH_REQUESTS(&xnb_unit_pvt.txf);
|
|
|
|
xnb_ring2pkt(&pkt, &xnb_unit_pvt.txb, xnb_unit_pvt.txb.req_cons);
|
|
|
|
pMbuf = xnb_pkt2mbufc(&pkt, xnb_unit_pvt.ifp);
|
|
n_entries = xnb_txpkt2gnttab(&pkt, pMbuf, xnb_unit_pvt.gnttab,
|
|
&xnb_unit_pvt.txb, DOMID_FIRST_RESERVED);
|
|
|
|
/* Update grant table's status fields as the hypervisor call would */
|
|
xnb_unit_pvt.gnttab[0].status = GNTST_okay;
|
|
|
|
xnb_update_mbufc(pMbuf, xnb_unit_pvt.gnttab, n_entries);
|
|
XNB_ASSERT(pMbuf->m_len == size);
|
|
XNB_ASSERT(pMbuf->m_pkthdr.len == size);
|
|
safe_m_freem(&pMbuf);
|
|
}
|
|
|
|
/**
|
|
* xnb_update_mbufc on a packet with two requests, that can fit into a single
|
|
* mbuf cluster
|
|
*/
|
|
static void
|
|
xnb_update_mbufc_2req(char *buffer, size_t buflen)
|
|
{
|
|
int n_entries;
|
|
struct xnb_pkt pkt;
|
|
struct mbuf *pMbuf;
|
|
|
|
struct netif_tx_request *req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->flags = NETTXF_more_data;
|
|
req->size = 1900;
|
|
req->gref = 7;
|
|
req->offset = 0;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->flags = 0;
|
|
req->size = 500;
|
|
req->gref = 8;
|
|
req->offset = 0;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
RING_PUSH_REQUESTS(&xnb_unit_pvt.txf);
|
|
|
|
xnb_ring2pkt(&pkt, &xnb_unit_pvt.txb, xnb_unit_pvt.txb.req_cons);
|
|
|
|
pMbuf = xnb_pkt2mbufc(&pkt, xnb_unit_pvt.ifp);
|
|
n_entries = xnb_txpkt2gnttab(&pkt, pMbuf, xnb_unit_pvt.gnttab,
|
|
&xnb_unit_pvt.txb, DOMID_FIRST_RESERVED);
|
|
|
|
/* Update grant table's status fields as the hypervisor call would */
|
|
xnb_unit_pvt.gnttab[0].status = GNTST_okay;
|
|
xnb_unit_pvt.gnttab[1].status = GNTST_okay;
|
|
|
|
xnb_update_mbufc(pMbuf, xnb_unit_pvt.gnttab, n_entries);
|
|
XNB_ASSERT(n_entries == 2);
|
|
XNB_ASSERT(pMbuf->m_pkthdr.len == 1900);
|
|
XNB_ASSERT(pMbuf->m_len == 1900);
|
|
|
|
safe_m_freem(&pMbuf);
|
|
}
|
|
|
|
/**
|
|
* xnb_update_mbufc on a single request that spans two mbuf clusters
|
|
*/
|
|
static void
|
|
xnb_update_mbufc_2cluster(char *buffer, size_t buflen)
|
|
{
|
|
int i;
|
|
int n_entries;
|
|
struct xnb_pkt pkt;
|
|
struct mbuf *pMbuf;
|
|
const uint16_t data_this_transaction = (MCLBYTES*2) + 1;
|
|
|
|
struct netif_tx_request *req = RING_GET_REQUEST(&xnb_unit_pvt.txf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->flags = 0;
|
|
req->size = data_this_transaction;
|
|
req->gref = 8;
|
|
req->offset = 0;
|
|
xnb_unit_pvt.txf.req_prod_pvt++;
|
|
|
|
RING_PUSH_REQUESTS(&xnb_unit_pvt.txf);
|
|
xnb_ring2pkt(&pkt, &xnb_unit_pvt.txb, xnb_unit_pvt.txb.req_cons);
|
|
|
|
pMbuf = xnb_pkt2mbufc(&pkt, xnb_unit_pvt.ifp);
|
|
n_entries = xnb_txpkt2gnttab(&pkt, pMbuf, xnb_unit_pvt.gnttab,
|
|
&xnb_unit_pvt.txb, DOMID_FIRST_RESERVED);
|
|
|
|
/* Update grant table's status fields */
|
|
for (i = 0; i < n_entries; i++) {
|
|
xnb_unit_pvt.gnttab[0].status = GNTST_okay;
|
|
}
|
|
xnb_update_mbufc(pMbuf, xnb_unit_pvt.gnttab, n_entries);
|
|
|
|
if (n_entries == 3) {
|
|
/* there should be three mbufs and three gnttab entries */
|
|
XNB_ASSERT(pMbuf->m_pkthdr.len == data_this_transaction);
|
|
XNB_ASSERT(pMbuf->m_len == MCLBYTES);
|
|
XNB_ASSERT(pMbuf->m_next->m_len == MCLBYTES);
|
|
XNB_ASSERT(pMbuf->m_next->m_next->m_len == 1);
|
|
} else if (n_entries == 2) {
|
|
/* there should be two mbufs and two gnttab entries */
|
|
XNB_ASSERT(n_entries == 2);
|
|
XNB_ASSERT(pMbuf->m_pkthdr.len == data_this_transaction);
|
|
XNB_ASSERT(pMbuf->m_len == 2 * MCLBYTES);
|
|
XNB_ASSERT(pMbuf->m_next->m_len == 1);
|
|
} else {
|
|
/* should never get here */
|
|
XNB_ASSERT(0);
|
|
}
|
|
safe_m_freem(&pMbuf);
|
|
}
|
|
|
|
/** xnb_mbufc2pkt on an empty mbufc */
|
|
static void
|
|
xnb_mbufc2pkt_empty(char *buffer, size_t buflen) {
|
|
struct xnb_pkt pkt;
|
|
int free_slots = 64;
|
|
struct mbuf *mbuf;
|
|
|
|
mbuf = m_get(M_WAITOK, MT_DATA);
|
|
/*
|
|
* note: it is illegal to set M_PKTHDR on a mbuf with no data. Doing so
|
|
* will cause m_freem to segfault
|
|
*/
|
|
XNB_ASSERT(mbuf->m_len == 0);
|
|
|
|
xnb_mbufc2pkt(mbuf, &pkt, 0, free_slots);
|
|
XNB_ASSERT(! xnb_pkt_is_valid(&pkt));
|
|
|
|
safe_m_freem(&mbuf);
|
|
}
|
|
|
|
/** xnb_mbufc2pkt on a short mbufc */
|
|
static void
|
|
xnb_mbufc2pkt_short(char *buffer, size_t buflen) {
|
|
struct xnb_pkt pkt;
|
|
size_t size = 128;
|
|
int free_slots = 64;
|
|
RING_IDX start = 9;
|
|
struct mbuf *mbuf;
|
|
|
|
mbuf = m_getm(NULL, size, M_WAITOK, MT_DATA);
|
|
mbuf->m_flags |= M_PKTHDR;
|
|
mbuf->m_pkthdr.len = size;
|
|
mbuf->m_len = size;
|
|
|
|
xnb_mbufc2pkt(mbuf, &pkt, start, free_slots);
|
|
XNB_ASSERT(xnb_pkt_is_valid(&pkt));
|
|
XNB_ASSERT(pkt.size == size);
|
|
XNB_ASSERT(pkt.car_size == size);
|
|
XNB_ASSERT(! (pkt.flags &
|
|
(NETRXF_more_data | NETRXF_extra_info)));
|
|
XNB_ASSERT(pkt.list_len == 1);
|
|
XNB_ASSERT(pkt.car == start);
|
|
|
|
safe_m_freem(&mbuf);
|
|
}
|
|
|
|
/** xnb_mbufc2pkt on a single mbuf with an mbuf cluster */
|
|
static void
|
|
xnb_mbufc2pkt_1cluster(char *buffer, size_t buflen) {
|
|
struct xnb_pkt pkt;
|
|
size_t size = MCLBYTES;
|
|
int free_slots = 32;
|
|
RING_IDX start = 12;
|
|
struct mbuf *mbuf;
|
|
|
|
mbuf = m_getm(NULL, size, M_WAITOK, MT_DATA);
|
|
mbuf->m_flags |= M_PKTHDR;
|
|
mbuf->m_pkthdr.len = size;
|
|
mbuf->m_len = size;
|
|
|
|
xnb_mbufc2pkt(mbuf, &pkt, start, free_slots);
|
|
XNB_ASSERT(xnb_pkt_is_valid(&pkt));
|
|
XNB_ASSERT(pkt.size == size);
|
|
XNB_ASSERT(pkt.car_size == size);
|
|
XNB_ASSERT(! (pkt.flags &
|
|
(NETRXF_more_data | NETRXF_extra_info)));
|
|
XNB_ASSERT(pkt.list_len == 1);
|
|
XNB_ASSERT(pkt.car == start);
|
|
|
|
safe_m_freem(&mbuf);
|
|
}
|
|
|
|
/** xnb_mbufc2pkt on a two-mbuf chain with short data regions */
|
|
static void
|
|
xnb_mbufc2pkt_2short(char *buffer, size_t buflen) {
|
|
struct xnb_pkt pkt;
|
|
size_t size1 = MHLEN - 5;
|
|
size_t size2 = MHLEN - 15;
|
|
int free_slots = 32;
|
|
RING_IDX start = 14;
|
|
struct mbuf *mbufc, *mbufc2;
|
|
|
|
mbufc = m_getm(NULL, size1, M_WAITOK, MT_DATA);
|
|
mbufc->m_flags |= M_PKTHDR;
|
|
if (mbufc == NULL) {
|
|
XNB_ASSERT(mbufc != NULL);
|
|
return;
|
|
}
|
|
|
|
mbufc2 = m_getm(mbufc, size2, M_WAITOK, MT_DATA);
|
|
if (mbufc2 == NULL) {
|
|
XNB_ASSERT(mbufc2 != NULL);
|
|
safe_m_freem(&mbufc);
|
|
return;
|
|
}
|
|
mbufc2->m_pkthdr.len = size1 + size2;
|
|
mbufc2->m_len = size1;
|
|
|
|
xnb_mbufc2pkt(mbufc2, &pkt, start, free_slots);
|
|
XNB_ASSERT(xnb_pkt_is_valid(&pkt));
|
|
XNB_ASSERT(pkt.size == size1 + size2);
|
|
XNB_ASSERT(pkt.car == start);
|
|
/*
|
|
* The second m_getm may allocate a new mbuf and append
|
|
* it to the chain, or it may simply extend the first mbuf.
|
|
*/
|
|
if (mbufc2->m_next != NULL) {
|
|
XNB_ASSERT(pkt.car_size == size1);
|
|
XNB_ASSERT(pkt.list_len == 1);
|
|
XNB_ASSERT(pkt.cdr == start + 1);
|
|
}
|
|
|
|
safe_m_freem(&mbufc2);
|
|
}
|
|
|
|
/** xnb_mbufc2pkt on a mbuf chain with >1 mbuf cluster */
|
|
static void
|
|
xnb_mbufc2pkt_long(char *buffer, size_t buflen) {
|
|
struct xnb_pkt pkt;
|
|
size_t size = 14 * MCLBYTES / 3;
|
|
size_t size_remaining;
|
|
int free_slots = 15;
|
|
RING_IDX start = 3;
|
|
struct mbuf *mbufc, *m;
|
|
|
|
mbufc = m_getm(NULL, size, M_WAITOK, MT_DATA);
|
|
mbufc->m_flags |= M_PKTHDR;
|
|
if (mbufc == NULL) {
|
|
XNB_ASSERT(mbufc != NULL);
|
|
return;
|
|
}
|
|
|
|
mbufc->m_pkthdr.len = size;
|
|
size_remaining = size;
|
|
for (m = mbufc; m != NULL; m = m->m_next) {
|
|
m->m_len = MAX(M_TRAILINGSPACE(m), size_remaining);
|
|
size_remaining -= m->m_len;
|
|
}
|
|
|
|
xnb_mbufc2pkt(mbufc, &pkt, start, free_slots);
|
|
XNB_ASSERT(xnb_pkt_is_valid(&pkt));
|
|
XNB_ASSERT(pkt.size == size);
|
|
XNB_ASSERT(pkt.car == start);
|
|
XNB_ASSERT(pkt.car_size = mbufc->m_len);
|
|
/*
|
|
* There should be >1 response in the packet, and there is no
|
|
* extra info.
|
|
*/
|
|
XNB_ASSERT(! (pkt.flags & NETRXF_extra_info));
|
|
XNB_ASSERT(pkt.cdr == pkt.car + 1);
|
|
|
|
safe_m_freem(&mbufc);
|
|
}
|
|
|
|
/** xnb_mbufc2pkt on a mbuf chain with >1 mbuf cluster and extra info */
|
|
static void
|
|
xnb_mbufc2pkt_extra(char *buffer, size_t buflen) {
|
|
struct xnb_pkt pkt;
|
|
size_t size = 14 * MCLBYTES / 3;
|
|
size_t size_remaining;
|
|
int free_slots = 15;
|
|
RING_IDX start = 3;
|
|
struct mbuf *mbufc, *m;
|
|
|
|
mbufc = m_getm(NULL, size, M_WAITOK, MT_DATA);
|
|
if (mbufc == NULL) {
|
|
XNB_ASSERT(mbufc != NULL);
|
|
return;
|
|
}
|
|
|
|
mbufc->m_flags |= M_PKTHDR;
|
|
mbufc->m_pkthdr.len = size;
|
|
mbufc->m_pkthdr.csum_flags |= CSUM_TSO;
|
|
mbufc->m_pkthdr.tso_segsz = TCP_MSS - 40;
|
|
size_remaining = size;
|
|
for (m = mbufc; m != NULL; m = m->m_next) {
|
|
m->m_len = MAX(M_TRAILINGSPACE(m), size_remaining);
|
|
size_remaining -= m->m_len;
|
|
}
|
|
|
|
xnb_mbufc2pkt(mbufc, &pkt, start, free_slots);
|
|
XNB_ASSERT(xnb_pkt_is_valid(&pkt));
|
|
XNB_ASSERT(pkt.size == size);
|
|
XNB_ASSERT(pkt.car == start);
|
|
XNB_ASSERT(pkt.car_size = mbufc->m_len);
|
|
/* There should be >1 response in the packet, there is extra info */
|
|
XNB_ASSERT(pkt.flags & NETRXF_extra_info);
|
|
XNB_ASSERT(pkt.flags & NETRXF_data_validated);
|
|
XNB_ASSERT(pkt.cdr == pkt.car + 2);
|
|
XNB_ASSERT(pkt.extra.u.gso.size = mbufc->m_pkthdr.tso_segsz);
|
|
XNB_ASSERT(pkt.extra.type == XEN_NETIF_EXTRA_TYPE_GSO);
|
|
XNB_ASSERT(! (pkt.extra.flags & XEN_NETIF_EXTRA_FLAG_MORE));
|
|
|
|
safe_m_freem(&mbufc);
|
|
}
|
|
|
|
/** xnb_mbufc2pkt with insufficient space in the ring */
|
|
static void
|
|
xnb_mbufc2pkt_nospace(char *buffer, size_t buflen) {
|
|
struct xnb_pkt pkt;
|
|
size_t size = 14 * MCLBYTES / 3;
|
|
size_t size_remaining;
|
|
int free_slots = 2;
|
|
RING_IDX start = 3;
|
|
struct mbuf *mbufc, *m;
|
|
int error;
|
|
|
|
mbufc = m_getm(NULL, size, M_WAITOK, MT_DATA);
|
|
mbufc->m_flags |= M_PKTHDR;
|
|
if (mbufc == NULL) {
|
|
XNB_ASSERT(mbufc != NULL);
|
|
return;
|
|
}
|
|
|
|
mbufc->m_pkthdr.len = size;
|
|
size_remaining = size;
|
|
for (m = mbufc; m != NULL; m = m->m_next) {
|
|
m->m_len = MAX(M_TRAILINGSPACE(m), size_remaining);
|
|
size_remaining -= m->m_len;
|
|
}
|
|
|
|
error = xnb_mbufc2pkt(mbufc, &pkt, start, free_slots);
|
|
XNB_ASSERT(error == EAGAIN);
|
|
XNB_ASSERT(! xnb_pkt_is_valid(&pkt));
|
|
|
|
safe_m_freem(&mbufc);
|
|
}
|
|
|
|
/**
|
|
* xnb_rxpkt2gnttab on an empty packet. Should return empty gnttab
|
|
*/
|
|
static void
|
|
xnb_rxpkt2gnttab_empty(char *buffer, size_t buflen)
|
|
{
|
|
struct xnb_pkt pkt;
|
|
int nr_entries;
|
|
int free_slots = 60;
|
|
struct mbuf *mbuf;
|
|
|
|
mbuf = m_get(M_WAITOK, MT_DATA);
|
|
|
|
xnb_mbufc2pkt(mbuf, &pkt, 0, free_slots);
|
|
nr_entries = xnb_rxpkt2gnttab(&pkt, mbuf, xnb_unit_pvt.gnttab,
|
|
&xnb_unit_pvt.rxb, DOMID_FIRST_RESERVED);
|
|
|
|
XNB_ASSERT(nr_entries == 0);
|
|
|
|
safe_m_freem(&mbuf);
|
|
}
|
|
|
|
/** xnb_rxpkt2gnttab on a short packet without extra data */
|
|
static void
|
|
xnb_rxpkt2gnttab_short(char *buffer, size_t buflen) {
|
|
struct xnb_pkt pkt;
|
|
int nr_entries;
|
|
size_t size = 128;
|
|
int free_slots = 60;
|
|
RING_IDX start = 9;
|
|
struct netif_rx_request *req;
|
|
struct mbuf *mbuf;
|
|
|
|
mbuf = m_getm(NULL, size, M_WAITOK, MT_DATA);
|
|
mbuf->m_flags |= M_PKTHDR;
|
|
mbuf->m_pkthdr.len = size;
|
|
mbuf->m_len = size;
|
|
|
|
xnb_mbufc2pkt(mbuf, &pkt, start, free_slots);
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.rxf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->gref = 7;
|
|
|
|
nr_entries = xnb_rxpkt2gnttab(&pkt, mbuf, xnb_unit_pvt.gnttab,
|
|
&xnb_unit_pvt.rxb, DOMID_FIRST_RESERVED);
|
|
|
|
XNB_ASSERT(nr_entries == 1);
|
|
XNB_ASSERT(xnb_unit_pvt.gnttab[0].len == size);
|
|
/* flags should indicate gref's for dest */
|
|
XNB_ASSERT(xnb_unit_pvt.gnttab[0].flags & GNTCOPY_dest_gref);
|
|
XNB_ASSERT(xnb_unit_pvt.gnttab[0].dest.offset == 0);
|
|
XNB_ASSERT(xnb_unit_pvt.gnttab[0].source.domid == DOMID_SELF);
|
|
XNB_ASSERT(xnb_unit_pvt.gnttab[0].source.offset == virt_to_offset(
|
|
mtod(mbuf, vm_offset_t)));
|
|
XNB_ASSERT(xnb_unit_pvt.gnttab[0].source.u.gmfn ==
|
|
virt_to_mfn(mtod(mbuf, vm_offset_t)));
|
|
XNB_ASSERT(xnb_unit_pvt.gnttab[0].dest.domid == DOMID_FIRST_RESERVED);
|
|
|
|
safe_m_freem(&mbuf);
|
|
}
|
|
|
|
/**
|
|
* xnb_rxpkt2gnttab on a packet with two different mbufs in a single chai
|
|
*/
|
|
static void
|
|
xnb_rxpkt2gnttab_2req(char *buffer, size_t buflen)
|
|
{
|
|
struct xnb_pkt pkt;
|
|
int nr_entries;
|
|
int i, num_mbufs;
|
|
size_t total_granted_size = 0;
|
|
size_t size = MJUMPAGESIZE + 1;
|
|
int free_slots = 60;
|
|
RING_IDX start = 11;
|
|
struct netif_rx_request *req;
|
|
struct mbuf *mbuf, *m;
|
|
|
|
mbuf = m_getm(NULL, size, M_WAITOK, MT_DATA);
|
|
mbuf->m_flags |= M_PKTHDR;
|
|
mbuf->m_pkthdr.len = size;
|
|
mbuf->m_len = size;
|
|
|
|
xnb_mbufc2pkt(mbuf, &pkt, start, free_slots);
|
|
|
|
for (i = 0, m=mbuf; m != NULL; i++, m = m->m_next) {
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.rxf,
|
|
xnb_unit_pvt.txf.req_prod_pvt);
|
|
req->gref = i;
|
|
req->id = 5;
|
|
}
|
|
num_mbufs = i;
|
|
|
|
nr_entries = xnb_rxpkt2gnttab(&pkt, mbuf, xnb_unit_pvt.gnttab,
|
|
&xnb_unit_pvt.rxb, DOMID_FIRST_RESERVED);
|
|
|
|
XNB_ASSERT(nr_entries >= num_mbufs);
|
|
for (i = 0; i < nr_entries; i++) {
|
|
int end_offset = xnb_unit_pvt.gnttab[i].len +
|
|
xnb_unit_pvt.gnttab[i].dest.offset;
|
|
XNB_ASSERT(end_offset <= PAGE_SIZE);
|
|
total_granted_size += xnb_unit_pvt.gnttab[i].len;
|
|
}
|
|
XNB_ASSERT(total_granted_size == size);
|
|
}
|
|
|
|
/**
|
|
* xnb_rxpkt2rsp on an empty packet. Shouldn't make any response
|
|
*/
|
|
static void
|
|
xnb_rxpkt2rsp_empty(char *buffer, size_t buflen)
|
|
{
|
|
struct xnb_pkt pkt;
|
|
int nr_entries;
|
|
int nr_reqs;
|
|
int free_slots = 60;
|
|
netif_rx_back_ring_t rxb_backup = xnb_unit_pvt.rxb;
|
|
netif_rx_sring_t rxs_backup = *xnb_unit_pvt.rxs;
|
|
struct mbuf *mbuf;
|
|
|
|
mbuf = m_get(M_WAITOK, MT_DATA);
|
|
|
|
xnb_mbufc2pkt(mbuf, &pkt, 0, free_slots);
|
|
nr_entries = xnb_rxpkt2gnttab(&pkt, mbuf, xnb_unit_pvt.gnttab,
|
|
&xnb_unit_pvt.rxb, DOMID_FIRST_RESERVED);
|
|
|
|
nr_reqs = xnb_rxpkt2rsp(&pkt, xnb_unit_pvt.gnttab, nr_entries,
|
|
&xnb_unit_pvt.rxb);
|
|
XNB_ASSERT(nr_reqs == 0);
|
|
XNB_ASSERT(
|
|
memcmp(&rxb_backup, &xnb_unit_pvt.rxb, sizeof(rxb_backup)) == 0);
|
|
XNB_ASSERT(
|
|
memcmp(&rxs_backup, xnb_unit_pvt.rxs, sizeof(rxs_backup)) == 0);
|
|
|
|
safe_m_freem(&mbuf);
|
|
}
|
|
|
|
/**
|
|
* xnb_rxpkt2rsp on a short packet with no extras
|
|
*/
|
|
static void
|
|
xnb_rxpkt2rsp_short(char *buffer, size_t buflen)
|
|
{
|
|
struct xnb_pkt pkt;
|
|
int nr_entries, nr_reqs;
|
|
size_t size = 128;
|
|
int free_slots = 60;
|
|
RING_IDX start = 5;
|
|
struct netif_rx_request *req;
|
|
struct netif_rx_response *rsp;
|
|
struct mbuf *mbuf;
|
|
|
|
mbuf = m_getm(NULL, size, M_WAITOK, MT_DATA);
|
|
mbuf->m_flags |= M_PKTHDR;
|
|
mbuf->m_pkthdr.len = size;
|
|
mbuf->m_len = size;
|
|
|
|
xnb_mbufc2pkt(mbuf, &pkt, start, free_slots);
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.rxf, start);
|
|
req->gref = 7;
|
|
xnb_unit_pvt.rxb.req_cons = start;
|
|
xnb_unit_pvt.rxb.rsp_prod_pvt = start;
|
|
xnb_unit_pvt.rxs->req_prod = start + 1;
|
|
xnb_unit_pvt.rxs->rsp_prod = start;
|
|
|
|
nr_entries = xnb_rxpkt2gnttab(&pkt, mbuf, xnb_unit_pvt.gnttab,
|
|
&xnb_unit_pvt.rxb, DOMID_FIRST_RESERVED);
|
|
|
|
nr_reqs = xnb_rxpkt2rsp(&pkt, xnb_unit_pvt.gnttab, nr_entries,
|
|
&xnb_unit_pvt.rxb);
|
|
|
|
XNB_ASSERT(nr_reqs == 1);
|
|
XNB_ASSERT(xnb_unit_pvt.rxb.rsp_prod_pvt == start + 1);
|
|
rsp = RING_GET_RESPONSE(&xnb_unit_pvt.rxb, start);
|
|
XNB_ASSERT(rsp->id == req->id);
|
|
XNB_ASSERT(rsp->offset == 0);
|
|
XNB_ASSERT((rsp->flags & (NETRXF_more_data | NETRXF_extra_info)) == 0);
|
|
XNB_ASSERT(rsp->status == size);
|
|
|
|
safe_m_freem(&mbuf);
|
|
}
|
|
|
|
/**
|
|
* xnb_rxpkt2rsp with extra data
|
|
*/
|
|
static void
|
|
xnb_rxpkt2rsp_extra(char *buffer, size_t buflen)
|
|
{
|
|
struct xnb_pkt pkt;
|
|
int nr_entries, nr_reqs;
|
|
size_t size = 14;
|
|
int free_slots = 15;
|
|
RING_IDX start = 3;
|
|
uint16_t id = 49;
|
|
uint16_t gref = 65;
|
|
uint16_t mss = TCP_MSS - 40;
|
|
struct mbuf *mbufc;
|
|
struct netif_rx_request *req;
|
|
struct netif_rx_response *rsp;
|
|
struct netif_extra_info *ext;
|
|
|
|
mbufc = m_getm(NULL, size, M_WAITOK, MT_DATA);
|
|
if (mbufc == NULL) {
|
|
XNB_ASSERT(mbufc != NULL);
|
|
return;
|
|
}
|
|
|
|
mbufc->m_flags |= M_PKTHDR;
|
|
mbufc->m_pkthdr.len = size;
|
|
mbufc->m_pkthdr.csum_flags |= CSUM_TSO;
|
|
mbufc->m_pkthdr.tso_segsz = mss;
|
|
mbufc->m_len = size;
|
|
|
|
xnb_mbufc2pkt(mbufc, &pkt, start, free_slots);
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.rxf, start);
|
|
req->id = id;
|
|
req->gref = gref;
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.rxf, start + 1);
|
|
req->id = id + 1;
|
|
req->gref = gref + 1;
|
|
xnb_unit_pvt.rxb.req_cons = start;
|
|
xnb_unit_pvt.rxb.rsp_prod_pvt = start;
|
|
xnb_unit_pvt.rxs->req_prod = start + 2;
|
|
xnb_unit_pvt.rxs->rsp_prod = start;
|
|
|
|
nr_entries = xnb_rxpkt2gnttab(&pkt, mbufc, xnb_unit_pvt.gnttab,
|
|
&xnb_unit_pvt.rxb, DOMID_FIRST_RESERVED);
|
|
|
|
nr_reqs = xnb_rxpkt2rsp(&pkt, xnb_unit_pvt.gnttab, nr_entries,
|
|
&xnb_unit_pvt.rxb);
|
|
|
|
XNB_ASSERT(nr_reqs == 2);
|
|
XNB_ASSERT(xnb_unit_pvt.rxb.rsp_prod_pvt == start + 2);
|
|
rsp = RING_GET_RESPONSE(&xnb_unit_pvt.rxb, start);
|
|
XNB_ASSERT(rsp->id == id);
|
|
XNB_ASSERT((rsp->flags & NETRXF_more_data) == 0);
|
|
XNB_ASSERT((rsp->flags & NETRXF_extra_info));
|
|
XNB_ASSERT((rsp->flags & NETRXF_data_validated));
|
|
XNB_ASSERT((rsp->flags & NETRXF_csum_blank));
|
|
XNB_ASSERT(rsp->status == size);
|
|
|
|
ext = (struct netif_extra_info*)
|
|
RING_GET_RESPONSE(&xnb_unit_pvt.rxb, start + 1);
|
|
XNB_ASSERT(ext->type == XEN_NETIF_EXTRA_TYPE_GSO);
|
|
XNB_ASSERT(! (ext->flags & XEN_NETIF_EXTRA_FLAG_MORE));
|
|
XNB_ASSERT(ext->u.gso.size == mss);
|
|
XNB_ASSERT(ext->u.gso.type == XEN_NETIF_EXTRA_TYPE_GSO);
|
|
|
|
safe_m_freem(&mbufc);
|
|
}
|
|
|
|
/**
|
|
* xnb_rxpkt2rsp on a packet with more than a pages's worth of data. It should
|
|
* generate two response slot
|
|
*/
|
|
static void
|
|
xnb_rxpkt2rsp_2slots(char *buffer, size_t buflen)
|
|
{
|
|
struct xnb_pkt pkt;
|
|
int nr_entries, nr_reqs;
|
|
size_t size = PAGE_SIZE + 100;
|
|
int free_slots = 3;
|
|
uint16_t id1 = 17;
|
|
uint16_t id2 = 37;
|
|
uint16_t gref1 = 24;
|
|
uint16_t gref2 = 34;
|
|
RING_IDX start = 15;
|
|
struct netif_rx_request *req;
|
|
struct netif_rx_response *rsp;
|
|
struct mbuf *mbuf;
|
|
|
|
mbuf = m_getm(NULL, size, M_WAITOK, MT_DATA);
|
|
mbuf->m_flags |= M_PKTHDR;
|
|
mbuf->m_pkthdr.len = size;
|
|
if (mbuf->m_next != NULL) {
|
|
size_t first_len = MIN(M_TRAILINGSPACE(mbuf), size);
|
|
mbuf->m_len = first_len;
|
|
mbuf->m_next->m_len = size - first_len;
|
|
|
|
} else {
|
|
mbuf->m_len = size;
|
|
}
|
|
|
|
xnb_mbufc2pkt(mbuf, &pkt, start, free_slots);
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.rxf, start);
|
|
req->gref = gref1;
|
|
req->id = id1;
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.rxf, start + 1);
|
|
req->gref = gref2;
|
|
req->id = id2;
|
|
xnb_unit_pvt.rxb.req_cons = start;
|
|
xnb_unit_pvt.rxb.rsp_prod_pvt = start;
|
|
xnb_unit_pvt.rxs->req_prod = start + 2;
|
|
xnb_unit_pvt.rxs->rsp_prod = start;
|
|
|
|
nr_entries = xnb_rxpkt2gnttab(&pkt, mbuf, xnb_unit_pvt.gnttab,
|
|
&xnb_unit_pvt.rxb, DOMID_FIRST_RESERVED);
|
|
|
|
nr_reqs = xnb_rxpkt2rsp(&pkt, xnb_unit_pvt.gnttab, nr_entries,
|
|
&xnb_unit_pvt.rxb);
|
|
|
|
XNB_ASSERT(nr_reqs == 2);
|
|
XNB_ASSERT(xnb_unit_pvt.rxb.rsp_prod_pvt == start + 2);
|
|
rsp = RING_GET_RESPONSE(&xnb_unit_pvt.rxb, start);
|
|
XNB_ASSERT(rsp->id == id1);
|
|
XNB_ASSERT(rsp->offset == 0);
|
|
XNB_ASSERT((rsp->flags & NETRXF_extra_info) == 0);
|
|
XNB_ASSERT(rsp->flags & NETRXF_more_data);
|
|
XNB_ASSERT(rsp->status == PAGE_SIZE);
|
|
|
|
rsp = RING_GET_RESPONSE(&xnb_unit_pvt.rxb, start + 1);
|
|
XNB_ASSERT(rsp->id == id2);
|
|
XNB_ASSERT(rsp->offset == 0);
|
|
XNB_ASSERT((rsp->flags & NETRXF_extra_info) == 0);
|
|
XNB_ASSERT(! (rsp->flags & NETRXF_more_data));
|
|
XNB_ASSERT(rsp->status == size - PAGE_SIZE);
|
|
|
|
safe_m_freem(&mbuf);
|
|
}
|
|
|
|
/** xnb_rxpkt2rsp on a grant table with two sub-page entries */
|
|
static void
|
|
xnb_rxpkt2rsp_2short(char *buffer, size_t buflen) {
|
|
struct xnb_pkt pkt;
|
|
int nr_reqs, nr_entries;
|
|
size_t size1 = MHLEN - 5;
|
|
size_t size2 = MHLEN - 15;
|
|
int free_slots = 32;
|
|
RING_IDX start = 14;
|
|
uint16_t id = 47;
|
|
uint16_t gref = 54;
|
|
struct netif_rx_request *req;
|
|
struct netif_rx_response *rsp;
|
|
struct mbuf *mbufc;
|
|
|
|
mbufc = m_getm(NULL, size1, M_WAITOK, MT_DATA);
|
|
mbufc->m_flags |= M_PKTHDR;
|
|
if (mbufc == NULL) {
|
|
XNB_ASSERT(mbufc != NULL);
|
|
return;
|
|
}
|
|
|
|
m_getm(mbufc, size2, M_WAITOK, MT_DATA);
|
|
XNB_ASSERT(mbufc->m_next != NULL);
|
|
mbufc->m_pkthdr.len = size1 + size2;
|
|
mbufc->m_len = size1;
|
|
mbufc->m_next->m_len = size2;
|
|
|
|
xnb_mbufc2pkt(mbufc, &pkt, start, free_slots);
|
|
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.rxf, start);
|
|
req->gref = gref;
|
|
req->id = id;
|
|
xnb_unit_pvt.rxb.req_cons = start;
|
|
xnb_unit_pvt.rxb.rsp_prod_pvt = start;
|
|
xnb_unit_pvt.rxs->req_prod = start + 1;
|
|
xnb_unit_pvt.rxs->rsp_prod = start;
|
|
|
|
nr_entries = xnb_rxpkt2gnttab(&pkt, mbufc, xnb_unit_pvt.gnttab,
|
|
&xnb_unit_pvt.rxb, DOMID_FIRST_RESERVED);
|
|
|
|
nr_reqs = xnb_rxpkt2rsp(&pkt, xnb_unit_pvt.gnttab, nr_entries,
|
|
&xnb_unit_pvt.rxb);
|
|
|
|
XNB_ASSERT(nr_entries == 2);
|
|
XNB_ASSERT(nr_reqs == 1);
|
|
rsp = RING_GET_RESPONSE(&xnb_unit_pvt.rxb, start);
|
|
XNB_ASSERT(rsp->id == id);
|
|
XNB_ASSERT(rsp->status == size1 + size2);
|
|
XNB_ASSERT(rsp->offset == 0);
|
|
XNB_ASSERT(! (rsp->flags & (NETRXF_more_data | NETRXF_extra_info)));
|
|
|
|
safe_m_freem(&mbufc);
|
|
}
|
|
|
|
/**
|
|
* xnb_rxpkt2rsp on a long packet with a hypervisor gnttab_copy error
|
|
* Note: this test will result in an error message being printed to the console
|
|
* such as:
|
|
* xnb(xnb_rxpkt2rsp:1720): Got error -1 for hypervisor gnttab_copy status
|
|
*/
|
|
static void
|
|
xnb_rxpkt2rsp_copyerror(char *buffer, size_t buflen)
|
|
{
|
|
struct xnb_pkt pkt;
|
|
int nr_entries, nr_reqs;
|
|
int id = 7;
|
|
int gref = 42;
|
|
uint16_t canary = 6859;
|
|
size_t size = 7 * MCLBYTES;
|
|
int free_slots = 9;
|
|
RING_IDX start = 2;
|
|
struct netif_rx_request *req;
|
|
struct netif_rx_response *rsp;
|
|
struct mbuf *mbuf;
|
|
|
|
mbuf = m_getm(NULL, size, M_WAITOK, MT_DATA);
|
|
mbuf->m_flags |= M_PKTHDR;
|
|
mbuf->m_pkthdr.len = size;
|
|
mbuf->m_len = size;
|
|
|
|
xnb_mbufc2pkt(mbuf, &pkt, start, free_slots);
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.rxf, start);
|
|
req->gref = gref;
|
|
req->id = id;
|
|
xnb_unit_pvt.rxb.req_cons = start;
|
|
xnb_unit_pvt.rxb.rsp_prod_pvt = start;
|
|
xnb_unit_pvt.rxs->req_prod = start + 1;
|
|
xnb_unit_pvt.rxs->rsp_prod = start;
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.rxf, start + 1);
|
|
req->gref = canary;
|
|
req->id = canary;
|
|
|
|
nr_entries = xnb_rxpkt2gnttab(&pkt, mbuf, xnb_unit_pvt.gnttab,
|
|
&xnb_unit_pvt.rxb, DOMID_FIRST_RESERVED);
|
|
/* Inject the error*/
|
|
xnb_unit_pvt.gnttab[2].status = GNTST_general_error;
|
|
|
|
nr_reqs = xnb_rxpkt2rsp(&pkt, xnb_unit_pvt.gnttab, nr_entries,
|
|
&xnb_unit_pvt.rxb);
|
|
|
|
XNB_ASSERT(nr_reqs == 1);
|
|
XNB_ASSERT(xnb_unit_pvt.rxb.rsp_prod_pvt == start + 1);
|
|
rsp = RING_GET_RESPONSE(&xnb_unit_pvt.rxb, start);
|
|
XNB_ASSERT(rsp->id == id);
|
|
XNB_ASSERT(rsp->status == NETIF_RSP_ERROR);
|
|
req = RING_GET_REQUEST(&xnb_unit_pvt.rxf, start + 1);
|
|
XNB_ASSERT(req->gref == canary);
|
|
XNB_ASSERT(req->id == canary);
|
|
|
|
safe_m_freem(&mbuf);
|
|
}
|
|
|
|
#if defined(INET) || defined(INET6)
|
|
/**
|
|
* xnb_add_mbuf_cksum on an ARP request packet
|
|
*/
|
|
static void
|
|
xnb_add_mbuf_cksum_arp(char *buffer, size_t buflen)
|
|
{
|
|
const size_t pkt_len = sizeof(struct ether_header) +
|
|
sizeof(struct ether_arp);
|
|
struct mbuf *mbufc;
|
|
struct ether_header *eh;
|
|
struct ether_arp *ep;
|
|
unsigned char pkt_orig[pkt_len];
|
|
|
|
mbufc = m_getm(NULL, pkt_len, M_WAITOK, MT_DATA);
|
|
/* Fill in an example arp request */
|
|
eh = mtod(mbufc, struct ether_header*);
|
|
eh->ether_dhost[0] = 0xff;
|
|
eh->ether_dhost[1] = 0xff;
|
|
eh->ether_dhost[2] = 0xff;
|
|
eh->ether_dhost[3] = 0xff;
|
|
eh->ether_dhost[4] = 0xff;
|
|
eh->ether_dhost[5] = 0xff;
|
|
eh->ether_shost[0] = 0x00;
|
|
eh->ether_shost[1] = 0x15;
|
|
eh->ether_shost[2] = 0x17;
|
|
eh->ether_shost[3] = 0xe9;
|
|
eh->ether_shost[4] = 0x30;
|
|
eh->ether_shost[5] = 0x68;
|
|
eh->ether_type = htons(ETHERTYPE_ARP);
|
|
ep = (struct ether_arp*)(eh + 1);
|
|
ep->ea_hdr.ar_hrd = htons(ARPHRD_ETHER);
|
|
ep->ea_hdr.ar_pro = htons(ETHERTYPE_IP);
|
|
ep->ea_hdr.ar_hln = 6;
|
|
ep->ea_hdr.ar_pln = 4;
|
|
ep->ea_hdr.ar_op = htons(ARPOP_REQUEST);
|
|
ep->arp_sha[0] = 0x00;
|
|
ep->arp_sha[1] = 0x15;
|
|
ep->arp_sha[2] = 0x17;
|
|
ep->arp_sha[3] = 0xe9;
|
|
ep->arp_sha[4] = 0x30;
|
|
ep->arp_sha[5] = 0x68;
|
|
ep->arp_spa[0] = 0xc0;
|
|
ep->arp_spa[1] = 0xa8;
|
|
ep->arp_spa[2] = 0x0a;
|
|
ep->arp_spa[3] = 0x04;
|
|
bzero(&(ep->arp_tha), ETHER_ADDR_LEN);
|
|
ep->arp_tpa[0] = 0xc0;
|
|
ep->arp_tpa[1] = 0xa8;
|
|
ep->arp_tpa[2] = 0x0a;
|
|
ep->arp_tpa[3] = 0x06;
|
|
|
|
/* fill in the length field */
|
|
mbufc->m_len = pkt_len;
|
|
mbufc->m_pkthdr.len = pkt_len;
|
|
/* indicate that the netfront uses hw-assisted checksums */
|
|
mbufc->m_pkthdr.csum_flags = CSUM_IP_CHECKED | CSUM_IP_VALID |
|
|
CSUM_DATA_VALID | CSUM_PSEUDO_HDR;
|
|
|
|
/* Make a backup copy of the packet */
|
|
bcopy(mtod(mbufc, const void*), pkt_orig, pkt_len);
|
|
|
|
/* Function under test */
|
|
xnb_add_mbuf_cksum(mbufc);
|
|
|
|
/* Verify that the packet's data did not change */
|
|
XNB_ASSERT(bcmp(mtod(mbufc, const void*), pkt_orig, pkt_len) == 0);
|
|
m_freem(mbufc);
|
|
}
|
|
|
|
/**
|
|
* Helper function that populates the ethernet header and IP header used by
|
|
* some of the xnb_add_mbuf_cksum unit tests. m must already be allocated
|
|
* and must be large enough
|
|
*/
|
|
static void
|
|
xnb_fill_eh_and_ip(struct mbuf *m, uint16_t ip_len, uint16_t ip_id,
|
|
uint16_t ip_p, uint16_t ip_off, uint16_t ip_sum)
|
|
{
|
|
struct ether_header *eh;
|
|
struct ip *iph;
|
|
|
|
eh = mtod(m, struct ether_header*);
|
|
eh->ether_dhost[0] = 0x00;
|
|
eh->ether_dhost[1] = 0x16;
|
|
eh->ether_dhost[2] = 0x3e;
|
|
eh->ether_dhost[3] = 0x23;
|
|
eh->ether_dhost[4] = 0x50;
|
|
eh->ether_dhost[5] = 0x0b;
|
|
eh->ether_shost[0] = 0x00;
|
|
eh->ether_shost[1] = 0x16;
|
|
eh->ether_shost[2] = 0x30;
|
|
eh->ether_shost[3] = 0x00;
|
|
eh->ether_shost[4] = 0x00;
|
|
eh->ether_shost[5] = 0x00;
|
|
eh->ether_type = htons(ETHERTYPE_IP);
|
|
iph = (struct ip*)(eh + 1);
|
|
iph->ip_hl = 0x5; /* 5 dwords == 20 bytes */
|
|
iph->ip_v = 4; /* IP v4 */
|
|
iph->ip_tos = 0;
|
|
iph->ip_len = htons(ip_len);
|
|
iph->ip_id = htons(ip_id);
|
|
iph->ip_off = htons(ip_off);
|
|
iph->ip_ttl = 64;
|
|
iph->ip_p = ip_p;
|
|
iph->ip_sum = htons(ip_sum);
|
|
iph->ip_src.s_addr = htonl(0xc0a80a04);
|
|
iph->ip_dst.s_addr = htonl(0xc0a80a05);
|
|
}
|
|
|
|
/**
|
|
* xnb_add_mbuf_cksum on an ICMP packet, based on a tcpdump of an actual
|
|
* ICMP packet
|
|
*/
|
|
static void
|
|
xnb_add_mbuf_cksum_icmp(char *buffer, size_t buflen)
|
|
{
|
|
const size_t icmp_len = 64; /* set by ping(1) */
|
|
const size_t pkt_len = sizeof(struct ether_header) +
|
|
sizeof(struct ip) + icmp_len;
|
|
struct mbuf *mbufc;
|
|
struct ether_header *eh;
|
|
struct ip *iph;
|
|
struct icmp *icmph;
|
|
unsigned char pkt_orig[icmp_len];
|
|
uint32_t *tv_field;
|
|
uint8_t *data_payload;
|
|
int i;
|
|
const uint16_t ICMP_CSUM = 0xaed7;
|
|
const uint16_t IP_CSUM = 0xe533;
|
|
|
|
mbufc = m_getm(NULL, pkt_len, M_WAITOK, MT_DATA);
|
|
/* Fill in an example ICMP ping request */
|
|
eh = mtod(mbufc, struct ether_header*);
|
|
xnb_fill_eh_and_ip(mbufc, 84, 28, IPPROTO_ICMP, 0, 0);
|
|
iph = (struct ip*)(eh + 1);
|
|
icmph = (struct icmp*)(iph + 1);
|
|
icmph->icmp_type = ICMP_ECHO;
|
|
icmph->icmp_code = 0;
|
|
icmph->icmp_cksum = htons(ICMP_CSUM);
|
|
icmph->icmp_id = htons(31492);
|
|
icmph->icmp_seq = htons(0);
|
|
/*
|
|
* ping(1) uses bcopy to insert a native-endian timeval after icmp_seq.
|
|
* For this test, we will set the bytes individually for portability.
|
|
*/
|
|
tv_field = (uint32_t*)(&(icmph->icmp_hun));
|
|
tv_field[0] = 0x4f02cfac;
|
|
tv_field[1] = 0x0007c46a;
|
|
/*
|
|
* Remainder of packet is an incrmenting 8 bit integer, starting with 8
|
|
*/
|
|
data_payload = (uint8_t*)(&tv_field[2]);
|
|
for (i = 8; i < 37; i++) {
|
|
*data_payload++ = i;
|
|
}
|
|
|
|
/* fill in the length field */
|
|
mbufc->m_len = pkt_len;
|
|
mbufc->m_pkthdr.len = pkt_len;
|
|
/* indicate that the netfront uses hw-assisted checksums */
|
|
mbufc->m_pkthdr.csum_flags = CSUM_IP_CHECKED | CSUM_IP_VALID |
|
|
CSUM_DATA_VALID | CSUM_PSEUDO_HDR;
|
|
|
|
bcopy(mtod(mbufc, const void*), pkt_orig, icmp_len);
|
|
/* Function under test */
|
|
xnb_add_mbuf_cksum(mbufc);
|
|
|
|
/* Check the IP checksum */
|
|
XNB_ASSERT(iph->ip_sum == htons(IP_CSUM));
|
|
|
|
/* Check that the ICMP packet did not change */
|
|
XNB_ASSERT(bcmp(icmph, pkt_orig, icmp_len));
|
|
m_freem(mbufc);
|
|
}
|
|
|
|
/**
|
|
* xnb_add_mbuf_cksum on a UDP packet, based on a tcpdump of an actual
|
|
* UDP packet
|
|
*/
|
|
static void
|
|
xnb_add_mbuf_cksum_udp(char *buffer, size_t buflen)
|
|
{
|
|
const size_t udp_len = 16;
|
|
const size_t pkt_len = sizeof(struct ether_header) +
|
|
sizeof(struct ip) + udp_len;
|
|
struct mbuf *mbufc;
|
|
struct ether_header *eh;
|
|
struct ip *iph;
|
|
struct udphdr *udp;
|
|
uint8_t *data_payload;
|
|
const uint16_t IP_CSUM = 0xe56b;
|
|
const uint16_t UDP_CSUM = 0xdde2;
|
|
|
|
mbufc = m_getm(NULL, pkt_len, M_WAITOK, MT_DATA);
|
|
/* Fill in an example UDP packet made by 'uname | nc -u <host> 2222 */
|
|
eh = mtod(mbufc, struct ether_header*);
|
|
xnb_fill_eh_and_ip(mbufc, 36, 4, IPPROTO_UDP, 0, 0xbaad);
|
|
iph = (struct ip*)(eh + 1);
|
|
udp = (struct udphdr*)(iph + 1);
|
|
udp->uh_sport = htons(0x51ae);
|
|
udp->uh_dport = htons(0x08ae);
|
|
udp->uh_ulen = htons(udp_len);
|
|
udp->uh_sum = htons(0xbaad); /* xnb_add_mbuf_cksum will fill this in */
|
|
data_payload = (uint8_t*)(udp + 1);
|
|
data_payload[0] = 'F';
|
|
data_payload[1] = 'r';
|
|
data_payload[2] = 'e';
|
|
data_payload[3] = 'e';
|
|
data_payload[4] = 'B';
|
|
data_payload[5] = 'S';
|
|
data_payload[6] = 'D';
|
|
data_payload[7] = '\n';
|
|
|
|
/* fill in the length field */
|
|
mbufc->m_len = pkt_len;
|
|
mbufc->m_pkthdr.len = pkt_len;
|
|
/* indicate that the netfront uses hw-assisted checksums */
|
|
mbufc->m_pkthdr.csum_flags = CSUM_IP_CHECKED | CSUM_IP_VALID |
|
|
CSUM_DATA_VALID | CSUM_PSEUDO_HDR;
|
|
|
|
/* Function under test */
|
|
xnb_add_mbuf_cksum(mbufc);
|
|
|
|
/* Check the checksums */
|
|
XNB_ASSERT(iph->ip_sum == htons(IP_CSUM));
|
|
XNB_ASSERT(udp->uh_sum == htons(UDP_CSUM));
|
|
|
|
m_freem(mbufc);
|
|
}
|
|
|
|
/**
|
|
* Helper function that populates a TCP packet used by all of the
|
|
* xnb_add_mbuf_cksum tcp unit tests. m must already be allocated and must be
|
|
* large enough
|
|
*/
|
|
static void
|
|
xnb_fill_tcp(struct mbuf *m)
|
|
{
|
|
struct ether_header *eh;
|
|
struct ip *iph;
|
|
struct tcphdr *tcp;
|
|
uint32_t *options;
|
|
uint8_t *data_payload;
|
|
|
|
/* Fill in an example TCP packet made by 'uname | nc <host> 2222' */
|
|
eh = mtod(m, struct ether_header*);
|
|
xnb_fill_eh_and_ip(m, 60, 8, IPPROTO_TCP, IP_DF, 0);
|
|
iph = (struct ip*)(eh + 1);
|
|
tcp = (struct tcphdr*)(iph + 1);
|
|
tcp->th_sport = htons(0x9cd9);
|
|
tcp->th_dport = htons(2222);
|
|
tcp->th_seq = htonl(0x00f72b10);
|
|
tcp->th_ack = htonl(0x7f37ba6c);
|
|
tcp->th_x2 = 0;
|
|
tcp->th_off = 8;
|
|
tcp->th_flags = 0x18;
|
|
tcp->th_win = htons(0x410);
|
|
/* th_sum is incorrect; will be inserted by function under test */
|
|
tcp->th_sum = htons(0xbaad);
|
|
tcp->th_urp = htons(0);
|
|
/*
|
|
* The following 12 bytes of options encode:
|
|
* [nop, nop, TS val 33247 ecr 3457687679]
|
|
*/
|
|
options = (uint32_t*)(tcp + 1);
|
|
options[0] = htonl(0x0101080a);
|
|
options[1] = htonl(0x000081df);
|
|
options[2] = htonl(0xce18207f);
|
|
data_payload = (uint8_t*)(&options[3]);
|
|
data_payload[0] = 'F';
|
|
data_payload[1] = 'r';
|
|
data_payload[2] = 'e';
|
|
data_payload[3] = 'e';
|
|
data_payload[4] = 'B';
|
|
data_payload[5] = 'S';
|
|
data_payload[6] = 'D';
|
|
data_payload[7] = '\n';
|
|
}
|
|
|
|
/**
|
|
* xnb_add_mbuf_cksum on a TCP packet, based on a tcpdump of an actual TCP
|
|
* packet
|
|
*/
|
|
static void
|
|
xnb_add_mbuf_cksum_tcp(char *buffer, size_t buflen)
|
|
{
|
|
const size_t payload_len = 8;
|
|
const size_t tcp_options_len = 12;
|
|
const size_t pkt_len = sizeof(struct ether_header) + sizeof(struct ip) +
|
|
sizeof(struct tcphdr) + tcp_options_len + payload_len;
|
|
struct mbuf *mbufc;
|
|
struct ether_header *eh;
|
|
struct ip *iph;
|
|
struct tcphdr *tcp;
|
|
const uint16_t IP_CSUM = 0xa55a;
|
|
const uint16_t TCP_CSUM = 0x2f64;
|
|
|
|
mbufc = m_getm(NULL, pkt_len, M_WAITOK, MT_DATA);
|
|
/* Fill in an example TCP packet made by 'uname | nc <host> 2222' */
|
|
xnb_fill_tcp(mbufc);
|
|
eh = mtod(mbufc, struct ether_header*);
|
|
iph = (struct ip*)(eh + 1);
|
|
tcp = (struct tcphdr*)(iph + 1);
|
|
|
|
/* fill in the length field */
|
|
mbufc->m_len = pkt_len;
|
|
mbufc->m_pkthdr.len = pkt_len;
|
|
/* indicate that the netfront uses hw-assisted checksums */
|
|
mbufc->m_pkthdr.csum_flags = CSUM_IP_CHECKED | CSUM_IP_VALID |
|
|
CSUM_DATA_VALID | CSUM_PSEUDO_HDR;
|
|
|
|
/* Function under test */
|
|
xnb_add_mbuf_cksum(mbufc);
|
|
|
|
/* Check the checksums */
|
|
XNB_ASSERT(iph->ip_sum == htons(IP_CSUM));
|
|
XNB_ASSERT(tcp->th_sum == htons(TCP_CSUM));
|
|
|
|
m_freem(mbufc);
|
|
}
|
|
|
|
/**
|
|
* xnb_add_mbuf_cksum on a TCP packet that does not use HW assisted checksums
|
|
*/
|
|
static void
|
|
xnb_add_mbuf_cksum_tcp_swcksum(char *buffer, size_t buflen)
|
|
{
|
|
const size_t payload_len = 8;
|
|
const size_t tcp_options_len = 12;
|
|
const size_t pkt_len = sizeof(struct ether_header) + sizeof(struct ip) +
|
|
sizeof(struct tcphdr) + tcp_options_len + payload_len;
|
|
struct mbuf *mbufc;
|
|
struct ether_header *eh;
|
|
struct ip *iph;
|
|
struct tcphdr *tcp;
|
|
/* Use deliberately bad checksums, and verify that they don't get */
|
|
/* corrected by xnb_add_mbuf_cksum */
|
|
const uint16_t IP_CSUM = 0xdead;
|
|
const uint16_t TCP_CSUM = 0xbeef;
|
|
|
|
mbufc = m_getm(NULL, pkt_len, M_WAITOK, MT_DATA);
|
|
/* Fill in an example TCP packet made by 'uname | nc <host> 2222' */
|
|
xnb_fill_tcp(mbufc);
|
|
eh = mtod(mbufc, struct ether_header*);
|
|
iph = (struct ip*)(eh + 1);
|
|
iph->ip_sum = htons(IP_CSUM);
|
|
tcp = (struct tcphdr*)(iph + 1);
|
|
tcp->th_sum = htons(TCP_CSUM);
|
|
|
|
/* fill in the length field */
|
|
mbufc->m_len = pkt_len;
|
|
mbufc->m_pkthdr.len = pkt_len;
|
|
/* indicate that the netfront does not use hw-assisted checksums */
|
|
mbufc->m_pkthdr.csum_flags = 0;
|
|
|
|
/* Function under test */
|
|
xnb_add_mbuf_cksum(mbufc);
|
|
|
|
/* Check that the checksums didn't change */
|
|
XNB_ASSERT(iph->ip_sum == htons(IP_CSUM));
|
|
XNB_ASSERT(tcp->th_sum == htons(TCP_CSUM));
|
|
|
|
m_freem(mbufc);
|
|
}
|
|
#endif /* INET || INET6 */
|
|
|
|
/**
|
|
* sscanf on unsigned chars
|
|
*/
|
|
static void
|
|
xnb_sscanf_hhu(char *buffer, size_t buflen)
|
|
{
|
|
const char mystr[] = "137";
|
|
uint8_t dest[12];
|
|
int i;
|
|
|
|
for (i = 0; i < 12; i++)
|
|
dest[i] = 'X';
|
|
|
|
sscanf(mystr, "%hhu", &dest[4]);
|
|
for (i = 0; i < 12; i++)
|
|
XNB_ASSERT(dest[i] == (i == 4 ? 137 : 'X'));
|
|
}
|
|
|
|
/**
|
|
* sscanf on signed chars
|
|
*/
|
|
static void
|
|
xnb_sscanf_hhd(char *buffer, size_t buflen)
|
|
{
|
|
const char mystr[] = "-27";
|
|
int8_t dest[12];
|
|
int i;
|
|
|
|
for (i = 0; i < 12; i++)
|
|
dest[i] = 'X';
|
|
|
|
sscanf(mystr, "%hhd", &dest[4]);
|
|
for (i = 0; i < 12; i++)
|
|
XNB_ASSERT(dest[i] == (i == 4 ? -27 : 'X'));
|
|
}
|
|
|
|
/**
|
|
* sscanf on signed long longs
|
|
*/
|
|
static void
|
|
xnb_sscanf_lld(char *buffer, size_t buflen)
|
|
{
|
|
const char mystr[] = "-123456789012345"; /* about -2**47 */
|
|
long long dest[3];
|
|
int i;
|
|
|
|
for (i = 0; i < 3; i++)
|
|
dest[i] = (long long)0xdeadbeefdeadbeef;
|
|
|
|
sscanf(mystr, "%lld", &dest[1]);
|
|
for (i = 0; i < 3; i++)
|
|
XNB_ASSERT(dest[i] == (i != 1 ? (long long)0xdeadbeefdeadbeef :
|
|
-123456789012345));
|
|
}
|
|
|
|
/**
|
|
* sscanf on unsigned long longs
|
|
*/
|
|
static void
|
|
xnb_sscanf_llu(char *buffer, size_t buflen)
|
|
{
|
|
const char mystr[] = "12802747070103273189";
|
|
unsigned long long dest[3];
|
|
int i;
|
|
|
|
for (i = 0; i < 3; i++)
|
|
dest[i] = (long long)0xdeadbeefdeadbeef;
|
|
|
|
sscanf(mystr, "%llu", &dest[1]);
|
|
for (i = 0; i < 3; i++)
|
|
XNB_ASSERT(dest[i] == (i != 1 ? (long long)0xdeadbeefdeadbeef :
|
|
12802747070103273189ull));
|
|
}
|
|
|
|
/**
|
|
* sscanf on unsigned short short n's
|
|
*/
|
|
static void
|
|
xnb_sscanf_hhn(char *buffer, size_t buflen)
|
|
{
|
|
const char mystr[] =
|
|
"000102030405060708090a0b0c0d0e0f101112131415161718191a1b1c1d1e1f"
|
|
"202122232425262728292a2b2c2d2e2f303132333435363738393a3b3c3d3e3f"
|
|
"404142434445464748494a4b4c4d4e4f505152535455565758595a5b5c5d5e5f";
|
|
unsigned char dest[12];
|
|
int i;
|
|
|
|
for (i = 0; i < 12; i++)
|
|
dest[i] = (unsigned char)'X';
|
|
|
|
sscanf(mystr,
|
|
"000102030405060708090a0b0c0d0e0f101112131415161718191a1b1c1d1e1f"
|
|
"202122232425262728292a2b2c2d2e2f303132333435363738393a3b3c3d3e3f"
|
|
"404142434445464748494a4b4c4d4e4f%hhn", &dest[4]);
|
|
for (i = 0; i < 12; i++)
|
|
XNB_ASSERT(dest[i] == (i == 4 ? 160 : 'X'));
|
|
}
|