Ensure that there are enough TX mbuf ring slots available before beginning
to dequeue a packet. The tx path was trying to ensure that enough Xenbus TX ring slots existed but it didn't check to see whether the mbuf TX ring slots were also available. They get freed in xn_txeof() which occurs after transmission, rather than earlier on in the process. (The same happens under Linux too.) Due to whatever reason (CPU use, scheduling, memory constraints, whatever) the mbuf TX ring may not have enough slots free and would allocate slot 0. This is used as the freelist head pointer to represent "free" mbuf TX ring slots; setting this to an actual mbuf value rather than an id crashes the code. This commit introduces some basic code to track the TX mbuf ring use and then (hopefully!) ensures that enough slots are free in said TX mbuf ring before it enters the actual work loop. A few notes: * Similar logic needs to be introduced to check there are enough actual slots available in the xenbuf TX ring. There's some logic which is invoked earlier but it doesn't hard-check against the number of available ring slots. Its trivial to do; I'll do it in a subsequent commit. * As I've now commented in the source, it is likely possible to deadlock the driver under certain conditions where the rings aren't receiving any changes (which I should enumerate) and thus Xen doesn't send any further software interrupts. I need to make sure that the timer(s) are running right and the queues are periodically kicked. PR: 134926
This commit is contained in:
parent
6732387afd
commit
8e9cac6f75
@ -176,6 +176,7 @@ static int xennet_get_responses(struct netfront_info *np,
|
||||
*/
|
||||
struct xn_chain_data {
|
||||
struct mbuf *xn_tx_chain[NET_TX_RING_SIZE+1];
|
||||
int xn_tx_chain_cnt;
|
||||
struct mbuf *xn_rx_chain[NET_RX_RING_SIZE+1];
|
||||
};
|
||||
|
||||
@ -727,6 +728,10 @@ netif_release_tx_bufs(struct netfront_info *np)
|
||||
np->grant_tx_ref[i]);
|
||||
np->grant_tx_ref[i] = GRANT_INVALID_REF;
|
||||
add_id_to_freelist(np->tx_mbufs, i);
|
||||
np->xn_cdata.xn_tx_chain_cnt--;
|
||||
if (np->xn_cdata.xn_tx_chain_cnt < 0) {
|
||||
panic("netif_release_tx_bufs: tx_chain_cnt must be >= 0");
|
||||
}
|
||||
m_freem(m);
|
||||
}
|
||||
}
|
||||
@ -1089,6 +1094,10 @@ xn_txeof(struct netfront_info *np)
|
||||
|
||||
np->xn_cdata.xn_tx_chain[id] = NULL;
|
||||
add_id_to_freelist(np->xn_cdata.xn_tx_chain, id);
|
||||
np->xn_cdata.xn_tx_chain_cnt--;
|
||||
if (np->xn_cdata.xn_tx_chain_cnt < 0) {
|
||||
panic("netif_release_tx_bufs: tx_chain_cnt must be >= 0");
|
||||
}
|
||||
m_free(m);
|
||||
}
|
||||
np->tx.rsp_cons = prod;
|
||||
@ -1417,7 +1426,6 @@ xn_start_locked(struct ifnet *ifp)
|
||||
ifp->if_drv_flags |= IFF_DRV_OACTIVE;
|
||||
break;
|
||||
}
|
||||
|
||||
|
||||
/*
|
||||
* Defragment the mbuf if necessary.
|
||||
@ -1433,6 +1441,36 @@ xn_start_locked(struct ifnet *ifp)
|
||||
m_head = m;
|
||||
}
|
||||
|
||||
/* Determine how many fragments now exist */
|
||||
for (m = m_head, nfrags = 0; m; m = m->m_next)
|
||||
nfrags++;
|
||||
|
||||
/*
|
||||
* Don't attempt to queue this packet if there aren't enough free entries in the chain.
|
||||
* There isn't a 1:1 correspondance between the mbuf TX ring and the xenbus TX ring.
|
||||
* xn_txeof() may need to be called to free up some slots.
|
||||
*
|
||||
* It is quite possible that this can be later eliminated if it turns out that partial
|
||||
* packets can be pushed into the ringbuffer, with fragments pushed in when further slots
|
||||
* free up.
|
||||
*
|
||||
* It is also quite possible that the driver will lock up - Xen may not send another
|
||||
* interrupt to kick the tx/rx processing if the xenbus RX ring is full and xenbus TX ring
|
||||
* is empty - no further TX work can be done until space is made in the TX mbuf ring and
|
||||
* the RX side may be waiting for TX data to continue. It is quite possible some timer
|
||||
* event should be created to kick TX/RX processing along in certain conditions.
|
||||
*/
|
||||
|
||||
/* its not +1 like the allocation because we need to keep slot [0] free for the freelist head */
|
||||
if (sc->xn_cdata.xn_tx_chain_cnt + nfrags >= NET_TX_RING_SIZE) {
|
||||
printf("xn_start_locked: xn_tx_chain_cnt (%d) + nfrags %d >= NET_TX_RING_SIZE (%d); must be full!\n",
|
||||
(int) sc->xn_cdata.xn_tx_chain_cnt, (int) nfrags, (int) NET_TX_RING_SIZE);
|
||||
IF_PREPEND(&ifp->if_snd, m_head);
|
||||
ifp->if_drv_flags |= IFF_DRV_OACTIVE;
|
||||
break;
|
||||
}
|
||||
|
||||
|
||||
/*
|
||||
* Start packing the mbufs in this chain into
|
||||
* the fragment pointers. Stop when we run out
|
||||
@ -1443,6 +1481,11 @@ xn_start_locked(struct ifnet *ifp)
|
||||
for (m = m_head; m; m = m->m_next) {
|
||||
tx = RING_GET_REQUEST(&sc->tx, i);
|
||||
id = get_id_from_freelist(sc->xn_cdata.xn_tx_chain);
|
||||
if (id == 0)
|
||||
panic("xn_start_locked: was allocated the freelist head!\n");
|
||||
sc->xn_cdata.xn_tx_chain_cnt++;
|
||||
if (sc->xn_cdata.xn_tx_chain_cnt >= NET_TX_RING_SIZE+1)
|
||||
panic("xn_start_locked: tx_chain_cnt must be < NET_TX_RING_SIZE+1\n");
|
||||
sc->xn_cdata.xn_tx_chain[id] = m;
|
||||
tx->id = id;
|
||||
ref = gnttab_claim_grant_reference(&sc->gref_tx_head);
|
||||
|
Loading…
Reference in New Issue
Block a user