numam-dpdk/drivers/net/bonding/rte_eth_bond_api.c
Wei Hu (Xavier) edf6489ea4 net/bonding: fix MAC address when switching active port
Currently, based on a active-backup bond device, when the link status of
the primary port changes from up to down, one slave port changes to the
primary port, but the new primary port's MAC address cannot change to
the bond device's MAC address. And we can't continue receive packets
whose destination MAC addresses are the same as the bond devices's MAC
address.

The current bonding PMD driver call mac_address_slaves_update function
to modify the MAC address of all slaves devices: the primary port using
bond device's MAC address, and other slaves devices using the respective
MAC address. We found that one error using primary_port instead of
current_primary_port in mac_address_slaves_update function.

On the other hand, The current bonding PMD driver sets slave devices's
MAC address according to the variable named current_primary_port. The
variable named current_primary_port changes in the following scenario:
1. Add the slave devices to bond, the first slave port will be regarded
   as the current_primary_port. If changing the order of adding the
   slave devices, the value of the variable named current_primary_port
   will be different.
2. The upper application specifies primary_port via calling the
   rte_eth_bond_primary_set API function.
3. Delete the primary slave device.
4. The link status of the primary port changes from up to down.

We have tested the above 4 cases and found that there are problems that
the new primary port's MAC address didn't change to the bond device's
MAC address when running case 3 and 4. When current_primary_port
changes, the new primary port's MAC address should change at the same
time. We also need to call mac_address_slaves_update function to update
MAC addresses in case
3 and 4.

Bugzilla ID: 256
Fixes: 2efb58cbab6e ("bond: new link bonding library")
Cc: stable@dpdk.org

Signed-off-by: Chunsong Feng <fengchunsong@huawei.com>
Signed-off-by: Wei Hu (Xavier) <xavier.huwei@huawei.com>
2020-07-17 18:21:21 +02:00

1054 lines
28 KiB
C

/* SPDX-License-Identifier: BSD-3-Clause
* Copyright(c) 2010-2017 Intel Corporation
*/
#include <string.h>
#include <rte_mbuf.h>
#include <rte_malloc.h>
#include <rte_ethdev_driver.h>
#include <rte_tcp.h>
#include <rte_bus_vdev.h>
#include <rte_kvargs.h>
#include "rte_eth_bond.h"
#include "eth_bond_private.h"
#include "eth_bond_8023ad_private.h"
int
check_for_bonded_ethdev(const struct rte_eth_dev *eth_dev)
{
/* Check valid pointer */
if (eth_dev == NULL ||
eth_dev->device == NULL ||
eth_dev->device->driver == NULL ||
eth_dev->device->driver->name == NULL)
return -1;
/* return 0 if driver name matches */
return eth_dev->device->driver->name != pmd_bond_drv.driver.name;
}
int
valid_bonded_port_id(uint16_t port_id)
{
RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -1);
return check_for_bonded_ethdev(&rte_eth_devices[port_id]);
}
int
check_for_master_bonded_ethdev(const struct rte_eth_dev *eth_dev)
{
int i;
struct bond_dev_private *internals;
if (check_for_bonded_ethdev(eth_dev) != 0)
return 0;
internals = eth_dev->data->dev_private;
/* Check if any of slave devices is a bonded device */
for (i = 0; i < internals->slave_count; i++)
if (valid_bonded_port_id(internals->slaves[i].port_id) == 0)
return 1;
return 0;
}
int
valid_slave_port_id(uint16_t port_id, uint8_t mode)
{
RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -1);
/* Verify that port_id refers to a non bonded port */
if (check_for_bonded_ethdev(&rte_eth_devices[port_id]) == 0 &&
mode == BONDING_MODE_8023AD) {
RTE_BOND_LOG(ERR, "Cannot add slave to bonded device in 802.3ad"
" mode as slave is also a bonded device, only "
"physical devices can be support in this mode.");
return -1;
}
return 0;
}
void
activate_slave(struct rte_eth_dev *eth_dev, uint16_t port_id)
{
struct bond_dev_private *internals = eth_dev->data->dev_private;
uint16_t active_count = internals->active_slave_count;
if (internals->mode == BONDING_MODE_8023AD)
bond_mode_8023ad_activate_slave(eth_dev, port_id);
if (internals->mode == BONDING_MODE_TLB
|| internals->mode == BONDING_MODE_ALB) {
internals->tlb_slaves_order[active_count] = port_id;
}
RTE_ASSERT(internals->active_slave_count <
(RTE_DIM(internals->active_slaves) - 1));
internals->active_slaves[internals->active_slave_count] = port_id;
internals->active_slave_count++;
if (internals->mode == BONDING_MODE_TLB)
bond_tlb_activate_slave(internals);
if (internals->mode == BONDING_MODE_ALB)
bond_mode_alb_client_list_upd(eth_dev);
}
void
deactivate_slave(struct rte_eth_dev *eth_dev, uint16_t port_id)
{
uint16_t slave_pos;
struct bond_dev_private *internals = eth_dev->data->dev_private;
uint16_t active_count = internals->active_slave_count;
if (internals->mode == BONDING_MODE_8023AD) {
bond_mode_8023ad_stop(eth_dev);
bond_mode_8023ad_deactivate_slave(eth_dev, port_id);
} else if (internals->mode == BONDING_MODE_TLB
|| internals->mode == BONDING_MODE_ALB)
bond_tlb_disable(internals);
slave_pos = find_slave_by_id(internals->active_slaves, active_count,
port_id);
/* If slave was not at the end of the list
* shift active slaves up active array list */
if (slave_pos < active_count) {
active_count--;
memmove(internals->active_slaves + slave_pos,
internals->active_slaves + slave_pos + 1,
(active_count - slave_pos) *
sizeof(internals->active_slaves[0]));
}
RTE_ASSERT(active_count < RTE_DIM(internals->active_slaves));
internals->active_slave_count = active_count;
/* Resetting active_slave when reaches to max
* no of slaves in active list
*/
if (internals->active_slave >= active_count)
internals->active_slave = 0;
if (eth_dev->data->dev_started) {
if (internals->mode == BONDING_MODE_8023AD) {
bond_mode_8023ad_start(eth_dev);
} else if (internals->mode == BONDING_MODE_TLB) {
bond_tlb_enable(internals);
} else if (internals->mode == BONDING_MODE_ALB) {
bond_tlb_enable(internals);
bond_mode_alb_client_list_upd(eth_dev);
}
}
}
int
rte_eth_bond_create(const char *name, uint8_t mode, uint8_t socket_id)
{
struct bond_dev_private *internals;
char devargs[52];
uint16_t port_id;
int ret;
if (name == NULL) {
RTE_BOND_LOG(ERR, "Invalid name specified");
return -EINVAL;
}
ret = snprintf(devargs, sizeof(devargs),
"driver=net_bonding,mode=%d,socket_id=%d", mode, socket_id);
if (ret < 0 || ret >= (int)sizeof(devargs))
return -ENOMEM;
ret = rte_vdev_init(name, devargs);
if (ret)
return ret;
ret = rte_eth_dev_get_port_by_name(name, &port_id);
RTE_ASSERT(!ret);
/*
* To make bond_ethdev_configure() happy we need to free the
* internals->kvlist here.
*
* Also see comment in bond_ethdev_configure().
*/
internals = rte_eth_devices[port_id].data->dev_private;
rte_kvargs_free(internals->kvlist);
internals->kvlist = NULL;
return port_id;
}
int
rte_eth_bond_free(const char *name)
{
return rte_vdev_uninit(name);
}
static int
slave_vlan_filter_set(uint16_t bonded_port_id, uint16_t slave_port_id)
{
struct rte_eth_dev *bonded_eth_dev;
struct bond_dev_private *internals;
int found;
int res = 0;
uint64_t slab = 0;
uint32_t pos = 0;
uint16_t first;
bonded_eth_dev = &rte_eth_devices[bonded_port_id];
if ((bonded_eth_dev->data->dev_conf.rxmode.offloads &
DEV_RX_OFFLOAD_VLAN_FILTER) == 0)
return 0;
internals = bonded_eth_dev->data->dev_private;
found = rte_bitmap_scan(internals->vlan_filter_bmp, &pos, &slab);
first = pos;
if (!found)
return 0;
do {
uint32_t i;
uint64_t mask;
for (i = 0, mask = 1;
i < RTE_BITMAP_SLAB_BIT_SIZE;
i ++, mask <<= 1) {
if (unlikely(slab & mask)) {
uint16_t vlan_id = pos + i;
res = rte_eth_dev_vlan_filter(slave_port_id,
vlan_id, 1);
}
}
found = rte_bitmap_scan(internals->vlan_filter_bmp,
&pos, &slab);
} while (found && first != pos && res == 0);
return res;
}
static int
slave_rte_flow_prepare(uint16_t slave_id, struct bond_dev_private *internals)
{
struct rte_flow *flow;
struct rte_flow_error ferror;
uint16_t slave_port_id = internals->slaves[slave_id].port_id;
if (internals->flow_isolated_valid != 0) {
rte_eth_dev_stop(slave_port_id);
if (rte_flow_isolate(slave_port_id, internals->flow_isolated,
&ferror)) {
RTE_BOND_LOG(ERR, "rte_flow_isolate failed for slave"
" %d: %s", slave_id, ferror.message ?
ferror.message : "(no stated reason)");
return -1;
}
}
TAILQ_FOREACH(flow, &internals->flow_list, next) {
flow->flows[slave_id] = rte_flow_create(slave_port_id,
flow->rule.attr,
flow->rule.pattern,
flow->rule.actions,
&ferror);
if (flow->flows[slave_id] == NULL) {
RTE_BOND_LOG(ERR, "Cannot create flow for slave"
" %d: %s", slave_id,
ferror.message ? ferror.message :
"(no stated reason)");
/* Destroy successful bond flows from the slave */
TAILQ_FOREACH(flow, &internals->flow_list, next) {
if (flow->flows[slave_id] != NULL) {
rte_flow_destroy(slave_port_id,
flow->flows[slave_id],
&ferror);
flow->flows[slave_id] = NULL;
}
}
return -1;
}
}
return 0;
}
static void
eth_bond_slave_inherit_dev_info_rx_first(struct bond_dev_private *internals,
const struct rte_eth_dev_info *di)
{
struct rte_eth_rxconf *rxconf_i = &internals->default_rxconf;
internals->reta_size = di->reta_size;
/* Inherit Rx offload capabilities from the first slave device */
internals->rx_offload_capa = di->rx_offload_capa;
internals->rx_queue_offload_capa = di->rx_queue_offload_capa;
internals->flow_type_rss_offloads = di->flow_type_rss_offloads;
/* Inherit maximum Rx packet size from the first slave device */
internals->candidate_max_rx_pktlen = di->max_rx_pktlen;
/* Inherit default Rx queue settings from the first slave device */
memcpy(rxconf_i, &di->default_rxconf, sizeof(*rxconf_i));
/*
* Turn off descriptor prefetch and writeback by default for all
* slave devices. Applications may tweak this setting if need be.
*/
rxconf_i->rx_thresh.pthresh = 0;
rxconf_i->rx_thresh.hthresh = 0;
rxconf_i->rx_thresh.wthresh = 0;
/* Setting this to zero should effectively enable default values */
rxconf_i->rx_free_thresh = 0;
/* Disable deferred start by default for all slave devices */
rxconf_i->rx_deferred_start = 0;
}
static void
eth_bond_slave_inherit_dev_info_tx_first(struct bond_dev_private *internals,
const struct rte_eth_dev_info *di)
{
struct rte_eth_txconf *txconf_i = &internals->default_txconf;
/* Inherit Tx offload capabilities from the first slave device */
internals->tx_offload_capa = di->tx_offload_capa;
internals->tx_queue_offload_capa = di->tx_queue_offload_capa;
/* Inherit default Tx queue settings from the first slave device */
memcpy(txconf_i, &di->default_txconf, sizeof(*txconf_i));
/*
* Turn off descriptor prefetch and writeback by default for all
* slave devices. Applications may tweak this setting if need be.
*/
txconf_i->tx_thresh.pthresh = 0;
txconf_i->tx_thresh.hthresh = 0;
txconf_i->tx_thresh.wthresh = 0;
/*
* Setting these parameters to zero assumes that default
* values will be configured implicitly by slave devices.
*/
txconf_i->tx_free_thresh = 0;
txconf_i->tx_rs_thresh = 0;
/* Disable deferred start by default for all slave devices */
txconf_i->tx_deferred_start = 0;
}
static void
eth_bond_slave_inherit_dev_info_rx_next(struct bond_dev_private *internals,
const struct rte_eth_dev_info *di)
{
struct rte_eth_rxconf *rxconf_i = &internals->default_rxconf;
const struct rte_eth_rxconf *rxconf = &di->default_rxconf;
internals->rx_offload_capa &= di->rx_offload_capa;
internals->rx_queue_offload_capa &= di->rx_queue_offload_capa;
internals->flow_type_rss_offloads &= di->flow_type_rss_offloads;
/*
* If at least one slave device suggests enabling this
* setting by default, enable it for all slave devices
* since disabling it may not be necessarily supported.
*/
if (rxconf->rx_drop_en == 1)
rxconf_i->rx_drop_en = 1;
/*
* Adding a new slave device may cause some of previously inherited
* offloads to be withdrawn from the internal rx_queue_offload_capa
* value. Thus, the new internal value of default Rx queue offloads
* has to be masked by rx_queue_offload_capa to make sure that only
* commonly supported offloads are preserved from both the previous
* value and the value being inhereted from the new slave device.
*/
rxconf_i->offloads = (rxconf_i->offloads | rxconf->offloads) &
internals->rx_queue_offload_capa;
/*
* RETA size is GCD of all slaves RETA sizes, so, if all sizes will be
* the power of 2, the lower one is GCD
*/
if (internals->reta_size > di->reta_size)
internals->reta_size = di->reta_size;
if (!internals->max_rx_pktlen &&
di->max_rx_pktlen < internals->candidate_max_rx_pktlen)
internals->candidate_max_rx_pktlen = di->max_rx_pktlen;
}
static void
eth_bond_slave_inherit_dev_info_tx_next(struct bond_dev_private *internals,
const struct rte_eth_dev_info *di)
{
struct rte_eth_txconf *txconf_i = &internals->default_txconf;
const struct rte_eth_txconf *txconf = &di->default_txconf;
internals->tx_offload_capa &= di->tx_offload_capa;
internals->tx_queue_offload_capa &= di->tx_queue_offload_capa;
/*
* Adding a new slave device may cause some of previously inherited
* offloads to be withdrawn from the internal tx_queue_offload_capa
* value. Thus, the new internal value of default Tx queue offloads
* has to be masked by tx_queue_offload_capa to make sure that only
* commonly supported offloads are preserved from both the previous
* value and the value being inhereted from the new slave device.
*/
txconf_i->offloads = (txconf_i->offloads | txconf->offloads) &
internals->tx_queue_offload_capa;
}
static void
eth_bond_slave_inherit_desc_lim_first(struct rte_eth_desc_lim *bond_desc_lim,
const struct rte_eth_desc_lim *slave_desc_lim)
{
memcpy(bond_desc_lim, slave_desc_lim, sizeof(*bond_desc_lim));
}
static int
eth_bond_slave_inherit_desc_lim_next(struct rte_eth_desc_lim *bond_desc_lim,
const struct rte_eth_desc_lim *slave_desc_lim)
{
bond_desc_lim->nb_max = RTE_MIN(bond_desc_lim->nb_max,
slave_desc_lim->nb_max);
bond_desc_lim->nb_min = RTE_MAX(bond_desc_lim->nb_min,
slave_desc_lim->nb_min);
bond_desc_lim->nb_align = RTE_MAX(bond_desc_lim->nb_align,
slave_desc_lim->nb_align);
if (bond_desc_lim->nb_min > bond_desc_lim->nb_max ||
bond_desc_lim->nb_align > bond_desc_lim->nb_max) {
RTE_BOND_LOG(ERR, "Failed to inherit descriptor limits");
return -EINVAL;
}
/* Treat maximum number of segments equal to 0 as unspecified */
if (slave_desc_lim->nb_seg_max != 0 &&
(bond_desc_lim->nb_seg_max == 0 ||
slave_desc_lim->nb_seg_max < bond_desc_lim->nb_seg_max))
bond_desc_lim->nb_seg_max = slave_desc_lim->nb_seg_max;
if (slave_desc_lim->nb_mtu_seg_max != 0 &&
(bond_desc_lim->nb_mtu_seg_max == 0 ||
slave_desc_lim->nb_mtu_seg_max < bond_desc_lim->nb_mtu_seg_max))
bond_desc_lim->nb_mtu_seg_max = slave_desc_lim->nb_mtu_seg_max;
return 0;
}
static int
__eth_bond_slave_add_lock_free(uint16_t bonded_port_id, uint16_t slave_port_id)
{
struct rte_eth_dev *bonded_eth_dev, *slave_eth_dev;
struct bond_dev_private *internals;
struct rte_eth_link link_props;
struct rte_eth_dev_info dev_info;
int ret;
bonded_eth_dev = &rte_eth_devices[bonded_port_id];
internals = bonded_eth_dev->data->dev_private;
if (valid_slave_port_id(slave_port_id, internals->mode) != 0)
return -1;
slave_eth_dev = &rte_eth_devices[slave_port_id];
if (slave_eth_dev->data->dev_flags & RTE_ETH_DEV_BONDED_SLAVE) {
RTE_BOND_LOG(ERR, "Slave device is already a slave of a bonded device");
return -1;
}
ret = rte_eth_dev_info_get(slave_port_id, &dev_info);
if (ret != 0) {
RTE_BOND_LOG(ERR,
"%s: Error during getting device (port %u) info: %s\n",
__func__, slave_port_id, strerror(-ret));
return ret;
}
if (dev_info.max_rx_pktlen < internals->max_rx_pktlen) {
RTE_BOND_LOG(ERR, "Slave (port %u) max_rx_pktlen too small",
slave_port_id);
return -1;
}
slave_add(internals, slave_eth_dev);
/* We need to store slaves reta_size to be able to synchronize RETA for all
* slave devices even if its sizes are different.
*/
internals->slaves[internals->slave_count].reta_size = dev_info.reta_size;
if (internals->slave_count < 1) {
/* if MAC is not user defined then use MAC of first slave add to
* bonded device */
if (!internals->user_defined_mac) {
if (mac_address_set(bonded_eth_dev,
slave_eth_dev->data->mac_addrs)) {
RTE_BOND_LOG(ERR, "Failed to set MAC address");
return -1;
}
}
/* Make primary slave */
internals->primary_port = slave_port_id;
internals->current_primary_port = slave_port_id;
/* Inherit queues settings from first slave */
internals->nb_rx_queues = slave_eth_dev->data->nb_rx_queues;
internals->nb_tx_queues = slave_eth_dev->data->nb_tx_queues;
eth_bond_slave_inherit_dev_info_rx_first(internals, &dev_info);
eth_bond_slave_inherit_dev_info_tx_first(internals, &dev_info);
eth_bond_slave_inherit_desc_lim_first(&internals->rx_desc_lim,
&dev_info.rx_desc_lim);
eth_bond_slave_inherit_desc_lim_first(&internals->tx_desc_lim,
&dev_info.tx_desc_lim);
} else {
int ret;
eth_bond_slave_inherit_dev_info_rx_next(internals, &dev_info);
eth_bond_slave_inherit_dev_info_tx_next(internals, &dev_info);
ret = eth_bond_slave_inherit_desc_lim_next(
&internals->rx_desc_lim, &dev_info.rx_desc_lim);
if (ret != 0)
return ret;
ret = eth_bond_slave_inherit_desc_lim_next(
&internals->tx_desc_lim, &dev_info.tx_desc_lim);
if (ret != 0)
return ret;
}
bonded_eth_dev->data->dev_conf.rx_adv_conf.rss_conf.rss_hf &=
internals->flow_type_rss_offloads;
if (slave_rte_flow_prepare(internals->slave_count, internals) != 0) {
RTE_BOND_LOG(ERR, "Failed to prepare new slave flows: port=%d",
slave_port_id);
return -1;
}
/* Add additional MAC addresses to the slave */
if (slave_add_mac_addresses(bonded_eth_dev, slave_port_id) != 0) {
RTE_BOND_LOG(ERR, "Failed to add mac address(es) to slave %hu",
slave_port_id);
return -1;
}
internals->slave_count++;
if (bonded_eth_dev->data->dev_started) {
if (slave_configure(bonded_eth_dev, slave_eth_dev) != 0) {
internals->slave_count--;
RTE_BOND_LOG(ERR, "rte_bond_slaves_configure: port=%d",
slave_port_id);
return -1;
}
}
/* Update all slave devices MACs */
mac_address_slaves_update(bonded_eth_dev);
/* Register link status change callback with bonded device pointer as
* argument*/
rte_eth_dev_callback_register(slave_port_id, RTE_ETH_EVENT_INTR_LSC,
bond_ethdev_lsc_event_callback, &bonded_eth_dev->data->port_id);
/* If bonded device is started then we can add the slave to our active
* slave array */
if (bonded_eth_dev->data->dev_started) {
ret = rte_eth_link_get_nowait(slave_port_id, &link_props);
if (ret < 0) {
rte_eth_dev_callback_unregister(slave_port_id,
RTE_ETH_EVENT_INTR_LSC,
bond_ethdev_lsc_event_callback,
&bonded_eth_dev->data->port_id);
internals->slave_count--;
RTE_BOND_LOG(ERR,
"Slave (port %u) link get failed: %s\n",
slave_port_id, rte_strerror(-ret));
return -1;
}
if (link_props.link_status == ETH_LINK_UP) {
if (internals->active_slave_count == 0 &&
!internals->user_defined_primary_port)
bond_ethdev_primary_set(internals,
slave_port_id);
}
}
/* Add slave details to bonded device */
slave_eth_dev->data->dev_flags |= RTE_ETH_DEV_BONDED_SLAVE;
slave_vlan_filter_set(bonded_port_id, slave_port_id);
return 0;
}
int
rte_eth_bond_slave_add(uint16_t bonded_port_id, uint16_t slave_port_id)
{
struct rte_eth_dev *bonded_eth_dev;
struct bond_dev_private *internals;
int retval;
/* Verify that port id's are valid bonded and slave ports */
if (valid_bonded_port_id(bonded_port_id) != 0)
return -1;
bonded_eth_dev = &rte_eth_devices[bonded_port_id];
internals = bonded_eth_dev->data->dev_private;
rte_spinlock_lock(&internals->lock);
retval = __eth_bond_slave_add_lock_free(bonded_port_id, slave_port_id);
rte_spinlock_unlock(&internals->lock);
return retval;
}
static int
__eth_bond_slave_remove_lock_free(uint16_t bonded_port_id,
uint16_t slave_port_id)
{
struct rte_eth_dev *bonded_eth_dev;
struct bond_dev_private *internals;
struct rte_eth_dev *slave_eth_dev;
struct rte_flow_error flow_error;
struct rte_flow *flow;
int i, slave_idx;
bonded_eth_dev = &rte_eth_devices[bonded_port_id];
internals = bonded_eth_dev->data->dev_private;
if (valid_slave_port_id(slave_port_id, internals->mode) < 0)
return -1;
/* first remove from active slave list */
slave_idx = find_slave_by_id(internals->active_slaves,
internals->active_slave_count, slave_port_id);
if (slave_idx < internals->active_slave_count)
deactivate_slave(bonded_eth_dev, slave_port_id);
slave_idx = -1;
/* now find in slave list */
for (i = 0; i < internals->slave_count; i++)
if (internals->slaves[i].port_id == slave_port_id) {
slave_idx = i;
break;
}
if (slave_idx < 0) {
RTE_BOND_LOG(ERR, "Couldn't find slave in port list, slave count %d",
internals->slave_count);
return -1;
}
/* Un-register link status change callback with bonded device pointer as
* argument*/
rte_eth_dev_callback_unregister(slave_port_id, RTE_ETH_EVENT_INTR_LSC,
bond_ethdev_lsc_event_callback,
&rte_eth_devices[bonded_port_id].data->port_id);
/* Restore original MAC address of slave device */
rte_eth_dev_default_mac_addr_set(slave_port_id,
&(internals->slaves[slave_idx].persisted_mac_addr));
/* remove additional MAC addresses from the slave */
slave_remove_mac_addresses(bonded_eth_dev, slave_port_id);
/*
* Remove bond device flows from slave device.
* Note: don't restore flow isolate mode.
*/
TAILQ_FOREACH(flow, &internals->flow_list, next) {
if (flow->flows[slave_idx] != NULL) {
rte_flow_destroy(slave_port_id, flow->flows[slave_idx],
&flow_error);
flow->flows[slave_idx] = NULL;
}
}
slave_eth_dev = &rte_eth_devices[slave_port_id];
slave_remove(internals, slave_eth_dev);
slave_eth_dev->data->dev_flags &= (~RTE_ETH_DEV_BONDED_SLAVE);
/* first slave in the active list will be the primary by default,
* otherwise use first device in list */
if (internals->current_primary_port == slave_port_id) {
if (internals->active_slave_count > 0)
internals->current_primary_port = internals->active_slaves[0];
else if (internals->slave_count > 0)
internals->current_primary_port = internals->slaves[0].port_id;
else
internals->primary_port = 0;
mac_address_slaves_update(bonded_eth_dev);
}
if (internals->active_slave_count < 1) {
/* if no slaves are any longer attached to bonded device and MAC is not
* user defined then clear MAC of bonded device as it will be reset
* when a new slave is added */
if (internals->slave_count < 1 && !internals->user_defined_mac)
memset(rte_eth_devices[bonded_port_id].data->mac_addrs, 0,
sizeof(*(rte_eth_devices[bonded_port_id].data->mac_addrs)));
}
if (internals->slave_count == 0) {
internals->rx_offload_capa = 0;
internals->tx_offload_capa = 0;
internals->rx_queue_offload_capa = 0;
internals->tx_queue_offload_capa = 0;
internals->flow_type_rss_offloads = ETH_RSS_PROTO_MASK;
internals->reta_size = 0;
internals->candidate_max_rx_pktlen = 0;
internals->max_rx_pktlen = 0;
}
return 0;
}
int
rte_eth_bond_slave_remove(uint16_t bonded_port_id, uint16_t slave_port_id)
{
struct rte_eth_dev *bonded_eth_dev;
struct bond_dev_private *internals;
int retval;
if (valid_bonded_port_id(bonded_port_id) != 0)
return -1;
bonded_eth_dev = &rte_eth_devices[bonded_port_id];
internals = bonded_eth_dev->data->dev_private;
rte_spinlock_lock(&internals->lock);
retval = __eth_bond_slave_remove_lock_free(bonded_port_id, slave_port_id);
rte_spinlock_unlock(&internals->lock);
return retval;
}
int
rte_eth_bond_mode_set(uint16_t bonded_port_id, uint8_t mode)
{
struct rte_eth_dev *bonded_eth_dev;
if (valid_bonded_port_id(bonded_port_id) != 0)
return -1;
bonded_eth_dev = &rte_eth_devices[bonded_port_id];
if (check_for_master_bonded_ethdev(bonded_eth_dev) != 0 &&
mode == BONDING_MODE_8023AD)
return -1;
return bond_ethdev_mode_set(bonded_eth_dev, mode);
}
int
rte_eth_bond_mode_get(uint16_t bonded_port_id)
{
struct bond_dev_private *internals;
if (valid_bonded_port_id(bonded_port_id) != 0)
return -1;
internals = rte_eth_devices[bonded_port_id].data->dev_private;
return internals->mode;
}
int
rte_eth_bond_primary_set(uint16_t bonded_port_id, uint16_t slave_port_id)
{
struct bond_dev_private *internals;
if (valid_bonded_port_id(bonded_port_id) != 0)
return -1;
internals = rte_eth_devices[bonded_port_id].data->dev_private;
if (valid_slave_port_id(slave_port_id, internals->mode) != 0)
return -1;
internals->user_defined_primary_port = 1;
internals->primary_port = slave_port_id;
bond_ethdev_primary_set(internals, slave_port_id);
return 0;
}
int
rte_eth_bond_primary_get(uint16_t bonded_port_id)
{
struct bond_dev_private *internals;
if (valid_bonded_port_id(bonded_port_id) != 0)
return -1;
internals = rte_eth_devices[bonded_port_id].data->dev_private;
if (internals->slave_count < 1)
return -1;
return internals->current_primary_port;
}
int
rte_eth_bond_slaves_get(uint16_t bonded_port_id, uint16_t slaves[],
uint16_t len)
{
struct bond_dev_private *internals;
uint16_t i;
if (valid_bonded_port_id(bonded_port_id) != 0)
return -1;
if (slaves == NULL)
return -1;
internals = rte_eth_devices[bonded_port_id].data->dev_private;
if (internals->slave_count > len)
return -1;
for (i = 0; i < internals->slave_count; i++)
slaves[i] = internals->slaves[i].port_id;
return internals->slave_count;
}
int
rte_eth_bond_active_slaves_get(uint16_t bonded_port_id, uint16_t slaves[],
uint16_t len)
{
struct bond_dev_private *internals;
if (valid_bonded_port_id(bonded_port_id) != 0)
return -1;
if (slaves == NULL)
return -1;
internals = rte_eth_devices[bonded_port_id].data->dev_private;
if (internals->active_slave_count > len)
return -1;
memcpy(slaves, internals->active_slaves,
internals->active_slave_count * sizeof(internals->active_slaves[0]));
return internals->active_slave_count;
}
int
rte_eth_bond_mac_address_set(uint16_t bonded_port_id,
struct rte_ether_addr *mac_addr)
{
struct rte_eth_dev *bonded_eth_dev;
struct bond_dev_private *internals;
if (valid_bonded_port_id(bonded_port_id) != 0)
return -1;
bonded_eth_dev = &rte_eth_devices[bonded_port_id];
internals = bonded_eth_dev->data->dev_private;
/* Set MAC Address of Bonded Device */
if (mac_address_set(bonded_eth_dev, mac_addr))
return -1;
internals->user_defined_mac = 1;
/* Update all slave devices MACs*/
if (internals->slave_count > 0)
return mac_address_slaves_update(bonded_eth_dev);
return 0;
}
int
rte_eth_bond_mac_address_reset(uint16_t bonded_port_id)
{
struct rte_eth_dev *bonded_eth_dev;
struct bond_dev_private *internals;
if (valid_bonded_port_id(bonded_port_id) != 0)
return -1;
bonded_eth_dev = &rte_eth_devices[bonded_port_id];
internals = bonded_eth_dev->data->dev_private;
internals->user_defined_mac = 0;
if (internals->slave_count > 0) {
int slave_port;
/* Get the primary slave location based on the primary port
* number as, while slave_add(), we will keep the primary
* slave based on slave_count,but not based on the primary port.
*/
for (slave_port = 0; slave_port < internals->slave_count;
slave_port++) {
if (internals->slaves[slave_port].port_id ==
internals->primary_port)
break;
}
/* Set MAC Address of Bonded Device */
if (mac_address_set(bonded_eth_dev,
&internals->slaves[slave_port].persisted_mac_addr)
!= 0) {
RTE_BOND_LOG(ERR, "Failed to set MAC address on bonded device");
return -1;
}
/* Update all slave devices MAC addresses */
return mac_address_slaves_update(bonded_eth_dev);
}
/* No need to update anything as no slaves present */
return 0;
}
int
rte_eth_bond_xmit_policy_set(uint16_t bonded_port_id, uint8_t policy)
{
struct bond_dev_private *internals;
if (valid_bonded_port_id(bonded_port_id) != 0)
return -1;
internals = rte_eth_devices[bonded_port_id].data->dev_private;
switch (policy) {
case BALANCE_XMIT_POLICY_LAYER2:
internals->balance_xmit_policy = policy;
internals->burst_xmit_hash = burst_xmit_l2_hash;
break;
case BALANCE_XMIT_POLICY_LAYER23:
internals->balance_xmit_policy = policy;
internals->burst_xmit_hash = burst_xmit_l23_hash;
break;
case BALANCE_XMIT_POLICY_LAYER34:
internals->balance_xmit_policy = policy;
internals->burst_xmit_hash = burst_xmit_l34_hash;
break;
default:
return -1;
}
return 0;
}
int
rte_eth_bond_xmit_policy_get(uint16_t bonded_port_id)
{
struct bond_dev_private *internals;
if (valid_bonded_port_id(bonded_port_id) != 0)
return -1;
internals = rte_eth_devices[bonded_port_id].data->dev_private;
return internals->balance_xmit_policy;
}
int
rte_eth_bond_link_monitoring_set(uint16_t bonded_port_id, uint32_t internal_ms)
{
struct bond_dev_private *internals;
if (valid_bonded_port_id(bonded_port_id) != 0)
return -1;
internals = rte_eth_devices[bonded_port_id].data->dev_private;
internals->link_status_polling_interval_ms = internal_ms;
return 0;
}
int
rte_eth_bond_link_monitoring_get(uint16_t bonded_port_id)
{
struct bond_dev_private *internals;
if (valid_bonded_port_id(bonded_port_id) != 0)
return -1;
internals = rte_eth_devices[bonded_port_id].data->dev_private;
return internals->link_status_polling_interval_ms;
}
int
rte_eth_bond_link_down_prop_delay_set(uint16_t bonded_port_id,
uint32_t delay_ms)
{
struct bond_dev_private *internals;
if (valid_bonded_port_id(bonded_port_id) != 0)
return -1;
internals = rte_eth_devices[bonded_port_id].data->dev_private;
internals->link_down_delay_ms = delay_ms;
return 0;
}
int
rte_eth_bond_link_down_prop_delay_get(uint16_t bonded_port_id)
{
struct bond_dev_private *internals;
if (valid_bonded_port_id(bonded_port_id) != 0)
return -1;
internals = rte_eth_devices[bonded_port_id].data->dev_private;
return internals->link_down_delay_ms;
}
int
rte_eth_bond_link_up_prop_delay_set(uint16_t bonded_port_id, uint32_t delay_ms)
{
struct bond_dev_private *internals;
if (valid_bonded_port_id(bonded_port_id) != 0)
return -1;
internals = rte_eth_devices[bonded_port_id].data->dev_private;
internals->link_up_delay_ms = delay_ms;
return 0;
}
int
rte_eth_bond_link_up_prop_delay_get(uint16_t bonded_port_id)
{
struct bond_dev_private *internals;
if (valid_bonded_port_id(bonded_port_id) != 0)
return -1;
internals = rte_eth_devices[bonded_port_id].data->dev_private;
return internals->link_up_delay_ms;
}