ixgbe: fix whitespace

Signed-off-by: Stephen Hemminger <stephen@networkplumber.org>
Acked-by: Helin Zhang <helin.zhang@intel.com>
This commit is contained in:
Stephen Hemminger 2016-01-12 20:54:10 -08:00 committed by Thomas Monjalon
parent 06554d381d
commit 02fb58d4c7
5 changed files with 42 additions and 42 deletions

View File

@ -46,7 +46,7 @@
__func__, __LINE__); \ __func__, __LINE__); \
return retval; \ return retval; \
} \ } \
} while(0) } while (0)
#define FUNC_PTR_OR_RET(func) do { \ #define FUNC_PTR_OR_RET(func) do { \
if ((func) == NULL) { \ if ((func) == NULL) { \
@ -54,7 +54,7 @@
__func__, __LINE__); \ __func__, __LINE__); \
return; \ return; \
} \ } \
} while(0) } while (0)
/** /**

View File

@ -392,19 +392,19 @@ static int ixgbe_dev_udp_tunnel_port_del(struct rte_eth_dev *dev,
uint32_t idx = (q) / (sizeof ((h)->bitmap[0]) * NBBY); \ uint32_t idx = (q) / (sizeof ((h)->bitmap[0]) * NBBY); \
uint32_t bit = (q) % (sizeof ((h)->bitmap[0]) * NBBY); \ uint32_t bit = (q) % (sizeof ((h)->bitmap[0]) * NBBY); \
(h)->bitmap[idx] |= 1 << bit;\ (h)->bitmap[idx] |= 1 << bit;\
}while(0) } while (0)
#define IXGBE_CLEAR_HWSTRIP(h, q) do{\ #define IXGBE_CLEAR_HWSTRIP(h, q) do{\
uint32_t idx = (q) / (sizeof ((h)->bitmap[0]) * NBBY); \ uint32_t idx = (q) / (sizeof ((h)->bitmap[0]) * NBBY); \
uint32_t bit = (q) % (sizeof ((h)->bitmap[0]) * NBBY); \ uint32_t bit = (q) % (sizeof ((h)->bitmap[0]) * NBBY); \
(h)->bitmap[idx] &= ~(1 << bit);\ (h)->bitmap[idx] &= ~(1 << bit);\
}while(0) } while (0)
#define IXGBE_GET_HWSTRIP(h, q, r) do{\ #define IXGBE_GET_HWSTRIP(h, q, r) do{\
uint32_t idx = (q) / (sizeof ((h)->bitmap[0]) * NBBY); \ uint32_t idx = (q) / (sizeof ((h)->bitmap[0]) * NBBY); \
uint32_t bit = (q) % (sizeof ((h)->bitmap[0]) * NBBY); \ uint32_t bit = (q) % (sizeof ((h)->bitmap[0]) * NBBY); \
(r) = (h)->bitmap[idx] >> bit & 1;\ (r) = (h)->bitmap[idx] >> bit & 1;\
}while(0) } while (0)
/* /*
* The set of PCI devices this driver supports * The set of PCI devices this driver supports
@ -819,7 +819,7 @@ ixgbe_reset_qstat_mappings(struct ixgbe_hw *hw)
{ {
uint32_t i; uint32_t i;
for(i = 0; i != IXGBE_NB_STAT_MAPPING_REGS; i++) { for (i = 0; i != IXGBE_NB_STAT_MAPPING_REGS; i++) {
IXGBE_WRITE_REG(hw, IXGBE_RQSMR(i), 0); IXGBE_WRITE_REG(hw, IXGBE_RQSMR(i), 0);
IXGBE_WRITE_REG(hw, IXGBE_TQSM(i), 0); IXGBE_WRITE_REG(hw, IXGBE_TQSM(i), 0);
} }
@ -1620,7 +1620,7 @@ ixgbe_vlan_hw_strip_bitmap_set(struct rte_eth_dev *dev, uint16_t queue, bool on)
struct ixgbe_hwstrip *hwstrip = struct ixgbe_hwstrip *hwstrip =
IXGBE_DEV_PRIVATE_TO_HWSTRIP_BITMAP(dev->data->dev_private); IXGBE_DEV_PRIVATE_TO_HWSTRIP_BITMAP(dev->data->dev_private);
if(queue >= IXGBE_MAX_RX_QUEUE_NUM) if (queue >= IXGBE_MAX_RX_QUEUE_NUM)
return; return;
if (on) if (on)
@ -1790,21 +1790,21 @@ ixgbe_vlan_hw_extend_enable(struct rte_eth_dev *dev)
static void static void
ixgbe_vlan_offload_set(struct rte_eth_dev *dev, int mask) ixgbe_vlan_offload_set(struct rte_eth_dev *dev, int mask)
{ {
if(mask & ETH_VLAN_STRIP_MASK){ if (mask & ETH_VLAN_STRIP_MASK) {
if (dev->data->dev_conf.rxmode.hw_vlan_strip) if (dev->data->dev_conf.rxmode.hw_vlan_strip)
ixgbe_vlan_hw_strip_enable_all(dev); ixgbe_vlan_hw_strip_enable_all(dev);
else else
ixgbe_vlan_hw_strip_disable_all(dev); ixgbe_vlan_hw_strip_disable_all(dev);
} }
if(mask & ETH_VLAN_FILTER_MASK){ if (mask & ETH_VLAN_FILTER_MASK) {
if (dev->data->dev_conf.rxmode.hw_vlan_filter) if (dev->data->dev_conf.rxmode.hw_vlan_filter)
ixgbe_vlan_hw_filter_enable(dev); ixgbe_vlan_hw_filter_enable(dev);
else else
ixgbe_vlan_hw_filter_disable(dev); ixgbe_vlan_hw_filter_disable(dev);
} }
if(mask & ETH_VLAN_EXTEND_MASK){ if (mask & ETH_VLAN_EXTEND_MASK) {
if (dev->data->dev_conf.rxmode.hw_vlan_extend) if (dev->data->dev_conf.rxmode.hw_vlan_extend)
ixgbe_vlan_hw_extend_enable(dev); ixgbe_vlan_hw_extend_enable(dev);
else else
@ -1819,7 +1819,7 @@ ixgbe_vmdq_vlan_hw_filter_enable(struct rte_eth_dev *dev)
IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private); IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);
/* VLNCTRL: enable vlan filtering and allow all vlan tags through */ /* VLNCTRL: enable vlan filtering and allow all vlan tags through */
uint32_t vlanctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL); uint32_t vlanctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL);
vlanctrl |= IXGBE_VLNCTRL_VFE ; /* enable vlan filters */ vlanctrl |= IXGBE_VLNCTRL_VFE; /* enable vlan filters */
IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, vlanctrl); IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, vlanctrl);
} }
@ -3464,13 +3464,13 @@ ixgbe_dcb_pfc_enable_generic(struct ixgbe_hw *hw,uint8_t tc_num)
/* Low water mark of zero causes XOFF floods */ /* Low water mark of zero causes XOFF floods */
if (hw->fc.current_mode & ixgbe_fc_tx_pause) { if (hw->fc.current_mode & ixgbe_fc_tx_pause) {
/* High/Low water can not be 0 */ /* High/Low water can not be 0 */
if( (!hw->fc.high_water[tc_num])|| (!hw->fc.low_water[tc_num])) { if ((!hw->fc.high_water[tc_num]) || (!hw->fc.low_water[tc_num])) {
PMD_INIT_LOG(ERR, "Invalid water mark configuration"); PMD_INIT_LOG(ERR, "Invalid water mark configuration");
ret_val = IXGBE_ERR_INVALID_LINK_SETTINGS; ret_val = IXGBE_ERR_INVALID_LINK_SETTINGS;
goto out; goto out;
} }
if(hw->fc.low_water[tc_num] >= hw->fc.high_water[tc_num]) { if (hw->fc.low_water[tc_num] >= hw->fc.high_water[tc_num]) {
PMD_INIT_LOG(ERR, "Invalid water mark configuration"); PMD_INIT_LOG(ERR, "Invalid water mark configuration");
ret_val = IXGBE_ERR_INVALID_LINK_SETTINGS; ret_val = IXGBE_ERR_INVALID_LINK_SETTINGS;
goto out; goto out;
@ -3584,7 +3584,7 @@ ixgbe_dcb_pfc_enable(struct rte_eth_dev *dev,uint8_t tc_num)
struct ixgbe_hw *hw = IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private); struct ixgbe_hw *hw = IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);
int32_t ret_val = IXGBE_NOT_IMPLEMENTED; int32_t ret_val = IXGBE_NOT_IMPLEMENTED;
if(hw->mac.type != ixgbe_mac_82598EB) { if (hw->mac.type != ixgbe_mac_82598EB) {
ret_val = ixgbe_dcb_pfc_enable_generic(hw,tc_num); ret_val = ixgbe_dcb_pfc_enable_generic(hw,tc_num);
} }
return ret_val; return ret_val;
@ -3999,10 +3999,10 @@ static void ixgbevf_set_vfta_all(struct rte_eth_dev *dev, bool on)
for (i = 0; i < IXGBE_VFTA_SIZE; i++){ for (i = 0; i < IXGBE_VFTA_SIZE; i++){
vfta = shadow_vfta->vfta[i]; vfta = shadow_vfta->vfta[i];
if(vfta){ if (vfta) {
mask = 1; mask = 1;
for (j = 0; j < 32; j++){ for (j = 0; j < 32; j++){
if(vfta & mask) if (vfta & mask)
ixgbe_set_vfta(hw, (i<<5)+j, 0, on); ixgbe_set_vfta(hw, (i<<5)+j, 0, on);
mask<<=1; mask<<=1;
} }
@ -4026,7 +4026,7 @@ ixgbevf_vlan_filter_set(struct rte_eth_dev *dev, uint16_t vlan_id, int on)
/* vind is not used in VF driver, set to 0, check ixgbe_set_vfta_vf */ /* vind is not used in VF driver, set to 0, check ixgbe_set_vfta_vf */
ret = ixgbe_set_vfta(hw, vlan_id, 0, !!on); ret = ixgbe_set_vfta(hw, vlan_id, 0, !!on);
if(ret){ if (ret) {
PMD_INIT_LOG(ERR, "Unable to set VF vlan"); PMD_INIT_LOG(ERR, "Unable to set VF vlan");
return ret; return ret;
} }
@ -4051,11 +4051,11 @@ ixgbevf_vlan_strip_queue_set(struct rte_eth_dev *dev, uint16_t queue, int on)
PMD_INIT_FUNC_TRACE(); PMD_INIT_FUNC_TRACE();
if(queue >= hw->mac.max_rx_queues) if (queue >= hw->mac.max_rx_queues)
return; return;
ctrl = IXGBE_READ_REG(hw, IXGBE_RXDCTL(queue)); ctrl = IXGBE_READ_REG(hw, IXGBE_RXDCTL(queue));
if(on) if (on)
ctrl |= IXGBE_RXDCTL_VME; ctrl |= IXGBE_RXDCTL_VME;
else else
ctrl &= ~IXGBE_RXDCTL_VME; ctrl &= ~IXGBE_RXDCTL_VME;
@ -4073,10 +4073,10 @@ ixgbevf_vlan_offload_set(struct rte_eth_dev *dev, int mask)
int on = 0; int on = 0;
/* VF function only support hw strip feature, others are not support */ /* VF function only support hw strip feature, others are not support */
if(mask & ETH_VLAN_STRIP_MASK){ if (mask & ETH_VLAN_STRIP_MASK) {
on = !!(dev->data->dev_conf.rxmode.hw_vlan_strip); on = !!(dev->data->dev_conf.rxmode.hw_vlan_strip);
for(i=0; i < hw->mac.max_rx_queues; i++) for (i = 0; i < hw->mac.max_rx_queues; i++)
ixgbevf_vlan_strip_queue_set(dev,i,on); ixgbevf_vlan_strip_queue_set(dev,i,on);
} }
} }
@ -4154,7 +4154,7 @@ ixgbe_uc_hash_table_set(struct rte_eth_dev *dev,struct ether_addr* mac_addr,
uta_shift = vector & ixgbe_uta_bit_mask; uta_shift = vector & ixgbe_uta_bit_mask;
rc = ((uta_info->uta_shadow[uta_idx] >> uta_shift & bit1) != 0); rc = ((uta_info->uta_shadow[uta_idx] >> uta_shift & bit1) != 0);
if(rc == on) if (rc == on)
return 0; return 0;
reg_val = IXGBE_READ_REG(hw, IXGBE_UTA(uta_idx)); reg_val = IXGBE_READ_REG(hw, IXGBE_UTA(uta_idx));
@ -4192,7 +4192,7 @@ ixgbe_uc_all_hash_table_set(struct rte_eth_dev *dev, uint8_t on)
if (hw->mac.type < ixgbe_mac_82599EB) if (hw->mac.type < ixgbe_mac_82599EB)
return -ENOTSUP; return -ENOTSUP;
if(on) { if (on) {
for (i = 0; i < ETH_VMDQ_NUM_UC_HASH_ARRAY; i++) { for (i = 0; i < ETH_VMDQ_NUM_UC_HASH_ARRAY; i++) {
uta_info->uta_shadow[i] = ~0; uta_info->uta_shadow[i] = ~0;
IXGBE_WRITE_REG(hw, IXGBE_UTA(i), ~0); IXGBE_WRITE_REG(hw, IXGBE_UTA(i), ~0);
@ -4385,7 +4385,7 @@ ixgbe_mirror_rule_set(struct rte_eth_dev *dev,
/* search vlan id related pool vlan filter index */ /* search vlan id related pool vlan filter index */
reg_index = ixgbe_find_vlvf_slot(hw, reg_index = ixgbe_find_vlvf_slot(hw,
mirror_conf->vlan.vlan_id[i]); mirror_conf->vlan.vlan_id[i]);
if(reg_index < 0) if (reg_index < 0)
return -EINVAL; return -EINVAL;
vlvf = IXGBE_READ_REG(hw, IXGBE_VLVF(reg_index)); vlvf = IXGBE_READ_REG(hw, IXGBE_VLVF(reg_index));
if ((vlvf & IXGBE_VLVF_VIEN) && if ((vlvf & IXGBE_VLVF_VIEN) &&
@ -4403,8 +4403,8 @@ ixgbe_mirror_rule_set(struct rte_eth_dev *dev,
mr_info->mr_conf[rule_id].vlan.vlan_mask = mr_info->mr_conf[rule_id].vlan.vlan_mask =
mirror_conf->vlan.vlan_mask; mirror_conf->vlan.vlan_mask;
for(i = 0 ;i < ETH_VMDQ_MAX_VLAN_FILTERS; i++) { for (i = 0; i < ETH_VMDQ_MAX_VLAN_FILTERS; i++) {
if(mirror_conf->vlan.vlan_mask & (1ULL << i)) if (mirror_conf->vlan.vlan_mask & (1ULL << i))
mr_info->mr_conf[rule_id].vlan.vlan_id[i] = mr_info->mr_conf[rule_id].vlan.vlan_id[i] =
mirror_conf->vlan.vlan_id[i]; mirror_conf->vlan.vlan_id[i];
} }
@ -4412,7 +4412,7 @@ ixgbe_mirror_rule_set(struct rte_eth_dev *dev,
mv_lsb = 0; mv_lsb = 0;
mv_msb = 0; mv_msb = 0;
mr_info->mr_conf[rule_id].vlan.vlan_mask = 0; mr_info->mr_conf[rule_id].vlan.vlan_mask = 0;
for(i = 0 ;i < ETH_VMDQ_MAX_VLAN_FILTERS; i++) for (i = 0; i < ETH_VMDQ_MAX_VLAN_FILTERS; i++)
mr_info->mr_conf[rule_id].vlan.vlan_id[i] = 0; mr_info->mr_conf[rule_id].vlan.vlan_id[i] = 0;
} }
} }

View File

@ -776,7 +776,7 @@ ixgbe_atr_compute_hash_82599(union ixgbe_atr_input *atr_input,
* *
* hi_hash_dword[31:0] ^= Stream[351:320]; * hi_hash_dword[31:0] ^= Stream[351:320];
* *
* if(key[0]) * if (key[0])
* hash[15:0] ^= Stream[15:0]; * hash[15:0] ^= Stream[15:0];
* *
* for (i = 0; i < 16; i++) { * for (i = 0; i < 16; i++) {

View File

@ -286,7 +286,7 @@ int ixgbe_pf_host_configure(struct rte_eth_dev *eth_dev)
* enable vlan filtering and allow all vlan tags through * enable vlan filtering and allow all vlan tags through
*/ */
vlanctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL); vlanctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL);
vlanctrl |= IXGBE_VLNCTRL_VFE ; /* enable vlan filters */ vlanctrl |= IXGBE_VLNCTRL_VFE; /* enable vlan filters */
IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, vlanctrl); IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, vlanctrl);
/* VFTA - enable all vlan filters */ /* VFTA - enable all vlan filters */

View File

@ -109,7 +109,7 @@ rte_rxmbuf_alloc(struct rte_mempool *mp)
*/ */
#define rte_ixgbe_prefetch(p) rte_prefetch0(p) #define rte_ixgbe_prefetch(p) rte_prefetch0(p)
#else #else
#define rte_ixgbe_prefetch(p) do {} while(0) #define rte_ixgbe_prefetch(p) do {} while (0)
#endif #endif
/********************************************************************* /*********************************************************************
@ -2909,7 +2909,7 @@ ixgbe_vmdq_dcb_configure(struct rte_eth_dev *dev)
pbsize = (uint16_t)(NIC_RX_BUFFER_SIZE / nb_tcs); pbsize = (uint16_t)(NIC_RX_BUFFER_SIZE / nb_tcs);
break; break;
} }
for (i = 0 ; i < nb_tcs; i++) { for (i = 0; i < nb_tcs; i++) {
uint32_t rxpbsize = IXGBE_READ_REG(hw, IXGBE_RXPBSIZE(i)); uint32_t rxpbsize = IXGBE_READ_REG(hw, IXGBE_RXPBSIZE(i));
rxpbsize &= (~(0x3FF << IXGBE_RXPBSIZE_SHIFT)); rxpbsize &= (~(0x3FF << IXGBE_RXPBSIZE_SHIFT));
/* clear 10 bits. */ /* clear 10 bits. */
@ -2955,7 +2955,7 @@ ixgbe_vmdq_dcb_configure(struct rte_eth_dev *dev)
/* VLNCTRL: enable vlan filtering and allow all vlan tags through */ /* VLNCTRL: enable vlan filtering and allow all vlan tags through */
vlanctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL); vlanctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL);
vlanctrl |= IXGBE_VLNCTRL_VFE ; /* enable vlan filters */ vlanctrl |= IXGBE_VLNCTRL_VFE; /* enable vlan filters */
IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, vlanctrl); IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, vlanctrl);
/* VFTA - enable all vlan filters */ /* VFTA - enable all vlan filters */
@ -3212,7 +3212,7 @@ ixgbe_dcb_rx_hw_config(struct ixgbe_hw *hw,
/* VLNCTRL: enable vlan filtering and allow all vlan tags through */ /* VLNCTRL: enable vlan filtering and allow all vlan tags through */
vlanctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL); vlanctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL);
vlanctrl |= IXGBE_VLNCTRL_VFE ; /* enable vlan filters */ vlanctrl |= IXGBE_VLNCTRL_VFE; /* enable vlan filters */
IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, vlanctrl); IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, vlanctrl);
/* VFTA - enable all vlan filters */ /* VFTA - enable all vlan filters */
@ -3352,7 +3352,7 @@ ixgbe_dcb_hw_configure(struct rte_eth_dev *dev,
nb_tcs = dcb_config->num_tcs.pfc_tcs; nb_tcs = dcb_config->num_tcs.pfc_tcs;
/* Unpack map */ /* Unpack map */
ixgbe_dcb_unpack_map_cee(dcb_config, IXGBE_DCB_RX_CONFIG, map); ixgbe_dcb_unpack_map_cee(dcb_config, IXGBE_DCB_RX_CONFIG, map);
if(nb_tcs == ETH_4_TCS) { if (nb_tcs == ETH_4_TCS) {
/* Avoid un-configured priority mapping to TC0 */ /* Avoid un-configured priority mapping to TC0 */
uint8_t j = 4; uint8_t j = 4;
uint8_t mask = 0xFF; uint8_t mask = 0xFF;
@ -3388,11 +3388,11 @@ ixgbe_dcb_hw_configure(struct rte_eth_dev *dev,
break; break;
} }
if(config_dcb_rx) { if (config_dcb_rx) {
/* Set RX buffer size */ /* Set RX buffer size */
pbsize = (uint16_t)(rx_buffer_size / nb_tcs); pbsize = (uint16_t)(rx_buffer_size / nb_tcs);
uint32_t rxpbsize = pbsize << IXGBE_RXPBSIZE_SHIFT; uint32_t rxpbsize = pbsize << IXGBE_RXPBSIZE_SHIFT;
for (i = 0 ; i < nb_tcs; i++) { for (i = 0; i < nb_tcs; i++) {
IXGBE_WRITE_REG(hw, IXGBE_RXPBSIZE(i), rxpbsize); IXGBE_WRITE_REG(hw, IXGBE_RXPBSIZE(i), rxpbsize);
} }
/* zero alloc all unused TCs */ /* zero alloc all unused TCs */
@ -3400,7 +3400,7 @@ ixgbe_dcb_hw_configure(struct rte_eth_dev *dev,
IXGBE_WRITE_REG(hw, IXGBE_RXPBSIZE(i), 0); IXGBE_WRITE_REG(hw, IXGBE_RXPBSIZE(i), 0);
} }
} }
if(config_dcb_tx) { if (config_dcb_tx) {
/* Only support an equally distributed Tx packet buffer strategy. */ /* Only support an equally distributed Tx packet buffer strategy. */
uint32_t txpktsize = IXGBE_TXPBSIZE_MAX / nb_tcs; uint32_t txpktsize = IXGBE_TXPBSIZE_MAX / nb_tcs;
uint32_t txpbthresh = (txpktsize / DCB_TX_PB) - IXGBE_TXPKT_SIZE_MAX; uint32_t txpbthresh = (txpktsize / DCB_TX_PB) - IXGBE_TXPKT_SIZE_MAX;
@ -3421,7 +3421,7 @@ ixgbe_dcb_hw_configure(struct rte_eth_dev *dev,
ixgbe_dcb_calculate_tc_credits_cee(hw, dcb_config,max_frame, ixgbe_dcb_calculate_tc_credits_cee(hw, dcb_config,max_frame,
IXGBE_DCB_RX_CONFIG); IXGBE_DCB_RX_CONFIG);
if(config_dcb_rx) { if (config_dcb_rx) {
/* Unpack CEE standard containers */ /* Unpack CEE standard containers */
ixgbe_dcb_unpack_refill_cee(dcb_config, IXGBE_DCB_RX_CONFIG, refill); ixgbe_dcb_unpack_refill_cee(dcb_config, IXGBE_DCB_RX_CONFIG, refill);
ixgbe_dcb_unpack_max_cee(dcb_config, max); ixgbe_dcb_unpack_max_cee(dcb_config, max);
@ -3431,7 +3431,7 @@ ixgbe_dcb_hw_configure(struct rte_eth_dev *dev,
ixgbe_dcb_hw_arbite_rx_config(hw,refill,max,bwgid,tsa,map); ixgbe_dcb_hw_arbite_rx_config(hw,refill,max,bwgid,tsa,map);
} }
if(config_dcb_tx) { if (config_dcb_tx) {
/* Unpack CEE standard containers */ /* Unpack CEE standard containers */
ixgbe_dcb_unpack_refill_cee(dcb_config, IXGBE_DCB_TX_CONFIG, refill); ixgbe_dcb_unpack_refill_cee(dcb_config, IXGBE_DCB_TX_CONFIG, refill);
ixgbe_dcb_unpack_max_cee(dcb_config, max); ixgbe_dcb_unpack_max_cee(dcb_config, max);
@ -3445,7 +3445,7 @@ ixgbe_dcb_hw_configure(struct rte_eth_dev *dev,
ixgbe_dcb_config_tc_stats_82599(hw, dcb_config); ixgbe_dcb_config_tc_stats_82599(hw, dcb_config);
/* Check if the PFC is supported */ /* Check if the PFC is supported */
if(dev->data->dev_conf.dcb_capability_en & ETH_DCB_PFC_SUPPORT) { if (dev->data->dev_conf.dcb_capability_en & ETH_DCB_PFC_SUPPORT) {
pbsize = (uint16_t)(rx_buffer_size / nb_tcs); pbsize = (uint16_t)(rx_buffer_size / nb_tcs);
for (i = 0; i < nb_tcs; i++) { for (i = 0; i < nb_tcs; i++) {
/* /*
@ -3459,7 +3459,7 @@ ixgbe_dcb_hw_configure(struct rte_eth_dev *dev,
tc->pfc = ixgbe_dcb_pfc_enabled; tc->pfc = ixgbe_dcb_pfc_enabled;
} }
ixgbe_dcb_unpack_pfc_cee(dcb_config, map, &pfc_en); ixgbe_dcb_unpack_pfc_cee(dcb_config, map, &pfc_en);
if(dcb_config->num_tcs.pfc_tcs == ETH_4_TCS) if (dcb_config->num_tcs.pfc_tcs == ETH_4_TCS)
pfc_en &= 0x0F; pfc_en &= 0x0F;
ret = ixgbe_dcb_config_pfc(hw, pfc_en, map); ret = ixgbe_dcb_config_pfc(hw, pfc_en, map);
} }
@ -3534,7 +3534,7 @@ ixgbe_vmdq_rx_hw_configure(struct rte_eth_dev *dev)
/* VLNCTRL: enable vlan filtering and allow all vlan tags through */ /* VLNCTRL: enable vlan filtering and allow all vlan tags through */
vlanctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL); vlanctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL);
vlanctrl |= IXGBE_VLNCTRL_VFE ; /* enable vlan filters */ vlanctrl |= IXGBE_VLNCTRL_VFE; /* enable vlan filters */
IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, vlanctrl); IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, vlanctrl);
/* VFTA - enable all vlan filters */ /* VFTA - enable all vlan filters */