320a9376e9
support for what their boot loader refers to as the "RSYS4GBE", of which there are two instances ("Data Processing Blocks") on the Radisys ATCA-7220.
2000 lines
70 KiB
C
2000 lines
70 KiB
C
/***********************license start***************
|
|
* Copyright (c) 2003-2010 Cavium Inc. (support@cavium.com). All rights
|
|
* reserved.
|
|
*
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions are
|
|
* met:
|
|
*
|
|
* * Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
*
|
|
* * Redistributions in binary form must reproduce the above
|
|
* copyright notice, this list of conditions and the following
|
|
* disclaimer in the documentation and/or other materials provided
|
|
* with the distribution.
|
|
|
|
* * Neither the name of Cavium Inc. nor the names of
|
|
* its contributors may be used to endorse or promote products
|
|
* derived from this software without specific prior written
|
|
* permission.
|
|
|
|
* This Software, including technical data, may be subject to U.S. export control
|
|
* laws, including the U.S. Export Administration Act and its associated
|
|
* regulations, and may be subject to export or import regulations in other
|
|
* countries.
|
|
|
|
* TO THE MAXIMUM EXTENT PERMITTED BY LAW, THE SOFTWARE IS PROVIDED "AS IS"
|
|
* AND WITH ALL FAULTS AND CAVIUM INC. MAKES NO PROMISES, REPRESENTATIONS OR
|
|
* WARRANTIES, EITHER EXPRESS, IMPLIED, STATUTORY, OR OTHERWISE, WITH RESPECT TO
|
|
* THE SOFTWARE, INCLUDING ITS CONDITION, ITS CONFORMITY TO ANY REPRESENTATION OR
|
|
* DESCRIPTION, OR THE EXISTENCE OF ANY LATENT OR PATENT DEFECTS, AND CAVIUM
|
|
* SPECIFICALLY DISCLAIMS ALL IMPLIED (IF ANY) WARRANTIES OF TITLE,
|
|
* MERCHANTABILITY, NONINFRINGEMENT, FITNESS FOR A PARTICULAR PURPOSE, LACK OF
|
|
* VIRUSES, ACCURACY OR COMPLETENESS, QUIET ENJOYMENT, QUIET POSSESSION OR
|
|
* CORRESPONDENCE TO DESCRIPTION. THE ENTIRE RISK ARISING OUT OF USE OR
|
|
* PERFORMANCE OF THE SOFTWARE LIES WITH YOU.
|
|
***********************license end**************************************/
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
/**
|
|
* @file
|
|
*
|
|
* Helper functions for common, but complicated tasks.
|
|
*
|
|
* <hr>$Revision: 70030 $<hr>
|
|
*/
|
|
#ifdef CVMX_BUILD_FOR_LINUX_KERNEL
|
|
#include <linux/module.h>
|
|
#include <asm/octeon/cvmx.h>
|
|
#include <asm/octeon/cvmx-config.h>
|
|
#include <asm/octeon/cvmx-bootmem.h>
|
|
#include <asm/octeon/cvmx-sriox-defs.h>
|
|
#include <asm/octeon/cvmx-npi-defs.h>
|
|
#include <asm/octeon/cvmx-mio-defs.h>
|
|
#include <asm/octeon/cvmx-pexp-defs.h>
|
|
#include <asm/octeon/cvmx-pip-defs.h>
|
|
#include <asm/octeon/cvmx-asxx-defs.h>
|
|
#include <asm/octeon/cvmx-gmxx-defs.h>
|
|
#include <asm/octeon/cvmx-smix-defs.h>
|
|
#include <asm/octeon/cvmx-dbg-defs.h>
|
|
#include <asm/octeon/cvmx-sso-defs.h>
|
|
|
|
#include <asm/octeon/cvmx-gmx.h>
|
|
#include <asm/octeon/cvmx-fpa.h>
|
|
#include <asm/octeon/cvmx-pip.h>
|
|
#include <asm/octeon/cvmx-pko.h>
|
|
#include <asm/octeon/cvmx-ipd.h>
|
|
#include <asm/octeon/cvmx-spi.h>
|
|
#include <asm/octeon/cvmx-clock.h>
|
|
#include <asm/octeon/cvmx-helper.h>
|
|
#include <asm/octeon/cvmx-helper-board.h>
|
|
#include <asm/octeon/cvmx-helper-errata.h>
|
|
#include <asm/octeon/cvmx-helper-cfg.h>
|
|
#else
|
|
#if !defined(__FreeBSD__) || !defined(_KERNEL)
|
|
#include "executive-config.h"
|
|
#endif
|
|
#include "cvmx.h"
|
|
#include "cvmx-sysinfo.h"
|
|
#include "cvmx-bootmem.h"
|
|
#include "cvmx-version.h"
|
|
#include "cvmx-helper-check-defines.h"
|
|
#include "cvmx-gmx.h"
|
|
#if !defined(__FreeBSD__) || !defined(_KERNEL)
|
|
#include "cvmx-error.h"
|
|
#include "cvmx-config.h"
|
|
#endif
|
|
|
|
#include "cvmx-fpa.h"
|
|
#include "cvmx-pip.h"
|
|
#include "cvmx-pko.h"
|
|
#include "cvmx-ipd.h"
|
|
#include "cvmx-spi.h"
|
|
#include "cvmx-helper.h"
|
|
#include "cvmx-helper-board.h"
|
|
#include "cvmx-helper-errata.h"
|
|
#include "cvmx-helper-cfg.h"
|
|
#endif
|
|
|
|
|
|
#ifdef CVMX_ENABLE_PKO_FUNCTIONS
|
|
|
|
/**
|
|
* cvmx_override_pko_queue_priority(int pko_port, uint64_t
|
|
* priorities[16]) is a function pointer. It is meant to allow
|
|
* customization of the PKO queue priorities based on the port
|
|
* number. Users should set this pointer to a function before
|
|
* calling any cvmx-helper operations.
|
|
*/
|
|
CVMX_SHARED void (*cvmx_override_pko_queue_priority)(int ipd_port,
|
|
uint64_t *priorities) = NULL;
|
|
#ifdef CVMX_BUILD_FOR_LINUX_KERNEL
|
|
EXPORT_SYMBOL(cvmx_override_pko_queue_priority);
|
|
#endif
|
|
|
|
/**
|
|
* cvmx_override_ipd_port_setup(int ipd_port) is a function
|
|
* pointer. It is meant to allow customization of the IPD
|
|
* port/port kind setup before packet input/output comes online.
|
|
* It is called after cvmx-helper does the default IPD configuration,
|
|
* but before IPD is enabled. Users should set this pointer to a
|
|
* function before calling any cvmx-helper operations.
|
|
*/
|
|
CVMX_SHARED void (*cvmx_override_ipd_port_setup)(int ipd_port) = NULL;
|
|
|
|
/**
|
|
* Return the number of interfaces the chip has. Each interface
|
|
* may have multiple ports. Most chips support two interfaces,
|
|
* but the CNX0XX and CNX1XX are exceptions. These only support
|
|
* one interface.
|
|
*
|
|
* @return Number of interfaces on chip
|
|
*/
|
|
int cvmx_helper_get_number_of_interfaces(void)
|
|
{
|
|
switch (cvmx_sysinfo_get()->board_type) {
|
|
#if defined(OCTEON_VENDOR_LANNER)
|
|
case CVMX_BOARD_TYPE_CUST_LANNER_MR955:
|
|
return 2;
|
|
case CVMX_BOARD_TYPE_CUST_LANNER_MR730:
|
|
return 1;
|
|
#endif
|
|
#if defined(OCTEON_VENDOR_RADISYS)
|
|
case CVMX_BOARD_TYPE_CUST_RADISYS_RSYS4GBE:
|
|
return 2;
|
|
#endif
|
|
default:
|
|
break;
|
|
}
|
|
|
|
if (OCTEON_IS_MODEL(OCTEON_CN68XX))
|
|
return 9;
|
|
else if (OCTEON_IS_MODEL(OCTEON_CN66XX))
|
|
if (OCTEON_IS_MODEL(OCTEON_CN66XX_PASS1_0))
|
|
return 7;
|
|
else
|
|
return 8;
|
|
else if (OCTEON_IS_MODEL(OCTEON_CN63XX))
|
|
return 6;
|
|
else if (OCTEON_IS_MODEL(OCTEON_CN56XX) || OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN61XX))
|
|
return 4;
|
|
else
|
|
return 3;
|
|
}
|
|
#ifdef CVMX_BUILD_FOR_LINUX_KERNEL
|
|
EXPORT_SYMBOL(cvmx_helper_get_number_of_interfaces);
|
|
#endif
|
|
|
|
|
|
/**
|
|
* Return the number of ports on an interface. Depending on the
|
|
* chip and configuration, this can be 1-16. A value of 0
|
|
* specifies that the interface doesn't exist or isn't usable.
|
|
*
|
|
* @param interface Interface to get the port count for
|
|
*
|
|
* @return Number of ports on interface. Can be Zero.
|
|
*/
|
|
int cvmx_helper_ports_on_interface(int interface)
|
|
{
|
|
if (octeon_has_feature(OCTEON_FEATURE_PKND))
|
|
return cvmx_helper_interface_enumerate(interface);
|
|
else
|
|
return __cvmx_helper_get_num_ipd_ports(interface);
|
|
}
|
|
#ifdef CVMX_BUILD_FOR_LINUX_KERNEL
|
|
EXPORT_SYMBOL(cvmx_helper_ports_on_interface);
|
|
#endif
|
|
|
|
|
|
/**
|
|
* Get the operating mode of an interface. Depending on the Octeon
|
|
* chip and configuration, this function returns an enumeration
|
|
* of the type of packet I/O supported by an interface.
|
|
*
|
|
* @param interface Interface to probe
|
|
*
|
|
* @return Mode of the interface. Unknown or unsupported interfaces return
|
|
* DISABLED.
|
|
*/
|
|
cvmx_helper_interface_mode_t cvmx_helper_interface_get_mode(int interface)
|
|
{
|
|
cvmx_gmxx_inf_mode_t mode;
|
|
|
|
if (OCTEON_IS_MODEL(OCTEON_CN68XX))
|
|
{
|
|
cvmx_mio_qlmx_cfg_t qlm_cfg;
|
|
switch(interface)
|
|
{
|
|
case 0:
|
|
qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(0));
|
|
/* QLM is disabled when QLM SPD is 15. */
|
|
if (qlm_cfg.s.qlm_spd == 15)
|
|
return CVMX_HELPER_INTERFACE_MODE_DISABLED;
|
|
|
|
if (qlm_cfg.s.qlm_cfg == 7)
|
|
return CVMX_HELPER_INTERFACE_MODE_RXAUI;
|
|
else if (qlm_cfg.s.qlm_cfg == 2)
|
|
return CVMX_HELPER_INTERFACE_MODE_SGMII;
|
|
else if (qlm_cfg.s.qlm_cfg == 3)
|
|
return CVMX_HELPER_INTERFACE_MODE_XAUI;
|
|
else
|
|
return CVMX_HELPER_INTERFACE_MODE_DISABLED;
|
|
break;
|
|
case 1:
|
|
qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(0));
|
|
/* QLM is disabled when QLM SPD is 15. */
|
|
if (qlm_cfg.s.qlm_spd == 15)
|
|
return CVMX_HELPER_INTERFACE_MODE_DISABLED;
|
|
|
|
if (qlm_cfg.s.qlm_cfg == 7)
|
|
return CVMX_HELPER_INTERFACE_MODE_RXAUI;
|
|
else
|
|
return CVMX_HELPER_INTERFACE_MODE_DISABLED;
|
|
break;
|
|
case 2:
|
|
case 3:
|
|
case 4:
|
|
qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(interface));
|
|
/* QLM is disabled when QLM SPD is 15. */
|
|
if (qlm_cfg.s.qlm_spd == 15)
|
|
return CVMX_HELPER_INTERFACE_MODE_DISABLED;
|
|
|
|
if (qlm_cfg.s.qlm_cfg == 2)
|
|
return CVMX_HELPER_INTERFACE_MODE_SGMII;
|
|
else if (qlm_cfg.s.qlm_cfg == 3)
|
|
return CVMX_HELPER_INTERFACE_MODE_XAUI;
|
|
else
|
|
return CVMX_HELPER_INTERFACE_MODE_DISABLED;
|
|
break;
|
|
case 5:
|
|
case 6:
|
|
qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(interface - 4));
|
|
/* QLM is disabled when QLM SPD is 15. */
|
|
if (qlm_cfg.s.qlm_spd == 15)
|
|
return CVMX_HELPER_INTERFACE_MODE_DISABLED;
|
|
|
|
if (qlm_cfg.s.qlm_cfg == 1)
|
|
{
|
|
return CVMX_HELPER_INTERFACE_MODE_ILK;
|
|
}
|
|
else
|
|
return CVMX_HELPER_INTERFACE_MODE_DISABLED;
|
|
break;
|
|
case 7:
|
|
qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(3));
|
|
/* QLM is disabled when QLM SPD is 15. */
|
|
if (qlm_cfg.s.qlm_spd == 15)
|
|
return CVMX_HELPER_INTERFACE_MODE_DISABLED;
|
|
else if (qlm_cfg.s.qlm_cfg != 0)
|
|
{
|
|
qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(1));
|
|
if (qlm_cfg.s.qlm_cfg != 0)
|
|
return CVMX_HELPER_INTERFACE_MODE_DISABLED;
|
|
}
|
|
return CVMX_HELPER_INTERFACE_MODE_NPI;
|
|
break;
|
|
case 8:
|
|
return CVMX_HELPER_INTERFACE_MODE_LOOP;
|
|
break;
|
|
default:
|
|
return CVMX_HELPER_INTERFACE_MODE_DISABLED;
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (interface == 2)
|
|
return CVMX_HELPER_INTERFACE_MODE_NPI;
|
|
|
|
if (interface == 3)
|
|
{
|
|
if (OCTEON_IS_MODEL(OCTEON_CN56XX)
|
|
|| OCTEON_IS_MODEL(OCTEON_CN52XX)
|
|
|| OCTEON_IS_MODEL(OCTEON_CN6XXX)
|
|
|| OCTEON_IS_MODEL(OCTEON_CNF7XXX))
|
|
return CVMX_HELPER_INTERFACE_MODE_LOOP;
|
|
else
|
|
return CVMX_HELPER_INTERFACE_MODE_DISABLED;
|
|
}
|
|
|
|
/* Only present in CN63XX & CN66XX Octeon model */
|
|
if ((OCTEON_IS_MODEL(OCTEON_CN63XX) && (interface == 4 || interface == 5))
|
|
|| (OCTEON_IS_MODEL(OCTEON_CN66XX) && interface >= 4 && interface <= 7))
|
|
{
|
|
cvmx_sriox_status_reg_t sriox_status_reg;
|
|
|
|
/* cn66xx pass1.0 has only 2 SRIO interfaces. */
|
|
if ((interface == 5 || interface == 7) && OCTEON_IS_MODEL(OCTEON_CN66XX_PASS1_0))
|
|
return CVMX_HELPER_INTERFACE_MODE_DISABLED;
|
|
|
|
sriox_status_reg.u64 = cvmx_read_csr(CVMX_SRIOX_STATUS_REG(interface-4));
|
|
if (sriox_status_reg.s.srio)
|
|
return CVMX_HELPER_INTERFACE_MODE_SRIO;
|
|
else
|
|
return CVMX_HELPER_INTERFACE_MODE_DISABLED;
|
|
}
|
|
|
|
/* Interface 5 always disabled in CN66XX */
|
|
if (OCTEON_IS_MODEL(OCTEON_CN66XX))
|
|
{
|
|
cvmx_mio_qlmx_cfg_t mio_qlm_cfg;
|
|
|
|
/* QLM2 is SGMII0 and QLM1 is SGMII1 */
|
|
if (interface == 0)
|
|
mio_qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(2));
|
|
else if (interface == 1)
|
|
mio_qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(1));
|
|
else
|
|
return CVMX_HELPER_INTERFACE_MODE_DISABLED;
|
|
|
|
if (mio_qlm_cfg.s.qlm_spd == 15)
|
|
return CVMX_HELPER_INTERFACE_MODE_DISABLED;
|
|
|
|
if (mio_qlm_cfg.s.qlm_cfg == 9)
|
|
return CVMX_HELPER_INTERFACE_MODE_SGMII;
|
|
else if (mio_qlm_cfg.s.qlm_cfg == 11)
|
|
return CVMX_HELPER_INTERFACE_MODE_XAUI;
|
|
else
|
|
return CVMX_HELPER_INTERFACE_MODE_DISABLED;
|
|
}
|
|
else if (OCTEON_IS_MODEL(OCTEON_CN61XX))
|
|
{
|
|
cvmx_mio_qlmx_cfg_t qlm_cfg;
|
|
|
|
if (interface == 0)
|
|
{
|
|
qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(2));
|
|
if (qlm_cfg.s.qlm_cfg == 2)
|
|
return CVMX_HELPER_INTERFACE_MODE_SGMII;
|
|
else if (qlm_cfg.s.qlm_cfg == 3)
|
|
return CVMX_HELPER_INTERFACE_MODE_XAUI;
|
|
else
|
|
return CVMX_HELPER_INTERFACE_MODE_DISABLED;
|
|
}
|
|
else if (interface == 1)
|
|
{
|
|
/* If QLM 1 is PEV0/PEM1 mode, them QLM0 cannot be SGMII/XAUI */
|
|
qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(1));
|
|
if (qlm_cfg.s.qlm_cfg == 1)
|
|
return CVMX_HELPER_INTERFACE_MODE_DISABLED;
|
|
|
|
qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(0));
|
|
if (qlm_cfg.s.qlm_cfg == 2)
|
|
return CVMX_HELPER_INTERFACE_MODE_SGMII;
|
|
else if (qlm_cfg.s.qlm_cfg == 3)
|
|
return CVMX_HELPER_INTERFACE_MODE_XAUI;
|
|
else
|
|
return CVMX_HELPER_INTERFACE_MODE_DISABLED;
|
|
}
|
|
}
|
|
|
|
if (interface == 0 && cvmx_sysinfo_get()->board_type == CVMX_BOARD_TYPE_CN3005_EVB_HS5 && cvmx_sysinfo_get()->board_rev_major == 1)
|
|
{
|
|
/* Lie about interface type of CN3005 board. This board has a switch on port 1 like
|
|
** the other evaluation boards, but it is connected over RGMII instead of GMII. Report
|
|
** GMII mode so that the speed is forced to 1 Gbit full duplex. Other than some initial configuration
|
|
** (which does not use the output of this function) there is no difference in setup between GMII and RGMII modes.
|
|
*/
|
|
return CVMX_HELPER_INTERFACE_MODE_GMII;
|
|
}
|
|
|
|
/* Interface 1 is always disabled on CN31XX and CN30XX */
|
|
if ((interface == 1)
|
|
&& (OCTEON_IS_MODEL(OCTEON_CN31XX)
|
|
|| OCTEON_IS_MODEL(OCTEON_CN30XX)
|
|
|| OCTEON_IS_MODEL(OCTEON_CN50XX)
|
|
|| OCTEON_IS_MODEL(OCTEON_CN52XX)
|
|
|| OCTEON_IS_MODEL(OCTEON_CN63XX)
|
|
|| OCTEON_IS_MODEL(OCTEON_CNF71XX)))
|
|
return CVMX_HELPER_INTERFACE_MODE_DISABLED;
|
|
|
|
mode.u64 = cvmx_read_csr(CVMX_GMXX_INF_MODE(interface));
|
|
|
|
if (OCTEON_IS_MODEL(OCTEON_CN56XX) || OCTEON_IS_MODEL(OCTEON_CN52XX))
|
|
{
|
|
switch(mode.cn56xx.mode)
|
|
{
|
|
case 0: return CVMX_HELPER_INTERFACE_MODE_DISABLED;
|
|
case 1: return CVMX_HELPER_INTERFACE_MODE_XAUI;
|
|
case 2: return CVMX_HELPER_INTERFACE_MODE_SGMII;
|
|
case 3: return CVMX_HELPER_INTERFACE_MODE_PICMG;
|
|
default:return CVMX_HELPER_INTERFACE_MODE_DISABLED;
|
|
}
|
|
}
|
|
else if (OCTEON_IS_MODEL(OCTEON_CN63XX) || OCTEON_IS_MODEL(OCTEON_CNF71XX))
|
|
{
|
|
switch(mode.cn63xx.mode)
|
|
{
|
|
case 0: return CVMX_HELPER_INTERFACE_MODE_SGMII;
|
|
case 1: return CVMX_HELPER_INTERFACE_MODE_XAUI;
|
|
default: return CVMX_HELPER_INTERFACE_MODE_DISABLED;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
if (!mode.s.en)
|
|
return CVMX_HELPER_INTERFACE_MODE_DISABLED;
|
|
|
|
if (mode.s.type)
|
|
{
|
|
if (OCTEON_IS_MODEL(OCTEON_CN38XX) || OCTEON_IS_MODEL(OCTEON_CN58XX))
|
|
return CVMX_HELPER_INTERFACE_MODE_SPI;
|
|
else
|
|
return CVMX_HELPER_INTERFACE_MODE_GMII;
|
|
}
|
|
else
|
|
return CVMX_HELPER_INTERFACE_MODE_RGMII;
|
|
}
|
|
}
|
|
#ifdef CVMX_BUILD_FOR_LINUX_KERNEL
|
|
EXPORT_SYMBOL(cvmx_helper_interface_get_mode);
|
|
#endif
|
|
|
|
/**
|
|
* @INTERNAL
|
|
* Configure the IPD/PIP tagging and QoS options for a specific
|
|
* port. This function determines the POW work queue entry
|
|
* contents for a port. The setup performed here is controlled by
|
|
* the defines in executive-config.h.
|
|
*
|
|
* @param ipd_port Port/Port kind to configure. This follows the IPD numbering,
|
|
* not the per interface numbering
|
|
*
|
|
* @return Zero on success, negative on failure
|
|
*/
|
|
static int __cvmx_helper_port_setup_ipd(int ipd_port)
|
|
{
|
|
cvmx_pip_prt_cfgx_t port_config;
|
|
cvmx_pip_prt_tagx_t tag_config;
|
|
|
|
if (octeon_has_feature(OCTEON_FEATURE_PKND))
|
|
{
|
|
int interface, index, pknd;
|
|
cvmx_pip_prt_cfgbx_t prt_cfgbx;
|
|
|
|
interface = cvmx_helper_get_interface_num(ipd_port);
|
|
index = cvmx_helper_get_interface_index_num(ipd_port);
|
|
pknd = cvmx_helper_get_pknd(interface, index);
|
|
|
|
port_config.u64 = cvmx_read_csr(CVMX_PIP_PRT_CFGX(pknd));
|
|
tag_config.u64 = cvmx_read_csr(CVMX_PIP_PRT_TAGX(pknd));
|
|
|
|
port_config.s.qos = pknd & 0x7;
|
|
|
|
/* Default BPID to use for packets on this port-kind */
|
|
prt_cfgbx.u64 = cvmx_read_csr(CVMX_PIP_PRT_CFGBX(pknd));
|
|
prt_cfgbx.s.bpid = pknd;
|
|
cvmx_write_csr(CVMX_PIP_PRT_CFGBX(pknd), prt_cfgbx.u64);
|
|
}
|
|
else
|
|
{
|
|
port_config.u64 = cvmx_read_csr(CVMX_PIP_PRT_CFGX(ipd_port));
|
|
tag_config.u64 = cvmx_read_csr(CVMX_PIP_PRT_TAGX(ipd_port));
|
|
|
|
/* Have each port go to a different POW queue */
|
|
port_config.s.qos = ipd_port & 0x7;
|
|
}
|
|
|
|
/* Process the headers and place the IP header in the work queue */
|
|
port_config.s.mode = CVMX_HELPER_INPUT_PORT_SKIP_MODE;
|
|
|
|
tag_config.s.ip6_src_flag = CVMX_HELPER_INPUT_TAG_IPV6_SRC_IP;
|
|
tag_config.s.ip6_dst_flag = CVMX_HELPER_INPUT_TAG_IPV6_DST_IP;
|
|
tag_config.s.ip6_sprt_flag = CVMX_HELPER_INPUT_TAG_IPV6_SRC_PORT;
|
|
tag_config.s.ip6_dprt_flag = CVMX_HELPER_INPUT_TAG_IPV6_DST_PORT;
|
|
tag_config.s.ip6_nxth_flag = CVMX_HELPER_INPUT_TAG_IPV6_NEXT_HEADER;
|
|
tag_config.s.ip4_src_flag = CVMX_HELPER_INPUT_TAG_IPV4_SRC_IP;
|
|
tag_config.s.ip4_dst_flag = CVMX_HELPER_INPUT_TAG_IPV4_DST_IP;
|
|
tag_config.s.ip4_sprt_flag = CVMX_HELPER_INPUT_TAG_IPV4_SRC_PORT;
|
|
tag_config.s.ip4_dprt_flag = CVMX_HELPER_INPUT_TAG_IPV4_DST_PORT;
|
|
tag_config.s.ip4_pctl_flag = CVMX_HELPER_INPUT_TAG_IPV4_PROTOCOL;
|
|
tag_config.s.inc_prt_flag = CVMX_HELPER_INPUT_TAG_INPUT_PORT;
|
|
tag_config.s.tcp6_tag_type = CVMX_HELPER_INPUT_TAG_TYPE;
|
|
tag_config.s.tcp4_tag_type = CVMX_HELPER_INPUT_TAG_TYPE;
|
|
tag_config.s.ip6_tag_type = CVMX_HELPER_INPUT_TAG_TYPE;
|
|
tag_config.s.ip4_tag_type = CVMX_HELPER_INPUT_TAG_TYPE;
|
|
tag_config.s.non_tag_type = CVMX_HELPER_INPUT_TAG_TYPE;
|
|
/* Put all packets in group 0. Other groups can be used by the app */
|
|
tag_config.s.grp = 0;
|
|
|
|
cvmx_pip_config_port(ipd_port, port_config, tag_config);
|
|
|
|
/* Give the user a chance to override our setting for each port */
|
|
if (cvmx_override_ipd_port_setup)
|
|
cvmx_override_ipd_port_setup(ipd_port);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* Enable or disable FCS stripping for all the ports on an interface.
|
|
*
|
|
* @param interface
|
|
* @param nports number of ports
|
|
* @param has_fcs 0 for disable and !0 for enable
|
|
*/
|
|
static int cvmx_helper_fcs_op(int interface, int nports, int has_fcs)
|
|
{
|
|
uint64_t port_bit;
|
|
int index;
|
|
int pknd;
|
|
cvmx_pip_sub_pkind_fcsx_t pkind_fcsx;
|
|
cvmx_pip_prt_cfgx_t port_cfg;
|
|
|
|
if (!octeon_has_feature(OCTEON_FEATURE_PKND))
|
|
return 0;
|
|
|
|
port_bit = 0;
|
|
for (index = 0; index < nports; index++)
|
|
port_bit |= ((uint64_t)1 << cvmx_helper_get_pknd(interface, index));
|
|
|
|
pkind_fcsx.u64 = cvmx_read_csr(CVMX_PIP_SUB_PKIND_FCSX(0));
|
|
if (has_fcs)
|
|
pkind_fcsx.s.port_bit |= port_bit;
|
|
else
|
|
pkind_fcsx.s.port_bit &= ~port_bit;
|
|
cvmx_write_csr(CVMX_PIP_SUB_PKIND_FCSX(0), pkind_fcsx.u64);
|
|
|
|
for (pknd = 0; pknd < 64; pknd++)
|
|
{
|
|
if ((1ull << pknd) & port_bit)
|
|
{
|
|
port_cfg.u64 = cvmx_read_csr(CVMX_PIP_PRT_CFGX(pknd));
|
|
port_cfg.s.crc_en = (has_fcs) ? 1 : 0;
|
|
cvmx_write_csr(CVMX_PIP_PRT_CFGX(pknd), port_cfg.u64);
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* Determine the actual number of hardware ports connected to an
|
|
* interface. It doesn't setup the ports or enable them.
|
|
*
|
|
* @param interface Interface to enumerate
|
|
*
|
|
* @return The number of ports on the interface, negative on failure
|
|
*/
|
|
int cvmx_helper_interface_enumerate(int interface)
|
|
{
|
|
switch (cvmx_helper_interface_get_mode(interface)) {
|
|
/* XAUI is a single high speed port */
|
|
case CVMX_HELPER_INTERFACE_MODE_XAUI:
|
|
case CVMX_HELPER_INTERFACE_MODE_RXAUI:
|
|
return __cvmx_helper_xaui_enumerate(interface);
|
|
/* RGMII/GMII/MII are all treated about the same. Most functions
|
|
refer to these ports as RGMII */
|
|
case CVMX_HELPER_INTERFACE_MODE_RGMII:
|
|
case CVMX_HELPER_INTERFACE_MODE_GMII:
|
|
return __cvmx_helper_rgmii_enumerate(interface);
|
|
/* SPI4 can have 1-16 ports depending on the device at the other end */
|
|
case CVMX_HELPER_INTERFACE_MODE_SPI:
|
|
return __cvmx_helper_spi_enumerate(interface);
|
|
/* SGMII can have 1-4 ports depending on how many are hooked up */
|
|
case CVMX_HELPER_INTERFACE_MODE_SGMII:
|
|
case CVMX_HELPER_INTERFACE_MODE_PICMG:
|
|
return __cvmx_helper_sgmii_enumerate(interface);
|
|
/* PCI target Network Packet Interface */
|
|
case CVMX_HELPER_INTERFACE_MODE_NPI:
|
|
return __cvmx_helper_npi_enumerate(interface);
|
|
/* Special loopback only ports. These are not the same
|
|
* as other ports in loopback mode */
|
|
case CVMX_HELPER_INTERFACE_MODE_LOOP:
|
|
return __cvmx_helper_loop_enumerate(interface);
|
|
/* SRIO has 2^N ports, where N is number of interfaces */
|
|
case CVMX_HELPER_INTERFACE_MODE_SRIO:
|
|
return __cvmx_helper_srio_enumerate(interface);
|
|
|
|
case CVMX_HELPER_INTERFACE_MODE_ILK:
|
|
return __cvmx_helper_ilk_enumerate(interface);
|
|
/* These types don't support ports to IPD/PKO */
|
|
case CVMX_HELPER_INTERFACE_MODE_DISABLED:
|
|
case CVMX_HELPER_INTERFACE_MODE_PCIE:
|
|
default:
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* This function probes an interface to determine the actual number of
|
|
* hardware ports connected to it. It does some setup the ports but
|
|
* doesn't enable them. The main goal here is to set the global
|
|
* interface_port_count[interface] correctly. Final hardware setup of
|
|
* the ports will be performed later.
|
|
*
|
|
* @param interface Interface to probe
|
|
*
|
|
* @return Zero on success, negative on failure
|
|
*/
|
|
int cvmx_helper_interface_probe(int interface)
|
|
{
|
|
/* At this stage in the game we don't want packets to be moving yet.
|
|
The following probe calls should perform hardware setup
|
|
needed to determine port counts. Receive must still be disabled */
|
|
int nports;
|
|
int has_fcs;
|
|
enum cvmx_pko_padding padding = CVMX_PKO_PADDING_NONE;
|
|
|
|
nports = -1;
|
|
has_fcs = 0;
|
|
switch (cvmx_helper_interface_get_mode(interface))
|
|
{
|
|
/* These types don't support ports to IPD/PKO */
|
|
case CVMX_HELPER_INTERFACE_MODE_DISABLED:
|
|
case CVMX_HELPER_INTERFACE_MODE_PCIE:
|
|
nports = 0;
|
|
break;
|
|
/* XAUI is a single high speed port */
|
|
case CVMX_HELPER_INTERFACE_MODE_XAUI:
|
|
case CVMX_HELPER_INTERFACE_MODE_RXAUI:
|
|
nports = __cvmx_helper_xaui_probe(interface);
|
|
has_fcs = 1;
|
|
padding = CVMX_PKO_PADDING_60;
|
|
break;
|
|
/* RGMII/GMII/MII are all treated about the same. Most functions
|
|
refer to these ports as RGMII */
|
|
case CVMX_HELPER_INTERFACE_MODE_RGMII:
|
|
case CVMX_HELPER_INTERFACE_MODE_GMII:
|
|
nports = __cvmx_helper_rgmii_probe(interface);
|
|
padding = CVMX_PKO_PADDING_60;
|
|
break;
|
|
/* SPI4 can have 1-16 ports depending on the device at the other end */
|
|
case CVMX_HELPER_INTERFACE_MODE_SPI:
|
|
nports = __cvmx_helper_spi_probe(interface);
|
|
padding = CVMX_PKO_PADDING_60;
|
|
break;
|
|
/* SGMII can have 1-4 ports depending on how many are hooked up */
|
|
case CVMX_HELPER_INTERFACE_MODE_SGMII:
|
|
padding = CVMX_PKO_PADDING_60;
|
|
case CVMX_HELPER_INTERFACE_MODE_PICMG:
|
|
nports = __cvmx_helper_sgmii_probe(interface);
|
|
has_fcs = 1;
|
|
break;
|
|
/* PCI target Network Packet Interface */
|
|
case CVMX_HELPER_INTERFACE_MODE_NPI:
|
|
nports = __cvmx_helper_npi_probe(interface);
|
|
break;
|
|
/* Special loopback only ports. These are not the same as other ports
|
|
in loopback mode */
|
|
case CVMX_HELPER_INTERFACE_MODE_LOOP:
|
|
nports = __cvmx_helper_loop_probe(interface);
|
|
break;
|
|
/* SRIO has 2^N ports, where N is number of interfaces */
|
|
case CVMX_HELPER_INTERFACE_MODE_SRIO:
|
|
nports = __cvmx_helper_srio_probe(interface);
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_ILK:
|
|
nports = __cvmx_helper_ilk_probe(interface);
|
|
has_fcs = 1;
|
|
padding = CVMX_PKO_PADDING_60;
|
|
break;
|
|
}
|
|
|
|
if (nports == -1)
|
|
return -1;
|
|
|
|
if (!octeon_has_feature(OCTEON_FEATURE_PKND))
|
|
has_fcs = 0;
|
|
|
|
nports = __cvmx_helper_board_interface_probe(interface, nports);
|
|
__cvmx_helper_init_interface(interface, nports, has_fcs, padding);
|
|
cvmx_helper_fcs_op(interface, nports, has_fcs);
|
|
|
|
/* Make sure all global variables propagate to other cores */
|
|
CVMX_SYNCWS;
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
/**
|
|
* @INTERNAL
|
|
* Setup the IPD/PIP for the ports on an interface. Packet
|
|
* classification and tagging are set for every port on the
|
|
* interface. The number of ports on the interface must already
|
|
* have been probed.
|
|
*
|
|
* @param interface Interface to setup IPD/PIP for
|
|
*
|
|
* @return Zero on success, negative on failure
|
|
*/
|
|
static int __cvmx_helper_interface_setup_ipd(int interface)
|
|
{
|
|
|
|
cvmx_helper_interface_mode_t mode;
|
|
int ipd_port = cvmx_helper_get_ipd_port(interface, 0);
|
|
int num_ports = cvmx_helper_ports_on_interface(interface);
|
|
int delta;
|
|
|
|
if (num_ports == CVMX_HELPER_CFG_INVALID_VALUE)
|
|
return 0;
|
|
|
|
mode = cvmx_helper_interface_get_mode(interface);
|
|
|
|
if (mode == CVMX_HELPER_INTERFACE_MODE_LOOP)
|
|
__cvmx_helper_loop_enable(interface);
|
|
|
|
delta = 1;
|
|
if (octeon_has_feature(OCTEON_FEATURE_PKND))
|
|
{
|
|
if (mode == CVMX_HELPER_INTERFACE_MODE_SGMII)
|
|
delta = 16;
|
|
}
|
|
|
|
while (num_ports--)
|
|
{
|
|
__cvmx_helper_port_setup_ipd(ipd_port);
|
|
ipd_port += delta;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
/**
|
|
* @INTERNAL
|
|
* Setup global setting for IPD/PIP not related to a specific
|
|
* interface or port. This must be called before IPD is enabled.
|
|
*
|
|
* @return Zero on success, negative on failure.
|
|
*/
|
|
static int __cvmx_helper_global_setup_ipd(void)
|
|
{
|
|
#ifndef CVMX_HELPER_IPD_DRAM_MODE
|
|
#define CVMX_HELPER_IPD_DRAM_MODE CVMX_IPD_OPC_MODE_STT
|
|
#endif
|
|
/* Setup the global packet input options */
|
|
cvmx_ipd_config(CVMX_FPA_PACKET_POOL_SIZE/8,
|
|
CVMX_HELPER_FIRST_MBUFF_SKIP/8,
|
|
CVMX_HELPER_NOT_FIRST_MBUFF_SKIP/8,
|
|
(CVMX_HELPER_FIRST_MBUFF_SKIP+8) / 128, /* The +8 is to account for the next ptr */
|
|
(CVMX_HELPER_NOT_FIRST_MBUFF_SKIP+8) / 128, /* The +8 is to account for the next ptr */
|
|
CVMX_FPA_WQE_POOL,
|
|
CVMX_HELPER_IPD_DRAM_MODE,
|
|
1);
|
|
return 0;
|
|
}
|
|
|
|
|
|
/**
|
|
* @INTERNAL
|
|
* Setup the PKO for the ports on an interface. The number of
|
|
* queues per port and the priority of each PKO output queue
|
|
* is set here. PKO must be disabled when this function is called.
|
|
*
|
|
* @param interface Interface to setup PKO for
|
|
*
|
|
* @return Zero on success, negative on failure
|
|
*/
|
|
static int __cvmx_helper_interface_setup_pko(int interface)
|
|
{
|
|
/* Each packet output queue has an associated priority. The higher the
|
|
priority, the more often it can send a packet. A priority of 8 means
|
|
it can send in all 8 rounds of contention. We're going to make each
|
|
queue one less than the last.
|
|
The vector of priorities has been extended to support CN5xxx CPUs,
|
|
where up to 16 queues can be associated to a port.
|
|
To keep backward compatibility we don't change the initial 8
|
|
priorities and replicate them in the second half.
|
|
With per-core PKO queues (PKO lockless operation) all queues have
|
|
the same priority. */
|
|
/* uint64_t priorities[16] = {8,7,6,5,4,3,2,1,8,7,6,5,4,3,2,1}; */
|
|
uint64_t priorities[16] = {[0 ... 15] = 8};
|
|
|
|
/* Setup the IPD/PIP and PKO for the ports discovered above. Here packet
|
|
classification, tagging and output priorities are set */
|
|
int ipd_port = cvmx_helper_get_ipd_port(interface, 0);
|
|
int num_ports = cvmx_helper_ports_on_interface(interface);
|
|
while (num_ports--)
|
|
{
|
|
/* Give the user a chance to override the per queue priorities */
|
|
if (cvmx_override_pko_queue_priority)
|
|
cvmx_override_pko_queue_priority(ipd_port, priorities);
|
|
|
|
cvmx_pko_config_port(ipd_port, cvmx_pko_get_base_queue_per_core(ipd_port, 0),
|
|
cvmx_pko_get_num_queues(ipd_port), priorities);
|
|
ipd_port++;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
|
|
/**
|
|
* @INTERNAL
|
|
* Setup global setting for PKO not related to a specific
|
|
* interface or port. This must be called before PKO is enabled.
|
|
*
|
|
* @return Zero on success, negative on failure.
|
|
*/
|
|
static int __cvmx_helper_global_setup_pko(void)
|
|
{
|
|
/* Disable tagwait FAU timeout. This needs to be done before anyone might
|
|
start packet output using tags */
|
|
cvmx_iob_fau_timeout_t fau_to;
|
|
fau_to.u64 = 0;
|
|
fau_to.s.tout_val = 0xfff;
|
|
fau_to.s.tout_enb = 0;
|
|
cvmx_write_csr(CVMX_IOB_FAU_TIMEOUT, fau_to.u64);
|
|
|
|
if (OCTEON_IS_MODEL(OCTEON_CN68XX)) {
|
|
cvmx_pko_reg_min_pkt_t min_pkt;
|
|
|
|
min_pkt.u64 = 0;
|
|
min_pkt.s.size1 = 59;
|
|
min_pkt.s.size2 = 59;
|
|
min_pkt.s.size3 = 59;
|
|
min_pkt.s.size4 = 59;
|
|
min_pkt.s.size5 = 59;
|
|
min_pkt.s.size6 = 59;
|
|
min_pkt.s.size7 = 59;
|
|
cvmx_write_csr(CVMX_PKO_REG_MIN_PKT, min_pkt.u64);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
/**
|
|
* @INTERNAL
|
|
* Setup global backpressure setting.
|
|
*
|
|
* @return Zero on success, negative on failure
|
|
*/
|
|
static int __cvmx_helper_global_setup_backpressure(void)
|
|
{
|
|
#if CVMX_HELPER_DISABLE_RGMII_BACKPRESSURE
|
|
/* Disable backpressure if configured to do so */
|
|
/* Disable backpressure (pause frame) generation */
|
|
int num_interfaces = cvmx_helper_get_number_of_interfaces();
|
|
int interface;
|
|
for (interface=0; interface<num_interfaces; interface++)
|
|
{
|
|
switch (cvmx_helper_interface_get_mode(interface))
|
|
{
|
|
case CVMX_HELPER_INTERFACE_MODE_DISABLED:
|
|
case CVMX_HELPER_INTERFACE_MODE_PCIE:
|
|
case CVMX_HELPER_INTERFACE_MODE_SRIO:
|
|
case CVMX_HELPER_INTERFACE_MODE_ILK:
|
|
case CVMX_HELPER_INTERFACE_MODE_NPI:
|
|
case CVMX_HELPER_INTERFACE_MODE_LOOP:
|
|
case CVMX_HELPER_INTERFACE_MODE_XAUI:
|
|
case CVMX_HELPER_INTERFACE_MODE_RXAUI:
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_RGMII:
|
|
case CVMX_HELPER_INTERFACE_MODE_GMII:
|
|
case CVMX_HELPER_INTERFACE_MODE_SPI:
|
|
case CVMX_HELPER_INTERFACE_MODE_SGMII:
|
|
case CVMX_HELPER_INTERFACE_MODE_PICMG:
|
|
cvmx_gmx_set_backpressure_override(interface, 0xf);
|
|
break;
|
|
}
|
|
}
|
|
//cvmx_dprintf("Disabling backpressure\n");
|
|
#endif
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* @INTERNAL
|
|
* Verify the per port IPD backpressure is aligned properly.
|
|
* @return Zero if working, non zero if misaligned
|
|
*/
|
|
static int __cvmx_helper_backpressure_is_misaligned(void)
|
|
{
|
|
uint64_t ipd_int_enb;
|
|
cvmx_ipd_ctl_status_t ipd_reg;
|
|
uint64_t bp_status0;
|
|
uint64_t bp_status1;
|
|
const int port0 = 0;
|
|
const int port1 = 16;
|
|
cvmx_helper_interface_mode_t mode0 = cvmx_helper_interface_get_mode(0);
|
|
cvmx_helper_interface_mode_t mode1 = cvmx_helper_interface_get_mode(1);
|
|
|
|
/* Disable error interrupts while we check backpressure */
|
|
ipd_int_enb = cvmx_read_csr(CVMX_IPD_INT_ENB);
|
|
cvmx_write_csr(CVMX_IPD_INT_ENB, 0);
|
|
|
|
/* Enable per port backpressure */
|
|
ipd_reg.u64 = cvmx_read_csr(CVMX_IPD_CTL_STATUS);
|
|
ipd_reg.s.pbp_en = 1;
|
|
cvmx_write_csr(CVMX_IPD_CTL_STATUS, ipd_reg.u64);
|
|
|
|
if (mode0 != CVMX_HELPER_INTERFACE_MODE_DISABLED)
|
|
{
|
|
/* Enable backpressure for port with a zero threshold */
|
|
cvmx_write_csr(CVMX_IPD_PORTX_BP_PAGE_CNT(port0), 1<<17);
|
|
/* Add 1000 to the page count to simulate packets coming in */
|
|
cvmx_write_csr(CVMX_IPD_SUB_PORT_BP_PAGE_CNT, (port0<<25) | 1000);
|
|
}
|
|
|
|
if (mode1 != CVMX_HELPER_INTERFACE_MODE_DISABLED)
|
|
{
|
|
/* Enable backpressure for port with a zero threshold */
|
|
cvmx_write_csr(CVMX_IPD_PORTX_BP_PAGE_CNT(port1), 1<<17);
|
|
/* Add 1000 to the page count to simulate packets coming in */
|
|
cvmx_write_csr(CVMX_IPD_SUB_PORT_BP_PAGE_CNT, (port1<<25) | 1000);
|
|
}
|
|
|
|
/* Wait 500 cycles for the BP to update */
|
|
cvmx_wait(500);
|
|
|
|
/* Read the BP state from the debug select register */
|
|
switch (mode0)
|
|
{
|
|
case CVMX_HELPER_INTERFACE_MODE_SPI:
|
|
cvmx_write_csr(CVMX_NPI_DBG_SELECT, 0x9004);
|
|
bp_status0 = cvmx_read_csr(CVMX_DBG_DATA);
|
|
bp_status0 = 0xffff & ~bp_status0;
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_RGMII:
|
|
case CVMX_HELPER_INTERFACE_MODE_GMII:
|
|
cvmx_write_csr(CVMX_NPI_DBG_SELECT, 0x0e00);
|
|
bp_status0 = 0xffff & cvmx_read_csr(CVMX_DBG_DATA);
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_XAUI:
|
|
case CVMX_HELPER_INTERFACE_MODE_SGMII:
|
|
case CVMX_HELPER_INTERFACE_MODE_PICMG:
|
|
cvmx_write_csr(CVMX_PEXP_NPEI_DBG_SELECT, 0x0e00);
|
|
bp_status0 = 0xffff & cvmx_read_csr(CVMX_PEXP_NPEI_DBG_DATA);
|
|
break;
|
|
default:
|
|
bp_status0 = 1<<port0;
|
|
break;
|
|
}
|
|
|
|
/* Read the BP state from the debug select register */
|
|
switch (mode1)
|
|
{
|
|
case CVMX_HELPER_INTERFACE_MODE_SPI:
|
|
cvmx_write_csr(CVMX_NPI_DBG_SELECT, 0x9804);
|
|
bp_status1 = cvmx_read_csr(CVMX_DBG_DATA);
|
|
bp_status1 = 0xffff & ~bp_status1;
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_RGMII:
|
|
case CVMX_HELPER_INTERFACE_MODE_GMII:
|
|
cvmx_write_csr(CVMX_NPI_DBG_SELECT, 0x1600);
|
|
bp_status1 = 0xffff & cvmx_read_csr(CVMX_DBG_DATA);
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_XAUI:
|
|
case CVMX_HELPER_INTERFACE_MODE_SGMII:
|
|
case CVMX_HELPER_INTERFACE_MODE_PICMG:
|
|
cvmx_write_csr(CVMX_PEXP_NPEI_DBG_SELECT, 0x1600);
|
|
bp_status1 = 0xffff & cvmx_read_csr(CVMX_PEXP_NPEI_DBG_DATA);
|
|
break;
|
|
default:
|
|
bp_status1 = 1<<(port1-16);
|
|
break;
|
|
}
|
|
|
|
if (mode0 != CVMX_HELPER_INTERFACE_MODE_DISABLED)
|
|
{
|
|
/* Shutdown BP */
|
|
cvmx_write_csr(CVMX_IPD_SUB_PORT_BP_PAGE_CNT, (port0<<25) | (0x1ffffff & -1000));
|
|
cvmx_write_csr(CVMX_IPD_PORTX_BP_PAGE_CNT(port0), 0);
|
|
}
|
|
|
|
if (mode1 != CVMX_HELPER_INTERFACE_MODE_DISABLED)
|
|
{
|
|
/* Shutdown BP */
|
|
cvmx_write_csr(CVMX_IPD_SUB_PORT_BP_PAGE_CNT, (port1<<25) | (0x1ffffff & -1000));
|
|
cvmx_write_csr(CVMX_IPD_PORTX_BP_PAGE_CNT(port1), 0);
|
|
}
|
|
|
|
/* Clear any error interrupts that might have been set */
|
|
cvmx_write_csr(CVMX_IPD_INT_SUM, 0x1f);
|
|
cvmx_write_csr(CVMX_IPD_INT_ENB, ipd_int_enb);
|
|
|
|
return ((bp_status0 != 1ull<<port0) || (bp_status1 != 1ull<<(port1-16)));
|
|
}
|
|
|
|
|
|
/**
|
|
* @INTERNAL
|
|
* Enable packet input/output from the hardware. This function is
|
|
* called after all internal setup is complete and IPD is enabled.
|
|
* After this function completes, packets will be accepted from the
|
|
* hardware ports. PKO should still be disabled to make sure packets
|
|
* aren't sent out partially setup hardware.
|
|
*
|
|
* @param interface Interface to enable
|
|
*
|
|
* @return Zero on success, negative on failure
|
|
*/
|
|
static int __cvmx_helper_packet_hardware_enable(int interface)
|
|
{
|
|
int result = 0;
|
|
switch (cvmx_helper_interface_get_mode(interface))
|
|
{
|
|
/* These types don't support ports to IPD/PKO */
|
|
case CVMX_HELPER_INTERFACE_MODE_DISABLED:
|
|
case CVMX_HELPER_INTERFACE_MODE_PCIE:
|
|
case CVMX_HELPER_INTERFACE_MODE_LOOP:
|
|
/* Nothing to do */
|
|
break;
|
|
/* XAUI is a single high speed port */
|
|
case CVMX_HELPER_INTERFACE_MODE_XAUI:
|
|
case CVMX_HELPER_INTERFACE_MODE_RXAUI:
|
|
result = __cvmx_helper_xaui_enable(interface);
|
|
break;
|
|
/* RGMII/GMII/MII are all treated about the same. Most functions
|
|
refer to these ports as RGMII */
|
|
case CVMX_HELPER_INTERFACE_MODE_RGMII:
|
|
case CVMX_HELPER_INTERFACE_MODE_GMII:
|
|
result = __cvmx_helper_rgmii_enable(interface);
|
|
break;
|
|
/* SPI4 can have 1-16 ports depending on the device at the other end */
|
|
case CVMX_HELPER_INTERFACE_MODE_SPI:
|
|
result = __cvmx_helper_spi_enable(interface);
|
|
break;
|
|
/* SGMII can have 1-4 ports depending on how many are hooked up */
|
|
case CVMX_HELPER_INTERFACE_MODE_SGMII:
|
|
case CVMX_HELPER_INTERFACE_MODE_PICMG:
|
|
result = __cvmx_helper_sgmii_enable(interface);
|
|
break;
|
|
/* PCI target Network Packet Interface */
|
|
case CVMX_HELPER_INTERFACE_MODE_NPI:
|
|
result = __cvmx_helper_npi_enable(interface);
|
|
break;
|
|
/* SRIO has 2^N ports, where N is number of interfaces */
|
|
case CVMX_HELPER_INTERFACE_MODE_SRIO:
|
|
result = __cvmx_helper_srio_enable(interface);
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_ILK:
|
|
result = __cvmx_helper_ilk_enable(interface);
|
|
break;
|
|
}
|
|
result |= __cvmx_helper_board_hardware_enable(interface);
|
|
return result;
|
|
}
|
|
|
|
|
|
/**
|
|
* Called after all internal packet IO paths are setup. This
|
|
* function enables IPD/PIP and begins packet input and output.
|
|
*
|
|
* @return Zero on success, negative on failure
|
|
*/
|
|
int cvmx_helper_ipd_and_packet_input_enable(void)
|
|
{
|
|
int num_interfaces;
|
|
int interface;
|
|
|
|
/* Enable IPD */
|
|
cvmx_ipd_enable();
|
|
|
|
/* Time to enable hardware ports packet input and output. Note that at this
|
|
point IPD/PIP must be fully functional and PKO must be disabled */
|
|
num_interfaces = cvmx_helper_get_number_of_interfaces();
|
|
for (interface=0; interface<num_interfaces; interface++)
|
|
{
|
|
if (cvmx_helper_ports_on_interface(interface) > 0)
|
|
{
|
|
//cvmx_dprintf("Enabling packet I/O on interface %d\n", interface);
|
|
__cvmx_helper_packet_hardware_enable(interface);
|
|
}
|
|
}
|
|
|
|
/* Finally enable PKO now that the entire path is up and running */
|
|
cvmx_pko_enable();
|
|
|
|
if ((OCTEON_IS_MODEL(OCTEON_CN31XX_PASS1) || OCTEON_IS_MODEL(OCTEON_CN30XX_PASS1)) &&
|
|
(cvmx_sysinfo_get()->board_type != CVMX_BOARD_TYPE_SIM))
|
|
__cvmx_helper_errata_fix_ipd_ptr_alignment();
|
|
return 0;
|
|
}
|
|
#ifdef CVMX_BUILD_FOR_LINUX_KERNEL
|
|
EXPORT_SYMBOL(cvmx_helper_ipd_and_packet_input_enable);
|
|
#endif
|
|
|
|
#define __CVMX_SSO_RWQ_SIZE 256
|
|
|
|
int cvmx_helper_initialize_sso(int wqe_entries)
|
|
{
|
|
int cvm_oct_sso_number_rwq_bufs;
|
|
char *mem;
|
|
int i;
|
|
cvmx_sso_cfg_t sso_cfg;
|
|
cvmx_fpa_fpfx_marks_t fpa_marks;
|
|
|
|
if (!OCTEON_IS_MODEL(OCTEON_CN68XX))
|
|
return 0;
|
|
|
|
/*
|
|
* CN68XX-P1 may reset with the wrong values, put in
|
|
* the correct values.
|
|
*/
|
|
fpa_marks.u64 = 0;
|
|
fpa_marks.s.fpf_wr = 0xa4;
|
|
fpa_marks.s.fpf_rd = 0x40;
|
|
cvmx_write_csr(CVMX_FPA_FPF8_MARKS, fpa_marks.u64);
|
|
|
|
cvm_oct_sso_number_rwq_bufs = ((wqe_entries - 1) / 26) + 1 + 48 + 8;
|
|
|
|
mem = cvmx_bootmem_alloc(__CVMX_SSO_RWQ_SIZE * cvm_oct_sso_number_rwq_bufs, CVMX_CACHE_LINE_SIZE);
|
|
if (mem == NULL) {
|
|
cvmx_dprintf("Out of memory initializing sso pool\n");
|
|
return -1;
|
|
}
|
|
/* Make sure RWI/RWO is disabled. */
|
|
sso_cfg.u64 = cvmx_read_csr(CVMX_SSO_CFG);
|
|
sso_cfg.s.rwen = 0;
|
|
cvmx_write_csr(CVMX_SSO_CFG, sso_cfg.u64);
|
|
|
|
for (i = cvm_oct_sso_number_rwq_bufs - 8; i > 0; i--) {
|
|
cvmx_sso_rwq_psh_fptr_t fptr;
|
|
|
|
for (;;) {
|
|
fptr.u64 = cvmx_read_csr(CVMX_SSO_RWQ_PSH_FPTR);
|
|
if (!fptr.s.full)
|
|
break;
|
|
cvmx_wait(1000);
|
|
}
|
|
fptr.s.fptr = cvmx_ptr_to_phys(mem) >> 7;
|
|
cvmx_write_csr(CVMX_SSO_RWQ_PSH_FPTR, fptr.u64);
|
|
mem = mem + __CVMX_SSO_RWQ_SIZE;
|
|
}
|
|
|
|
for (i = 0; i < 8; i++) {
|
|
cvmx_sso_rwq_head_ptrx_t head_ptr;
|
|
cvmx_sso_rwq_tail_ptrx_t tail_ptr;
|
|
|
|
head_ptr.u64 = 0;
|
|
tail_ptr.u64 = 0;
|
|
head_ptr.s.ptr = cvmx_ptr_to_phys(mem) >> 7;
|
|
tail_ptr.s.ptr = head_ptr.s.ptr;
|
|
cvmx_write_csr(CVMX_SSO_RWQ_HEAD_PTRX(i), head_ptr.u64);
|
|
cvmx_write_csr(CVMX_SSO_RWQ_TAIL_PTRX(i), tail_ptr.u64);
|
|
mem = mem + __CVMX_SSO_RWQ_SIZE;
|
|
}
|
|
|
|
sso_cfg.u64 = cvmx_read_csr(CVMX_SSO_CFG);
|
|
sso_cfg.s.rwen = 1;
|
|
sso_cfg.s.dwb = cvmx_helper_cfg_opt_get(CVMX_HELPER_CFG_OPT_USE_DWB);
|
|
sso_cfg.s.rwq_byp_dis = 0;
|
|
sso_cfg.s.rwio_byp_dis = 0;
|
|
cvmx_write_csr(CVMX_SSO_CFG, sso_cfg.u64);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int cvmx_helper_uninitialize_sso(void)
|
|
{
|
|
cvmx_fpa_quex_available_t queue_available;
|
|
cvmx_sso_cfg_t sso_cfg;
|
|
cvmx_sso_rwq_pop_fptr_t pop_fptr;
|
|
cvmx_sso_rwq_psh_fptr_t fptr;
|
|
cvmx_sso_fpage_cnt_t fpage_cnt;
|
|
int num_to_transfer, i;
|
|
char *mem;
|
|
|
|
if (!OCTEON_IS_MODEL(OCTEON_CN68XX))
|
|
return 0;
|
|
|
|
sso_cfg.u64 = cvmx_read_csr(CVMX_SSO_CFG);
|
|
sso_cfg.s.rwen = 0;
|
|
sso_cfg.s.rwq_byp_dis = 1;
|
|
cvmx_write_csr(CVMX_SSO_CFG, sso_cfg.u64);
|
|
cvmx_read_csr(CVMX_SSO_CFG);
|
|
queue_available.u64 = cvmx_read_csr(CVMX_FPA_QUEX_AVAILABLE(8));
|
|
|
|
/* Make CVMX_FPA_QUEX_AVAILABLE(8) % 16 == 0*/
|
|
for (num_to_transfer = (16 - queue_available.s.que_siz) % 16;
|
|
num_to_transfer > 0; num_to_transfer--) {
|
|
do {
|
|
pop_fptr.u64 = cvmx_read_csr(CVMX_SSO_RWQ_POP_FPTR);
|
|
} while (!pop_fptr.s.val);
|
|
for (;;) {
|
|
fptr.u64 = cvmx_read_csr(CVMX_SSO_RWQ_PSH_FPTR);
|
|
if (!fptr.s.full)
|
|
break;
|
|
cvmx_wait(1000);
|
|
}
|
|
fptr.s.fptr = pop_fptr.s.fptr;
|
|
cvmx_write_csr(CVMX_SSO_RWQ_PSH_FPTR, fptr.u64);
|
|
}
|
|
cvmx_read_csr(CVMX_SSO_CFG);
|
|
|
|
do {
|
|
queue_available.u64 = cvmx_read_csr(CVMX_FPA_QUEX_AVAILABLE(8));
|
|
} while (queue_available.s.que_siz % 16);
|
|
|
|
sso_cfg.s.rwen = 1;
|
|
sso_cfg.s.rwq_byp_dis = 0;
|
|
cvmx_write_csr(CVMX_SSO_CFG, sso_cfg.u64);
|
|
|
|
for (i = 0; i < 8; i++) {
|
|
cvmx_sso_rwq_head_ptrx_t head_ptr;
|
|
cvmx_sso_rwq_tail_ptrx_t tail_ptr;
|
|
|
|
head_ptr.u64 = cvmx_read_csr(CVMX_SSO_RWQ_HEAD_PTRX(i));
|
|
tail_ptr.u64 = cvmx_read_csr(CVMX_SSO_RWQ_TAIL_PTRX(i));
|
|
if (head_ptr.s.ptr != tail_ptr.s.ptr) {
|
|
cvmx_dprintf("head_ptr.s.ptr != tail_ptr.s.ptr, idx: %d\n", i);
|
|
}
|
|
|
|
mem = cvmx_phys_to_ptr(((uint64_t)head_ptr.s.ptr) << 7);
|
|
/* Leak the memory */
|
|
}
|
|
|
|
do {
|
|
do {
|
|
pop_fptr.u64 = cvmx_read_csr(CVMX_SSO_RWQ_POP_FPTR);
|
|
if (pop_fptr.s.val) {
|
|
mem = cvmx_phys_to_ptr(((uint64_t)pop_fptr.s.fptr) << 7);
|
|
/* Leak the memory */
|
|
}
|
|
} while (pop_fptr.s.val);
|
|
fpage_cnt.u64 = cvmx_read_csr(CVMX_SSO_FPAGE_CNT);
|
|
} while (fpage_cnt.s.fpage_cnt);
|
|
|
|
sso_cfg.s.rwen = 0;
|
|
sso_cfg.s.rwq_byp_dis = 0;
|
|
cvmx_write_csr(CVMX_SSO_CFG, sso_cfg.u64);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* Initialize the PIP, IPD, and PKO hardware to support
|
|
* simple priority based queues for the ethernet ports. Each
|
|
* port is configured with a number of priority queues based
|
|
* on CVMX_PKO_QUEUES_PER_PORT_* where each queue is lower
|
|
* priority than the previous.
|
|
*
|
|
* @return Zero on success, non-zero on failure
|
|
*/
|
|
int cvmx_helper_initialize_packet_io_global(void)
|
|
{
|
|
int result = 0;
|
|
int interface;
|
|
cvmx_l2c_cfg_t l2c_cfg;
|
|
cvmx_smix_en_t smix_en;
|
|
const int num_interfaces = cvmx_helper_get_number_of_interfaces();
|
|
|
|
/* CN52XX pass 1: Due to a bug in 2nd order CDR, it needs to be disabled */
|
|
if (OCTEON_IS_MODEL(OCTEON_CN52XX_PASS1_0))
|
|
__cvmx_helper_errata_qlm_disable_2nd_order_cdr(1);
|
|
|
|
/* Tell L2 to give the IOB statically higher priority compared to the
|
|
cores. This avoids conditions where IO blocks might be starved under
|
|
very high L2 loads */
|
|
if (OCTEON_IS_MODEL(OCTEON_CN6XXX) || OCTEON_IS_MODEL(OCTEON_CNF7XXX))
|
|
{
|
|
cvmx_l2c_ctl_t l2c_ctl;
|
|
l2c_ctl.u64 = cvmx_read_csr(CVMX_L2C_CTL);
|
|
l2c_ctl.s.rsp_arb_mode = 1;
|
|
l2c_ctl.s.xmc_arb_mode = 0;
|
|
cvmx_write_csr(CVMX_L2C_CTL, l2c_ctl.u64);
|
|
}
|
|
else
|
|
{
|
|
l2c_cfg.u64 = cvmx_read_csr(CVMX_L2C_CFG);
|
|
l2c_cfg.s.lrf_arb_mode = 0;
|
|
l2c_cfg.s.rfb_arb_mode = 0;
|
|
cvmx_write_csr(CVMX_L2C_CFG, l2c_cfg.u64);
|
|
}
|
|
|
|
if (cvmx_sysinfo_get()->board_type != CVMX_BOARD_TYPE_SIM)
|
|
{
|
|
int smi_inf = 1;
|
|
int i;
|
|
|
|
/* Newer chips have more than one SMI/MDIO interface */
|
|
if (OCTEON_IS_MODEL(OCTEON_CN68XX))
|
|
smi_inf = 4;
|
|
else if (!OCTEON_IS_MODEL(OCTEON_CN3XXX)
|
|
&& !OCTEON_IS_MODEL(OCTEON_CN58XX)
|
|
&& !OCTEON_IS_MODEL(OCTEON_CN50XX))
|
|
smi_inf = 2;
|
|
|
|
for (i = 0; i < smi_inf; i++)
|
|
{
|
|
/* Make sure SMI/MDIO is enabled so we can query PHYs */
|
|
smix_en.u64 = cvmx_read_csr(CVMX_SMIX_EN(i));
|
|
if (!smix_en.s.en)
|
|
{
|
|
smix_en.s.en = 1;
|
|
cvmx_write_csr(CVMX_SMIX_EN(i), smix_en.u64);
|
|
}
|
|
}
|
|
}
|
|
|
|
__cvmx_helper_cfg_init();
|
|
|
|
for (interface=0; interface<num_interfaces; interface++)
|
|
result |= cvmx_helper_interface_probe(interface);
|
|
|
|
cvmx_pko_initialize_global();
|
|
for (interface=0; interface<num_interfaces; interface++)
|
|
{
|
|
if (cvmx_helper_ports_on_interface(interface) > 0)
|
|
cvmx_dprintf("Interface %d has %d ports (%s)\n",
|
|
interface, cvmx_helper_ports_on_interface(interface),
|
|
cvmx_helper_interface_mode_to_string(cvmx_helper_interface_get_mode(interface)));
|
|
result |= __cvmx_helper_interface_setup_ipd(interface);
|
|
if (!OCTEON_IS_MODEL(OCTEON_CN68XX))
|
|
result |= __cvmx_helper_interface_setup_pko(interface);
|
|
}
|
|
|
|
result |= __cvmx_helper_global_setup_ipd();
|
|
result |= __cvmx_helper_global_setup_pko();
|
|
|
|
/* Enable any flow control and backpressure */
|
|
result |= __cvmx_helper_global_setup_backpressure();
|
|
|
|
#if CVMX_HELPER_ENABLE_IPD
|
|
result |= cvmx_helper_ipd_and_packet_input_enable();
|
|
#endif
|
|
return result;
|
|
}
|
|
#ifdef CVMX_BUILD_FOR_LINUX_KERNEL
|
|
EXPORT_SYMBOL(cvmx_helper_initialize_packet_io_global);
|
|
#endif
|
|
|
|
|
|
/**
|
|
* Does core local initialization for packet io
|
|
*
|
|
* @return Zero on success, non-zero on failure
|
|
*/
|
|
int cvmx_helper_initialize_packet_io_local(void)
|
|
{
|
|
return cvmx_pko_initialize_local();
|
|
}
|
|
|
|
/**
|
|
* wait for the pko queue to drain
|
|
*
|
|
* @param queue a valid pko queue
|
|
* @return count is the length of the queue after calling this
|
|
* function
|
|
*/
|
|
static int cvmx_helper_wait_pko_queue_drain(int queue)
|
|
{
|
|
const int timeout = 5; /* Wait up to 5 seconds for timeouts */
|
|
int count;
|
|
uint64_t start_cycle, stop_cycle;
|
|
|
|
count = cvmx_cmd_queue_length(CVMX_CMD_QUEUE_PKO(queue));
|
|
start_cycle = cvmx_get_cycle();
|
|
stop_cycle = start_cycle + cvmx_clock_get_rate(CVMX_CLOCK_CORE) * timeout;
|
|
while (count && (cvmx_get_cycle() < stop_cycle))
|
|
{
|
|
cvmx_wait(10000);
|
|
count = cvmx_cmd_queue_length(CVMX_CMD_QUEUE_PKO(queue));
|
|
}
|
|
|
|
return count;
|
|
}
|
|
|
|
struct cvmx_buffer_list {
|
|
struct cvmx_buffer_list *next;
|
|
};
|
|
|
|
/**
|
|
* Undo the initialization performed in
|
|
* cvmx_helper_initialize_packet_io_global(). After calling this routine and the
|
|
* local version on each core, packet IO for Octeon will be disabled and placed
|
|
* in the initial reset state. It will then be safe to call the initialize
|
|
* later on. Note that this routine does not empty the FPA pools. It frees all
|
|
* buffers used by the packet IO hardware to the FPA so a function emptying the
|
|
* FPA after shutdown should find all packet buffers in the FPA.
|
|
*
|
|
* @return Zero on success, negative on failure.
|
|
*/
|
|
int cvmx_helper_shutdown_packet_io_global(void)
|
|
{
|
|
const int timeout = 5; /* Wait up to 5 seconds for timeouts */
|
|
int result = 0;
|
|
int num_interfaces;
|
|
int interface;
|
|
int num_ports;
|
|
int index;
|
|
struct cvmx_buffer_list *pool0_buffers;
|
|
struct cvmx_buffer_list *pool0_buffers_tail;
|
|
cvmx_wqe_t *work;
|
|
|
|
/* Step 1: Disable all backpressure */
|
|
for (interface=0; interface<CVMX_HELPER_MAX_GMX; interface++)
|
|
if (cvmx_helper_interface_get_mode(interface) != CVMX_HELPER_INTERFACE_MODE_DISABLED)
|
|
cvmx_gmx_set_backpressure_override(interface, 0xf);
|
|
|
|
step2:
|
|
/* Step 2: Wait for the PKO queues to drain */
|
|
if (octeon_has_feature(OCTEON_FEATURE_PKND))
|
|
{
|
|
int queue, max_queue;
|
|
|
|
max_queue = __cvmx_helper_cfg_pko_max_queue();
|
|
for (queue = 0; queue < max_queue; queue++)
|
|
{
|
|
if (cvmx_helper_wait_pko_queue_drain(queue))
|
|
{
|
|
result = -1;
|
|
goto step3;
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
num_interfaces = cvmx_helper_get_number_of_interfaces();
|
|
for (interface=0; interface<num_interfaces; interface++)
|
|
{
|
|
num_ports = cvmx_helper_ports_on_interface(interface);
|
|
for (index=0; index<num_ports; index++)
|
|
{
|
|
int pko_port = cvmx_helper_get_ipd_port(interface, index);
|
|
int queue = cvmx_pko_get_base_queue(pko_port);
|
|
int max_queue = queue + cvmx_pko_get_num_queues(pko_port);
|
|
while (queue < max_queue)
|
|
{
|
|
if (cvmx_helper_wait_pko_queue_drain(queue))
|
|
{
|
|
result = -1;
|
|
goto step3;
|
|
}
|
|
queue++;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
step3:
|
|
/* Step 3: Disable TX and RX on all ports */
|
|
for (interface=0; interface<CVMX_HELPER_MAX_GMX; interface++)
|
|
{
|
|
switch (cvmx_helper_interface_get_mode(interface))
|
|
{
|
|
case CVMX_HELPER_INTERFACE_MODE_DISABLED:
|
|
case CVMX_HELPER_INTERFACE_MODE_PCIE:
|
|
/* Not a packet interface */
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_NPI:
|
|
case CVMX_HELPER_INTERFACE_MODE_SRIO:
|
|
case CVMX_HELPER_INTERFACE_MODE_ILK:
|
|
/* We don't handle the NPI/NPEI/SRIO packet engines. The caller
|
|
must know these are idle */
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_LOOP:
|
|
/* Nothing needed. Once PKO is idle, the loopback devices
|
|
must be idle */
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_SPI:
|
|
/* SPI cannot be disabled from Octeon. It is the responsibility
|
|
of the caller to make sure SPI is idle before doing
|
|
shutdown */
|
|
/* Fall through and do the same processing as RGMII/GMII */
|
|
case CVMX_HELPER_INTERFACE_MODE_GMII:
|
|
case CVMX_HELPER_INTERFACE_MODE_RGMII:
|
|
/* Disable outermost RX at the ASX block */
|
|
cvmx_write_csr(CVMX_ASXX_RX_PRT_EN(interface), 0);
|
|
num_ports = cvmx_helper_ports_on_interface(interface);
|
|
if (num_ports > 4)
|
|
num_ports = 4;
|
|
for (index=0; index<num_ports; index++)
|
|
{
|
|
cvmx_gmxx_prtx_cfg_t gmx_cfg;
|
|
gmx_cfg.u64 = cvmx_read_csr(CVMX_GMXX_PRTX_CFG(index, interface));
|
|
gmx_cfg.s.en = 0;
|
|
cvmx_write_csr(CVMX_GMXX_PRTX_CFG(index, interface), gmx_cfg.u64);
|
|
/* Poll the GMX state machine waiting for it to become idle */
|
|
cvmx_write_csr(CVMX_NPI_DBG_SELECT, interface*0x800 + index*0x100 + 0x880);
|
|
if (CVMX_WAIT_FOR_FIELD64(CVMX_DBG_DATA, cvmx_dbg_data_t, data&7, ==, 0, timeout*1000000))
|
|
{
|
|
cvmx_dprintf("GMX RX path timeout waiting for idle\n");
|
|
result = -1;
|
|
}
|
|
if (CVMX_WAIT_FOR_FIELD64(CVMX_DBG_DATA, cvmx_dbg_data_t, data&0xf, ==, 0, timeout*1000000))
|
|
{
|
|
cvmx_dprintf("GMX TX path timeout waiting for idle\n");
|
|
result = -1;
|
|
}
|
|
}
|
|
/* Disable outermost TX at the ASX block */
|
|
cvmx_write_csr(CVMX_ASXX_TX_PRT_EN(interface), 0);
|
|
/* Disable interrupts for interface */
|
|
cvmx_write_csr(CVMX_ASXX_INT_EN(interface), 0);
|
|
cvmx_write_csr(CVMX_GMXX_TX_INT_EN(interface), 0);
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_XAUI:
|
|
case CVMX_HELPER_INTERFACE_MODE_RXAUI:
|
|
case CVMX_HELPER_INTERFACE_MODE_SGMII:
|
|
case CVMX_HELPER_INTERFACE_MODE_PICMG:
|
|
num_ports = cvmx_helper_ports_on_interface(interface);
|
|
if (num_ports > 4)
|
|
num_ports = 4;
|
|
for (index=0; index<num_ports; index++)
|
|
{
|
|
cvmx_gmxx_prtx_cfg_t gmx_cfg;
|
|
gmx_cfg.u64 = cvmx_read_csr(CVMX_GMXX_PRTX_CFG(index, interface));
|
|
gmx_cfg.s.en = 0;
|
|
cvmx_write_csr(CVMX_GMXX_PRTX_CFG(index, interface), gmx_cfg.u64);
|
|
if (CVMX_WAIT_FOR_FIELD64(CVMX_GMXX_PRTX_CFG(index, interface), cvmx_gmxx_prtx_cfg_t, rx_idle, ==, 1, timeout*1000000))
|
|
{
|
|
cvmx_dprintf("GMX RX path timeout waiting for idle\n");
|
|
result = -1;
|
|
}
|
|
if (CVMX_WAIT_FOR_FIELD64(CVMX_GMXX_PRTX_CFG(index, interface), cvmx_gmxx_prtx_cfg_t, tx_idle, ==, 1, timeout*1000000))
|
|
{
|
|
cvmx_dprintf("GMX TX path timeout waiting for idle\n");
|
|
result = -1;
|
|
}
|
|
}
|
|
break;
|
|
}
|
|
}
|
|
|
|
/* Step 4: Retrieve all packets from the POW and free them */
|
|
while ((work = cvmx_pow_work_request_sync(CVMX_POW_WAIT)))
|
|
{
|
|
cvmx_helper_free_packet_data(work);
|
|
cvmx_fpa_free(work, CVMX_FPA_WQE_POOL, 0);
|
|
}
|
|
|
|
/* Step 4b: Special workaround for pass 2 errata */
|
|
if (OCTEON_IS_MODEL(OCTEON_CN38XX_PASS2))
|
|
{
|
|
cvmx_ipd_ptr_count_t ipd_cnt;
|
|
int to_add;
|
|
ipd_cnt.u64 = cvmx_read_csr(CVMX_IPD_PTR_COUNT);
|
|
to_add = (ipd_cnt.s.wqev_cnt + ipd_cnt.s.wqe_pcnt) & 0x7;
|
|
if (to_add)
|
|
{
|
|
int port = -1;
|
|
cvmx_dprintf("Aligning CN38XX pass 2 IPD counters\n");
|
|
if (cvmx_helper_interface_get_mode(0) == CVMX_HELPER_INTERFACE_MODE_RGMII)
|
|
port = 0;
|
|
else if (cvmx_helper_interface_get_mode(1) == CVMX_HELPER_INTERFACE_MODE_RGMII)
|
|
port = 16;
|
|
|
|
if (port != -1)
|
|
{
|
|
char *buffer = cvmx_fpa_alloc(CVMX_FPA_PACKET_POOL);
|
|
if (buffer)
|
|
{
|
|
int queue = cvmx_pko_get_base_queue(port);
|
|
cvmx_pko_command_word0_t pko_command;
|
|
cvmx_buf_ptr_t packet;
|
|
uint64_t start_cycle;
|
|
uint64_t stop_cycle;
|
|
|
|
/* Populate a minimal packet */
|
|
memset(buffer, 0xff, 6);
|
|
memset(buffer+6, 0, 54);
|
|
pko_command.u64 = 0;
|
|
pko_command.s.dontfree = 1;
|
|
pko_command.s.total_bytes = 60;
|
|
pko_command.s.segs = 1;
|
|
packet.u64 = 0;
|
|
packet.s.addr = cvmx_ptr_to_phys(buffer);
|
|
packet.s.size = CVMX_FPA_PACKET_POOL_SIZE;
|
|
__cvmx_helper_rgmii_configure_loopback(port, 1, 0);
|
|
while (to_add--)
|
|
{
|
|
cvmx_pko_send_packet_prepare(port, queue, CVMX_PKO_LOCK_CMD_QUEUE);
|
|
if (cvmx_pko_send_packet_finish(port, queue, pko_command, packet, CVMX_PKO_LOCK_CMD_QUEUE))
|
|
{
|
|
cvmx_dprintf("ERROR: Unable to align IPD counters (PKO failed)\n");
|
|
break;
|
|
}
|
|
}
|
|
cvmx_fpa_free(buffer, CVMX_FPA_PACKET_POOL, 0);
|
|
|
|
/* Wait for the packets to loop back */
|
|
start_cycle = cvmx_get_cycle();
|
|
stop_cycle = start_cycle + cvmx_clock_get_rate(CVMX_CLOCK_CORE) * timeout;
|
|
while (cvmx_cmd_queue_length(CVMX_CMD_QUEUE_PKO(queue)) &&
|
|
(cvmx_get_cycle() < stop_cycle))
|
|
{
|
|
cvmx_wait(1000);
|
|
}
|
|
cvmx_wait(1000);
|
|
__cvmx_helper_rgmii_configure_loopback(port, 0, 0);
|
|
if (to_add == -1)
|
|
goto step2;
|
|
}
|
|
else
|
|
cvmx_dprintf("ERROR: Unable to align IPD counters (Packet pool empty)\n");
|
|
}
|
|
else
|
|
cvmx_dprintf("ERROR: Unable to align IPD counters\n");
|
|
}
|
|
}
|
|
|
|
/* Step 5 */
|
|
cvmx_ipd_disable();
|
|
|
|
/* Step 6: Drain all prefetched buffers from IPD/PIP. Note that IPD/PIP
|
|
have not been reset yet */
|
|
__cvmx_ipd_free_ptr();
|
|
|
|
/* Step 7: Free the PKO command buffers and put PKO in reset */
|
|
cvmx_pko_shutdown();
|
|
|
|
/* Step 8: Disable MAC address filtering */
|
|
for (interface=0; interface<CVMX_HELPER_MAX_GMX; interface++)
|
|
{
|
|
switch (cvmx_helper_interface_get_mode(interface))
|
|
{
|
|
case CVMX_HELPER_INTERFACE_MODE_DISABLED:
|
|
case CVMX_HELPER_INTERFACE_MODE_PCIE:
|
|
case CVMX_HELPER_INTERFACE_MODE_SRIO:
|
|
case CVMX_HELPER_INTERFACE_MODE_ILK:
|
|
case CVMX_HELPER_INTERFACE_MODE_NPI:
|
|
case CVMX_HELPER_INTERFACE_MODE_LOOP:
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_XAUI:
|
|
case CVMX_HELPER_INTERFACE_MODE_RXAUI:
|
|
case CVMX_HELPER_INTERFACE_MODE_GMII:
|
|
case CVMX_HELPER_INTERFACE_MODE_RGMII:
|
|
case CVMX_HELPER_INTERFACE_MODE_SPI:
|
|
case CVMX_HELPER_INTERFACE_MODE_SGMII:
|
|
case CVMX_HELPER_INTERFACE_MODE_PICMG:
|
|
num_ports = cvmx_helper_ports_on_interface(interface);
|
|
if (num_ports > 4)
|
|
num_ports = 4;
|
|
for (index=0; index<num_ports; index++)
|
|
{
|
|
cvmx_write_csr(CVMX_GMXX_RXX_ADR_CTL(index, interface), 1);
|
|
cvmx_write_csr(CVMX_GMXX_RXX_ADR_CAM_EN(index, interface), 0);
|
|
cvmx_write_csr(CVMX_GMXX_RXX_ADR_CAM0(index, interface), 0);
|
|
cvmx_write_csr(CVMX_GMXX_RXX_ADR_CAM1(index, interface), 0);
|
|
cvmx_write_csr(CVMX_GMXX_RXX_ADR_CAM2(index, interface), 0);
|
|
cvmx_write_csr(CVMX_GMXX_RXX_ADR_CAM3(index, interface), 0);
|
|
cvmx_write_csr(CVMX_GMXX_RXX_ADR_CAM4(index, interface), 0);
|
|
cvmx_write_csr(CVMX_GMXX_RXX_ADR_CAM5(index, interface), 0);
|
|
}
|
|
break;
|
|
}
|
|
}
|
|
|
|
/* Step 9: Drain all FPA buffers out of pool 0 before we reset
|
|
* IPD/PIP. This is needed to keep IPD_QUE0_FREE_PAGE_CNT in
|
|
* sync. We temporarily keep the buffers in the pool0_buffers
|
|
* list.
|
|
*/
|
|
pool0_buffers = NULL;
|
|
pool0_buffers_tail = NULL;
|
|
while (1)
|
|
{
|
|
struct cvmx_buffer_list *buffer = cvmx_fpa_alloc(0);
|
|
if (buffer) {
|
|
buffer->next = NULL;
|
|
|
|
if (pool0_buffers == NULL)
|
|
pool0_buffers = buffer;
|
|
else
|
|
pool0_buffers_tail->next = buffer;
|
|
|
|
pool0_buffers_tail = buffer;
|
|
}
|
|
else
|
|
break;
|
|
}
|
|
|
|
/* Step 10: Reset IPD and PIP */
|
|
{
|
|
cvmx_ipd_ctl_status_t ipd_ctl_status;
|
|
ipd_ctl_status.u64 = cvmx_read_csr(CVMX_IPD_CTL_STATUS);
|
|
ipd_ctl_status.s.reset = 1;
|
|
cvmx_write_csr(CVMX_IPD_CTL_STATUS, ipd_ctl_status.u64);
|
|
|
|
if ((cvmx_sysinfo_get()->board_type != CVMX_BOARD_TYPE_SIM) &&
|
|
(OCTEON_IS_MODEL(OCTEON_CN3XXX) || OCTEON_IS_MODEL(OCTEON_CN5XXX)))
|
|
{
|
|
/* only try 1000 times. Normally if this works it will happen in
|
|
** the first 50 loops. */
|
|
int max_loops = 1000;
|
|
int loop = 0;
|
|
/* Per port backpressure counters can get misaligned after an
|
|
IPD reset. This code realigns them by performing repeated
|
|
resets. See IPD-13473 */
|
|
cvmx_wait(100);
|
|
if (__cvmx_helper_backpressure_is_misaligned())
|
|
{
|
|
cvmx_dprintf("Starting to align per port backpressure counters.\n");
|
|
while (__cvmx_helper_backpressure_is_misaligned() && (loop++ < max_loops))
|
|
{
|
|
cvmx_write_csr(CVMX_IPD_CTL_STATUS, ipd_ctl_status.u64);
|
|
cvmx_wait(123);
|
|
}
|
|
if (loop < max_loops)
|
|
cvmx_dprintf("Completed aligning per port backpressure counters (%d loops).\n", loop);
|
|
else
|
|
{
|
|
cvmx_dprintf("ERROR: unable to align per port backpressure counters.\n");
|
|
/* For now, don't hang.... */
|
|
}
|
|
}
|
|
}
|
|
|
|
/* PIP_SFT_RST not present in CN38XXp{1,2} */
|
|
if (!OCTEON_IS_MODEL(OCTEON_CN38XX_PASS2))
|
|
{
|
|
cvmx_pip_sft_rst_t pip_sft_rst;
|
|
pip_sft_rst.u64 = cvmx_read_csr(CVMX_PIP_SFT_RST);
|
|
pip_sft_rst.s.rst = 1;
|
|
cvmx_write_csr(CVMX_PIP_SFT_RST, pip_sft_rst.u64);
|
|
}
|
|
|
|
/* Make sure IPD has finished reset. */
|
|
if (OCTEON_IS_MODEL(OCTEON_CN6XXX))
|
|
{
|
|
if (CVMX_WAIT_FOR_FIELD64(CVMX_IPD_CTL_STATUS, cvmx_ipd_ctl_status_t, rst_done, ==, 0, 1000))
|
|
{
|
|
cvmx_dprintf("IPD reset timeout waiting for idle\n");
|
|
result = -1;
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Step 11: Restore the FPA buffers into pool 0 */
|
|
while (pool0_buffers) {
|
|
struct cvmx_buffer_list *n = pool0_buffers->next;
|
|
cvmx_fpa_free(pool0_buffers, 0, 0);
|
|
pool0_buffers = n;
|
|
}
|
|
|
|
/* Step 12: Release interface structures */
|
|
__cvmx_helper_shutdown_interfaces();
|
|
|
|
return result;
|
|
}
|
|
#ifdef CVMX_BUILD_FOR_LINUX_KERNEL
|
|
EXPORT_SYMBOL(cvmx_helper_shutdown_packet_io_global);
|
|
#endif
|
|
|
|
|
|
/**
|
|
* Does core local shutdown of packet io
|
|
*
|
|
* @return Zero on success, non-zero on failure
|
|
*/
|
|
int cvmx_helper_shutdown_packet_io_local(void)
|
|
{
|
|
/* Currently there is nothing to do per core. This may change in
|
|
the future */
|
|
return 0;
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
* Auto configure an IPD/PKO port link state and speed. This
|
|
* function basically does the equivalent of:
|
|
* cvmx_helper_link_set(ipd_port, cvmx_helper_link_get(ipd_port));
|
|
*
|
|
* @param ipd_port IPD/PKO port to auto configure
|
|
*
|
|
* @return Link state after configure
|
|
*/
|
|
cvmx_helper_link_info_t cvmx_helper_link_autoconf(int ipd_port)
|
|
{
|
|
cvmx_helper_link_info_t link_info;
|
|
int interface = cvmx_helper_get_interface_num(ipd_port);
|
|
int index = cvmx_helper_get_interface_index_num(ipd_port);
|
|
|
|
if (index >= cvmx_helper_ports_on_interface(interface))
|
|
{
|
|
link_info.u64 = 0;
|
|
return link_info;
|
|
}
|
|
|
|
link_info = cvmx_helper_link_get(ipd_port);
|
|
if (link_info.u64 == (__cvmx_helper_get_link_info(interface, index)).u64)
|
|
return link_info;
|
|
|
|
#if !defined(CVMX_BUILD_FOR_LINUX_KERNEL) && !defined(CVMX_BUILD_FOR_FREEBSD_KERNEL)
|
|
if (!link_info.s.link_up)
|
|
cvmx_error_disable_group(CVMX_ERROR_GROUP_ETHERNET, ipd_port);
|
|
#endif
|
|
|
|
/* If we fail to set the link speed, port_link_info will not change */
|
|
cvmx_helper_link_set(ipd_port, link_info);
|
|
|
|
#if !defined(CVMX_BUILD_FOR_LINUX_KERNEL) && !defined(CVMX_BUILD_FOR_FREEBSD_KERNEL)
|
|
if (link_info.s.link_up)
|
|
cvmx_error_enable_group(CVMX_ERROR_GROUP_ETHERNET, ipd_port);
|
|
#endif
|
|
|
|
return link_info;
|
|
}
|
|
#ifdef CVMX_BUILD_FOR_LINUX_KERNEL
|
|
EXPORT_SYMBOL(cvmx_helper_link_autoconf);
|
|
#endif
|
|
|
|
/**
|
|
* Return the link state of an IPD/PKO port as returned by
|
|
* auto negotiation. The result of this function may not match
|
|
* Octeon's link config if auto negotiation has changed since
|
|
* the last call to cvmx_helper_link_set().
|
|
*
|
|
* @param ipd_port IPD/PKO port to query
|
|
*
|
|
* @return Link state
|
|
*/
|
|
cvmx_helper_link_info_t cvmx_helper_link_get(int ipd_port)
|
|
{
|
|
cvmx_helper_link_info_t result;
|
|
int interface = cvmx_helper_get_interface_num(ipd_port);
|
|
int index = cvmx_helper_get_interface_index_num(ipd_port);
|
|
|
|
/* The default result will be a down link unless the code below
|
|
changes it */
|
|
result.u64 = 0;
|
|
|
|
if (index >= cvmx_helper_ports_on_interface(interface))
|
|
return result;
|
|
|
|
switch (cvmx_helper_interface_get_mode(interface))
|
|
{
|
|
case CVMX_HELPER_INTERFACE_MODE_DISABLED:
|
|
case CVMX_HELPER_INTERFACE_MODE_PCIE:
|
|
/* Network links are not supported */
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_XAUI:
|
|
case CVMX_HELPER_INTERFACE_MODE_RXAUI:
|
|
result = __cvmx_helper_xaui_link_get(ipd_port);
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_GMII:
|
|
if (index == 0)
|
|
result = __cvmx_helper_rgmii_link_get(ipd_port);
|
|
else
|
|
{
|
|
result.s.full_duplex = 1;
|
|
result.s.link_up = 1;
|
|
result.s.speed = 1000;
|
|
}
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_RGMII:
|
|
result = __cvmx_helper_rgmii_link_get(ipd_port);
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_SPI:
|
|
result = __cvmx_helper_spi_link_get(ipd_port);
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_SGMII:
|
|
case CVMX_HELPER_INTERFACE_MODE_PICMG:
|
|
result = __cvmx_helper_sgmii_link_get(ipd_port);
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_SRIO:
|
|
result = __cvmx_helper_srio_link_get(ipd_port);
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_ILK:
|
|
result = __cvmx_helper_ilk_link_get(ipd_port);
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_NPI:
|
|
case CVMX_HELPER_INTERFACE_MODE_LOOP:
|
|
/* Network links are not supported */
|
|
break;
|
|
}
|
|
return result;
|
|
}
|
|
#ifdef CVMX_BUILD_FOR_LINUX_KERNEL
|
|
EXPORT_SYMBOL(cvmx_helper_link_get);
|
|
#endif
|
|
|
|
|
|
/**
|
|
* Configure an IPD/PKO port for the specified link state. This
|
|
* function does not influence auto negotiation at the PHY level.
|
|
* The passed link state must always match the link state returned
|
|
* by cvmx_helper_link_get(). It is normally best to use
|
|
* cvmx_helper_link_autoconf() instead.
|
|
*
|
|
* @param ipd_port IPD/PKO port to configure
|
|
* @param link_info The new link state
|
|
*
|
|
* @return Zero on success, negative on failure
|
|
*/
|
|
int cvmx_helper_link_set(int ipd_port, cvmx_helper_link_info_t link_info)
|
|
{
|
|
int result = -1;
|
|
int interface = cvmx_helper_get_interface_num(ipd_port);
|
|
int index = cvmx_helper_get_interface_index_num(ipd_port);
|
|
|
|
if (index >= cvmx_helper_ports_on_interface(interface))
|
|
return -1;
|
|
|
|
switch (cvmx_helper_interface_get_mode(interface))
|
|
{
|
|
case CVMX_HELPER_INTERFACE_MODE_DISABLED:
|
|
case CVMX_HELPER_INTERFACE_MODE_PCIE:
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_XAUI:
|
|
case CVMX_HELPER_INTERFACE_MODE_RXAUI:
|
|
result = __cvmx_helper_xaui_link_set(ipd_port, link_info);
|
|
break;
|
|
/* RGMII/GMII/MII are all treated about the same. Most functions
|
|
refer to these ports as RGMII */
|
|
case CVMX_HELPER_INTERFACE_MODE_RGMII:
|
|
case CVMX_HELPER_INTERFACE_MODE_GMII:
|
|
result = __cvmx_helper_rgmii_link_set(ipd_port, link_info);
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_SPI:
|
|
result = __cvmx_helper_spi_link_set(ipd_port, link_info);
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_SGMII:
|
|
case CVMX_HELPER_INTERFACE_MODE_PICMG:
|
|
result = __cvmx_helper_sgmii_link_set(ipd_port, link_info);
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_SRIO:
|
|
result = __cvmx_helper_srio_link_set(ipd_port, link_info);
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_ILK:
|
|
result = __cvmx_helper_ilk_link_set(ipd_port, link_info);
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_NPI:
|
|
case CVMX_HELPER_INTERFACE_MODE_LOOP:
|
|
break;
|
|
}
|
|
/* Set the port_link_info here so that the link status is updated
|
|
no matter how cvmx_helper_link_set is called. We don't change
|
|
the value if link_set failed */
|
|
if (result == 0)
|
|
__cvmx_helper_set_link_info(interface, index, link_info);
|
|
return result;
|
|
}
|
|
#ifdef CVMX_BUILD_FOR_LINUX_KERNEL
|
|
EXPORT_SYMBOL(cvmx_helper_link_set);
|
|
#endif
|
|
|
|
|
|
/**
|
|
* Configure a port for internal and/or external loopback. Internal loopback
|
|
* causes packets sent by the port to be received by Octeon. External loopback
|
|
* causes packets received from the wire to sent out again.
|
|
*
|
|
* @param ipd_port IPD/PKO port to loopback.
|
|
* @param enable_internal
|
|
* Non zero if you want internal loopback
|
|
* @param enable_external
|
|
* Non zero if you want external loopback
|
|
*
|
|
* @return Zero on success, negative on failure.
|
|
*/
|
|
int cvmx_helper_configure_loopback(int ipd_port, int enable_internal, int enable_external)
|
|
{
|
|
int result = -1;
|
|
int interface = cvmx_helper_get_interface_num(ipd_port);
|
|
int index = cvmx_helper_get_interface_index_num(ipd_port);
|
|
|
|
if (index >= cvmx_helper_ports_on_interface(interface))
|
|
return -1;
|
|
|
|
switch (cvmx_helper_interface_get_mode(interface))
|
|
{
|
|
case CVMX_HELPER_INTERFACE_MODE_DISABLED:
|
|
case CVMX_HELPER_INTERFACE_MODE_PCIE:
|
|
case CVMX_HELPER_INTERFACE_MODE_SRIO:
|
|
case CVMX_HELPER_INTERFACE_MODE_ILK:
|
|
case CVMX_HELPER_INTERFACE_MODE_SPI:
|
|
case CVMX_HELPER_INTERFACE_MODE_NPI:
|
|
case CVMX_HELPER_INTERFACE_MODE_LOOP:
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_XAUI:
|
|
case CVMX_HELPER_INTERFACE_MODE_RXAUI:
|
|
result = __cvmx_helper_xaui_configure_loopback(ipd_port, enable_internal, enable_external);
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_RGMII:
|
|
case CVMX_HELPER_INTERFACE_MODE_GMII:
|
|
result = __cvmx_helper_rgmii_configure_loopback(ipd_port, enable_internal, enable_external);
|
|
break;
|
|
case CVMX_HELPER_INTERFACE_MODE_SGMII:
|
|
case CVMX_HELPER_INTERFACE_MODE_PICMG:
|
|
result = __cvmx_helper_sgmii_configure_loopback(ipd_port, enable_internal, enable_external);
|
|
break;
|
|
}
|
|
return result;
|
|
}
|
|
|
|
#endif /* CVMX_ENABLE_PKO_FUNCTIONS */
|