freebsd-skq/sys/contrib/octeon-sdk/cvmx-srio.c

1623 lines
63 KiB
C
Raw Normal View History

/***********************license start***************
* Copyright (c) 2003-2010 Cavium Inc. (support@cavium.com). All rights
* reserved.
*
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions are
* met:
*
* * Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
*
* * Redistributions in binary form must reproduce the above
* copyright notice, this list of conditions and the following
* disclaimer in the documentation and/or other materials provided
* with the distribution.
* * Neither the name of Cavium Inc. nor the names of
* its contributors may be used to endorse or promote products
* derived from this software without specific prior written
* permission.
* This Software, including technical data, may be subject to U.S. export control
* laws, including the U.S. Export Administration Act and its associated
* regulations, and may be subject to export or import regulations in other
* countries.
* TO THE MAXIMUM EXTENT PERMITTED BY LAW, THE SOFTWARE IS PROVIDED "AS IS"
* AND WITH ALL FAULTS AND CAVIUM INC. MAKES NO PROMISES, REPRESENTATIONS OR
* WARRANTIES, EITHER EXPRESS, IMPLIED, STATUTORY, OR OTHERWISE, WITH RESPECT TO
* THE SOFTWARE, INCLUDING ITS CONDITION, ITS CONFORMITY TO ANY REPRESENTATION OR
* DESCRIPTION, OR THE EXISTENCE OF ANY LATENT OR PATENT DEFECTS, AND CAVIUM
* SPECIFICALLY DISCLAIMS ALL IMPLIED (IF ANY) WARRANTIES OF TITLE,
* MERCHANTABILITY, NONINFRINGEMENT, FITNESS FOR A PARTICULAR PURPOSE, LACK OF
* VIRUSES, ACCURACY OR COMPLETENESS, QUIET ENJOYMENT, QUIET POSSESSION OR
* CORRESPONDENCE TO DESCRIPTION. THE ENTIRE RISK ARISING OUT OF USE OR
* PERFORMANCE OF THE SOFTWARE LIES WITH YOU.
***********************license end**************************************/
/**
* @file
*
* Interface to SRIO
*
* <hr>$Revision: 41586 $<hr>
*/
#ifdef CVMX_BUILD_FOR_LINUX_KERNEL
#include <asm/octeon/cvmx.h>
#include <asm/octeon/cvmx-srio.h>
#include <asm/octeon/cvmx-clock.h>
#include <asm/octeon/cvmx-atomic.h>
#ifdef CONFIG_CAVIUM_DECODE_RSL
#include <asm/octeon/cvmx-error.h>
#endif
#include <asm/octeon/cvmx-sriox-defs.h>
#include <asm/octeon/cvmx-sriomaintx-defs.h>
#include <asm/octeon/cvmx-sli-defs.h>
#include <asm/octeon/cvmx-dpi-defs.h>
#include <asm/octeon/cvmx-pexp-defs.h>
#include <asm/octeon/cvmx-helper.h>
#include <asm/octeon/cvmx-qlm.h>
#else
#include "cvmx.h"
#include "cvmx-srio.h"
#include "cvmx-clock.h"
#include "cvmx-helper.h"
#ifndef CVMX_BUILD_FOR_LINUX_HOST
#include "cvmx-atomic.h"
#if !defined(CVMX_BUILD_FOR_FREEBSD_KERNEL)
#include "cvmx-error.h"
#endif
#include "cvmx-helper-errata.h"
#endif
#include "cvmx-qlm.h"
#include "cvmx-helper.h"
#endif
#define CVMX_SRIO_CONFIG_TIMEOUT 10000 /* 10ms */
#define CVMX_SRIO_DOORBELL_TIMEOUT 10000 /* 10ms */
#define CVMX_SRIO_CONFIG_PRIORITY 0
#define ULL unsigned long long
typedef union
{
uint64_t u64;
struct
{
#ifdef __BIG_ENDIAN_BITFIELD
uint64_t upper : 2; /* Normally 2 for XKPHYS */
uint64_t reserved_49_61 : 13; /* Must be zero */
uint64_t io : 1; /* 1 for IO space access */
uint64_t did : 5; /* DID = 3 */
uint64_t subdid : 3; /* SubDID = 3-6 */
uint64_t reserved_36_39 : 4; /* Must be zero */
uint64_t se : 2; /* SubDID extender */
uint64_t reserved_32_33 : 2; /* Must be zero */
uint64_t hopcount : 8; /* Hopcount */
uint64_t address : 24; /* Mem address */
#else
uint64_t address : 24;
uint64_t hopcount : 8;
uint64_t reserved_32_33 : 2;
uint64_t se : 2;
uint64_t reserved_36_39 : 4;
uint64_t subdid : 3;
uint64_t did : 5;
uint64_t io : 1;
uint64_t reserved_49_61 : 13;
uint64_t upper : 2;
#endif
} config;
struct
{
#ifdef __BIG_ENDIAN_BITFIELD
uint64_t upper : 2; /* Normally 2 for XKPHYS */
uint64_t reserved_49_61 : 13; /* Must be zero */
uint64_t io : 1; /* 1 for IO space access */
uint64_t did : 5; /* DID = 3 */
uint64_t subdid : 3; /* SubDID = 3-6 */
uint64_t reserved_36_39 : 4; /* Must be zero */
uint64_t se : 2; /* SubDID extender */
uint64_t address : 34; /* Mem address */
#else
uint64_t address : 34;
uint64_t se : 2;
uint64_t reserved_36_39 : 4;
uint64_t subdid : 3;
uint64_t did : 5;
uint64_t io : 1;
uint64_t reserved_49_61 : 13;
uint64_t upper : 2;
#endif
} mem;
} cvmx_sli_address_t;
typedef struct
{
cvmx_srio_initialize_flags_t flags;
int32_t subidx_ref_count[16]; /* Reference count for SLI_MEM_ACCESS_SUBID[12-27]. Index=X-12 */
int32_t s2m_ref_count[16]; /* Reference count for SRIOX_S2M_TYPE[0-15]. */
} __cvmx_srio_state_t;
static CVMX_SHARED __cvmx_srio_state_t __cvmx_srio_state[4];
#ifndef CVMX_BUILD_FOR_LINUX_HOST
/**
* @INTERNAL
* Allocate a SRIOX_S2M_TYPEX register for mapping a remote SRIO
* device's address range into Octeons SLI address space. Reference
* counting is used to allow sharing of duplicate setups. The current
* implementation treats reads and writes as paired, but this could be
* changed if we have trouble running out of indexes.
*
* @param srio_port SRIO port device is on
* @param s2m SRIOX_S2M_TYPEX setup required
*
* @return Index of CSR, or negative on failure
*/
static int __cvmx_srio_alloc_s2m(int srio_port, cvmx_sriox_s2m_typex_t s2m)
{
int s2m_index;
/* Search through the S2M_TYPE registers looking for an unsed one or one
setup the way we need it */
for (s2m_index=0; s2m_index<16; s2m_index++)
{
/* Increment ref count by 2 since we count read and write
independently. We might need a more complicated search in the
future */
int ref_count = cvmx_atomic_fetch_and_add32(&__cvmx_srio_state[srio_port].s2m_ref_count[s2m_index], 2);
if (ref_count == 0)
{
/* Unused location. Write our value */
cvmx_write_csr(CVMX_SRIOX_S2M_TYPEX(s2m_index, srio_port), s2m.u64);
/* Read back to make sure the update is complete */
cvmx_read_csr(CVMX_SRIOX_S2M_TYPEX(s2m_index, srio_port));
return s2m_index;
}
else
{
/* In use, see if we can use it */
if (cvmx_read_csr(CVMX_SRIOX_S2M_TYPEX(s2m_index, srio_port)) == s2m.u64)
return s2m_index;
else
cvmx_atomic_add32(&__cvmx_srio_state[srio_port].s2m_ref_count[s2m_index], -2);
}
}
cvmx_dprintf("SRIO%d: Unable to find free SRIOX_S2M_TYPEX\n", srio_port);
return -1;
}
/**
* @INTERNAL
* Free a handle allocated by __cvmx_srio_alloc_s2m
*
* @param srio_port SRIO port
* @param index Index to free
*/
static void __cvmx_srio_free_s2m(int srio_port, int index)
{
/* Read to force pending transactions to complete */
cvmx_read_csr(CVMX_SRIOX_S2M_TYPEX(index, srio_port));
cvmx_atomic_add32(&__cvmx_srio_state[srio_port].s2m_ref_count[index], -2);
}
/**
* @INTERNAL
* Allocate a SLI SubID to map a region of memory. Reference
* counting is used to allow sharing of duplicate setups.
*
* @param subid SLI_MEM_ACCESS_SUBIDX we need an index for
*
* @return Index of CSR, or negative on failure
*/
static int __cvmx_srio_alloc_subid(cvmx_sli_mem_access_subidx_t subid)
{
int mem_index;
/* Search through the mem access subid registers looking for an unsed one
or one setup the way we need it. PCIe uses the low indexes, so search
backwards */
for (mem_index=27; mem_index>=12; mem_index--)
{
int ref_count = cvmx_atomic_fetch_and_add32(&__cvmx_srio_state[0].subidx_ref_count[mem_index-12], 1);
if (ref_count == 0)
{
/* Unused location. Write our value */
cvmx_write_csr(CVMX_PEXP_SLI_MEM_ACCESS_SUBIDX(mem_index), subid.u64);
/* Read back the value to make sure the update is complete */
cvmx_read_csr(CVMX_PEXP_SLI_MEM_ACCESS_SUBIDX(mem_index));
return mem_index;
}
else
{
/* In use, see if we can use it */
if (cvmx_read_csr(CVMX_PEXP_SLI_MEM_ACCESS_SUBIDX(mem_index)) == subid.u64)
return mem_index;
else
cvmx_atomic_add32(&__cvmx_srio_state[0].subidx_ref_count[mem_index-12], -1);
}
}
cvmx_dprintf("SRIO: Unable to find free SLI_MEM_ACCESS_SUBIDX\n");
return -1;
}
/**
* @INTERNAL
* Free a handle allocated by __cvmx_srio_alloc_subid
*
* @param index Index to free
*/
static void __cvmx_srio_free_subid(int index)
{
/* Read to force pending transactions to complete */
cvmx_read_csr(CVMX_PEXP_SLI_MEM_ACCESS_SUBIDX(index));
cvmx_atomic_add32(&__cvmx_srio_state[0].subidx_ref_count[index-12], -1);
}
#endif
/**
* @INTERNAL
* Read 32bits from a local port
*
* @param srio_port SRIO port the device is on
* @param offset Offset in config space. This must be a multiple of 32 bits.
* @param result Result of the read. This will be unmodified on failure.
*
* @return Zero on success, negative on failure.
*/
static int __cvmx_srio_local_read32(int srio_port, uint32_t offset, uint32_t *result)
{
cvmx_sriox_maint_op_t maint_op;
cvmx_sriox_maint_rd_data_t maint_rd_data;
maint_op.u64 = 0;
maint_op.s.op = 0; /* Read */
maint_op.s.addr = offset;
/* Make sure SRIO isn't already busy */
if (CVMX_WAIT_FOR_FIELD64(CVMX_SRIOX_MAINT_OP(srio_port), cvmx_sriox_maint_op_t, pending, ==, 0, CVMX_SRIO_CONFIG_TIMEOUT))
{
cvmx_dprintf("SRIO%d: Pending bit stuck before config read\n", srio_port);
return -1;
}
/* Issue the read to the hardware */
cvmx_write_csr(CVMX_SRIOX_MAINT_OP(srio_port), maint_op.u64);
/* Wait for the hardware to complete the operation */
if (CVMX_WAIT_FOR_FIELD64(CVMX_SRIOX_MAINT_OP(srio_port), cvmx_sriox_maint_op_t, pending, ==, 0, CVMX_SRIO_CONFIG_TIMEOUT))
{
cvmx_dprintf("SRIO%d: Config read timeout\n", srio_port);
return -1;
}
/* Display and error and return if the operation failed to issue */
maint_op.u64 = cvmx_read_csr(CVMX_SRIOX_MAINT_OP(srio_port));
if (maint_op.s.fail)
{
cvmx_dprintf("SRIO%d: Config read addressing error (offset=0x%x)\n", srio_port, (unsigned int)offset);
return -1;
}
/* Wait for the read data to become valid */
if (CVMX_WAIT_FOR_FIELD64(CVMX_SRIOX_MAINT_RD_DATA(srio_port), cvmx_sriox_maint_rd_data_t, valid, ==, 1, CVMX_SRIO_CONFIG_TIMEOUT))
{
cvmx_dprintf("SRIO%d: Config read data timeout\n", srio_port);
return -1;
}
/* Get the read data */
maint_rd_data.u64 = cvmx_read_csr(CVMX_SRIOX_MAINT_RD_DATA(srio_port));
*result = maint_rd_data.s.rd_data;
return 0;
}
/**
* @INTERNAL
* Write 32bits to a local port
* @param srio_port SRIO port the device is on
* @param offset Offset in config space. This must be a multiple of 32 bits.
* @param data Data to write.
*
* @return Zero on success, negative on failure.
*/
static int __cvmx_srio_local_write32(int srio_port, uint32_t offset, uint32_t data)
{
cvmx_sriox_maint_op_t maint_op;
maint_op.u64 = 0;
maint_op.s.wr_data = data;
maint_op.s.op = 1; /* Write */
maint_op.s.addr = offset;
/* Make sure SRIO isn't already busy */
if (CVMX_WAIT_FOR_FIELD64(CVMX_SRIOX_MAINT_OP(srio_port), cvmx_sriox_maint_op_t, pending, ==, 0, CVMX_SRIO_CONFIG_TIMEOUT))
{
cvmx_dprintf("SRIO%d: Pending bit stuck before config write\n", srio_port);
return -1;
}
/* Issue the write to the hardware */
cvmx_write_csr(CVMX_SRIOX_MAINT_OP(srio_port), maint_op.u64);
/* Wait for the hardware to complete the operation */
if (CVMX_WAIT_FOR_FIELD64(CVMX_SRIOX_MAINT_OP(srio_port), cvmx_sriox_maint_op_t, pending, ==, 0, CVMX_SRIO_CONFIG_TIMEOUT))
{
cvmx_dprintf("SRIO%d: Config write timeout\n", srio_port);
return -1;
}
/* Display and error and return if the operation failed to issue */
maint_op.u64 = cvmx_read_csr(CVMX_SRIOX_MAINT_OP(srio_port));
if (maint_op.s.fail)
{
cvmx_dprintf("SRIO%d: Config write addressing error (offset=0x%x)\n", srio_port, (unsigned int)offset);
return -1;
}
return 0;
}
/**
* Reset SRIO to link partner
*
* @param srio_port SRIO port to initialize
*
* @return Zero on success
*/
int cvmx_srio_link_rst(int srio_port)
{
cvmx_sriomaintx_port_0_link_resp_t link_resp;
if (OCTEON_IS_MODEL(OCTEON_CN63XX_PASS1_X))
return -1;
/* Generate a symbol reset to the link partner by writing 0x3. */
if (cvmx_srio_config_write32(srio_port, 0, -1, 0, 0,
CVMX_SRIOMAINTX_PORT_0_LINK_REQ(srio_port), 3))
return -1;
if (cvmx_srio_config_read32(srio_port, 0, -1, 0, 0,
CVMX_SRIOMAINTX_PORT_0_LINK_RESP(srio_port), &link_resp.u32))
return -1;
/* Poll until link partner has received the reset. */
while (link_resp.s.valid == 0)
{
//cvmx_dprintf("Waiting for Link Response\n");
if (cvmx_srio_config_read32(srio_port, 0, -1, 0, 0,
CVMX_SRIOMAINTX_PORT_0_LINK_RESP(srio_port), &link_resp.u32))
return -1;
}
/* Valid response, Asserting MAC reset */
cvmx_write_csr(CVMX_CIU_SOFT_PRST, 0x1);
cvmx_wait(10);
/* De-asserting MAC Reset */
cvmx_write_csr(CVMX_CIU_SOFT_PRST, 0x0);
return 0;
}
/**
* Initialize a SRIO port for use.
*
* @param srio_port SRIO port to initialize
* @param flags Optional flags
*
* @return Zero on success
*/
int cvmx_srio_initialize(int srio_port, cvmx_srio_initialize_flags_t flags)
{
cvmx_sriomaintx_port_lt_ctl_t port_lt_ctl;
cvmx_sriomaintx_port_rt_ctl_t port_rt_ctl;
cvmx_sriomaintx_port_0_ctl_t port_0_ctl;
cvmx_sriomaintx_core_enables_t core_enables;
cvmx_sriomaintx_port_gen_ctl_t port_gen_ctl;
cvmx_sriox_status_reg_t sriox_status_reg;
cvmx_mio_rst_ctlx_t mio_rst_ctl;
cvmx_sriox_imsg_vport_thr_t sriox_imsg_vport_thr;
cvmx_dpi_sli_prtx_cfg_t prt_cfg;
cvmx_sli_s2m_portx_ctl_t sli_s2m_portx_ctl;
cvmx_sli_mem_access_ctl_t sli_mem_access_ctl;
cvmx_sriomaintx_port_0_ctl2_t port_0_ctl2;
sriox_status_reg.u64 = cvmx_read_csr(CVMX_SRIOX_STATUS_REG(srio_port));
if (OCTEON_IS_MODEL(OCTEON_CN66XX))
{
/* All SRIO ports are connected to QLM0 */
int status = cvmx_qlm_get_status(0);
if (status < 4 || status > 6)
{
cvmx_dprintf("SRIO%d: Initialization called on a port not in SRIO mode\n", srio_port);
return -1;
}
}
else if (!sriox_status_reg.s.srio)
{
cvmx_dprintf("SRIO%d: Initialization called on a port not in SRIO mode\n", srio_port);
return -1;
}
__cvmx_srio_state[srio_port].flags = flags;
/* CN63XX Pass 1.0 errata G-14395 requires the QLM De-emphasis be
programmed */
if (OCTEON_IS_MODEL(OCTEON_CN63XX_PASS1_0))
{
if (srio_port)
{
cvmx_ciu_qlm1_t ciu_qlm;
ciu_qlm.u64 = cvmx_read_csr(CVMX_CIU_QLM1);
ciu_qlm.s.txbypass = 1;
ciu_qlm.s.txdeemph = 5;
ciu_qlm.s.txmargin = 0x17;
cvmx_write_csr(CVMX_CIU_QLM1, ciu_qlm.u64);
}
else
{
cvmx_ciu_qlm0_t ciu_qlm;
ciu_qlm.u64 = cvmx_read_csr(CVMX_CIU_QLM0);
ciu_qlm.s.txbypass = 1;
ciu_qlm.s.txdeemph = 5;
ciu_qlm.s.txmargin = 0x17;
cvmx_write_csr(CVMX_CIU_QLM0, ciu_qlm.u64);
}
}
/* Don't receive or drive reset signals for the SRIO QLM */
if (OCTEON_IS_MODEL(OCTEON_CN66XX))
{
/* The reset signals are available only for srio_port == 0. */
if (srio_port == 0 || (OCTEON_IS_MODEL(OCTEON_CN66XX_PASS1_2) && srio_port == 1))
{
cvmx_mio_rst_cntlx_t mio_rst_cntl;
mio_rst_cntl.u64 = cvmx_read_csr(CVMX_MIO_RST_CNTLX(srio_port));
mio_rst_cntl.s.rst_drv = 0;
mio_rst_cntl.s.rst_rcv = 0;
mio_rst_cntl.s.rst_chip = 0;
cvmx_write_csr(CVMX_MIO_RST_CNTLX(srio_port), mio_rst_cntl.u64);
}
/* MIO_RST_CNTL2<prtmode> is initialized to 0 on cold reset */
mio_rst_ctl.u64 = cvmx_read_csr(CVMX_MIO_RST_CNTLX(srio_port));
}
else
{
mio_rst_ctl.u64 = cvmx_read_csr(CVMX_MIO_RST_CTLX(srio_port));
mio_rst_ctl.s.rst_drv = 0;
mio_rst_ctl.s.rst_rcv = 0;
mio_rst_ctl.s.rst_chip = 0;
cvmx_write_csr(CVMX_MIO_RST_CTLX(srio_port), mio_rst_ctl.u64);
mio_rst_ctl.u64 = cvmx_read_csr(CVMX_MIO_RST_CTLX(srio_port));
}
cvmx_dprintf("SRIO%d: Port in %s mode\n", srio_port,
(mio_rst_ctl.s.prtmode) ? "host" : "endpoint");
/* Bring the port out of reset if necessary */
switch (srio_port)
{
case 0:
{
cvmx_ciu_soft_prst_t prst;
prst.u64 = cvmx_read_csr(CVMX_CIU_SOFT_PRST);
if (prst.s.soft_prst)
{
prst.s.soft_prst = 0;
cvmx_write_csr(CVMX_CIU_SOFT_PRST, prst.u64);
/* Wait up to 250ms for the port to come out of reset */
if (CVMX_WAIT_FOR_FIELD64(CVMX_SRIOX_STATUS_REG(srio_port), cvmx_sriox_status_reg_t, access, ==, 1, 250000))
return -1;
}
break;
}
case 1:
{
cvmx_ciu_soft_prst1_t prst;
prst.u64 = cvmx_read_csr(CVMX_CIU_SOFT_PRST1);
if (prst.s.soft_prst)
{
prst.s.soft_prst = 0;
cvmx_write_csr(CVMX_CIU_SOFT_PRST1, prst.u64);
/* Wait up to 250ms for the port to come out of reset */
if (CVMX_WAIT_FOR_FIELD64(CVMX_SRIOX_STATUS_REG(srio_port), cvmx_sriox_status_reg_t, access, ==, 1, 250000))
return -1;
}
break;
}
case 2:
{
cvmx_ciu_soft_prst2_t prst;
prst.u64 = cvmx_read_csr(CVMX_CIU_SOFT_PRST2);
if (prst.s.soft_prst)
{
prst.s.soft_prst = 0;
cvmx_write_csr(CVMX_CIU_SOFT_PRST2, prst.u64);
/* Wait up to 250ms for the port to come out of reset */
if (CVMX_WAIT_FOR_FIELD64(CVMX_SRIOX_STATUS_REG(srio_port), cvmx_sriox_status_reg_t, access, ==, 1, 250000))
return -1;
}
break;
}
}
/* Disable the link while we make changes */
if (__cvmx_srio_local_read32(srio_port, CVMX_SRIOMAINTX_PORT_0_CTL(srio_port), &port_0_ctl.u32))
return -1;
port_0_ctl.s.o_enable = 0;
port_0_ctl.s.i_enable = 0;
port_0_ctl.s.prt_lock = 1;
port_0_ctl.s.disable = 0;
if (__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_PORT_0_CTL(srio_port), port_0_ctl.u32))
return -1;
/* CN63XX Pass 2.0 and 2.1 errata G-15273 requires the QLM De-emphasis be
programmed when using a 156.25Mhz ref clock */
if (OCTEON_IS_MODEL(OCTEON_CN63XX_PASS2_0) ||
OCTEON_IS_MODEL(OCTEON_CN63XX_PASS2_1))
{
cvmx_mio_rst_boot_t mio_rst_boot;
cvmx_sriomaintx_lane_x_status_0_t lane_x_status;
/* Read the QLM config and speed pins */
mio_rst_boot.u64 = cvmx_read_csr(CVMX_MIO_RST_BOOT);
if (__cvmx_srio_local_read32(srio_port, CVMX_SRIOMAINTX_LANE_X_STATUS_0(0, srio_port), &lane_x_status.u32))
return -1;
if (srio_port)
{
cvmx_ciu_qlm1_t ciu_qlm;
ciu_qlm.u64 = cvmx_read_csr(CVMX_CIU_QLM1);
switch (mio_rst_boot.cn63xx.qlm1_spd)
{
case 0x4: /* 1.25 Gbaud, 156.25MHz */
ciu_qlm.s.txbypass = 1;
ciu_qlm.s.txdeemph = 0x0;
ciu_qlm.s.txmargin = (lane_x_status.s.rx_type == 0) ? 0x11 : 0x1c; /* short or med/long */
break;
case 0xb: /* 5.0 Gbaud, 156.25MHz */
ciu_qlm.s.txbypass = 1;
ciu_qlm.s.txdeemph = (lane_x_status.s.rx_type == 0) ? 0xa : 0xf; /* short or med/long */
ciu_qlm.s.txmargin = (lane_x_status.s.rx_type == 0) ? 0xf : 0x1a; /* short or med/long */
break;
}
cvmx_write_csr(CVMX_CIU_QLM1, ciu_qlm.u64);
}
else
{
cvmx_ciu_qlm0_t ciu_qlm;
ciu_qlm.u64 = cvmx_read_csr(CVMX_CIU_QLM0);
switch (mio_rst_boot.cn63xx.qlm0_spd)
{
case 0x4: /* 1.25 Gbaud, 156.25MHz */
ciu_qlm.s.txbypass = 1;
ciu_qlm.s.txdeemph = 0x0;
ciu_qlm.s.txmargin = (lane_x_status.s.rx_type == 0) ? 0x11 : 0x1c; /* short or med/long */
break;
case 0xb: /* 5.0 Gbaud, 156.25MHz */
ciu_qlm.s.txbypass = 1;
ciu_qlm.s.txdeemph = (lane_x_status.s.rx_type == 0) ? 0xa : 0xf; /* short or med/long */
ciu_qlm.s.txmargin = (lane_x_status.s.rx_type == 0) ? 0xf : 0x1a; /* short or med/long */
break;
}
cvmx_write_csr(CVMX_CIU_QLM0, ciu_qlm.u64);
}
}
/* Errata SRIO-14485: Link speed is reported incorrectly in CN63XX
pass 1.x */
if (OCTEON_IS_MODEL(OCTEON_CN63XX_PASS1_X))
{
cvmx_sriomaintx_port_0_ctl2_t port_0_ctl2;
if (__cvmx_srio_local_read32(srio_port, CVMX_SRIOMAINTX_PORT_0_CTL2(srio_port), &port_0_ctl2.u32))
return -1;
if (port_0_ctl2.s.enb_500g)
{
port_0_ctl2.u32 = 0;
port_0_ctl2.s.enb_625g = 1;
}
else if (port_0_ctl2.s.enb_312g)
{
port_0_ctl2.u32 = 0;
port_0_ctl2.s.enb_500g = 1;
}
else if (port_0_ctl2.s.enb_250g)
{
port_0_ctl2.u32 = 0;
port_0_ctl2.s.enb_312g = 1;
}
else if (port_0_ctl2.s.enb_125g)
{
port_0_ctl2.u32 = 0;
port_0_ctl2.s.enb_250g = 1;
}
else
{
port_0_ctl2.u32 = 0;
port_0_ctl2.s.enb_125g = 1;
}
if (__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_PORT_0_CTL2(srio_port), port_0_ctl2.u32))
return -1;
}
/* Errata SRIO-15351: Turn off SRIOMAINTX_MAC_CTRL[TYPE_MRG] as it may
cause packet ACCEPT to be lost */
if (OCTEON_IS_MODEL(OCTEON_CN63XX_PASS2_0) || OCTEON_IS_MODEL(OCTEON_CN63XX_PASS2_1))
{
cvmx_sriomaintx_mac_ctrl_t mac_ctrl;
if (__cvmx_srio_local_read32(srio_port, CVMX_SRIOMAINTX_MAC_CTRL(srio_port), &mac_ctrl.u32))
return -1;
mac_ctrl.s.type_mrg = 0;
if (__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_MAC_CTRL(srio_port), mac_ctrl.u32))
return -1;
}
/* Set the link layer timeout to 1ms. The default is too high and causes
core bus errors */
if (__cvmx_srio_local_read32(srio_port, CVMX_SRIOMAINTX_PORT_LT_CTL(srio_port), &port_lt_ctl.u32))
return -1;
port_lt_ctl.s.timeout = 1000000 / 200; /* 1ms = 1000000ns / 200ns */
if (__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_PORT_LT_CTL(srio_port), port_lt_ctl.u32))
return -1;
/* Set the logical layer timeout to 100ms. The default is too high and causes
core bus errors */
if (__cvmx_srio_local_read32(srio_port, CVMX_SRIOMAINTX_PORT_RT_CTL(srio_port), &port_rt_ctl.u32))
return -1;
port_rt_ctl.s.timeout = 100000000 / 200; /* 100ms = 100000000ns / 200ns */
if (__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_PORT_RT_CTL(srio_port), port_rt_ctl.u32))
return -1;
/* Allow memory and doorbells. Messaging is enabled later */
if (__cvmx_srio_local_read32(srio_port, CVMX_SRIOMAINTX_CORE_ENABLES(srio_port), &core_enables.u32))
return -1;
core_enables.s.doorbell = 1;
core_enables.s.memory = 1;
if (__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_CORE_ENABLES(srio_port), core_enables.u32))
return -1;
/* Allow us to master transactions */
if (__cvmx_srio_local_read32(srio_port, CVMX_SRIOMAINTX_PORT_GEN_CTL(srio_port), &port_gen_ctl.u32))
return -1;
port_gen_ctl.s.menable = 1;
if (__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_PORT_GEN_CTL(srio_port), port_gen_ctl.u32))
return -1;
/* Set the MRRS and MPS for optimal SRIO performance */
prt_cfg.u64 = cvmx_read_csr(CVMX_DPI_SLI_PRTX_CFG(srio_port));
prt_cfg.s.mps = 1;
prt_cfg.s.mrrs = 1;
prt_cfg.s.molr = 32;
if (OCTEON_IS_MODEL(OCTEON_CN66XX))
prt_cfg.s.molr = ((prt_cfg.s.qlm_cfg == 1 || prt_cfg.s.qlm_cfg == 3) ? 8
: (prt_cfg.s.qlm_cfg == 4 || prt_cfg.s.qlm_cfg == 6) ? 16
: 32);
cvmx_write_csr(CVMX_DPI_SLI_PRTX_CFG(srio_port), prt_cfg.u64);
sli_s2m_portx_ctl.u64 = cvmx_read_csr(CVMX_PEXP_SLI_S2M_PORTX_CTL(srio_port));
sli_s2m_portx_ctl.s.mrrs = 1;
cvmx_write_csr(CVMX_PEXP_SLI_S2M_PORTX_CTL(srio_port), sli_s2m_portx_ctl.u64);
/* Setup RX messaging thresholds */
sriox_imsg_vport_thr.u64 = cvmx_read_csr(CVMX_SRIOX_IMSG_VPORT_THR(srio_port));
if (OCTEON_IS_MODEL(OCTEON_CN66XX))
sriox_imsg_vport_thr.s.max_tot = ((prt_cfg.s.qlm_cfg == 1 || prt_cfg.s.qlm_cfg == 3) ? 44 : 46);
else
sriox_imsg_vport_thr.s.max_tot = 48;
sriox_imsg_vport_thr.s.max_s1 = 24;
sriox_imsg_vport_thr.s.max_s0 = 24;
sriox_imsg_vport_thr.s.sp_vport = 1;
sriox_imsg_vport_thr.s.buf_thr = 4;
sriox_imsg_vport_thr.s.max_p1 = 12;
sriox_imsg_vport_thr.s.max_p0 = 12;
cvmx_write_csr(CVMX_SRIOX_IMSG_VPORT_THR(srio_port), sriox_imsg_vport_thr.u64);
/* Setup RX messaging thresholds for other virtual ports. */
if (OCTEON_IS_MODEL(OCTEON_CN66XX))
{
cvmx_sriox_imsg_vport_thr2_t sriox_imsg_vport_thr2;
sriox_imsg_vport_thr2.u64 = cvmx_read_csr(CVMX_SRIOX_IMSG_VPORT_THR2(srio_port));
sriox_imsg_vport_thr2.s.max_s2 = 24;
sriox_imsg_vport_thr2.s.max_s3 = 24;
cvmx_write_csr(CVMX_SRIOX_IMSG_VPORT_THR2(srio_port), sriox_imsg_vport_thr2.u64);
}
/* Errata SRIO-X: SRIO error behavior may not be optimal in CN63XX pass 1.x */
if (OCTEON_IS_MODEL(OCTEON_CN63XX_PASS1_X))
{
cvmx_sriox_tx_ctrl_t sriox_tx_ctrl;
sriox_tx_ctrl.u64 = cvmx_read_csr(CVMX_SRIOX_TX_CTRL(srio_port));
sriox_tx_ctrl.s.tag_th2 = 2;
sriox_tx_ctrl.s.tag_th1 = 3;
sriox_tx_ctrl.s.tag_th0 = 4;
cvmx_write_csr(CVMX_SRIOX_TX_CTRL(srio_port), sriox_tx_ctrl.u64);
}
/* Errata SLI-15954: SLI relaxed order issues */
if (OCTEON_IS_MODEL(OCTEON_CN66XX_PASS1_X))
{
cvmx_sli_ctl_portx_t sli_ctl_portx;
sli_ctl_portx.u64 = cvmx_read_csr(CVMX_PEXP_SLI_CTL_PORTX(srio_port));
sli_ctl_portx.s.ptlp_ro = 1; /* Set to same value for all MACs. */
sli_ctl_portx.s.ctlp_ro = 1; /* Set to same value for all MACs. */
sli_ctl_portx.s.wait_com = 0; /* So that no inbound stores wait for a commit */
sli_ctl_portx.s.waitl_com = 0;
cvmx_write_csr(CVMX_PEXP_SLI_CTL_PORTX(srio_port), sli_ctl_portx.u64);
}
if (!OCTEON_IS_MODEL(OCTEON_CN63XX_PASS1_X))
{
/* Clear the ACK state */
if (__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_PORT_0_LOCAL_ACKID(srio_port), 0))
return -1;
}
/* Bring the link down, then up, by writing to the SRIO port's
PORT_0_CTL2 CSR. */
if (__cvmx_srio_local_read32(srio_port, CVMX_SRIOMAINTX_PORT_0_CTL2(srio_port), &port_0_ctl2.u32))
return -1;
if (__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_PORT_0_CTL2(srio_port), port_0_ctl2.u32))
return -1;
/* Clear any pending interrupts */
cvmx_write_csr(CVMX_SRIOX_INT_REG(srio_port), cvmx_read_csr(CVMX_SRIOX_INT_REG(srio_port)));
/* Enable error reporting */
#if (!defined(CVMX_BUILD_FOR_LINUX_HOST) && !defined(CVMX_BUILD_FOR_LINUX_KERNEL) && !defined(CVMX_BUILD_FOR_FREEBSD_KERNEL)) || defined(CONFIG_CAVIUM_DECODE_RSL)
cvmx_error_enable_group(CVMX_ERROR_GROUP_SRIO, srio_port);
#endif
/* Finally enable the link */
if (__cvmx_srio_local_read32(srio_port, CVMX_SRIOMAINTX_PORT_0_CTL(srio_port), &port_0_ctl.u32))
return -1;
port_0_ctl.s.o_enable = 1;
port_0_ctl.s.i_enable = 1;
port_0_ctl.s.disable = 0;
port_0_ctl.s.prt_lock = 0;
if (__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_PORT_0_CTL(srio_port), port_0_ctl.u32))
return -1;
/* Store merge control (SLI_MEM_ACCESS_CTL[TIMER,MAX_WORD]) */
sli_mem_access_ctl.u64 = cvmx_read_csr(CVMX_PEXP_SLI_MEM_ACCESS_CTL);
sli_mem_access_ctl.s.max_word = 0; /* Allow 16 words to combine */
sli_mem_access_ctl.s.timer = 127; /* Wait up to 127 cycles for more data */
cvmx_write_csr(CVMX_PEXP_SLI_MEM_ACCESS_CTL, sli_mem_access_ctl.u64);
/* FIXME: Disable sending a link request when the SRIO link is
brought up. For unknown reasons this code causes issues with some SRIO
devices. As we currently don't support hotplug in software, this code
should never be needed. Without link down/up events, the ACKs should
start off and stay synchronized */
#if 0
/* Ask for a link and align our ACK state. CN63XXp1 didn't support this */
if (!OCTEON_IS_MODEL(OCTEON_CN63XX_PASS1_X))
{
uint64_t stop_cycle;
cvmx_sriomaintx_port_0_err_stat_t sriomaintx_port_0_err_stat;
/* Clear the SLI_CTL_PORTX[DIS_PORT[ bit to re-enable traffic-flow
to the SRIO MACs. */
cvmx_write_csr(CVMX_PEXP_SLI_CTL_PORTX(srio_port), cvmx_read_csr(CVMX_PEXP_SLI_CTL_PORTX(srio_port)));
/* Wait a little to see if the link comes up */
stop_cycle = cvmx_clock_get_rate(CVMX_CLOCK_CORE)/4 + cvmx_clock_get_count(CVMX_CLOCK_CORE);
do
{
/* Read the port link status */
if (__cvmx_srio_local_read32(srio_port, CVMX_SRIOMAINTX_PORT_0_ERR_STAT(srio_port), &sriomaintx_port_0_err_stat.u32))
return -1;
} while (!sriomaintx_port_0_err_stat.s.pt_ok && (cvmx_clock_get_count(CVMX_CLOCK_CORE) < stop_cycle));
/* Send link request if link is up */
if (sriomaintx_port_0_err_stat.s.pt_ok)
{
cvmx_sriomaintx_port_0_link_req_t link_req;
cvmx_sriomaintx_port_0_link_resp_t link_resp;
link_req.u32 = 0;
link_req.s.cmd = 4;
/* Send the request */
if (__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_PORT_0_LINK_REQ(srio_port), link_req.u32))
return -1;
/* Wait for the response */
stop_cycle = cvmx_clock_get_rate(CVMX_CLOCK_CORE)/8 + cvmx_clock_get_count(CVMX_CLOCK_CORE);
do
{
if (__cvmx_srio_local_read32(srio_port, CVMX_SRIOMAINTX_PORT_0_LINK_RESP(srio_port), &link_resp.u32))
return -1;
} while (!link_resp.s.valid && (cvmx_clock_get_count(CVMX_CLOCK_CORE) < stop_cycle));
/* Set our ACK state if we got a response */
if (link_resp.s.valid)
{
cvmx_sriomaintx_port_0_local_ackid_t local_ackid;
local_ackid.u32 = 0;
local_ackid.s.i_ackid = 0;
local_ackid.s.e_ackid = link_resp.s.ackid;
local_ackid.s.o_ackid = link_resp.s.ackid;
if (__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_PORT_0_LOCAL_ACKID(srio_port), local_ackid.u32))
return -1;
}
else
return -1;
}
}
#endif
return 0;
}
/**
* Read 32bits from a Device's config space
*
* @param srio_port SRIO port the device is on
* @param srcid_index
* Which SRIO source ID to use. 0 = Primary, 1 = Secondary
* @param destid RapidIO device ID, or -1 for the local Octeon.
* @param is16bit Non zero if the transactions should use 16bit device IDs. Zero
* if transactions should use 8bit device IDs.
* @param hopcount Number of hops to the remote device. Use 0 for the local Octeon.
* @param offset Offset in config space. This must be a multiple of 32 bits.
* @param result Result of the read. This will be unmodified on failure.
*
* @return Zero on success, negative on failure.
*/
int cvmx_srio_config_read32(int srio_port, int srcid_index, int destid,
int is16bit, uint8_t hopcount, uint32_t offset,
uint32_t *result)
{
if (destid == -1)
{
int status = __cvmx_srio_local_read32(srio_port, offset, result);
if ((status == 0) && (__cvmx_srio_state[srio_port].flags & CVMX_SRIO_INITIALIZE_DEBUG))
cvmx_dprintf("SRIO%d: Local read [0x%06x] <= 0x%08x\n", srio_port, (unsigned int)offset, (unsigned int)*result);
return status;
}
else
{
if (OCTEON_IS_MODEL(OCTEON_CN63XX_PASS1_X))
{
int return_code;
uint32_t pkt = 0;
uint32_t sourceid;
uint64_t stop_cycle;
char rx_buffer[64];
/* Tell the user */
if (__cvmx_srio_state[srio_port].flags & CVMX_SRIO_INITIALIZE_DEBUG)
cvmx_dprintf("SRIO%d: Remote read [id=0x%04x hop=%3d offset=0x%06x] <= ", srio_port, destid, hopcount, (unsigned int)offset);
/* Read the proper source ID */
if (srcid_index)
__cvmx_srio_local_read32(srio_port, CVMX_SRIOMAINTX_SEC_DEV_ID(srio_port), &sourceid);
else
__cvmx_srio_local_read32(srio_port, CVMX_SRIOMAINTX_PRI_DEV_ID(srio_port), &sourceid);
if (is16bit)
{
/* Use the 16bit source ID */
sourceid &= 0xffff;
/* MAINT Reads are 11 bytes */
__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_IR_SP_TX_CTRL(srio_port), 11<<16);
pkt |= CVMX_SRIO_CONFIG_PRIORITY << 30; /* priority [31:30] */
pkt |= 1 << 28; /* tt [29:28] */
pkt |= 0x8 << 24; /* ftype [27:24] */
pkt |= destid << 8; /* destID [23:8] */
pkt |= sourceid >> 8; /* sourceID [7:0] */
__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_IR_SP_TX_DATA(srio_port), pkt);
pkt = 0;
pkt |= sourceid << 24; /* sourceID [31:24] */
pkt |= 0 << 20; /* transaction [23:20] */
pkt |= 8 << 16; /* rdsize [19:16] */
pkt |= 0xc0 << 8; /* srcTID [15:8] */
pkt |= hopcount; /* hopcount [7:0] */
__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_IR_SP_TX_DATA(srio_port), pkt);
pkt = 0;
pkt |= offset << 8; /* offset [31:11, wdptr[10], reserved[9:8] */
__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_IR_SP_TX_DATA(srio_port), pkt);
}
else
{
/* Use the 8bit source ID */
sourceid = (sourceid >> 16) & 0xff;
/* MAINT Reads are 9 bytes */
__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_IR_SP_TX_CTRL(srio_port), 9<<16);
pkt |= CVMX_SRIO_CONFIG_PRIORITY << 30; /* priority [31:30] */
pkt |= 0 << 28; /* tt [29:28] */
pkt |= 0x8 << 24; /* ftype [27:24] */
pkt |= destid << 16; /* destID [23:16] */
pkt |= sourceid << 8; /* sourceID [15:8] */
pkt |= 0 << 4; /* transaction [7:4] */
pkt |= 8 << 0; /* rdsize [3:0] */
__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_IR_SP_TX_DATA(srio_port), pkt);
pkt = 0;
pkt |= 0xc0 << 24; /* srcTID [31:24] */
pkt |= hopcount << 16; /* hopcount [23:16] */
pkt |= offset >> 8; /* offset [15:0] */
__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_IR_SP_TX_DATA(srio_port), pkt);
pkt = 0;
pkt |= offset << 24; /* offset [31:27, wdptr[26], reserved[25:24] */
__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_IR_SP_TX_DATA(srio_port), pkt);
}
stop_cycle = cvmx_clock_get_rate(CVMX_CLOCK_CORE)/10 + cvmx_clock_get_count(CVMX_CLOCK_CORE);
do
{
return_code = cvmx_srio_receive_spf(srio_port, rx_buffer, sizeof(rx_buffer));
if ((return_code == 0) && (cvmx_clock_get_count(CVMX_CLOCK_CORE) > stop_cycle))
{
if (__cvmx_srio_state[srio_port].flags & CVMX_SRIO_INITIALIZE_DEBUG)
cvmx_dprintf("timeout\n");
return_code = -1;
}
} while (return_code == 0);
if (return_code == ((is16bit) ? 23 : 19))
{
if (is16bit)
{
if (offset & 4)
*result = *(uint32_t*)(rx_buffer + 15);
else
*result = *(uint32_t*)(rx_buffer + 11);
}
else
{
if (offset & 4)
*result = *(uint32_t*)(rx_buffer + 13);
else
*result = *(uint32_t*)(rx_buffer + 9);
}
if (__cvmx_srio_state[srio_port].flags & CVMX_SRIO_INITIALIZE_DEBUG)
cvmx_dprintf("0x%08x\n", (unsigned int)*result);
return_code = 0;
}
else
{
*result = 0xffffffff;
return_code = -1;
}
return return_code;
}
else
{
#if !defined(CVMX_BUILD_FOR_LINUX_HOST)
uint64_t physical;
physical = cvmx_srio_physical_map(srio_port,
CVMX_SRIO_WRITE_MODE_MAINTENANCE, CVMX_SRIO_CONFIG_PRIORITY,
CVMX_SRIO_READ_MODE_MAINTENANCE, CVMX_SRIO_CONFIG_PRIORITY,
srcid_index, destid, is16bit, offset + (hopcount<<24), 4);
if (!physical)
return -1;
if (__cvmx_srio_state[srio_port].flags & CVMX_SRIO_INITIALIZE_DEBUG)
cvmx_dprintf("SRIO%d: Remote read [id=0x%04x hop=%3d offset=0x%06x] <= ", srio_port, destid, hopcount, (unsigned int)offset);
/* Finally do the maintenance read to complete the config request */
*result = cvmx_read64_uint32(CVMX_ADD_IO_SEG(physical));
cvmx_srio_physical_unmap(physical, 4);
if (__cvmx_srio_state[srio_port].flags & CVMX_SRIO_INITIALIZE_DEBUG)
cvmx_dprintf("0x%08x\n", (unsigned int)*result);
return 0;
#else
return -1;
#endif
}
}
}
#ifdef CVMX_BUILD_FOR_LINUX_KERNEL
EXPORT_SYMBOL(cvmx_srio_config_read32);
#endif
/**
* Write 32bits to a Device's config space
*
* @param srio_port SRIO port the device is on
* @param srcid_index
* Which SRIO source ID to use. 0 = Primary, 1 = Secondary
* @param destid RapidIO device ID, or -1 for the local Octeon.
* @param is16bit Non zero if the transactions should use 16bit device IDs. Zero
* if transactions should use 8bit device IDs.
* @param hopcount Number of hops to the remote device. Use 0 for the local Octeon.
* @param offset Offset in config space. This must be a multiple of 32 bits.
* @param data Data to write.
*
* @return Zero on success, negative on failure.
*/
int cvmx_srio_config_write32(int srio_port, int srcid_index, int destid,
int is16bit, uint8_t hopcount, uint32_t offset,
uint32_t data)
{
if (destid == -1)
{
if (__cvmx_srio_state[srio_port].flags & CVMX_SRIO_INITIALIZE_DEBUG)
cvmx_dprintf("SRIO%d: Local write[0x%06x] => 0x%08x\n", srio_port, (unsigned int)offset, (unsigned int)data);
return __cvmx_srio_local_write32(srio_port, offset, data);
}
else
{
if (OCTEON_IS_MODEL(OCTEON_CN63XX_PASS1_X))
{
int return_code;
uint32_t pkt = 0;
uint32_t sourceid;
uint64_t stop_cycle;
char rx_buffer[64];
/* Tell the user */
if (__cvmx_srio_state[srio_port].flags & CVMX_SRIO_INITIALIZE_DEBUG)
cvmx_dprintf("SRIO%d: Remote write[id=0x%04x hop=%3d offset=0x%06x] => 0x%08x\n", srio_port, destid, hopcount, (unsigned int)offset, (unsigned int)data);
/* Read the proper source ID */
if (srcid_index)
__cvmx_srio_local_read32(srio_port, CVMX_SRIOMAINTX_SEC_DEV_ID(srio_port), &sourceid);
else
__cvmx_srio_local_read32(srio_port, CVMX_SRIOMAINTX_PRI_DEV_ID(srio_port), &sourceid);
if (is16bit)
{
/* Use the 16bit source ID */
sourceid &= 0xffff;
/* MAINT Writes are 19 bytes */
__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_IR_SP_TX_CTRL(srio_port), 19<<16);
pkt |= CVMX_SRIO_CONFIG_PRIORITY << 30; /* priority [31:30] */
pkt |= 1 << 28; /* tt [29:28] */
pkt |= 0x8 << 24; /* ftype [27:24] */
pkt |= destid << 8; /* destID [23:8] */
pkt |= sourceid >> 8; /* sourceID [7:0] */
__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_IR_SP_TX_DATA(srio_port), pkt);
pkt = 0;
pkt |= sourceid << 24; /* sourceID [31:24] */
pkt |= 1 << 20; /* transaction [23:20] */
pkt |= 8 << 16; /* wrsize [19:16] */
pkt |= 0xc0 << 8; /* srcTID [15:8] */
pkt |= hopcount; /* hopcount [7:0] */
__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_IR_SP_TX_DATA(srio_port), pkt);
pkt = 0;
pkt |= offset << 8; /* offset [31:11, wdptr[10], reserved[9:8] */
if ((offset & 4) == 0)
pkt |= 0xff & (data >> 24); /* data [7:0] */
__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_IR_SP_TX_DATA(srio_port), pkt);
if (offset & 4)
{
pkt = 0xff & (data >> 24);
__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_IR_SP_TX_DATA(srio_port), pkt);
pkt = data << 8;
__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_IR_SP_TX_DATA(srio_port), pkt);
}
else
{
pkt = data << 8;
__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_IR_SP_TX_DATA(srio_port), pkt);
__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_IR_SP_TX_DATA(srio_port), 0);
}
}
else
{
/* Use the 8bit source ID */
sourceid = (sourceid >> 16) & 0xff;
/* MAINT Writes are 17 bytes */
__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_IR_SP_TX_CTRL(srio_port), 17<<16);
pkt |= CVMX_SRIO_CONFIG_PRIORITY << 30; /* priority [31:30] */
pkt |= 0 << 28; /* tt [29:28] */
pkt |= 0x8 << 24; /* ftype [27:24] */
pkt |= destid << 16; /* destID [23:16] */
pkt |= sourceid << 8; /* sourceID [15:8] */
pkt |= 1 << 4; /* transaction [7:4] */
pkt |= 8 << 0; /* wrsize [3:0] */
__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_IR_SP_TX_DATA(srio_port), pkt);
pkt = 0;
pkt |= 0xc0 << 24; /* srcTID [31:24] */
pkt |= hopcount << 16; /* hopcount [23:16] */
pkt |= offset >> 8; /* offset [15:0] */
__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_IR_SP_TX_DATA(srio_port), pkt);
pkt = 0;
pkt |= offset << 24; /* offset [31:27, wdptr[26], reserved[25:24] */
if (offset & 4)
{
__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_IR_SP_TX_DATA(srio_port), pkt);
pkt = data >> 8;
__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_IR_SP_TX_DATA(srio_port), pkt);
pkt = data << 24;
__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_IR_SP_TX_DATA(srio_port), pkt);
}
else
{
pkt |= data >> 8; /* data [23:0] */
__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_IR_SP_TX_DATA(srio_port), pkt);
pkt = data << 24; /* data [31:24] */
__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_IR_SP_TX_DATA(srio_port), pkt);
__cvmx_srio_local_write32(srio_port, CVMX_SRIOMAINTX_IR_SP_TX_DATA(srio_port), 0);
}
}
stop_cycle = cvmx_clock_get_rate(CVMX_CLOCK_CORE)/10 + cvmx_clock_get_count(CVMX_CLOCK_CORE);
do
{
return_code = cvmx_srio_receive_spf(srio_port, rx_buffer, sizeof(rx_buffer));
if ((return_code == 0) && (cvmx_clock_get_count(CVMX_CLOCK_CORE) > stop_cycle))
{
if (__cvmx_srio_state[srio_port].flags & CVMX_SRIO_INITIALIZE_DEBUG)
cvmx_dprintf("timeout\n");
return_code = -1;
}
} while (return_code == 0);
if (return_code == ((is16bit) ? 15 : 11))
return_code = 0;
else
{
cvmx_dprintf("SRIO%d: Remote write failed\n", srio_port);
return_code = -1;
}
return return_code;
}
else
{
#if !defined(CVMX_BUILD_FOR_LINUX_HOST)
uint64_t physical = cvmx_srio_physical_map(srio_port,
CVMX_SRIO_WRITE_MODE_MAINTENANCE, CVMX_SRIO_CONFIG_PRIORITY,
CVMX_SRIO_READ_MODE_MAINTENANCE, CVMX_SRIO_CONFIG_PRIORITY,
srcid_index, destid, is16bit, offset + (hopcount<<24), 4);
if (!physical)
return -1;
if (__cvmx_srio_state[srio_port].flags & CVMX_SRIO_INITIALIZE_DEBUG)
cvmx_dprintf("SRIO%d: Remote write[id=0x%04x hop=%3d offset=0x%06x] => 0x%08x\n", srio_port, destid, hopcount, (unsigned int)offset, (unsigned int)data);
/* Finally do the maintenance write to complete the config request */
cvmx_write64_uint32(CVMX_ADD_IO_SEG(physical), data);
return cvmx_srio_physical_unmap(physical, 4);
#else
return -1;
#endif
}
}
}
/**
* Send a RapidIO doorbell to a remote device
*
* @param srio_port SRIO port the device is on
* @param srcid_index
* Which SRIO source ID to use. 0 = Primary, 1 = Secondary
* @param destid RapidIO device ID.
* @param is16bit Non zero if the transactions should use 16bit device IDs. Zero
* if transactions should use 8bit device IDs.
* @param priority Doorbell priority (0-3)
* @param data Data for doorbell.
*
* @return Zero on success, negative on failure.
*/
int cvmx_srio_send_doorbell(int srio_port, int srcid_index, int destid, int is16bit, int priority, uint16_t data)
{
cvmx_sriox_tx_bell_t tx_bell;
tx_bell.u64 = 0;
tx_bell.s.data = data;
tx_bell.s.dest_id = destid;
tx_bell.s.src_id = srcid_index;
tx_bell.s.id16 = !!is16bit;
tx_bell.s.priority = priority;
/* Make sure the previous doorbell has completed */
if (CVMX_WAIT_FOR_FIELD64(CVMX_SRIOX_TX_BELL(srio_port), cvmx_sriox_tx_bell_t, pending, ==, 0, CVMX_SRIO_DOORBELL_TIMEOUT))
{
cvmx_dprintf("SRIO%d: Pending bit stuck before doorbell\n", srio_port);
return -1;
}
if (__cvmx_srio_state[srio_port].flags & CVMX_SRIO_INITIALIZE_DEBUG)
cvmx_dprintf("SRIO%d: Send doorbell destid=0x%x, priority=%d, data=0x%x\n", srio_port, destid, priority, 0xffff & data);
/* Send the doorbell. We don't wait for it to complete. The next doorbell
may delay on the pending bit, but this gives the caller the ability to
do other stuff while the doorbell processes */
cvmx_write_csr(CVMX_SRIOX_TX_BELL(srio_port), tx_bell.u64);
return 0;
}
#ifdef CVMX_BUILD_FOR_LINUX_KERNEL
EXPORT_SYMBOL(cvmx_srio_send_doorbell);
#endif
/**
* Get the status of the last doorbell sent. If the dooorbell
* hardware is done, then the status is cleared to get ready for
* the next doorbell (or retry).
*
* @param srio_port SRIO port to check doorbell on
*
* @return Doorbell status
*/
cvmx_srio_doorbell_status_t cvmx_srio_send_doorbell_status(int srio_port)
{
cvmx_sriox_tx_bell_t tx_bell;
cvmx_sriox_tx_bell_info_t tx_bell_info;
cvmx_sriox_int_reg_t int_reg;
cvmx_sriox_int_reg_t int_reg_clear;
/* Return busy if the doorbell is still processing */
tx_bell.u64 = cvmx_read_csr(CVMX_SRIOX_TX_BELL(srio_port));
if (tx_bell.s.pending)
return CVMX_SRIO_DOORBELL_BUSY;
/* Read and clear the TX doorbell interrupts */
int_reg.u64 = cvmx_read_csr(CVMX_SRIOX_INT_REG(srio_port));
int_reg_clear.u64 = 0;
int_reg_clear.s.bell_err = int_reg.s.bell_err;
int_reg_clear.s.txbell = int_reg.s.txbell;
cvmx_write_csr(CVMX_SRIOX_INT_REG(srio_port), int_reg_clear.u64);
/* Check for errors */
if (int_reg.s.bell_err)
{
if (__cvmx_srio_state[srio_port].flags & CVMX_SRIO_INITIALIZE_DEBUG)
cvmx_dprintf("SRIO%d: Send doorbell failed\n", srio_port);
tx_bell_info.u64 = cvmx_read_csr(CVMX_SRIOX_TX_BELL_INFO(srio_port));
if (tx_bell_info.s.timeout)
return CVMX_SRIO_DOORBELL_TMOUT;
if (tx_bell_info.s.error)
return CVMX_SRIO_DOORBELL_ERROR;
if (tx_bell_info.s.retry)
return CVMX_SRIO_DOORBELL_RETRY;
}
/* Check if we're done */
if (int_reg.s.txbell)
return CVMX_SRIO_DOORBELL_DONE;
/* No doorbell found */
return CVMX_SRIO_DOORBELL_NONE;
}
/**
* Read a received doorbell and report data about it.
*
* @param srio_port SRIO port to check for the received doorbell
* @param destid_index
* Which Octeon destination ID was the doorbell for
* @param sequence_num
* Sequence number of doorbell (32bits)
* @param srcid RapidIO source ID of the doorbell sender
* @param priority Priority of the doorbell (0-3)
* @param is16bit Non zero if the transactions should use 16bit device IDs. Zero
* if transactions should use 8bit device IDs.
* @param data Data in the doorbell (16 bits)
*
* @return Doorbell status. Either DONE, NONE, or ERROR.
*/
cvmx_srio_doorbell_status_t cvmx_srio_receive_doorbell(int srio_port,
int *destid_index, uint32_t *sequence_num, int *srcid, int *priority,
int *is16bit, uint16_t *data)
{
cvmx_sriox_rx_bell_seq_t rx_bell_seq;
cvmx_sriox_rx_bell_t rx_bell;
/* Check if there are any pending doorbells */
rx_bell_seq.u64 = cvmx_read_csr(CVMX_SRIOX_RX_BELL_SEQ(srio_port));
if (!rx_bell_seq.s.count)
return CVMX_SRIO_DOORBELL_NONE;
/* Read the doorbell and write our return parameters */
rx_bell.u64 = cvmx_read_csr(CVMX_SRIOX_RX_BELL(srio_port));
*sequence_num = rx_bell_seq.s.seq;
*srcid = rx_bell.s.src_id;
*priority = rx_bell.s.priority;
*is16bit = rx_bell.s.id16;
*data = rx_bell.s.data;
*destid_index = rx_bell.s.dest_id;
if (__cvmx_srio_state[srio_port].flags & CVMX_SRIO_INITIALIZE_DEBUG)
cvmx_dprintf("SRIO%d: Receive doorbell sequence=0x%x, srcid=0x%x, priority=%d, data=0x%x\n",
srio_port, rx_bell_seq.s.seq, rx_bell.s.src_id, rx_bell.s.priority, rx_bell.s.data);
return CVMX_SRIO_DOORBELL_DONE;
}
/**
* Receive a packet from the Soft Packet FIFO (SPF).
*
* @param srio_port SRIO port to read the packet from.
* @param buffer Buffer to receive the packet.
* @param buffer_length
* Length of the buffer in bytes.
*
* @return Returns the length of the packet read. Negative on failure.
* Zero if no packets are available.
*/
int cvmx_srio_receive_spf(int srio_port, void *buffer, int buffer_length)
{
uint32_t *ptr = (uint32_t *)buffer;
cvmx_sriomaintx_ir_sp_rx_stat_t sriomaintx_ir_sp_rx_stat;
/* Read the SFP status */
if (__cvmx_srio_local_read32(srio_port, CVMX_SRIOMAINTX_IR_SP_RX_STAT(srio_port), &sriomaintx_ir_sp_rx_stat.u32))
return -1;
/* Return zero if there isn't a packet available */
if (sriomaintx_ir_sp_rx_stat.s.buffers < 1)
return 0;
if (__cvmx_srio_state[srio_port].flags & CVMX_SRIO_INITIALIZE_DEBUG)
cvmx_dprintf("SRIO%d: Soft packet FIFO received %d bytes", srio_port, sriomaintx_ir_sp_rx_stat.s.octets);
/* Return error if the packet is larger than our buffer */
if (sriomaintx_ir_sp_rx_stat.s.octets > buffer_length)
return -1;
/* Read out the packet four bytes at a time */
buffer_length = sriomaintx_ir_sp_rx_stat.s.octets;
while (buffer_length > 0)
{
__cvmx_srio_local_read32(srio_port, CVMX_SRIOMAINTX_IR_SP_RX_DATA(srio_port), ptr);
if (__cvmx_srio_state[srio_port].flags & CVMX_SRIO_INITIALIZE_DEBUG)
cvmx_dprintf(" %08x", (unsigned int)*ptr);
ptr++;
buffer_length-=4;
}
if (__cvmx_srio_state[srio_port].flags & CVMX_SRIO_INITIALIZE_DEBUG)
cvmx_dprintf("\n");
/* Return the number of bytes in the buffer */
return sriomaintx_ir_sp_rx_stat.s.octets;
}
#ifndef CVMX_BUILD_FOR_LINUX_HOST
/**
* Map a remote device's memory region into Octeon's physical
* address area. The caller can then map this into a core using
* the TLB or XKPHYS.
*
* @param srio_port SRIO port to map the device on
* @param write_op Type of operation to perform on a write to the device.
* Normally should be CVMX_SRIO_WRITE_MODE_AUTO.
* @param write_priority
* SRIO priority of writes (0-3)
* @param read_op Type of operation to perform on reads to the device.
* Normally should be CVMX_SRIO_READ_MODE_NORMAL.
* @param read_priority
* SRIO priority of reads (0-3)
* @param srcid_index
* Which SRIO source ID to use. 0 = Primary, 1 = Secondary
* @param destid RapidIO device ID.
* @param is16bit Non zero if the transactions should use 16bit device IDs. Zero
* if transactions should use 8bit device IDs.
* @param base Device base address to start the mapping
* @param size Size of the mapping in bytes
*
* @return Octeon 64bit physical address that accesses the remote device,
* or zero on failure.
*/
uint64_t cvmx_srio_physical_map(int srio_port, cvmx_srio_write_mode_t write_op,
int write_priority, cvmx_srio_read_mode_t read_op, int read_priority,
int srcid_index, int destid, int is16bit, uint64_t base, uint64_t size)
{
cvmx_sriox_s2m_typex_t needed_s2m_type;
cvmx_sli_mem_access_subidx_t needed_subid;
int s2m_index;
int subdid;
cvmx_sli_address_t sli_address;
/* We currently don't support mapping regions that span a 34 bit boundary.
Keeping track of multiple regions to span 34 bits is hard and not
likely to be needed */
if (((base+size-1)>>34) != (base>>34))
{
cvmx_dprintf("SRIO%d: Failed to map range 0x%llx-0x%llx spanning a 34bit boundary\n",
srio_port, (ULL)base, (ULL)base+size-1);
return 0;
}
/* Build the S2M_TYPE we are going to need */
needed_s2m_type.u64 = 0;
needed_s2m_type.s.wr_op = write_op;
needed_s2m_type.s.rd_op = read_op;
needed_s2m_type.s.wr_prior = write_priority;
needed_s2m_type.s.rd_prior = read_priority;
needed_s2m_type.s.src_id = srcid_index;
needed_s2m_type.s.id16 = !!is16bit;
/* Build the needed SubID config */
needed_subid.u64 = 0;
needed_subid.s.port = srio_port;
needed_subid.s.nmerge = 0;
/* FIXME: We might want to use the device ID swapping modes so the device
ID is part of the lower address bits. This would allow many more
devices to share S2M_TYPE indexes. This would require "base+size-1"
to fit in bits [17:0] or bits[25:0] for 8 bits of device ID */
if (base < (1ull<<34))
{
needed_subid.cn63xx.ba = destid;
needed_s2m_type.s.iaow_sel = 0;
}
else if (base < (1ull<<42))
{
needed_subid.cn63xx.ba = (base>>34) & 0xff;
needed_subid.cn63xx.ba |= ((uint64_t)destid & 0xff) << (42-34);
needed_subid.cn63xx.ba |= (((uint64_t)destid>>8) & 0xff) << (51-34);
needed_s2m_type.s.iaow_sel = 1;
}
else
{
if (destid>>8)
{
cvmx_dprintf("SRIO%d: Attempt to map 16bit device ID 0x%x using 66bit addressing\n", srio_port, destid);
return 0;
}
if (base>>50)
{
cvmx_dprintf("SRIO%d: Attempt to map address 0x%llx using 66bit addressing\n", srio_port, (ULL)base);
return 0;
}
needed_subid.cn63xx.ba = (base>>34) & 0xffff;
needed_subid.cn63xx.ba |= ((uint64_t)destid & 0xff) << (51-34);
needed_s2m_type.s.iaow_sel = 2;
}
/* Find a S2M_TYPE index to use. If this fails return 0 */
s2m_index = __cvmx_srio_alloc_s2m(srio_port, needed_s2m_type);
if (s2m_index == -1)
return 0;
/* Attach the SubID to the S2M_TYPE index */
needed_subid.s.rtype = s2m_index & 3;
needed_subid.s.wtype = s2m_index & 3;
needed_subid.cn63xx.ba |= (((uint64_t)s2m_index >> 2) & 1) << (50-34);
needed_subid.cn63xx.ba |= (((uint64_t)s2m_index >> 3) & 1) << (59-34);
/* Allocate a SubID for use */
subdid = __cvmx_srio_alloc_subid(needed_subid);
if (subdid == -1)
{
/* Free the s2m_index as we aren't using it */
__cvmx_srio_free_s2m(srio_port, s2m_index);
return 0;
}
/* Build the final core physical address */
sli_address.u64 = 0;
sli_address.mem.io = 1;
sli_address.mem.did = 3;
sli_address.mem.subdid = subdid>>2;
sli_address.mem.se = subdid & 3;
sli_address.mem.address = base; /* Bits[33:0] of full address */
return sli_address.u64;
}
/**
* Unmap a physical address window created by cvmx_srio_phys_map().
*
* @param physical_address
* Physical address returned by cvmx_srio_phys_map().
* @param size Size used on original call.
*
* @return Zero on success, negative on failure.
*/
int cvmx_srio_physical_unmap(uint64_t physical_address, uint64_t size)
{
cvmx_sli_mem_access_subidx_t subid;
int subdid = (physical_address >> 40) & 7;
int extender = (physical_address >> 34) & 3;
int mem_index = subdid * 4 + extender;
int read_s2m_type;
/* Get the subid setup so we can figure out where this mapping was for */
subid.u64 = cvmx_read_csr(CVMX_PEXP_SLI_MEM_ACCESS_SUBIDX(mem_index));
/* Type[0] is mapped to the Relaxed Ordering
Type[1] is mapped to the No Snoop
Type[2] is mapped directly to bit 50 of the SLI address
Type[3] is mapped directly to bit 59 of the SLI address */
read_s2m_type = ((subid.cn63xx.ba>>(50-34))&1<<2) | ((subid.cn63xx.ba>>(59-34))&1<<3);
read_s2m_type |= subid.s.rtype;
__cvmx_srio_free_subid(mem_index);
__cvmx_srio_free_s2m(subid.s.port, read_s2m_type);
return 0;
}
#ifdef CVMX_ENABLE_PKO_FUNCTIONS
/**
* fill out outbound message descriptor
*
* @param port pip/ipd port number
* @param buf_ptr pointer to a buffer pointer. the buffer pointer points
* to a chain of buffers that hold an outbound srio packet.
* the packet can take the format of (1) a pip/ipd inbound
* message or (2) an application-generated outbound message
* @param desc_ptr pointer to an outbound message descriptor. should be null
* if *buf_ptr is in the format (1)
*
* @return 0 on success; negative of failure.
*/
int cvmx_srio_omsg_desc (uint64_t port, cvmx_buf_ptr_t *buf_ptr,
cvmx_srio_tx_message_header_t *desc_ptr)
{
int ret_val = -1;
int intf_num;
cvmx_helper_interface_mode_t imode;
uint64_t *desc_addr, *hdr_addr;
cvmx_srio_rx_message_header_t rx_msg_hdr;
cvmx_srio_tx_message_header_t *tx_msg_hdr_ptr;
if (buf_ptr == NULL)
return ret_val;
/* check if port is an srio port */
intf_num = cvmx_helper_get_interface_num (port);
imode = cvmx_helper_interface_get_mode (intf_num);
if (imode != CVMX_HELPER_INTERFACE_MODE_SRIO)
return ret_val;
/* app-generated outbound message. descriptor space pre-allocated */
if (desc_ptr != NULL)
{
desc_addr = (uint64_t *) cvmx_phys_to_ptr ((*buf_ptr).s.addr);
*desc_addr = *(uint64_t *) desc_ptr;
ret_val = 0;
return ret_val;
}
/* pip/ipd inbound message. 16-byte srio message header is present */
hdr_addr = (uint64_t *) cvmx_phys_to_ptr ((*buf_ptr).s.addr);
rx_msg_hdr.word0.u64 = *hdr_addr;
/* adjust buffer pointer to get rid of srio message header word 0 */
(*buf_ptr).s.addr += 8;
(*buf_ptr).s.size -= 8; /* last buffer or not */
if ((*buf_ptr).s.addr >> 7 > ((*buf_ptr).s.addr - 8) >> 7)
(*buf_ptr).s.back++;
tx_msg_hdr_ptr = (cvmx_srio_tx_message_header_t *)
cvmx_phys_to_ptr ((*buf_ptr).s.addr);
/* transfer values from rx to tx */
tx_msg_hdr_ptr->s.prio = rx_msg_hdr.word0.s.prio;
tx_msg_hdr_ptr->s.tt = rx_msg_hdr.word0.s.tt; /* called id in hrm */
tx_msg_hdr_ptr->s.sis = rx_msg_hdr.word0.s.dis;
tx_msg_hdr_ptr->s.ssize = rx_msg_hdr.word0.s.ssize;
tx_msg_hdr_ptr->s.did = rx_msg_hdr.word0.s.sid;
tx_msg_hdr_ptr->s.mbox = rx_msg_hdr.word0.s.mbox;
/* other values we have to decide */
tx_msg_hdr_ptr->s.xmbox = 0; /* multi-segement in general */
tx_msg_hdr_ptr->s.letter = 0; /* fake like traffic gen */
tx_msg_hdr_ptr->s.lns = 0; /* not use sriox_omsg_ctrly[] */
tx_msg_hdr_ptr->s.intr = 1; /* get status */
ret_val = 0;
return ret_val;
}
#endif
#endif