freebsd-dev/sys/contrib/octeon-sdk/cvmx-npei-defs.h
Juli Mallett 04b6fa8330 Merge Cavium Octeon SDK 2.0 Simple Executive; this brings some fixes and new
facilities as well as support for the Octeon 2 family of SoCs.

XXX Note that with our antediluvian assembler, we can't support some Octeon 2
    instructions and fall back to using the old ones instead.
2010-11-28 08:18:16 +00:00

7677 lines
379 KiB
C

/***********************license start***************
* Copyright (c) 2003-2010 Cavium Networks (support@cavium.com). All rights
* reserved.
*
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions are
* met:
*
* * Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
*
* * Redistributions in binary form must reproduce the above
* copyright notice, this list of conditions and the following
* disclaimer in the documentation and/or other materials provided
* with the distribution.
* * Neither the name of Cavium Networks nor the names of
* its contributors may be used to endorse or promote products
* derived from this software without specific prior written
* permission.
* This Software, including technical data, may be subject to U.S. export control
* laws, including the U.S. Export Administration Act and its associated
* regulations, and may be subject to export or import regulations in other
* countries.
* TO THE MAXIMUM EXTENT PERMITTED BY LAW, THE SOFTWARE IS PROVIDED "AS IS"
* AND WITH ALL FAULTS AND CAVIUM NETWORKS MAKES NO PROMISES, REPRESENTATIONS OR
* WARRANTIES, EITHER EXPRESS, IMPLIED, STATUTORY, OR OTHERWISE, WITH RESPECT TO
* THE SOFTWARE, INCLUDING ITS CONDITION, ITS CONFORMITY TO ANY REPRESENTATION OR
* DESCRIPTION, OR THE EXISTENCE OF ANY LATENT OR PATENT DEFECTS, AND CAVIUM
* SPECIFICALLY DISCLAIMS ALL IMPLIED (IF ANY) WARRANTIES OF TITLE,
* MERCHANTABILITY, NONINFRINGEMENT, FITNESS FOR A PARTICULAR PURPOSE, LACK OF
* VIRUSES, ACCURACY OR COMPLETENESS, QUIET ENJOYMENT, QUIET POSSESSION OR
* CORRESPONDENCE TO DESCRIPTION. THE ENTIRE RISK ARISING OUT OF USE OR
* PERFORMANCE OF THE SOFTWARE LIES WITH YOU.
***********************license end**************************************/
/**
* cvmx-npei-defs.h
*
* Configuration and status register (CSR) type definitions for
* Octeon npei.
*
* This file is auto generated. Do not edit.
*
* <hr>$Revision$<hr>
*
*/
#ifndef __CVMX_NPEI_TYPEDEFS_H__
#define __CVMX_NPEI_TYPEDEFS_H__
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_NPEI_BAR1_INDEXX(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 31))) ||
(OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 31)))))
cvmx_warn("CVMX_NPEI_BAR1_INDEXX(%lu) is invalid on this chip\n", offset);
return 0x0000000000000000ull + ((offset) & 31) * 16;
}
#else
#define CVMX_NPEI_BAR1_INDEXX(offset) (0x0000000000000000ull + ((offset) & 31) * 16)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_BIST_STATUS CVMX_NPEI_BIST_STATUS_FUNC()
static inline uint64_t CVMX_NPEI_BIST_STATUS_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_BIST_STATUS not supported on this chip\n");
return 0x0000000000000580ull;
}
#else
#define CVMX_NPEI_BIST_STATUS (0x0000000000000580ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_BIST_STATUS2 CVMX_NPEI_BIST_STATUS2_FUNC()
static inline uint64_t CVMX_NPEI_BIST_STATUS2_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_BIST_STATUS2 not supported on this chip\n");
return 0x0000000000000680ull;
}
#else
#define CVMX_NPEI_BIST_STATUS2 (0x0000000000000680ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_CTL_PORT0 CVMX_NPEI_CTL_PORT0_FUNC()
static inline uint64_t CVMX_NPEI_CTL_PORT0_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_CTL_PORT0 not supported on this chip\n");
return 0x0000000000000250ull;
}
#else
#define CVMX_NPEI_CTL_PORT0 (0x0000000000000250ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_CTL_PORT1 CVMX_NPEI_CTL_PORT1_FUNC()
static inline uint64_t CVMX_NPEI_CTL_PORT1_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_CTL_PORT1 not supported on this chip\n");
return 0x0000000000000260ull;
}
#else
#define CVMX_NPEI_CTL_PORT1 (0x0000000000000260ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_CTL_STATUS CVMX_NPEI_CTL_STATUS_FUNC()
static inline uint64_t CVMX_NPEI_CTL_STATUS_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_CTL_STATUS not supported on this chip\n");
return 0x0000000000000570ull;
}
#else
#define CVMX_NPEI_CTL_STATUS (0x0000000000000570ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_CTL_STATUS2 CVMX_NPEI_CTL_STATUS2_FUNC()
static inline uint64_t CVMX_NPEI_CTL_STATUS2_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_CTL_STATUS2 not supported on this chip\n");
return 0x0000000000003C00ull;
}
#else
#define CVMX_NPEI_CTL_STATUS2 (0x0000000000003C00ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_DATA_OUT_CNT CVMX_NPEI_DATA_OUT_CNT_FUNC()
static inline uint64_t CVMX_NPEI_DATA_OUT_CNT_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_DATA_OUT_CNT not supported on this chip\n");
return 0x00000000000005F0ull;
}
#else
#define CVMX_NPEI_DATA_OUT_CNT (0x00000000000005F0ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_DBG_DATA CVMX_NPEI_DBG_DATA_FUNC()
static inline uint64_t CVMX_NPEI_DBG_DATA_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_DBG_DATA not supported on this chip\n");
return 0x0000000000000510ull;
}
#else
#define CVMX_NPEI_DBG_DATA (0x0000000000000510ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_DBG_SELECT CVMX_NPEI_DBG_SELECT_FUNC()
static inline uint64_t CVMX_NPEI_DBG_SELECT_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_DBG_SELECT not supported on this chip\n");
return 0x0000000000000500ull;
}
#else
#define CVMX_NPEI_DBG_SELECT (0x0000000000000500ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_DMA0_INT_LEVEL CVMX_NPEI_DMA0_INT_LEVEL_FUNC()
static inline uint64_t CVMX_NPEI_DMA0_INT_LEVEL_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_DMA0_INT_LEVEL not supported on this chip\n");
return 0x00000000000005C0ull;
}
#else
#define CVMX_NPEI_DMA0_INT_LEVEL (0x00000000000005C0ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_DMA1_INT_LEVEL CVMX_NPEI_DMA1_INT_LEVEL_FUNC()
static inline uint64_t CVMX_NPEI_DMA1_INT_LEVEL_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_DMA1_INT_LEVEL not supported on this chip\n");
return 0x00000000000005D0ull;
}
#else
#define CVMX_NPEI_DMA1_INT_LEVEL (0x00000000000005D0ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_NPEI_DMAX_COUNTS(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 4))) ||
(OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 4)))))
cvmx_warn("CVMX_NPEI_DMAX_COUNTS(%lu) is invalid on this chip\n", offset);
return 0x0000000000000450ull + ((offset) & 7) * 16;
}
#else
#define CVMX_NPEI_DMAX_COUNTS(offset) (0x0000000000000450ull + ((offset) & 7) * 16)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_NPEI_DMAX_DBELL(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 4))) ||
(OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 4)))))
cvmx_warn("CVMX_NPEI_DMAX_DBELL(%lu) is invalid on this chip\n", offset);
return 0x00000000000003B0ull + ((offset) & 7) * 16;
}
#else
#define CVMX_NPEI_DMAX_DBELL(offset) (0x00000000000003B0ull + ((offset) & 7) * 16)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_NPEI_DMAX_IBUFF_SADDR(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 4))) ||
(OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 4)))))
cvmx_warn("CVMX_NPEI_DMAX_IBUFF_SADDR(%lu) is invalid on this chip\n", offset);
return 0x0000000000000400ull + ((offset) & 7) * 16;
}
#else
#define CVMX_NPEI_DMAX_IBUFF_SADDR(offset) (0x0000000000000400ull + ((offset) & 7) * 16)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_NPEI_DMAX_NADDR(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 4))) ||
(OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 4)))))
cvmx_warn("CVMX_NPEI_DMAX_NADDR(%lu) is invalid on this chip\n", offset);
return 0x00000000000004A0ull + ((offset) & 7) * 16;
}
#else
#define CVMX_NPEI_DMAX_NADDR(offset) (0x00000000000004A0ull + ((offset) & 7) * 16)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_DMA_CNTS CVMX_NPEI_DMA_CNTS_FUNC()
static inline uint64_t CVMX_NPEI_DMA_CNTS_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_DMA_CNTS not supported on this chip\n");
return 0x00000000000005E0ull;
}
#else
#define CVMX_NPEI_DMA_CNTS (0x00000000000005E0ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_DMA_CONTROL CVMX_NPEI_DMA_CONTROL_FUNC()
static inline uint64_t CVMX_NPEI_DMA_CONTROL_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_DMA_CONTROL not supported on this chip\n");
return 0x00000000000003A0ull;
}
#else
#define CVMX_NPEI_DMA_CONTROL (0x00000000000003A0ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_DMA_PCIE_REQ_NUM CVMX_NPEI_DMA_PCIE_REQ_NUM_FUNC()
static inline uint64_t CVMX_NPEI_DMA_PCIE_REQ_NUM_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_DMA_PCIE_REQ_NUM not supported on this chip\n");
return 0x00000000000005B0ull;
}
#else
#define CVMX_NPEI_DMA_PCIE_REQ_NUM (0x00000000000005B0ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_DMA_STATE1 CVMX_NPEI_DMA_STATE1_FUNC()
static inline uint64_t CVMX_NPEI_DMA_STATE1_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX)))
cvmx_warn("CVMX_NPEI_DMA_STATE1 not supported on this chip\n");
return 0x00000000000006C0ull;
}
#else
#define CVMX_NPEI_DMA_STATE1 (0x00000000000006C0ull)
#endif
#define CVMX_NPEI_DMA_STATE1_P1 (0x0000000000000680ull)
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_DMA_STATE2 CVMX_NPEI_DMA_STATE2_FUNC()
static inline uint64_t CVMX_NPEI_DMA_STATE2_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX)))
cvmx_warn("CVMX_NPEI_DMA_STATE2 not supported on this chip\n");
return 0x00000000000006D0ull;
}
#else
#define CVMX_NPEI_DMA_STATE2 (0x00000000000006D0ull)
#endif
#define CVMX_NPEI_DMA_STATE2_P1 (0x0000000000000690ull)
#define CVMX_NPEI_DMA_STATE3_P1 (0x00000000000006A0ull)
#define CVMX_NPEI_DMA_STATE4_P1 (0x00000000000006B0ull)
#define CVMX_NPEI_DMA_STATE5_P1 (0x00000000000006C0ull)
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_INT_A_ENB CVMX_NPEI_INT_A_ENB_FUNC()
static inline uint64_t CVMX_NPEI_INT_A_ENB_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_INT_A_ENB not supported on this chip\n");
return 0x0000000000000560ull;
}
#else
#define CVMX_NPEI_INT_A_ENB (0x0000000000000560ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_INT_A_ENB2 CVMX_NPEI_INT_A_ENB2_FUNC()
static inline uint64_t CVMX_NPEI_INT_A_ENB2_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_INT_A_ENB2 not supported on this chip\n");
return 0x0000000000003CE0ull;
}
#else
#define CVMX_NPEI_INT_A_ENB2 (0x0000000000003CE0ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_INT_A_SUM CVMX_NPEI_INT_A_SUM_FUNC()
static inline uint64_t CVMX_NPEI_INT_A_SUM_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_INT_A_SUM not supported on this chip\n");
return 0x0000000000000550ull;
}
#else
#define CVMX_NPEI_INT_A_SUM (0x0000000000000550ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_INT_ENB CVMX_NPEI_INT_ENB_FUNC()
static inline uint64_t CVMX_NPEI_INT_ENB_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_INT_ENB not supported on this chip\n");
return 0x0000000000000540ull;
}
#else
#define CVMX_NPEI_INT_ENB (0x0000000000000540ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_INT_ENB2 CVMX_NPEI_INT_ENB2_FUNC()
static inline uint64_t CVMX_NPEI_INT_ENB2_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_INT_ENB2 not supported on this chip\n");
return 0x0000000000003CD0ull;
}
#else
#define CVMX_NPEI_INT_ENB2 (0x0000000000003CD0ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_INT_INFO CVMX_NPEI_INT_INFO_FUNC()
static inline uint64_t CVMX_NPEI_INT_INFO_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_INT_INFO not supported on this chip\n");
return 0x0000000000000590ull;
}
#else
#define CVMX_NPEI_INT_INFO (0x0000000000000590ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_INT_SUM CVMX_NPEI_INT_SUM_FUNC()
static inline uint64_t CVMX_NPEI_INT_SUM_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_INT_SUM not supported on this chip\n");
return 0x0000000000000530ull;
}
#else
#define CVMX_NPEI_INT_SUM (0x0000000000000530ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_INT_SUM2 CVMX_NPEI_INT_SUM2_FUNC()
static inline uint64_t CVMX_NPEI_INT_SUM2_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_INT_SUM2 not supported on this chip\n");
return 0x0000000000003CC0ull;
}
#else
#define CVMX_NPEI_INT_SUM2 (0x0000000000003CC0ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_LAST_WIN_RDATA0 CVMX_NPEI_LAST_WIN_RDATA0_FUNC()
static inline uint64_t CVMX_NPEI_LAST_WIN_RDATA0_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_LAST_WIN_RDATA0 not supported on this chip\n");
return 0x0000000000000600ull;
}
#else
#define CVMX_NPEI_LAST_WIN_RDATA0 (0x0000000000000600ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_LAST_WIN_RDATA1 CVMX_NPEI_LAST_WIN_RDATA1_FUNC()
static inline uint64_t CVMX_NPEI_LAST_WIN_RDATA1_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_LAST_WIN_RDATA1 not supported on this chip\n");
return 0x0000000000000610ull;
}
#else
#define CVMX_NPEI_LAST_WIN_RDATA1 (0x0000000000000610ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_MEM_ACCESS_CTL CVMX_NPEI_MEM_ACCESS_CTL_FUNC()
static inline uint64_t CVMX_NPEI_MEM_ACCESS_CTL_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_MEM_ACCESS_CTL not supported on this chip\n");
return 0x00000000000004F0ull;
}
#else
#define CVMX_NPEI_MEM_ACCESS_CTL (0x00000000000004F0ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_NPEI_MEM_ACCESS_SUBIDX(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN52XX) && (((offset >= 12) && (offset <= 27)))) ||
(OCTEON_IS_MODEL(OCTEON_CN56XX) && (((offset >= 12) && (offset <= 27))))))
cvmx_warn("CVMX_NPEI_MEM_ACCESS_SUBIDX(%lu) is invalid on this chip\n", offset);
return 0x0000000000000340ull + ((offset) & 31) * 16 - 16*12;
}
#else
#define CVMX_NPEI_MEM_ACCESS_SUBIDX(offset) (0x0000000000000340ull + ((offset) & 31) * 16 - 16*12)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_MSI_ENB0 CVMX_NPEI_MSI_ENB0_FUNC()
static inline uint64_t CVMX_NPEI_MSI_ENB0_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_MSI_ENB0 not supported on this chip\n");
return 0x0000000000003C50ull;
}
#else
#define CVMX_NPEI_MSI_ENB0 (0x0000000000003C50ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_MSI_ENB1 CVMX_NPEI_MSI_ENB1_FUNC()
static inline uint64_t CVMX_NPEI_MSI_ENB1_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_MSI_ENB1 not supported on this chip\n");
return 0x0000000000003C60ull;
}
#else
#define CVMX_NPEI_MSI_ENB1 (0x0000000000003C60ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_MSI_ENB2 CVMX_NPEI_MSI_ENB2_FUNC()
static inline uint64_t CVMX_NPEI_MSI_ENB2_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_MSI_ENB2 not supported on this chip\n");
return 0x0000000000003C70ull;
}
#else
#define CVMX_NPEI_MSI_ENB2 (0x0000000000003C70ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_MSI_ENB3 CVMX_NPEI_MSI_ENB3_FUNC()
static inline uint64_t CVMX_NPEI_MSI_ENB3_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_MSI_ENB3 not supported on this chip\n");
return 0x0000000000003C80ull;
}
#else
#define CVMX_NPEI_MSI_ENB3 (0x0000000000003C80ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_MSI_RCV0 CVMX_NPEI_MSI_RCV0_FUNC()
static inline uint64_t CVMX_NPEI_MSI_RCV0_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_MSI_RCV0 not supported on this chip\n");
return 0x0000000000003C10ull;
}
#else
#define CVMX_NPEI_MSI_RCV0 (0x0000000000003C10ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_MSI_RCV1 CVMX_NPEI_MSI_RCV1_FUNC()
static inline uint64_t CVMX_NPEI_MSI_RCV1_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_MSI_RCV1 not supported on this chip\n");
return 0x0000000000003C20ull;
}
#else
#define CVMX_NPEI_MSI_RCV1 (0x0000000000003C20ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_MSI_RCV2 CVMX_NPEI_MSI_RCV2_FUNC()
static inline uint64_t CVMX_NPEI_MSI_RCV2_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_MSI_RCV2 not supported on this chip\n");
return 0x0000000000003C30ull;
}
#else
#define CVMX_NPEI_MSI_RCV2 (0x0000000000003C30ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_MSI_RCV3 CVMX_NPEI_MSI_RCV3_FUNC()
static inline uint64_t CVMX_NPEI_MSI_RCV3_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_MSI_RCV3 not supported on this chip\n");
return 0x0000000000003C40ull;
}
#else
#define CVMX_NPEI_MSI_RCV3 (0x0000000000003C40ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_MSI_RD_MAP CVMX_NPEI_MSI_RD_MAP_FUNC()
static inline uint64_t CVMX_NPEI_MSI_RD_MAP_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_MSI_RD_MAP not supported on this chip\n");
return 0x0000000000003CA0ull;
}
#else
#define CVMX_NPEI_MSI_RD_MAP (0x0000000000003CA0ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_MSI_W1C_ENB0 CVMX_NPEI_MSI_W1C_ENB0_FUNC()
static inline uint64_t CVMX_NPEI_MSI_W1C_ENB0_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_MSI_W1C_ENB0 not supported on this chip\n");
return 0x0000000000003CF0ull;
}
#else
#define CVMX_NPEI_MSI_W1C_ENB0 (0x0000000000003CF0ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_MSI_W1C_ENB1 CVMX_NPEI_MSI_W1C_ENB1_FUNC()
static inline uint64_t CVMX_NPEI_MSI_W1C_ENB1_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_MSI_W1C_ENB1 not supported on this chip\n");
return 0x0000000000003D00ull;
}
#else
#define CVMX_NPEI_MSI_W1C_ENB1 (0x0000000000003D00ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_MSI_W1C_ENB2 CVMX_NPEI_MSI_W1C_ENB2_FUNC()
static inline uint64_t CVMX_NPEI_MSI_W1C_ENB2_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_MSI_W1C_ENB2 not supported on this chip\n");
return 0x0000000000003D10ull;
}
#else
#define CVMX_NPEI_MSI_W1C_ENB2 (0x0000000000003D10ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_MSI_W1C_ENB3 CVMX_NPEI_MSI_W1C_ENB3_FUNC()
static inline uint64_t CVMX_NPEI_MSI_W1C_ENB3_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_MSI_W1C_ENB3 not supported on this chip\n");
return 0x0000000000003D20ull;
}
#else
#define CVMX_NPEI_MSI_W1C_ENB3 (0x0000000000003D20ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_MSI_W1S_ENB0 CVMX_NPEI_MSI_W1S_ENB0_FUNC()
static inline uint64_t CVMX_NPEI_MSI_W1S_ENB0_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_MSI_W1S_ENB0 not supported on this chip\n");
return 0x0000000000003D30ull;
}
#else
#define CVMX_NPEI_MSI_W1S_ENB0 (0x0000000000003D30ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_MSI_W1S_ENB1 CVMX_NPEI_MSI_W1S_ENB1_FUNC()
static inline uint64_t CVMX_NPEI_MSI_W1S_ENB1_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_MSI_W1S_ENB1 not supported on this chip\n");
return 0x0000000000003D40ull;
}
#else
#define CVMX_NPEI_MSI_W1S_ENB1 (0x0000000000003D40ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_MSI_W1S_ENB2 CVMX_NPEI_MSI_W1S_ENB2_FUNC()
static inline uint64_t CVMX_NPEI_MSI_W1S_ENB2_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_MSI_W1S_ENB2 not supported on this chip\n");
return 0x0000000000003D50ull;
}
#else
#define CVMX_NPEI_MSI_W1S_ENB2 (0x0000000000003D50ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_MSI_W1S_ENB3 CVMX_NPEI_MSI_W1S_ENB3_FUNC()
static inline uint64_t CVMX_NPEI_MSI_W1S_ENB3_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_MSI_W1S_ENB3 not supported on this chip\n");
return 0x0000000000003D60ull;
}
#else
#define CVMX_NPEI_MSI_W1S_ENB3 (0x0000000000003D60ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_MSI_WR_MAP CVMX_NPEI_MSI_WR_MAP_FUNC()
static inline uint64_t CVMX_NPEI_MSI_WR_MAP_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_MSI_WR_MAP not supported on this chip\n");
return 0x0000000000003C90ull;
}
#else
#define CVMX_NPEI_MSI_WR_MAP (0x0000000000003C90ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PCIE_CREDIT_CNT CVMX_NPEI_PCIE_CREDIT_CNT_FUNC()
static inline uint64_t CVMX_NPEI_PCIE_CREDIT_CNT_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PCIE_CREDIT_CNT not supported on this chip\n");
return 0x0000000000003D70ull;
}
#else
#define CVMX_NPEI_PCIE_CREDIT_CNT (0x0000000000003D70ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PCIE_MSI_RCV CVMX_NPEI_PCIE_MSI_RCV_FUNC()
static inline uint64_t CVMX_NPEI_PCIE_MSI_RCV_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PCIE_MSI_RCV not supported on this chip\n");
return 0x0000000000003CB0ull;
}
#else
#define CVMX_NPEI_PCIE_MSI_RCV (0x0000000000003CB0ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PCIE_MSI_RCV_B1 CVMX_NPEI_PCIE_MSI_RCV_B1_FUNC()
static inline uint64_t CVMX_NPEI_PCIE_MSI_RCV_B1_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PCIE_MSI_RCV_B1 not supported on this chip\n");
return 0x0000000000000650ull;
}
#else
#define CVMX_NPEI_PCIE_MSI_RCV_B1 (0x0000000000000650ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PCIE_MSI_RCV_B2 CVMX_NPEI_PCIE_MSI_RCV_B2_FUNC()
static inline uint64_t CVMX_NPEI_PCIE_MSI_RCV_B2_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PCIE_MSI_RCV_B2 not supported on this chip\n");
return 0x0000000000000660ull;
}
#else
#define CVMX_NPEI_PCIE_MSI_RCV_B2 (0x0000000000000660ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PCIE_MSI_RCV_B3 CVMX_NPEI_PCIE_MSI_RCV_B3_FUNC()
static inline uint64_t CVMX_NPEI_PCIE_MSI_RCV_B3_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PCIE_MSI_RCV_B3 not supported on this chip\n");
return 0x0000000000000670ull;
}
#else
#define CVMX_NPEI_PCIE_MSI_RCV_B3 (0x0000000000000670ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_NPEI_PKTX_CNTS(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 31))) ||
(OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 31)))))
cvmx_warn("CVMX_NPEI_PKTX_CNTS(%lu) is invalid on this chip\n", offset);
return 0x0000000000002400ull + ((offset) & 31) * 16;
}
#else
#define CVMX_NPEI_PKTX_CNTS(offset) (0x0000000000002400ull + ((offset) & 31) * 16)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_NPEI_PKTX_INSTR_BADDR(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 31))) ||
(OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 31)))))
cvmx_warn("CVMX_NPEI_PKTX_INSTR_BADDR(%lu) is invalid on this chip\n", offset);
return 0x0000000000002800ull + ((offset) & 31) * 16;
}
#else
#define CVMX_NPEI_PKTX_INSTR_BADDR(offset) (0x0000000000002800ull + ((offset) & 31) * 16)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_NPEI_PKTX_INSTR_BAOFF_DBELL(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 31))) ||
(OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 31)))))
cvmx_warn("CVMX_NPEI_PKTX_INSTR_BAOFF_DBELL(%lu) is invalid on this chip\n", offset);
return 0x0000000000002C00ull + ((offset) & 31) * 16;
}
#else
#define CVMX_NPEI_PKTX_INSTR_BAOFF_DBELL(offset) (0x0000000000002C00ull + ((offset) & 31) * 16)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_NPEI_PKTX_INSTR_FIFO_RSIZE(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 31))) ||
(OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 31)))))
cvmx_warn("CVMX_NPEI_PKTX_INSTR_FIFO_RSIZE(%lu) is invalid on this chip\n", offset);
return 0x0000000000003000ull + ((offset) & 31) * 16;
}
#else
#define CVMX_NPEI_PKTX_INSTR_FIFO_RSIZE(offset) (0x0000000000003000ull + ((offset) & 31) * 16)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_NPEI_PKTX_INSTR_HEADER(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 31))) ||
(OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 31)))))
cvmx_warn("CVMX_NPEI_PKTX_INSTR_HEADER(%lu) is invalid on this chip\n", offset);
return 0x0000000000003400ull + ((offset) & 31) * 16;
}
#else
#define CVMX_NPEI_PKTX_INSTR_HEADER(offset) (0x0000000000003400ull + ((offset) & 31) * 16)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_NPEI_PKTX_IN_BP(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 31))) ||
(OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 31)))))
cvmx_warn("CVMX_NPEI_PKTX_IN_BP(%lu) is invalid on this chip\n", offset);
return 0x0000000000003800ull + ((offset) & 31) * 16;
}
#else
#define CVMX_NPEI_PKTX_IN_BP(offset) (0x0000000000003800ull + ((offset) & 31) * 16)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_NPEI_PKTX_SLIST_BADDR(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 31))) ||
(OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 31)))))
cvmx_warn("CVMX_NPEI_PKTX_SLIST_BADDR(%lu) is invalid on this chip\n", offset);
return 0x0000000000001400ull + ((offset) & 31) * 16;
}
#else
#define CVMX_NPEI_PKTX_SLIST_BADDR(offset) (0x0000000000001400ull + ((offset) & 31) * 16)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_NPEI_PKTX_SLIST_BAOFF_DBELL(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 31))) ||
(OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 31)))))
cvmx_warn("CVMX_NPEI_PKTX_SLIST_BAOFF_DBELL(%lu) is invalid on this chip\n", offset);
return 0x0000000000001800ull + ((offset) & 31) * 16;
}
#else
#define CVMX_NPEI_PKTX_SLIST_BAOFF_DBELL(offset) (0x0000000000001800ull + ((offset) & 31) * 16)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_NPEI_PKTX_SLIST_FIFO_RSIZE(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 31))) ||
(OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 31)))))
cvmx_warn("CVMX_NPEI_PKTX_SLIST_FIFO_RSIZE(%lu) is invalid on this chip\n", offset);
return 0x0000000000001C00ull + ((offset) & 31) * 16;
}
#else
#define CVMX_NPEI_PKTX_SLIST_FIFO_RSIZE(offset) (0x0000000000001C00ull + ((offset) & 31) * 16)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PKT_CNT_INT CVMX_NPEI_PKT_CNT_INT_FUNC()
static inline uint64_t CVMX_NPEI_PKT_CNT_INT_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PKT_CNT_INT not supported on this chip\n");
return 0x0000000000001110ull;
}
#else
#define CVMX_NPEI_PKT_CNT_INT (0x0000000000001110ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PKT_CNT_INT_ENB CVMX_NPEI_PKT_CNT_INT_ENB_FUNC()
static inline uint64_t CVMX_NPEI_PKT_CNT_INT_ENB_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PKT_CNT_INT_ENB not supported on this chip\n");
return 0x0000000000001130ull;
}
#else
#define CVMX_NPEI_PKT_CNT_INT_ENB (0x0000000000001130ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PKT_DATA_OUT_ES CVMX_NPEI_PKT_DATA_OUT_ES_FUNC()
static inline uint64_t CVMX_NPEI_PKT_DATA_OUT_ES_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PKT_DATA_OUT_ES not supported on this chip\n");
return 0x00000000000010B0ull;
}
#else
#define CVMX_NPEI_PKT_DATA_OUT_ES (0x00000000000010B0ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PKT_DATA_OUT_NS CVMX_NPEI_PKT_DATA_OUT_NS_FUNC()
static inline uint64_t CVMX_NPEI_PKT_DATA_OUT_NS_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PKT_DATA_OUT_NS not supported on this chip\n");
return 0x00000000000010A0ull;
}
#else
#define CVMX_NPEI_PKT_DATA_OUT_NS (0x00000000000010A0ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PKT_DATA_OUT_ROR CVMX_NPEI_PKT_DATA_OUT_ROR_FUNC()
static inline uint64_t CVMX_NPEI_PKT_DATA_OUT_ROR_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PKT_DATA_OUT_ROR not supported on this chip\n");
return 0x0000000000001090ull;
}
#else
#define CVMX_NPEI_PKT_DATA_OUT_ROR (0x0000000000001090ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PKT_DPADDR CVMX_NPEI_PKT_DPADDR_FUNC()
static inline uint64_t CVMX_NPEI_PKT_DPADDR_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PKT_DPADDR not supported on this chip\n");
return 0x0000000000001080ull;
}
#else
#define CVMX_NPEI_PKT_DPADDR (0x0000000000001080ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PKT_INPUT_CONTROL CVMX_NPEI_PKT_INPUT_CONTROL_FUNC()
static inline uint64_t CVMX_NPEI_PKT_INPUT_CONTROL_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PKT_INPUT_CONTROL not supported on this chip\n");
return 0x0000000000001150ull;
}
#else
#define CVMX_NPEI_PKT_INPUT_CONTROL (0x0000000000001150ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PKT_INSTR_ENB CVMX_NPEI_PKT_INSTR_ENB_FUNC()
static inline uint64_t CVMX_NPEI_PKT_INSTR_ENB_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PKT_INSTR_ENB not supported on this chip\n");
return 0x0000000000001000ull;
}
#else
#define CVMX_NPEI_PKT_INSTR_ENB (0x0000000000001000ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PKT_INSTR_RD_SIZE CVMX_NPEI_PKT_INSTR_RD_SIZE_FUNC()
static inline uint64_t CVMX_NPEI_PKT_INSTR_RD_SIZE_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PKT_INSTR_RD_SIZE not supported on this chip\n");
return 0x0000000000001190ull;
}
#else
#define CVMX_NPEI_PKT_INSTR_RD_SIZE (0x0000000000001190ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PKT_INSTR_SIZE CVMX_NPEI_PKT_INSTR_SIZE_FUNC()
static inline uint64_t CVMX_NPEI_PKT_INSTR_SIZE_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PKT_INSTR_SIZE not supported on this chip\n");
return 0x0000000000001020ull;
}
#else
#define CVMX_NPEI_PKT_INSTR_SIZE (0x0000000000001020ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PKT_INT_LEVELS CVMX_NPEI_PKT_INT_LEVELS_FUNC()
static inline uint64_t CVMX_NPEI_PKT_INT_LEVELS_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PKT_INT_LEVELS not supported on this chip\n");
return 0x0000000000001100ull;
}
#else
#define CVMX_NPEI_PKT_INT_LEVELS (0x0000000000001100ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PKT_IN_BP CVMX_NPEI_PKT_IN_BP_FUNC()
static inline uint64_t CVMX_NPEI_PKT_IN_BP_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PKT_IN_BP not supported on this chip\n");
return 0x00000000000006B0ull;
}
#else
#define CVMX_NPEI_PKT_IN_BP (0x00000000000006B0ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_NPEI_PKT_IN_DONEX_CNTS(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 31))) ||
(OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 31)))))
cvmx_warn("CVMX_NPEI_PKT_IN_DONEX_CNTS(%lu) is invalid on this chip\n", offset);
return 0x0000000000002000ull + ((offset) & 31) * 16;
}
#else
#define CVMX_NPEI_PKT_IN_DONEX_CNTS(offset) (0x0000000000002000ull + ((offset) & 31) * 16)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PKT_IN_INSTR_COUNTS CVMX_NPEI_PKT_IN_INSTR_COUNTS_FUNC()
static inline uint64_t CVMX_NPEI_PKT_IN_INSTR_COUNTS_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PKT_IN_INSTR_COUNTS not supported on this chip\n");
return 0x00000000000006A0ull;
}
#else
#define CVMX_NPEI_PKT_IN_INSTR_COUNTS (0x00000000000006A0ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PKT_IN_PCIE_PORT CVMX_NPEI_PKT_IN_PCIE_PORT_FUNC()
static inline uint64_t CVMX_NPEI_PKT_IN_PCIE_PORT_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PKT_IN_PCIE_PORT not supported on this chip\n");
return 0x00000000000011A0ull;
}
#else
#define CVMX_NPEI_PKT_IN_PCIE_PORT (0x00000000000011A0ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PKT_IPTR CVMX_NPEI_PKT_IPTR_FUNC()
static inline uint64_t CVMX_NPEI_PKT_IPTR_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PKT_IPTR not supported on this chip\n");
return 0x0000000000001070ull;
}
#else
#define CVMX_NPEI_PKT_IPTR (0x0000000000001070ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PKT_OUTPUT_WMARK CVMX_NPEI_PKT_OUTPUT_WMARK_FUNC()
static inline uint64_t CVMX_NPEI_PKT_OUTPUT_WMARK_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PKT_OUTPUT_WMARK not supported on this chip\n");
return 0x0000000000001160ull;
}
#else
#define CVMX_NPEI_PKT_OUTPUT_WMARK (0x0000000000001160ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PKT_OUT_BMODE CVMX_NPEI_PKT_OUT_BMODE_FUNC()
static inline uint64_t CVMX_NPEI_PKT_OUT_BMODE_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PKT_OUT_BMODE not supported on this chip\n");
return 0x00000000000010D0ull;
}
#else
#define CVMX_NPEI_PKT_OUT_BMODE (0x00000000000010D0ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PKT_OUT_ENB CVMX_NPEI_PKT_OUT_ENB_FUNC()
static inline uint64_t CVMX_NPEI_PKT_OUT_ENB_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PKT_OUT_ENB not supported on this chip\n");
return 0x0000000000001010ull;
}
#else
#define CVMX_NPEI_PKT_OUT_ENB (0x0000000000001010ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PKT_PCIE_PORT CVMX_NPEI_PKT_PCIE_PORT_FUNC()
static inline uint64_t CVMX_NPEI_PKT_PCIE_PORT_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PKT_PCIE_PORT not supported on this chip\n");
return 0x00000000000010E0ull;
}
#else
#define CVMX_NPEI_PKT_PCIE_PORT (0x00000000000010E0ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PKT_PORT_IN_RST CVMX_NPEI_PKT_PORT_IN_RST_FUNC()
static inline uint64_t CVMX_NPEI_PKT_PORT_IN_RST_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PKT_PORT_IN_RST not supported on this chip\n");
return 0x0000000000000690ull;
}
#else
#define CVMX_NPEI_PKT_PORT_IN_RST (0x0000000000000690ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PKT_SLIST_ES CVMX_NPEI_PKT_SLIST_ES_FUNC()
static inline uint64_t CVMX_NPEI_PKT_SLIST_ES_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PKT_SLIST_ES not supported on this chip\n");
return 0x0000000000001050ull;
}
#else
#define CVMX_NPEI_PKT_SLIST_ES (0x0000000000001050ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PKT_SLIST_ID_SIZE CVMX_NPEI_PKT_SLIST_ID_SIZE_FUNC()
static inline uint64_t CVMX_NPEI_PKT_SLIST_ID_SIZE_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PKT_SLIST_ID_SIZE not supported on this chip\n");
return 0x0000000000001180ull;
}
#else
#define CVMX_NPEI_PKT_SLIST_ID_SIZE (0x0000000000001180ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PKT_SLIST_NS CVMX_NPEI_PKT_SLIST_NS_FUNC()
static inline uint64_t CVMX_NPEI_PKT_SLIST_NS_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PKT_SLIST_NS not supported on this chip\n");
return 0x0000000000001040ull;
}
#else
#define CVMX_NPEI_PKT_SLIST_NS (0x0000000000001040ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PKT_SLIST_ROR CVMX_NPEI_PKT_SLIST_ROR_FUNC()
static inline uint64_t CVMX_NPEI_PKT_SLIST_ROR_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PKT_SLIST_ROR not supported on this chip\n");
return 0x0000000000001030ull;
}
#else
#define CVMX_NPEI_PKT_SLIST_ROR (0x0000000000001030ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PKT_TIME_INT CVMX_NPEI_PKT_TIME_INT_FUNC()
static inline uint64_t CVMX_NPEI_PKT_TIME_INT_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PKT_TIME_INT not supported on this chip\n");
return 0x0000000000001120ull;
}
#else
#define CVMX_NPEI_PKT_TIME_INT (0x0000000000001120ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_PKT_TIME_INT_ENB CVMX_NPEI_PKT_TIME_INT_ENB_FUNC()
static inline uint64_t CVMX_NPEI_PKT_TIME_INT_ENB_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_PKT_TIME_INT_ENB not supported on this chip\n");
return 0x0000000000001140ull;
}
#else
#define CVMX_NPEI_PKT_TIME_INT_ENB (0x0000000000001140ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_RSL_INT_BLOCKS CVMX_NPEI_RSL_INT_BLOCKS_FUNC()
static inline uint64_t CVMX_NPEI_RSL_INT_BLOCKS_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_RSL_INT_BLOCKS not supported on this chip\n");
return 0x0000000000000520ull;
}
#else
#define CVMX_NPEI_RSL_INT_BLOCKS (0x0000000000000520ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_SCRATCH_1 CVMX_NPEI_SCRATCH_1_FUNC()
static inline uint64_t CVMX_NPEI_SCRATCH_1_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_SCRATCH_1 not supported on this chip\n");
return 0x0000000000000270ull;
}
#else
#define CVMX_NPEI_SCRATCH_1 (0x0000000000000270ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_STATE1 CVMX_NPEI_STATE1_FUNC()
static inline uint64_t CVMX_NPEI_STATE1_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_STATE1 not supported on this chip\n");
return 0x0000000000000620ull;
}
#else
#define CVMX_NPEI_STATE1 (0x0000000000000620ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_STATE2 CVMX_NPEI_STATE2_FUNC()
static inline uint64_t CVMX_NPEI_STATE2_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_STATE2 not supported on this chip\n");
return 0x0000000000000630ull;
}
#else
#define CVMX_NPEI_STATE2 (0x0000000000000630ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_STATE3 CVMX_NPEI_STATE3_FUNC()
static inline uint64_t CVMX_NPEI_STATE3_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_STATE3 not supported on this chip\n");
return 0x0000000000000640ull;
}
#else
#define CVMX_NPEI_STATE3 (0x0000000000000640ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_WINDOW_CTL CVMX_NPEI_WINDOW_CTL_FUNC()
static inline uint64_t CVMX_NPEI_WINDOW_CTL_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_WINDOW_CTL not supported on this chip\n");
return 0x0000000000000380ull;
}
#else
#define CVMX_NPEI_WINDOW_CTL (0x0000000000000380ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_WIN_RD_ADDR CVMX_NPEI_WIN_RD_ADDR_FUNC()
static inline uint64_t CVMX_NPEI_WIN_RD_ADDR_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_WIN_RD_ADDR not supported on this chip\n");
return 0x0000000000000210ull;
}
#else
#define CVMX_NPEI_WIN_RD_ADDR (0x0000000000000210ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_WIN_RD_DATA CVMX_NPEI_WIN_RD_DATA_FUNC()
static inline uint64_t CVMX_NPEI_WIN_RD_DATA_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_WIN_RD_DATA not supported on this chip\n");
return 0x0000000000000240ull;
}
#else
#define CVMX_NPEI_WIN_RD_DATA (0x0000000000000240ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_WIN_WR_ADDR CVMX_NPEI_WIN_WR_ADDR_FUNC()
static inline uint64_t CVMX_NPEI_WIN_WR_ADDR_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_WIN_WR_ADDR not supported on this chip\n");
return 0x0000000000000200ull;
}
#else
#define CVMX_NPEI_WIN_WR_ADDR (0x0000000000000200ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_WIN_WR_DATA CVMX_NPEI_WIN_WR_DATA_FUNC()
static inline uint64_t CVMX_NPEI_WIN_WR_DATA_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_WIN_WR_DATA not supported on this chip\n");
return 0x0000000000000220ull;
}
#else
#define CVMX_NPEI_WIN_WR_DATA (0x0000000000000220ull)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
#define CVMX_NPEI_WIN_WR_MASK CVMX_NPEI_WIN_WR_MASK_FUNC()
static inline uint64_t CVMX_NPEI_WIN_WR_MASK_FUNC(void)
{
if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX)))
cvmx_warn("CVMX_NPEI_WIN_WR_MASK not supported on this chip\n");
return 0x0000000000000230ull;
}
#else
#define CVMX_NPEI_WIN_WR_MASK (0x0000000000000230ull)
#endif
/**
* cvmx_npei_bar1_index#
*
* Total Address is 16Kb; 0x0000 - 0x3fff, 0x000 - 0x7fe(Reg, every other 8B)
*
* General 5kb; 0x0000 - 0x13ff, 0x000 - 0x27e(Reg-General)
* PktMem 10Kb; 0x1400 - 0x3bff, 0x280 - 0x77e(Reg-General-Packet)
* Rsvd 1Kb; 0x3c00 - 0x3fff, 0x780 - 0x7fe(Reg-NCB Only Mode)
* == NPEI_PKT_CNT_INT_ENB[PORT]
* == NPEI_PKT_TIME_INT_ENB[PORT]
* == NPEI_PKT_CNT_INT[PORT]
* == NPEI_PKT_TIME_INT[PORT]
* == NPEI_PKT_PCIE_PORT[PP]
* == NPEI_PKT_SLIST_ROR[ROR]
* == NPEI_PKT_SLIST_ROR[NSR] ?
* == NPEI_PKT_SLIST_ES[ES]
* == NPEI_PKTn_SLIST_BAOFF_DBELL[AOFF]
* == NPEI_PKTn_SLIST_BAOFF_DBELL[DBELL]
* == NPEI_PKTn_CNTS[CNT]
* NPEI_CTL_STATUS[OUTn_ENB] == NPEI_PKT_OUT_ENB[ENB]
* NPEI_BASE_ADDRESS_OUTPUTn[BADDR] == NPEI_PKTn_SLIST_BADDR[ADDR]
* NPEI_DESC_OUTPUTn[SIZE] == NPEI_PKTn_SLIST_FIFO_RSIZE[RSIZE]
* NPEI_Pn_DBPAIR_ADDR[NADDR] == NPEI_PKTn_SLIST_BADDR[ADDR] + NPEI_PKTn_SLIST_BAOFF_DBELL[AOFF]
* NPEI_PKT_CREDITSn[PTR_CNT] == NPEI_PKTn_SLIST_BAOFF_DBELL[DBELL]
* NPEI_P0_PAIR_CNTS[AVAIL] == NPEI_PKTn_SLIST_BAOFF_DBELL[DBELL]
* NPEI_P0_PAIR_CNTS[FCNT] ==
* NPEI_PKTS_SENTn[PKT_CNT] == NPEI_PKTn_CNTS[CNT]
* NPEI_OUTPUT_CONTROL[Pn_BMODE] == NPEI_PKT_OUT_BMODE[BMODE]
* NPEI_PKT_CREDITSn[PKT_CNT] == NPEI_PKTn_CNTS[CNT]
* NPEI_BUFF_SIZE_OUTPUTn[BSIZE] == NPEI_PKT_SLIST_ID_SIZE[BSIZE]
* NPEI_BUFF_SIZE_OUTPUTn[ISIZE] == NPEI_PKT_SLIST_ID_SIZE[ISIZE]
* NPEI_OUTPUT_CONTROL[On_CSRM] == NPEI_PKT_DPADDR[DPTR] & NPEI_PKT_OUT_USE_IPTR[PORT]
* NPEI_OUTPUT_CONTROL[On_ES] == NPEI_PKT_DATA_OUT_ES[ES]
* NPEI_OUTPUT_CONTROL[On_NS] == NPEI_PKT_DATA_OUT_NS[NSR] ?
* NPEI_OUTPUT_CONTROL[On_RO] == NPEI_PKT_DATA_OUT_ROR[ROR]
* NPEI_PKTS_SENT_INT_LEVn[PKT_CNT] == NPEI_PKT_INT_LEVELS[CNT]
* NPEI_PKTS_SENT_TIMEn[PKT_TIME] == NPEI_PKT_INT_LEVELS[TIME]
* NPEI_OUTPUT_CONTROL[IPTR_On] == NPEI_PKT_IPTR[IPTR]
* NPEI_PCIE_PORT_OUTPUT[] == NPEI_PKT_PCIE_PORT[PP]
*
* NPEI_BAR1_INDEXX = NPEI BAR1 IndexX Register
*
* Contains address index and control bits for access to memory ranges of BAR-1. Index is build from supplied address [25:22].
* NPEI_BAR1_INDEX0 through NPEI_BAR1_INDEX15 is used for transactions orginating with PCIE-PORT0 and NPEI_BAR1_INDEX16
* through NPEI_BAR1_INDEX31 is used for transactions originating with PCIE-PORT1.
*/
union cvmx_npei_bar1_indexx
{
uint32_t u32;
struct cvmx_npei_bar1_indexx_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint32_t reserved_18_31 : 14;
uint32_t addr_idx : 14; /**< Address bits [35:22] sent to L2C */
uint32_t ca : 1; /**< Set '1' when access is not to be cached in L2. */
uint32_t end_swp : 2; /**< Endian Swap Mode */
uint32_t addr_v : 1; /**< Set '1' when the selected address range is valid. */
#else
uint32_t addr_v : 1;
uint32_t end_swp : 2;
uint32_t ca : 1;
uint32_t addr_idx : 14;
uint32_t reserved_18_31 : 14;
#endif
} s;
struct cvmx_npei_bar1_indexx_s cn52xx;
struct cvmx_npei_bar1_indexx_s cn52xxp1;
struct cvmx_npei_bar1_indexx_s cn56xx;
struct cvmx_npei_bar1_indexx_s cn56xxp1;
};
typedef union cvmx_npei_bar1_indexx cvmx_npei_bar1_indexx_t;
/**
* cvmx_npei_bist_status
*
* NPEI_BIST_STATUS = NPI's BIST Status Register
*
* Results from BIST runs of NPEI's memories.
*/
union cvmx_npei_bist_status
{
uint64_t u64;
struct cvmx_npei_bist_status_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t pkt_rdf : 1; /**< BIST Status for PKT Read FIFO */
uint64_t reserved_60_62 : 3;
uint64_t pcr_gim : 1; /**< BIST Status for PKT Gather Instr MEM */
uint64_t pkt_pif : 1; /**< BIST Status for PKT INB FIFO */
uint64_t pcsr_int : 1; /**< BIST Status for PKT pout_int_bstatus */
uint64_t pcsr_im : 1; /**< BIST Status for PKT pcsr_instr_mem_bstatus */
uint64_t pcsr_cnt : 1; /**< BIST Status for PKT pin_cnt_bstatus */
uint64_t pcsr_id : 1; /**< BIST Status for PKT pcsr_in_done_bstatus */
uint64_t pcsr_sl : 1; /**< BIST Status for PKT pcsr_slist_bstatus */
uint64_t reserved_50_52 : 3;
uint64_t pkt_ind : 1; /**< BIST Status for PKT Instruction Done MEM */
uint64_t pkt_slm : 1; /**< BIST Status for PKT SList MEM */
uint64_t reserved_36_47 : 12;
uint64_t d0_pst : 1; /**< BIST Status for DMA0 Pcie Store */
uint64_t d1_pst : 1; /**< BIST Status for DMA1 Pcie Store */
uint64_t d2_pst : 1; /**< BIST Status for DMA2 Pcie Store */
uint64_t d3_pst : 1; /**< BIST Status for DMA3 Pcie Store */
uint64_t reserved_31_31 : 1;
uint64_t n2p0_c : 1; /**< BIST Status for N2P Port0 Cmd */
uint64_t n2p0_o : 1; /**< BIST Status for N2P Port0 Data */
uint64_t n2p1_c : 1; /**< BIST Status for N2P Port1 Cmd */
uint64_t n2p1_o : 1; /**< BIST Status for N2P Port1 Data */
uint64_t cpl_p0 : 1; /**< BIST Status for CPL Port 0 */
uint64_t cpl_p1 : 1; /**< BIST Status for CPL Port 1 */
uint64_t p2n1_po : 1; /**< BIST Status for P2N Port1 P Order */
uint64_t p2n1_no : 1; /**< BIST Status for P2N Port1 N Order */
uint64_t p2n1_co : 1; /**< BIST Status for P2N Port1 C Order */
uint64_t p2n0_po : 1; /**< BIST Status for P2N Port0 P Order */
uint64_t p2n0_no : 1; /**< BIST Status for P2N Port0 N Order */
uint64_t p2n0_co : 1; /**< BIST Status for P2N Port0 C Order */
uint64_t p2n0_c0 : 1; /**< BIST Status for P2N Port0 C0 */
uint64_t p2n0_c1 : 1; /**< BIST Status for P2N Port0 C1 */
uint64_t p2n0_n : 1; /**< BIST Status for P2N Port0 N */
uint64_t p2n0_p0 : 1; /**< BIST Status for P2N Port0 P0 */
uint64_t p2n0_p1 : 1; /**< BIST Status for P2N Port0 P1 */
uint64_t p2n1_c0 : 1; /**< BIST Status for P2N Port1 C0 */
uint64_t p2n1_c1 : 1; /**< BIST Status for P2N Port1 C1 */
uint64_t p2n1_n : 1; /**< BIST Status for P2N Port1 N */
uint64_t p2n1_p0 : 1; /**< BIST Status for P2N Port1 P0 */
uint64_t p2n1_p1 : 1; /**< BIST Status for P2N Port1 P1 */
uint64_t csm0 : 1; /**< BIST Status for CSM0 */
uint64_t csm1 : 1; /**< BIST Status for CSM1 */
uint64_t dif0 : 1; /**< BIST Status for DMA Instr0 */
uint64_t dif1 : 1; /**< BIST Status for DMA Instr0 */
uint64_t dif2 : 1; /**< BIST Status for DMA Instr0 */
uint64_t dif3 : 1; /**< BIST Status for DMA Instr0 */
uint64_t reserved_2_2 : 1;
uint64_t msi : 1; /**< BIST Status for MSI Memory Map */
uint64_t ncb_cmd : 1; /**< BIST Status for NCB Outbound Commands */
#else
uint64_t ncb_cmd : 1;
uint64_t msi : 1;
uint64_t reserved_2_2 : 1;
uint64_t dif3 : 1;
uint64_t dif2 : 1;
uint64_t dif1 : 1;
uint64_t dif0 : 1;
uint64_t csm1 : 1;
uint64_t csm0 : 1;
uint64_t p2n1_p1 : 1;
uint64_t p2n1_p0 : 1;
uint64_t p2n1_n : 1;
uint64_t p2n1_c1 : 1;
uint64_t p2n1_c0 : 1;
uint64_t p2n0_p1 : 1;
uint64_t p2n0_p0 : 1;
uint64_t p2n0_n : 1;
uint64_t p2n0_c1 : 1;
uint64_t p2n0_c0 : 1;
uint64_t p2n0_co : 1;
uint64_t p2n0_no : 1;
uint64_t p2n0_po : 1;
uint64_t p2n1_co : 1;
uint64_t p2n1_no : 1;
uint64_t p2n1_po : 1;
uint64_t cpl_p1 : 1;
uint64_t cpl_p0 : 1;
uint64_t n2p1_o : 1;
uint64_t n2p1_c : 1;
uint64_t n2p0_o : 1;
uint64_t n2p0_c : 1;
uint64_t reserved_31_31 : 1;
uint64_t d3_pst : 1;
uint64_t d2_pst : 1;
uint64_t d1_pst : 1;
uint64_t d0_pst : 1;
uint64_t reserved_36_47 : 12;
uint64_t pkt_slm : 1;
uint64_t pkt_ind : 1;
uint64_t reserved_50_52 : 3;
uint64_t pcsr_sl : 1;
uint64_t pcsr_id : 1;
uint64_t pcsr_cnt : 1;
uint64_t pcsr_im : 1;
uint64_t pcsr_int : 1;
uint64_t pkt_pif : 1;
uint64_t pcr_gim : 1;
uint64_t reserved_60_62 : 3;
uint64_t pkt_rdf : 1;
#endif
} s;
struct cvmx_npei_bist_status_cn52xx
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t pkt_rdf : 1; /**< BIST Status for PKT Read FIFO */
uint64_t reserved_60_62 : 3;
uint64_t pcr_gim : 1; /**< BIST Status for PKT Gather Instr MEM */
uint64_t pkt_pif : 1; /**< BIST Status for PKT INB FIFO */
uint64_t pcsr_int : 1; /**< BIST Status for PKT OUTB Interrupt MEM */
uint64_t pcsr_im : 1; /**< BIST Status for PKT CSR Instr MEM */
uint64_t pcsr_cnt : 1; /**< BIST Status for PKT INB Count MEM */
uint64_t pcsr_id : 1; /**< BIST Status for PKT INB Instr Done MEM */
uint64_t pcsr_sl : 1; /**< BIST Status for PKT OUTB SLIST MEM */
uint64_t pkt_imem : 1; /**< BIST Status for PKT OUTB IFIFO */
uint64_t pkt_pfm : 1; /**< BIST Status for PKT Front MEM */
uint64_t pkt_pof : 1; /**< BIST Status for PKT OUTB FIFO */
uint64_t reserved_48_49 : 2;
uint64_t pkt_pop0 : 1; /**< BIST Status for PKT OUTB Slist0 */
uint64_t pkt_pop1 : 1; /**< BIST Status for PKT OUTB Slist1 */
uint64_t d0_mem : 1; /**< BIST Status for DMA MEM 0 */
uint64_t d1_mem : 1; /**< BIST Status for DMA MEM 1 */
uint64_t d2_mem : 1; /**< BIST Status for DMA MEM 2 */
uint64_t d3_mem : 1; /**< BIST Status for DMA MEM 3 */
uint64_t d4_mem : 1; /**< BIST Status for DMA MEM 4 */
uint64_t ds_mem : 1; /**< BIST Status for DMA Memory */
uint64_t reserved_36_39 : 4;
uint64_t d0_pst : 1; /**< BIST Status for DMA0 Pcie Store */
uint64_t d1_pst : 1; /**< BIST Status for DMA1 Pcie Store */
uint64_t d2_pst : 1; /**< BIST Status for DMA2 Pcie Store */
uint64_t d3_pst : 1; /**< BIST Status for DMA3 Pcie Store */
uint64_t d4_pst : 1; /**< BIST Status for DMA4 Pcie Store */
uint64_t n2p0_c : 1; /**< BIST Status for N2P Port0 Cmd */
uint64_t n2p0_o : 1; /**< BIST Status for N2P Port0 Data */
uint64_t n2p1_c : 1; /**< BIST Status for N2P Port1 Cmd */
uint64_t n2p1_o : 1; /**< BIST Status for N2P Port1 Data */
uint64_t cpl_p0 : 1; /**< BIST Status for CPL Port 0 */
uint64_t cpl_p1 : 1; /**< BIST Status for CPL Port 1 */
uint64_t p2n1_po : 1; /**< BIST Status for P2N Port1 P Order */
uint64_t p2n1_no : 1; /**< BIST Status for P2N Port1 N Order */
uint64_t p2n1_co : 1; /**< BIST Status for P2N Port1 C Order */
uint64_t p2n0_po : 1; /**< BIST Status for P2N Port0 P Order */
uint64_t p2n0_no : 1; /**< BIST Status for P2N Port0 N Order */
uint64_t p2n0_co : 1; /**< BIST Status for P2N Port0 C Order */
uint64_t p2n0_c0 : 1; /**< BIST Status for P2N Port0 C0 */
uint64_t p2n0_c1 : 1; /**< BIST Status for P2N Port0 C1 */
uint64_t p2n0_n : 1; /**< BIST Status for P2N Port0 N */
uint64_t p2n0_p0 : 1; /**< BIST Status for P2N Port0 P0 */
uint64_t p2n0_p1 : 1; /**< BIST Status for P2N Port0 P1 */
uint64_t p2n1_c0 : 1; /**< BIST Status for P2N Port1 C0 */
uint64_t p2n1_c1 : 1; /**< BIST Status for P2N Port1 C1 */
uint64_t p2n1_n : 1; /**< BIST Status for P2N Port1 N */
uint64_t p2n1_p0 : 1; /**< BIST Status for P2N Port1 P0 */
uint64_t p2n1_p1 : 1; /**< BIST Status for P2N Port1 P1 */
uint64_t csm0 : 1; /**< BIST Status for CSM0 */
uint64_t csm1 : 1; /**< BIST Status for CSM1 */
uint64_t dif0 : 1; /**< BIST Status for DMA Instr0 */
uint64_t dif1 : 1; /**< BIST Status for DMA Instr0 */
uint64_t dif2 : 1; /**< BIST Status for DMA Instr0 */
uint64_t dif3 : 1; /**< BIST Status for DMA Instr0 */
uint64_t dif4 : 1; /**< BIST Status for DMA Instr0 */
uint64_t msi : 1; /**< BIST Status for MSI Memory Map */
uint64_t ncb_cmd : 1; /**< BIST Status for NCB Outbound Commands */
#else
uint64_t ncb_cmd : 1;
uint64_t msi : 1;
uint64_t dif4 : 1;
uint64_t dif3 : 1;
uint64_t dif2 : 1;
uint64_t dif1 : 1;
uint64_t dif0 : 1;
uint64_t csm1 : 1;
uint64_t csm0 : 1;
uint64_t p2n1_p1 : 1;
uint64_t p2n1_p0 : 1;
uint64_t p2n1_n : 1;
uint64_t p2n1_c1 : 1;
uint64_t p2n1_c0 : 1;
uint64_t p2n0_p1 : 1;
uint64_t p2n0_p0 : 1;
uint64_t p2n0_n : 1;
uint64_t p2n0_c1 : 1;
uint64_t p2n0_c0 : 1;
uint64_t p2n0_co : 1;
uint64_t p2n0_no : 1;
uint64_t p2n0_po : 1;
uint64_t p2n1_co : 1;
uint64_t p2n1_no : 1;
uint64_t p2n1_po : 1;
uint64_t cpl_p1 : 1;
uint64_t cpl_p0 : 1;
uint64_t n2p1_o : 1;
uint64_t n2p1_c : 1;
uint64_t n2p0_o : 1;
uint64_t n2p0_c : 1;
uint64_t d4_pst : 1;
uint64_t d3_pst : 1;
uint64_t d2_pst : 1;
uint64_t d1_pst : 1;
uint64_t d0_pst : 1;
uint64_t reserved_36_39 : 4;
uint64_t ds_mem : 1;
uint64_t d4_mem : 1;
uint64_t d3_mem : 1;
uint64_t d2_mem : 1;
uint64_t d1_mem : 1;
uint64_t d0_mem : 1;
uint64_t pkt_pop1 : 1;
uint64_t pkt_pop0 : 1;
uint64_t reserved_48_49 : 2;
uint64_t pkt_pof : 1;
uint64_t pkt_pfm : 1;
uint64_t pkt_imem : 1;
uint64_t pcsr_sl : 1;
uint64_t pcsr_id : 1;
uint64_t pcsr_cnt : 1;
uint64_t pcsr_im : 1;
uint64_t pcsr_int : 1;
uint64_t pkt_pif : 1;
uint64_t pcr_gim : 1;
uint64_t reserved_60_62 : 3;
uint64_t pkt_rdf : 1;
#endif
} cn52xx;
struct cvmx_npei_bist_status_cn52xxp1
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_46_63 : 18;
uint64_t d0_mem0 : 1; /**< BIST Status for DMA0 Memory */
uint64_t d1_mem1 : 1; /**< BIST Status for DMA1 Memory */
uint64_t d2_mem2 : 1; /**< BIST Status for DMA2 Memory */
uint64_t d3_mem3 : 1; /**< BIST Status for DMA3 Memory */
uint64_t dr0_mem : 1; /**< BIST Status for DMA0 Store */
uint64_t d0_mem : 1; /**< BIST Status for DMA0 Memory */
uint64_t d1_mem : 1; /**< BIST Status for DMA1 Memory */
uint64_t d2_mem : 1; /**< BIST Status for DMA2 Memory */
uint64_t d3_mem : 1; /**< BIST Status for DMA3 Memory */
uint64_t dr1_mem : 1; /**< BIST Status for DMA1 Store */
uint64_t d0_pst : 1; /**< BIST Status for DMA0 Pcie Store */
uint64_t d1_pst : 1; /**< BIST Status for DMA1 Pcie Store */
uint64_t d2_pst : 1; /**< BIST Status for DMA2 Pcie Store */
uint64_t d3_pst : 1; /**< BIST Status for DMA3 Pcie Store */
uint64_t dr2_mem : 1; /**< BIST Status for DMA2 Store */
uint64_t n2p0_c : 1; /**< BIST Status for N2P Port0 Cmd */
uint64_t n2p0_o : 1; /**< BIST Status for N2P Port0 Data */
uint64_t n2p1_c : 1; /**< BIST Status for N2P Port1 Cmd */
uint64_t n2p1_o : 1; /**< BIST Status for N2P Port1 Data */
uint64_t cpl_p0 : 1; /**< BIST Status for CPL Port 0 */
uint64_t cpl_p1 : 1; /**< BIST Status for CPL Port 1 */
uint64_t p2n1_po : 1; /**< BIST Status for P2N Port1 P Order */
uint64_t p2n1_no : 1; /**< BIST Status for P2N Port1 N Order */
uint64_t p2n1_co : 1; /**< BIST Status for P2N Port1 C Order */
uint64_t p2n0_po : 1; /**< BIST Status for P2N Port0 P Order */
uint64_t p2n0_no : 1; /**< BIST Status for P2N Port0 N Order */
uint64_t p2n0_co : 1; /**< BIST Status for P2N Port0 C Order */
uint64_t p2n0_c0 : 1; /**< BIST Status for P2N Port0 C0 */
uint64_t p2n0_c1 : 1; /**< BIST Status for P2N Port0 C1 */
uint64_t p2n0_n : 1; /**< BIST Status for P2N Port0 N */
uint64_t p2n0_p0 : 1; /**< BIST Status for P2N Port0 P0 */
uint64_t p2n0_p1 : 1; /**< BIST Status for P2N Port0 P1 */
uint64_t p2n1_c0 : 1; /**< BIST Status for P2N Port1 C0 */
uint64_t p2n1_c1 : 1; /**< BIST Status for P2N Port1 C1 */
uint64_t p2n1_n : 1; /**< BIST Status for P2N Port1 N */
uint64_t p2n1_p0 : 1; /**< BIST Status for P2N Port1 P0 */
uint64_t p2n1_p1 : 1; /**< BIST Status for P2N Port1 P1 */
uint64_t csm0 : 1; /**< BIST Status for CSM0 */
uint64_t csm1 : 1; /**< BIST Status for CSM1 */
uint64_t dif0 : 1; /**< BIST Status for DMA Instr0 */
uint64_t dif1 : 1; /**< BIST Status for DMA Instr0 */
uint64_t dif2 : 1; /**< BIST Status for DMA Instr0 */
uint64_t dif3 : 1; /**< BIST Status for DMA Instr0 */
uint64_t dr3_mem : 1; /**< BIST Status for DMA3 Store */
uint64_t msi : 1; /**< BIST Status for MSI Memory Map */
uint64_t ncb_cmd : 1; /**< BIST Status for NCB Outbound Commands */
#else
uint64_t ncb_cmd : 1;
uint64_t msi : 1;
uint64_t dr3_mem : 1;
uint64_t dif3 : 1;
uint64_t dif2 : 1;
uint64_t dif1 : 1;
uint64_t dif0 : 1;
uint64_t csm1 : 1;
uint64_t csm0 : 1;
uint64_t p2n1_p1 : 1;
uint64_t p2n1_p0 : 1;
uint64_t p2n1_n : 1;
uint64_t p2n1_c1 : 1;
uint64_t p2n1_c0 : 1;
uint64_t p2n0_p1 : 1;
uint64_t p2n0_p0 : 1;
uint64_t p2n0_n : 1;
uint64_t p2n0_c1 : 1;
uint64_t p2n0_c0 : 1;
uint64_t p2n0_co : 1;
uint64_t p2n0_no : 1;
uint64_t p2n0_po : 1;
uint64_t p2n1_co : 1;
uint64_t p2n1_no : 1;
uint64_t p2n1_po : 1;
uint64_t cpl_p1 : 1;
uint64_t cpl_p0 : 1;
uint64_t n2p1_o : 1;
uint64_t n2p1_c : 1;
uint64_t n2p0_o : 1;
uint64_t n2p0_c : 1;
uint64_t dr2_mem : 1;
uint64_t d3_pst : 1;
uint64_t d2_pst : 1;
uint64_t d1_pst : 1;
uint64_t d0_pst : 1;
uint64_t dr1_mem : 1;
uint64_t d3_mem : 1;
uint64_t d2_mem : 1;
uint64_t d1_mem : 1;
uint64_t d0_mem : 1;
uint64_t dr0_mem : 1;
uint64_t d3_mem3 : 1;
uint64_t d2_mem2 : 1;
uint64_t d1_mem1 : 1;
uint64_t d0_mem0 : 1;
uint64_t reserved_46_63 : 18;
#endif
} cn52xxp1;
struct cvmx_npei_bist_status_cn52xx cn56xx;
struct cvmx_npei_bist_status_cn56xxp1
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_58_63 : 6;
uint64_t pcsr_int : 1; /**< BIST Status for PKT pout_int_bstatus */
uint64_t pcsr_im : 1; /**< BIST Status for PKT pcsr_instr_mem_bstatus */
uint64_t pcsr_cnt : 1; /**< BIST Status for PKT pin_cnt_bstatus */
uint64_t pcsr_id : 1; /**< BIST Status for PKT pcsr_in_done_bstatus */
uint64_t pcsr_sl : 1; /**< BIST Status for PKT pcsr_slist_bstatus */
uint64_t pkt_pout : 1; /**< BIST Status for PKT OUT Count MEM */
uint64_t pkt_imem : 1; /**< BIST Status for PKT Instruction MEM */
uint64_t pkt_cntm : 1; /**< BIST Status for PKT Count MEM */
uint64_t pkt_ind : 1; /**< BIST Status for PKT Instruction Done MEM */
uint64_t pkt_slm : 1; /**< BIST Status for PKT SList MEM */
uint64_t pkt_odf : 1; /**< BIST Status for PKT Output Data FIFO */
uint64_t pkt_oif : 1; /**< BIST Status for PKT Output INFO FIFO */
uint64_t pkt_out : 1; /**< BIST Status for PKT Output FIFO */
uint64_t pkt_i0 : 1; /**< BIST Status for PKT Instr0 */
uint64_t pkt_i1 : 1; /**< BIST Status for PKT Instr1 */
uint64_t pkt_s0 : 1; /**< BIST Status for PKT Slist0 */
uint64_t pkt_s1 : 1; /**< BIST Status for PKT Slist1 */
uint64_t d0_mem : 1; /**< BIST Status for DMA0 Memory */
uint64_t d1_mem : 1; /**< BIST Status for DMA1 Memory */
uint64_t d2_mem : 1; /**< BIST Status for DMA2 Memory */
uint64_t d3_mem : 1; /**< BIST Status for DMA3 Memory */
uint64_t d4_mem : 1; /**< BIST Status for DMA4 Memory */
uint64_t d0_pst : 1; /**< BIST Status for DMA0 Pcie Store */
uint64_t d1_pst : 1; /**< BIST Status for DMA1 Pcie Store */
uint64_t d2_pst : 1; /**< BIST Status for DMA2 Pcie Store */
uint64_t d3_pst : 1; /**< BIST Status for DMA3 Pcie Store */
uint64_t d4_pst : 1; /**< BIST Status for DMA4 Pcie Store */
uint64_t n2p0_c : 1; /**< BIST Status for N2P Port0 Cmd */
uint64_t n2p0_o : 1; /**< BIST Status for N2P Port0 Data */
uint64_t n2p1_c : 1; /**< BIST Status for N2P Port1 Cmd */
uint64_t n2p1_o : 1; /**< BIST Status for N2P Port1 Data */
uint64_t cpl_p0 : 1; /**< BIST Status for CPL Port 0 */
uint64_t cpl_p1 : 1; /**< BIST Status for CPL Port 1 */
uint64_t p2n1_po : 1; /**< BIST Status for P2N Port1 P Order */
uint64_t p2n1_no : 1; /**< BIST Status for P2N Port1 N Order */
uint64_t p2n1_co : 1; /**< BIST Status for P2N Port1 C Order */
uint64_t p2n0_po : 1; /**< BIST Status for P2N Port0 P Order */
uint64_t p2n0_no : 1; /**< BIST Status for P2N Port0 N Order */
uint64_t p2n0_co : 1; /**< BIST Status for P2N Port0 C Order */
uint64_t p2n0_c0 : 1; /**< BIST Status for P2N Port0 C0 */
uint64_t p2n0_c1 : 1; /**< BIST Status for P2N Port0 C1 */
uint64_t p2n0_n : 1; /**< BIST Status for P2N Port0 N */
uint64_t p2n0_p0 : 1; /**< BIST Status for P2N Port0 P0 */
uint64_t p2n0_p1 : 1; /**< BIST Status for P2N Port0 P1 */
uint64_t p2n1_c0 : 1; /**< BIST Status for P2N Port1 C0 */
uint64_t p2n1_c1 : 1; /**< BIST Status for P2N Port1 C1 */
uint64_t p2n1_n : 1; /**< BIST Status for P2N Port1 N */
uint64_t p2n1_p0 : 1; /**< BIST Status for P2N Port1 P0 */
uint64_t p2n1_p1 : 1; /**< BIST Status for P2N Port1 P1 */
uint64_t csm0 : 1; /**< BIST Status for CSM0 */
uint64_t csm1 : 1; /**< BIST Status for CSM1 */
uint64_t dif0 : 1; /**< BIST Status for DMA Instr0 */
uint64_t dif1 : 1; /**< BIST Status for DMA Instr0 */
uint64_t dif2 : 1; /**< BIST Status for DMA Instr0 */
uint64_t dif3 : 1; /**< BIST Status for DMA Instr0 */
uint64_t dif4 : 1; /**< BIST Status for DMA Instr0 */
uint64_t msi : 1; /**< BIST Status for MSI Memory Map */
uint64_t ncb_cmd : 1; /**< BIST Status for NCB Outbound Commands */
#else
uint64_t ncb_cmd : 1;
uint64_t msi : 1;
uint64_t dif4 : 1;
uint64_t dif3 : 1;
uint64_t dif2 : 1;
uint64_t dif1 : 1;
uint64_t dif0 : 1;
uint64_t csm1 : 1;
uint64_t csm0 : 1;
uint64_t p2n1_p1 : 1;
uint64_t p2n1_p0 : 1;
uint64_t p2n1_n : 1;
uint64_t p2n1_c1 : 1;
uint64_t p2n1_c0 : 1;
uint64_t p2n0_p1 : 1;
uint64_t p2n0_p0 : 1;
uint64_t p2n0_n : 1;
uint64_t p2n0_c1 : 1;
uint64_t p2n0_c0 : 1;
uint64_t p2n0_co : 1;
uint64_t p2n0_no : 1;
uint64_t p2n0_po : 1;
uint64_t p2n1_co : 1;
uint64_t p2n1_no : 1;
uint64_t p2n1_po : 1;
uint64_t cpl_p1 : 1;
uint64_t cpl_p0 : 1;
uint64_t n2p1_o : 1;
uint64_t n2p1_c : 1;
uint64_t n2p0_o : 1;
uint64_t n2p0_c : 1;
uint64_t d4_pst : 1;
uint64_t d3_pst : 1;
uint64_t d2_pst : 1;
uint64_t d1_pst : 1;
uint64_t d0_pst : 1;
uint64_t d4_mem : 1;
uint64_t d3_mem : 1;
uint64_t d2_mem : 1;
uint64_t d1_mem : 1;
uint64_t d0_mem : 1;
uint64_t pkt_s1 : 1;
uint64_t pkt_s0 : 1;
uint64_t pkt_i1 : 1;
uint64_t pkt_i0 : 1;
uint64_t pkt_out : 1;
uint64_t pkt_oif : 1;
uint64_t pkt_odf : 1;
uint64_t pkt_slm : 1;
uint64_t pkt_ind : 1;
uint64_t pkt_cntm : 1;
uint64_t pkt_imem : 1;
uint64_t pkt_pout : 1;
uint64_t pcsr_sl : 1;
uint64_t pcsr_id : 1;
uint64_t pcsr_cnt : 1;
uint64_t pcsr_im : 1;
uint64_t pcsr_int : 1;
uint64_t reserved_58_63 : 6;
#endif
} cn56xxp1;
};
typedef union cvmx_npei_bist_status cvmx_npei_bist_status_t;
/**
* cvmx_npei_bist_status2
*
* NPEI_BIST_STATUS2 = NPI's BIST Status Register2
*
* Results from BIST runs of NPEI's memories.
*/
union cvmx_npei_bist_status2
{
uint64_t u64;
struct cvmx_npei_bist_status2_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_14_63 : 50;
uint64_t prd_tag : 1; /**< BIST Status for DMA PCIE RD Tag MEM */
uint64_t prd_st0 : 1; /**< BIST Status for DMA PCIE RD state MEM 0 */
uint64_t prd_st1 : 1; /**< BIST Status for DMA PCIE RD state MEM 1 */
uint64_t prd_err : 1; /**< BIST Status for DMA PCIE RD ERR state MEM */
uint64_t nrd_st : 1; /**< BIST Status for DMA L2C RD state MEM */
uint64_t nwe_st : 1; /**< BIST Status for DMA L2C WR state MEM */
uint64_t nwe_wr0 : 1; /**< BIST Status for DMA L2C WR MEM 0 */
uint64_t nwe_wr1 : 1; /**< BIST Status for DMA L2C WR MEM 1 */
uint64_t pkt_rd : 1; /**< BIST Status for Inbound PKT MEM */
uint64_t psc_p0 : 1; /**< BIST Status for PSC TLP 0 MEM */
uint64_t psc_p1 : 1; /**< BIST Status for PSC TLP 1 MEM */
uint64_t pkt_gd : 1; /**< BIST Status for PKT OUTB Gather Data FIFO */
uint64_t pkt_gl : 1; /**< BIST Status for PKT_OUTB Gather List FIFO */
uint64_t pkt_blk : 1; /**< BIST Status for PKT OUTB Blocked FIFO */
#else
uint64_t pkt_blk : 1;
uint64_t pkt_gl : 1;
uint64_t pkt_gd : 1;
uint64_t psc_p1 : 1;
uint64_t psc_p0 : 1;
uint64_t pkt_rd : 1;
uint64_t nwe_wr1 : 1;
uint64_t nwe_wr0 : 1;
uint64_t nwe_st : 1;
uint64_t nrd_st : 1;
uint64_t prd_err : 1;
uint64_t prd_st1 : 1;
uint64_t prd_st0 : 1;
uint64_t prd_tag : 1;
uint64_t reserved_14_63 : 50;
#endif
} s;
struct cvmx_npei_bist_status2_s cn52xx;
struct cvmx_npei_bist_status2_s cn56xx;
};
typedef union cvmx_npei_bist_status2 cvmx_npei_bist_status2_t;
/**
* cvmx_npei_ctl_port0
*
* NPEI_CTL_PORT0 = NPEI's Control Port 0
*
* Contains control for access for Port0
*/
union cvmx_npei_ctl_port0
{
uint64_t u64;
struct cvmx_npei_ctl_port0_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_21_63 : 43;
uint64_t waitl_com : 1; /**< When set '1' casues the NPI to wait for a commit
from the L2C before sending additional completions
to the L2C from the PCIe.
Set this for more conservative behavior. Clear
this for more aggressive, higher-performance
behavior */
uint64_t intd : 1; /**< When '0' Intd wire asserted. Before mapping. */
uint64_t intc : 1; /**< When '0' Intc wire asserted. Before mapping. */
uint64_t intb : 1; /**< When '0' Intb wire asserted. Before mapping. */
uint64_t inta : 1; /**< When '0' Inta wire asserted. Before mapping. */
uint64_t intd_map : 2; /**< Maps INTD to INTA(00), INTB(01), INTC(10) or
INTD (11). */
uint64_t intc_map : 2; /**< Maps INTC to INTA(00), INTB(01), INTC(10) or
INTD (11). */
uint64_t intb_map : 2; /**< Maps INTB to INTA(00), INTB(01), INTC(10) or
INTD (11). */
uint64_t inta_map : 2; /**< Maps INTA to INTA(00), INTB(01), INTC(10) or
INTD (11). */
uint64_t ctlp_ro : 1; /**< Relaxed ordering enable for Completion TLPS. */
uint64_t reserved_6_6 : 1;
uint64_t ptlp_ro : 1; /**< Relaxed ordering enable for Posted TLPS. */
uint64_t bar2_enb : 1; /**< When set '1' BAR2 is enable and will respond when
clear '0' BAR2 access will cause UR responses. */
uint64_t bar2_esx : 2; /**< Value will be XORed with pci-address[37:36] to
determine the endian swap mode. */
uint64_t bar2_cax : 1; /**< Value will be XORed with pcie-address[38] to
determine the L2 cache attribute.
Not cached in L2 if XOR result is 1 */
uint64_t wait_com : 1; /**< When set '1' casues the NPI to wait for a commit
from the L2C before sending additional stores to
the L2C from the PCIe.
Most applications will not notice a difference, so
should not set this bit. Setting the bit is more
conservative on ordering, lower performance */
#else
uint64_t wait_com : 1;
uint64_t bar2_cax : 1;
uint64_t bar2_esx : 2;
uint64_t bar2_enb : 1;
uint64_t ptlp_ro : 1;
uint64_t reserved_6_6 : 1;
uint64_t ctlp_ro : 1;
uint64_t inta_map : 2;
uint64_t intb_map : 2;
uint64_t intc_map : 2;
uint64_t intd_map : 2;
uint64_t inta : 1;
uint64_t intb : 1;
uint64_t intc : 1;
uint64_t intd : 1;
uint64_t waitl_com : 1;
uint64_t reserved_21_63 : 43;
#endif
} s;
struct cvmx_npei_ctl_port0_s cn52xx;
struct cvmx_npei_ctl_port0_s cn52xxp1;
struct cvmx_npei_ctl_port0_s cn56xx;
struct cvmx_npei_ctl_port0_s cn56xxp1;
};
typedef union cvmx_npei_ctl_port0 cvmx_npei_ctl_port0_t;
/**
* cvmx_npei_ctl_port1
*
* NPEI_CTL_PORT1 = NPEI's Control Port1
*
* Contains control for access for Port1
*/
union cvmx_npei_ctl_port1
{
uint64_t u64;
struct cvmx_npei_ctl_port1_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_21_63 : 43;
uint64_t waitl_com : 1; /**< When set '1' casues the NPI to wait for a commit
from the L2C before sending additional completions
to the L2C from the PCIe.
Set this for more conservative behavior. Clear
this for more aggressive, higher-performance */
uint64_t intd : 1; /**< When '0' Intd wire asserted. Before mapping. */
uint64_t intc : 1; /**< When '0' Intc wire asserted. Before mapping. */
uint64_t intb : 1; /**< When '0' Intv wire asserted. Before mapping. */
uint64_t inta : 1; /**< When '0' Inta wire asserted. Before mapping. */
uint64_t intd_map : 2; /**< Maps INTD to INTA(00), INTB(01), INTC(10) or
INTD (11). */
uint64_t intc_map : 2; /**< Maps INTC to INTA(00), INTB(01), INTC(10) or
INTD (11). */
uint64_t intb_map : 2; /**< Maps INTB to INTA(00), INTB(01), INTC(10) or
INTD (11). */
uint64_t inta_map : 2; /**< Maps INTA to INTA(00), INTB(01), INTC(10) or
INTD (11). */
uint64_t ctlp_ro : 1; /**< Relaxed ordering enable for Completion TLPS. */
uint64_t reserved_6_6 : 1;
uint64_t ptlp_ro : 1; /**< Relaxed ordering enable for Posted TLPS. */
uint64_t bar2_enb : 1; /**< When set '1' BAR2 is enable and will respond when
clear '0' BAR2 access will cause UR responses. */
uint64_t bar2_esx : 2; /**< Value will be XORed with pci-address[37:36] to
determine the endian swap mode. */
uint64_t bar2_cax : 1; /**< Value will be XORed with pcie-address[38] to
determine the L2 cache attribute.
Not cached in L2 if XOR result is 1 */
uint64_t wait_com : 1; /**< When set '1' casues the NPI to wait for a commit
from the L2C before sending additional stores to
the L2C from the PCIe.
Most applications will not notice a difference, so
should not set this bit. Setting the bit is more
conservative on ordering, lower performance */
#else
uint64_t wait_com : 1;
uint64_t bar2_cax : 1;
uint64_t bar2_esx : 2;
uint64_t bar2_enb : 1;
uint64_t ptlp_ro : 1;
uint64_t reserved_6_6 : 1;
uint64_t ctlp_ro : 1;
uint64_t inta_map : 2;
uint64_t intb_map : 2;
uint64_t intc_map : 2;
uint64_t intd_map : 2;
uint64_t inta : 1;
uint64_t intb : 1;
uint64_t intc : 1;
uint64_t intd : 1;
uint64_t waitl_com : 1;
uint64_t reserved_21_63 : 43;
#endif
} s;
struct cvmx_npei_ctl_port1_s cn52xx;
struct cvmx_npei_ctl_port1_s cn52xxp1;
struct cvmx_npei_ctl_port1_s cn56xx;
struct cvmx_npei_ctl_port1_s cn56xxp1;
};
typedef union cvmx_npei_ctl_port1 cvmx_npei_ctl_port1_t;
/**
* cvmx_npei_ctl_status
*
* NPEI_CTL_STATUS = NPEI Control Status Register
*
* Contains control and status for NPEI. Writes to this register are not oSrdered with writes/reads to the PCIe Memory space.
* To ensure that a write has completed the user must read the register before making an access(i.e. PCIe memory space)
* that requires the value of this register to be updated.
*/
union cvmx_npei_ctl_status
{
uint64_t u64;
struct cvmx_npei_ctl_status_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_44_63 : 20;
uint64_t p1_ntags : 6; /**< Number of tags avaiable for PCIe Port1.
In RC mode 1 tag is needed for each outbound TLP
that requires a CPL TLP. In Endpoint mode the
number of tags required for a TLP request is
1 per 64-bytes of CPL data + 1.
This field should only be written as part of
reset sequence, before issuing any reads, CFGs, or
IO transactions from the core(s). */
uint64_t p0_ntags : 6; /**< Number of tags avaiable for PCIe Port0.
In RC mode 1 tag is needed for each outbound TLP
that requires a CPL TLP. In Endpoint mode the
number of tags required for a TLP request is
1 per 64-bytes of CPL data + 1.
This field should only be written as part of
reset sequence, before issuing any reads, CFGs, or
IO transactions from the core(s). */
uint64_t cfg_rtry : 16; /**< The time x 0x10000 in core clocks to wait for a
CPL to a CFG RD that does not carry a Retry Status.
Until such time that the timeout occurs and Retry
Status is received for a CFG RD, the Read CFG Read
will be resent. A value of 0 disables retries and
treats a CPL Retry as a CPL UR. */
uint64_t ring_en : 1; /**< When '0' forces "relative Q position" received
from PKO to be zero, and replicates the back-
pressure indication for the first ring attached
to a PKO port across all the rings attached to a
PKO port. When '1' backpressure is on a per
port/ring. */
uint64_t lnk_rst : 1; /**< Set when PCIe Core 0 request a link reset due to
link down state. This bit is only reset on raw
reset so it can be read for state to determine if
a reset occured. Bit is cleared when a '1' is
written to this field. */
uint64_t arb : 1; /**< PCIe switch arbitration mode. '0' == fixed priority
NPEI, PCIe0, then PCIe1. '1' == round robin. */
uint64_t pkt_bp : 4; /**< Unused */
uint64_t host_mode : 1; /**< Host mode */
uint64_t chip_rev : 8; /**< The chip revision. */
#else
uint64_t chip_rev : 8;
uint64_t host_mode : 1;
uint64_t pkt_bp : 4;
uint64_t arb : 1;
uint64_t lnk_rst : 1;
uint64_t ring_en : 1;
uint64_t cfg_rtry : 16;
uint64_t p0_ntags : 6;
uint64_t p1_ntags : 6;
uint64_t reserved_44_63 : 20;
#endif
} s;
struct cvmx_npei_ctl_status_s cn52xx;
struct cvmx_npei_ctl_status_cn52xxp1
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_44_63 : 20;
uint64_t p1_ntags : 6; /**< Number of tags avaiable for PCIe Port1.
In RC mode 1 tag is needed for each outbound TLP
that requires a CPL TLP. In Endpoint mode the
number of tags required for a TLP request is
1 per 64-bytes of CPL data + 1.
This field should only be written as part of
reset sequence, before issuing any reads, CFGs, or
IO transactions from the core(s). */
uint64_t p0_ntags : 6; /**< Number of tags avaiable for PCIe Port0.
In RC mode 1 tag is needed for each outbound TLP
that requires a CPL TLP. In Endpoint mode the
number of tags required for a TLP request is
1 per 64-bytes of CPL data + 1.
This field should only be written as part of
reset sequence, before issuing any reads, CFGs, or
IO transactions from the core(s). */
uint64_t cfg_rtry : 16; /**< The time x 0x10000 in core clocks to wait for a
CPL to a CFG RD that does not carry a Retry Status.
Until such time that the timeout occurs and Retry
Status is received for a CFG RD, the Read CFG Read
will be resent. A value of 0 disables retries and
treats a CPL Retry as a CPL UR. */
uint64_t reserved_15_15 : 1;
uint64_t lnk_rst : 1; /**< Set when PCIe Core 0 request a link reset due to
link down state. This bit is only reset on raw
reset so it can be read for state to determine if
a reset occured. Bit is cleared when a '1' is
written to this field. */
uint64_t arb : 1; /**< PCIe switch arbitration mode. '0' == fixed priority
NPEI, PCIe0, then PCIe1. '1' == round robin. */
uint64_t reserved_9_12 : 4;
uint64_t host_mode : 1; /**< Host mode */
uint64_t chip_rev : 8; /**< The chip revision. */
#else
uint64_t chip_rev : 8;
uint64_t host_mode : 1;
uint64_t reserved_9_12 : 4;
uint64_t arb : 1;
uint64_t lnk_rst : 1;
uint64_t reserved_15_15 : 1;
uint64_t cfg_rtry : 16;
uint64_t p0_ntags : 6;
uint64_t p1_ntags : 6;
uint64_t reserved_44_63 : 20;
#endif
} cn52xxp1;
struct cvmx_npei_ctl_status_s cn56xx;
struct cvmx_npei_ctl_status_cn56xxp1
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_15_63 : 49;
uint64_t lnk_rst : 1; /**< Set when PCIe Core 0 request a link reset due to
link down state. This bit is only reset on raw
reset so it can be read for state to determine if
a reset occured. Bit is cleared when a '1' is
written to this field. */
uint64_t arb : 1; /**< PCIe switch arbitration mode. '0' == fixed priority
NPEI, PCIe0, then PCIe1. '1' == round robin. */
uint64_t pkt_bp : 4; /**< Unused */
uint64_t host_mode : 1; /**< Host mode */
uint64_t chip_rev : 8; /**< The chip revision. */
#else
uint64_t chip_rev : 8;
uint64_t host_mode : 1;
uint64_t pkt_bp : 4;
uint64_t arb : 1;
uint64_t lnk_rst : 1;
uint64_t reserved_15_63 : 49;
#endif
} cn56xxp1;
};
typedef union cvmx_npei_ctl_status cvmx_npei_ctl_status_t;
/**
* cvmx_npei_ctl_status2
*
* NPEI_CTL_STATUS2 = NPEI's Control Status2 Register
*
* Contains control and status for NPEI.
* Writes to this register are not ordered with writes/reads to the PCI Memory space.
* To ensure that a write has completed the user must read the register before
* making an access(i.e. PCI memory space) that requires the value of this register to be updated.
*/
union cvmx_npei_ctl_status2
{
uint64_t u64;
struct cvmx_npei_ctl_status2_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_16_63 : 48;
uint64_t mps : 1; /**< Max Payload Size
0 = 128B
1 = 256B
Note: PCIE*_CFG030[MPS] must be set to the same
value for proper function. */
uint64_t mrrs : 3; /**< Max Read Request Size
0 = 128B
1 = 256B
2 = 512B
3 = 1024B
4 = 2048B
5 = 4096B
Note: This field must not exceed the desired
max read request size. This means this field
should not exceed PCIE*_CFG030[MRRS]. */
uint64_t c1_w_flt : 1; /**< When '1' enables the window filter for reads and
writes using the window registers.
PCIE-Port1.
Unfilter writes are:
MIO, SubId0
MIO, SubId7
NPEI, SubId0
NPEI, SubId7
POW, SubId7
IPD, SubId7
USBN0, SubId7
Unfiltered Reads are:
MIO, SubId0
MIO, SubId7
NPEI, SubId0
NPEI, SubId7
POW, SubId1
POW, SubId2
POW, SubId3
POW, SubId7
IPD, SubId7
USBN0, SubId7 */
uint64_t c0_w_flt : 1; /**< When '1' enables the window filter for reads and
writes using the window registers.
PCIE-Port0.
Unfilter writes are:
MIO, SubId0
MIO, SubId7
NPEI, SubId0
NPEI, SubId7
POW, SubId7
IPD, SubId7
USBN0, SubId7
Unfiltered Reads are:
MIO, SubId0
MIO, SubId7
NPEI, SubId0
NPEI, SubId7
POW, SubId1
POW, SubId2
POW, SubId3
POW, SubId7
IPD, SubId7
USBN0, SubId7 */
uint64_t c1_b1_s : 3; /**< Pcie-Port1, Bar1 Size. 1 == 64MB, 2 == 128MB,
3 == 256MB, 4 == 512MB, 5 == 1024MB, 6 == 2048MB,
0 and 7 are reserved. */
uint64_t c0_b1_s : 3; /**< Pcie-Port0, Bar1 Size. 1 == 64MB, 2 == 128MB,
3 == 256MB, 4 == 512MB, 5 == 1024MB, 6 == 2048MB,
0 and 7 are reserved. */
uint64_t c1_wi_d : 1; /**< When set '1' disables access to the Window
Registers from the PCIe-Port1. */
uint64_t c1_b0_d : 1; /**< When set '1' disables access from PCIe-Port1 to
BAR-0 address offsets: Less Than 0x270,
Greater than 0x270 AND less than 0x0520, 0x3BC0,
0x3CD0. */
uint64_t c0_wi_d : 1; /**< When set '1' disables access to the Window
Registers from the PCIe-Port0. */
uint64_t c0_b0_d : 1; /**< When set '1' disables access from PCIe-Port0 to
BAR-0 address offsets: Less Than 0x270,
Greater than 0x270 AND less than 0x0520, 0x3BC0,
0x3CD0. */
#else
uint64_t c0_b0_d : 1;
uint64_t c0_wi_d : 1;
uint64_t c1_b0_d : 1;
uint64_t c1_wi_d : 1;
uint64_t c0_b1_s : 3;
uint64_t c1_b1_s : 3;
uint64_t c0_w_flt : 1;
uint64_t c1_w_flt : 1;
uint64_t mrrs : 3;
uint64_t mps : 1;
uint64_t reserved_16_63 : 48;
#endif
} s;
struct cvmx_npei_ctl_status2_s cn52xx;
struct cvmx_npei_ctl_status2_s cn52xxp1;
struct cvmx_npei_ctl_status2_s cn56xx;
struct cvmx_npei_ctl_status2_s cn56xxp1;
};
typedef union cvmx_npei_ctl_status2 cvmx_npei_ctl_status2_t;
/**
* cvmx_npei_data_out_cnt
*
* NPEI_DATA_OUT_CNT = NPEI DATA OUT COUNT
*
* The EXEC data out fifo-count and the data unload counter.
*/
union cvmx_npei_data_out_cnt
{
uint64_t u64;
struct cvmx_npei_data_out_cnt_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_44_63 : 20;
uint64_t p1_ucnt : 16; /**< PCIE-Port1 Fifo Unload Count. This counter is
incremented by '1' every time a word is removed
from the Data Out FIFO, whose count is shown in
P0_FCNT. */
uint64_t p1_fcnt : 6; /**< PCIE-Port1 Data Out Fifo Count. Number of address
data words to be sent out the PCIe port presently
buffered in the FIFO. */
uint64_t p0_ucnt : 16; /**< PCIE-Port0 Fifo Unload Count. This counter is
incremented by '1' every time a word is removed
from the Data Out FIFO, whose count is shown in
P0_FCNT. */
uint64_t p0_fcnt : 6; /**< PCIE-Port0 Data Out Fifo Count. Number of address
data words to be sent out the PCIe port presently
buffered in the FIFO. */
#else
uint64_t p0_fcnt : 6;
uint64_t p0_ucnt : 16;
uint64_t p1_fcnt : 6;
uint64_t p1_ucnt : 16;
uint64_t reserved_44_63 : 20;
#endif
} s;
struct cvmx_npei_data_out_cnt_s cn52xx;
struct cvmx_npei_data_out_cnt_s cn52xxp1;
struct cvmx_npei_data_out_cnt_s cn56xx;
struct cvmx_npei_data_out_cnt_s cn56xxp1;
};
typedef union cvmx_npei_data_out_cnt cvmx_npei_data_out_cnt_t;
/**
* cvmx_npei_dbg_data
*
* NPEI_DBG_DATA = NPEI Debug Data Register
*
* Value returned on the debug-data lines from the RSLs
*/
union cvmx_npei_dbg_data
{
uint64_t u64;
struct cvmx_npei_dbg_data_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_28_63 : 36;
uint64_t qlm0_rev_lanes : 1; /**< Lane reversal for PCIe port 0 */
uint64_t reserved_25_26 : 2;
uint64_t qlm1_spd : 2; /**< Sets the QLM1 frequency
0=1.25 Gbaud
1=2.5 Gbaud
2=3.125 Gbaud
3=3.75 Gbaud */
uint64_t c_mul : 5; /**< PLL_MUL pins sampled at DCOK assertion
Core frequency = 50MHz*C_MUL */
uint64_t dsel_ext : 1; /**< Allows changes in the external pins to set the
debug select value. */
uint64_t data : 17; /**< Value on the debug data lines. */
#else
uint64_t data : 17;
uint64_t dsel_ext : 1;
uint64_t c_mul : 5;
uint64_t qlm1_spd : 2;
uint64_t reserved_25_26 : 2;
uint64_t qlm0_rev_lanes : 1;
uint64_t reserved_28_63 : 36;
#endif
} s;
struct cvmx_npei_dbg_data_cn52xx
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_29_63 : 35;
uint64_t qlm0_link_width : 1; /**< Link width of PCIe port 0
0 = PCIe port 0 is 2 lanes,
2 lane PCIe port 1 exists
1 = PCIe port 0 is 4 lanes,
PCIe port 1 does not exist */
uint64_t qlm0_rev_lanes : 1; /**< Lane reversal for PCIe port 0 */
uint64_t qlm1_mode : 2; /**< Sets the QLM1 Mode
0=Reserved
1=XAUI
2=SGMII
3=PICMG */
uint64_t qlm1_spd : 2; /**< Sets the QLM1 frequency
0=1.25 Gbaud
1=2.5 Gbaud
2=3.125 Gbaud
3=3.75 Gbaud */
uint64_t c_mul : 5; /**< PLL_MUL pins sampled at DCOK assertion
Core frequency = 50MHz*C_MUL */
uint64_t dsel_ext : 1; /**< Allows changes in the external pins to set the
debug select value. */
uint64_t data : 17; /**< Value on the debug data lines. */
#else
uint64_t data : 17;
uint64_t dsel_ext : 1;
uint64_t c_mul : 5;
uint64_t qlm1_spd : 2;
uint64_t qlm1_mode : 2;
uint64_t qlm0_rev_lanes : 1;
uint64_t qlm0_link_width : 1;
uint64_t reserved_29_63 : 35;
#endif
} cn52xx;
struct cvmx_npei_dbg_data_cn52xx cn52xxp1;
struct cvmx_npei_dbg_data_cn56xx
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_29_63 : 35;
uint64_t qlm2_rev_lanes : 1; /**< Lane reversal for PCIe port 1 */
uint64_t qlm0_rev_lanes : 1; /**< Lane reversal for PCIe port 0 */
uint64_t qlm3_spd : 2; /**< Sets the QLM3 frequency
0=1.25 Gbaud
1=2.5 Gbaud
2=3.125 Gbaud
3=3.75 Gbaud */
uint64_t qlm1_spd : 2; /**< Sets the QLM1 frequency
0=1.25 Gbaud
1=2.5 Gbaud
2=3.125 Gbaud
3=3.75 Gbaud */
uint64_t c_mul : 5; /**< PLL_MUL pins sampled at DCOK assertion
Core frequency = 50MHz*C_MUL */
uint64_t dsel_ext : 1; /**< Allows changes in the external pins to set the
debug select value. */
uint64_t data : 17; /**< Value on the debug data lines. */
#else
uint64_t data : 17;
uint64_t dsel_ext : 1;
uint64_t c_mul : 5;
uint64_t qlm1_spd : 2;
uint64_t qlm3_spd : 2;
uint64_t qlm0_rev_lanes : 1;
uint64_t qlm2_rev_lanes : 1;
uint64_t reserved_29_63 : 35;
#endif
} cn56xx;
struct cvmx_npei_dbg_data_cn56xx cn56xxp1;
};
typedef union cvmx_npei_dbg_data cvmx_npei_dbg_data_t;
/**
* cvmx_npei_dbg_select
*
* NPEI_DBG_SELECT = Debug Select Register
*
* Contains the debug select value last written to the RSLs.
*/
union cvmx_npei_dbg_select
{
uint64_t u64;
struct cvmx_npei_dbg_select_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_16_63 : 48;
uint64_t dbg_sel : 16; /**< When this register is written its value is sent to
all RSLs. */
#else
uint64_t dbg_sel : 16;
uint64_t reserved_16_63 : 48;
#endif
} s;
struct cvmx_npei_dbg_select_s cn52xx;
struct cvmx_npei_dbg_select_s cn52xxp1;
struct cvmx_npei_dbg_select_s cn56xx;
struct cvmx_npei_dbg_select_s cn56xxp1;
};
typedef union cvmx_npei_dbg_select cvmx_npei_dbg_select_t;
/**
* cvmx_npei_dma#_counts
*
* NPEI_DMA[0..4]_COUNTS = DMA Instruction Counts
*
* Values for determing the number of instructions for DMA[0..4] in the NPEI.
*/
union cvmx_npei_dmax_counts
{
uint64_t u64;
struct cvmx_npei_dmax_counts_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_39_63 : 25;
uint64_t fcnt : 7; /**< Number of words in the Instruction FIFO. */
uint64_t dbell : 32; /**< Number of available words of Instructions to read. */
#else
uint64_t dbell : 32;
uint64_t fcnt : 7;
uint64_t reserved_39_63 : 25;
#endif
} s;
struct cvmx_npei_dmax_counts_s cn52xx;
struct cvmx_npei_dmax_counts_s cn52xxp1;
struct cvmx_npei_dmax_counts_s cn56xx;
struct cvmx_npei_dmax_counts_s cn56xxp1;
};
typedef union cvmx_npei_dmax_counts cvmx_npei_dmax_counts_t;
/**
* cvmx_npei_dma#_dbell
*
* NPEI_DMA_DBELL[0..4] = DMA Door Bell
*
* The door bell register for DMA[0..4] queue.
*/
union cvmx_npei_dmax_dbell
{
uint32_t u32;
struct cvmx_npei_dmax_dbell_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint32_t reserved_16_31 : 16;
uint32_t dbell : 16; /**< The value written to this register is added to the
number of 8byte words to be read and processes for
the low priority dma queue. */
#else
uint32_t dbell : 16;
uint32_t reserved_16_31 : 16;
#endif
} s;
struct cvmx_npei_dmax_dbell_s cn52xx;
struct cvmx_npei_dmax_dbell_s cn52xxp1;
struct cvmx_npei_dmax_dbell_s cn56xx;
struct cvmx_npei_dmax_dbell_s cn56xxp1;
};
typedef union cvmx_npei_dmax_dbell cvmx_npei_dmax_dbell_t;
/**
* cvmx_npei_dma#_ibuff_saddr
*
* NPEI_DMA[0..4]_IBUFF_SADDR = DMA Instruction Buffer Starting Address
*
* The address to start reading Instructions from for DMA[0..4].
*/
union cvmx_npei_dmax_ibuff_saddr
{
uint64_t u64;
struct cvmx_npei_dmax_ibuff_saddr_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_37_63 : 27;
uint64_t idle : 1; /**< DMA Engine IDLE state */
uint64_t saddr : 29; /**< The 128 byte aligned starting address to read the
first instruction. SADDR is address bit 35:7 of the
first instructions address. */
uint64_t reserved_0_6 : 7;
#else
uint64_t reserved_0_6 : 7;
uint64_t saddr : 29;
uint64_t idle : 1;
uint64_t reserved_37_63 : 27;
#endif
} s;
struct cvmx_npei_dmax_ibuff_saddr_s cn52xx;
struct cvmx_npei_dmax_ibuff_saddr_cn52xxp1
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_36_63 : 28;
uint64_t saddr : 29; /**< The 128 byte aligned starting address to read the
first instruction. SADDR is address bit 35:7 of the
first instructions address. */
uint64_t reserved_0_6 : 7;
#else
uint64_t reserved_0_6 : 7;
uint64_t saddr : 29;
uint64_t reserved_36_63 : 28;
#endif
} cn52xxp1;
struct cvmx_npei_dmax_ibuff_saddr_s cn56xx;
struct cvmx_npei_dmax_ibuff_saddr_cn52xxp1 cn56xxp1;
};
typedef union cvmx_npei_dmax_ibuff_saddr cvmx_npei_dmax_ibuff_saddr_t;
/**
* cvmx_npei_dma#_naddr
*
* NPEI_DMA[0..4]_NADDR = DMA Next Ichunk Address
*
* Place NPEI will read the next Ichunk data from. This is valid when state is 0
*/
union cvmx_npei_dmax_naddr
{
uint64_t u64;
struct cvmx_npei_dmax_naddr_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_36_63 : 28;
uint64_t addr : 36; /**< The next L2C address to read DMA# instructions
from. */
#else
uint64_t addr : 36;
uint64_t reserved_36_63 : 28;
#endif
} s;
struct cvmx_npei_dmax_naddr_s cn52xx;
struct cvmx_npei_dmax_naddr_s cn52xxp1;
struct cvmx_npei_dmax_naddr_s cn56xx;
struct cvmx_npei_dmax_naddr_s cn56xxp1;
};
typedef union cvmx_npei_dmax_naddr cvmx_npei_dmax_naddr_t;
/**
* cvmx_npei_dma0_int_level
*
* NPEI_DMA0_INT_LEVEL = NPEI DMA0 Interrupt Level
*
* Thresholds for DMA count and timer interrupts for DMA0.
*/
union cvmx_npei_dma0_int_level
{
uint64_t u64;
struct cvmx_npei_dma0_int_level_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t time : 32; /**< Whenever the DMA_CNT0 timer exceeds
this value, NPEI_INT_SUM[DTIME0] is set.
The DMA_CNT0 timer increments every core clock
whenever NPEI_DMA_CNTS[DMA0]!=0, and is cleared
when NPEI_INT_SUM[DTIME0] is written with one. */
uint64_t cnt : 32; /**< Whenever NPEI_DMA_CNTS[DMA0] exceeds this value,
NPEI_INT_SUM[DCNT0] is set. */
#else
uint64_t cnt : 32;
uint64_t time : 32;
#endif
} s;
struct cvmx_npei_dma0_int_level_s cn52xx;
struct cvmx_npei_dma0_int_level_s cn52xxp1;
struct cvmx_npei_dma0_int_level_s cn56xx;
struct cvmx_npei_dma0_int_level_s cn56xxp1;
};
typedef union cvmx_npei_dma0_int_level cvmx_npei_dma0_int_level_t;
/**
* cvmx_npei_dma1_int_level
*
* NPEI_DMA1_INT_LEVEL = NPEI DMA1 Interrupt Level
*
* Thresholds for DMA count and timer interrupts for DMA1.
*/
union cvmx_npei_dma1_int_level
{
uint64_t u64;
struct cvmx_npei_dma1_int_level_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t time : 32; /**< Whenever the DMA_CNT1 timer exceeds
this value, NPEI_INT_SUM[DTIME1] is set.
The DMA_CNT1 timer increments every core clock
whenever NPEI_DMA_CNTS[DMA1]!=0, and is cleared
when NPEI_INT_SUM[DTIME1] is written with one. */
uint64_t cnt : 32; /**< Whenever NPEI_DMA_CNTS[DMA1] exceeds this value,
NPEI_INT_SUM[DCNT1] is set. */
#else
uint64_t cnt : 32;
uint64_t time : 32;
#endif
} s;
struct cvmx_npei_dma1_int_level_s cn52xx;
struct cvmx_npei_dma1_int_level_s cn52xxp1;
struct cvmx_npei_dma1_int_level_s cn56xx;
struct cvmx_npei_dma1_int_level_s cn56xxp1;
};
typedef union cvmx_npei_dma1_int_level cvmx_npei_dma1_int_level_t;
/**
* cvmx_npei_dma_cnts
*
* NPEI_DMA_CNTS = NPEI DMA Count
*
* The DMA Count values for DMA0 and DMA1.
*/
union cvmx_npei_dma_cnts
{
uint64_t u64;
struct cvmx_npei_dma_cnts_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t dma1 : 32; /**< The DMA counter 1.
Writing this field will cause the written value to
be subtracted from DMA1. SW should use a 4-byte
write to access this field so as not to change the
value of other fields in this register.
HW will optionally increment this field after
it completes an OUTBOUND or EXTERNAL-ONLY DMA
instruction. These increments may cause interrupts.
Refer to NPEI_DMA1_INT_LEVEL and
NPEI_INT_SUM[DCNT1,DTIME1]. */
uint64_t dma0 : 32; /**< The DMA counter 0.
Writing this field will cause the written value to
be subtracted from DMA0. SW should use a 4-byte
write to access this field so as not to change the
value of other fields in this register.
HW will optionally increment this field after
it completes an OUTBOUND or EXTERNAL-ONLY DMA
instruction. These increments may cause interrupts.
Refer to NPEI_DMA0_INT_LEVEL and
NPEI_INT_SUM[DCNT0,DTIME0]. */
#else
uint64_t dma0 : 32;
uint64_t dma1 : 32;
#endif
} s;
struct cvmx_npei_dma_cnts_s cn52xx;
struct cvmx_npei_dma_cnts_s cn52xxp1;
struct cvmx_npei_dma_cnts_s cn56xx;
struct cvmx_npei_dma_cnts_s cn56xxp1;
};
typedef union cvmx_npei_dma_cnts cvmx_npei_dma_cnts_t;
/**
* cvmx_npei_dma_control
*
* NPEI_DMA_CONTROL = DMA Control Register
*
* Controls operation of the DMA IN/OUT.
*/
union cvmx_npei_dma_control
{
uint64_t u64;
struct cvmx_npei_dma_control_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_40_63 : 24;
uint64_t p_32b_m : 1; /**< DMA PCIE 32-bit word read disable bit
When 0, enable the feature */
uint64_t dma4_enb : 1; /**< DMA# enable. Enables the operation of the DMA
engine. After being enabled a DMA engine should not
be dis-abled while processing instructions. */
uint64_t dma3_enb : 1; /**< DMA# enable. Enables the operation of the DMA
engine. After being enabled a DMA engine should not
be dis-abled while processing instructions. */
uint64_t dma2_enb : 1; /**< DMA# enable. Enables the operation of the DMA
engine. After being enabled a DMA engine should not
be dis-abled while processing instructions. */
uint64_t dma1_enb : 1; /**< DMA# enable. Enables the operation of the DMA
engine. After being enabled a DMA engine should not
be dis-abled while processing instructions. */
uint64_t dma0_enb : 1; /**< DMA# enable. Enables the operation of the DMA
engine. After being enabled a DMA engine should not
be dis-abled while processing instructions. */
uint64_t b0_lend : 1; /**< When set '1' and the NPEI is in the mode to write
0 to L2C memory when a DMA is done, the address
to be written to will be treated as a Little
Endian address. */
uint64_t dwb_denb : 1; /**< When set '1' the NPEI will send a value in the DWB
field for a free page operation for the memory
that contained the data. */
uint64_t dwb_ichk : 9; /**< When Instruction Chunks for DMA operations are freed
this value is used for the DWB field of the
operation. */
uint64_t fpa_que : 3; /**< The FPA queue that the instruction-chunk page will
be returned to when used. */
uint64_t o_add1 : 1; /**< When set '1' 1 will be added to the DMA counters,
if '0' then the number of bytes in the dma transfer
will be added to the count register. */
uint64_t o_ro : 1; /**< Relaxed Ordering Mode for DMA. */
uint64_t o_ns : 1; /**< Nosnoop For DMA. */
uint64_t o_es : 2; /**< Endian Swap Mode for DMA. */
uint64_t o_mode : 1; /**< Select PCI_POINTER MODE to be used.
'1' use pointer values for address and register
values for RO, ES, and NS, '0' use register
values for address and pointer values for
RO, ES, and NS. */
uint64_t csize : 14; /**< The size in words of the DMA Instruction Chunk.
This value should only be written once. After
writing this value a new value will not be
recognized until the end of the DMA I-Chunk is
reached. */
#else
uint64_t csize : 14;
uint64_t o_mode : 1;
uint64_t o_es : 2;
uint64_t o_ns : 1;
uint64_t o_ro : 1;
uint64_t o_add1 : 1;
uint64_t fpa_que : 3;
uint64_t dwb_ichk : 9;
uint64_t dwb_denb : 1;
uint64_t b0_lend : 1;
uint64_t dma0_enb : 1;
uint64_t dma1_enb : 1;
uint64_t dma2_enb : 1;
uint64_t dma3_enb : 1;
uint64_t dma4_enb : 1;
uint64_t p_32b_m : 1;
uint64_t reserved_40_63 : 24;
#endif
} s;
struct cvmx_npei_dma_control_s cn52xx;
struct cvmx_npei_dma_control_cn52xxp1
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_38_63 : 26;
uint64_t dma3_enb : 1; /**< DMA# enable. Enables the operation of the DMA
engine. After being enabled a DMA engine should not
be dis-abled while processing instructions. */
uint64_t dma2_enb : 1; /**< DMA# enable. Enables the operation of the DMA
engine. After being enabled a DMA engine should not
be dis-abled while processing instructions. */
uint64_t dma1_enb : 1; /**< DMA# enable. Enables the operation of the DMA
engine. After being enabled a DMA engine should not
be dis-abled while processing instructions. */
uint64_t dma0_enb : 1; /**< DMA# enable. Enables the operation of the DMA
engine. After being enabled a DMA engine should not
be dis-abled while processing instructions. */
uint64_t b0_lend : 1; /**< When set '1' and the NPEI is in the mode to write
0 to L2C memory when a DMA is done, the address
to be written to will be treated as a Little
Endian address. */
uint64_t dwb_denb : 1; /**< When set '1' the NPEI will send a value in the DWB
field for a free page operation for the memory
that contained the data. */
uint64_t dwb_ichk : 9; /**< When Instruction Chunks for DMA operations are freed
this value is used for the DWB field of the
operation. */
uint64_t fpa_que : 3; /**< The FPA queue that the instruction-chunk page will
be returned to when used. */
uint64_t o_add1 : 1; /**< When set '1' 1 will be added to the DMA counters,
if '0' then the number of bytes in the dma transfer
will be added to the count register. */
uint64_t o_ro : 1; /**< Relaxed Ordering Mode for DMA. */
uint64_t o_ns : 1; /**< Nosnoop For DMA. */
uint64_t o_es : 2; /**< Endian Swap Mode for DMA. */
uint64_t o_mode : 1; /**< Select PCI_POINTER MODE to be used.
'1' use pointer values for address and register
values for RO, ES, and NS, '0' use register
values for address and pointer values for
RO, ES, and NS. */
uint64_t csize : 14; /**< The size in words of the DMA Instruction Chunk.
This value should only be written once. After
writing this value a new value will not be
recognized until the end of the DMA I-Chunk is
reached. */
#else
uint64_t csize : 14;
uint64_t o_mode : 1;
uint64_t o_es : 2;
uint64_t o_ns : 1;
uint64_t o_ro : 1;
uint64_t o_add1 : 1;
uint64_t fpa_que : 3;
uint64_t dwb_ichk : 9;
uint64_t dwb_denb : 1;
uint64_t b0_lend : 1;
uint64_t dma0_enb : 1;
uint64_t dma1_enb : 1;
uint64_t dma2_enb : 1;
uint64_t dma3_enb : 1;
uint64_t reserved_38_63 : 26;
#endif
} cn52xxp1;
struct cvmx_npei_dma_control_s cn56xx;
struct cvmx_npei_dma_control_cn56xxp1
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_39_63 : 25;
uint64_t dma4_enb : 1; /**< DMA# enable. Enables the operation of the DMA
engine. After being enabled a DMA engine should not
be dis-abled while processing instructions. */
uint64_t dma3_enb : 1; /**< DMA# enable. Enables the operation of the DMA
engine. After being enabled a DMA engine should not
be dis-abled while processing instructions. */
uint64_t dma2_enb : 1; /**< DMA# enable. Enables the operation of the DMA
engine. After being enabled a DMA engine should not
be dis-abled while processing instructions. */
uint64_t dma1_enb : 1; /**< DMA# enable. Enables the operation of the DMA
engine. After being enabled a DMA engine should not
be dis-abled while processing instructions. */
uint64_t dma0_enb : 1; /**< DMA# enable. Enables the operation of the DMA
engine. After being enabled a DMA engine should not
be dis-abled while processing instructions. */
uint64_t b0_lend : 1; /**< When set '1' and the NPEI is in the mode to write
0 to L2C memory when a DMA is done, the address
to be written to will be treated as a Little
Endian address. */
uint64_t dwb_denb : 1; /**< When set '1' the NPEI will send a value in the DWB
field for a free page operation for the memory
that contained the data. */
uint64_t dwb_ichk : 9; /**< When Instruction Chunks for DMA operations are freed
this value is used for the DWB field of the
operation. */
uint64_t fpa_que : 3; /**< The FPA queue that the instruction-chunk page will
be returned to when used. */
uint64_t o_add1 : 1; /**< When set '1' 1 will be added to the DMA counters,
if '0' then the number of bytes in the dma transfer
will be added to the count register. */
uint64_t o_ro : 1; /**< Relaxed Ordering Mode for DMA. */
uint64_t o_ns : 1; /**< Nosnoop For DMA. */
uint64_t o_es : 2; /**< Endian Swap Mode for DMA. */
uint64_t o_mode : 1; /**< Select PCI_POINTER MODE to be used.
'1' use pointer values for address and register
values for RO, ES, and NS, '0' use register
values for address and pointer values for
RO, ES, and NS. */
uint64_t csize : 14; /**< The size in words of the DMA Instruction Chunk.
This value should only be written once. After
writing this value a new value will not be
recognized until the end of the DMA I-Chunk is
reached. */
#else
uint64_t csize : 14;
uint64_t o_mode : 1;
uint64_t o_es : 2;
uint64_t o_ns : 1;
uint64_t o_ro : 1;
uint64_t o_add1 : 1;
uint64_t fpa_que : 3;
uint64_t dwb_ichk : 9;
uint64_t dwb_denb : 1;
uint64_t b0_lend : 1;
uint64_t dma0_enb : 1;
uint64_t dma1_enb : 1;
uint64_t dma2_enb : 1;
uint64_t dma3_enb : 1;
uint64_t dma4_enb : 1;
uint64_t reserved_39_63 : 25;
#endif
} cn56xxp1;
};
typedef union cvmx_npei_dma_control cvmx_npei_dma_control_t;
/**
* cvmx_npei_dma_pcie_req_num
*
* NPEI_DMA_PCIE_REQ_NUM = NPEI DMA PCIE Outstanding Read Request Number
*
* Outstanding PCIE read request number for DMAs and Packet, maximum number is 16
*/
union cvmx_npei_dma_pcie_req_num
{
uint64_t u64;
struct cvmx_npei_dma_pcie_req_num_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t dma_arb : 1; /**< DMA_PKT Read Request Arbitration
- 1: DMA0-4 and PKT are round robin. i.e.
DMA0-DMA1-DMA2-DMA3-DMA4-PKT...
- 0: DMA0-4 are round robin, pkt gets selected
half the time. i.e.
DMA0-PKT-DMA1-PKT-DMA2-PKT-DMA3-PKT-DMA4-PKT... */
uint64_t reserved_53_62 : 10;
uint64_t pkt_cnt : 5; /**< PKT outstanding PCIE Read Request Number for each
PCIe port
When PKT_CNT=x, for each PCIe port, the number
of outstanding PCIe memory space reads by the PCIe
packet input/output will not exceed x.
Valid Number is between 1 and 16 */
uint64_t reserved_45_47 : 3;
uint64_t dma4_cnt : 5; /**< DMA4 outstanding PCIE Read Request Number
When DMA4_CNT=x, the number of outstanding PCIe
memory space reads by the PCIe DMA engine 4
will not exceed x.
Valid Number is between 1 and 16 */
uint64_t reserved_37_39 : 3;
uint64_t dma3_cnt : 5; /**< DMA3 outstanding PCIE Read Request Number
When DMA3_CNT=x, the number of outstanding PCIe
memory space reads by the PCIe DMA engine 3
will not exceed x.
Valid Number is between 1 and 16 */
uint64_t reserved_29_31 : 3;
uint64_t dma2_cnt : 5; /**< DMA2 outstanding PCIE Read Request Number
When DMA2_CNT=x, the number of outstanding PCIe
memory space reads by the PCIe DMA engine 2
will not exceed x.
Valid Number is between 1 and 16 */
uint64_t reserved_21_23 : 3;
uint64_t dma1_cnt : 5; /**< DMA1 outstanding PCIE Read Request Number
When DMA1_CNT=x, the number of outstanding PCIe
memory space reads by the PCIe DMA engine 1
will not exceed x.
Valid Number is between 1 and 16 */
uint64_t reserved_13_15 : 3;
uint64_t dma0_cnt : 5; /**< DMA0 outstanding PCIE Read Request Number
When DMA0_CNT=x, the number of outstanding PCIe
memory space reads by the PCIe DMA engine 0
will not exceed x.
Valid Number is between 1 and 16 */
uint64_t reserved_5_7 : 3;
uint64_t dma_cnt : 5; /**< Total outstanding PCIE Read Request Number for each
PCIe port
When DMA_CNT=x, for each PCIe port, the total
number of outstanding PCIe memory space reads
by the PCIe DMA engines and packet input/output
will not exceed x.
Valid Number is between 1 and 16 */
#else
uint64_t dma_cnt : 5;
uint64_t reserved_5_7 : 3;
uint64_t dma0_cnt : 5;
uint64_t reserved_13_15 : 3;
uint64_t dma1_cnt : 5;
uint64_t reserved_21_23 : 3;
uint64_t dma2_cnt : 5;
uint64_t reserved_29_31 : 3;
uint64_t dma3_cnt : 5;
uint64_t reserved_37_39 : 3;
uint64_t dma4_cnt : 5;
uint64_t reserved_45_47 : 3;
uint64_t pkt_cnt : 5;
uint64_t reserved_53_62 : 10;
uint64_t dma_arb : 1;
#endif
} s;
struct cvmx_npei_dma_pcie_req_num_s cn52xx;
struct cvmx_npei_dma_pcie_req_num_s cn56xx;
};
typedef union cvmx_npei_dma_pcie_req_num cvmx_npei_dma_pcie_req_num_t;
/**
* cvmx_npei_dma_state1
*
* NPEI_DMA_STATE1 = NPI's DMA State 1
*
* Results from DMA state register 1
*/
union cvmx_npei_dma_state1
{
uint64_t u64;
struct cvmx_npei_dma_state1_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_40_63 : 24;
uint64_t d4_dwe : 8; /**< DMA4 PICe Write State */
uint64_t d3_dwe : 8; /**< DMA3 PICe Write State */
uint64_t d2_dwe : 8; /**< DMA2 PICe Write State */
uint64_t d1_dwe : 8; /**< DMA1 PICe Write State */
uint64_t d0_dwe : 8; /**< DMA0 PICe Write State */
#else
uint64_t d0_dwe : 8;
uint64_t d1_dwe : 8;
uint64_t d2_dwe : 8;
uint64_t d3_dwe : 8;
uint64_t d4_dwe : 8;
uint64_t reserved_40_63 : 24;
#endif
} s;
struct cvmx_npei_dma_state1_s cn52xx;
};
typedef union cvmx_npei_dma_state1 cvmx_npei_dma_state1_t;
/**
* cvmx_npei_dma_state1_p1
*
* NPEI_DMA_STATE1_P1 = NPEI DMA Request and Instruction State
*
* DMA engine Debug information.
*/
union cvmx_npei_dma_state1_p1
{
uint64_t u64;
struct cvmx_npei_dma_state1_p1_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_60_63 : 4;
uint64_t d0_difst : 7; /**< DMA engine 0 dif instruction read state */
uint64_t d1_difst : 7; /**< DMA engine 1 dif instruction read state */
uint64_t d2_difst : 7; /**< DMA engine 2 dif instruction read state */
uint64_t d3_difst : 7; /**< DMA engine 3 dif instruction read state */
uint64_t d4_difst : 7; /**< DMA engine 4 dif instruction read state */
uint64_t d0_reqst : 5; /**< DMA engine 0 request data state */
uint64_t d1_reqst : 5; /**< DMA engine 1 request data state */
uint64_t d2_reqst : 5; /**< DMA engine 2 request data state */
uint64_t d3_reqst : 5; /**< DMA engine 3 request data state */
uint64_t d4_reqst : 5; /**< DMA engine 4 request data state */
#else
uint64_t d4_reqst : 5;
uint64_t d3_reqst : 5;
uint64_t d2_reqst : 5;
uint64_t d1_reqst : 5;
uint64_t d0_reqst : 5;
uint64_t d4_difst : 7;
uint64_t d3_difst : 7;
uint64_t d2_difst : 7;
uint64_t d1_difst : 7;
uint64_t d0_difst : 7;
uint64_t reserved_60_63 : 4;
#endif
} s;
struct cvmx_npei_dma_state1_p1_cn52xxp1
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_60_63 : 4;
uint64_t d0_difst : 7; /**< DMA engine 0 dif instruction read state */
uint64_t d1_difst : 7; /**< DMA engine 1 dif instruction read state */
uint64_t d2_difst : 7; /**< DMA engine 2 dif instruction read state */
uint64_t d3_difst : 7; /**< DMA engine 3 dif instruction read state */
uint64_t reserved_25_31 : 7;
uint64_t d0_reqst : 5; /**< DMA engine 0 request data state */
uint64_t d1_reqst : 5; /**< DMA engine 1 request data state */
uint64_t d2_reqst : 5; /**< DMA engine 2 request data state */
uint64_t d3_reqst : 5; /**< DMA engine 3 request data state */
uint64_t reserved_0_4 : 5;
#else
uint64_t reserved_0_4 : 5;
uint64_t d3_reqst : 5;
uint64_t d2_reqst : 5;
uint64_t d1_reqst : 5;
uint64_t d0_reqst : 5;
uint64_t reserved_25_31 : 7;
uint64_t d3_difst : 7;
uint64_t d2_difst : 7;
uint64_t d1_difst : 7;
uint64_t d0_difst : 7;
uint64_t reserved_60_63 : 4;
#endif
} cn52xxp1;
struct cvmx_npei_dma_state1_p1_s cn56xxp1;
};
typedef union cvmx_npei_dma_state1_p1 cvmx_npei_dma_state1_p1_t;
/**
* cvmx_npei_dma_state2
*
* NPEI_DMA_STATE2 = NPI's DMA State 2
*
* Results from DMA state register 2
*/
union cvmx_npei_dma_state2
{
uint64_t u64;
struct cvmx_npei_dma_state2_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_28_63 : 36;
uint64_t ndwe : 4; /**< DMA L2C Write State */
uint64_t reserved_21_23 : 3;
uint64_t ndre : 5; /**< DMA L2C Read State */
uint64_t reserved_10_15 : 6;
uint64_t prd : 10; /**< DMA PICe Read State */
#else
uint64_t prd : 10;
uint64_t reserved_10_15 : 6;
uint64_t ndre : 5;
uint64_t reserved_21_23 : 3;
uint64_t ndwe : 4;
uint64_t reserved_28_63 : 36;
#endif
} s;
struct cvmx_npei_dma_state2_s cn52xx;
};
typedef union cvmx_npei_dma_state2 cvmx_npei_dma_state2_t;
/**
* cvmx_npei_dma_state2_p1
*
* NPEI_DMA_STATE2_P1 = NPEI DMA Instruction Fetch State
*
* DMA engine Debug information.
*/
union cvmx_npei_dma_state2_p1
{
uint64_t u64;
struct cvmx_npei_dma_state2_p1_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_45_63 : 19;
uint64_t d0_dffst : 9; /**< DMA engine 0 dif instruction fetch state */
uint64_t d1_dffst : 9; /**< DMA engine 1 dif instruction fetch state */
uint64_t d2_dffst : 9; /**< DMA engine 2 dif instruction fetch state */
uint64_t d3_dffst : 9; /**< DMA engine 3 dif instruction fetch state */
uint64_t d4_dffst : 9; /**< DMA engine 4 dif instruction fetch state */
#else
uint64_t d4_dffst : 9;
uint64_t d3_dffst : 9;
uint64_t d2_dffst : 9;
uint64_t d1_dffst : 9;
uint64_t d0_dffst : 9;
uint64_t reserved_45_63 : 19;
#endif
} s;
struct cvmx_npei_dma_state2_p1_cn52xxp1
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_45_63 : 19;
uint64_t d0_dffst : 9; /**< DMA engine 0 dif instruction fetch state */
uint64_t d1_dffst : 9; /**< DMA engine 1 dif instruction fetch state */
uint64_t d2_dffst : 9; /**< DMA engine 2 dif instruction fetch state */
uint64_t d3_dffst : 9; /**< DMA engine 3 dif instruction fetch state */
uint64_t reserved_0_8 : 9;
#else
uint64_t reserved_0_8 : 9;
uint64_t d3_dffst : 9;
uint64_t d2_dffst : 9;
uint64_t d1_dffst : 9;
uint64_t d0_dffst : 9;
uint64_t reserved_45_63 : 19;
#endif
} cn52xxp1;
struct cvmx_npei_dma_state2_p1_s cn56xxp1;
};
typedef union cvmx_npei_dma_state2_p1 cvmx_npei_dma_state2_p1_t;
/**
* cvmx_npei_dma_state3_p1
*
* NPEI_DMA_STATE3_P1 = NPEI DMA DRE State
*
* DMA engine Debug information.
*/
union cvmx_npei_dma_state3_p1
{
uint64_t u64;
struct cvmx_npei_dma_state3_p1_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_60_63 : 4;
uint64_t d0_drest : 15; /**< DMA engine 0 dre state */
uint64_t d1_drest : 15; /**< DMA engine 1 dre state */
uint64_t d2_drest : 15; /**< DMA engine 2 dre state */
uint64_t d3_drest : 15; /**< DMA engine 3 dre state */
#else
uint64_t d3_drest : 15;
uint64_t d2_drest : 15;
uint64_t d1_drest : 15;
uint64_t d0_drest : 15;
uint64_t reserved_60_63 : 4;
#endif
} s;
struct cvmx_npei_dma_state3_p1_s cn52xxp1;
struct cvmx_npei_dma_state3_p1_s cn56xxp1;
};
typedef union cvmx_npei_dma_state3_p1 cvmx_npei_dma_state3_p1_t;
/**
* cvmx_npei_dma_state4_p1
*
* NPEI_DMA_STATE4_P1 = NPEI DMA DWE State
*
* DMA engine Debug information.
*/
union cvmx_npei_dma_state4_p1
{
uint64_t u64;
struct cvmx_npei_dma_state4_p1_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_52_63 : 12;
uint64_t d0_dwest : 13; /**< DMA engine 0 dwe state */
uint64_t d1_dwest : 13; /**< DMA engine 1 dwe state */
uint64_t d2_dwest : 13; /**< DMA engine 2 dwe state */
uint64_t d3_dwest : 13; /**< DMA engine 3 dwe state */
#else
uint64_t d3_dwest : 13;
uint64_t d2_dwest : 13;
uint64_t d1_dwest : 13;
uint64_t d0_dwest : 13;
uint64_t reserved_52_63 : 12;
#endif
} s;
struct cvmx_npei_dma_state4_p1_s cn52xxp1;
struct cvmx_npei_dma_state4_p1_s cn56xxp1;
};
typedef union cvmx_npei_dma_state4_p1 cvmx_npei_dma_state4_p1_t;
/**
* cvmx_npei_dma_state5_p1
*
* NPEI_DMA_STATE5_P1 = NPEI DMA DWE and DRE State
*
* DMA engine Debug information.
*/
union cvmx_npei_dma_state5_p1
{
uint64_t u64;
struct cvmx_npei_dma_state5_p1_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_28_63 : 36;
uint64_t d4_drest : 15; /**< DMA engine 4 dre state */
uint64_t d4_dwest : 13; /**< DMA engine 4 dwe state */
#else
uint64_t d4_dwest : 13;
uint64_t d4_drest : 15;
uint64_t reserved_28_63 : 36;
#endif
} s;
struct cvmx_npei_dma_state5_p1_s cn56xxp1;
};
typedef union cvmx_npei_dma_state5_p1 cvmx_npei_dma_state5_p1_t;
/**
* cvmx_npei_int_a_enb
*
* NPEI_INTERRUPT_A_ENB = NPI's Interrupt A Enable Register
*
* Used to allow the generation of interrupts (MSI/INTA) to the PCIe CoresUsed to enable the various interrupting conditions of NPEI
*/
union cvmx_npei_int_a_enb
{
uint64_t u64;
struct cvmx_npei_int_a_enb_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_10_63 : 54;
uint64_t pout_err : 1; /**< Enables NPEI_INT_A_SUM[9] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t pin_bp : 1; /**< Enables NPEI_INT_A_SUM[8] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t p1_rdlk : 1; /**< Enables NPEI_INT_A_SUM[7] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t p0_rdlk : 1; /**< Enables NPEI_INT_A_SUM[6] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t pgl_err : 1; /**< Enables NPEI_INT_A_SUM[5] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t pdi_err : 1; /**< Enables NPEI_INT_A_SUM[4] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t pop_err : 1; /**< Enables NPEI_INT_A_SUM[3] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t pins_err : 1; /**< Enables NPEI_INT_A_SUM[2] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dma1_cpl : 1; /**< Enables NPEI_INT_A_SUM[1] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dma0_cpl : 1; /**< Enables NPEI_INT_A_SUM[0] to generate an
interrupt to the PCIE core for MSI/inta. */
#else
uint64_t dma0_cpl : 1;
uint64_t dma1_cpl : 1;
uint64_t pins_err : 1;
uint64_t pop_err : 1;
uint64_t pdi_err : 1;
uint64_t pgl_err : 1;
uint64_t p0_rdlk : 1;
uint64_t p1_rdlk : 1;
uint64_t pin_bp : 1;
uint64_t pout_err : 1;
uint64_t reserved_10_63 : 54;
#endif
} s;
struct cvmx_npei_int_a_enb_s cn52xx;
struct cvmx_npei_int_a_enb_cn52xxp1
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_2_63 : 62;
uint64_t dma1_cpl : 1; /**< Enables NPEI_INT_A_SUM[1] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dma0_cpl : 1; /**< Enables NPEI_INT_A_SUM[0] to generate an
interrupt to the PCIE core for MSI/inta. */
#else
uint64_t dma0_cpl : 1;
uint64_t dma1_cpl : 1;
uint64_t reserved_2_63 : 62;
#endif
} cn52xxp1;
struct cvmx_npei_int_a_enb_s cn56xx;
};
typedef union cvmx_npei_int_a_enb cvmx_npei_int_a_enb_t;
/**
* cvmx_npei_int_a_enb2
*
* NPEI_INTERRUPT_A_ENB2 = NPEI's Interrupt A Enable2 Register
*
* Used to enable the various interrupting conditions of NPEI
*/
union cvmx_npei_int_a_enb2
{
uint64_t u64;
struct cvmx_npei_int_a_enb2_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_10_63 : 54;
uint64_t pout_err : 1; /**< Enables NPEI_INT_A_SUM[9] to generate an
interrupt on the RSL. */
uint64_t pin_bp : 1; /**< Enables NPEI_INT_A_SUM[8] to generate an
interrupt on the RSL. */
uint64_t p1_rdlk : 1; /**< Enables NPEI_INT_A_SUM[7] to generate an
interrupt on the RSL. */
uint64_t p0_rdlk : 1; /**< Enables NPEI_INT_A_SUM[6] to generate an
interrupt on the RSL. */
uint64_t pgl_err : 1; /**< Enables NPEI_INT_A_SUM[5] to generate an
interrupt on the RSL. */
uint64_t pdi_err : 1; /**< Enables NPEI_INT_A_SUM[4] to generate an
interrupt on the RSL. */
uint64_t pop_err : 1; /**< Enables NPEI_INT_A_SUM[3] to generate an
interrupt on the RSL. */
uint64_t pins_err : 1; /**< Enables NPEI_INT_A_SUM[2] to generate an
interrupt on the RSL. */
uint64_t dma1_cpl : 1; /**< Enables NPEI_INT_A_SUM[1] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dma0_cpl : 1; /**< Enables NPEI_INT_A_SUM[0] to generate an
interrupt to the PCIE core for MSI/inta. */
#else
uint64_t dma0_cpl : 1;
uint64_t dma1_cpl : 1;
uint64_t pins_err : 1;
uint64_t pop_err : 1;
uint64_t pdi_err : 1;
uint64_t pgl_err : 1;
uint64_t p0_rdlk : 1;
uint64_t p1_rdlk : 1;
uint64_t pin_bp : 1;
uint64_t pout_err : 1;
uint64_t reserved_10_63 : 54;
#endif
} s;
struct cvmx_npei_int_a_enb2_s cn52xx;
struct cvmx_npei_int_a_enb2_cn52xxp1
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_2_63 : 62;
uint64_t dma1_cpl : 1; /**< Enables NPEI_INT_A_SUM[1] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dma0_cpl : 1; /**< Enables NPEI_INT_A_SUM[0] to generate an
interrupt to the PCIE core for MSI/inta. */
#else
uint64_t dma0_cpl : 1;
uint64_t dma1_cpl : 1;
uint64_t reserved_2_63 : 62;
#endif
} cn52xxp1;
struct cvmx_npei_int_a_enb2_s cn56xx;
};
typedef union cvmx_npei_int_a_enb2 cvmx_npei_int_a_enb2_t;
/**
* cvmx_npei_int_a_sum
*
* NPEI_INTERRUPT_A_SUM = NPI Interrupt A Summary Register
*
* Set when an interrupt condition occurs, write '1' to clear. When an interrupt bitin this register is set and
* the cooresponding bit in the NPEI_INT_A_ENB register is set, then NPEI_INT_SUM[61] will be set.
*/
union cvmx_npei_int_a_sum
{
uint64_t u64;
struct cvmx_npei_int_a_sum_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_10_63 : 54;
uint64_t pout_err : 1; /**< Set when PKO sends packet data with the error bit
set. */
uint64_t pin_bp : 1; /**< Packet input count has exceeded the WMARK.
See NPEI_PKT_IN_BP */
uint64_t p1_rdlk : 1; /**< PCIe port 1 received a read lock. */
uint64_t p0_rdlk : 1; /**< PCIe port 0 received a read lock. */
uint64_t pgl_err : 1; /**< When a read error occurs on a packet gather list
read this bit is set. */
uint64_t pdi_err : 1; /**< When a read error occurs on a packet data read
this bit is set. */
uint64_t pop_err : 1; /**< When a read error occurs on a packet scatter
pointer pair this bit is set. */
uint64_t pins_err : 1; /**< When a read error occurs on a packet instruction
this bit is set. */
uint64_t dma1_cpl : 1; /**< Set each time any PCIe DMA engine recieves a UR/CA
response from PCIe Port 1 */
uint64_t dma0_cpl : 1; /**< Set each time any PCIe DMA engine recieves a UR/CA
response from PCIe Port 0 */
#else
uint64_t dma0_cpl : 1;
uint64_t dma1_cpl : 1;
uint64_t pins_err : 1;
uint64_t pop_err : 1;
uint64_t pdi_err : 1;
uint64_t pgl_err : 1;
uint64_t p0_rdlk : 1;
uint64_t p1_rdlk : 1;
uint64_t pin_bp : 1;
uint64_t pout_err : 1;
uint64_t reserved_10_63 : 54;
#endif
} s;
struct cvmx_npei_int_a_sum_s cn52xx;
struct cvmx_npei_int_a_sum_cn52xxp1
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_2_63 : 62;
uint64_t dma1_cpl : 1; /**< Set each time any PCIe DMA engine recieves a UR/CA
response from PCIe Port 1 */
uint64_t dma0_cpl : 1; /**< Set each time any PCIe DMA engine recieves a UR/CA
response from PCIe Port 0 */
#else
uint64_t dma0_cpl : 1;
uint64_t dma1_cpl : 1;
uint64_t reserved_2_63 : 62;
#endif
} cn52xxp1;
struct cvmx_npei_int_a_sum_s cn56xx;
};
typedef union cvmx_npei_int_a_sum cvmx_npei_int_a_sum_t;
/**
* cvmx_npei_int_enb
*
* NPEI_INTERRUPT_ENB = NPI's Interrupt Enable Register
*
* Used to allow the generation of interrupts (MSI/INTA) to the PCIe CoresUsed to enable the various interrupting conditions of NPI
*/
union cvmx_npei_int_enb
{
uint64_t u64;
struct cvmx_npei_int_enb_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t mio_inta : 1; /**< Enables NPEI_INT_SUM[63] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t reserved_62_62 : 1;
uint64_t int_a : 1; /**< Enables NPEI_INT_SUM[61] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_ldwn : 1; /**< Enables NPEI_INT_SUM[60] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_ldwn : 1; /**< Enables NPEI_INT_SUM[59] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_exc : 1; /**< Enables NPEI_INT_SUM[58] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_exc : 1; /**< Enables NPEI_INT_SUM[57] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_up_wf : 1; /**< Enables NPEI_INT_SUM[56] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_up_wf : 1; /**< Enables NPEI_INT_SUM[55] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_un_wf : 1; /**< Enables NPEI_INT_SUM[54] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_un_wf : 1; /**< Enables NPEI_INT_SUM[53] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_un_bx : 1; /**< Enables NPEI_INT_SUM[52] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_un_wi : 1; /**< Enables NPEI_INT_SUM[51] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_un_b2 : 1; /**< Enables NPEI_INT_SUM[50] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_un_b1 : 1; /**< Enables NPEI_INT_SUM[49] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_un_b0 : 1; /**< Enables NPEI_INT_SUM[48] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_up_bx : 1; /**< Enables NPEI_INT_SUM[47] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_up_wi : 1; /**< Enables NPEI_INT_SUM[46] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_up_b2 : 1; /**< Enables NPEI_INT_SUM[45] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_up_b1 : 1; /**< Enables NPEI_INT_SUM[44] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_up_b0 : 1; /**< Enables NPEI_INT_SUM[43] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_un_bx : 1; /**< Enables NPEI_INT_SUM[42] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_un_wi : 1; /**< Enables NPEI_INT_SUM[41] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_un_b2 : 1; /**< Enables NPEI_INT_SUM[40] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_un_b1 : 1; /**< Enables NPEI_INT_SUM[39] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_un_b0 : 1; /**< Enables NPEI_INT_SUM[38] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_up_bx : 1; /**< Enables NPEI_INT_SUM[37] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_up_wi : 1; /**< Enables NPEI_INT_SUM[36] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_up_b2 : 1; /**< Enables NPEI_INT_SUM[35] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_up_b1 : 1; /**< Enables NPEI_INT_SUM[34] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_up_b0 : 1; /**< Enables NPEI_INT_SUM[33] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_hpint : 1; /**< Enables NPEI_INT_SUM[32] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_pmei : 1; /**< Enables NPEI_INT_SUM[31] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_wake : 1; /**< Enables NPEI_INT_SUM[30] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t crs1_dr : 1; /**< Enables NPEI_INT_SUM[29] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_se : 1; /**< Enables NPEI_INT_SUM[28] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t crs1_er : 1; /**< Enables NPEI_INT_SUM[27] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_aeri : 1; /**< Enables NPEI_INT_SUM[26] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_hpint : 1; /**< Enables NPEI_INT_SUM[25] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_pmei : 1; /**< Enables NPEI_INT_SUM[24] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_wake : 1; /**< Enables NPEI_INT_SUM[23] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t crs0_dr : 1; /**< Enables NPEI_INT_SUM[22] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_se : 1; /**< Enables NPEI_INT_SUM[21] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t crs0_er : 1; /**< Enables NPEI_INT_SUM[20] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_aeri : 1; /**< Enables NPEI_INT_SUM[19] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t ptime : 1; /**< Enables NPEI_INT_SUM[18] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t pcnt : 1; /**< Enables NPEI_INT_SUM[17] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t pidbof : 1; /**< Enables NPEI_INT_SUM[16] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t psldbof : 1; /**< Enables NPEI_INT_SUM[15] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dtime1 : 1; /**< Enables NPEI_INT_SUM[14] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dtime0 : 1; /**< Enables NPEI_INT_SUM[13] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dcnt1 : 1; /**< Enables NPEI_INT_SUM[12] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dcnt0 : 1; /**< Enables NPEI_INT_SUM[11] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dma1fi : 1; /**< Enables NPEI_INT_SUM[10] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dma0fi : 1; /**< Enables NPEI_INT_SUM[9] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dma4dbo : 1; /**< Enables NPEI_INT_SUM[8] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dma3dbo : 1; /**< Enables NPEI_INT_SUM[7] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dma2dbo : 1; /**< Enables NPEI_INT_SUM[6] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dma1dbo : 1; /**< Enables NPEI_INT_SUM[5] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dma0dbo : 1; /**< Enables NPEI_INT_SUM[4] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t iob2big : 1; /**< Enables NPEI_INT_SUM[3] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t bar0_to : 1; /**< Enables NPEI_INT_SUM[2] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t rml_wto : 1; /**< Enables NPEI_INT_SUM[1] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t rml_rto : 1; /**< Enables NPEI_INT_SUM[0] to generate an
interrupt to the PCIE core for MSI/inta. */
#else
uint64_t rml_rto : 1;
uint64_t rml_wto : 1;
uint64_t bar0_to : 1;
uint64_t iob2big : 1;
uint64_t dma0dbo : 1;
uint64_t dma1dbo : 1;
uint64_t dma2dbo : 1;
uint64_t dma3dbo : 1;
uint64_t dma4dbo : 1;
uint64_t dma0fi : 1;
uint64_t dma1fi : 1;
uint64_t dcnt0 : 1;
uint64_t dcnt1 : 1;
uint64_t dtime0 : 1;
uint64_t dtime1 : 1;
uint64_t psldbof : 1;
uint64_t pidbof : 1;
uint64_t pcnt : 1;
uint64_t ptime : 1;
uint64_t c0_aeri : 1;
uint64_t crs0_er : 1;
uint64_t c0_se : 1;
uint64_t crs0_dr : 1;
uint64_t c0_wake : 1;
uint64_t c0_pmei : 1;
uint64_t c0_hpint : 1;
uint64_t c1_aeri : 1;
uint64_t crs1_er : 1;
uint64_t c1_se : 1;
uint64_t crs1_dr : 1;
uint64_t c1_wake : 1;
uint64_t c1_pmei : 1;
uint64_t c1_hpint : 1;
uint64_t c0_up_b0 : 1;
uint64_t c0_up_b1 : 1;
uint64_t c0_up_b2 : 1;
uint64_t c0_up_wi : 1;
uint64_t c0_up_bx : 1;
uint64_t c0_un_b0 : 1;
uint64_t c0_un_b1 : 1;
uint64_t c0_un_b2 : 1;
uint64_t c0_un_wi : 1;
uint64_t c0_un_bx : 1;
uint64_t c1_up_b0 : 1;
uint64_t c1_up_b1 : 1;
uint64_t c1_up_b2 : 1;
uint64_t c1_up_wi : 1;
uint64_t c1_up_bx : 1;
uint64_t c1_un_b0 : 1;
uint64_t c1_un_b1 : 1;
uint64_t c1_un_b2 : 1;
uint64_t c1_un_wi : 1;
uint64_t c1_un_bx : 1;
uint64_t c0_un_wf : 1;
uint64_t c1_un_wf : 1;
uint64_t c0_up_wf : 1;
uint64_t c1_up_wf : 1;
uint64_t c0_exc : 1;
uint64_t c1_exc : 1;
uint64_t c0_ldwn : 1;
uint64_t c1_ldwn : 1;
uint64_t int_a : 1;
uint64_t reserved_62_62 : 1;
uint64_t mio_inta : 1;
#endif
} s;
struct cvmx_npei_int_enb_s cn52xx;
struct cvmx_npei_int_enb_cn52xxp1
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t mio_inta : 1; /**< Enables NPEI_INT_SUM[63] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t reserved_62_62 : 1;
uint64_t int_a : 1; /**< Enables NPEI_INT_SUM[61] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_ldwn : 1; /**< Enables NPEI_INT_SUM[60] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_ldwn : 1; /**< Enables NPEI_INT_SUM[59] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_exc : 1; /**< Enables NPEI_INT_SUM[58] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_exc : 1; /**< Enables NPEI_INT_SUM[57] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_up_wf : 1; /**< Enables NPEI_INT_SUM[56] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_up_wf : 1; /**< Enables NPEI_INT_SUM[55] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_un_wf : 1; /**< Enables NPEI_INT_SUM[54] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_un_wf : 1; /**< Enables NPEI_INT_SUM[53] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_un_bx : 1; /**< Enables NPEI_INT_SUM[52] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_un_wi : 1; /**< Enables NPEI_INT_SUM[51] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_un_b2 : 1; /**< Enables NPEI_INT_SUM[50] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_un_b1 : 1; /**< Enables NPEI_INT_SUM[49] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_un_b0 : 1; /**< Enables NPEI_INT_SUM[48] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_up_bx : 1; /**< Enables NPEI_INT_SUM[47] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_up_wi : 1; /**< Enables NPEI_INT_SUM[46] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_up_b2 : 1; /**< Enables NPEI_INT_SUM[45] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_up_b1 : 1; /**< Enables NPEI_INT_SUM[44] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_up_b0 : 1; /**< Enables NPEI_INT_SUM[43] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_un_bx : 1; /**< Enables NPEI_INT_SUM[42] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_un_wi : 1; /**< Enables NPEI_INT_SUM[41] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_un_b2 : 1; /**< Enables NPEI_INT_SUM[40] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_un_b1 : 1; /**< Enables NPEI_INT_SUM[39] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_un_b0 : 1; /**< Enables NPEI_INT_SUM[38] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_up_bx : 1; /**< Enables NPEI_INT_SUM[37] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_up_wi : 1; /**< Enables NPEI_INT_SUM[36] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_up_b2 : 1; /**< Enables NPEI_INT_SUM[35] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_up_b1 : 1; /**< Enables NPEI_INT_SUM[34] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_up_b0 : 1; /**< Enables NPEI_INT_SUM[33] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_hpint : 1; /**< Enables NPEI_INT_SUM[32] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_pmei : 1; /**< Enables NPEI_INT_SUM[31] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_wake : 1; /**< Enables NPEI_INT_SUM[30] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t crs1_dr : 1; /**< Enables NPEI_INT_SUM[29] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_se : 1; /**< Enables NPEI_INT_SUM[28] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t crs1_er : 1; /**< Enables NPEI_INT_SUM[27] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_aeri : 1; /**< Enables NPEI_INT_SUM[26] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_hpint : 1; /**< Enables NPEI_INT_SUM[25] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_pmei : 1; /**< Enables NPEI_INT_SUM[24] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_wake : 1; /**< Enables NPEI_INT_SUM[23] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t crs0_dr : 1; /**< Enables NPEI_INT_SUM[22] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_se : 1; /**< Enables NPEI_INT_SUM[21] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t crs0_er : 1; /**< Enables NPEI_INT_SUM[20] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_aeri : 1; /**< Enables NPEI_INT_SUM[19] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t ptime : 1; /**< Enables NPEI_INT_SUM[18] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t pcnt : 1; /**< Enables NPEI_INT_SUM[17] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t pidbof : 1; /**< Enables NPEI_INT_SUM[16] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t psldbof : 1; /**< Enables NPEI_INT_SUM[15] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dtime1 : 1; /**< Enables NPEI_INT_SUM[14] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dtime0 : 1; /**< Enables NPEI_INT_SUM[13] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dcnt1 : 1; /**< Enables NPEI_INT_SUM[12] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dcnt0 : 1; /**< Enables NPEI_INT_SUM[11] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dma1fi : 1; /**< Enables NPEI_INT_SUM[10] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dma0fi : 1; /**< Enables NPEI_INT_SUM[9] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t reserved_8_8 : 1;
uint64_t dma3dbo : 1; /**< Enables NPEI_INT_SUM[7] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dma2dbo : 1; /**< Enables NPEI_INT_SUM[6] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dma1dbo : 1; /**< Enables NPEI_INT_SUM[5] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dma0dbo : 1; /**< Enables NPEI_INT_SUM[4] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t iob2big : 1; /**< Enables NPEI_INT_SUM[3] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t bar0_to : 1; /**< Enables NPEI_INT_SUM[2] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t rml_wto : 1; /**< Enables NPEI_INT_SUM[1] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t rml_rto : 1; /**< Enables NPEI_INT_SUM[0] to generate an
interrupt to the PCIE core for MSI/inta. */
#else
uint64_t rml_rto : 1;
uint64_t rml_wto : 1;
uint64_t bar0_to : 1;
uint64_t iob2big : 1;
uint64_t dma0dbo : 1;
uint64_t dma1dbo : 1;
uint64_t dma2dbo : 1;
uint64_t dma3dbo : 1;
uint64_t reserved_8_8 : 1;
uint64_t dma0fi : 1;
uint64_t dma1fi : 1;
uint64_t dcnt0 : 1;
uint64_t dcnt1 : 1;
uint64_t dtime0 : 1;
uint64_t dtime1 : 1;
uint64_t psldbof : 1;
uint64_t pidbof : 1;
uint64_t pcnt : 1;
uint64_t ptime : 1;
uint64_t c0_aeri : 1;
uint64_t crs0_er : 1;
uint64_t c0_se : 1;
uint64_t crs0_dr : 1;
uint64_t c0_wake : 1;
uint64_t c0_pmei : 1;
uint64_t c0_hpint : 1;
uint64_t c1_aeri : 1;
uint64_t crs1_er : 1;
uint64_t c1_se : 1;
uint64_t crs1_dr : 1;
uint64_t c1_wake : 1;
uint64_t c1_pmei : 1;
uint64_t c1_hpint : 1;
uint64_t c0_up_b0 : 1;
uint64_t c0_up_b1 : 1;
uint64_t c0_up_b2 : 1;
uint64_t c0_up_wi : 1;
uint64_t c0_up_bx : 1;
uint64_t c0_un_b0 : 1;
uint64_t c0_un_b1 : 1;
uint64_t c0_un_b2 : 1;
uint64_t c0_un_wi : 1;
uint64_t c0_un_bx : 1;
uint64_t c1_up_b0 : 1;
uint64_t c1_up_b1 : 1;
uint64_t c1_up_b2 : 1;
uint64_t c1_up_wi : 1;
uint64_t c1_up_bx : 1;
uint64_t c1_un_b0 : 1;
uint64_t c1_un_b1 : 1;
uint64_t c1_un_b2 : 1;
uint64_t c1_un_wi : 1;
uint64_t c1_un_bx : 1;
uint64_t c0_un_wf : 1;
uint64_t c1_un_wf : 1;
uint64_t c0_up_wf : 1;
uint64_t c1_up_wf : 1;
uint64_t c0_exc : 1;
uint64_t c1_exc : 1;
uint64_t c0_ldwn : 1;
uint64_t c1_ldwn : 1;
uint64_t int_a : 1;
uint64_t reserved_62_62 : 1;
uint64_t mio_inta : 1;
#endif
} cn52xxp1;
struct cvmx_npei_int_enb_s cn56xx;
struct cvmx_npei_int_enb_cn56xxp1
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t mio_inta : 1; /**< Enables NPEI_INT_SUM[63] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t reserved_61_62 : 2;
uint64_t c1_ldwn : 1; /**< Enables NPEI_INT_SUM[60] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_ldwn : 1; /**< Enables NPEI_INT_SUM[59] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_exc : 1; /**< Enables NPEI_INT_SUM[58] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_exc : 1; /**< Enables NPEI_INT_SUM[57] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_up_wf : 1; /**< Enables NPEI_INT_SUM[56] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_up_wf : 1; /**< Enables NPEI_INT_SUM[55] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_un_wf : 1; /**< Enables NPEI_INT_SUM[54] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_un_wf : 1; /**< Enables NPEI_INT_SUM[53] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_un_bx : 1; /**< Enables NPEI_INT_SUM[52] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_un_wi : 1; /**< Enables NPEI_INT_SUM[51] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_un_b2 : 1; /**< Enables NPEI_INT_SUM[50] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_un_b1 : 1; /**< Enables NPEI_INT_SUM[49] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_un_b0 : 1; /**< Enables NPEI_INT_SUM[48] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_up_bx : 1; /**< Enables NPEI_INT_SUM[47] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_up_wi : 1; /**< Enables NPEI_INT_SUM[46] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_up_b2 : 1; /**< Enables NPEI_INT_SUM[45] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_up_b1 : 1; /**< Enables NPEI_INT_SUM[44] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_up_b0 : 1; /**< Enables NPEI_INT_SUM[43] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_un_bx : 1; /**< Enables NPEI_INT_SUM[42] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_un_wi : 1; /**< Enables NPEI_INT_SUM[41] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_un_b2 : 1; /**< Enables NPEI_INT_SUM[40] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_un_b1 : 1; /**< Enables NPEI_INT_SUM[39] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_un_b0 : 1; /**< Enables NPEI_INT_SUM[38] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_up_bx : 1; /**< Enables NPEI_INT_SUM[37] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_up_wi : 1; /**< Enables NPEI_INT_SUM[36] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_up_b2 : 1; /**< Enables NPEI_INT_SUM[35] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_up_b1 : 1; /**< Enables NPEI_INT_SUM[34] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_up_b0 : 1; /**< Enables NPEI_INT_SUM[33] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_hpint : 1; /**< Enables NPEI_INT_SUM[32] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_pmei : 1; /**< Enables NPEI_INT_SUM[31] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c1_wake : 1; /**< Enables NPEI_INT_SUM[30] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t reserved_29_29 : 1;
uint64_t c1_se : 1; /**< Enables NPEI_INT_SUM[28] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t reserved_27_27 : 1;
uint64_t c1_aeri : 1; /**< Enables NPEI_INT_SUM[26] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_hpint : 1; /**< Enables NPEI_INT_SUM[25] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_pmei : 1; /**< Enables NPEI_INT_SUM[24] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t c0_wake : 1; /**< Enables NPEI_INT_SUM[23] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t reserved_22_22 : 1;
uint64_t c0_se : 1; /**< Enables NPEI_INT_SUM[21] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t reserved_20_20 : 1;
uint64_t c0_aeri : 1; /**< Enables NPEI_INT_SUM[19] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t ptime : 1; /**< Enables NPEI_INT_SUM[18] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t pcnt : 1; /**< Enables NPEI_INT_SUM[17] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t pidbof : 1; /**< Enables NPEI_INT_SUM[16] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t psldbof : 1; /**< Enables NPEI_INT_SUM[15] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dtime1 : 1; /**< Enables NPEI_INT_SUM[14] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dtime0 : 1; /**< Enables NPEI_INT_SUM[13] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dcnt1 : 1; /**< Enables NPEI_INT_SUM[12] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dcnt0 : 1; /**< Enables NPEI_INT_SUM[11] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dma1fi : 1; /**< Enables NPEI_INT_SUM[10] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dma0fi : 1; /**< Enables NPEI_INT_SUM[9] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dma4dbo : 1; /**< Enables NPEI_INT_SUM[8] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dma3dbo : 1; /**< Enables NPEI_INT_SUM[7] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dma2dbo : 1; /**< Enables NPEI_INT_SUM[6] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dma1dbo : 1; /**< Enables NPEI_INT_SUM[5] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t dma0dbo : 1; /**< Enables NPEI_INT_SUM[4] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t iob2big : 1; /**< Enables NPEI_INT_SUM[3] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t bar0_to : 1; /**< Enables NPEI_INT_SUM[2] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t rml_wto : 1; /**< Enables NPEI_INT_SUM[1] to generate an
interrupt to the PCIE core for MSI/inta. */
uint64_t rml_rto : 1; /**< Enables NPEI_INT_SUM[0] to generate an
interrupt to the PCIE core for MSI/inta. */
#else
uint64_t rml_rto : 1;
uint64_t rml_wto : 1;
uint64_t bar0_to : 1;
uint64_t iob2big : 1;
uint64_t dma0dbo : 1;
uint64_t dma1dbo : 1;
uint64_t dma2dbo : 1;
uint64_t dma3dbo : 1;
uint64_t dma4dbo : 1;
uint64_t dma0fi : 1;
uint64_t dma1fi : 1;
uint64_t dcnt0 : 1;
uint64_t dcnt1 : 1;
uint64_t dtime0 : 1;
uint64_t dtime1 : 1;
uint64_t psldbof : 1;
uint64_t pidbof : 1;
uint64_t pcnt : 1;
uint64_t ptime : 1;
uint64_t c0_aeri : 1;
uint64_t reserved_20_20 : 1;
uint64_t c0_se : 1;
uint64_t reserved_22_22 : 1;
uint64_t c0_wake : 1;
uint64_t c0_pmei : 1;
uint64_t c0_hpint : 1;
uint64_t c1_aeri : 1;
uint64_t reserved_27_27 : 1;
uint64_t c1_se : 1;
uint64_t reserved_29_29 : 1;
uint64_t c1_wake : 1;
uint64_t c1_pmei : 1;
uint64_t c1_hpint : 1;
uint64_t c0_up_b0 : 1;
uint64_t c0_up_b1 : 1;
uint64_t c0_up_b2 : 1;
uint64_t c0_up_wi : 1;
uint64_t c0_up_bx : 1;
uint64_t c0_un_b0 : 1;
uint64_t c0_un_b1 : 1;
uint64_t c0_un_b2 : 1;
uint64_t c0_un_wi : 1;
uint64_t c0_un_bx : 1;
uint64_t c1_up_b0 : 1;
uint64_t c1_up_b1 : 1;
uint64_t c1_up_b2 : 1;
uint64_t c1_up_wi : 1;
uint64_t c1_up_bx : 1;
uint64_t c1_un_b0 : 1;
uint64_t c1_un_b1 : 1;
uint64_t c1_un_b2 : 1;
uint64_t c1_un_wi : 1;
uint64_t c1_un_bx : 1;
uint64_t c0_un_wf : 1;
uint64_t c1_un_wf : 1;
uint64_t c0_up_wf : 1;
uint64_t c1_up_wf : 1;
uint64_t c0_exc : 1;
uint64_t c1_exc : 1;
uint64_t c0_ldwn : 1;
uint64_t c1_ldwn : 1;
uint64_t reserved_61_62 : 2;
uint64_t mio_inta : 1;
#endif
} cn56xxp1;
};
typedef union cvmx_npei_int_enb cvmx_npei_int_enb_t;
/**
* cvmx_npei_int_enb2
*
* NPEI_INTERRUPT_ENB2 = NPI's Interrupt Enable2 Register
*
* Used to enable the various interrupting conditions of NPI
*/
union cvmx_npei_int_enb2
{
uint64_t u64;
struct cvmx_npei_int_enb2_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_62_63 : 2;
uint64_t int_a : 1; /**< Enables NPEI_INT_SUM2[61] to generate an
interrupt on the RSL. */
uint64_t c1_ldwn : 1; /**< Enables NPEI_INT_SUM[60] to generate an
interrupt on the RSL. */
uint64_t c0_ldwn : 1; /**< Enables NPEI_INT_SUM[59] to generate an
interrupt on the RSL. */
uint64_t c1_exc : 1; /**< Enables NPEI_INT_SUM[58] to generate an
interrupt on the RSL. */
uint64_t c0_exc : 1; /**< Enables NPEI_INT_SUM[57] to generate an
interrupt on the RSL. */
uint64_t c1_up_wf : 1; /**< Enables NPEI_INT_SUM[56] to generate an
interrupt on the RSL. */
uint64_t c0_up_wf : 1; /**< Enables NPEI_INT_SUM[55] to generate an
interrupt on the RSL. */
uint64_t c1_un_wf : 1; /**< Enables NPEI_INT_SUM[54] to generate an
interrupt on the RSL. */
uint64_t c0_un_wf : 1; /**< Enables NPEI_INT_SUM[53] to generate an
interrupt on the RSL. */
uint64_t c1_un_bx : 1; /**< Enables NPEI_INT_SUM[52] to generate an
interrupt on the RSL. */
uint64_t c1_un_wi : 1; /**< Enables NPEI_INT_SUM[51] to generate an
interrupt on the RSL. */
uint64_t c1_un_b2 : 1; /**< Enables NPEI_INT_SUM[50] to generate an
interrupt on the RSL. */
uint64_t c1_un_b1 : 1; /**< Enables NPEI_INT_SUM[49] to generate an
interrupt on the RSL. */
uint64_t c1_un_b0 : 1; /**< Enables NPEI_INT_SUM[48] to generate an
interrupt on the RSL. */
uint64_t c1_up_bx : 1; /**< Enables NPEI_INT_SUM[47] to generate an
interrupt on the RSL. */
uint64_t c1_up_wi : 1; /**< Enables NPEI_INT_SUM[46] to generate an
interrupt on the RSL. */
uint64_t c1_up_b2 : 1; /**< Enables NPEI_INT_SUM[45] to generate an
interrupt on the RSL. */
uint64_t c1_up_b1 : 1; /**< Enables NPEI_INT_SUM[44] to generate an
interrupt on the RSL. */
uint64_t c1_up_b0 : 1; /**< Enables NPEI_INT_SUM[43] to generate an
interrupt on the RSL. */
uint64_t c0_un_bx : 1; /**< Enables NPEI_INT_SUM[42] to generate an
interrupt on the RSL. */
uint64_t c0_un_wi : 1; /**< Enables NPEI_INT_SUM[41] to generate an
interrupt on the RSL. */
uint64_t c0_un_b2 : 1; /**< Enables NPEI_INT_SUM[40] to generate an
interrupt on the RSL. */
uint64_t c0_un_b1 : 1; /**< Enables NPEI_INT_SUM[39] to generate an
interrupt on the RSL. */
uint64_t c0_un_b0 : 1; /**< Enables NPEI_INT_SUM[38] to generate an
interrupt on the RSL. */
uint64_t c0_up_bx : 1; /**< Enables NPEI_INT_SUM[37] to generate an
interrupt on the RSL. */
uint64_t c0_up_wi : 1; /**< Enables NPEI_INT_SUM[36] to generate an
interrupt on the RSL. */
uint64_t c0_up_b2 : 1; /**< Enables NPEI_INT_SUM[35] to generate an
interrupt on the RSL. */
uint64_t c0_up_b1 : 1; /**< Enables NPEI_INT_SUM[34] to generate an
interrupt on the RSL. */
uint64_t c0_up_b0 : 1; /**< Enables NPEI_INT_SUM[33] to generate an
interrupt on the RSL. */
uint64_t c1_hpint : 1; /**< Enables NPEI_INT_SUM[32] to generate an
interrupt on the RSL. */
uint64_t c1_pmei : 1; /**< Enables NPEI_INT_SUM[31] to generate an
interrupt on the RSL. */
uint64_t c1_wake : 1; /**< Enables NPEI_INT_SUM[30] to generate an
interrupt on the RSL. */
uint64_t crs1_dr : 1; /**< Enables NPEI_INT_SUM2[29] to generate an
interrupt on the RSL. */
uint64_t c1_se : 1; /**< Enables NPEI_INT_SUM[28] to generate an
interrupt on the RSL. */
uint64_t crs1_er : 1; /**< Enables NPEI_INT_SUM2[27] to generate an
interrupt on the RSL. */
uint64_t c1_aeri : 1; /**< Enables NPEI_INT_SUM[26] to generate an
interrupt on the RSL. */
uint64_t c0_hpint : 1; /**< Enables NPEI_INT_SUM[25] to generate an
interrupt on the RSL. */
uint64_t c0_pmei : 1; /**< Enables NPEI_INT_SUM[24] to generate an
interrupt on the RSL. */
uint64_t c0_wake : 1; /**< Enables NPEI_INT_SUM[23] to generate an
interrupt on the RSL. */
uint64_t crs0_dr : 1; /**< Enables NPEI_INT_SUM2[22] to generate an
interrupt on the RSL. */
uint64_t c0_se : 1; /**< Enables NPEI_INT_SUM[21] to generate an
interrupt on the RSL. */
uint64_t crs0_er : 1; /**< Enables NPEI_INT_SUM2[20] to generate an
interrupt on the RSL. */
uint64_t c0_aeri : 1; /**< Enables NPEI_INT_SUM[19] to generate an
interrupt on the RSL. */
uint64_t ptime : 1; /**< Enables NPEI_INT_SUM[18] to generate an
interrupt on the RSL. */
uint64_t pcnt : 1; /**< Enables NPEI_INT_SUM[17] to generate an
interrupt on the RSL. */
uint64_t pidbof : 1; /**< Enables NPEI_INT_SUM[16] to generate an
interrupt on the RSL. */
uint64_t psldbof : 1; /**< Enables NPEI_INT_SUM[15] to generate an
interrupt on the RSL. */
uint64_t dtime1 : 1; /**< Enables NPEI_INT_SUM[14] to generate an
interrupt on the RSL. */
uint64_t dtime0 : 1; /**< Enables NPEI_INT_SUM[13] to generate an
interrupt on the RSL. */
uint64_t dcnt1 : 1; /**< Enables NPEI_INT_SUM[12] to generate an
interrupt on the RSL. */
uint64_t dcnt0 : 1; /**< Enables NPEI_INT_SUM[11] to generate an
interrupt on the RSL. */
uint64_t dma1fi : 1; /**< Enables NPEI_INT_SUM[10] to generate an
interrupt on the RSL. */
uint64_t dma0fi : 1; /**< Enables NPEI_INT_SUM[9] to generate an
interrupt on the RSL. */
uint64_t dma4dbo : 1; /**< Enables NPEI_INT_SUM[8] to generate an
interrupt on the RSL. */
uint64_t dma3dbo : 1; /**< Enables NPEI_INT_SUM[7] to generate an
interrupt on the RSL. */
uint64_t dma2dbo : 1; /**< Enables NPEI_INT_SUM[6] to generate an
interrupt on the RSL. */
uint64_t dma1dbo : 1; /**< Enables NPEI_INT_SUM[5] to generate an
interrupt on the RSL. */
uint64_t dma0dbo : 1; /**< Enables NPEI_INT_SUM[4] to generate an
interrupt on the RSL. */
uint64_t iob2big : 1; /**< Enables NPEI_INT_SUM[3] to generate an
interrupt on the RSL. */
uint64_t bar0_to : 1; /**< Enables NPEI_INT_SUM[2] to generate an
interrupt on the RSL. */
uint64_t rml_wto : 1; /**< Enables NPEI_INT_SUM[1] to generate an
interrupt on the RSL. */
uint64_t rml_rto : 1; /**< Enables NPEI_INT_UM[0] to generate an
interrupt on the RSL. */
#else
uint64_t rml_rto : 1;
uint64_t rml_wto : 1;
uint64_t bar0_to : 1;
uint64_t iob2big : 1;
uint64_t dma0dbo : 1;
uint64_t dma1dbo : 1;
uint64_t dma2dbo : 1;
uint64_t dma3dbo : 1;
uint64_t dma4dbo : 1;
uint64_t dma0fi : 1;
uint64_t dma1fi : 1;
uint64_t dcnt0 : 1;
uint64_t dcnt1 : 1;
uint64_t dtime0 : 1;
uint64_t dtime1 : 1;
uint64_t psldbof : 1;
uint64_t pidbof : 1;
uint64_t pcnt : 1;
uint64_t ptime : 1;
uint64_t c0_aeri : 1;
uint64_t crs0_er : 1;
uint64_t c0_se : 1;
uint64_t crs0_dr : 1;
uint64_t c0_wake : 1;
uint64_t c0_pmei : 1;
uint64_t c0_hpint : 1;
uint64_t c1_aeri : 1;
uint64_t crs1_er : 1;
uint64_t c1_se : 1;
uint64_t crs1_dr : 1;
uint64_t c1_wake : 1;
uint64_t c1_pmei : 1;
uint64_t c1_hpint : 1;
uint64_t c0_up_b0 : 1;
uint64_t c0_up_b1 : 1;
uint64_t c0_up_b2 : 1;
uint64_t c0_up_wi : 1;
uint64_t c0_up_bx : 1;
uint64_t c0_un_b0 : 1;
uint64_t c0_un_b1 : 1;
uint64_t c0_un_b2 : 1;
uint64_t c0_un_wi : 1;
uint64_t c0_un_bx : 1;
uint64_t c1_up_b0 : 1;
uint64_t c1_up_b1 : 1;
uint64_t c1_up_b2 : 1;
uint64_t c1_up_wi : 1;
uint64_t c1_up_bx : 1;
uint64_t c1_un_b0 : 1;
uint64_t c1_un_b1 : 1;
uint64_t c1_un_b2 : 1;
uint64_t c1_un_wi : 1;
uint64_t c1_un_bx : 1;
uint64_t c0_un_wf : 1;
uint64_t c1_un_wf : 1;
uint64_t c0_up_wf : 1;
uint64_t c1_up_wf : 1;
uint64_t c0_exc : 1;
uint64_t c1_exc : 1;
uint64_t c0_ldwn : 1;
uint64_t c1_ldwn : 1;
uint64_t int_a : 1;
uint64_t reserved_62_63 : 2;
#endif
} s;
struct cvmx_npei_int_enb2_s cn52xx;
struct cvmx_npei_int_enb2_cn52xxp1
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_62_63 : 2;
uint64_t int_a : 1; /**< Enables NPEI_INT_SUM2[61] to generate an
interrupt on the RSL. */
uint64_t c1_ldwn : 1; /**< Enables NPEI_INT_SUM2[60] to generate an
interrupt on the RSL. */
uint64_t c0_ldwn : 1; /**< Enables NPEI_INT_SUM2[59] to generate an
interrupt on the RSL. */
uint64_t c1_exc : 1; /**< Enables NPEI_INT_SUM2[58] to generate an
interrupt on the RSL. */
uint64_t c0_exc : 1; /**< Enables NPEI_INT_SUM2[57] to generate an
interrupt on the RSL. */
uint64_t c1_up_wf : 1; /**< Enables NPEI_INT_SUM2[56] to generate an
interrupt on the RSL. */
uint64_t c0_up_wf : 1; /**< Enables NPEI_INT_SUM2[55] to generate an
interrupt on the RSL. */
uint64_t c1_un_wf : 1; /**< Enables NPEI_INT_SUM2[54] to generate an
interrupt on the RSL. */
uint64_t c0_un_wf : 1; /**< Enables NPEI_INT_SUM2[53] to generate an
interrupt on the RSL. */
uint64_t c1_un_bx : 1; /**< Enables NPEI_INT_SUM2[52] to generate an
interrupt on the RSL. */
uint64_t c1_un_wi : 1; /**< Enables NPEI_INT_SUM2[51] to generate an
interrupt on the RSL. */
uint64_t c1_un_b2 : 1; /**< Enables NPEI_INT_SUM2[50] to generate an
interrupt on the RSL. */
uint64_t c1_un_b1 : 1; /**< Enables NPEI_INT_SUM2[49] to generate an
interrupt on the RSL. */
uint64_t c1_un_b0 : 1; /**< Enables NPEI_INT_SUM2[48] to generate an
interrupt on the RSL. */
uint64_t c1_up_bx : 1; /**< Enables NPEI_INT_SUM2[47] to generate an
interrupt on the RSL. */
uint64_t c1_up_wi : 1; /**< Enables NPEI_INT_SUM2[46] to generate an
interrupt on the RSL. */
uint64_t c1_up_b2 : 1; /**< Enables NPEI_INT_SUM2[45] to generate an
interrupt on the RSL. */
uint64_t c1_up_b1 : 1; /**< Enables NPEI_INT_SUM2[44] to generate an
interrupt on the RSL. */
uint64_t c1_up_b0 : 1; /**< Enables NPEI_INT_SUM2[43] to generate an
interrupt on the RSL. */
uint64_t c0_un_bx : 1; /**< Enables NPEI_INT_SUM2[42] to generate an
interrupt on the RSL. */
uint64_t c0_un_wi : 1; /**< Enables NPEI_INT_SUM2[41] to generate an
interrupt on the RSL. */
uint64_t c0_un_b2 : 1; /**< Enables NPEI_INT_SUM2[40] to generate an
interrupt on the RSL. */
uint64_t c0_un_b1 : 1; /**< Enables NPEI_INT_SUM2[39] to generate an
interrupt on the RSL. */
uint64_t c0_un_b0 : 1; /**< Enables NPEI_INT_SUM2[38] to generate an
interrupt on the RSL. */
uint64_t c0_up_bx : 1; /**< Enables NPEI_INT_SUM2[37] to generate an
interrupt on the RSL. */
uint64_t c0_up_wi : 1; /**< Enables NPEI_INT_SUM2[36] to generate an
interrupt on the RSL. */
uint64_t c0_up_b2 : 1; /**< Enables NPEI_INT_SUM2[35] to generate an
interrupt on the RSL. */
uint64_t c0_up_b1 : 1; /**< Enables NPEI_INT_SUM2[34] to generate an
interrupt on the RSL. */
uint64_t c0_up_b0 : 1; /**< Enables NPEI_INT_SUM2[33] to generate an
interrupt on the RSL. */
uint64_t c1_hpint : 1; /**< Enables NPEI_INT_SUM2[32] to generate an
interrupt on the RSL. */
uint64_t c1_pmei : 1; /**< Enables NPEI_INT_SUM2[31] to generate an
interrupt on the RSL. */
uint64_t c1_wake : 1; /**< Enables NPEI_INT_SUM2[30] to generate an
interrupt on the RSL. */
uint64_t crs1_dr : 1; /**< Enables NPEI_INT_SUM2[29] to generate an
interrupt on the RSL. */
uint64_t c1_se : 1; /**< Enables NPEI_INT_SUM2[28] to generate an
interrupt on the RSL. */
uint64_t crs1_er : 1; /**< Enables NPEI_INT_SUM2[27] to generate an
interrupt on the RSL. */
uint64_t c1_aeri : 1; /**< Enables NPEI_INT_SUM2[26] to generate an
interrupt on the RSL. */
uint64_t c0_hpint : 1; /**< Enables NPEI_INT_SUM2[25] to generate an
interrupt on the RSL. */
uint64_t c0_pmei : 1; /**< Enables NPEI_INT_SUM2[24] to generate an
interrupt on the RSL. */
uint64_t c0_wake : 1; /**< Enables NPEI_INT_SUM2[23] to generate an
interrupt on the RSL. */
uint64_t crs0_dr : 1; /**< Enables NPEI_INT_SUM2[22] to generate an
interrupt on the RSL. */
uint64_t c0_se : 1; /**< Enables NPEI_INT_SUM2[21] to generate an
interrupt on the RSL. */
uint64_t crs0_er : 1; /**< Enables NPEI_INT_SUM2[20] to generate an
interrupt on the RSL. */
uint64_t c0_aeri : 1; /**< Enables NPEI_INT_SUM2[19] to generate an
interrupt on the RSL. */
uint64_t ptime : 1; /**< Enables NPEI_INT_SUM2[18] to generate an
interrupt on the RSL. */
uint64_t pcnt : 1; /**< Enables NPEI_INT_SUM2[17] to generate an
interrupt on the RSL. */
uint64_t pidbof : 1; /**< Enables NPEI_INT_SUM2[16] to generate an
interrupt on the RSL. */
uint64_t psldbof : 1; /**< Enables NPEI_INT_SUM2[15] to generate an
interrupt on the RSL. */
uint64_t dtime1 : 1; /**< Enables NPEI_INT_SUM2[14] to generate an
interrupt on the RSL. */
uint64_t dtime0 : 1; /**< Enables NPEI_INT_SUM2[13] to generate an
interrupt on the RSL. */
uint64_t dcnt1 : 1; /**< Enables NPEI_INT_SUM2[12] to generate an
interrupt on the RSL. */
uint64_t dcnt0 : 1; /**< Enables NPEI_INT_SUM2[11] to generate an
interrupt on the RSL. */
uint64_t dma1fi : 1; /**< Enables NPEI_INT_SUM2[10] to generate an
interrupt on the RSL. */
uint64_t dma0fi : 1; /**< Enables NPEI_INT_SUM2[9] to generate an
interrupt on the RSL. */
uint64_t reserved_8_8 : 1;
uint64_t dma3dbo : 1; /**< Enables NPEI_INT_SUM2[7] to generate an
interrupt on the RSL. */
uint64_t dma2dbo : 1; /**< Enables NPEI_INT_SUM2[6] to generate an
interrupt on the RSL. */
uint64_t dma1dbo : 1; /**< Enables NPEI_INT_SUM2[5] to generate an
interrupt on the RSL. */
uint64_t dma0dbo : 1; /**< Enables NPEI_INT_SUM2[4] to generate an
interrupt on the RSL. */
uint64_t iob2big : 1; /**< Enables NPEI_INT_SUM2[3] to generate an
interrupt on the RSL. */
uint64_t bar0_to : 1; /**< Enables NPEI_INT_SUM2[2] to generate an
interrupt on the RSL. */
uint64_t rml_wto : 1; /**< Enables NPEI_INT_SUM2[1] to generate an
interrupt on the RSL. */
uint64_t rml_rto : 1; /**< Enables NPEI_INT_SUM2[0] to generate an
interrupt on the RSL. */
#else
uint64_t rml_rto : 1;
uint64_t rml_wto : 1;
uint64_t bar0_to : 1;
uint64_t iob2big : 1;
uint64_t dma0dbo : 1;
uint64_t dma1dbo : 1;
uint64_t dma2dbo : 1;
uint64_t dma3dbo : 1;
uint64_t reserved_8_8 : 1;
uint64_t dma0fi : 1;
uint64_t dma1fi : 1;
uint64_t dcnt0 : 1;
uint64_t dcnt1 : 1;
uint64_t dtime0 : 1;
uint64_t dtime1 : 1;
uint64_t psldbof : 1;
uint64_t pidbof : 1;
uint64_t pcnt : 1;
uint64_t ptime : 1;
uint64_t c0_aeri : 1;
uint64_t crs0_er : 1;
uint64_t c0_se : 1;
uint64_t crs0_dr : 1;
uint64_t c0_wake : 1;
uint64_t c0_pmei : 1;
uint64_t c0_hpint : 1;
uint64_t c1_aeri : 1;
uint64_t crs1_er : 1;
uint64_t c1_se : 1;
uint64_t crs1_dr : 1;
uint64_t c1_wake : 1;
uint64_t c1_pmei : 1;
uint64_t c1_hpint : 1;
uint64_t c0_up_b0 : 1;
uint64_t c0_up_b1 : 1;
uint64_t c0_up_b2 : 1;
uint64_t c0_up_wi : 1;
uint64_t c0_up_bx : 1;
uint64_t c0_un_b0 : 1;
uint64_t c0_un_b1 : 1;
uint64_t c0_un_b2 : 1;
uint64_t c0_un_wi : 1;
uint64_t c0_un_bx : 1;
uint64_t c1_up_b0 : 1;
uint64_t c1_up_b1 : 1;
uint64_t c1_up_b2 : 1;
uint64_t c1_up_wi : 1;
uint64_t c1_up_bx : 1;
uint64_t c1_un_b0 : 1;
uint64_t c1_un_b1 : 1;
uint64_t c1_un_b2 : 1;
uint64_t c1_un_wi : 1;
uint64_t c1_un_bx : 1;
uint64_t c0_un_wf : 1;
uint64_t c1_un_wf : 1;
uint64_t c0_up_wf : 1;
uint64_t c1_up_wf : 1;
uint64_t c0_exc : 1;
uint64_t c1_exc : 1;
uint64_t c0_ldwn : 1;
uint64_t c1_ldwn : 1;
uint64_t int_a : 1;
uint64_t reserved_62_63 : 2;
#endif
} cn52xxp1;
struct cvmx_npei_int_enb2_s cn56xx;
struct cvmx_npei_int_enb2_cn56xxp1
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_61_63 : 3;
uint64_t c1_ldwn : 1; /**< Enables NPEI_INT_SUM[60] to generate an
interrupt on the RSL. */
uint64_t c0_ldwn : 1; /**< Enables NPEI_INT_SUM[59] to generate an
interrupt on the RSL. */
uint64_t c1_exc : 1; /**< Enables NPEI_INT_SUM[58] to generate an
interrupt on the RSL. */
uint64_t c0_exc : 1; /**< Enables NPEI_INT_SUM[57] to generate an
interrupt on the RSL. */
uint64_t c1_up_wf : 1; /**< Enables NPEI_INT_SUM[56] to generate an
interrupt on the RSL. */
uint64_t c0_up_wf : 1; /**< Enables NPEI_INT_SUM[55] to generate an
interrupt on the RSL. */
uint64_t c1_un_wf : 1; /**< Enables NPEI_INT_SUM[54] to generate an
interrupt on the RSL. */
uint64_t c0_un_wf : 1; /**< Enables NPEI_INT_SUM[53] to generate an
interrupt on the RSL. */
uint64_t c1_un_bx : 1; /**< Enables NPEI_INT_SUM[52] to generate an
interrupt on the RSL. */
uint64_t c1_un_wi : 1; /**< Enables NPEI_INT_SUM[51] to generate an
interrupt on the RSL. */
uint64_t c1_un_b2 : 1; /**< Enables NPEI_INT_SUM[50] to generate an
interrupt on the RSL. */
uint64_t c1_un_b1 : 1; /**< Enables NPEI_INT_SUM[49] to generate an
interrupt on the RSL. */
uint64_t c1_un_b0 : 1; /**< Enables NPEI_INT_SUM[48] to generate an
interrupt on the RSL. */
uint64_t c1_up_bx : 1; /**< Enables NPEI_INT_SUM[47] to generate an
interrupt on the RSL. */
uint64_t c1_up_wi : 1; /**< Enables NPEI_INT_SUM[46] to generate an
interrupt on the RSL. */
uint64_t c1_up_b2 : 1; /**< Enables NPEI_INT_SUM[45] to generate an
interrupt on the RSL. */
uint64_t c1_up_b1 : 1; /**< Enables NPEI_INT_SUM[44] to generate an
interrupt on the RSL. */
uint64_t c1_up_b0 : 1; /**< Enables NPEI_INT_SUM[43] to generate an
interrupt on the RSL. */
uint64_t c0_un_bx : 1; /**< Enables NPEI_INT_SUM[42] to generate an
interrupt on the RSL. */
uint64_t c0_un_wi : 1; /**< Enables NPEI_INT_SUM[41] to generate an
interrupt on the RSL. */
uint64_t c0_un_b2 : 1; /**< Enables NPEI_INT_SUM[40] to generate an
interrupt on the RSL. */
uint64_t c0_un_b1 : 1; /**< Enables NPEI_INT_SUM[39] to generate an
interrupt on the RSL. */
uint64_t c0_un_b0 : 1; /**< Enables NPEI_INT_SUM[38] to generate an
interrupt on the RSL. */
uint64_t c0_up_bx : 1; /**< Enables NPEI_INT_SUM[37] to generate an
interrupt on the RSL. */
uint64_t c0_up_wi : 1; /**< Enables NPEI_INT_SUM[36] to generate an
interrupt on the RSL. */
uint64_t c0_up_b2 : 1; /**< Enables NPEI_INT_SUM[35] to generate an
interrupt on the RSL. */
uint64_t c0_up_b1 : 1; /**< Enables NPEI_INT_SUM[34] to generate an
interrupt on the RSL. */
uint64_t c0_up_b0 : 1; /**< Enables NPEI_INT_SUM[33] to generate an
interrupt on the RSL. */
uint64_t c1_hpint : 1; /**< Enables NPEI_INT_SUM[32] to generate an
interrupt on the RSL. */
uint64_t c1_pmei : 1; /**< Enables NPEI_INT_SUM[31] to generate an
interrupt on the RSL. */
uint64_t c1_wake : 1; /**< Enables NPEI_INT_SUM[30] to generate an
interrupt on the RSL. */
uint64_t reserved_29_29 : 1;
uint64_t c1_se : 1; /**< Enables NPEI_INT_SUM[28] to generate an
interrupt on the RSL. */
uint64_t reserved_27_27 : 1;
uint64_t c1_aeri : 1; /**< Enables NPEI_INT_SUM[26] to generate an
interrupt on the RSL. */
uint64_t c0_hpint : 1; /**< Enables NPEI_INT_SUM[25] to generate an
interrupt on the RSL. */
uint64_t c0_pmei : 1; /**< Enables NPEI_INT_SUM[24] to generate an
interrupt on the RSL. */
uint64_t c0_wake : 1; /**< Enables NPEI_INT_SUM[23] to generate an
interrupt on the RSL. */
uint64_t reserved_22_22 : 1;
uint64_t c0_se : 1; /**< Enables NPEI_INT_SUM[21] to generate an
interrupt on the RSL. */
uint64_t reserved_20_20 : 1;
uint64_t c0_aeri : 1; /**< Enables NPEI_INT_SUM[19] to generate an
interrupt on the RSL. */
uint64_t ptime : 1; /**< Enables NPEI_INT_SUM[18] to generate an
interrupt on the RSL. */
uint64_t pcnt : 1; /**< Enables NPEI_INT_SUM[17] to generate an
interrupt on the RSL. */
uint64_t pidbof : 1; /**< Enables NPEI_INT_SUM[16] to generate an
interrupt on the RSL. */
uint64_t psldbof : 1; /**< Enables NPEI_INT_SUM[15] to generate an
interrupt on the RSL. */
uint64_t dtime1 : 1; /**< Enables NPEI_INT_SUM[14] to generate an
interrupt on the RSL. */
uint64_t dtime0 : 1; /**< Enables NPEI_INT_SUM[13] to generate an
interrupt on the RSL. */
uint64_t dcnt1 : 1; /**< Enables NPEI_INT_SUM[12] to generate an
interrupt on the RSL. */
uint64_t dcnt0 : 1; /**< Enables NPEI_INT_SUM[11] to generate an
interrupt on the RSL. */
uint64_t dma1fi : 1; /**< Enables NPEI_INT_SUM[10] to generate an
interrupt on the RSL. */
uint64_t dma0fi : 1; /**< Enables NPEI_INT_SUM[9] to generate an
interrupt on the RSL. */
uint64_t dma4dbo : 1; /**< Enables NPEI_INT_SUM[8] to generate an
interrupt on the RSL. */
uint64_t dma3dbo : 1; /**< Enables NPEI_INT_SUM[7] to generate an
interrupt on the RSL. */
uint64_t dma2dbo : 1; /**< Enables NPEI_INT_SUM[6] to generate an
interrupt on the RSL. */
uint64_t dma1dbo : 1; /**< Enables NPEI_INT_SUM[5] to generate an
interrupt on the RSL. */
uint64_t dma0dbo : 1; /**< Enables NPEI_INT_SUM[4] to generate an
interrupt on the RSL. */
uint64_t iob2big : 1; /**< Enables NPEI_INT_SUM[3] to generate an
interrupt on the RSL. */
uint64_t bar0_to : 1; /**< Enables NPEI_INT_SUM[2] to generate an
interrupt on the RSL. */
uint64_t rml_wto : 1; /**< Enables NPEI_INT_SUM[1] to generate an
interrupt on the RSL. */
uint64_t rml_rto : 1; /**< Enables NPEI_INT_UM[0] to generate an
interrupt on the RSL. */
#else
uint64_t rml_rto : 1;
uint64_t rml_wto : 1;
uint64_t bar0_to : 1;
uint64_t iob2big : 1;
uint64_t dma0dbo : 1;
uint64_t dma1dbo : 1;
uint64_t dma2dbo : 1;
uint64_t dma3dbo : 1;
uint64_t dma4dbo : 1;
uint64_t dma0fi : 1;
uint64_t dma1fi : 1;
uint64_t dcnt0 : 1;
uint64_t dcnt1 : 1;
uint64_t dtime0 : 1;
uint64_t dtime1 : 1;
uint64_t psldbof : 1;
uint64_t pidbof : 1;
uint64_t pcnt : 1;
uint64_t ptime : 1;
uint64_t c0_aeri : 1;
uint64_t reserved_20_20 : 1;
uint64_t c0_se : 1;
uint64_t reserved_22_22 : 1;
uint64_t c0_wake : 1;
uint64_t c0_pmei : 1;
uint64_t c0_hpint : 1;
uint64_t c1_aeri : 1;
uint64_t reserved_27_27 : 1;
uint64_t c1_se : 1;
uint64_t reserved_29_29 : 1;
uint64_t c1_wake : 1;
uint64_t c1_pmei : 1;
uint64_t c1_hpint : 1;
uint64_t c0_up_b0 : 1;
uint64_t c0_up_b1 : 1;
uint64_t c0_up_b2 : 1;
uint64_t c0_up_wi : 1;
uint64_t c0_up_bx : 1;
uint64_t c0_un_b0 : 1;
uint64_t c0_un_b1 : 1;
uint64_t c0_un_b2 : 1;
uint64_t c0_un_wi : 1;
uint64_t c0_un_bx : 1;
uint64_t c1_up_b0 : 1;
uint64_t c1_up_b1 : 1;
uint64_t c1_up_b2 : 1;
uint64_t c1_up_wi : 1;
uint64_t c1_up_bx : 1;
uint64_t c1_un_b0 : 1;
uint64_t c1_un_b1 : 1;
uint64_t c1_un_b2 : 1;
uint64_t c1_un_wi : 1;
uint64_t c1_un_bx : 1;
uint64_t c0_un_wf : 1;
uint64_t c1_un_wf : 1;
uint64_t c0_up_wf : 1;
uint64_t c1_up_wf : 1;
uint64_t c0_exc : 1;
uint64_t c1_exc : 1;
uint64_t c0_ldwn : 1;
uint64_t c1_ldwn : 1;
uint64_t reserved_61_63 : 3;
#endif
} cn56xxp1;
};
typedef union cvmx_npei_int_enb2 cvmx_npei_int_enb2_t;
/**
* cvmx_npei_int_info
*
* NPEI_INT_INFO = NPI Interrupt Information
*
* Contains information about some of the interrupt condition that can occur in the NPEI_INTERRUPT_SUM register.
*/
union cvmx_npei_int_info
{
uint64_t u64;
struct cvmx_npei_int_info_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_12_63 : 52;
uint64_t pidbof : 6; /**< Field set when the NPEI_INTERRUPT_SUM[PIDBOF] bit
is set. This field when set will not change again
unitl NPEI_INTERRUPT_SUM[PIDBOF] is cleared. */
uint64_t psldbof : 6; /**< Field set when the NPEI_INTERRUPT_SUM[PSLDBOF] bit
is set. This field when set will not change again
unitl NPEI_INTERRUPT_SUM[PSLDBOF] is cleared. */
#else
uint64_t psldbof : 6;
uint64_t pidbof : 6;
uint64_t reserved_12_63 : 52;
#endif
} s;
struct cvmx_npei_int_info_s cn52xx;
struct cvmx_npei_int_info_s cn56xx;
struct cvmx_npei_int_info_s cn56xxp1;
};
typedef union cvmx_npei_int_info cvmx_npei_int_info_t;
/**
* cvmx_npei_int_sum
*
* NPEI_INTERRUPT_SUM = NPI Interrupt Summary Register
*
* Set when an interrupt condition occurs, write '1' to clear.
*
* HACK: These used to exist, how are TO handled?
* <3> PO0_2SML R/W1C 0x0 0 The packet being sent out on Port0 is smaller $R NS
* than the NPI_BUFF_SIZE_OUTPUT0[ISIZE] field.
* <7> I0_RTOUT R/W1C 0x0 0 Port-0 had a read timeout while attempting to $R NS
* read instructions.
* <15> P0_RTOUT R/W1C 0x0 0 Port-0 had a read timeout while attempting to $R NS
* read packet data.
* <23> G0_RTOUT R/W1C 0x0 0 Port-0 had a read timeout while attempting to $R NS
* read a gather list.
* <31> P0_PTOUT R/W1C 0x0 0 Port-0 output had a read timeout on a DATA/INFO $R NS
* pair.
*/
union cvmx_npei_int_sum
{
uint64_t u64;
struct cvmx_npei_int_sum_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t mio_inta : 1; /**< Interrupt from MIO. */
uint64_t reserved_62_62 : 1;
uint64_t int_a : 1; /**< Set when a bit in the NPEI_INT_A_SUM register and
the cooresponding bit in the NPEI_INT_A_ENB
register is set. */
uint64_t c1_ldwn : 1; /**< Reset request due to link1 down status. */
uint64_t c0_ldwn : 1; /**< Reset request due to link0 down status. */
uint64_t c1_exc : 1; /**< Set when the PESC1_DBG_INFO register has a bit
set and its cooresponding PESC1_DBG_INFO_EN bit
is set. */
uint64_t c0_exc : 1; /**< Set when the PESC0_DBG_INFO register has a bit
set and its cooresponding PESC0_DBG_INFO_EN bit
is set. */
uint64_t c1_up_wf : 1; /**< Received Unsupported P-TLP for filtered window
register. Core1. */
uint64_t c0_up_wf : 1; /**< Received Unsupported P-TLP for filtered window
register. Core0. */
uint64_t c1_un_wf : 1; /**< Received Unsupported N-TLP for filtered window
register. Core1. */
uint64_t c0_un_wf : 1; /**< Received Unsupported N-TLP for filtered window
register. Core0. */
uint64_t c1_un_bx : 1; /**< Received Unsupported N-TLP for unknown Bar.
Core 1. */
uint64_t c1_un_wi : 1; /**< Received Unsupported N-TLP for Window Register.
Core 1. */
uint64_t c1_un_b2 : 1; /**< Received Unsupported N-TLP for Bar2.
Core 1. */
uint64_t c1_un_b1 : 1; /**< Received Unsupported N-TLP for Bar1.
Core 1. */
uint64_t c1_un_b0 : 1; /**< Received Unsupported N-TLP for Bar0.
Core 1. */
uint64_t c1_up_bx : 1; /**< Received Unsupported P-TLP for unknown Bar.
Core 1. */
uint64_t c1_up_wi : 1; /**< Received Unsupported P-TLP for Window Register.
Core 1. */
uint64_t c1_up_b2 : 1; /**< Received Unsupported P-TLP for Bar2.
Core 1. */
uint64_t c1_up_b1 : 1; /**< Received Unsupported P-TLP for Bar1.
Core 1. */
uint64_t c1_up_b0 : 1; /**< Received Unsupported P-TLP for Bar0.
Core 1. */
uint64_t c0_un_bx : 1; /**< Received Unsupported N-TLP for unknown Bar.
Core 0. */
uint64_t c0_un_wi : 1; /**< Received Unsupported N-TLP for Window Register.
Core 0. */
uint64_t c0_un_b2 : 1; /**< Received Unsupported N-TLP for Bar2.
Core 0. */
uint64_t c0_un_b1 : 1; /**< Received Unsupported N-TLP for Bar1.
Core 0. */
uint64_t c0_un_b0 : 1; /**< Received Unsupported N-TLP for Bar0.
Core 0. */
uint64_t c0_up_bx : 1; /**< Received Unsupported P-TLP for unknown Bar.
Core 0. */
uint64_t c0_up_wi : 1; /**< Received Unsupported P-TLP for Window Register.
Core 0. */
uint64_t c0_up_b2 : 1; /**< Received Unsupported P-TLP for Bar2.
Core 0. */
uint64_t c0_up_b1 : 1; /**< Received Unsupported P-TLP for Bar1.
Core 0. */
uint64_t c0_up_b0 : 1; /**< Received Unsupported P-TLP for Bar0.
Core 0. */
uint64_t c1_hpint : 1; /**< Hot-Plug Interrupt.
Pcie Core 1 (hp_int).
This interrupt will only be generated when
PCIERC1_CFG034[DLLS_C] is generated. Hot plug is
not supported. */
uint64_t c1_pmei : 1; /**< PME Interrupt.
Pcie Core 1. (cfg_pme_int) */
uint64_t c1_wake : 1; /**< Wake up from Power Management Unit.
Pcie Core 1. (wake_n)
Octeon will never generate this interrupt. */
uint64_t crs1_dr : 1; /**< Had a CRS when Retries were disabled. */
uint64_t c1_se : 1; /**< System Error, RC Mode Only.
Pcie Core 1. (cfg_sys_err_rc) */
uint64_t crs1_er : 1; /**< Had a CRS Timeout when Retries were enabled. */
uint64_t c1_aeri : 1; /**< Advanced Error Reporting Interrupt, RC Mode Only.
Pcie Core 1. */
uint64_t c0_hpint : 1; /**< Hot-Plug Interrupt.
Pcie Core 0 (hp_int).
This interrupt will only be generated when
PCIERC0_CFG034[DLLS_C] is generated. Hot plug is
not supported. */
uint64_t c0_pmei : 1; /**< PME Interrupt.
Pcie Core 0. (cfg_pme_int) */
uint64_t c0_wake : 1; /**< Wake up from Power Management Unit.
Pcie Core 0. (wake_n)
Octeon will never generate this interrupt. */
uint64_t crs0_dr : 1; /**< Had a CRS when Retries were disabled. */
uint64_t c0_se : 1; /**< System Error, RC Mode Only.
Pcie Core 0. (cfg_sys_err_rc) */
uint64_t crs0_er : 1; /**< Had a CRS Timeout when Retries were enabled. */
uint64_t c0_aeri : 1; /**< Advanced Error Reporting Interrupt, RC Mode Only.
Pcie Core 0 (cfg_aer_rc_err_int). */
uint64_t ptime : 1; /**< Packet Timer has an interrupt. Which rings can
be found in NPEI_PKT_TIME_INT. */
uint64_t pcnt : 1; /**< Packet Counter has an interrupt. Which rings can
be found in NPEI_PKT_CNT_INT. */
uint64_t pidbof : 1; /**< Packet Instruction Doorbell count overflowed. Which
doorbell can be found in NPEI_INT_INFO[PIDBOF] */
uint64_t psldbof : 1; /**< Packet Scatterlist Doorbell count overflowed. Which
doorbell can be found in NPEI_INT_INFO[PSLDBOF] */
uint64_t dtime1 : 1; /**< Whenever NPEI_DMA_CNTS[DMA1] is not 0, the
DMA_CNT1 timer increments every core clock. When
DMA_CNT1 timer exceeds NPEI_DMA1_INT_LEVEL[TIME],
this bit is set. Writing a '1' to this bit also
clears the DMA_CNT1 timer. */
uint64_t dtime0 : 1; /**< Whenever NPEI_DMA_CNTS[DMA0] is not 0, the
DMA_CNT0 timer increments every core clock. When
DMA_CNT0 timer exceeds NPEI_DMA0_INT_LEVEL[TIME],
this bit is set. Writing a '1' to this bit also
clears the DMA_CNT0 timer. */
uint64_t dcnt1 : 1; /**< This bit indicates that NPEI_DMA_CNTS[DMA1] was/is
greater than NPEI_DMA1_INT_LEVEL[CNT]. */
uint64_t dcnt0 : 1; /**< This bit indicates that NPEI_DMA_CNTS[DMA0] was/is
greater than NPEI_DMA0_INT_LEVEL[CNT]. */
uint64_t dma1fi : 1; /**< DMA0 set Forced Interrupt. */
uint64_t dma0fi : 1; /**< DMA0 set Forced Interrupt. */
uint64_t dma4dbo : 1; /**< DMA4 doorbell overflow.
Bit[32] of the doorbell count was set. */
uint64_t dma3dbo : 1; /**< DMA3 doorbell overflow.
Bit[32] of the doorbell count was set. */
uint64_t dma2dbo : 1; /**< DMA2 doorbell overflow.
Bit[32] of the doorbell count was set. */
uint64_t dma1dbo : 1; /**< DMA1 doorbell overflow.
Bit[32] of the doorbell count was set. */
uint64_t dma0dbo : 1; /**< DMA0 doorbell overflow.
Bit[32] of the doorbell count was set. */
uint64_t iob2big : 1; /**< A requested IOBDMA is to large. */
uint64_t bar0_to : 1; /**< BAR0 R/W to a NCB device did not receive
read-data/commit in 0xffff core clocks. */
uint64_t rml_wto : 1; /**< RML write did not get commit in 0xffff core clocks. */
uint64_t rml_rto : 1; /**< RML read did not return data in 0xffff core clocks. */
#else
uint64_t rml_rto : 1;
uint64_t rml_wto : 1;
uint64_t bar0_to : 1;
uint64_t iob2big : 1;
uint64_t dma0dbo : 1;
uint64_t dma1dbo : 1;
uint64_t dma2dbo : 1;
uint64_t dma3dbo : 1;
uint64_t dma4dbo : 1;
uint64_t dma0fi : 1;
uint64_t dma1fi : 1;
uint64_t dcnt0 : 1;
uint64_t dcnt1 : 1;
uint64_t dtime0 : 1;
uint64_t dtime1 : 1;
uint64_t psldbof : 1;
uint64_t pidbof : 1;
uint64_t pcnt : 1;
uint64_t ptime : 1;
uint64_t c0_aeri : 1;
uint64_t crs0_er : 1;
uint64_t c0_se : 1;
uint64_t crs0_dr : 1;
uint64_t c0_wake : 1;
uint64_t c0_pmei : 1;
uint64_t c0_hpint : 1;
uint64_t c1_aeri : 1;
uint64_t crs1_er : 1;
uint64_t c1_se : 1;
uint64_t crs1_dr : 1;
uint64_t c1_wake : 1;
uint64_t c1_pmei : 1;
uint64_t c1_hpint : 1;
uint64_t c0_up_b0 : 1;
uint64_t c0_up_b1 : 1;
uint64_t c0_up_b2 : 1;
uint64_t c0_up_wi : 1;
uint64_t c0_up_bx : 1;
uint64_t c0_un_b0 : 1;
uint64_t c0_un_b1 : 1;
uint64_t c0_un_b2 : 1;
uint64_t c0_un_wi : 1;
uint64_t c0_un_bx : 1;
uint64_t c1_up_b0 : 1;
uint64_t c1_up_b1 : 1;
uint64_t c1_up_b2 : 1;
uint64_t c1_up_wi : 1;
uint64_t c1_up_bx : 1;
uint64_t c1_un_b0 : 1;
uint64_t c1_un_b1 : 1;
uint64_t c1_un_b2 : 1;
uint64_t c1_un_wi : 1;
uint64_t c1_un_bx : 1;
uint64_t c0_un_wf : 1;
uint64_t c1_un_wf : 1;
uint64_t c0_up_wf : 1;
uint64_t c1_up_wf : 1;
uint64_t c0_exc : 1;
uint64_t c1_exc : 1;
uint64_t c0_ldwn : 1;
uint64_t c1_ldwn : 1;
uint64_t int_a : 1;
uint64_t reserved_62_62 : 1;
uint64_t mio_inta : 1;
#endif
} s;
struct cvmx_npei_int_sum_s cn52xx;
struct cvmx_npei_int_sum_cn52xxp1
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t mio_inta : 1; /**< Interrupt from MIO. */
uint64_t reserved_62_62 : 1;
uint64_t int_a : 1; /**< Set when a bit in the NPEI_INT_A_SUM register and
the cooresponding bit in the NPEI_INT_A_ENB
register is set. */
uint64_t c1_ldwn : 1; /**< Reset request due to link1 down status. */
uint64_t c0_ldwn : 1; /**< Reset request due to link0 down status. */
uint64_t c1_exc : 1; /**< Set when the PESC1_DBG_INFO register has a bit
set and its cooresponding PESC1_DBG_INFO_EN bit
is set. */
uint64_t c0_exc : 1; /**< Set when the PESC0_DBG_INFO register has a bit
set and its cooresponding PESC0_DBG_INFO_EN bit
is set. */
uint64_t c1_up_wf : 1; /**< Received Unsupported P-TLP for filtered window
register. Core1. */
uint64_t c0_up_wf : 1; /**< Received Unsupported P-TLP for filtered window
register. Core0. */
uint64_t c1_un_wf : 1; /**< Received Unsupported N-TLP for filtered window
register. Core1. */
uint64_t c0_un_wf : 1; /**< Received Unsupported N-TLP for filtered window
register. Core0. */
uint64_t c1_un_bx : 1; /**< Received Unsupported N-TLP for unknown Bar.
Core 1. */
uint64_t c1_un_wi : 1; /**< Received Unsupported N-TLP for Window Register.
Core 1. */
uint64_t c1_un_b2 : 1; /**< Received Unsupported N-TLP for Bar2.
Core 1. */
uint64_t c1_un_b1 : 1; /**< Received Unsupported N-TLP for Bar1.
Core 1. */
uint64_t c1_un_b0 : 1; /**< Received Unsupported N-TLP for Bar0.
Core 1. */
uint64_t c1_up_bx : 1; /**< Received Unsupported P-TLP for unknown Bar.
Core 1. */
uint64_t c1_up_wi : 1; /**< Received Unsupported P-TLP for Window Register.
Core 1. */
uint64_t c1_up_b2 : 1; /**< Received Unsupported P-TLP for Bar2.
Core 1. */
uint64_t c1_up_b1 : 1; /**< Received Unsupported P-TLP for Bar1.
Core 1. */
uint64_t c1_up_b0 : 1; /**< Received Unsupported P-TLP for Bar0.
Core 1. */
uint64_t c0_un_bx : 1; /**< Received Unsupported N-TLP for unknown Bar.
Core 0. */
uint64_t c0_un_wi : 1; /**< Received Unsupported N-TLP for Window Register.
Core 0. */
uint64_t c0_un_b2 : 1; /**< Received Unsupported N-TLP for Bar2.
Core 0. */
uint64_t c0_un_b1 : 1; /**< Received Unsupported N-TLP for Bar1.
Core 0. */
uint64_t c0_un_b0 : 1; /**< Received Unsupported N-TLP for Bar0.
Core 0. */
uint64_t c0_up_bx : 1; /**< Received Unsupported P-TLP for unknown Bar.
Core 0. */
uint64_t c0_up_wi : 1; /**< Received Unsupported P-TLP for Window Register.
Core 0. */
uint64_t c0_up_b2 : 1; /**< Received Unsupported P-TLP for Bar2.
Core 0. */
uint64_t c0_up_b1 : 1; /**< Received Unsupported P-TLP for Bar1.
Core 0. */
uint64_t c0_up_b0 : 1; /**< Received Unsupported P-TLP for Bar0.
Core 0. */
uint64_t c1_hpint : 1; /**< Hot-Plug Interrupt.
Pcie Core 1 (hp_int).
This interrupt will only be generated when
PCIERC1_CFG034[DLLS_C] is generated. Hot plug is
not supported. */
uint64_t c1_pmei : 1; /**< PME Interrupt.
Pcie Core 1. (cfg_pme_int) */
uint64_t c1_wake : 1; /**< Wake up from Power Management Unit.
Pcie Core 1. (wake_n)
Octeon will never generate this interrupt. */
uint64_t crs1_dr : 1; /**< Had a CRS when Retries were disabled. */
uint64_t c1_se : 1; /**< System Error, RC Mode Only.
Pcie Core 1. (cfg_sys_err_rc) */
uint64_t crs1_er : 1; /**< Had a CRS Timeout when Retries were enabled. */
uint64_t c1_aeri : 1; /**< Advanced Error Reporting Interrupt, RC Mode Only.
Pcie Core 1. */
uint64_t c0_hpint : 1; /**< Hot-Plug Interrupt.
Pcie Core 0 (hp_int).
This interrupt will only be generated when
PCIERC0_CFG034[DLLS_C] is generated. Hot plug is
not supported. */
uint64_t c0_pmei : 1; /**< PME Interrupt.
Pcie Core 0. (cfg_pme_int) */
uint64_t c0_wake : 1; /**< Wake up from Power Management Unit.
Pcie Core 0. (wake_n)
Octeon will never generate this interrupt. */
uint64_t crs0_dr : 1; /**< Had a CRS when Retries were disabled. */
uint64_t c0_se : 1; /**< System Error, RC Mode Only.
Pcie Core 0. (cfg_sys_err_rc) */
uint64_t crs0_er : 1; /**< Had a CRS Timeout when Retries were enabled. */
uint64_t c0_aeri : 1; /**< Advanced Error Reporting Interrupt, RC Mode Only.
Pcie Core 0 (cfg_aer_rc_err_int). */
uint64_t reserved_15_18 : 4;
uint64_t dtime1 : 1; /**< Whenever NPEI_DMA_CNTS[DMA1] is not 0, the
DMA_CNT1 timer increments every core clock. When
DMA_CNT1 timer exceeds NPEI_DMA1_INT_LEVEL[TIME],
this bit is set. Writing a '1' to this bit also
clears the DMA_CNT1 timer. */
uint64_t dtime0 : 1; /**< Whenever NPEI_DMA_CNTS[DMA0] is not 0, the
DMA_CNT0 timer increments every core clock. When
DMA_CNT0 timer exceeds NPEI_DMA0_INT_LEVEL[TIME],
this bit is set. Writing a '1' to this bit also
clears the DMA_CNT0 timer. */
uint64_t dcnt1 : 1; /**< This bit indicates that NPEI_DMA_CNTS[DMA1] was/is
greater than NPEI_DMA1_INT_LEVEL[CNT]. */
uint64_t dcnt0 : 1; /**< This bit indicates that NPEI_DMA_CNTS[DMA0] was/is
greater than NPEI_DMA0_INT_LEVEL[CNT]. */
uint64_t dma1fi : 1; /**< DMA0 set Forced Interrupt. */
uint64_t dma0fi : 1; /**< DMA0 set Forced Interrupt. */
uint64_t reserved_8_8 : 1;
uint64_t dma3dbo : 1; /**< DMA3 doorbell count overflow.
Bit[32] of the doorbell count was set. */
uint64_t dma2dbo : 1; /**< DMA2 doorbell count overflow.
Bit[32] of the doorbell count was set. */
uint64_t dma1dbo : 1; /**< DMA1 doorbell count overflow.
Bit[32] of the doorbell count was set. */
uint64_t dma0dbo : 1; /**< DMA0 doorbell count overflow.
Bit[32] of the doorbell count was set. */
uint64_t iob2big : 1; /**< A requested IOBDMA is to large. */
uint64_t bar0_to : 1; /**< BAR0 R/W to a NCB device did not receive
read-data/commit in 0xffff core clocks. */
uint64_t rml_wto : 1; /**< RML write did not get commit in 0xffff core clocks. */
uint64_t rml_rto : 1; /**< RML read did not return data in 0xffff core clocks. */
#else
uint64_t rml_rto : 1;
uint64_t rml_wto : 1;
uint64_t bar0_to : 1;
uint64_t iob2big : 1;
uint64_t dma0dbo : 1;
uint64_t dma1dbo : 1;
uint64_t dma2dbo : 1;
uint64_t dma3dbo : 1;
uint64_t reserved_8_8 : 1;
uint64_t dma0fi : 1;
uint64_t dma1fi : 1;
uint64_t dcnt0 : 1;
uint64_t dcnt1 : 1;
uint64_t dtime0 : 1;
uint64_t dtime1 : 1;
uint64_t reserved_15_18 : 4;
uint64_t c0_aeri : 1;
uint64_t crs0_er : 1;
uint64_t c0_se : 1;
uint64_t crs0_dr : 1;
uint64_t c0_wake : 1;
uint64_t c0_pmei : 1;
uint64_t c0_hpint : 1;
uint64_t c1_aeri : 1;
uint64_t crs1_er : 1;
uint64_t c1_se : 1;
uint64_t crs1_dr : 1;
uint64_t c1_wake : 1;
uint64_t c1_pmei : 1;
uint64_t c1_hpint : 1;
uint64_t c0_up_b0 : 1;
uint64_t c0_up_b1 : 1;
uint64_t c0_up_b2 : 1;
uint64_t c0_up_wi : 1;
uint64_t c0_up_bx : 1;
uint64_t c0_un_b0 : 1;
uint64_t c0_un_b1 : 1;
uint64_t c0_un_b2 : 1;
uint64_t c0_un_wi : 1;
uint64_t c0_un_bx : 1;
uint64_t c1_up_b0 : 1;
uint64_t c1_up_b1 : 1;
uint64_t c1_up_b2 : 1;
uint64_t c1_up_wi : 1;
uint64_t c1_up_bx : 1;
uint64_t c1_un_b0 : 1;
uint64_t c1_un_b1 : 1;
uint64_t c1_un_b2 : 1;
uint64_t c1_un_wi : 1;
uint64_t c1_un_bx : 1;
uint64_t c0_un_wf : 1;
uint64_t c1_un_wf : 1;
uint64_t c0_up_wf : 1;
uint64_t c1_up_wf : 1;
uint64_t c0_exc : 1;
uint64_t c1_exc : 1;
uint64_t c0_ldwn : 1;
uint64_t c1_ldwn : 1;
uint64_t int_a : 1;
uint64_t reserved_62_62 : 1;
uint64_t mio_inta : 1;
#endif
} cn52xxp1;
struct cvmx_npei_int_sum_s cn56xx;
struct cvmx_npei_int_sum_cn56xxp1
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t mio_inta : 1; /**< Interrupt from MIO. */
uint64_t reserved_61_62 : 2;
uint64_t c1_ldwn : 1; /**< Reset request due to link1 down status. */
uint64_t c0_ldwn : 1; /**< Reset request due to link0 down status. */
uint64_t c1_exc : 1; /**< Set when the PESC1_DBG_INFO register has a bit
set and its cooresponding PESC1_DBG_INFO_EN bit
is set. */
uint64_t c0_exc : 1; /**< Set when the PESC0_DBG_INFO register has a bit
set and its cooresponding PESC0_DBG_INFO_EN bit
is set. */
uint64_t c1_up_wf : 1; /**< Received Unsupported P-TLP for filtered window
register. Core1. */
uint64_t c0_up_wf : 1; /**< Received Unsupported P-TLP for filtered window
register. Core0. */
uint64_t c1_un_wf : 1; /**< Received Unsupported N-TLP for filtered window
register. Core1. */
uint64_t c0_un_wf : 1; /**< Received Unsupported N-TLP for filtered window
register. Core0. */
uint64_t c1_un_bx : 1; /**< Received Unsupported N-TLP for unknown Bar.
Core 1. */
uint64_t c1_un_wi : 1; /**< Received Unsupported N-TLP for Window Register.
Core 1. */
uint64_t c1_un_b2 : 1; /**< Received Unsupported N-TLP for Bar2.
Core 1. */
uint64_t c1_un_b1 : 1; /**< Received Unsupported N-TLP for Bar1.
Core 1. */
uint64_t c1_un_b0 : 1; /**< Received Unsupported N-TLP for Bar0.
Core 1. */
uint64_t c1_up_bx : 1; /**< Received Unsupported P-TLP for unknown Bar.
Core 1. */
uint64_t c1_up_wi : 1; /**< Received Unsupported P-TLP for Window Register.
Core 1. */
uint64_t c1_up_b2 : 1; /**< Received Unsupported P-TLP for Bar2.
Core 1. */
uint64_t c1_up_b1 : 1; /**< Received Unsupported P-TLP for Bar1.
Core 1. */
uint64_t c1_up_b0 : 1; /**< Received Unsupported P-TLP for Bar0.
Core 1. */
uint64_t c0_un_bx : 1; /**< Received Unsupported N-TLP for unknown Bar.
Core 0. */
uint64_t c0_un_wi : 1; /**< Received Unsupported N-TLP for Window Register.
Core 0. */
uint64_t c0_un_b2 : 1; /**< Received Unsupported N-TLP for Bar2.
Core 0. */
uint64_t c0_un_b1 : 1; /**< Received Unsupported N-TLP for Bar1.
Core 0. */
uint64_t c0_un_b0 : 1; /**< Received Unsupported N-TLP for Bar0.
Core 0. */
uint64_t c0_up_bx : 1; /**< Received Unsupported P-TLP for unknown Bar.
Core 0. */
uint64_t c0_up_wi : 1; /**< Received Unsupported P-TLP for Window Register.
Core 0. */
uint64_t c0_up_b2 : 1; /**< Received Unsupported P-TLP for Bar2.
Core 0. */
uint64_t c0_up_b1 : 1; /**< Received Unsupported P-TLP for Bar1.
Core 0. */
uint64_t c0_up_b0 : 1; /**< Received Unsupported P-TLP for Bar0.
Core 0. */
uint64_t c1_hpint : 1; /**< Hot-Plug Interrupt.
Pcie Core 1 (hp_int).
This interrupt will only be generated when
PCIERC1_CFG034[DLLS_C] is generated. Hot plug is
not supported. */
uint64_t c1_pmei : 1; /**< PME Interrupt.
Pcie Core 1. (cfg_pme_int) */
uint64_t c1_wake : 1; /**< Wake up from Power Management Unit.
Pcie Core 1. (wake_n)
Octeon will never generate this interrupt. */
uint64_t reserved_29_29 : 1;
uint64_t c1_se : 1; /**< System Error, RC Mode Only.
Pcie Core 1. (cfg_sys_err_rc) */
uint64_t reserved_27_27 : 1;
uint64_t c1_aeri : 1; /**< Advanced Error Reporting Interrupt, RC Mode Only.
Pcie Core 1. */
uint64_t c0_hpint : 1; /**< Hot-Plug Interrupt.
Pcie Core 0 (hp_int).
This interrupt will only be generated when
PCIERC0_CFG034[DLLS_C] is generated. Hot plug is
not supported. */
uint64_t c0_pmei : 1; /**< PME Interrupt.
Pcie Core 0. (cfg_pme_int) */
uint64_t c0_wake : 1; /**< Wake up from Power Management Unit.
Pcie Core 0. (wake_n)
Octeon will never generate this interrupt. */
uint64_t reserved_22_22 : 1;
uint64_t c0_se : 1; /**< System Error, RC Mode Only.
Pcie Core 0. (cfg_sys_err_rc) */
uint64_t reserved_20_20 : 1;
uint64_t c0_aeri : 1; /**< Advanced Error Reporting Interrupt, RC Mode Only.
Pcie Core 0 (cfg_aer_rc_err_int). */
uint64_t reserved_15_18 : 4;
uint64_t dtime1 : 1; /**< Whenever NPEI_DMA_CNTS[DMA1] is not 0, the
DMA_CNT1 timer increments every core clock. When
DMA_CNT1 timer exceeds NPEI_DMA1_INT_LEVEL[TIME],
this bit is set. Writing a '1' to this bit also
clears the DMA_CNT1 timer. */
uint64_t dtime0 : 1; /**< Whenever NPEI_DMA_CNTS[DMA0] is not 0, the
DMA_CNT0 timer increments every core clock. When
DMA_CNT0 timer exceeds NPEI_DMA0_INT_LEVEL[TIME],
this bit is set. Writing a '1' to this bit also
clears the DMA_CNT0 timer. */
uint64_t dcnt1 : 1; /**< This bit indicates that NPEI_DMA_CNTS[DMA1] was/is
greater than NPEI_DMA1_INT_LEVEL[CNT]. */
uint64_t dcnt0 : 1; /**< This bit indicates that NPEI_DMA_CNTS[DMA0] was/is
greater than NPEI_DMA0_INT_LEVEL[CNT]. */
uint64_t dma1fi : 1; /**< DMA0 set Forced Interrupt. */
uint64_t dma0fi : 1; /**< DMA0 set Forced Interrupt. */
uint64_t dma4dbo : 1; /**< DMA4 doorbell overflow.
Bit[32] of the doorbell count was set. */
uint64_t dma3dbo : 1; /**< DMA3 doorbell overflow.
Bit[32] of the doorbell count was set. */
uint64_t dma2dbo : 1; /**< DMA2 doorbell overflow.
Bit[32] of the doorbell count was set. */
uint64_t dma1dbo : 1; /**< DMA1 doorbell overflow.
Bit[32] of the doorbell count was set. */
uint64_t dma0dbo : 1; /**< DMA0 doorbell overflow.
Bit[32] of the doorbell count was set. */
uint64_t iob2big : 1; /**< A requested IOBDMA is to large. */
uint64_t bar0_to : 1; /**< BAR0 R/W to a NCB device did not receive
read-data/commit in 0xffff core clocks. */
uint64_t rml_wto : 1; /**< RML write did not get commit in 0xffff core clocks. */
uint64_t rml_rto : 1; /**< RML read did not return data in 0xffff core clocks. */
#else
uint64_t rml_rto : 1;
uint64_t rml_wto : 1;
uint64_t bar0_to : 1;
uint64_t iob2big : 1;
uint64_t dma0dbo : 1;
uint64_t dma1dbo : 1;
uint64_t dma2dbo : 1;
uint64_t dma3dbo : 1;
uint64_t dma4dbo : 1;
uint64_t dma0fi : 1;
uint64_t dma1fi : 1;
uint64_t dcnt0 : 1;
uint64_t dcnt1 : 1;
uint64_t dtime0 : 1;
uint64_t dtime1 : 1;
uint64_t reserved_15_18 : 4;
uint64_t c0_aeri : 1;
uint64_t reserved_20_20 : 1;
uint64_t c0_se : 1;
uint64_t reserved_22_22 : 1;
uint64_t c0_wake : 1;
uint64_t c0_pmei : 1;
uint64_t c0_hpint : 1;
uint64_t c1_aeri : 1;
uint64_t reserved_27_27 : 1;
uint64_t c1_se : 1;
uint64_t reserved_29_29 : 1;
uint64_t c1_wake : 1;
uint64_t c1_pmei : 1;
uint64_t c1_hpint : 1;
uint64_t c0_up_b0 : 1;
uint64_t c0_up_b1 : 1;
uint64_t c0_up_b2 : 1;
uint64_t c0_up_wi : 1;
uint64_t c0_up_bx : 1;
uint64_t c0_un_b0 : 1;
uint64_t c0_un_b1 : 1;
uint64_t c0_un_b2 : 1;
uint64_t c0_un_wi : 1;
uint64_t c0_un_bx : 1;
uint64_t c1_up_b0 : 1;
uint64_t c1_up_b1 : 1;
uint64_t c1_up_b2 : 1;
uint64_t c1_up_wi : 1;
uint64_t c1_up_bx : 1;
uint64_t c1_un_b0 : 1;
uint64_t c1_un_b1 : 1;
uint64_t c1_un_b2 : 1;
uint64_t c1_un_wi : 1;
uint64_t c1_un_bx : 1;
uint64_t c0_un_wf : 1;
uint64_t c1_un_wf : 1;
uint64_t c0_up_wf : 1;
uint64_t c1_up_wf : 1;
uint64_t c0_exc : 1;
uint64_t c1_exc : 1;
uint64_t c0_ldwn : 1;
uint64_t c1_ldwn : 1;
uint64_t reserved_61_62 : 2;
uint64_t mio_inta : 1;
#endif
} cn56xxp1;
};
typedef union cvmx_npei_int_sum cvmx_npei_int_sum_t;
/**
* cvmx_npei_int_sum2
*
* NPEI_INTERRUPT_SUM2 = NPI Interrupt Summary2 Register
*
* This is a read only copy of the NPEI_INTERRUPT_SUM register with bit variances.
*/
union cvmx_npei_int_sum2
{
uint64_t u64;
struct cvmx_npei_int_sum2_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t mio_inta : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t reserved_62_62 : 1;
uint64_t int_a : 1; /**< Set when a bit in the NPEI_INT_A_SUM register and
the cooresponding bit in the NPEI_INT_A_ENB2
register is set. */
uint64_t c1_ldwn : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c0_ldwn : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c1_exc : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c0_exc : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c1_up_wf : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c0_up_wf : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c1_un_wf : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c0_un_wf : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c1_un_bx : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c1_un_wi : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c1_un_b2 : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c1_un_b1 : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c1_un_b0 : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c1_up_bx : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c1_up_wi : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c1_up_b2 : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c1_up_b1 : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c1_up_b0 : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c0_un_bx : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c0_un_wi : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c0_un_b2 : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c0_un_b1 : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c0_un_b0 : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c0_up_bx : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c0_up_wi : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c0_up_b2 : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c0_up_b1 : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c0_up_b0 : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c1_hpint : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c1_pmei : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c1_wake : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t crs1_dr : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c1_se : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t crs1_er : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c1_aeri : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c0_hpint : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c0_pmei : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c0_wake : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t crs0_dr : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c0_se : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t crs0_er : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t c0_aeri : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t reserved_15_18 : 4;
uint64_t dtime1 : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t dtime0 : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t dcnt1 : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t dcnt0 : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t dma1fi : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t dma0fi : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t reserved_8_8 : 1;
uint64_t dma3dbo : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t dma2dbo : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t dma1dbo : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t dma0dbo : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t iob2big : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t bar0_to : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t rml_wto : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
uint64_t rml_rto : 1; /**< Equal to the cooresponding bit if the
NPEI_INT_SUM register. */
#else
uint64_t rml_rto : 1;
uint64_t rml_wto : 1;
uint64_t bar0_to : 1;
uint64_t iob2big : 1;
uint64_t dma0dbo : 1;
uint64_t dma1dbo : 1;
uint64_t dma2dbo : 1;
uint64_t dma3dbo : 1;
uint64_t reserved_8_8 : 1;
uint64_t dma0fi : 1;
uint64_t dma1fi : 1;
uint64_t dcnt0 : 1;
uint64_t dcnt1 : 1;
uint64_t dtime0 : 1;
uint64_t dtime1 : 1;
uint64_t reserved_15_18 : 4;
uint64_t c0_aeri : 1;
uint64_t crs0_er : 1;
uint64_t c0_se : 1;
uint64_t crs0_dr : 1;
uint64_t c0_wake : 1;
uint64_t c0_pmei : 1;
uint64_t c0_hpint : 1;
uint64_t c1_aeri : 1;
uint64_t crs1_er : 1;
uint64_t c1_se : 1;
uint64_t crs1_dr : 1;
uint64_t c1_wake : 1;
uint64_t c1_pmei : 1;
uint64_t c1_hpint : 1;
uint64_t c0_up_b0 : 1;
uint64_t c0_up_b1 : 1;
uint64_t c0_up_b2 : 1;
uint64_t c0_up_wi : 1;
uint64_t c0_up_bx : 1;
uint64_t c0_un_b0 : 1;
uint64_t c0_un_b1 : 1;
uint64_t c0_un_b2 : 1;
uint64_t c0_un_wi : 1;
uint64_t c0_un_bx : 1;
uint64_t c1_up_b0 : 1;
uint64_t c1_up_b1 : 1;
uint64_t c1_up_b2 : 1;
uint64_t c1_up_wi : 1;
uint64_t c1_up_bx : 1;
uint64_t c1_un_b0 : 1;
uint64_t c1_un_b1 : 1;
uint64_t c1_un_b2 : 1;
uint64_t c1_un_wi : 1;
uint64_t c1_un_bx : 1;
uint64_t c0_un_wf : 1;
uint64_t c1_un_wf : 1;
uint64_t c0_up_wf : 1;
uint64_t c1_up_wf : 1;
uint64_t c0_exc : 1;
uint64_t c1_exc : 1;
uint64_t c0_ldwn : 1;
uint64_t c1_ldwn : 1;
uint64_t int_a : 1;
uint64_t reserved_62_62 : 1;
uint64_t mio_inta : 1;
#endif
} s;
struct cvmx_npei_int_sum2_s cn52xx;
struct cvmx_npei_int_sum2_s cn52xxp1;
struct cvmx_npei_int_sum2_s cn56xx;
};
typedef union cvmx_npei_int_sum2 cvmx_npei_int_sum2_t;
/**
* cvmx_npei_last_win_rdata0
*
* NPEI_LAST_WIN_RDATA0 = NPEI Last Window Read Data Port0
*
* The data from the last initiated window read.
*/
union cvmx_npei_last_win_rdata0
{
uint64_t u64;
struct cvmx_npei_last_win_rdata0_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t data : 64; /**< Last window read data. */
#else
uint64_t data : 64;
#endif
} s;
struct cvmx_npei_last_win_rdata0_s cn52xx;
struct cvmx_npei_last_win_rdata0_s cn52xxp1;
struct cvmx_npei_last_win_rdata0_s cn56xx;
struct cvmx_npei_last_win_rdata0_s cn56xxp1;
};
typedef union cvmx_npei_last_win_rdata0 cvmx_npei_last_win_rdata0_t;
/**
* cvmx_npei_last_win_rdata1
*
* NPEI_LAST_WIN_RDATA1 = NPEI Last Window Read Data Port1
*
* The data from the last initiated window read.
*/
union cvmx_npei_last_win_rdata1
{
uint64_t u64;
struct cvmx_npei_last_win_rdata1_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t data : 64; /**< Last window read data. */
#else
uint64_t data : 64;
#endif
} s;
struct cvmx_npei_last_win_rdata1_s cn52xx;
struct cvmx_npei_last_win_rdata1_s cn52xxp1;
struct cvmx_npei_last_win_rdata1_s cn56xx;
struct cvmx_npei_last_win_rdata1_s cn56xxp1;
};
typedef union cvmx_npei_last_win_rdata1 cvmx_npei_last_win_rdata1_t;
/**
* cvmx_npei_mem_access_ctl
*
* NPEI_MEM_ACCESS_CTL = NPEI's Memory Access Control
*
* Contains control for access to the PCIe address space.
*/
union cvmx_npei_mem_access_ctl
{
uint64_t u64;
struct cvmx_npei_mem_access_ctl_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_14_63 : 50;
uint64_t max_word : 4; /**< The maximum number of words to merge into a single
write operation from the PPs to the PCIe. Legal
values are 1 to 16, where a '0' is treated as 16. */
uint64_t timer : 10; /**< When the NPEI starts a PP to PCIe write it waits
no longer than the value of TIMER in eclks to
merge additional writes from the PPs into 1
large write. The values for this field is 1 to
1024 where a value of '0' is treated as 1024. */
#else
uint64_t timer : 10;
uint64_t max_word : 4;
uint64_t reserved_14_63 : 50;
#endif
} s;
struct cvmx_npei_mem_access_ctl_s cn52xx;
struct cvmx_npei_mem_access_ctl_s cn52xxp1;
struct cvmx_npei_mem_access_ctl_s cn56xx;
struct cvmx_npei_mem_access_ctl_s cn56xxp1;
};
typedef union cvmx_npei_mem_access_ctl cvmx_npei_mem_access_ctl_t;
/**
* cvmx_npei_mem_access_subid#
*
* NPEI_MEM_ACCESS_SUBIDX = NPEI Memory Access SubidX Register
*
* Contains address index and control bits for access to memory from Core PPs.
*/
union cvmx_npei_mem_access_subidx
{
uint64_t u64;
struct cvmx_npei_mem_access_subidx_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_42_63 : 22;
uint64_t zero : 1; /**< Causes all byte reads to be zero length reads.
Returns to the EXEC a zero for all read data. */
uint64_t port : 2; /**< Port the request is sent to. */
uint64_t nmerge : 1; /**< No merging is allowed in this window. */
uint64_t esr : 2; /**< Endian-swap for Reads. */
uint64_t esw : 2; /**< Endian-swap for Writes. */
uint64_t nsr : 1; /**< No Snoop for Reads. */
uint64_t nsw : 1; /**< No Snoop for Writes. */
uint64_t ror : 1; /**< Relaxed Ordering for Reads. */
uint64_t row : 1; /**< Relaxed Ordering for Writes. */
uint64_t ba : 30; /**< PCIe Adddress Bits <63:34>. */
#else
uint64_t ba : 30;
uint64_t row : 1;
uint64_t ror : 1;
uint64_t nsw : 1;
uint64_t nsr : 1;
uint64_t esw : 2;
uint64_t esr : 2;
uint64_t nmerge : 1;
uint64_t port : 2;
uint64_t zero : 1;
uint64_t reserved_42_63 : 22;
#endif
} s;
struct cvmx_npei_mem_access_subidx_s cn52xx;
struct cvmx_npei_mem_access_subidx_s cn52xxp1;
struct cvmx_npei_mem_access_subidx_s cn56xx;
struct cvmx_npei_mem_access_subidx_s cn56xxp1;
};
typedef union cvmx_npei_mem_access_subidx cvmx_npei_mem_access_subidx_t;
/**
* cvmx_npei_msi_enb0
*
* NPEI_MSI_ENB0 = NPEI MSI Enable0
*
* Used to enable the interrupt generation for the bits in the NPEI_MSI_RCV0.
*/
union cvmx_npei_msi_enb0
{
uint64_t u64;
struct cvmx_npei_msi_enb0_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t enb : 64; /**< Enables bit [63:0] of NPEI_MSI_RCV0. */
#else
uint64_t enb : 64;
#endif
} s;
struct cvmx_npei_msi_enb0_s cn52xx;
struct cvmx_npei_msi_enb0_s cn52xxp1;
struct cvmx_npei_msi_enb0_s cn56xx;
struct cvmx_npei_msi_enb0_s cn56xxp1;
};
typedef union cvmx_npei_msi_enb0 cvmx_npei_msi_enb0_t;
/**
* cvmx_npei_msi_enb1
*
* NPEI_MSI_ENB1 = NPEI MSI Enable1
*
* Used to enable the interrupt generation for the bits in the NPEI_MSI_RCV1.
*/
union cvmx_npei_msi_enb1
{
uint64_t u64;
struct cvmx_npei_msi_enb1_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t enb : 64; /**< Enables bit [63:0] of NPEI_MSI_RCV1. */
#else
uint64_t enb : 64;
#endif
} s;
struct cvmx_npei_msi_enb1_s cn52xx;
struct cvmx_npei_msi_enb1_s cn52xxp1;
struct cvmx_npei_msi_enb1_s cn56xx;
struct cvmx_npei_msi_enb1_s cn56xxp1;
};
typedef union cvmx_npei_msi_enb1 cvmx_npei_msi_enb1_t;
/**
* cvmx_npei_msi_enb2
*
* NPEI_MSI_ENB2 = NPEI MSI Enable2
*
* Used to enable the interrupt generation for the bits in the NPEI_MSI_RCV2.
*/
union cvmx_npei_msi_enb2
{
uint64_t u64;
struct cvmx_npei_msi_enb2_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t enb : 64; /**< Enables bit [63:0] of NPEI_MSI_RCV2. */
#else
uint64_t enb : 64;
#endif
} s;
struct cvmx_npei_msi_enb2_s cn52xx;
struct cvmx_npei_msi_enb2_s cn52xxp1;
struct cvmx_npei_msi_enb2_s cn56xx;
struct cvmx_npei_msi_enb2_s cn56xxp1;
};
typedef union cvmx_npei_msi_enb2 cvmx_npei_msi_enb2_t;
/**
* cvmx_npei_msi_enb3
*
* NPEI_MSI_ENB3 = NPEI MSI Enable3
*
* Used to enable the interrupt generation for the bits in the NPEI_MSI_RCV3.
*/
union cvmx_npei_msi_enb3
{
uint64_t u64;
struct cvmx_npei_msi_enb3_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t enb : 64; /**< Enables bit [63:0] of NPEI_MSI_RCV3. */
#else
uint64_t enb : 64;
#endif
} s;
struct cvmx_npei_msi_enb3_s cn52xx;
struct cvmx_npei_msi_enb3_s cn52xxp1;
struct cvmx_npei_msi_enb3_s cn56xx;
struct cvmx_npei_msi_enb3_s cn56xxp1;
};
typedef union cvmx_npei_msi_enb3 cvmx_npei_msi_enb3_t;
/**
* cvmx_npei_msi_rcv0
*
* NPEI_MSI_RCV0 = NPEI MSI Receive0
*
* Contains bits [63:0] of the 256 bits oof MSI interrupts.
*/
union cvmx_npei_msi_rcv0
{
uint64_t u64;
struct cvmx_npei_msi_rcv0_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t intr : 64; /**< Bits 63-0 of the 256 bits of MSI interrupt. */
#else
uint64_t intr : 64;
#endif
} s;
struct cvmx_npei_msi_rcv0_s cn52xx;
struct cvmx_npei_msi_rcv0_s cn52xxp1;
struct cvmx_npei_msi_rcv0_s cn56xx;
struct cvmx_npei_msi_rcv0_s cn56xxp1;
};
typedef union cvmx_npei_msi_rcv0 cvmx_npei_msi_rcv0_t;
/**
* cvmx_npei_msi_rcv1
*
* NPEI_MSI_RCV1 = NPEI MSI Receive1
*
* Contains bits [127:64] of the 256 bits oof MSI interrupts.
*/
union cvmx_npei_msi_rcv1
{
uint64_t u64;
struct cvmx_npei_msi_rcv1_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t intr : 64; /**< Bits 127-64 of the 256 bits of MSI interrupt. */
#else
uint64_t intr : 64;
#endif
} s;
struct cvmx_npei_msi_rcv1_s cn52xx;
struct cvmx_npei_msi_rcv1_s cn52xxp1;
struct cvmx_npei_msi_rcv1_s cn56xx;
struct cvmx_npei_msi_rcv1_s cn56xxp1;
};
typedef union cvmx_npei_msi_rcv1 cvmx_npei_msi_rcv1_t;
/**
* cvmx_npei_msi_rcv2
*
* NPEI_MSI_RCV2 = NPEI MSI Receive2
*
* Contains bits [191:128] of the 256 bits oof MSI interrupts.
*/
union cvmx_npei_msi_rcv2
{
uint64_t u64;
struct cvmx_npei_msi_rcv2_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t intr : 64; /**< Bits 191-128 of the 256 bits of MSI interrupt. */
#else
uint64_t intr : 64;
#endif
} s;
struct cvmx_npei_msi_rcv2_s cn52xx;
struct cvmx_npei_msi_rcv2_s cn52xxp1;
struct cvmx_npei_msi_rcv2_s cn56xx;
struct cvmx_npei_msi_rcv2_s cn56xxp1;
};
typedef union cvmx_npei_msi_rcv2 cvmx_npei_msi_rcv2_t;
/**
* cvmx_npei_msi_rcv3
*
* NPEI_MSI_RCV3 = NPEI MSI Receive3
*
* Contains bits [255:192] of the 256 bits oof MSI interrupts.
*/
union cvmx_npei_msi_rcv3
{
uint64_t u64;
struct cvmx_npei_msi_rcv3_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t intr : 64; /**< Bits 255-192 of the 256 bits of MSI interrupt. */
#else
uint64_t intr : 64;
#endif
} s;
struct cvmx_npei_msi_rcv3_s cn52xx;
struct cvmx_npei_msi_rcv3_s cn52xxp1;
struct cvmx_npei_msi_rcv3_s cn56xx;
struct cvmx_npei_msi_rcv3_s cn56xxp1;
};
typedef union cvmx_npei_msi_rcv3 cvmx_npei_msi_rcv3_t;
/**
* cvmx_npei_msi_rd_map
*
* NPEI_MSI_RD_MAP = NPEI MSI Read MAP
*
* Used to read the mapping function of the NPEI_PCIE_MSI_RCV to NPEI_MSI_RCV registers.
*/
union cvmx_npei_msi_rd_map
{
uint64_t u64;
struct cvmx_npei_msi_rd_map_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_16_63 : 48;
uint64_t rd_int : 8; /**< The value of the map at the location PREVIOUSLY
written to the MSI_INT field of this register. */
uint64_t msi_int : 8; /**< Selects the value that would be received when the
NPEI_PCIE_MSI_RCV register is written. */
#else
uint64_t msi_int : 8;
uint64_t rd_int : 8;
uint64_t reserved_16_63 : 48;
#endif
} s;
struct cvmx_npei_msi_rd_map_s cn52xx;
struct cvmx_npei_msi_rd_map_s cn52xxp1;
struct cvmx_npei_msi_rd_map_s cn56xx;
struct cvmx_npei_msi_rd_map_s cn56xxp1;
};
typedef union cvmx_npei_msi_rd_map cvmx_npei_msi_rd_map_t;
/**
* cvmx_npei_msi_w1c_enb0
*
* NPEI_MSI_W1C_ENB0 = NPEI MSI Write 1 To Clear Enable0
*
* Used to clear bits in NPEI_MSI_ENB0. This is a PASS2 register.
*/
union cvmx_npei_msi_w1c_enb0
{
uint64_t u64;
struct cvmx_npei_msi_w1c_enb0_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t clr : 64; /**< A write of '1' to a vector will clear the
cooresponding bit in NPEI_MSI_ENB0.
A read to this address will return 0. */
#else
uint64_t clr : 64;
#endif
} s;
struct cvmx_npei_msi_w1c_enb0_s cn52xx;
struct cvmx_npei_msi_w1c_enb0_s cn56xx;
};
typedef union cvmx_npei_msi_w1c_enb0 cvmx_npei_msi_w1c_enb0_t;
/**
* cvmx_npei_msi_w1c_enb1
*
* NPEI_MSI_W1C_ENB1 = NPEI MSI Write 1 To Clear Enable1
*
* Used to clear bits in NPEI_MSI_ENB1. This is a PASS2 register.
*/
union cvmx_npei_msi_w1c_enb1
{
uint64_t u64;
struct cvmx_npei_msi_w1c_enb1_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t clr : 64; /**< A write of '1' to a vector will clear the
cooresponding bit in NPEI_MSI_ENB1.
A read to this address will return 0. */
#else
uint64_t clr : 64;
#endif
} s;
struct cvmx_npei_msi_w1c_enb1_s cn52xx;
struct cvmx_npei_msi_w1c_enb1_s cn56xx;
};
typedef union cvmx_npei_msi_w1c_enb1 cvmx_npei_msi_w1c_enb1_t;
/**
* cvmx_npei_msi_w1c_enb2
*
* NPEI_MSI_W1C_ENB2 = NPEI MSI Write 1 To Clear Enable2
*
* Used to clear bits in NPEI_MSI_ENB2. This is a PASS2 register.
*/
union cvmx_npei_msi_w1c_enb2
{
uint64_t u64;
struct cvmx_npei_msi_w1c_enb2_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t clr : 64; /**< A write of '1' to a vector will clear the
cooresponding bit in NPEI_MSI_ENB2.
A read to this address will return 0. */
#else
uint64_t clr : 64;
#endif
} s;
struct cvmx_npei_msi_w1c_enb2_s cn52xx;
struct cvmx_npei_msi_w1c_enb2_s cn56xx;
};
typedef union cvmx_npei_msi_w1c_enb2 cvmx_npei_msi_w1c_enb2_t;
/**
* cvmx_npei_msi_w1c_enb3
*
* NPEI_MSI_W1C_ENB3 = NPEI MSI Write 1 To Clear Enable3
*
* Used to clear bits in NPEI_MSI_ENB3. This is a PASS2 register.
*/
union cvmx_npei_msi_w1c_enb3
{
uint64_t u64;
struct cvmx_npei_msi_w1c_enb3_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t clr : 64; /**< A write of '1' to a vector will clear the
cooresponding bit in NPEI_MSI_ENB3.
A read to this address will return 0. */
#else
uint64_t clr : 64;
#endif
} s;
struct cvmx_npei_msi_w1c_enb3_s cn52xx;
struct cvmx_npei_msi_w1c_enb3_s cn56xx;
};
typedef union cvmx_npei_msi_w1c_enb3 cvmx_npei_msi_w1c_enb3_t;
/**
* cvmx_npei_msi_w1s_enb0
*
* NPEI_MSI_W1S_ENB0 = NPEI MSI Write 1 To Set Enable0
*
* Used to set bits in NPEI_MSI_ENB0. This is a PASS2 register.
*/
union cvmx_npei_msi_w1s_enb0
{
uint64_t u64;
struct cvmx_npei_msi_w1s_enb0_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t set : 64; /**< A write of '1' to a vector will set the
cooresponding bit in NPEI_MSI_ENB0.
A read to this address will return 0. */
#else
uint64_t set : 64;
#endif
} s;
struct cvmx_npei_msi_w1s_enb0_s cn52xx;
struct cvmx_npei_msi_w1s_enb0_s cn56xx;
};
typedef union cvmx_npei_msi_w1s_enb0 cvmx_npei_msi_w1s_enb0_t;
/**
* cvmx_npei_msi_w1s_enb1
*
* NPEI_MSI_W1S_ENB0 = NPEI MSI Write 1 To Set Enable1
*
* Used to set bits in NPEI_MSI_ENB1. This is a PASS2 register.
*/
union cvmx_npei_msi_w1s_enb1
{
uint64_t u64;
struct cvmx_npei_msi_w1s_enb1_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t set : 64; /**< A write of '1' to a vector will set the
cooresponding bit in NPEI_MSI_ENB1.
A read to this address will return 0. */
#else
uint64_t set : 64;
#endif
} s;
struct cvmx_npei_msi_w1s_enb1_s cn52xx;
struct cvmx_npei_msi_w1s_enb1_s cn56xx;
};
typedef union cvmx_npei_msi_w1s_enb1 cvmx_npei_msi_w1s_enb1_t;
/**
* cvmx_npei_msi_w1s_enb2
*
* NPEI_MSI_W1S_ENB2 = NPEI MSI Write 1 To Set Enable2
*
* Used to set bits in NPEI_MSI_ENB2. This is a PASS2 register.
*/
union cvmx_npei_msi_w1s_enb2
{
uint64_t u64;
struct cvmx_npei_msi_w1s_enb2_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t set : 64; /**< A write of '1' to a vector will set the
cooresponding bit in NPEI_MSI_ENB2.
A read to this address will return 0. */
#else
uint64_t set : 64;
#endif
} s;
struct cvmx_npei_msi_w1s_enb2_s cn52xx;
struct cvmx_npei_msi_w1s_enb2_s cn56xx;
};
typedef union cvmx_npei_msi_w1s_enb2 cvmx_npei_msi_w1s_enb2_t;
/**
* cvmx_npei_msi_w1s_enb3
*
* NPEI_MSI_W1S_ENB3 = NPEI MSI Write 1 To Set Enable3
*
* Used to set bits in NPEI_MSI_ENB3. This is a PASS2 register.
*/
union cvmx_npei_msi_w1s_enb3
{
uint64_t u64;
struct cvmx_npei_msi_w1s_enb3_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t set : 64; /**< A write of '1' to a vector will set the
cooresponding bit in NPEI_MSI_ENB3.
A read to this address will return 0. */
#else
uint64_t set : 64;
#endif
} s;
struct cvmx_npei_msi_w1s_enb3_s cn52xx;
struct cvmx_npei_msi_w1s_enb3_s cn56xx;
};
typedef union cvmx_npei_msi_w1s_enb3 cvmx_npei_msi_w1s_enb3_t;
/**
* cvmx_npei_msi_wr_map
*
* NPEI_MSI_WR_MAP = NPEI MSI Write MAP
*
* Used to write the mapping function of the NPEI_PCIE_MSI_RCV to NPEI_MSI_RCV registers.
*/
union cvmx_npei_msi_wr_map
{
uint64_t u64;
struct cvmx_npei_msi_wr_map_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_16_63 : 48;
uint64_t ciu_int : 8; /**< Selects which bit in the NPEI_MSI_RCV# (0-255)
will be set when the value specified in the
MSI_INT of this register is recevied during a
write to the NPEI_PCIE_MSI_RCV register. */
uint64_t msi_int : 8; /**< Selects the value that would be received when the
NPEI_PCIE_MSI_RCV register is written. */
#else
uint64_t msi_int : 8;
uint64_t ciu_int : 8;
uint64_t reserved_16_63 : 48;
#endif
} s;
struct cvmx_npei_msi_wr_map_s cn52xx;
struct cvmx_npei_msi_wr_map_s cn52xxp1;
struct cvmx_npei_msi_wr_map_s cn56xx;
struct cvmx_npei_msi_wr_map_s cn56xxp1;
};
typedef union cvmx_npei_msi_wr_map cvmx_npei_msi_wr_map_t;
/**
* cvmx_npei_pcie_credit_cnt
*
* NPEI_PCIE_CREDIT_CNT = NPEI PCIE Credit Count
*
* Contains the number of credits for the pcie port FIFOs used by the NPEI. This value needs to be set BEFORE PCIe traffic
* flow from NPEI to PCIE Ports starts. A write to this register will cause the credit counts in the NPEI for the two
* PCIE ports to be reset to the value in this register.
*/
union cvmx_npei_pcie_credit_cnt
{
uint64_t u64;
struct cvmx_npei_pcie_credit_cnt_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_48_63 : 16;
uint64_t p1_ccnt : 8; /**< Port1 C-TLP FIFO Credits.
Legal values are 0x25 to 0x80. */
uint64_t p1_ncnt : 8; /**< Port1 N-TLP FIFO Credits.
Legal values are 0x5 to 0x10. */
uint64_t p1_pcnt : 8; /**< Port1 P-TLP FIFO Credits.
Legal values are 0x25 to 0x80. */
uint64_t p0_ccnt : 8; /**< Port0 C-TLP FIFO Credits.
Legal values are 0x25 to 0x80. */
uint64_t p0_ncnt : 8; /**< Port0 N-TLP FIFO Credits.
Legal values are 0x5 to 0x10. */
uint64_t p0_pcnt : 8; /**< Port0 P-TLP FIFO Credits.
Legal values are 0x25 to 0x80. */
#else
uint64_t p0_pcnt : 8;
uint64_t p0_ncnt : 8;
uint64_t p0_ccnt : 8;
uint64_t p1_pcnt : 8;
uint64_t p1_ncnt : 8;
uint64_t p1_ccnt : 8;
uint64_t reserved_48_63 : 16;
#endif
} s;
struct cvmx_npei_pcie_credit_cnt_s cn52xx;
struct cvmx_npei_pcie_credit_cnt_s cn56xx;
};
typedef union cvmx_npei_pcie_credit_cnt cvmx_npei_pcie_credit_cnt_t;
/**
* cvmx_npei_pcie_msi_rcv
*
* NPEI_PCIE_MSI_RCV = NPEI PCIe MSI Receive
*
* Register where MSI writes are directed from the PCIe.
*/
union cvmx_npei_pcie_msi_rcv
{
uint64_t u64;
struct cvmx_npei_pcie_msi_rcv_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_8_63 : 56;
uint64_t intr : 8; /**< A write to this register will result in a bit in
one of the NPEI_MSI_RCV# registers being set.
Which bit is set is dependent on the previously
written using the NPEI_MSI_WR_MAP register or if
not previously written the reset value of the MAP. */
#else
uint64_t intr : 8;
uint64_t reserved_8_63 : 56;
#endif
} s;
struct cvmx_npei_pcie_msi_rcv_s cn52xx;
struct cvmx_npei_pcie_msi_rcv_s cn52xxp1;
struct cvmx_npei_pcie_msi_rcv_s cn56xx;
struct cvmx_npei_pcie_msi_rcv_s cn56xxp1;
};
typedef union cvmx_npei_pcie_msi_rcv cvmx_npei_pcie_msi_rcv_t;
/**
* cvmx_npei_pcie_msi_rcv_b1
*
* NPEI_PCIE_MSI_RCV_B1 = NPEI PCIe MSI Receive Byte 1
*
* Register where MSI writes are directed from the PCIe.
*/
union cvmx_npei_pcie_msi_rcv_b1
{
uint64_t u64;
struct cvmx_npei_pcie_msi_rcv_b1_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_16_63 : 48;
uint64_t intr : 8; /**< A write to this register will result in a bit in
one of the NPEI_MSI_RCV# registers being set.
Which bit is set is dependent on the previously
written using the NPEI_MSI_WR_MAP register or if
not previously written the reset value of the MAP. */
uint64_t reserved_0_7 : 8;
#else
uint64_t reserved_0_7 : 8;
uint64_t intr : 8;
uint64_t reserved_16_63 : 48;
#endif
} s;
struct cvmx_npei_pcie_msi_rcv_b1_s cn52xx;
struct cvmx_npei_pcie_msi_rcv_b1_s cn52xxp1;
struct cvmx_npei_pcie_msi_rcv_b1_s cn56xx;
struct cvmx_npei_pcie_msi_rcv_b1_s cn56xxp1;
};
typedef union cvmx_npei_pcie_msi_rcv_b1 cvmx_npei_pcie_msi_rcv_b1_t;
/**
* cvmx_npei_pcie_msi_rcv_b2
*
* NPEI_PCIE_MSI_RCV_B2 = NPEI PCIe MSI Receive Byte 2
*
* Register where MSI writes are directed from the PCIe.
*/
union cvmx_npei_pcie_msi_rcv_b2
{
uint64_t u64;
struct cvmx_npei_pcie_msi_rcv_b2_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_24_63 : 40;
uint64_t intr : 8; /**< A write to this register will result in a bit in
one of the NPEI_MSI_RCV# registers being set.
Which bit is set is dependent on the previously
written using the NPEI_MSI_WR_MAP register or if
not previously written the reset value of the MAP. */
uint64_t reserved_0_15 : 16;
#else
uint64_t reserved_0_15 : 16;
uint64_t intr : 8;
uint64_t reserved_24_63 : 40;
#endif
} s;
struct cvmx_npei_pcie_msi_rcv_b2_s cn52xx;
struct cvmx_npei_pcie_msi_rcv_b2_s cn52xxp1;
struct cvmx_npei_pcie_msi_rcv_b2_s cn56xx;
struct cvmx_npei_pcie_msi_rcv_b2_s cn56xxp1;
};
typedef union cvmx_npei_pcie_msi_rcv_b2 cvmx_npei_pcie_msi_rcv_b2_t;
/**
* cvmx_npei_pcie_msi_rcv_b3
*
* NPEI_PCIE_MSI_RCV_B3 = NPEI PCIe MSI Receive Byte 3
*
* Register where MSI writes are directed from the PCIe.
*/
union cvmx_npei_pcie_msi_rcv_b3
{
uint64_t u64;
struct cvmx_npei_pcie_msi_rcv_b3_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_32_63 : 32;
uint64_t intr : 8; /**< A write to this register will result in a bit in
one of the NPEI_MSI_RCV# registers being set.
Which bit is set is dependent on the previously
written using the NPEI_MSI_WR_MAP register or if
not previously written the reset value of the MAP. */
uint64_t reserved_0_23 : 24;
#else
uint64_t reserved_0_23 : 24;
uint64_t intr : 8;
uint64_t reserved_32_63 : 32;
#endif
} s;
struct cvmx_npei_pcie_msi_rcv_b3_s cn52xx;
struct cvmx_npei_pcie_msi_rcv_b3_s cn52xxp1;
struct cvmx_npei_pcie_msi_rcv_b3_s cn56xx;
struct cvmx_npei_pcie_msi_rcv_b3_s cn56xxp1;
};
typedef union cvmx_npei_pcie_msi_rcv_b3 cvmx_npei_pcie_msi_rcv_b3_t;
/**
* cvmx_npei_pkt#_cnts
*
* NPEI_PKT[0..31]_CNTS = NPEI Packet ring# Counts
*
* The counters for output rings.
*/
union cvmx_npei_pktx_cnts
{
uint64_t u64;
struct cvmx_npei_pktx_cnts_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_54_63 : 10;
uint64_t timer : 22; /**< Timer incremented every 1024 core clocks
when NPEI_PKTS#_CNTS[CNT] is non zero. Field
cleared when NPEI_PKTS#_CNTS[CNT] goes to 0.
Field is also cleared when NPEI_PKT_TIME_INT is
cleared.
The first increment of this count can occur
between 0 to 1023 core clocks. */
uint64_t cnt : 32; /**< ring counter. This field is incremented as
packets are sent out and decremented in response to
writes to this field.
When NPEI_PKT_OUT_BMODE is '0' a value of 1 is
added to the register for each packet, when '1'
and the info-pointer is NOT used the length of the
packet plus 8 is added, when '1' and info-pointer
mode IS used the packet length is added to this
field. */
#else
uint64_t cnt : 32;
uint64_t timer : 22;
uint64_t reserved_54_63 : 10;
#endif
} s;
struct cvmx_npei_pktx_cnts_s cn52xx;
struct cvmx_npei_pktx_cnts_s cn56xx;
};
typedef union cvmx_npei_pktx_cnts cvmx_npei_pktx_cnts_t;
/**
* cvmx_npei_pkt#_in_bp
*
* NPEI_PKT[0..31]_IN_BP = NPEI Packet ring# Input Backpressure
*
* The counters and thresholds for input packets to apply backpressure to processing of the packets.
*/
union cvmx_npei_pktx_in_bp
{
uint64_t u64;
struct cvmx_npei_pktx_in_bp_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t wmark : 32; /**< When CNT is greater than this threshold no more
packets will be processed for this ring.
When writing this field of the NPEI_PKT#_IN_BP
register, use a 4-byte write so as to not write
any other field of this register. */
uint64_t cnt : 32; /**< ring counter. This field is incremented by one
whenever OCTEON receives, buffers, and creates a
work queue entry for a packet that arrives by the
cooresponding input ring. A write to this field
will be subtracted from the field value.
When writing this field of the NPEI_PKT#_IN_BP
register, use a 4-byte write so as to not write
any other field of this register. */
#else
uint64_t cnt : 32;
uint64_t wmark : 32;
#endif
} s;
struct cvmx_npei_pktx_in_bp_s cn52xx;
struct cvmx_npei_pktx_in_bp_s cn56xx;
};
typedef union cvmx_npei_pktx_in_bp cvmx_npei_pktx_in_bp_t;
/**
* cvmx_npei_pkt#_instr_baddr
*
* NPEI_PKT[0..31]_INSTR_BADDR = NPEI Packet ring# Instruction Base Address
*
* Start of Instruction for input packets.
*/
union cvmx_npei_pktx_instr_baddr
{
uint64_t u64;
struct cvmx_npei_pktx_instr_baddr_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t addr : 61; /**< Base address for Instructions. */
uint64_t reserved_0_2 : 3;
#else
uint64_t reserved_0_2 : 3;
uint64_t addr : 61;
#endif
} s;
struct cvmx_npei_pktx_instr_baddr_s cn52xx;
struct cvmx_npei_pktx_instr_baddr_s cn56xx;
};
typedef union cvmx_npei_pktx_instr_baddr cvmx_npei_pktx_instr_baddr_t;
/**
* cvmx_npei_pkt#_instr_baoff_dbell
*
* NPEI_PKT[0..31]_INSTR_BAOFF_DBELL = NPEI Packet ring# Instruction Base Address Offset and Doorbell
*
* The doorbell and base address offset for next read.
*/
union cvmx_npei_pktx_instr_baoff_dbell
{
uint64_t u64;
struct cvmx_npei_pktx_instr_baoff_dbell_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t aoff : 32; /**< The offset from the NPEI_PKT[0..31]_INSTR_BADDR
where the next instruction will be read. */
uint64_t dbell : 32; /**< Instruction doorbell count. Writes to this field
will increment the value here. Reads will return
present value. A write of 0xffffffff will set the
DBELL and AOFF fields to '0'. */
#else
uint64_t dbell : 32;
uint64_t aoff : 32;
#endif
} s;
struct cvmx_npei_pktx_instr_baoff_dbell_s cn52xx;
struct cvmx_npei_pktx_instr_baoff_dbell_s cn56xx;
};
typedef union cvmx_npei_pktx_instr_baoff_dbell cvmx_npei_pktx_instr_baoff_dbell_t;
/**
* cvmx_npei_pkt#_instr_fifo_rsize
*
* NPEI_PKT[0..31]_INSTR_FIFO_RSIZE = NPEI Packet ring# Instruction FIFO and Ring Size.
*
* Fifo field and ring size for Instructions.
*/
union cvmx_npei_pktx_instr_fifo_rsize
{
uint64_t u64;
struct cvmx_npei_pktx_instr_fifo_rsize_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t max : 9; /**< Max Fifo Size. */
uint64_t rrp : 9; /**< Fifo read pointer. */
uint64_t wrp : 9; /**< Fifo write pointer. */
uint64_t fcnt : 5; /**< Fifo count. */
uint64_t rsize : 32; /**< Instruction ring size. */
#else
uint64_t rsize : 32;
uint64_t fcnt : 5;
uint64_t wrp : 9;
uint64_t rrp : 9;
uint64_t max : 9;
#endif
} s;
struct cvmx_npei_pktx_instr_fifo_rsize_s cn52xx;
struct cvmx_npei_pktx_instr_fifo_rsize_s cn56xx;
};
typedef union cvmx_npei_pktx_instr_fifo_rsize cvmx_npei_pktx_instr_fifo_rsize_t;
/**
* cvmx_npei_pkt#_instr_header
*
* NPEI_PKT[0..31]_INSTR_HEADER = NPEI Packet ring# Instruction Header.
*
* VAlues used to build input packet header.
*/
union cvmx_npei_pktx_instr_header
{
uint64_t u64;
struct cvmx_npei_pktx_instr_header_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_44_63 : 20;
uint64_t pbp : 1; /**< Enable Packet-by-packet mode. */
uint64_t reserved_38_42 : 5;
uint64_t rparmode : 2; /**< Parse Mode. Used when packet is raw and PBP==0. */
uint64_t reserved_35_35 : 1;
uint64_t rskp_len : 7; /**< Skip Length. Used when packet is raw and PBP==0. */
uint64_t reserved_22_27 : 6;
uint64_t use_ihdr : 1; /**< When set '1' the instruction header will be sent
as part of the packet data, regardless of the
value of bit [63] of the instruction header.
USE_IHDR must be set whenever PBP is set. */
uint64_t reserved_16_20 : 5;
uint64_t par_mode : 2; /**< Parse Mode. Used when USE_IHDR is set and packet
is not raw and PBP is not set. */
uint64_t reserved_13_13 : 1;
uint64_t skp_len : 7; /**< Skip Length. Used when USE_IHDR is set and packet
is not raw and PBP is not set. */
uint64_t reserved_0_5 : 6;
#else
uint64_t reserved_0_5 : 6;
uint64_t skp_len : 7;
uint64_t reserved_13_13 : 1;
uint64_t par_mode : 2;
uint64_t reserved_16_20 : 5;
uint64_t use_ihdr : 1;
uint64_t reserved_22_27 : 6;
uint64_t rskp_len : 7;
uint64_t reserved_35_35 : 1;
uint64_t rparmode : 2;
uint64_t reserved_38_42 : 5;
uint64_t pbp : 1;
uint64_t reserved_44_63 : 20;
#endif
} s;
struct cvmx_npei_pktx_instr_header_s cn52xx;
struct cvmx_npei_pktx_instr_header_s cn56xx;
};
typedef union cvmx_npei_pktx_instr_header cvmx_npei_pktx_instr_header_t;
/**
* cvmx_npei_pkt#_slist_baddr
*
* NPEI_PKT[0..31]_SLIST_BADDR = NPEI Packet ring# Scatter List Base Address
*
* Start of Scatter List for output packet pointers - MUST be 16 byte alligned
*/
union cvmx_npei_pktx_slist_baddr
{
uint64_t u64;
struct cvmx_npei_pktx_slist_baddr_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t addr : 60; /**< Base address for scatter list pointers. */
uint64_t reserved_0_3 : 4;
#else
uint64_t reserved_0_3 : 4;
uint64_t addr : 60;
#endif
} s;
struct cvmx_npei_pktx_slist_baddr_s cn52xx;
struct cvmx_npei_pktx_slist_baddr_s cn56xx;
};
typedef union cvmx_npei_pktx_slist_baddr cvmx_npei_pktx_slist_baddr_t;
/**
* cvmx_npei_pkt#_slist_baoff_dbell
*
* NPEI_PKT[0..31]_SLIST_BAOFF_DBELL = NPEI Packet ring# Scatter List Base Address Offset and Doorbell
*
* The doorbell and base address offset for next read.
*/
union cvmx_npei_pktx_slist_baoff_dbell
{
uint64_t u64;
struct cvmx_npei_pktx_slist_baoff_dbell_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t aoff : 32; /**< The offset from the NPEI_PKT[0..31]_SLIST_BADDR
where the next SList pointer will be read.
A write of 0xFFFFFFFF to the DBELL field will
clear DBELL and AOFF */
uint64_t dbell : 32; /**< Scatter list doorbell count. Writes to this field
will increment the value here. Reads will return
present value. The value of this field is
decremented as read operations are ISSUED for
scatter pointers.
A write of 0xFFFFFFFF will clear DBELL and AOFF */
#else
uint64_t dbell : 32;
uint64_t aoff : 32;
#endif
} s;
struct cvmx_npei_pktx_slist_baoff_dbell_s cn52xx;
struct cvmx_npei_pktx_slist_baoff_dbell_s cn56xx;
};
typedef union cvmx_npei_pktx_slist_baoff_dbell cvmx_npei_pktx_slist_baoff_dbell_t;
/**
* cvmx_npei_pkt#_slist_fifo_rsize
*
* NPEI_PKT[0..31]_SLIST_FIFO_RSIZE = NPEI Packet ring# Scatter List FIFO and Ring Size.
*
* The number of scatter pointer pairs in the scatter list.
*/
union cvmx_npei_pktx_slist_fifo_rsize
{
uint64_t u64;
struct cvmx_npei_pktx_slist_fifo_rsize_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_32_63 : 32;
uint64_t rsize : 32; /**< The number of scatter pointer pairs contained in
the scatter list ring. */
#else
uint64_t rsize : 32;
uint64_t reserved_32_63 : 32;
#endif
} s;
struct cvmx_npei_pktx_slist_fifo_rsize_s cn52xx;
struct cvmx_npei_pktx_slist_fifo_rsize_s cn56xx;
};
typedef union cvmx_npei_pktx_slist_fifo_rsize cvmx_npei_pktx_slist_fifo_rsize_t;
/**
* cvmx_npei_pkt_cnt_int
*
* NPEI_PKT_CNT_INT = NPI Packet Counter Interrupt
*
* The packets rings that are interrupting because of Packet Counters.
*/
union cvmx_npei_pkt_cnt_int
{
uint64_t u64;
struct cvmx_npei_pkt_cnt_int_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_32_63 : 32;
uint64_t port : 32; /**< Bit vector cooresponding to ring number is set when
NPEI_PKT#_CNTS[CNT] is greater
than NPEI_PKT_INT_LEVELS[CNT]. */
#else
uint64_t port : 32;
uint64_t reserved_32_63 : 32;
#endif
} s;
struct cvmx_npei_pkt_cnt_int_s cn52xx;
struct cvmx_npei_pkt_cnt_int_s cn56xx;
};
typedef union cvmx_npei_pkt_cnt_int cvmx_npei_pkt_cnt_int_t;
/**
* cvmx_npei_pkt_cnt_int_enb
*
* NPEI_PKT_CNT_INT_ENB = NPI Packet Counter Interrupt Enable
*
* Enable for the packets rings that are interrupting because of Packet Counters.
*/
union cvmx_npei_pkt_cnt_int_enb
{
uint64_t u64;
struct cvmx_npei_pkt_cnt_int_enb_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_32_63 : 32;
uint64_t port : 32; /**< Bit vector cooresponding to ring number when set
allows NPEI_PKT_CNT_INT to generate an interrupt. */
#else
uint64_t port : 32;
uint64_t reserved_32_63 : 32;
#endif
} s;
struct cvmx_npei_pkt_cnt_int_enb_s cn52xx;
struct cvmx_npei_pkt_cnt_int_enb_s cn56xx;
};
typedef union cvmx_npei_pkt_cnt_int_enb cvmx_npei_pkt_cnt_int_enb_t;
/**
* cvmx_npei_pkt_data_out_es
*
* NPEI_PKT_DATA_OUT_ES = NPEI's Packet Data Out Endian Swap
*
* The Endian Swap for writing Data Out.
*/
union cvmx_npei_pkt_data_out_es
{
uint64_t u64;
struct cvmx_npei_pkt_data_out_es_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t es : 64; /**< The endian swap mode for Packet rings 0 through 31.
Two bits are used per ring (i.e. ring 0 [1:0],
ring 1 [3:2], ....). */
#else
uint64_t es : 64;
#endif
} s;
struct cvmx_npei_pkt_data_out_es_s cn52xx;
struct cvmx_npei_pkt_data_out_es_s cn56xx;
};
typedef union cvmx_npei_pkt_data_out_es cvmx_npei_pkt_data_out_es_t;
/**
* cvmx_npei_pkt_data_out_ns
*
* NPEI_PKT_DATA_OUT_NS = NPEI's Packet Data Out No Snoop
*
* The NS field for the TLP when writing packet data.
*/
union cvmx_npei_pkt_data_out_ns
{
uint64_t u64;
struct cvmx_npei_pkt_data_out_ns_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_32_63 : 32;
uint64_t nsr : 32; /**< When asserted '1' the vector bit cooresponding
to the Packet-ring will enable NS in TLP header. */
#else
uint64_t nsr : 32;
uint64_t reserved_32_63 : 32;
#endif
} s;
struct cvmx_npei_pkt_data_out_ns_s cn52xx;
struct cvmx_npei_pkt_data_out_ns_s cn56xx;
};
typedef union cvmx_npei_pkt_data_out_ns cvmx_npei_pkt_data_out_ns_t;
/**
* cvmx_npei_pkt_data_out_ror
*
* NPEI_PKT_DATA_OUT_ROR = NPEI's Packet Data Out Relaxed Ordering
*
* The ROR field for the TLP when writing Packet Data.
*/
union cvmx_npei_pkt_data_out_ror
{
uint64_t u64;
struct cvmx_npei_pkt_data_out_ror_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_32_63 : 32;
uint64_t ror : 32; /**< When asserted '1' the vector bit cooresponding
to the Packet-ring will enable ROR in TLP header. */
#else
uint64_t ror : 32;
uint64_t reserved_32_63 : 32;
#endif
} s;
struct cvmx_npei_pkt_data_out_ror_s cn52xx;
struct cvmx_npei_pkt_data_out_ror_s cn56xx;
};
typedef union cvmx_npei_pkt_data_out_ror cvmx_npei_pkt_data_out_ror_t;
/**
* cvmx_npei_pkt_dpaddr
*
* NPEI_PKT_DPADDR = NPEI's Packet Data Pointer Addr
*
* Used to detemine address and attributes for packet data writes.
*/
union cvmx_npei_pkt_dpaddr
{
uint64_t u64;
struct cvmx_npei_pkt_dpaddr_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_32_63 : 32;
uint64_t dptr : 32; /**< When asserted '1' the vector bit cooresponding
to the Packet-ring will use:
the address[63:60] to write packet data
comes from the DPTR[63:60] in the scatter-list
pair and the RO, NS, ES values come from the O0_ES,
O0_NS, O0_RO. When '0' the RO == DPTR[60],
NS == DPTR[61], ES == DPTR[63:62], the address the
packet will be written to is ADDR[63:60] ==
O0_ES[1:0], O0_NS, O0_RO. */
#else
uint64_t dptr : 32;
uint64_t reserved_32_63 : 32;
#endif
} s;
struct cvmx_npei_pkt_dpaddr_s cn52xx;
struct cvmx_npei_pkt_dpaddr_s cn56xx;
};
typedef union cvmx_npei_pkt_dpaddr cvmx_npei_pkt_dpaddr_t;
/**
* cvmx_npei_pkt_in_bp
*
* NPEI_PKT_IN_BP = NPEI Packet Input Backpressure
*
* Which input rings have backpressure applied.
*/
union cvmx_npei_pkt_in_bp
{
uint64_t u64;
struct cvmx_npei_pkt_in_bp_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_32_63 : 32;
uint64_t bp : 32; /**< A packet input ring that has its count greater
than its WMARK will have backpressure applied.
Each of the 32 bits coorespond to an input ring.
When '1' that ring has backpressure applied an
will fetch no more instructions, but will process
any previously fetched instructions. */
#else
uint64_t bp : 32;
uint64_t reserved_32_63 : 32;
#endif
} s;
struct cvmx_npei_pkt_in_bp_s cn52xx;
struct cvmx_npei_pkt_in_bp_s cn56xx;
};
typedef union cvmx_npei_pkt_in_bp cvmx_npei_pkt_in_bp_t;
/**
* cvmx_npei_pkt_in_done#_cnts
*
* NPEI_PKT_IN_DONE[0..31]_CNTS = NPEI Instruction Done ring# Counts
*
* Counters for instructions completed on Input rings.
*/
union cvmx_npei_pkt_in_donex_cnts
{
uint64_t u64;
struct cvmx_npei_pkt_in_donex_cnts_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_32_63 : 32;
uint64_t cnt : 32; /**< This field is incrmented by '1' when an instruction
is completed. This field is incremented as the
last of the data is read from the PCIe. */
#else
uint64_t cnt : 32;
uint64_t reserved_32_63 : 32;
#endif
} s;
struct cvmx_npei_pkt_in_donex_cnts_s cn52xx;
struct cvmx_npei_pkt_in_donex_cnts_s cn56xx;
};
typedef union cvmx_npei_pkt_in_donex_cnts cvmx_npei_pkt_in_donex_cnts_t;
/**
* cvmx_npei_pkt_in_instr_counts
*
* NPEI_PKT_IN_INSTR_COUNTS = NPEI Packet Input Instrutction Counts
*
* Keeps track of the number of instructions read into the FIFO and Packets sent to IPD.
*/
union cvmx_npei_pkt_in_instr_counts
{
uint64_t u64;
struct cvmx_npei_pkt_in_instr_counts_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t wr_cnt : 32; /**< Shows the number of packets sent to the IPD. */
uint64_t rd_cnt : 32; /**< Shows the value of instructions that have had reads
issued for them.
to the Packet-ring is in reset. */
#else
uint64_t rd_cnt : 32;
uint64_t wr_cnt : 32;
#endif
} s;
struct cvmx_npei_pkt_in_instr_counts_s cn52xx;
struct cvmx_npei_pkt_in_instr_counts_s cn56xx;
};
typedef union cvmx_npei_pkt_in_instr_counts cvmx_npei_pkt_in_instr_counts_t;
/**
* cvmx_npei_pkt_in_pcie_port
*
* NPEI_PKT_IN_PCIE_PORT = NPEI's Packet In To PCIe Port Assignment
*
* Assigns Packet Input rings to PCIe ports.
*/
union cvmx_npei_pkt_in_pcie_port
{
uint64_t u64;
struct cvmx_npei_pkt_in_pcie_port_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t pp : 64; /**< The PCIe port that the Packet ring number is
assigned. Two bits are used per ring (i.e. ring 0
[1:0], ring 1 [3:2], ....). A value of '0 means
that the Packetring is assign to PCIe Port 0, a '1'
PCIe Port 1, '2' and '3' are reserved. */
#else
uint64_t pp : 64;
#endif
} s;
struct cvmx_npei_pkt_in_pcie_port_s cn52xx;
struct cvmx_npei_pkt_in_pcie_port_s cn56xx;
};
typedef union cvmx_npei_pkt_in_pcie_port cvmx_npei_pkt_in_pcie_port_t;
/**
* cvmx_npei_pkt_input_control
*
* NPEI_PKT_INPUT_CONTROL = NPEI's Packet Input Control
*
* Control for reads for gather list and instructions.
*/
union cvmx_npei_pkt_input_control
{
uint64_t u64;
struct cvmx_npei_pkt_input_control_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_23_63 : 41;
uint64_t pkt_rr : 1; /**< When set '1' the input packet selection will be
made with a Round Robin arbitration. When '0'
the input packet ring is fixed in priority,
where the lower ring number has higher priority. */
uint64_t pbp_dhi : 13; /**< Field when in [PBP] is set to be used in
calculating a DPTR. */
uint64_t d_nsr : 1; /**< Enables '1' NoSnoop for reading of
gather data. */
uint64_t d_esr : 2; /**< The Endian-Swap-Mode for reading of
gather data. */
uint64_t d_ror : 1; /**< Enables '1' Relaxed Ordering for reading of
gather data. */
uint64_t use_csr : 1; /**< When set '1' the csr value will be used for
ROR, ESR, and NSR. When clear '0' the value in
DPTR will be used. In turn the bits not used for
ROR, ESR, and NSR, will be used for bits [63:60]
of the address used to fetch packet data. */
uint64_t nsr : 1; /**< Enables '1' NoSnoop for reading of
gather list and gather instruction. */
uint64_t esr : 2; /**< The Endian-Swap-Mode for reading of
gather list and gather instruction. */
uint64_t ror : 1; /**< Enables '1' Relaxed Ordering for reading of
gather list and gather instruction. */
#else
uint64_t ror : 1;
uint64_t esr : 2;
uint64_t nsr : 1;
uint64_t use_csr : 1;
uint64_t d_ror : 1;
uint64_t d_esr : 2;
uint64_t d_nsr : 1;
uint64_t pbp_dhi : 13;
uint64_t pkt_rr : 1;
uint64_t reserved_23_63 : 41;
#endif
} s;
struct cvmx_npei_pkt_input_control_s cn52xx;
struct cvmx_npei_pkt_input_control_s cn56xx;
};
typedef union cvmx_npei_pkt_input_control cvmx_npei_pkt_input_control_t;
/**
* cvmx_npei_pkt_instr_enb
*
* NPEI_PKT_INSTR_ENB = NPEI's Packet Instruction Enable
*
* Enables the instruction fetch for a Packet-ring.
*/
union cvmx_npei_pkt_instr_enb
{
uint64_t u64;
struct cvmx_npei_pkt_instr_enb_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_32_63 : 32;
uint64_t enb : 32; /**< When asserted '1' the vector bit cooresponding
to the Packet-ring is enabled. */
#else
uint64_t enb : 32;
uint64_t reserved_32_63 : 32;
#endif
} s;
struct cvmx_npei_pkt_instr_enb_s cn52xx;
struct cvmx_npei_pkt_instr_enb_s cn56xx;
};
typedef union cvmx_npei_pkt_instr_enb cvmx_npei_pkt_instr_enb_t;
/**
* cvmx_npei_pkt_instr_rd_size
*
* NPEI_PKT_INSTR_RD_SIZE = NPEI Instruction Read Size
*
* The number of instruction allowed to be read at one time.
*/
union cvmx_npei_pkt_instr_rd_size
{
uint64_t u64;
struct cvmx_npei_pkt_instr_rd_size_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t rdsize : 64; /**< Number of instructions to be read in one PCIe read
request for the 4 PKOport - 8 rings. Every two bits
(i.e. 1:0, 3:2, 5:4..) are assign to the port/ring
combinations.
- 15:0 PKOPort0,Ring 7..0 31:16 PKOPort1,Ring 7..0
- 47:32 PKOPort2,Ring 7..0 63:48 PKOPort3,Ring 7..0
Two bit value are:
0 - 1 Instruction
1 - 2 Instructions
2 - 3 Instructions
3 - 4 Instructions */
#else
uint64_t rdsize : 64;
#endif
} s;
struct cvmx_npei_pkt_instr_rd_size_s cn52xx;
struct cvmx_npei_pkt_instr_rd_size_s cn56xx;
};
typedef union cvmx_npei_pkt_instr_rd_size cvmx_npei_pkt_instr_rd_size_t;
/**
* cvmx_npei_pkt_instr_size
*
* NPEI_PKT_INSTR_SIZE = NPEI's Packet Instruction Size
*
* Determines if instructions are 64 or 32 byte in size for a Packet-ring.
*/
union cvmx_npei_pkt_instr_size
{
uint64_t u64;
struct cvmx_npei_pkt_instr_size_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_32_63 : 32;
uint64_t is_64b : 32; /**< When asserted '1' the vector bit cooresponding
to the Packet-ring is a 64-byte instruction. */
#else
uint64_t is_64b : 32;
uint64_t reserved_32_63 : 32;
#endif
} s;
struct cvmx_npei_pkt_instr_size_s cn52xx;
struct cvmx_npei_pkt_instr_size_s cn56xx;
};
typedef union cvmx_npei_pkt_instr_size cvmx_npei_pkt_instr_size_t;
/**
* cvmx_npei_pkt_int_levels
*
* 0x90F0 reserved NPEI_PKT_PCIE_PORT2
*
*
* NPEI_PKT_INT_LEVELS = NPEI's Packet Interrupt Levels
*
* Output packet interrupt levels.
*/
union cvmx_npei_pkt_int_levels
{
uint64_t u64;
struct cvmx_npei_pkt_int_levels_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_54_63 : 10;
uint64_t time : 22; /**< When NPEI_PKT#_CNTS[TIMER] is greater than this
value an interrupt is generated. */
uint64_t cnt : 32; /**< When NPEI_PKT#_CNTS[CNT] becomes
greater than this value an interrupt is generated. */
#else
uint64_t cnt : 32;
uint64_t time : 22;
uint64_t reserved_54_63 : 10;
#endif
} s;
struct cvmx_npei_pkt_int_levels_s cn52xx;
struct cvmx_npei_pkt_int_levels_s cn56xx;
};
typedef union cvmx_npei_pkt_int_levels cvmx_npei_pkt_int_levels_t;
/**
* cvmx_npei_pkt_iptr
*
* NPEI_PKT_IPTR = NPEI's Packet Info Poitner
*
* Controls using the Info-Pointer to store length and data.
*/
union cvmx_npei_pkt_iptr
{
uint64_t u64;
struct cvmx_npei_pkt_iptr_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_32_63 : 32;
uint64_t iptr : 32; /**< When asserted '1' the vector bit cooresponding
to the Packet-ring will use the Info-Pointer to
store length and data. */
#else
uint64_t iptr : 32;
uint64_t reserved_32_63 : 32;
#endif
} s;
struct cvmx_npei_pkt_iptr_s cn52xx;
struct cvmx_npei_pkt_iptr_s cn56xx;
};
typedef union cvmx_npei_pkt_iptr cvmx_npei_pkt_iptr_t;
/**
* cvmx_npei_pkt_out_bmode
*
* NPEI_PKT_OUT_BMODE = NPEI's Packet Out Byte Mode
*
* Control the updating of the NPEI_PKT#_CNT register.
*/
union cvmx_npei_pkt_out_bmode
{
uint64_t u64;
struct cvmx_npei_pkt_out_bmode_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_32_63 : 32;
uint64_t bmode : 32; /**< When asserted '1' the vector bit cooresponding
to the Packet-ring will have its NPEI_PKT#_CNT
register updated with the number of bytes in the
packet sent, when '0' the register will have a
value of '1' added. */
#else
uint64_t bmode : 32;
uint64_t reserved_32_63 : 32;
#endif
} s;
struct cvmx_npei_pkt_out_bmode_s cn52xx;
struct cvmx_npei_pkt_out_bmode_s cn56xx;
};
typedef union cvmx_npei_pkt_out_bmode cvmx_npei_pkt_out_bmode_t;
/**
* cvmx_npei_pkt_out_enb
*
* NPEI_PKT_OUT_ENB = NPEI's Packet Output Enable
*
* Enables the output packet engines.
*/
union cvmx_npei_pkt_out_enb
{
uint64_t u64;
struct cvmx_npei_pkt_out_enb_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_32_63 : 32;
uint64_t enb : 32; /**< When asserted '1' the vector bit cooresponding
to the Packet-ring is enabled.
If an error occurs on reading pointers for an
output ring, the ring will be disabled by clearing
the bit associated with the ring to '0'. */
#else
uint64_t enb : 32;
uint64_t reserved_32_63 : 32;
#endif
} s;
struct cvmx_npei_pkt_out_enb_s cn52xx;
struct cvmx_npei_pkt_out_enb_s cn56xx;
};
typedef union cvmx_npei_pkt_out_enb cvmx_npei_pkt_out_enb_t;
/**
* cvmx_npei_pkt_output_wmark
*
* NPEI_PKT_OUTPUT_WMARK = NPEI's Packet Output Water Mark
*
* Value that when the NPEI_PKT#_SLIST_BAOFF_DBELL[DBELL] value is less then that backpressure for the rings will be applied.
*/
union cvmx_npei_pkt_output_wmark
{
uint64_t u64;
struct cvmx_npei_pkt_output_wmark_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_32_63 : 32;
uint64_t wmark : 32; /**< Value when DBELL count drops below backpressure
for the ring will be applied to the PKO. */
#else
uint64_t wmark : 32;
uint64_t reserved_32_63 : 32;
#endif
} s;
struct cvmx_npei_pkt_output_wmark_s cn52xx;
struct cvmx_npei_pkt_output_wmark_s cn56xx;
};
typedef union cvmx_npei_pkt_output_wmark cvmx_npei_pkt_output_wmark_t;
/**
* cvmx_npei_pkt_pcie_port
*
* NPEI_PKT_PCIE_PORT = NPEI's Packet To PCIe Port Assignment
*
* Assigns Packet Ports to PCIe ports.
*/
union cvmx_npei_pkt_pcie_port
{
uint64_t u64;
struct cvmx_npei_pkt_pcie_port_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t pp : 64; /**< The PCIe port that the Packet ring number is
assigned. Two bits are used per ring (i.e. ring 0
[1:0], ring 1 [3:2], ....). A value of '0 means
that the Packetring is assign to PCIe Port 0, a '1'
PCIe Port 1, '2' and '3' are reserved. */
#else
uint64_t pp : 64;
#endif
} s;
struct cvmx_npei_pkt_pcie_port_s cn52xx;
struct cvmx_npei_pkt_pcie_port_s cn56xx;
};
typedef union cvmx_npei_pkt_pcie_port cvmx_npei_pkt_pcie_port_t;
/**
* cvmx_npei_pkt_port_in_rst
*
* NPEI_PKT_PORT_IN_RST = NPEI Packet Port In Reset
*
* Vector bits related to ring-port for ones that are reset.
*/
union cvmx_npei_pkt_port_in_rst
{
uint64_t u64;
struct cvmx_npei_pkt_port_in_rst_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t in_rst : 32; /**< When asserted '1' the vector bit cooresponding
to the inbound Packet-ring is in reset. */
uint64_t out_rst : 32; /**< When asserted '1' the vector bit cooresponding
to the outbound Packet-ring is in reset. */
#else
uint64_t out_rst : 32;
uint64_t in_rst : 32;
#endif
} s;
struct cvmx_npei_pkt_port_in_rst_s cn52xx;
struct cvmx_npei_pkt_port_in_rst_s cn56xx;
};
typedef union cvmx_npei_pkt_port_in_rst cvmx_npei_pkt_port_in_rst_t;
/**
* cvmx_npei_pkt_slist_es
*
* NPEI_PKT_SLIST_ES = NPEI's Packet Scatter List Endian Swap
*
* The Endian Swap for Scatter List Read.
*/
union cvmx_npei_pkt_slist_es
{
uint64_t u64;
struct cvmx_npei_pkt_slist_es_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t es : 64; /**< The endian swap mode for Packet rings 0 through 31.
Two bits are used per ring (i.e. ring 0 [1:0],
ring 1 [3:2], ....). */
#else
uint64_t es : 64;
#endif
} s;
struct cvmx_npei_pkt_slist_es_s cn52xx;
struct cvmx_npei_pkt_slist_es_s cn56xx;
};
typedef union cvmx_npei_pkt_slist_es cvmx_npei_pkt_slist_es_t;
/**
* cvmx_npei_pkt_slist_id_size
*
* NPEI_PKT_SLIST_ID_SIZE = NPEI Packet Scatter List Info and Data Size
*
* The Size of the information and data fields pointed to by Scatter List pointers.
*/
union cvmx_npei_pkt_slist_id_size
{
uint64_t u64;
struct cvmx_npei_pkt_slist_id_size_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_23_63 : 41;
uint64_t isize : 7; /**< Information size. Legal sizes are 0 to 120. */
uint64_t bsize : 16; /**< Data size. */
#else
uint64_t bsize : 16;
uint64_t isize : 7;
uint64_t reserved_23_63 : 41;
#endif
} s;
struct cvmx_npei_pkt_slist_id_size_s cn52xx;
struct cvmx_npei_pkt_slist_id_size_s cn56xx;
};
typedef union cvmx_npei_pkt_slist_id_size cvmx_npei_pkt_slist_id_size_t;
/**
* cvmx_npei_pkt_slist_ns
*
* NPEI_PKT_SLIST_NS = NPEI's Packet Scatter List No Snoop
*
* The NS field for the TLP when fetching Scatter List.
*/
union cvmx_npei_pkt_slist_ns
{
uint64_t u64;
struct cvmx_npei_pkt_slist_ns_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_32_63 : 32;
uint64_t nsr : 32; /**< When asserted '1' the vector bit cooresponding
to the Packet-ring will enable NS in TLP header. */
#else
uint64_t nsr : 32;
uint64_t reserved_32_63 : 32;
#endif
} s;
struct cvmx_npei_pkt_slist_ns_s cn52xx;
struct cvmx_npei_pkt_slist_ns_s cn56xx;
};
typedef union cvmx_npei_pkt_slist_ns cvmx_npei_pkt_slist_ns_t;
/**
* cvmx_npei_pkt_slist_ror
*
* NPEI_PKT_SLIST_ROR = NPEI's Packet Scatter List Relaxed Ordering
*
* The ROR field for the TLP when fetching Scatter List.
*/
union cvmx_npei_pkt_slist_ror
{
uint64_t u64;
struct cvmx_npei_pkt_slist_ror_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_32_63 : 32;
uint64_t ror : 32; /**< When asserted '1' the vector bit cooresponding
to the Packet-ring will enable ROR in TLP header. */
#else
uint64_t ror : 32;
uint64_t reserved_32_63 : 32;
#endif
} s;
struct cvmx_npei_pkt_slist_ror_s cn52xx;
struct cvmx_npei_pkt_slist_ror_s cn56xx;
};
typedef union cvmx_npei_pkt_slist_ror cvmx_npei_pkt_slist_ror_t;
/**
* cvmx_npei_pkt_time_int
*
* NPEI_PKT_TIME_INT = NPEI Packet Timer Interrupt
*
* The packets rings that are interrupting because of Packet Timers.
*/
union cvmx_npei_pkt_time_int
{
uint64_t u64;
struct cvmx_npei_pkt_time_int_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_32_63 : 32;
uint64_t port : 32; /**< Bit vector cooresponding to ring number is set when
NPEI_PKT#_CNTS[TIMER] is greater than
NPEI_PKT_INT_LEVELS[TIME]. */
#else
uint64_t port : 32;
uint64_t reserved_32_63 : 32;
#endif
} s;
struct cvmx_npei_pkt_time_int_s cn52xx;
struct cvmx_npei_pkt_time_int_s cn56xx;
};
typedef union cvmx_npei_pkt_time_int cvmx_npei_pkt_time_int_t;
/**
* cvmx_npei_pkt_time_int_enb
*
* NPEI_PKT_TIME_INT_ENB = NPEI Packet Timer Interrupt Enable
*
* The packets rings that are interrupting because of Packet Timers.
*/
union cvmx_npei_pkt_time_int_enb
{
uint64_t u64;
struct cvmx_npei_pkt_time_int_enb_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_32_63 : 32;
uint64_t port : 32; /**< Bit vector cooresponding to ring number when set
allows NPEI_PKT_TIME_INT to generate an interrupt. */
#else
uint64_t port : 32;
uint64_t reserved_32_63 : 32;
#endif
} s;
struct cvmx_npei_pkt_time_int_enb_s cn52xx;
struct cvmx_npei_pkt_time_int_enb_s cn56xx;
};
typedef union cvmx_npei_pkt_time_int_enb cvmx_npei_pkt_time_int_enb_t;
/**
* cvmx_npei_rsl_int_blocks
*
* NPEI_RSL_INT_BLOCKS = NPEI RSL Interrupt Blocks Register
*
* Reading this register will return a vector with a bit set '1' for a corresponding RSL block
* that presently has an interrupt pending. The Field Description below supplies the name of the
* register that software should read to find out why that intterupt bit is set.
*/
union cvmx_npei_rsl_int_blocks
{
uint64_t u64;
struct cvmx_npei_rsl_int_blocks_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_31_63 : 33;
uint64_t iob : 1; /**< IOB_INT_SUM */
uint64_t lmc1 : 1; /**< LMC1_MEM_CFG0 */
uint64_t agl : 1; /**< AGL_GMX_RX0_INT_REG & AGL_GMX_TX_INT_REG */
uint64_t reserved_24_27 : 4;
uint64_t asxpcs1 : 1; /**< PCS1_INT*_REG */
uint64_t asxpcs0 : 1; /**< PCS0_INT*_REG */
uint64_t reserved_21_21 : 1;
uint64_t pip : 1; /**< PIP_INT_REG. */
uint64_t spx1 : 1; /**< Always reads as zero */
uint64_t spx0 : 1; /**< Always reads as zero */
uint64_t lmc0 : 1; /**< LMC0_MEM_CFG0 */
uint64_t l2c : 1; /**< L2C_INT_STAT */
uint64_t usb1 : 1; /**< Always reads as zero */
uint64_t rad : 1; /**< RAD_REG_ERROR */
uint64_t usb : 1; /**< USBN0_INT_SUM */
uint64_t pow : 1; /**< POW_ECC_ERR */
uint64_t tim : 1; /**< TIM_REG_ERROR */
uint64_t pko : 1; /**< PKO_REG_ERROR */
uint64_t ipd : 1; /**< IPD_INT_SUM */
uint64_t reserved_8_8 : 1;
uint64_t zip : 1; /**< ZIP_ERROR */
uint64_t dfa : 1; /**< Always reads as zero */
uint64_t fpa : 1; /**< FPA_INT_SUM */
uint64_t key : 1; /**< KEY_INT_SUM */
uint64_t npei : 1; /**< NPEI_INT_SUM */
uint64_t gmx1 : 1; /**< GMX1_RX*_INT_REG & GMX1_TX_INT_REG */
uint64_t gmx0 : 1; /**< GMX0_RX*_INT_REG & GMX0_TX_INT_REG */
uint64_t mio : 1; /**< MIO_BOOT_ERR */
#else
uint64_t mio : 1;
uint64_t gmx0 : 1;
uint64_t gmx1 : 1;
uint64_t npei : 1;
uint64_t key : 1;
uint64_t fpa : 1;
uint64_t dfa : 1;
uint64_t zip : 1;
uint64_t reserved_8_8 : 1;
uint64_t ipd : 1;
uint64_t pko : 1;
uint64_t tim : 1;
uint64_t pow : 1;
uint64_t usb : 1;
uint64_t rad : 1;
uint64_t usb1 : 1;
uint64_t l2c : 1;
uint64_t lmc0 : 1;
uint64_t spx0 : 1;
uint64_t spx1 : 1;
uint64_t pip : 1;
uint64_t reserved_21_21 : 1;
uint64_t asxpcs0 : 1;
uint64_t asxpcs1 : 1;
uint64_t reserved_24_27 : 4;
uint64_t agl : 1;
uint64_t lmc1 : 1;
uint64_t iob : 1;
uint64_t reserved_31_63 : 33;
#endif
} s;
struct cvmx_npei_rsl_int_blocks_s cn52xx;
struct cvmx_npei_rsl_int_blocks_s cn52xxp1;
struct cvmx_npei_rsl_int_blocks_s cn56xx;
struct cvmx_npei_rsl_int_blocks_s cn56xxp1;
};
typedef union cvmx_npei_rsl_int_blocks cvmx_npei_rsl_int_blocks_t;
/**
* cvmx_npei_scratch_1
*
* NPEI_SCRATCH_1 = NPEI's Scratch 1
*
* A general purpose 64 bit register for SW use.
*/
union cvmx_npei_scratch_1
{
uint64_t u64;
struct cvmx_npei_scratch_1_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t data : 64; /**< The value in this register is totaly SW dependent. */
#else
uint64_t data : 64;
#endif
} s;
struct cvmx_npei_scratch_1_s cn52xx;
struct cvmx_npei_scratch_1_s cn52xxp1;
struct cvmx_npei_scratch_1_s cn56xx;
struct cvmx_npei_scratch_1_s cn56xxp1;
};
typedef union cvmx_npei_scratch_1 cvmx_npei_scratch_1_t;
/**
* cvmx_npei_state1
*
* NPEI_STATE1 = NPEI State 1
*
* State machines in NPEI. For debug.
*/
union cvmx_npei_state1
{
uint64_t u64;
struct cvmx_npei_state1_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t cpl1 : 12; /**< CPL1 State */
uint64_t cpl0 : 12; /**< CPL0 State */
uint64_t arb : 1; /**< ARB State */
uint64_t csr : 39; /**< CSR State */
#else
uint64_t csr : 39;
uint64_t arb : 1;
uint64_t cpl0 : 12;
uint64_t cpl1 : 12;
#endif
} s;
struct cvmx_npei_state1_s cn52xx;
struct cvmx_npei_state1_s cn52xxp1;
struct cvmx_npei_state1_s cn56xx;
struct cvmx_npei_state1_s cn56xxp1;
};
typedef union cvmx_npei_state1 cvmx_npei_state1_t;
/**
* cvmx_npei_state2
*
* NPEI_STATE2 = NPEI State 2
*
* State machines in NPEI. For debug.
*/
union cvmx_npei_state2
{
uint64_t u64;
struct cvmx_npei_state2_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_48_63 : 16;
uint64_t npei : 1; /**< NPEI State */
uint64_t rac : 1; /**< RAC State */
uint64_t csm1 : 15; /**< CSM1 State */
uint64_t csm0 : 15; /**< CSM0 State */
uint64_t nnp0 : 8; /**< NNP0 State */
uint64_t nnd : 8; /**< NND State */
#else
uint64_t nnd : 8;
uint64_t nnp0 : 8;
uint64_t csm0 : 15;
uint64_t csm1 : 15;
uint64_t rac : 1;
uint64_t npei : 1;
uint64_t reserved_48_63 : 16;
#endif
} s;
struct cvmx_npei_state2_s cn52xx;
struct cvmx_npei_state2_s cn52xxp1;
struct cvmx_npei_state2_s cn56xx;
struct cvmx_npei_state2_s cn56xxp1;
};
typedef union cvmx_npei_state2 cvmx_npei_state2_t;
/**
* cvmx_npei_state3
*
* NPEI_STATE3 = NPEI State 3
*
* State machines in NPEI. For debug.
*/
union cvmx_npei_state3
{
uint64_t u64;
struct cvmx_npei_state3_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_56_63 : 8;
uint64_t psm1 : 15; /**< PSM1 State */
uint64_t psm0 : 15; /**< PSM0 State */
uint64_t nsm1 : 13; /**< NSM1 State */
uint64_t nsm0 : 13; /**< NSM0 State */
#else
uint64_t nsm0 : 13;
uint64_t nsm1 : 13;
uint64_t psm0 : 15;
uint64_t psm1 : 15;
uint64_t reserved_56_63 : 8;
#endif
} s;
struct cvmx_npei_state3_s cn52xx;
struct cvmx_npei_state3_s cn52xxp1;
struct cvmx_npei_state3_s cn56xx;
struct cvmx_npei_state3_s cn56xxp1;
};
typedef union cvmx_npei_state3 cvmx_npei_state3_t;
/**
* cvmx_npei_win_rd_addr
*
* NPEI_WIN_RD_ADDR = NPEI Window Read Address Register
*
* The address to be read when the NPEI_WIN_RD_DATA register is read.
*/
union cvmx_npei_win_rd_addr
{
uint64_t u64;
struct cvmx_npei_win_rd_addr_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_51_63 : 13;
uint64_t ld_cmd : 2; /**< The load command sent wit hthe read.
0x0 == Load 8-bytes, 0x1 == Load 4-bytes,
0x2 == Load 2-bytes, 0x3 == Load 1-bytes, */
uint64_t iobit : 1; /**< A 1 or 0 can be written here but this will always
read as '0'. */
uint64_t rd_addr : 48; /**< The address to be read from. Whenever the LSB of
this register is written, the Read Operation will
take place.
[47:40] = NCB_ID
[39:0] = Address
When [47:43] == NPI & [42:0] == 0 bits [39:0] are:
[39:32] == x, Not Used
[31:27] == RSL_ID
[12:0] == RSL Register Offset */
#else
uint64_t rd_addr : 48;
uint64_t iobit : 1;
uint64_t ld_cmd : 2;
uint64_t reserved_51_63 : 13;
#endif
} s;
struct cvmx_npei_win_rd_addr_s cn52xx;
struct cvmx_npei_win_rd_addr_s cn52xxp1;
struct cvmx_npei_win_rd_addr_s cn56xx;
struct cvmx_npei_win_rd_addr_s cn56xxp1;
};
typedef union cvmx_npei_win_rd_addr cvmx_npei_win_rd_addr_t;
/**
* cvmx_npei_win_rd_data
*
* NPEI_WIN_RD_DATA = NPEI Window Read Data Register
*
* Reading this register causes a window read operation to take place. Address read is taht contained in the NPEI_WIN_RD_ADDR
* register.
*/
union cvmx_npei_win_rd_data
{
uint64_t u64;
struct cvmx_npei_win_rd_data_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t rd_data : 64; /**< The read data. */
#else
uint64_t rd_data : 64;
#endif
} s;
struct cvmx_npei_win_rd_data_s cn52xx;
struct cvmx_npei_win_rd_data_s cn52xxp1;
struct cvmx_npei_win_rd_data_s cn56xx;
struct cvmx_npei_win_rd_data_s cn56xxp1;
};
typedef union cvmx_npei_win_rd_data cvmx_npei_win_rd_data_t;
/**
* cvmx_npei_win_wr_addr
*
* NPEI_WIN_WR_ADDR = NPEI Window Write Address Register
*
* Contains the address to be writen to when a write operation is started by writing the
* NPEI_WIN_WR_DATA register (see below).
*
* Notes:
* Even though address bit [2] can be set, it should always be kept to '0'.
*
*/
union cvmx_npei_win_wr_addr
{
uint64_t u64;
struct cvmx_npei_win_wr_addr_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_49_63 : 15;
uint64_t iobit : 1; /**< A 1 or 0 can be written here but this will always
read as '0'. */
uint64_t wr_addr : 46; /**< The address that will be written to when the
NPEI_WIN_WR_DATA register is written.
[47:40] = NCB_ID
[39:3] = Address
When [47:43] == NPI & [42:0] == 0 bits [39:0] are:
[39:32] == x, Not Used
[31:27] == RSL_ID
[12:2] == RSL Register Offset
[1:0] == x, Not Used */
uint64_t reserved_0_1 : 2;
#else
uint64_t reserved_0_1 : 2;
uint64_t wr_addr : 46;
uint64_t iobit : 1;
uint64_t reserved_49_63 : 15;
#endif
} s;
struct cvmx_npei_win_wr_addr_s cn52xx;
struct cvmx_npei_win_wr_addr_s cn52xxp1;
struct cvmx_npei_win_wr_addr_s cn56xx;
struct cvmx_npei_win_wr_addr_s cn56xxp1;
};
typedef union cvmx_npei_win_wr_addr cvmx_npei_win_wr_addr_t;
/**
* cvmx_npei_win_wr_data
*
* NPEI_WIN_WR_DATA = NPEI Window Write Data Register
*
* Contains the data to write to the address located in the NPEI_WIN_WR_ADDR Register.
* Writing the least-significant-byte of this register will cause a write operation to take place.
*/
union cvmx_npei_win_wr_data
{
uint64_t u64;
struct cvmx_npei_win_wr_data_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t wr_data : 64; /**< The data to be written. Whenever the LSB of this
register is written, the Window Write will take
place. */
#else
uint64_t wr_data : 64;
#endif
} s;
struct cvmx_npei_win_wr_data_s cn52xx;
struct cvmx_npei_win_wr_data_s cn52xxp1;
struct cvmx_npei_win_wr_data_s cn56xx;
struct cvmx_npei_win_wr_data_s cn56xxp1;
};
typedef union cvmx_npei_win_wr_data cvmx_npei_win_wr_data_t;
/**
* cvmx_npei_win_wr_mask
*
* NPEI_WIN_WR_MASK = NPEI Window Write Mask Register
*
* Contains the mask for the data in the NPEI_WIN_WR_DATA Register.
*/
union cvmx_npei_win_wr_mask
{
uint64_t u64;
struct cvmx_npei_win_wr_mask_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_8_63 : 56;
uint64_t wr_mask : 8; /**< The data to be written. When a bit is '0'
the corresponding byte will be written. */
#else
uint64_t wr_mask : 8;
uint64_t reserved_8_63 : 56;
#endif
} s;
struct cvmx_npei_win_wr_mask_s cn52xx;
struct cvmx_npei_win_wr_mask_s cn52xxp1;
struct cvmx_npei_win_wr_mask_s cn56xx;
struct cvmx_npei_win_wr_mask_s cn56xxp1;
};
typedef union cvmx_npei_win_wr_mask cvmx_npei_win_wr_mask_t;
/**
* cvmx_npei_window_ctl
*
* NPEI_WINDOW_CTL = NPEI's Window Control
*
* The name of this register is misleading. The timeout value is used for BAR0 access from PCIE0 and PCIE1.
* Any access to the regigisters on the RML will timeout as 0xFFFF clock cycle. At time of timeout the next
* RML access will start, and interrupt will be set, and in the case of reads no data will be returned.
*
* The value of this register should be set to a minimum of 0x200000 to ensure that a timeout to an RML register
* occurs on the RML 0xFFFF timer before the timeout for a BAR0 access from the PCIE#.
*/
union cvmx_npei_window_ctl
{
uint64_t u64;
struct cvmx_npei_window_ctl_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_32_63 : 32;
uint64_t time : 32; /**< Time to wait in core clocks to wait for a
BAR0 access to completeon the NCB
before timing out. A value of 0 will cause no
timeouts. A minimum value of 0x200000 should be
used when this register is not set to 0x0. */
#else
uint64_t time : 32;
uint64_t reserved_32_63 : 32;
#endif
} s;
struct cvmx_npei_window_ctl_s cn52xx;
struct cvmx_npei_window_ctl_s cn52xxp1;
struct cvmx_npei_window_ctl_s cn56xx;
struct cvmx_npei_window_ctl_s cn56xxp1;
};
typedef union cvmx_npei_window_ctl cvmx_npei_window_ctl_t;
#endif