freebsd-nq/sys/contrib/octeon-sdk/cvmx-pcmx-defs.h
Juli Mallett 04b6fa8330 Merge Cavium Octeon SDK 2.0 Simple Executive; this brings some fixes and new
facilities as well as support for the Octeon 2 family of SoCs.

XXX Note that with our antediluvian assembler, we can't support some Octeon 2
    instructions and fall back to using the old ones instead.
2010-11-28 08:18:16 +00:00

1083 lines
39 KiB
C

/***********************license start***************
* Copyright (c) 2003-2010 Cavium Networks (support@cavium.com). All rights
* reserved.
*
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions are
* met:
*
* * Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
*
* * Redistributions in binary form must reproduce the above
* copyright notice, this list of conditions and the following
* disclaimer in the documentation and/or other materials provided
* with the distribution.
* * Neither the name of Cavium Networks nor the names of
* its contributors may be used to endorse or promote products
* derived from this software without specific prior written
* permission.
* This Software, including technical data, may be subject to U.S. export control
* laws, including the U.S. Export Administration Act and its associated
* regulations, and may be subject to export or import regulations in other
* countries.
* TO THE MAXIMUM EXTENT PERMITTED BY LAW, THE SOFTWARE IS PROVIDED "AS IS"
* AND WITH ALL FAULTS AND CAVIUM NETWORKS MAKES NO PROMISES, REPRESENTATIONS OR
* WARRANTIES, EITHER EXPRESS, IMPLIED, STATUTORY, OR OTHERWISE, WITH RESPECT TO
* THE SOFTWARE, INCLUDING ITS CONDITION, ITS CONFORMITY TO ANY REPRESENTATION OR
* DESCRIPTION, OR THE EXISTENCE OF ANY LATENT OR PATENT DEFECTS, AND CAVIUM
* SPECIFICALLY DISCLAIMS ALL IMPLIED (IF ANY) WARRANTIES OF TITLE,
* MERCHANTABILITY, NONINFRINGEMENT, FITNESS FOR A PARTICULAR PURPOSE, LACK OF
* VIRUSES, ACCURACY OR COMPLETENESS, QUIET ENJOYMENT, QUIET POSSESSION OR
* CORRESPONDENCE TO DESCRIPTION. THE ENTIRE RISK ARISING OUT OF USE OR
* PERFORMANCE OF THE SOFTWARE LIES WITH YOU.
***********************license end**************************************/
/**
* cvmx-pcmx-defs.h
*
* Configuration and status register (CSR) type definitions for
* Octeon pcmx.
*
* This file is auto generated. Do not edit.
*
* <hr>$Revision$<hr>
*
*/
#ifndef __CVMX_PCMX_TYPEDEFS_H__
#define __CVMX_PCMX_TYPEDEFS_H__
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_PCMX_DMA_CFG(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN30XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN31XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN50XX) && ((offset <= 3)))))
cvmx_warn("CVMX_PCMX_DMA_CFG(%lu) is invalid on this chip\n", offset);
return CVMX_ADD_IO_SEG(0x0001070000010018ull) + ((offset) & 3) * 16384;
}
#else
#define CVMX_PCMX_DMA_CFG(offset) (CVMX_ADD_IO_SEG(0x0001070000010018ull) + ((offset) & 3) * 16384)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_PCMX_INT_ENA(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN30XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN31XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN50XX) && ((offset <= 3)))))
cvmx_warn("CVMX_PCMX_INT_ENA(%lu) is invalid on this chip\n", offset);
return CVMX_ADD_IO_SEG(0x0001070000010020ull) + ((offset) & 3) * 16384;
}
#else
#define CVMX_PCMX_INT_ENA(offset) (CVMX_ADD_IO_SEG(0x0001070000010020ull) + ((offset) & 3) * 16384)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_PCMX_INT_SUM(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN30XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN31XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN50XX) && ((offset <= 3)))))
cvmx_warn("CVMX_PCMX_INT_SUM(%lu) is invalid on this chip\n", offset);
return CVMX_ADD_IO_SEG(0x0001070000010028ull) + ((offset) & 3) * 16384;
}
#else
#define CVMX_PCMX_INT_SUM(offset) (CVMX_ADD_IO_SEG(0x0001070000010028ull) + ((offset) & 3) * 16384)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_PCMX_RXADDR(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN30XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN31XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN50XX) && ((offset <= 3)))))
cvmx_warn("CVMX_PCMX_RXADDR(%lu) is invalid on this chip\n", offset);
return CVMX_ADD_IO_SEG(0x0001070000010068ull) + ((offset) & 3) * 16384;
}
#else
#define CVMX_PCMX_RXADDR(offset) (CVMX_ADD_IO_SEG(0x0001070000010068ull) + ((offset) & 3) * 16384)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_PCMX_RXCNT(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN30XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN31XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN50XX) && ((offset <= 3)))))
cvmx_warn("CVMX_PCMX_RXCNT(%lu) is invalid on this chip\n", offset);
return CVMX_ADD_IO_SEG(0x0001070000010060ull) + ((offset) & 3) * 16384;
}
#else
#define CVMX_PCMX_RXCNT(offset) (CVMX_ADD_IO_SEG(0x0001070000010060ull) + ((offset) & 3) * 16384)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_PCMX_RXMSK0(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN30XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN31XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN50XX) && ((offset <= 3)))))
cvmx_warn("CVMX_PCMX_RXMSK0(%lu) is invalid on this chip\n", offset);
return CVMX_ADD_IO_SEG(0x00010700000100C0ull) + ((offset) & 3) * 16384;
}
#else
#define CVMX_PCMX_RXMSK0(offset) (CVMX_ADD_IO_SEG(0x00010700000100C0ull) + ((offset) & 3) * 16384)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_PCMX_RXMSK1(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN30XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN31XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN50XX) && ((offset <= 3)))))
cvmx_warn("CVMX_PCMX_RXMSK1(%lu) is invalid on this chip\n", offset);
return CVMX_ADD_IO_SEG(0x00010700000100C8ull) + ((offset) & 3) * 16384;
}
#else
#define CVMX_PCMX_RXMSK1(offset) (CVMX_ADD_IO_SEG(0x00010700000100C8ull) + ((offset) & 3) * 16384)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_PCMX_RXMSK2(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN30XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN31XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN50XX) && ((offset <= 3)))))
cvmx_warn("CVMX_PCMX_RXMSK2(%lu) is invalid on this chip\n", offset);
return CVMX_ADD_IO_SEG(0x00010700000100D0ull) + ((offset) & 3) * 16384;
}
#else
#define CVMX_PCMX_RXMSK2(offset) (CVMX_ADD_IO_SEG(0x00010700000100D0ull) + ((offset) & 3) * 16384)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_PCMX_RXMSK3(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN30XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN31XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN50XX) && ((offset <= 3)))))
cvmx_warn("CVMX_PCMX_RXMSK3(%lu) is invalid on this chip\n", offset);
return CVMX_ADD_IO_SEG(0x00010700000100D8ull) + ((offset) & 3) * 16384;
}
#else
#define CVMX_PCMX_RXMSK3(offset) (CVMX_ADD_IO_SEG(0x00010700000100D8ull) + ((offset) & 3) * 16384)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_PCMX_RXMSK4(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN30XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN31XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN50XX) && ((offset <= 3)))))
cvmx_warn("CVMX_PCMX_RXMSK4(%lu) is invalid on this chip\n", offset);
return CVMX_ADD_IO_SEG(0x00010700000100E0ull) + ((offset) & 3) * 16384;
}
#else
#define CVMX_PCMX_RXMSK4(offset) (CVMX_ADD_IO_SEG(0x00010700000100E0ull) + ((offset) & 3) * 16384)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_PCMX_RXMSK5(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN30XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN31XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN50XX) && ((offset <= 3)))))
cvmx_warn("CVMX_PCMX_RXMSK5(%lu) is invalid on this chip\n", offset);
return CVMX_ADD_IO_SEG(0x00010700000100E8ull) + ((offset) & 3) * 16384;
}
#else
#define CVMX_PCMX_RXMSK5(offset) (CVMX_ADD_IO_SEG(0x00010700000100E8ull) + ((offset) & 3) * 16384)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_PCMX_RXMSK6(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN30XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN31XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN50XX) && ((offset <= 3)))))
cvmx_warn("CVMX_PCMX_RXMSK6(%lu) is invalid on this chip\n", offset);
return CVMX_ADD_IO_SEG(0x00010700000100F0ull) + ((offset) & 3) * 16384;
}
#else
#define CVMX_PCMX_RXMSK6(offset) (CVMX_ADD_IO_SEG(0x00010700000100F0ull) + ((offset) & 3) * 16384)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_PCMX_RXMSK7(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN30XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN31XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN50XX) && ((offset <= 3)))))
cvmx_warn("CVMX_PCMX_RXMSK7(%lu) is invalid on this chip\n", offset);
return CVMX_ADD_IO_SEG(0x00010700000100F8ull) + ((offset) & 3) * 16384;
}
#else
#define CVMX_PCMX_RXMSK7(offset) (CVMX_ADD_IO_SEG(0x00010700000100F8ull) + ((offset) & 3) * 16384)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_PCMX_RXSTART(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN30XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN31XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN50XX) && ((offset <= 3)))))
cvmx_warn("CVMX_PCMX_RXSTART(%lu) is invalid on this chip\n", offset);
return CVMX_ADD_IO_SEG(0x0001070000010058ull) + ((offset) & 3) * 16384;
}
#else
#define CVMX_PCMX_RXSTART(offset) (CVMX_ADD_IO_SEG(0x0001070000010058ull) + ((offset) & 3) * 16384)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_PCMX_TDM_CFG(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN30XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN31XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN50XX) && ((offset <= 3)))))
cvmx_warn("CVMX_PCMX_TDM_CFG(%lu) is invalid on this chip\n", offset);
return CVMX_ADD_IO_SEG(0x0001070000010010ull) + ((offset) & 3) * 16384;
}
#else
#define CVMX_PCMX_TDM_CFG(offset) (CVMX_ADD_IO_SEG(0x0001070000010010ull) + ((offset) & 3) * 16384)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_PCMX_TDM_DBG(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN30XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN31XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN50XX) && ((offset <= 3)))))
cvmx_warn("CVMX_PCMX_TDM_DBG(%lu) is invalid on this chip\n", offset);
return CVMX_ADD_IO_SEG(0x0001070000010030ull) + ((offset) & 3) * 16384;
}
#else
#define CVMX_PCMX_TDM_DBG(offset) (CVMX_ADD_IO_SEG(0x0001070000010030ull) + ((offset) & 3) * 16384)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_PCMX_TXADDR(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN30XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN31XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN50XX) && ((offset <= 3)))))
cvmx_warn("CVMX_PCMX_TXADDR(%lu) is invalid on this chip\n", offset);
return CVMX_ADD_IO_SEG(0x0001070000010050ull) + ((offset) & 3) * 16384;
}
#else
#define CVMX_PCMX_TXADDR(offset) (CVMX_ADD_IO_SEG(0x0001070000010050ull) + ((offset) & 3) * 16384)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_PCMX_TXCNT(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN30XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN31XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN50XX) && ((offset <= 3)))))
cvmx_warn("CVMX_PCMX_TXCNT(%lu) is invalid on this chip\n", offset);
return CVMX_ADD_IO_SEG(0x0001070000010048ull) + ((offset) & 3) * 16384;
}
#else
#define CVMX_PCMX_TXCNT(offset) (CVMX_ADD_IO_SEG(0x0001070000010048ull) + ((offset) & 3) * 16384)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_PCMX_TXMSK0(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN30XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN31XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN50XX) && ((offset <= 3)))))
cvmx_warn("CVMX_PCMX_TXMSK0(%lu) is invalid on this chip\n", offset);
return CVMX_ADD_IO_SEG(0x0001070000010080ull) + ((offset) & 3) * 16384;
}
#else
#define CVMX_PCMX_TXMSK0(offset) (CVMX_ADD_IO_SEG(0x0001070000010080ull) + ((offset) & 3) * 16384)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_PCMX_TXMSK1(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN30XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN31XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN50XX) && ((offset <= 3)))))
cvmx_warn("CVMX_PCMX_TXMSK1(%lu) is invalid on this chip\n", offset);
return CVMX_ADD_IO_SEG(0x0001070000010088ull) + ((offset) & 3) * 16384;
}
#else
#define CVMX_PCMX_TXMSK1(offset) (CVMX_ADD_IO_SEG(0x0001070000010088ull) + ((offset) & 3) * 16384)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_PCMX_TXMSK2(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN30XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN31XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN50XX) && ((offset <= 3)))))
cvmx_warn("CVMX_PCMX_TXMSK2(%lu) is invalid on this chip\n", offset);
return CVMX_ADD_IO_SEG(0x0001070000010090ull) + ((offset) & 3) * 16384;
}
#else
#define CVMX_PCMX_TXMSK2(offset) (CVMX_ADD_IO_SEG(0x0001070000010090ull) + ((offset) & 3) * 16384)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_PCMX_TXMSK3(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN30XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN31XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN50XX) && ((offset <= 3)))))
cvmx_warn("CVMX_PCMX_TXMSK3(%lu) is invalid on this chip\n", offset);
return CVMX_ADD_IO_SEG(0x0001070000010098ull) + ((offset) & 3) * 16384;
}
#else
#define CVMX_PCMX_TXMSK3(offset) (CVMX_ADD_IO_SEG(0x0001070000010098ull) + ((offset) & 3) * 16384)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_PCMX_TXMSK4(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN30XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN31XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN50XX) && ((offset <= 3)))))
cvmx_warn("CVMX_PCMX_TXMSK4(%lu) is invalid on this chip\n", offset);
return CVMX_ADD_IO_SEG(0x00010700000100A0ull) + ((offset) & 3) * 16384;
}
#else
#define CVMX_PCMX_TXMSK4(offset) (CVMX_ADD_IO_SEG(0x00010700000100A0ull) + ((offset) & 3) * 16384)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_PCMX_TXMSK5(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN30XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN31XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN50XX) && ((offset <= 3)))))
cvmx_warn("CVMX_PCMX_TXMSK5(%lu) is invalid on this chip\n", offset);
return CVMX_ADD_IO_SEG(0x00010700000100A8ull) + ((offset) & 3) * 16384;
}
#else
#define CVMX_PCMX_TXMSK5(offset) (CVMX_ADD_IO_SEG(0x00010700000100A8ull) + ((offset) & 3) * 16384)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_PCMX_TXMSK6(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN30XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN31XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN50XX) && ((offset <= 3)))))
cvmx_warn("CVMX_PCMX_TXMSK6(%lu) is invalid on this chip\n", offset);
return CVMX_ADD_IO_SEG(0x00010700000100B0ull) + ((offset) & 3) * 16384;
}
#else
#define CVMX_PCMX_TXMSK6(offset) (CVMX_ADD_IO_SEG(0x00010700000100B0ull) + ((offset) & 3) * 16384)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_PCMX_TXMSK7(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN30XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN31XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN50XX) && ((offset <= 3)))))
cvmx_warn("CVMX_PCMX_TXMSK7(%lu) is invalid on this chip\n", offset);
return CVMX_ADD_IO_SEG(0x00010700000100B8ull) + ((offset) & 3) * 16384;
}
#else
#define CVMX_PCMX_TXMSK7(offset) (CVMX_ADD_IO_SEG(0x00010700000100B8ull) + ((offset) & 3) * 16384)
#endif
#if CVMX_ENABLE_CSR_ADDRESS_CHECKING
static inline uint64_t CVMX_PCMX_TXSTART(unsigned long offset)
{
if (!(
(OCTEON_IS_MODEL(OCTEON_CN30XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN31XX) && ((offset <= 3))) ||
(OCTEON_IS_MODEL(OCTEON_CN50XX) && ((offset <= 3)))))
cvmx_warn("CVMX_PCMX_TXSTART(%lu) is invalid on this chip\n", offset);
return CVMX_ADD_IO_SEG(0x0001070000010040ull) + ((offset) & 3) * 16384;
}
#else
#define CVMX_PCMX_TXSTART(offset) (CVMX_ADD_IO_SEG(0x0001070000010040ull) + ((offset) & 3) * 16384)
#endif
/**
* cvmx_pcm#_dma_cfg
*/
union cvmx_pcmx_dma_cfg
{
uint64_t u64;
struct cvmx_pcmx_dma_cfg_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t rdpend : 1; /**< If 0, no L2C read responses pending
1, L2C read responses are outstanding
NOTE: When restarting after stopping a running TDM
engine, software must wait for RDPEND to read 0
before writing PCMn_TDM_CFG[ENABLE] to a 1 */
uint64_t reserved_54_62 : 9;
uint64_t rxslots : 10; /**< Number of 8-bit slots to receive per frame
(number of slots in a receive superframe) */
uint64_t reserved_42_43 : 2;
uint64_t txslots : 10; /**< Number of 8-bit slots to transmit per frame
(number of slots in a transmit superframe) */
uint64_t reserved_30_31 : 2;
uint64_t rxst : 10; /**< Number of frame writes for interrupt */
uint64_t reserved_19_19 : 1;
uint64_t useldt : 1; /**< If 0, use LDI command to read from L2C
1, use LDT command to read from L2C */
uint64_t txrd : 10; /**< Number of frame reads for interrupt */
uint64_t fetchsiz : 4; /**< FETCHSIZ+1 timeslots are read when threshold is
reached. */
uint64_t thresh : 4; /**< If number of bytes remaining in the DMA fifo is <=
THRESH, initiate a fetch of timeslot data from the
transmit memory region.
NOTE: there are only 16B of buffer for each engine
so the seetings for FETCHSIZ and THRESH must be
such that the buffer will not be overrun:
THRESH + min(FETCHSIZ + 1,TXSLOTS) MUST BE <= 16 */
#else
uint64_t thresh : 4;
uint64_t fetchsiz : 4;
uint64_t txrd : 10;
uint64_t useldt : 1;
uint64_t reserved_19_19 : 1;
uint64_t rxst : 10;
uint64_t reserved_30_31 : 2;
uint64_t txslots : 10;
uint64_t reserved_42_43 : 2;
uint64_t rxslots : 10;
uint64_t reserved_54_62 : 9;
uint64_t rdpend : 1;
#endif
} s;
struct cvmx_pcmx_dma_cfg_s cn30xx;
struct cvmx_pcmx_dma_cfg_s cn31xx;
struct cvmx_pcmx_dma_cfg_s cn50xx;
};
typedef union cvmx_pcmx_dma_cfg cvmx_pcmx_dma_cfg_t;
/**
* cvmx_pcm#_int_ena
*/
union cvmx_pcmx_int_ena
{
uint64_t u64;
struct cvmx_pcmx_int_ena_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_8_63 : 56;
uint64_t rxovf : 1; /**< Enable interrupt if RX byte overflows */
uint64_t txempty : 1; /**< Enable interrupt on TX byte empty */
uint64_t txrd : 1; /**< Enable DMA engine frame read interrupts */
uint64_t txwrap : 1; /**< Enable TX region wrap interrupts */
uint64_t rxst : 1; /**< Enable DMA engine frame store interrupts */
uint64_t rxwrap : 1; /**< Enable RX region wrap interrupts */
uint64_t fsyncextra : 1; /**< Enable FSYNC extra interrupts
NOTE: FSYNCEXTRA errors are defined as an FSYNC
found in the "wrong" spot of a frame given the
programming of PCMn_CLK_CFG[NUMSLOTS] and
PCMn_CLK_CFG[EXTRABIT]. */
uint64_t fsyncmissed : 1; /**< Enable FSYNC missed interrupts
NOTE: FSYNCMISSED errors are defined as an FSYNC
missing from the correct spot in a frame given
the programming of PCMn_CLK_CFG[NUMSLOTS] and
PCMn_CLK_CFG[EXTRABIT]. */
#else
uint64_t fsyncmissed : 1;
uint64_t fsyncextra : 1;
uint64_t rxwrap : 1;
uint64_t rxst : 1;
uint64_t txwrap : 1;
uint64_t txrd : 1;
uint64_t txempty : 1;
uint64_t rxovf : 1;
uint64_t reserved_8_63 : 56;
#endif
} s;
struct cvmx_pcmx_int_ena_s cn30xx;
struct cvmx_pcmx_int_ena_s cn31xx;
struct cvmx_pcmx_int_ena_s cn50xx;
};
typedef union cvmx_pcmx_int_ena cvmx_pcmx_int_ena_t;
/**
* cvmx_pcm#_int_sum
*/
union cvmx_pcmx_int_sum
{
uint64_t u64;
struct cvmx_pcmx_int_sum_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_8_63 : 56;
uint64_t rxovf : 1; /**< RX byte overflowed */
uint64_t txempty : 1; /**< TX byte was empty when sampled */
uint64_t txrd : 1; /**< DMA engine frame read interrupt occurred */
uint64_t txwrap : 1; /**< TX region wrap interrupt occurred */
uint64_t rxst : 1; /**< DMA engine frame store interrupt occurred */
uint64_t rxwrap : 1; /**< RX region wrap interrupt occurred */
uint64_t fsyncextra : 1; /**< FSYNC extra interrupt occurred */
uint64_t fsyncmissed : 1; /**< FSYNC missed interrupt occurred */
#else
uint64_t fsyncmissed : 1;
uint64_t fsyncextra : 1;
uint64_t rxwrap : 1;
uint64_t rxst : 1;
uint64_t txwrap : 1;
uint64_t txrd : 1;
uint64_t txempty : 1;
uint64_t rxovf : 1;
uint64_t reserved_8_63 : 56;
#endif
} s;
struct cvmx_pcmx_int_sum_s cn30xx;
struct cvmx_pcmx_int_sum_s cn31xx;
struct cvmx_pcmx_int_sum_s cn50xx;
};
typedef union cvmx_pcmx_int_sum cvmx_pcmx_int_sum_t;
/**
* cvmx_pcm#_rxaddr
*/
union cvmx_pcmx_rxaddr
{
uint64_t u64;
struct cvmx_pcmx_rxaddr_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_36_63 : 28;
uint64_t addr : 36; /**< Address of the next write to the receive memory
region */
#else
uint64_t addr : 36;
uint64_t reserved_36_63 : 28;
#endif
} s;
struct cvmx_pcmx_rxaddr_s cn30xx;
struct cvmx_pcmx_rxaddr_s cn31xx;
struct cvmx_pcmx_rxaddr_s cn50xx;
};
typedef union cvmx_pcmx_rxaddr cvmx_pcmx_rxaddr_t;
/**
* cvmx_pcm#_rxcnt
*/
union cvmx_pcmx_rxcnt
{
uint64_t u64;
struct cvmx_pcmx_rxcnt_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_16_63 : 48;
uint64_t cnt : 16; /**< Number of superframes in receive memory region */
#else
uint64_t cnt : 16;
uint64_t reserved_16_63 : 48;
#endif
} s;
struct cvmx_pcmx_rxcnt_s cn30xx;
struct cvmx_pcmx_rxcnt_s cn31xx;
struct cvmx_pcmx_rxcnt_s cn50xx;
};
typedef union cvmx_pcmx_rxcnt cvmx_pcmx_rxcnt_t;
/**
* cvmx_pcm#_rxmsk0
*/
union cvmx_pcmx_rxmsk0
{
uint64_t u64;
struct cvmx_pcmx_rxmsk0_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t mask : 64; /**< Receive mask bits for slots 63 to 0
(1 means transmit, 0 means don't transmit) */
#else
uint64_t mask : 64;
#endif
} s;
struct cvmx_pcmx_rxmsk0_s cn30xx;
struct cvmx_pcmx_rxmsk0_s cn31xx;
struct cvmx_pcmx_rxmsk0_s cn50xx;
};
typedef union cvmx_pcmx_rxmsk0 cvmx_pcmx_rxmsk0_t;
/**
* cvmx_pcm#_rxmsk1
*/
union cvmx_pcmx_rxmsk1
{
uint64_t u64;
struct cvmx_pcmx_rxmsk1_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t mask : 64; /**< Receive mask bits for slots 127 to 64
(1 means transmit, 0 means don't transmit) */
#else
uint64_t mask : 64;
#endif
} s;
struct cvmx_pcmx_rxmsk1_s cn30xx;
struct cvmx_pcmx_rxmsk1_s cn31xx;
struct cvmx_pcmx_rxmsk1_s cn50xx;
};
typedef union cvmx_pcmx_rxmsk1 cvmx_pcmx_rxmsk1_t;
/**
* cvmx_pcm#_rxmsk2
*/
union cvmx_pcmx_rxmsk2
{
uint64_t u64;
struct cvmx_pcmx_rxmsk2_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t mask : 64; /**< Receive mask bits for slots 191 to 128
(1 means transmit, 0 means don't transmit) */
#else
uint64_t mask : 64;
#endif
} s;
struct cvmx_pcmx_rxmsk2_s cn30xx;
struct cvmx_pcmx_rxmsk2_s cn31xx;
struct cvmx_pcmx_rxmsk2_s cn50xx;
};
typedef union cvmx_pcmx_rxmsk2 cvmx_pcmx_rxmsk2_t;
/**
* cvmx_pcm#_rxmsk3
*/
union cvmx_pcmx_rxmsk3
{
uint64_t u64;
struct cvmx_pcmx_rxmsk3_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t mask : 64; /**< Receive mask bits for slots 255 to 192
(1 means transmit, 0 means don't transmit) */
#else
uint64_t mask : 64;
#endif
} s;
struct cvmx_pcmx_rxmsk3_s cn30xx;
struct cvmx_pcmx_rxmsk3_s cn31xx;
struct cvmx_pcmx_rxmsk3_s cn50xx;
};
typedef union cvmx_pcmx_rxmsk3 cvmx_pcmx_rxmsk3_t;
/**
* cvmx_pcm#_rxmsk4
*/
union cvmx_pcmx_rxmsk4
{
uint64_t u64;
struct cvmx_pcmx_rxmsk4_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t mask : 64; /**< Receive mask bits for slots 319 to 256
(1 means transmit, 0 means don't transmit) */
#else
uint64_t mask : 64;
#endif
} s;
struct cvmx_pcmx_rxmsk4_s cn30xx;
struct cvmx_pcmx_rxmsk4_s cn31xx;
struct cvmx_pcmx_rxmsk4_s cn50xx;
};
typedef union cvmx_pcmx_rxmsk4 cvmx_pcmx_rxmsk4_t;
/**
* cvmx_pcm#_rxmsk5
*/
union cvmx_pcmx_rxmsk5
{
uint64_t u64;
struct cvmx_pcmx_rxmsk5_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t mask : 64; /**< Receive mask bits for slots 383 to 320
(1 means transmit, 0 means don't transmit) */
#else
uint64_t mask : 64;
#endif
} s;
struct cvmx_pcmx_rxmsk5_s cn30xx;
struct cvmx_pcmx_rxmsk5_s cn31xx;
struct cvmx_pcmx_rxmsk5_s cn50xx;
};
typedef union cvmx_pcmx_rxmsk5 cvmx_pcmx_rxmsk5_t;
/**
* cvmx_pcm#_rxmsk6
*/
union cvmx_pcmx_rxmsk6
{
uint64_t u64;
struct cvmx_pcmx_rxmsk6_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t mask : 64; /**< Receive mask bits for slots 447 to 384
(1 means transmit, 0 means don't transmit) */
#else
uint64_t mask : 64;
#endif
} s;
struct cvmx_pcmx_rxmsk6_s cn30xx;
struct cvmx_pcmx_rxmsk6_s cn31xx;
struct cvmx_pcmx_rxmsk6_s cn50xx;
};
typedef union cvmx_pcmx_rxmsk6 cvmx_pcmx_rxmsk6_t;
/**
* cvmx_pcm#_rxmsk7
*/
union cvmx_pcmx_rxmsk7
{
uint64_t u64;
struct cvmx_pcmx_rxmsk7_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t mask : 64; /**< Receive mask bits for slots 511 to 448
(1 means transmit, 0 means don't transmit) */
#else
uint64_t mask : 64;
#endif
} s;
struct cvmx_pcmx_rxmsk7_s cn30xx;
struct cvmx_pcmx_rxmsk7_s cn31xx;
struct cvmx_pcmx_rxmsk7_s cn50xx;
};
typedef union cvmx_pcmx_rxmsk7 cvmx_pcmx_rxmsk7_t;
/**
* cvmx_pcm#_rxstart
*/
union cvmx_pcmx_rxstart
{
uint64_t u64;
struct cvmx_pcmx_rxstart_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_36_63 : 28;
uint64_t addr : 33; /**< Starting address for the receive memory region */
uint64_t reserved_0_2 : 3;
#else
uint64_t reserved_0_2 : 3;
uint64_t addr : 33;
uint64_t reserved_36_63 : 28;
#endif
} s;
struct cvmx_pcmx_rxstart_s cn30xx;
struct cvmx_pcmx_rxstart_s cn31xx;
struct cvmx_pcmx_rxstart_s cn50xx;
};
typedef union cvmx_pcmx_rxstart cvmx_pcmx_rxstart_t;
/**
* cvmx_pcm#_tdm_cfg
*/
union cvmx_pcmx_tdm_cfg
{
uint64_t u64;
struct cvmx_pcmx_tdm_cfg_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t drvtim : 16; /**< Number of ECLKs from start of bit time to stop
driving last bit of timeslot (if not driving next
timeslot) */
uint64_t samppt : 16; /**< Number of ECLKs from start of bit time to sample
data bit. */
uint64_t reserved_3_31 : 29;
uint64_t lsbfirst : 1; /**< If 0, shift/receive MSB first
1, shift/receive LSB first */
uint64_t useclk1 : 1; /**< If 0, this PCM is based on BCLK/FSYNC0
1, this PCM is based on BCLK/FSYNC1 */
uint64_t enable : 1; /**< If 1, PCM is enabled, otherwise pins are GPIOs
NOTE: when TDM is disabled by detection of an
FSYNC error all transmission and reception is
halted. In addition, PCMn_TX/RXADDR are updated
to point to the position at which the error was
detected. */
#else
uint64_t enable : 1;
uint64_t useclk1 : 1;
uint64_t lsbfirst : 1;
uint64_t reserved_3_31 : 29;
uint64_t samppt : 16;
uint64_t drvtim : 16;
#endif
} s;
struct cvmx_pcmx_tdm_cfg_s cn30xx;
struct cvmx_pcmx_tdm_cfg_s cn31xx;
struct cvmx_pcmx_tdm_cfg_s cn50xx;
};
typedef union cvmx_pcmx_tdm_cfg cvmx_pcmx_tdm_cfg_t;
/**
* cvmx_pcm#_tdm_dbg
*/
union cvmx_pcmx_tdm_dbg
{
uint64_t u64;
struct cvmx_pcmx_tdm_dbg_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t debuginfo : 64; /**< Miscellaneous debug information */
#else
uint64_t debuginfo : 64;
#endif
} s;
struct cvmx_pcmx_tdm_dbg_s cn30xx;
struct cvmx_pcmx_tdm_dbg_s cn31xx;
struct cvmx_pcmx_tdm_dbg_s cn50xx;
};
typedef union cvmx_pcmx_tdm_dbg cvmx_pcmx_tdm_dbg_t;
/**
* cvmx_pcm#_txaddr
*/
union cvmx_pcmx_txaddr
{
uint64_t u64;
struct cvmx_pcmx_txaddr_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_36_63 : 28;
uint64_t addr : 33; /**< Address of the next read from the transmit memory
region */
uint64_t fram : 3; /**< Frame offset
NOTE: this is used to extract the correct byte from
each 64b word read from the transmit memory region */
#else
uint64_t fram : 3;
uint64_t addr : 33;
uint64_t reserved_36_63 : 28;
#endif
} s;
struct cvmx_pcmx_txaddr_s cn30xx;
struct cvmx_pcmx_txaddr_s cn31xx;
struct cvmx_pcmx_txaddr_s cn50xx;
};
typedef union cvmx_pcmx_txaddr cvmx_pcmx_txaddr_t;
/**
* cvmx_pcm#_txcnt
*/
union cvmx_pcmx_txcnt
{
uint64_t u64;
struct cvmx_pcmx_txcnt_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_16_63 : 48;
uint64_t cnt : 16; /**< Number of superframes in transmit memory region */
#else
uint64_t cnt : 16;
uint64_t reserved_16_63 : 48;
#endif
} s;
struct cvmx_pcmx_txcnt_s cn30xx;
struct cvmx_pcmx_txcnt_s cn31xx;
struct cvmx_pcmx_txcnt_s cn50xx;
};
typedef union cvmx_pcmx_txcnt cvmx_pcmx_txcnt_t;
/**
* cvmx_pcm#_txmsk0
*/
union cvmx_pcmx_txmsk0
{
uint64_t u64;
struct cvmx_pcmx_txmsk0_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t mask : 64; /**< Transmit mask bits for slots 63 to 0
(1 means transmit, 0 means don't transmit) */
#else
uint64_t mask : 64;
#endif
} s;
struct cvmx_pcmx_txmsk0_s cn30xx;
struct cvmx_pcmx_txmsk0_s cn31xx;
struct cvmx_pcmx_txmsk0_s cn50xx;
};
typedef union cvmx_pcmx_txmsk0 cvmx_pcmx_txmsk0_t;
/**
* cvmx_pcm#_txmsk1
*/
union cvmx_pcmx_txmsk1
{
uint64_t u64;
struct cvmx_pcmx_txmsk1_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t mask : 64; /**< Transmit mask bits for slots 127 to 64
(1 means transmit, 0 means don't transmit) */
#else
uint64_t mask : 64;
#endif
} s;
struct cvmx_pcmx_txmsk1_s cn30xx;
struct cvmx_pcmx_txmsk1_s cn31xx;
struct cvmx_pcmx_txmsk1_s cn50xx;
};
typedef union cvmx_pcmx_txmsk1 cvmx_pcmx_txmsk1_t;
/**
* cvmx_pcm#_txmsk2
*/
union cvmx_pcmx_txmsk2
{
uint64_t u64;
struct cvmx_pcmx_txmsk2_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t mask : 64; /**< Transmit mask bits for slots 191 to 128
(1 means transmit, 0 means don't transmit) */
#else
uint64_t mask : 64;
#endif
} s;
struct cvmx_pcmx_txmsk2_s cn30xx;
struct cvmx_pcmx_txmsk2_s cn31xx;
struct cvmx_pcmx_txmsk2_s cn50xx;
};
typedef union cvmx_pcmx_txmsk2 cvmx_pcmx_txmsk2_t;
/**
* cvmx_pcm#_txmsk3
*/
union cvmx_pcmx_txmsk3
{
uint64_t u64;
struct cvmx_pcmx_txmsk3_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t mask : 64; /**< Transmit mask bits for slots 255 to 192
(1 means transmit, 0 means don't transmit) */
#else
uint64_t mask : 64;
#endif
} s;
struct cvmx_pcmx_txmsk3_s cn30xx;
struct cvmx_pcmx_txmsk3_s cn31xx;
struct cvmx_pcmx_txmsk3_s cn50xx;
};
typedef union cvmx_pcmx_txmsk3 cvmx_pcmx_txmsk3_t;
/**
* cvmx_pcm#_txmsk4
*/
union cvmx_pcmx_txmsk4
{
uint64_t u64;
struct cvmx_pcmx_txmsk4_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t mask : 64; /**< Transmit mask bits for slots 319 to 256
(1 means transmit, 0 means don't transmit) */
#else
uint64_t mask : 64;
#endif
} s;
struct cvmx_pcmx_txmsk4_s cn30xx;
struct cvmx_pcmx_txmsk4_s cn31xx;
struct cvmx_pcmx_txmsk4_s cn50xx;
};
typedef union cvmx_pcmx_txmsk4 cvmx_pcmx_txmsk4_t;
/**
* cvmx_pcm#_txmsk5
*/
union cvmx_pcmx_txmsk5
{
uint64_t u64;
struct cvmx_pcmx_txmsk5_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t mask : 64; /**< Transmit mask bits for slots 383 to 320
(1 means transmit, 0 means don't transmit) */
#else
uint64_t mask : 64;
#endif
} s;
struct cvmx_pcmx_txmsk5_s cn30xx;
struct cvmx_pcmx_txmsk5_s cn31xx;
struct cvmx_pcmx_txmsk5_s cn50xx;
};
typedef union cvmx_pcmx_txmsk5 cvmx_pcmx_txmsk5_t;
/**
* cvmx_pcm#_txmsk6
*/
union cvmx_pcmx_txmsk6
{
uint64_t u64;
struct cvmx_pcmx_txmsk6_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t mask : 64; /**< Transmit mask bits for slots 447 to 384
(1 means transmit, 0 means don't transmit) */
#else
uint64_t mask : 64;
#endif
} s;
struct cvmx_pcmx_txmsk6_s cn30xx;
struct cvmx_pcmx_txmsk6_s cn31xx;
struct cvmx_pcmx_txmsk6_s cn50xx;
};
typedef union cvmx_pcmx_txmsk6 cvmx_pcmx_txmsk6_t;
/**
* cvmx_pcm#_txmsk7
*/
union cvmx_pcmx_txmsk7
{
uint64_t u64;
struct cvmx_pcmx_txmsk7_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t mask : 64; /**< Transmit mask bits for slots 511 to 448
(1 means transmit, 0 means don't transmit) */
#else
uint64_t mask : 64;
#endif
} s;
struct cvmx_pcmx_txmsk7_s cn30xx;
struct cvmx_pcmx_txmsk7_s cn31xx;
struct cvmx_pcmx_txmsk7_s cn50xx;
};
typedef union cvmx_pcmx_txmsk7 cvmx_pcmx_txmsk7_t;
/**
* cvmx_pcm#_txstart
*/
union cvmx_pcmx_txstart
{
uint64_t u64;
struct cvmx_pcmx_txstart_s
{
#if __BYTE_ORDER == __BIG_ENDIAN
uint64_t reserved_36_63 : 28;
uint64_t addr : 33; /**< Starting address for the transmit memory region */
uint64_t reserved_0_2 : 3;
#else
uint64_t reserved_0_2 : 3;
uint64_t addr : 33;
uint64_t reserved_36_63 : 28;
#endif
} s;
struct cvmx_pcmx_txstart_s cn30xx;
struct cvmx_pcmx_txstart_s cn31xx;
struct cvmx_pcmx_txstart_s cn50xx;
};
typedef union cvmx_pcmx_txstart cvmx_pcmx_txstart_t;
#endif