diff options
Diffstat (limited to 'sys/contrib/octeon-sdk/cvmx-npei-defs.h')
-rw-r--r-- | sys/contrib/octeon-sdk/cvmx-npei-defs.h | 7676 |
1 files changed, 7676 insertions, 0 deletions
diff --git a/sys/contrib/octeon-sdk/cvmx-npei-defs.h b/sys/contrib/octeon-sdk/cvmx-npei-defs.h new file mode 100644 index 0000000..72155f8 --- /dev/null +++ b/sys/contrib/octeon-sdk/cvmx-npei-defs.h @@ -0,0 +1,7676 @@ +/***********************license start*************** + * Copyright (c) 2003-2010 Cavium Networks (support@cavium.com). All rights + * reserved. + * + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions are + * met: + * + * * Redistributions of source code must retain the above copyright + * notice, this list of conditions and the following disclaimer. + * + * * Redistributions in binary form must reproduce the above + * copyright notice, this list of conditions and the following + * disclaimer in the documentation and/or other materials provided + * with the distribution. + + * * Neither the name of Cavium Networks nor the names of + * its contributors may be used to endorse or promote products + * derived from this software without specific prior written + * permission. + + * This Software, including technical data, may be subject to U.S. export control + * laws, including the U.S. Export Administration Act and its associated + * regulations, and may be subject to export or import regulations in other + * countries. + + * TO THE MAXIMUM EXTENT PERMITTED BY LAW, THE SOFTWARE IS PROVIDED "AS IS" + * AND WITH ALL FAULTS AND CAVIUM NETWORKS MAKES NO PROMISES, REPRESENTATIONS OR + * WARRANTIES, EITHER EXPRESS, IMPLIED, STATUTORY, OR OTHERWISE, WITH RESPECT TO + * THE SOFTWARE, INCLUDING ITS CONDITION, ITS CONFORMITY TO ANY REPRESENTATION OR + * DESCRIPTION, OR THE EXISTENCE OF ANY LATENT OR PATENT DEFECTS, AND CAVIUM + * SPECIFICALLY DISCLAIMS ALL IMPLIED (IF ANY) WARRANTIES OF TITLE, + * MERCHANTABILITY, NONINFRINGEMENT, FITNESS FOR A PARTICULAR PURPOSE, LACK OF + * VIRUSES, ACCURACY OR COMPLETENESS, QUIET ENJOYMENT, QUIET POSSESSION OR + * CORRESPONDENCE TO DESCRIPTION. THE ENTIRE RISK ARISING OUT OF USE OR + * PERFORMANCE OF THE SOFTWARE LIES WITH YOU. + ***********************license end**************************************/ + + +/** + * cvmx-npei-defs.h + * + * Configuration and status register (CSR) type definitions for + * Octeon npei. + * + * This file is auto generated. Do not edit. + * + * <hr>$Revision$<hr> + * + */ +#ifndef __CVMX_NPEI_TYPEDEFS_H__ +#define __CVMX_NPEI_TYPEDEFS_H__ + +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +static inline uint64_t CVMX_NPEI_BAR1_INDEXX(unsigned long offset) +{ + if (!( + (OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 31))) || + (OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 31))))) + cvmx_warn("CVMX_NPEI_BAR1_INDEXX(%lu) is invalid on this chip\n", offset); + return 0x0000000000000000ull + ((offset) & 31) * 16; +} +#else +#define CVMX_NPEI_BAR1_INDEXX(offset) (0x0000000000000000ull + ((offset) & 31) * 16) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_BIST_STATUS CVMX_NPEI_BIST_STATUS_FUNC() +static inline uint64_t CVMX_NPEI_BIST_STATUS_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_BIST_STATUS not supported on this chip\n"); + return 0x0000000000000580ull; +} +#else +#define CVMX_NPEI_BIST_STATUS (0x0000000000000580ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_BIST_STATUS2 CVMX_NPEI_BIST_STATUS2_FUNC() +static inline uint64_t CVMX_NPEI_BIST_STATUS2_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_BIST_STATUS2 not supported on this chip\n"); + return 0x0000000000000680ull; +} +#else +#define CVMX_NPEI_BIST_STATUS2 (0x0000000000000680ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_CTL_PORT0 CVMX_NPEI_CTL_PORT0_FUNC() +static inline uint64_t CVMX_NPEI_CTL_PORT0_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_CTL_PORT0 not supported on this chip\n"); + return 0x0000000000000250ull; +} +#else +#define CVMX_NPEI_CTL_PORT0 (0x0000000000000250ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_CTL_PORT1 CVMX_NPEI_CTL_PORT1_FUNC() +static inline uint64_t CVMX_NPEI_CTL_PORT1_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_CTL_PORT1 not supported on this chip\n"); + return 0x0000000000000260ull; +} +#else +#define CVMX_NPEI_CTL_PORT1 (0x0000000000000260ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_CTL_STATUS CVMX_NPEI_CTL_STATUS_FUNC() +static inline uint64_t CVMX_NPEI_CTL_STATUS_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_CTL_STATUS not supported on this chip\n"); + return 0x0000000000000570ull; +} +#else +#define CVMX_NPEI_CTL_STATUS (0x0000000000000570ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_CTL_STATUS2 CVMX_NPEI_CTL_STATUS2_FUNC() +static inline uint64_t CVMX_NPEI_CTL_STATUS2_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_CTL_STATUS2 not supported on this chip\n"); + return 0x0000000000003C00ull; +} +#else +#define CVMX_NPEI_CTL_STATUS2 (0x0000000000003C00ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_DATA_OUT_CNT CVMX_NPEI_DATA_OUT_CNT_FUNC() +static inline uint64_t CVMX_NPEI_DATA_OUT_CNT_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_DATA_OUT_CNT not supported on this chip\n"); + return 0x00000000000005F0ull; +} +#else +#define CVMX_NPEI_DATA_OUT_CNT (0x00000000000005F0ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_DBG_DATA CVMX_NPEI_DBG_DATA_FUNC() +static inline uint64_t CVMX_NPEI_DBG_DATA_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_DBG_DATA not supported on this chip\n"); + return 0x0000000000000510ull; +} +#else +#define CVMX_NPEI_DBG_DATA (0x0000000000000510ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_DBG_SELECT CVMX_NPEI_DBG_SELECT_FUNC() +static inline uint64_t CVMX_NPEI_DBG_SELECT_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_DBG_SELECT not supported on this chip\n"); + return 0x0000000000000500ull; +} +#else +#define CVMX_NPEI_DBG_SELECT (0x0000000000000500ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_DMA0_INT_LEVEL CVMX_NPEI_DMA0_INT_LEVEL_FUNC() +static inline uint64_t CVMX_NPEI_DMA0_INT_LEVEL_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_DMA0_INT_LEVEL not supported on this chip\n"); + return 0x00000000000005C0ull; +} +#else +#define CVMX_NPEI_DMA0_INT_LEVEL (0x00000000000005C0ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_DMA1_INT_LEVEL CVMX_NPEI_DMA1_INT_LEVEL_FUNC() +static inline uint64_t CVMX_NPEI_DMA1_INT_LEVEL_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_DMA1_INT_LEVEL not supported on this chip\n"); + return 0x00000000000005D0ull; +} +#else +#define CVMX_NPEI_DMA1_INT_LEVEL (0x00000000000005D0ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +static inline uint64_t CVMX_NPEI_DMAX_COUNTS(unsigned long offset) +{ + if (!( + (OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 4))) || + (OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 4))))) + cvmx_warn("CVMX_NPEI_DMAX_COUNTS(%lu) is invalid on this chip\n", offset); + return 0x0000000000000450ull + ((offset) & 7) * 16; +} +#else +#define CVMX_NPEI_DMAX_COUNTS(offset) (0x0000000000000450ull + ((offset) & 7) * 16) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +static inline uint64_t CVMX_NPEI_DMAX_DBELL(unsigned long offset) +{ + if (!( + (OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 4))) || + (OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 4))))) + cvmx_warn("CVMX_NPEI_DMAX_DBELL(%lu) is invalid on this chip\n", offset); + return 0x00000000000003B0ull + ((offset) & 7) * 16; +} +#else +#define CVMX_NPEI_DMAX_DBELL(offset) (0x00000000000003B0ull + ((offset) & 7) * 16) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +static inline uint64_t CVMX_NPEI_DMAX_IBUFF_SADDR(unsigned long offset) +{ + if (!( + (OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 4))) || + (OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 4))))) + cvmx_warn("CVMX_NPEI_DMAX_IBUFF_SADDR(%lu) is invalid on this chip\n", offset); + return 0x0000000000000400ull + ((offset) & 7) * 16; +} +#else +#define CVMX_NPEI_DMAX_IBUFF_SADDR(offset) (0x0000000000000400ull + ((offset) & 7) * 16) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +static inline uint64_t CVMX_NPEI_DMAX_NADDR(unsigned long offset) +{ + if (!( + (OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 4))) || + (OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 4))))) + cvmx_warn("CVMX_NPEI_DMAX_NADDR(%lu) is invalid on this chip\n", offset); + return 0x00000000000004A0ull + ((offset) & 7) * 16; +} +#else +#define CVMX_NPEI_DMAX_NADDR(offset) (0x00000000000004A0ull + ((offset) & 7) * 16) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_DMA_CNTS CVMX_NPEI_DMA_CNTS_FUNC() +static inline uint64_t CVMX_NPEI_DMA_CNTS_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_DMA_CNTS not supported on this chip\n"); + return 0x00000000000005E0ull; +} +#else +#define CVMX_NPEI_DMA_CNTS (0x00000000000005E0ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_DMA_CONTROL CVMX_NPEI_DMA_CONTROL_FUNC() +static inline uint64_t CVMX_NPEI_DMA_CONTROL_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_DMA_CONTROL not supported on this chip\n"); + return 0x00000000000003A0ull; +} +#else +#define CVMX_NPEI_DMA_CONTROL (0x00000000000003A0ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_DMA_PCIE_REQ_NUM CVMX_NPEI_DMA_PCIE_REQ_NUM_FUNC() +static inline uint64_t CVMX_NPEI_DMA_PCIE_REQ_NUM_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_DMA_PCIE_REQ_NUM not supported on this chip\n"); + return 0x00000000000005B0ull; +} +#else +#define CVMX_NPEI_DMA_PCIE_REQ_NUM (0x00000000000005B0ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_DMA_STATE1 CVMX_NPEI_DMA_STATE1_FUNC() +static inline uint64_t CVMX_NPEI_DMA_STATE1_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX))) + cvmx_warn("CVMX_NPEI_DMA_STATE1 not supported on this chip\n"); + return 0x00000000000006C0ull; +} +#else +#define CVMX_NPEI_DMA_STATE1 (0x00000000000006C0ull) +#endif +#define CVMX_NPEI_DMA_STATE1_P1 (0x0000000000000680ull) +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_DMA_STATE2 CVMX_NPEI_DMA_STATE2_FUNC() +static inline uint64_t CVMX_NPEI_DMA_STATE2_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX))) + cvmx_warn("CVMX_NPEI_DMA_STATE2 not supported on this chip\n"); + return 0x00000000000006D0ull; +} +#else +#define CVMX_NPEI_DMA_STATE2 (0x00000000000006D0ull) +#endif +#define CVMX_NPEI_DMA_STATE2_P1 (0x0000000000000690ull) +#define CVMX_NPEI_DMA_STATE3_P1 (0x00000000000006A0ull) +#define CVMX_NPEI_DMA_STATE4_P1 (0x00000000000006B0ull) +#define CVMX_NPEI_DMA_STATE5_P1 (0x00000000000006C0ull) +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_INT_A_ENB CVMX_NPEI_INT_A_ENB_FUNC() +static inline uint64_t CVMX_NPEI_INT_A_ENB_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_INT_A_ENB not supported on this chip\n"); + return 0x0000000000000560ull; +} +#else +#define CVMX_NPEI_INT_A_ENB (0x0000000000000560ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_INT_A_ENB2 CVMX_NPEI_INT_A_ENB2_FUNC() +static inline uint64_t CVMX_NPEI_INT_A_ENB2_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_INT_A_ENB2 not supported on this chip\n"); + return 0x0000000000003CE0ull; +} +#else +#define CVMX_NPEI_INT_A_ENB2 (0x0000000000003CE0ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_INT_A_SUM CVMX_NPEI_INT_A_SUM_FUNC() +static inline uint64_t CVMX_NPEI_INT_A_SUM_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_INT_A_SUM not supported on this chip\n"); + return 0x0000000000000550ull; +} +#else +#define CVMX_NPEI_INT_A_SUM (0x0000000000000550ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_INT_ENB CVMX_NPEI_INT_ENB_FUNC() +static inline uint64_t CVMX_NPEI_INT_ENB_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_INT_ENB not supported on this chip\n"); + return 0x0000000000000540ull; +} +#else +#define CVMX_NPEI_INT_ENB (0x0000000000000540ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_INT_ENB2 CVMX_NPEI_INT_ENB2_FUNC() +static inline uint64_t CVMX_NPEI_INT_ENB2_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_INT_ENB2 not supported on this chip\n"); + return 0x0000000000003CD0ull; +} +#else +#define CVMX_NPEI_INT_ENB2 (0x0000000000003CD0ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_INT_INFO CVMX_NPEI_INT_INFO_FUNC() +static inline uint64_t CVMX_NPEI_INT_INFO_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_INT_INFO not supported on this chip\n"); + return 0x0000000000000590ull; +} +#else +#define CVMX_NPEI_INT_INFO (0x0000000000000590ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_INT_SUM CVMX_NPEI_INT_SUM_FUNC() +static inline uint64_t CVMX_NPEI_INT_SUM_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_INT_SUM not supported on this chip\n"); + return 0x0000000000000530ull; +} +#else +#define CVMX_NPEI_INT_SUM (0x0000000000000530ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_INT_SUM2 CVMX_NPEI_INT_SUM2_FUNC() +static inline uint64_t CVMX_NPEI_INT_SUM2_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_INT_SUM2 not supported on this chip\n"); + return 0x0000000000003CC0ull; +} +#else +#define CVMX_NPEI_INT_SUM2 (0x0000000000003CC0ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_LAST_WIN_RDATA0 CVMX_NPEI_LAST_WIN_RDATA0_FUNC() +static inline uint64_t CVMX_NPEI_LAST_WIN_RDATA0_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_LAST_WIN_RDATA0 not supported on this chip\n"); + return 0x0000000000000600ull; +} +#else +#define CVMX_NPEI_LAST_WIN_RDATA0 (0x0000000000000600ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_LAST_WIN_RDATA1 CVMX_NPEI_LAST_WIN_RDATA1_FUNC() +static inline uint64_t CVMX_NPEI_LAST_WIN_RDATA1_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_LAST_WIN_RDATA1 not supported on this chip\n"); + return 0x0000000000000610ull; +} +#else +#define CVMX_NPEI_LAST_WIN_RDATA1 (0x0000000000000610ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_MEM_ACCESS_CTL CVMX_NPEI_MEM_ACCESS_CTL_FUNC() +static inline uint64_t CVMX_NPEI_MEM_ACCESS_CTL_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_MEM_ACCESS_CTL not supported on this chip\n"); + return 0x00000000000004F0ull; +} +#else +#define CVMX_NPEI_MEM_ACCESS_CTL (0x00000000000004F0ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +static inline uint64_t CVMX_NPEI_MEM_ACCESS_SUBIDX(unsigned long offset) +{ + if (!( + (OCTEON_IS_MODEL(OCTEON_CN52XX) && (((offset >= 12) && (offset <= 27)))) || + (OCTEON_IS_MODEL(OCTEON_CN56XX) && (((offset >= 12) && (offset <= 27)))))) + cvmx_warn("CVMX_NPEI_MEM_ACCESS_SUBIDX(%lu) is invalid on this chip\n", offset); + return 0x0000000000000340ull + ((offset) & 31) * 16 - 16*12; +} +#else +#define CVMX_NPEI_MEM_ACCESS_SUBIDX(offset) (0x0000000000000340ull + ((offset) & 31) * 16 - 16*12) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_MSI_ENB0 CVMX_NPEI_MSI_ENB0_FUNC() +static inline uint64_t CVMX_NPEI_MSI_ENB0_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_MSI_ENB0 not supported on this chip\n"); + return 0x0000000000003C50ull; +} +#else +#define CVMX_NPEI_MSI_ENB0 (0x0000000000003C50ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_MSI_ENB1 CVMX_NPEI_MSI_ENB1_FUNC() +static inline uint64_t CVMX_NPEI_MSI_ENB1_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_MSI_ENB1 not supported on this chip\n"); + return 0x0000000000003C60ull; +} +#else +#define CVMX_NPEI_MSI_ENB1 (0x0000000000003C60ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_MSI_ENB2 CVMX_NPEI_MSI_ENB2_FUNC() +static inline uint64_t CVMX_NPEI_MSI_ENB2_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_MSI_ENB2 not supported on this chip\n"); + return 0x0000000000003C70ull; +} +#else +#define CVMX_NPEI_MSI_ENB2 (0x0000000000003C70ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_MSI_ENB3 CVMX_NPEI_MSI_ENB3_FUNC() +static inline uint64_t CVMX_NPEI_MSI_ENB3_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_MSI_ENB3 not supported on this chip\n"); + return 0x0000000000003C80ull; +} +#else +#define CVMX_NPEI_MSI_ENB3 (0x0000000000003C80ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_MSI_RCV0 CVMX_NPEI_MSI_RCV0_FUNC() +static inline uint64_t CVMX_NPEI_MSI_RCV0_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_MSI_RCV0 not supported on this chip\n"); + return 0x0000000000003C10ull; +} +#else +#define CVMX_NPEI_MSI_RCV0 (0x0000000000003C10ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_MSI_RCV1 CVMX_NPEI_MSI_RCV1_FUNC() +static inline uint64_t CVMX_NPEI_MSI_RCV1_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_MSI_RCV1 not supported on this chip\n"); + return 0x0000000000003C20ull; +} +#else +#define CVMX_NPEI_MSI_RCV1 (0x0000000000003C20ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_MSI_RCV2 CVMX_NPEI_MSI_RCV2_FUNC() +static inline uint64_t CVMX_NPEI_MSI_RCV2_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_MSI_RCV2 not supported on this chip\n"); + return 0x0000000000003C30ull; +} +#else +#define CVMX_NPEI_MSI_RCV2 (0x0000000000003C30ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_MSI_RCV3 CVMX_NPEI_MSI_RCV3_FUNC() +static inline uint64_t CVMX_NPEI_MSI_RCV3_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_MSI_RCV3 not supported on this chip\n"); + return 0x0000000000003C40ull; +} +#else +#define CVMX_NPEI_MSI_RCV3 (0x0000000000003C40ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_MSI_RD_MAP CVMX_NPEI_MSI_RD_MAP_FUNC() +static inline uint64_t CVMX_NPEI_MSI_RD_MAP_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_MSI_RD_MAP not supported on this chip\n"); + return 0x0000000000003CA0ull; +} +#else +#define CVMX_NPEI_MSI_RD_MAP (0x0000000000003CA0ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_MSI_W1C_ENB0 CVMX_NPEI_MSI_W1C_ENB0_FUNC() +static inline uint64_t CVMX_NPEI_MSI_W1C_ENB0_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_MSI_W1C_ENB0 not supported on this chip\n"); + return 0x0000000000003CF0ull; +} +#else +#define CVMX_NPEI_MSI_W1C_ENB0 (0x0000000000003CF0ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_MSI_W1C_ENB1 CVMX_NPEI_MSI_W1C_ENB1_FUNC() +static inline uint64_t CVMX_NPEI_MSI_W1C_ENB1_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_MSI_W1C_ENB1 not supported on this chip\n"); + return 0x0000000000003D00ull; +} +#else +#define CVMX_NPEI_MSI_W1C_ENB1 (0x0000000000003D00ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_MSI_W1C_ENB2 CVMX_NPEI_MSI_W1C_ENB2_FUNC() +static inline uint64_t CVMX_NPEI_MSI_W1C_ENB2_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_MSI_W1C_ENB2 not supported on this chip\n"); + return 0x0000000000003D10ull; +} +#else +#define CVMX_NPEI_MSI_W1C_ENB2 (0x0000000000003D10ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_MSI_W1C_ENB3 CVMX_NPEI_MSI_W1C_ENB3_FUNC() +static inline uint64_t CVMX_NPEI_MSI_W1C_ENB3_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_MSI_W1C_ENB3 not supported on this chip\n"); + return 0x0000000000003D20ull; +} +#else +#define CVMX_NPEI_MSI_W1C_ENB3 (0x0000000000003D20ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_MSI_W1S_ENB0 CVMX_NPEI_MSI_W1S_ENB0_FUNC() +static inline uint64_t CVMX_NPEI_MSI_W1S_ENB0_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_MSI_W1S_ENB0 not supported on this chip\n"); + return 0x0000000000003D30ull; +} +#else +#define CVMX_NPEI_MSI_W1S_ENB0 (0x0000000000003D30ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_MSI_W1S_ENB1 CVMX_NPEI_MSI_W1S_ENB1_FUNC() +static inline uint64_t CVMX_NPEI_MSI_W1S_ENB1_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_MSI_W1S_ENB1 not supported on this chip\n"); + return 0x0000000000003D40ull; +} +#else +#define CVMX_NPEI_MSI_W1S_ENB1 (0x0000000000003D40ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_MSI_W1S_ENB2 CVMX_NPEI_MSI_W1S_ENB2_FUNC() +static inline uint64_t CVMX_NPEI_MSI_W1S_ENB2_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_MSI_W1S_ENB2 not supported on this chip\n"); + return 0x0000000000003D50ull; +} +#else +#define CVMX_NPEI_MSI_W1S_ENB2 (0x0000000000003D50ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_MSI_W1S_ENB3 CVMX_NPEI_MSI_W1S_ENB3_FUNC() +static inline uint64_t CVMX_NPEI_MSI_W1S_ENB3_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_MSI_W1S_ENB3 not supported on this chip\n"); + return 0x0000000000003D60ull; +} +#else +#define CVMX_NPEI_MSI_W1S_ENB3 (0x0000000000003D60ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_MSI_WR_MAP CVMX_NPEI_MSI_WR_MAP_FUNC() +static inline uint64_t CVMX_NPEI_MSI_WR_MAP_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_MSI_WR_MAP not supported on this chip\n"); + return 0x0000000000003C90ull; +} +#else +#define CVMX_NPEI_MSI_WR_MAP (0x0000000000003C90ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PCIE_CREDIT_CNT CVMX_NPEI_PCIE_CREDIT_CNT_FUNC() +static inline uint64_t CVMX_NPEI_PCIE_CREDIT_CNT_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PCIE_CREDIT_CNT not supported on this chip\n"); + return 0x0000000000003D70ull; +} +#else +#define CVMX_NPEI_PCIE_CREDIT_CNT (0x0000000000003D70ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PCIE_MSI_RCV CVMX_NPEI_PCIE_MSI_RCV_FUNC() +static inline uint64_t CVMX_NPEI_PCIE_MSI_RCV_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PCIE_MSI_RCV not supported on this chip\n"); + return 0x0000000000003CB0ull; +} +#else +#define CVMX_NPEI_PCIE_MSI_RCV (0x0000000000003CB0ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PCIE_MSI_RCV_B1 CVMX_NPEI_PCIE_MSI_RCV_B1_FUNC() +static inline uint64_t CVMX_NPEI_PCIE_MSI_RCV_B1_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PCIE_MSI_RCV_B1 not supported on this chip\n"); + return 0x0000000000000650ull; +} +#else +#define CVMX_NPEI_PCIE_MSI_RCV_B1 (0x0000000000000650ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PCIE_MSI_RCV_B2 CVMX_NPEI_PCIE_MSI_RCV_B2_FUNC() +static inline uint64_t CVMX_NPEI_PCIE_MSI_RCV_B2_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PCIE_MSI_RCV_B2 not supported on this chip\n"); + return 0x0000000000000660ull; +} +#else +#define CVMX_NPEI_PCIE_MSI_RCV_B2 (0x0000000000000660ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PCIE_MSI_RCV_B3 CVMX_NPEI_PCIE_MSI_RCV_B3_FUNC() +static inline uint64_t CVMX_NPEI_PCIE_MSI_RCV_B3_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PCIE_MSI_RCV_B3 not supported on this chip\n"); + return 0x0000000000000670ull; +} +#else +#define CVMX_NPEI_PCIE_MSI_RCV_B3 (0x0000000000000670ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +static inline uint64_t CVMX_NPEI_PKTX_CNTS(unsigned long offset) +{ + if (!( + (OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 31))) || + (OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 31))))) + cvmx_warn("CVMX_NPEI_PKTX_CNTS(%lu) is invalid on this chip\n", offset); + return 0x0000000000002400ull + ((offset) & 31) * 16; +} +#else +#define CVMX_NPEI_PKTX_CNTS(offset) (0x0000000000002400ull + ((offset) & 31) * 16) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +static inline uint64_t CVMX_NPEI_PKTX_INSTR_BADDR(unsigned long offset) +{ + if (!( + (OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 31))) || + (OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 31))))) + cvmx_warn("CVMX_NPEI_PKTX_INSTR_BADDR(%lu) is invalid on this chip\n", offset); + return 0x0000000000002800ull + ((offset) & 31) * 16; +} +#else +#define CVMX_NPEI_PKTX_INSTR_BADDR(offset) (0x0000000000002800ull + ((offset) & 31) * 16) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +static inline uint64_t CVMX_NPEI_PKTX_INSTR_BAOFF_DBELL(unsigned long offset) +{ + if (!( + (OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 31))) || + (OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 31))))) + cvmx_warn("CVMX_NPEI_PKTX_INSTR_BAOFF_DBELL(%lu) is invalid on this chip\n", offset); + return 0x0000000000002C00ull + ((offset) & 31) * 16; +} +#else +#define CVMX_NPEI_PKTX_INSTR_BAOFF_DBELL(offset) (0x0000000000002C00ull + ((offset) & 31) * 16) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +static inline uint64_t CVMX_NPEI_PKTX_INSTR_FIFO_RSIZE(unsigned long offset) +{ + if (!( + (OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 31))) || + (OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 31))))) + cvmx_warn("CVMX_NPEI_PKTX_INSTR_FIFO_RSIZE(%lu) is invalid on this chip\n", offset); + return 0x0000000000003000ull + ((offset) & 31) * 16; +} +#else +#define CVMX_NPEI_PKTX_INSTR_FIFO_RSIZE(offset) (0x0000000000003000ull + ((offset) & 31) * 16) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +static inline uint64_t CVMX_NPEI_PKTX_INSTR_HEADER(unsigned long offset) +{ + if (!( + (OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 31))) || + (OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 31))))) + cvmx_warn("CVMX_NPEI_PKTX_INSTR_HEADER(%lu) is invalid on this chip\n", offset); + return 0x0000000000003400ull + ((offset) & 31) * 16; +} +#else +#define CVMX_NPEI_PKTX_INSTR_HEADER(offset) (0x0000000000003400ull + ((offset) & 31) * 16) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +static inline uint64_t CVMX_NPEI_PKTX_IN_BP(unsigned long offset) +{ + if (!( + (OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 31))) || + (OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 31))))) + cvmx_warn("CVMX_NPEI_PKTX_IN_BP(%lu) is invalid on this chip\n", offset); + return 0x0000000000003800ull + ((offset) & 31) * 16; +} +#else +#define CVMX_NPEI_PKTX_IN_BP(offset) (0x0000000000003800ull + ((offset) & 31) * 16) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +static inline uint64_t CVMX_NPEI_PKTX_SLIST_BADDR(unsigned long offset) +{ + if (!( + (OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 31))) || + (OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 31))))) + cvmx_warn("CVMX_NPEI_PKTX_SLIST_BADDR(%lu) is invalid on this chip\n", offset); + return 0x0000000000001400ull + ((offset) & 31) * 16; +} +#else +#define CVMX_NPEI_PKTX_SLIST_BADDR(offset) (0x0000000000001400ull + ((offset) & 31) * 16) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +static inline uint64_t CVMX_NPEI_PKTX_SLIST_BAOFF_DBELL(unsigned long offset) +{ + if (!( + (OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 31))) || + (OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 31))))) + cvmx_warn("CVMX_NPEI_PKTX_SLIST_BAOFF_DBELL(%lu) is invalid on this chip\n", offset); + return 0x0000000000001800ull + ((offset) & 31) * 16; +} +#else +#define CVMX_NPEI_PKTX_SLIST_BAOFF_DBELL(offset) (0x0000000000001800ull + ((offset) & 31) * 16) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +static inline uint64_t CVMX_NPEI_PKTX_SLIST_FIFO_RSIZE(unsigned long offset) +{ + if (!( + (OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 31))) || + (OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 31))))) + cvmx_warn("CVMX_NPEI_PKTX_SLIST_FIFO_RSIZE(%lu) is invalid on this chip\n", offset); + return 0x0000000000001C00ull + ((offset) & 31) * 16; +} +#else +#define CVMX_NPEI_PKTX_SLIST_FIFO_RSIZE(offset) (0x0000000000001C00ull + ((offset) & 31) * 16) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PKT_CNT_INT CVMX_NPEI_PKT_CNT_INT_FUNC() +static inline uint64_t CVMX_NPEI_PKT_CNT_INT_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PKT_CNT_INT not supported on this chip\n"); + return 0x0000000000001110ull; +} +#else +#define CVMX_NPEI_PKT_CNT_INT (0x0000000000001110ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PKT_CNT_INT_ENB CVMX_NPEI_PKT_CNT_INT_ENB_FUNC() +static inline uint64_t CVMX_NPEI_PKT_CNT_INT_ENB_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PKT_CNT_INT_ENB not supported on this chip\n"); + return 0x0000000000001130ull; +} +#else +#define CVMX_NPEI_PKT_CNT_INT_ENB (0x0000000000001130ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PKT_DATA_OUT_ES CVMX_NPEI_PKT_DATA_OUT_ES_FUNC() +static inline uint64_t CVMX_NPEI_PKT_DATA_OUT_ES_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PKT_DATA_OUT_ES not supported on this chip\n"); + return 0x00000000000010B0ull; +} +#else +#define CVMX_NPEI_PKT_DATA_OUT_ES (0x00000000000010B0ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PKT_DATA_OUT_NS CVMX_NPEI_PKT_DATA_OUT_NS_FUNC() +static inline uint64_t CVMX_NPEI_PKT_DATA_OUT_NS_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PKT_DATA_OUT_NS not supported on this chip\n"); + return 0x00000000000010A0ull; +} +#else +#define CVMX_NPEI_PKT_DATA_OUT_NS (0x00000000000010A0ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PKT_DATA_OUT_ROR CVMX_NPEI_PKT_DATA_OUT_ROR_FUNC() +static inline uint64_t CVMX_NPEI_PKT_DATA_OUT_ROR_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PKT_DATA_OUT_ROR not supported on this chip\n"); + return 0x0000000000001090ull; +} +#else +#define CVMX_NPEI_PKT_DATA_OUT_ROR (0x0000000000001090ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PKT_DPADDR CVMX_NPEI_PKT_DPADDR_FUNC() +static inline uint64_t CVMX_NPEI_PKT_DPADDR_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PKT_DPADDR not supported on this chip\n"); + return 0x0000000000001080ull; +} +#else +#define CVMX_NPEI_PKT_DPADDR (0x0000000000001080ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PKT_INPUT_CONTROL CVMX_NPEI_PKT_INPUT_CONTROL_FUNC() +static inline uint64_t CVMX_NPEI_PKT_INPUT_CONTROL_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PKT_INPUT_CONTROL not supported on this chip\n"); + return 0x0000000000001150ull; +} +#else +#define CVMX_NPEI_PKT_INPUT_CONTROL (0x0000000000001150ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PKT_INSTR_ENB CVMX_NPEI_PKT_INSTR_ENB_FUNC() +static inline uint64_t CVMX_NPEI_PKT_INSTR_ENB_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PKT_INSTR_ENB not supported on this chip\n"); + return 0x0000000000001000ull; +} +#else +#define CVMX_NPEI_PKT_INSTR_ENB (0x0000000000001000ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PKT_INSTR_RD_SIZE CVMX_NPEI_PKT_INSTR_RD_SIZE_FUNC() +static inline uint64_t CVMX_NPEI_PKT_INSTR_RD_SIZE_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PKT_INSTR_RD_SIZE not supported on this chip\n"); + return 0x0000000000001190ull; +} +#else +#define CVMX_NPEI_PKT_INSTR_RD_SIZE (0x0000000000001190ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PKT_INSTR_SIZE CVMX_NPEI_PKT_INSTR_SIZE_FUNC() +static inline uint64_t CVMX_NPEI_PKT_INSTR_SIZE_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PKT_INSTR_SIZE not supported on this chip\n"); + return 0x0000000000001020ull; +} +#else +#define CVMX_NPEI_PKT_INSTR_SIZE (0x0000000000001020ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PKT_INT_LEVELS CVMX_NPEI_PKT_INT_LEVELS_FUNC() +static inline uint64_t CVMX_NPEI_PKT_INT_LEVELS_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PKT_INT_LEVELS not supported on this chip\n"); + return 0x0000000000001100ull; +} +#else +#define CVMX_NPEI_PKT_INT_LEVELS (0x0000000000001100ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PKT_IN_BP CVMX_NPEI_PKT_IN_BP_FUNC() +static inline uint64_t CVMX_NPEI_PKT_IN_BP_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PKT_IN_BP not supported on this chip\n"); + return 0x00000000000006B0ull; +} +#else +#define CVMX_NPEI_PKT_IN_BP (0x00000000000006B0ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +static inline uint64_t CVMX_NPEI_PKT_IN_DONEX_CNTS(unsigned long offset) +{ + if (!( + (OCTEON_IS_MODEL(OCTEON_CN52XX) && ((offset <= 31))) || + (OCTEON_IS_MODEL(OCTEON_CN56XX) && ((offset <= 31))))) + cvmx_warn("CVMX_NPEI_PKT_IN_DONEX_CNTS(%lu) is invalid on this chip\n", offset); + return 0x0000000000002000ull + ((offset) & 31) * 16; +} +#else +#define CVMX_NPEI_PKT_IN_DONEX_CNTS(offset) (0x0000000000002000ull + ((offset) & 31) * 16) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PKT_IN_INSTR_COUNTS CVMX_NPEI_PKT_IN_INSTR_COUNTS_FUNC() +static inline uint64_t CVMX_NPEI_PKT_IN_INSTR_COUNTS_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PKT_IN_INSTR_COUNTS not supported on this chip\n"); + return 0x00000000000006A0ull; +} +#else +#define CVMX_NPEI_PKT_IN_INSTR_COUNTS (0x00000000000006A0ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PKT_IN_PCIE_PORT CVMX_NPEI_PKT_IN_PCIE_PORT_FUNC() +static inline uint64_t CVMX_NPEI_PKT_IN_PCIE_PORT_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PKT_IN_PCIE_PORT not supported on this chip\n"); + return 0x00000000000011A0ull; +} +#else +#define CVMX_NPEI_PKT_IN_PCIE_PORT (0x00000000000011A0ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PKT_IPTR CVMX_NPEI_PKT_IPTR_FUNC() +static inline uint64_t CVMX_NPEI_PKT_IPTR_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PKT_IPTR not supported on this chip\n"); + return 0x0000000000001070ull; +} +#else +#define CVMX_NPEI_PKT_IPTR (0x0000000000001070ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PKT_OUTPUT_WMARK CVMX_NPEI_PKT_OUTPUT_WMARK_FUNC() +static inline uint64_t CVMX_NPEI_PKT_OUTPUT_WMARK_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PKT_OUTPUT_WMARK not supported on this chip\n"); + return 0x0000000000001160ull; +} +#else +#define CVMX_NPEI_PKT_OUTPUT_WMARK (0x0000000000001160ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PKT_OUT_BMODE CVMX_NPEI_PKT_OUT_BMODE_FUNC() +static inline uint64_t CVMX_NPEI_PKT_OUT_BMODE_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PKT_OUT_BMODE not supported on this chip\n"); + return 0x00000000000010D0ull; +} +#else +#define CVMX_NPEI_PKT_OUT_BMODE (0x00000000000010D0ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PKT_OUT_ENB CVMX_NPEI_PKT_OUT_ENB_FUNC() +static inline uint64_t CVMX_NPEI_PKT_OUT_ENB_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PKT_OUT_ENB not supported on this chip\n"); + return 0x0000000000001010ull; +} +#else +#define CVMX_NPEI_PKT_OUT_ENB (0x0000000000001010ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PKT_PCIE_PORT CVMX_NPEI_PKT_PCIE_PORT_FUNC() +static inline uint64_t CVMX_NPEI_PKT_PCIE_PORT_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PKT_PCIE_PORT not supported on this chip\n"); + return 0x00000000000010E0ull; +} +#else +#define CVMX_NPEI_PKT_PCIE_PORT (0x00000000000010E0ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PKT_PORT_IN_RST CVMX_NPEI_PKT_PORT_IN_RST_FUNC() +static inline uint64_t CVMX_NPEI_PKT_PORT_IN_RST_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PKT_PORT_IN_RST not supported on this chip\n"); + return 0x0000000000000690ull; +} +#else +#define CVMX_NPEI_PKT_PORT_IN_RST (0x0000000000000690ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PKT_SLIST_ES CVMX_NPEI_PKT_SLIST_ES_FUNC() +static inline uint64_t CVMX_NPEI_PKT_SLIST_ES_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PKT_SLIST_ES not supported on this chip\n"); + return 0x0000000000001050ull; +} +#else +#define CVMX_NPEI_PKT_SLIST_ES (0x0000000000001050ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PKT_SLIST_ID_SIZE CVMX_NPEI_PKT_SLIST_ID_SIZE_FUNC() +static inline uint64_t CVMX_NPEI_PKT_SLIST_ID_SIZE_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PKT_SLIST_ID_SIZE not supported on this chip\n"); + return 0x0000000000001180ull; +} +#else +#define CVMX_NPEI_PKT_SLIST_ID_SIZE (0x0000000000001180ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PKT_SLIST_NS CVMX_NPEI_PKT_SLIST_NS_FUNC() +static inline uint64_t CVMX_NPEI_PKT_SLIST_NS_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PKT_SLIST_NS not supported on this chip\n"); + return 0x0000000000001040ull; +} +#else +#define CVMX_NPEI_PKT_SLIST_NS (0x0000000000001040ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PKT_SLIST_ROR CVMX_NPEI_PKT_SLIST_ROR_FUNC() +static inline uint64_t CVMX_NPEI_PKT_SLIST_ROR_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PKT_SLIST_ROR not supported on this chip\n"); + return 0x0000000000001030ull; +} +#else +#define CVMX_NPEI_PKT_SLIST_ROR (0x0000000000001030ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PKT_TIME_INT CVMX_NPEI_PKT_TIME_INT_FUNC() +static inline uint64_t CVMX_NPEI_PKT_TIME_INT_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PKT_TIME_INT not supported on this chip\n"); + return 0x0000000000001120ull; +} +#else +#define CVMX_NPEI_PKT_TIME_INT (0x0000000000001120ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_PKT_TIME_INT_ENB CVMX_NPEI_PKT_TIME_INT_ENB_FUNC() +static inline uint64_t CVMX_NPEI_PKT_TIME_INT_ENB_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_PKT_TIME_INT_ENB not supported on this chip\n"); + return 0x0000000000001140ull; +} +#else +#define CVMX_NPEI_PKT_TIME_INT_ENB (0x0000000000001140ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_RSL_INT_BLOCKS CVMX_NPEI_RSL_INT_BLOCKS_FUNC() +static inline uint64_t CVMX_NPEI_RSL_INT_BLOCKS_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_RSL_INT_BLOCKS not supported on this chip\n"); + return 0x0000000000000520ull; +} +#else +#define CVMX_NPEI_RSL_INT_BLOCKS (0x0000000000000520ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_SCRATCH_1 CVMX_NPEI_SCRATCH_1_FUNC() +static inline uint64_t CVMX_NPEI_SCRATCH_1_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_SCRATCH_1 not supported on this chip\n"); + return 0x0000000000000270ull; +} +#else +#define CVMX_NPEI_SCRATCH_1 (0x0000000000000270ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_STATE1 CVMX_NPEI_STATE1_FUNC() +static inline uint64_t CVMX_NPEI_STATE1_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_STATE1 not supported on this chip\n"); + return 0x0000000000000620ull; +} +#else +#define CVMX_NPEI_STATE1 (0x0000000000000620ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_STATE2 CVMX_NPEI_STATE2_FUNC() +static inline uint64_t CVMX_NPEI_STATE2_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_STATE2 not supported on this chip\n"); + return 0x0000000000000630ull; +} +#else +#define CVMX_NPEI_STATE2 (0x0000000000000630ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_STATE3 CVMX_NPEI_STATE3_FUNC() +static inline uint64_t CVMX_NPEI_STATE3_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_STATE3 not supported on this chip\n"); + return 0x0000000000000640ull; +} +#else +#define CVMX_NPEI_STATE3 (0x0000000000000640ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_WINDOW_CTL CVMX_NPEI_WINDOW_CTL_FUNC() +static inline uint64_t CVMX_NPEI_WINDOW_CTL_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_WINDOW_CTL not supported on this chip\n"); + return 0x0000000000000380ull; +} +#else +#define CVMX_NPEI_WINDOW_CTL (0x0000000000000380ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_WIN_RD_ADDR CVMX_NPEI_WIN_RD_ADDR_FUNC() +static inline uint64_t CVMX_NPEI_WIN_RD_ADDR_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_WIN_RD_ADDR not supported on this chip\n"); + return 0x0000000000000210ull; +} +#else +#define CVMX_NPEI_WIN_RD_ADDR (0x0000000000000210ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_WIN_RD_DATA CVMX_NPEI_WIN_RD_DATA_FUNC() +static inline uint64_t CVMX_NPEI_WIN_RD_DATA_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_WIN_RD_DATA not supported on this chip\n"); + return 0x0000000000000240ull; +} +#else +#define CVMX_NPEI_WIN_RD_DATA (0x0000000000000240ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_WIN_WR_ADDR CVMX_NPEI_WIN_WR_ADDR_FUNC() +static inline uint64_t CVMX_NPEI_WIN_WR_ADDR_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_WIN_WR_ADDR not supported on this chip\n"); + return 0x0000000000000200ull; +} +#else +#define CVMX_NPEI_WIN_WR_ADDR (0x0000000000000200ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_WIN_WR_DATA CVMX_NPEI_WIN_WR_DATA_FUNC() +static inline uint64_t CVMX_NPEI_WIN_WR_DATA_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_WIN_WR_DATA not supported on this chip\n"); + return 0x0000000000000220ull; +} +#else +#define CVMX_NPEI_WIN_WR_DATA (0x0000000000000220ull) +#endif +#if CVMX_ENABLE_CSR_ADDRESS_CHECKING +#define CVMX_NPEI_WIN_WR_MASK CVMX_NPEI_WIN_WR_MASK_FUNC() +static inline uint64_t CVMX_NPEI_WIN_WR_MASK_FUNC(void) +{ + if (!(OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX))) + cvmx_warn("CVMX_NPEI_WIN_WR_MASK not supported on this chip\n"); + return 0x0000000000000230ull; +} +#else +#define CVMX_NPEI_WIN_WR_MASK (0x0000000000000230ull) +#endif + +/** + * cvmx_npei_bar1_index# + * + * Total Address is 16Kb; 0x0000 - 0x3fff, 0x000 - 0x7fe(Reg, every other 8B) + * + * General 5kb; 0x0000 - 0x13ff, 0x000 - 0x27e(Reg-General) + * PktMem 10Kb; 0x1400 - 0x3bff, 0x280 - 0x77e(Reg-General-Packet) + * Rsvd 1Kb; 0x3c00 - 0x3fff, 0x780 - 0x7fe(Reg-NCB Only Mode) + * == NPEI_PKT_CNT_INT_ENB[PORT] + * == NPEI_PKT_TIME_INT_ENB[PORT] + * == NPEI_PKT_CNT_INT[PORT] + * == NPEI_PKT_TIME_INT[PORT] + * == NPEI_PKT_PCIE_PORT[PP] + * == NPEI_PKT_SLIST_ROR[ROR] + * == NPEI_PKT_SLIST_ROR[NSR] ? + * == NPEI_PKT_SLIST_ES[ES] + * == NPEI_PKTn_SLIST_BAOFF_DBELL[AOFF] + * == NPEI_PKTn_SLIST_BAOFF_DBELL[DBELL] + * == NPEI_PKTn_CNTS[CNT] + * NPEI_CTL_STATUS[OUTn_ENB] == NPEI_PKT_OUT_ENB[ENB] + * NPEI_BASE_ADDRESS_OUTPUTn[BADDR] == NPEI_PKTn_SLIST_BADDR[ADDR] + * NPEI_DESC_OUTPUTn[SIZE] == NPEI_PKTn_SLIST_FIFO_RSIZE[RSIZE] + * NPEI_Pn_DBPAIR_ADDR[NADDR] == NPEI_PKTn_SLIST_BADDR[ADDR] + NPEI_PKTn_SLIST_BAOFF_DBELL[AOFF] + * NPEI_PKT_CREDITSn[PTR_CNT] == NPEI_PKTn_SLIST_BAOFF_DBELL[DBELL] + * NPEI_P0_PAIR_CNTS[AVAIL] == NPEI_PKTn_SLIST_BAOFF_DBELL[DBELL] + * NPEI_P0_PAIR_CNTS[FCNT] == + * NPEI_PKTS_SENTn[PKT_CNT] == NPEI_PKTn_CNTS[CNT] + * NPEI_OUTPUT_CONTROL[Pn_BMODE] == NPEI_PKT_OUT_BMODE[BMODE] + * NPEI_PKT_CREDITSn[PKT_CNT] == NPEI_PKTn_CNTS[CNT] + * NPEI_BUFF_SIZE_OUTPUTn[BSIZE] == NPEI_PKT_SLIST_ID_SIZE[BSIZE] + * NPEI_BUFF_SIZE_OUTPUTn[ISIZE] == NPEI_PKT_SLIST_ID_SIZE[ISIZE] + * NPEI_OUTPUT_CONTROL[On_CSRM] == NPEI_PKT_DPADDR[DPTR] & NPEI_PKT_OUT_USE_IPTR[PORT] + * NPEI_OUTPUT_CONTROL[On_ES] == NPEI_PKT_DATA_OUT_ES[ES] + * NPEI_OUTPUT_CONTROL[On_NS] == NPEI_PKT_DATA_OUT_NS[NSR] ? + * NPEI_OUTPUT_CONTROL[On_RO] == NPEI_PKT_DATA_OUT_ROR[ROR] + * NPEI_PKTS_SENT_INT_LEVn[PKT_CNT] == NPEI_PKT_INT_LEVELS[CNT] + * NPEI_PKTS_SENT_TIMEn[PKT_TIME] == NPEI_PKT_INT_LEVELS[TIME] + * NPEI_OUTPUT_CONTROL[IPTR_On] == NPEI_PKT_IPTR[IPTR] + * NPEI_PCIE_PORT_OUTPUT[] == NPEI_PKT_PCIE_PORT[PP] + * + * NPEI_BAR1_INDEXX = NPEI BAR1 IndexX Register + * + * Contains address index and control bits for access to memory ranges of BAR-1. Index is build from supplied address [25:22]. + * NPEI_BAR1_INDEX0 through NPEI_BAR1_INDEX15 is used for transactions orginating with PCIE-PORT0 and NPEI_BAR1_INDEX16 + * through NPEI_BAR1_INDEX31 is used for transactions originating with PCIE-PORT1. + */ +union cvmx_npei_bar1_indexx +{ + uint32_t u32; + struct cvmx_npei_bar1_indexx_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint32_t reserved_18_31 : 14; + uint32_t addr_idx : 14; /**< Address bits [35:22] sent to L2C */ + uint32_t ca : 1; /**< Set '1' when access is not to be cached in L2. */ + uint32_t end_swp : 2; /**< Endian Swap Mode */ + uint32_t addr_v : 1; /**< Set '1' when the selected address range is valid. */ +#else + uint32_t addr_v : 1; + uint32_t end_swp : 2; + uint32_t ca : 1; + uint32_t addr_idx : 14; + uint32_t reserved_18_31 : 14; +#endif + } s; + struct cvmx_npei_bar1_indexx_s cn52xx; + struct cvmx_npei_bar1_indexx_s cn52xxp1; + struct cvmx_npei_bar1_indexx_s cn56xx; + struct cvmx_npei_bar1_indexx_s cn56xxp1; +}; +typedef union cvmx_npei_bar1_indexx cvmx_npei_bar1_indexx_t; + +/** + * cvmx_npei_bist_status + * + * NPEI_BIST_STATUS = NPI's BIST Status Register + * + * Results from BIST runs of NPEI's memories. + */ +union cvmx_npei_bist_status +{ + uint64_t u64; + struct cvmx_npei_bist_status_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t pkt_rdf : 1; /**< BIST Status for PKT Read FIFO */ + uint64_t reserved_60_62 : 3; + uint64_t pcr_gim : 1; /**< BIST Status for PKT Gather Instr MEM */ + uint64_t pkt_pif : 1; /**< BIST Status for PKT INB FIFO */ + uint64_t pcsr_int : 1; /**< BIST Status for PKT pout_int_bstatus */ + uint64_t pcsr_im : 1; /**< BIST Status for PKT pcsr_instr_mem_bstatus */ + uint64_t pcsr_cnt : 1; /**< BIST Status for PKT pin_cnt_bstatus */ + uint64_t pcsr_id : 1; /**< BIST Status for PKT pcsr_in_done_bstatus */ + uint64_t pcsr_sl : 1; /**< BIST Status for PKT pcsr_slist_bstatus */ + uint64_t reserved_50_52 : 3; + uint64_t pkt_ind : 1; /**< BIST Status for PKT Instruction Done MEM */ + uint64_t pkt_slm : 1; /**< BIST Status for PKT SList MEM */ + uint64_t reserved_36_47 : 12; + uint64_t d0_pst : 1; /**< BIST Status for DMA0 Pcie Store */ + uint64_t d1_pst : 1; /**< BIST Status for DMA1 Pcie Store */ + uint64_t d2_pst : 1; /**< BIST Status for DMA2 Pcie Store */ + uint64_t d3_pst : 1; /**< BIST Status for DMA3 Pcie Store */ + uint64_t reserved_31_31 : 1; + uint64_t n2p0_c : 1; /**< BIST Status for N2P Port0 Cmd */ + uint64_t n2p0_o : 1; /**< BIST Status for N2P Port0 Data */ + uint64_t n2p1_c : 1; /**< BIST Status for N2P Port1 Cmd */ + uint64_t n2p1_o : 1; /**< BIST Status for N2P Port1 Data */ + uint64_t cpl_p0 : 1; /**< BIST Status for CPL Port 0 */ + uint64_t cpl_p1 : 1; /**< BIST Status for CPL Port 1 */ + uint64_t p2n1_po : 1; /**< BIST Status for P2N Port1 P Order */ + uint64_t p2n1_no : 1; /**< BIST Status for P2N Port1 N Order */ + uint64_t p2n1_co : 1; /**< BIST Status for P2N Port1 C Order */ + uint64_t p2n0_po : 1; /**< BIST Status for P2N Port0 P Order */ + uint64_t p2n0_no : 1; /**< BIST Status for P2N Port0 N Order */ + uint64_t p2n0_co : 1; /**< BIST Status for P2N Port0 C Order */ + uint64_t p2n0_c0 : 1; /**< BIST Status for P2N Port0 C0 */ + uint64_t p2n0_c1 : 1; /**< BIST Status for P2N Port0 C1 */ + uint64_t p2n0_n : 1; /**< BIST Status for P2N Port0 N */ + uint64_t p2n0_p0 : 1; /**< BIST Status for P2N Port0 P0 */ + uint64_t p2n0_p1 : 1; /**< BIST Status for P2N Port0 P1 */ + uint64_t p2n1_c0 : 1; /**< BIST Status for P2N Port1 C0 */ + uint64_t p2n1_c1 : 1; /**< BIST Status for P2N Port1 C1 */ + uint64_t p2n1_n : 1; /**< BIST Status for P2N Port1 N */ + uint64_t p2n1_p0 : 1; /**< BIST Status for P2N Port1 P0 */ + uint64_t p2n1_p1 : 1; /**< BIST Status for P2N Port1 P1 */ + uint64_t csm0 : 1; /**< BIST Status for CSM0 */ + uint64_t csm1 : 1; /**< BIST Status for CSM1 */ + uint64_t dif0 : 1; /**< BIST Status for DMA Instr0 */ + uint64_t dif1 : 1; /**< BIST Status for DMA Instr0 */ + uint64_t dif2 : 1; /**< BIST Status for DMA Instr0 */ + uint64_t dif3 : 1; /**< BIST Status for DMA Instr0 */ + uint64_t reserved_2_2 : 1; + uint64_t msi : 1; /**< BIST Status for MSI Memory Map */ + uint64_t ncb_cmd : 1; /**< BIST Status for NCB Outbound Commands */ +#else + uint64_t ncb_cmd : 1; + uint64_t msi : 1; + uint64_t reserved_2_2 : 1; + uint64_t dif3 : 1; + uint64_t dif2 : 1; + uint64_t dif1 : 1; + uint64_t dif0 : 1; + uint64_t csm1 : 1; + uint64_t csm0 : 1; + uint64_t p2n1_p1 : 1; + uint64_t p2n1_p0 : 1; + uint64_t p2n1_n : 1; + uint64_t p2n1_c1 : 1; + uint64_t p2n1_c0 : 1; + uint64_t p2n0_p1 : 1; + uint64_t p2n0_p0 : 1; + uint64_t p2n0_n : 1; + uint64_t p2n0_c1 : 1; + uint64_t p2n0_c0 : 1; + uint64_t p2n0_co : 1; + uint64_t p2n0_no : 1; + uint64_t p2n0_po : 1; + uint64_t p2n1_co : 1; + uint64_t p2n1_no : 1; + uint64_t p2n1_po : 1; + uint64_t cpl_p1 : 1; + uint64_t cpl_p0 : 1; + uint64_t n2p1_o : 1; + uint64_t n2p1_c : 1; + uint64_t n2p0_o : 1; + uint64_t n2p0_c : 1; + uint64_t reserved_31_31 : 1; + uint64_t d3_pst : 1; + uint64_t d2_pst : 1; + uint64_t d1_pst : 1; + uint64_t d0_pst : 1; + uint64_t reserved_36_47 : 12; + uint64_t pkt_slm : 1; + uint64_t pkt_ind : 1; + uint64_t reserved_50_52 : 3; + uint64_t pcsr_sl : 1; + uint64_t pcsr_id : 1; + uint64_t pcsr_cnt : 1; + uint64_t pcsr_im : 1; + uint64_t pcsr_int : 1; + uint64_t pkt_pif : 1; + uint64_t pcr_gim : 1; + uint64_t reserved_60_62 : 3; + uint64_t pkt_rdf : 1; +#endif + } s; + struct cvmx_npei_bist_status_cn52xx + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t pkt_rdf : 1; /**< BIST Status for PKT Read FIFO */ + uint64_t reserved_60_62 : 3; + uint64_t pcr_gim : 1; /**< BIST Status for PKT Gather Instr MEM */ + uint64_t pkt_pif : 1; /**< BIST Status for PKT INB FIFO */ + uint64_t pcsr_int : 1; /**< BIST Status for PKT OUTB Interrupt MEM */ + uint64_t pcsr_im : 1; /**< BIST Status for PKT CSR Instr MEM */ + uint64_t pcsr_cnt : 1; /**< BIST Status for PKT INB Count MEM */ + uint64_t pcsr_id : 1; /**< BIST Status for PKT INB Instr Done MEM */ + uint64_t pcsr_sl : 1; /**< BIST Status for PKT OUTB SLIST MEM */ + uint64_t pkt_imem : 1; /**< BIST Status for PKT OUTB IFIFO */ + uint64_t pkt_pfm : 1; /**< BIST Status for PKT Front MEM */ + uint64_t pkt_pof : 1; /**< BIST Status for PKT OUTB FIFO */ + uint64_t reserved_48_49 : 2; + uint64_t pkt_pop0 : 1; /**< BIST Status for PKT OUTB Slist0 */ + uint64_t pkt_pop1 : 1; /**< BIST Status for PKT OUTB Slist1 */ + uint64_t d0_mem : 1; /**< BIST Status for DMA MEM 0 */ + uint64_t d1_mem : 1; /**< BIST Status for DMA MEM 1 */ + uint64_t d2_mem : 1; /**< BIST Status for DMA MEM 2 */ + uint64_t d3_mem : 1; /**< BIST Status for DMA MEM 3 */ + uint64_t d4_mem : 1; /**< BIST Status for DMA MEM 4 */ + uint64_t ds_mem : 1; /**< BIST Status for DMA Memory */ + uint64_t reserved_36_39 : 4; + uint64_t d0_pst : 1; /**< BIST Status for DMA0 Pcie Store */ + uint64_t d1_pst : 1; /**< BIST Status for DMA1 Pcie Store */ + uint64_t d2_pst : 1; /**< BIST Status for DMA2 Pcie Store */ + uint64_t d3_pst : 1; /**< BIST Status for DMA3 Pcie Store */ + uint64_t d4_pst : 1; /**< BIST Status for DMA4 Pcie Store */ + uint64_t n2p0_c : 1; /**< BIST Status for N2P Port0 Cmd */ + uint64_t n2p0_o : 1; /**< BIST Status for N2P Port0 Data */ + uint64_t n2p1_c : 1; /**< BIST Status for N2P Port1 Cmd */ + uint64_t n2p1_o : 1; /**< BIST Status for N2P Port1 Data */ + uint64_t cpl_p0 : 1; /**< BIST Status for CPL Port 0 */ + uint64_t cpl_p1 : 1; /**< BIST Status for CPL Port 1 */ + uint64_t p2n1_po : 1; /**< BIST Status for P2N Port1 P Order */ + uint64_t p2n1_no : 1; /**< BIST Status for P2N Port1 N Order */ + uint64_t p2n1_co : 1; /**< BIST Status for P2N Port1 C Order */ + uint64_t p2n0_po : 1; /**< BIST Status for P2N Port0 P Order */ + uint64_t p2n0_no : 1; /**< BIST Status for P2N Port0 N Order */ + uint64_t p2n0_co : 1; /**< BIST Status for P2N Port0 C Order */ + uint64_t p2n0_c0 : 1; /**< BIST Status for P2N Port0 C0 */ + uint64_t p2n0_c1 : 1; /**< BIST Status for P2N Port0 C1 */ + uint64_t p2n0_n : 1; /**< BIST Status for P2N Port0 N */ + uint64_t p2n0_p0 : 1; /**< BIST Status for P2N Port0 P0 */ + uint64_t p2n0_p1 : 1; /**< BIST Status for P2N Port0 P1 */ + uint64_t p2n1_c0 : 1; /**< BIST Status for P2N Port1 C0 */ + uint64_t p2n1_c1 : 1; /**< BIST Status for P2N Port1 C1 */ + uint64_t p2n1_n : 1; /**< BIST Status for P2N Port1 N */ + uint64_t p2n1_p0 : 1; /**< BIST Status for P2N Port1 P0 */ + uint64_t p2n1_p1 : 1; /**< BIST Status for P2N Port1 P1 */ + uint64_t csm0 : 1; /**< BIST Status for CSM0 */ + uint64_t csm1 : 1; /**< BIST Status for CSM1 */ + uint64_t dif0 : 1; /**< BIST Status for DMA Instr0 */ + uint64_t dif1 : 1; /**< BIST Status for DMA Instr0 */ + uint64_t dif2 : 1; /**< BIST Status for DMA Instr0 */ + uint64_t dif3 : 1; /**< BIST Status for DMA Instr0 */ + uint64_t dif4 : 1; /**< BIST Status for DMA Instr0 */ + uint64_t msi : 1; /**< BIST Status for MSI Memory Map */ + uint64_t ncb_cmd : 1; /**< BIST Status for NCB Outbound Commands */ +#else + uint64_t ncb_cmd : 1; + uint64_t msi : 1; + uint64_t dif4 : 1; + uint64_t dif3 : 1; + uint64_t dif2 : 1; + uint64_t dif1 : 1; + uint64_t dif0 : 1; + uint64_t csm1 : 1; + uint64_t csm0 : 1; + uint64_t p2n1_p1 : 1; + uint64_t p2n1_p0 : 1; + uint64_t p2n1_n : 1; + uint64_t p2n1_c1 : 1; + uint64_t p2n1_c0 : 1; + uint64_t p2n0_p1 : 1; + uint64_t p2n0_p0 : 1; + uint64_t p2n0_n : 1; + uint64_t p2n0_c1 : 1; + uint64_t p2n0_c0 : 1; + uint64_t p2n0_co : 1; + uint64_t p2n0_no : 1; + uint64_t p2n0_po : 1; + uint64_t p2n1_co : 1; + uint64_t p2n1_no : 1; + uint64_t p2n1_po : 1; + uint64_t cpl_p1 : 1; + uint64_t cpl_p0 : 1; + uint64_t n2p1_o : 1; + uint64_t n2p1_c : 1; + uint64_t n2p0_o : 1; + uint64_t n2p0_c : 1; + uint64_t d4_pst : 1; + uint64_t d3_pst : 1; + uint64_t d2_pst : 1; + uint64_t d1_pst : 1; + uint64_t d0_pst : 1; + uint64_t reserved_36_39 : 4; + uint64_t ds_mem : 1; + uint64_t d4_mem : 1; + uint64_t d3_mem : 1; + uint64_t d2_mem : 1; + uint64_t d1_mem : 1; + uint64_t d0_mem : 1; + uint64_t pkt_pop1 : 1; + uint64_t pkt_pop0 : 1; + uint64_t reserved_48_49 : 2; + uint64_t pkt_pof : 1; + uint64_t pkt_pfm : 1; + uint64_t pkt_imem : 1; + uint64_t pcsr_sl : 1; + uint64_t pcsr_id : 1; + uint64_t pcsr_cnt : 1; + uint64_t pcsr_im : 1; + uint64_t pcsr_int : 1; + uint64_t pkt_pif : 1; + uint64_t pcr_gim : 1; + uint64_t reserved_60_62 : 3; + uint64_t pkt_rdf : 1; +#endif + } cn52xx; + struct cvmx_npei_bist_status_cn52xxp1 + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_46_63 : 18; + uint64_t d0_mem0 : 1; /**< BIST Status for DMA0 Memory */ + uint64_t d1_mem1 : 1; /**< BIST Status for DMA1 Memory */ + uint64_t d2_mem2 : 1; /**< BIST Status for DMA2 Memory */ + uint64_t d3_mem3 : 1; /**< BIST Status for DMA3 Memory */ + uint64_t dr0_mem : 1; /**< BIST Status for DMA0 Store */ + uint64_t d0_mem : 1; /**< BIST Status for DMA0 Memory */ + uint64_t d1_mem : 1; /**< BIST Status for DMA1 Memory */ + uint64_t d2_mem : 1; /**< BIST Status for DMA2 Memory */ + uint64_t d3_mem : 1; /**< BIST Status for DMA3 Memory */ + uint64_t dr1_mem : 1; /**< BIST Status for DMA1 Store */ + uint64_t d0_pst : 1; /**< BIST Status for DMA0 Pcie Store */ + uint64_t d1_pst : 1; /**< BIST Status for DMA1 Pcie Store */ + uint64_t d2_pst : 1; /**< BIST Status for DMA2 Pcie Store */ + uint64_t d3_pst : 1; /**< BIST Status for DMA3 Pcie Store */ + uint64_t dr2_mem : 1; /**< BIST Status for DMA2 Store */ + uint64_t n2p0_c : 1; /**< BIST Status for N2P Port0 Cmd */ + uint64_t n2p0_o : 1; /**< BIST Status for N2P Port0 Data */ + uint64_t n2p1_c : 1; /**< BIST Status for N2P Port1 Cmd */ + uint64_t n2p1_o : 1; /**< BIST Status for N2P Port1 Data */ + uint64_t cpl_p0 : 1; /**< BIST Status for CPL Port 0 */ + uint64_t cpl_p1 : 1; /**< BIST Status for CPL Port 1 */ + uint64_t p2n1_po : 1; /**< BIST Status for P2N Port1 P Order */ + uint64_t p2n1_no : 1; /**< BIST Status for P2N Port1 N Order */ + uint64_t p2n1_co : 1; /**< BIST Status for P2N Port1 C Order */ + uint64_t p2n0_po : 1; /**< BIST Status for P2N Port0 P Order */ + uint64_t p2n0_no : 1; /**< BIST Status for P2N Port0 N Order */ + uint64_t p2n0_co : 1; /**< BIST Status for P2N Port0 C Order */ + uint64_t p2n0_c0 : 1; /**< BIST Status for P2N Port0 C0 */ + uint64_t p2n0_c1 : 1; /**< BIST Status for P2N Port0 C1 */ + uint64_t p2n0_n : 1; /**< BIST Status for P2N Port0 N */ + uint64_t p2n0_p0 : 1; /**< BIST Status for P2N Port0 P0 */ + uint64_t p2n0_p1 : 1; /**< BIST Status for P2N Port0 P1 */ + uint64_t p2n1_c0 : 1; /**< BIST Status for P2N Port1 C0 */ + uint64_t p2n1_c1 : 1; /**< BIST Status for P2N Port1 C1 */ + uint64_t p2n1_n : 1; /**< BIST Status for P2N Port1 N */ + uint64_t p2n1_p0 : 1; /**< BIST Status for P2N Port1 P0 */ + uint64_t p2n1_p1 : 1; /**< BIST Status for P2N Port1 P1 */ + uint64_t csm0 : 1; /**< BIST Status for CSM0 */ + uint64_t csm1 : 1; /**< BIST Status for CSM1 */ + uint64_t dif0 : 1; /**< BIST Status for DMA Instr0 */ + uint64_t dif1 : 1; /**< BIST Status for DMA Instr0 */ + uint64_t dif2 : 1; /**< BIST Status for DMA Instr0 */ + uint64_t dif3 : 1; /**< BIST Status for DMA Instr0 */ + uint64_t dr3_mem : 1; /**< BIST Status for DMA3 Store */ + uint64_t msi : 1; /**< BIST Status for MSI Memory Map */ + uint64_t ncb_cmd : 1; /**< BIST Status for NCB Outbound Commands */ +#else + uint64_t ncb_cmd : 1; + uint64_t msi : 1; + uint64_t dr3_mem : 1; + uint64_t dif3 : 1; + uint64_t dif2 : 1; + uint64_t dif1 : 1; + uint64_t dif0 : 1; + uint64_t csm1 : 1; + uint64_t csm0 : 1; + uint64_t p2n1_p1 : 1; + uint64_t p2n1_p0 : 1; + uint64_t p2n1_n : 1; + uint64_t p2n1_c1 : 1; + uint64_t p2n1_c0 : 1; + uint64_t p2n0_p1 : 1; + uint64_t p2n0_p0 : 1; + uint64_t p2n0_n : 1; + uint64_t p2n0_c1 : 1; + uint64_t p2n0_c0 : 1; + uint64_t p2n0_co : 1; + uint64_t p2n0_no : 1; + uint64_t p2n0_po : 1; + uint64_t p2n1_co : 1; + uint64_t p2n1_no : 1; + uint64_t p2n1_po : 1; + uint64_t cpl_p1 : 1; + uint64_t cpl_p0 : 1; + uint64_t n2p1_o : 1; + uint64_t n2p1_c : 1; + uint64_t n2p0_o : 1; + uint64_t n2p0_c : 1; + uint64_t dr2_mem : 1; + uint64_t d3_pst : 1; + uint64_t d2_pst : 1; + uint64_t d1_pst : 1; + uint64_t d0_pst : 1; + uint64_t dr1_mem : 1; + uint64_t d3_mem : 1; + uint64_t d2_mem : 1; + uint64_t d1_mem : 1; + uint64_t d0_mem : 1; + uint64_t dr0_mem : 1; + uint64_t d3_mem3 : 1; + uint64_t d2_mem2 : 1; + uint64_t d1_mem1 : 1; + uint64_t d0_mem0 : 1; + uint64_t reserved_46_63 : 18; +#endif + } cn52xxp1; + struct cvmx_npei_bist_status_cn52xx cn56xx; + struct cvmx_npei_bist_status_cn56xxp1 + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_58_63 : 6; + uint64_t pcsr_int : 1; /**< BIST Status for PKT pout_int_bstatus */ + uint64_t pcsr_im : 1; /**< BIST Status for PKT pcsr_instr_mem_bstatus */ + uint64_t pcsr_cnt : 1; /**< BIST Status for PKT pin_cnt_bstatus */ + uint64_t pcsr_id : 1; /**< BIST Status for PKT pcsr_in_done_bstatus */ + uint64_t pcsr_sl : 1; /**< BIST Status for PKT pcsr_slist_bstatus */ + uint64_t pkt_pout : 1; /**< BIST Status for PKT OUT Count MEM */ + uint64_t pkt_imem : 1; /**< BIST Status for PKT Instruction MEM */ + uint64_t pkt_cntm : 1; /**< BIST Status for PKT Count MEM */ + uint64_t pkt_ind : 1; /**< BIST Status for PKT Instruction Done MEM */ + uint64_t pkt_slm : 1; /**< BIST Status for PKT SList MEM */ + uint64_t pkt_odf : 1; /**< BIST Status for PKT Output Data FIFO */ + uint64_t pkt_oif : 1; /**< BIST Status for PKT Output INFO FIFO */ + uint64_t pkt_out : 1; /**< BIST Status for PKT Output FIFO */ + uint64_t pkt_i0 : 1; /**< BIST Status for PKT Instr0 */ + uint64_t pkt_i1 : 1; /**< BIST Status for PKT Instr1 */ + uint64_t pkt_s0 : 1; /**< BIST Status for PKT Slist0 */ + uint64_t pkt_s1 : 1; /**< BIST Status for PKT Slist1 */ + uint64_t d0_mem : 1; /**< BIST Status for DMA0 Memory */ + uint64_t d1_mem : 1; /**< BIST Status for DMA1 Memory */ + uint64_t d2_mem : 1; /**< BIST Status for DMA2 Memory */ + uint64_t d3_mem : 1; /**< BIST Status for DMA3 Memory */ + uint64_t d4_mem : 1; /**< BIST Status for DMA4 Memory */ + uint64_t d0_pst : 1; /**< BIST Status for DMA0 Pcie Store */ + uint64_t d1_pst : 1; /**< BIST Status for DMA1 Pcie Store */ + uint64_t d2_pst : 1; /**< BIST Status for DMA2 Pcie Store */ + uint64_t d3_pst : 1; /**< BIST Status for DMA3 Pcie Store */ + uint64_t d4_pst : 1; /**< BIST Status for DMA4 Pcie Store */ + uint64_t n2p0_c : 1; /**< BIST Status for N2P Port0 Cmd */ + uint64_t n2p0_o : 1; /**< BIST Status for N2P Port0 Data */ + uint64_t n2p1_c : 1; /**< BIST Status for N2P Port1 Cmd */ + uint64_t n2p1_o : 1; /**< BIST Status for N2P Port1 Data */ + uint64_t cpl_p0 : 1; /**< BIST Status for CPL Port 0 */ + uint64_t cpl_p1 : 1; /**< BIST Status for CPL Port 1 */ + uint64_t p2n1_po : 1; /**< BIST Status for P2N Port1 P Order */ + uint64_t p2n1_no : 1; /**< BIST Status for P2N Port1 N Order */ + uint64_t p2n1_co : 1; /**< BIST Status for P2N Port1 C Order */ + uint64_t p2n0_po : 1; /**< BIST Status for P2N Port0 P Order */ + uint64_t p2n0_no : 1; /**< BIST Status for P2N Port0 N Order */ + uint64_t p2n0_co : 1; /**< BIST Status for P2N Port0 C Order */ + uint64_t p2n0_c0 : 1; /**< BIST Status for P2N Port0 C0 */ + uint64_t p2n0_c1 : 1; /**< BIST Status for P2N Port0 C1 */ + uint64_t p2n0_n : 1; /**< BIST Status for P2N Port0 N */ + uint64_t p2n0_p0 : 1; /**< BIST Status for P2N Port0 P0 */ + uint64_t p2n0_p1 : 1; /**< BIST Status for P2N Port0 P1 */ + uint64_t p2n1_c0 : 1; /**< BIST Status for P2N Port1 C0 */ + uint64_t p2n1_c1 : 1; /**< BIST Status for P2N Port1 C1 */ + uint64_t p2n1_n : 1; /**< BIST Status for P2N Port1 N */ + uint64_t p2n1_p0 : 1; /**< BIST Status for P2N Port1 P0 */ + uint64_t p2n1_p1 : 1; /**< BIST Status for P2N Port1 P1 */ + uint64_t csm0 : 1; /**< BIST Status for CSM0 */ + uint64_t csm1 : 1; /**< BIST Status for CSM1 */ + uint64_t dif0 : 1; /**< BIST Status for DMA Instr0 */ + uint64_t dif1 : 1; /**< BIST Status for DMA Instr0 */ + uint64_t dif2 : 1; /**< BIST Status for DMA Instr0 */ + uint64_t dif3 : 1; /**< BIST Status for DMA Instr0 */ + uint64_t dif4 : 1; /**< BIST Status for DMA Instr0 */ + uint64_t msi : 1; /**< BIST Status for MSI Memory Map */ + uint64_t ncb_cmd : 1; /**< BIST Status for NCB Outbound Commands */ +#else + uint64_t ncb_cmd : 1; + uint64_t msi : 1; + uint64_t dif4 : 1; + uint64_t dif3 : 1; + uint64_t dif2 : 1; + uint64_t dif1 : 1; + uint64_t dif0 : 1; + uint64_t csm1 : 1; + uint64_t csm0 : 1; + uint64_t p2n1_p1 : 1; + uint64_t p2n1_p0 : 1; + uint64_t p2n1_n : 1; + uint64_t p2n1_c1 : 1; + uint64_t p2n1_c0 : 1; + uint64_t p2n0_p1 : 1; + uint64_t p2n0_p0 : 1; + uint64_t p2n0_n : 1; + uint64_t p2n0_c1 : 1; + uint64_t p2n0_c0 : 1; + uint64_t p2n0_co : 1; + uint64_t p2n0_no : 1; + uint64_t p2n0_po : 1; + uint64_t p2n1_co : 1; + uint64_t p2n1_no : 1; + uint64_t p2n1_po : 1; + uint64_t cpl_p1 : 1; + uint64_t cpl_p0 : 1; + uint64_t n2p1_o : 1; + uint64_t n2p1_c : 1; + uint64_t n2p0_o : 1; + uint64_t n2p0_c : 1; + uint64_t d4_pst : 1; + uint64_t d3_pst : 1; + uint64_t d2_pst : 1; + uint64_t d1_pst : 1; + uint64_t d0_pst : 1; + uint64_t d4_mem : 1; + uint64_t d3_mem : 1; + uint64_t d2_mem : 1; + uint64_t d1_mem : 1; + uint64_t d0_mem : 1; + uint64_t pkt_s1 : 1; + uint64_t pkt_s0 : 1; + uint64_t pkt_i1 : 1; + uint64_t pkt_i0 : 1; + uint64_t pkt_out : 1; + uint64_t pkt_oif : 1; + uint64_t pkt_odf : 1; + uint64_t pkt_slm : 1; + uint64_t pkt_ind : 1; + uint64_t pkt_cntm : 1; + uint64_t pkt_imem : 1; + uint64_t pkt_pout : 1; + uint64_t pcsr_sl : 1; + uint64_t pcsr_id : 1; + uint64_t pcsr_cnt : 1; + uint64_t pcsr_im : 1; + uint64_t pcsr_int : 1; + uint64_t reserved_58_63 : 6; +#endif + } cn56xxp1; +}; +typedef union cvmx_npei_bist_status cvmx_npei_bist_status_t; + +/** + * cvmx_npei_bist_status2 + * + * NPEI_BIST_STATUS2 = NPI's BIST Status Register2 + * + * Results from BIST runs of NPEI's memories. + */ +union cvmx_npei_bist_status2 +{ + uint64_t u64; + struct cvmx_npei_bist_status2_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_14_63 : 50; + uint64_t prd_tag : 1; /**< BIST Status for DMA PCIE RD Tag MEM */ + uint64_t prd_st0 : 1; /**< BIST Status for DMA PCIE RD state MEM 0 */ + uint64_t prd_st1 : 1; /**< BIST Status for DMA PCIE RD state MEM 1 */ + uint64_t prd_err : 1; /**< BIST Status for DMA PCIE RD ERR state MEM */ + uint64_t nrd_st : 1; /**< BIST Status for DMA L2C RD state MEM */ + uint64_t nwe_st : 1; /**< BIST Status for DMA L2C WR state MEM */ + uint64_t nwe_wr0 : 1; /**< BIST Status for DMA L2C WR MEM 0 */ + uint64_t nwe_wr1 : 1; /**< BIST Status for DMA L2C WR MEM 1 */ + uint64_t pkt_rd : 1; /**< BIST Status for Inbound PKT MEM */ + uint64_t psc_p0 : 1; /**< BIST Status for PSC TLP 0 MEM */ + uint64_t psc_p1 : 1; /**< BIST Status for PSC TLP 1 MEM */ + uint64_t pkt_gd : 1; /**< BIST Status for PKT OUTB Gather Data FIFO */ + uint64_t pkt_gl : 1; /**< BIST Status for PKT_OUTB Gather List FIFO */ + uint64_t pkt_blk : 1; /**< BIST Status for PKT OUTB Blocked FIFO */ +#else + uint64_t pkt_blk : 1; + uint64_t pkt_gl : 1; + uint64_t pkt_gd : 1; + uint64_t psc_p1 : 1; + uint64_t psc_p0 : 1; + uint64_t pkt_rd : 1; + uint64_t nwe_wr1 : 1; + uint64_t nwe_wr0 : 1; + uint64_t nwe_st : 1; + uint64_t nrd_st : 1; + uint64_t prd_err : 1; + uint64_t prd_st1 : 1; + uint64_t prd_st0 : 1; + uint64_t prd_tag : 1; + uint64_t reserved_14_63 : 50; +#endif + } s; + struct cvmx_npei_bist_status2_s cn52xx; + struct cvmx_npei_bist_status2_s cn56xx; +}; +typedef union cvmx_npei_bist_status2 cvmx_npei_bist_status2_t; + +/** + * cvmx_npei_ctl_port0 + * + * NPEI_CTL_PORT0 = NPEI's Control Port 0 + * + * Contains control for access for Port0 + */ +union cvmx_npei_ctl_port0 +{ + uint64_t u64; + struct cvmx_npei_ctl_port0_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_21_63 : 43; + uint64_t waitl_com : 1; /**< When set '1' casues the NPI to wait for a commit + from the L2C before sending additional completions + to the L2C from the PCIe. + Set this for more conservative behavior. Clear + this for more aggressive, higher-performance + behavior */ + uint64_t intd : 1; /**< When '0' Intd wire asserted. Before mapping. */ + uint64_t intc : 1; /**< When '0' Intc wire asserted. Before mapping. */ + uint64_t intb : 1; /**< When '0' Intb wire asserted. Before mapping. */ + uint64_t inta : 1; /**< When '0' Inta wire asserted. Before mapping. */ + uint64_t intd_map : 2; /**< Maps INTD to INTA(00), INTB(01), INTC(10) or + INTD (11). */ + uint64_t intc_map : 2; /**< Maps INTC to INTA(00), INTB(01), INTC(10) or + INTD (11). */ + uint64_t intb_map : 2; /**< Maps INTB to INTA(00), INTB(01), INTC(10) or + INTD (11). */ + uint64_t inta_map : 2; /**< Maps INTA to INTA(00), INTB(01), INTC(10) or + INTD (11). */ + uint64_t ctlp_ro : 1; /**< Relaxed ordering enable for Completion TLPS. */ + uint64_t reserved_6_6 : 1; + uint64_t ptlp_ro : 1; /**< Relaxed ordering enable for Posted TLPS. */ + uint64_t bar2_enb : 1; /**< When set '1' BAR2 is enable and will respond when + clear '0' BAR2 access will cause UR responses. */ + uint64_t bar2_esx : 2; /**< Value will be XORed with pci-address[37:36] to + determine the endian swap mode. */ + uint64_t bar2_cax : 1; /**< Value will be XORed with pcie-address[38] to + determine the L2 cache attribute. + Not cached in L2 if XOR result is 1 */ + uint64_t wait_com : 1; /**< When set '1' casues the NPI to wait for a commit + from the L2C before sending additional stores to + the L2C from the PCIe. + Most applications will not notice a difference, so + should not set this bit. Setting the bit is more + conservative on ordering, lower performance */ +#else + uint64_t wait_com : 1; + uint64_t bar2_cax : 1; + uint64_t bar2_esx : 2; + uint64_t bar2_enb : 1; + uint64_t ptlp_ro : 1; + uint64_t reserved_6_6 : 1; + uint64_t ctlp_ro : 1; + uint64_t inta_map : 2; + uint64_t intb_map : 2; + uint64_t intc_map : 2; + uint64_t intd_map : 2; + uint64_t inta : 1; + uint64_t intb : 1; + uint64_t intc : 1; + uint64_t intd : 1; + uint64_t waitl_com : 1; + uint64_t reserved_21_63 : 43; +#endif + } s; + struct cvmx_npei_ctl_port0_s cn52xx; + struct cvmx_npei_ctl_port0_s cn52xxp1; + struct cvmx_npei_ctl_port0_s cn56xx; + struct cvmx_npei_ctl_port0_s cn56xxp1; +}; +typedef union cvmx_npei_ctl_port0 cvmx_npei_ctl_port0_t; + +/** + * cvmx_npei_ctl_port1 + * + * NPEI_CTL_PORT1 = NPEI's Control Port1 + * + * Contains control for access for Port1 + */ +union cvmx_npei_ctl_port1 +{ + uint64_t u64; + struct cvmx_npei_ctl_port1_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_21_63 : 43; + uint64_t waitl_com : 1; /**< When set '1' casues the NPI to wait for a commit + from the L2C before sending additional completions + to the L2C from the PCIe. + Set this for more conservative behavior. Clear + this for more aggressive, higher-performance */ + uint64_t intd : 1; /**< When '0' Intd wire asserted. Before mapping. */ + uint64_t intc : 1; /**< When '0' Intc wire asserted. Before mapping. */ + uint64_t intb : 1; /**< When '0' Intv wire asserted. Before mapping. */ + uint64_t inta : 1; /**< When '0' Inta wire asserted. Before mapping. */ + uint64_t intd_map : 2; /**< Maps INTD to INTA(00), INTB(01), INTC(10) or + INTD (11). */ + uint64_t intc_map : 2; /**< Maps INTC to INTA(00), INTB(01), INTC(10) or + INTD (11). */ + uint64_t intb_map : 2; /**< Maps INTB to INTA(00), INTB(01), INTC(10) or + INTD (11). */ + uint64_t inta_map : 2; /**< Maps INTA to INTA(00), INTB(01), INTC(10) or + INTD (11). */ + uint64_t ctlp_ro : 1; /**< Relaxed ordering enable for Completion TLPS. */ + uint64_t reserved_6_6 : 1; + uint64_t ptlp_ro : 1; /**< Relaxed ordering enable for Posted TLPS. */ + uint64_t bar2_enb : 1; /**< When set '1' BAR2 is enable and will respond when + clear '0' BAR2 access will cause UR responses. */ + uint64_t bar2_esx : 2; /**< Value will be XORed with pci-address[37:36] to + determine the endian swap mode. */ + uint64_t bar2_cax : 1; /**< Value will be XORed with pcie-address[38] to + determine the L2 cache attribute. + Not cached in L2 if XOR result is 1 */ + uint64_t wait_com : 1; /**< When set '1' casues the NPI to wait for a commit + from the L2C before sending additional stores to + the L2C from the PCIe. + Most applications will not notice a difference, so + should not set this bit. Setting the bit is more + conservative on ordering, lower performance */ +#else + uint64_t wait_com : 1; + uint64_t bar2_cax : 1; + uint64_t bar2_esx : 2; + uint64_t bar2_enb : 1; + uint64_t ptlp_ro : 1; + uint64_t reserved_6_6 : 1; + uint64_t ctlp_ro : 1; + uint64_t inta_map : 2; + uint64_t intb_map : 2; + uint64_t intc_map : 2; + uint64_t intd_map : 2; + uint64_t inta : 1; + uint64_t intb : 1; + uint64_t intc : 1; + uint64_t intd : 1; + uint64_t waitl_com : 1; + uint64_t reserved_21_63 : 43; +#endif + } s; + struct cvmx_npei_ctl_port1_s cn52xx; + struct cvmx_npei_ctl_port1_s cn52xxp1; + struct cvmx_npei_ctl_port1_s cn56xx; + struct cvmx_npei_ctl_port1_s cn56xxp1; +}; +typedef union cvmx_npei_ctl_port1 cvmx_npei_ctl_port1_t; + +/** + * cvmx_npei_ctl_status + * + * NPEI_CTL_STATUS = NPEI Control Status Register + * + * Contains control and status for NPEI. Writes to this register are not oSrdered with writes/reads to the PCIe Memory space. + * To ensure that a write has completed the user must read the register before making an access(i.e. PCIe memory space) + * that requires the value of this register to be updated. + */ +union cvmx_npei_ctl_status +{ + uint64_t u64; + struct cvmx_npei_ctl_status_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_44_63 : 20; + uint64_t p1_ntags : 6; /**< Number of tags avaiable for PCIe Port1. + In RC mode 1 tag is needed for each outbound TLP + that requires a CPL TLP. In Endpoint mode the + number of tags required for a TLP request is + 1 per 64-bytes of CPL data + 1. + This field should only be written as part of + reset sequence, before issuing any reads, CFGs, or + IO transactions from the core(s). */ + uint64_t p0_ntags : 6; /**< Number of tags avaiable for PCIe Port0. + In RC mode 1 tag is needed for each outbound TLP + that requires a CPL TLP. In Endpoint mode the + number of tags required for a TLP request is + 1 per 64-bytes of CPL data + 1. + This field should only be written as part of + reset sequence, before issuing any reads, CFGs, or + IO transactions from the core(s). */ + uint64_t cfg_rtry : 16; /**< The time x 0x10000 in core clocks to wait for a + CPL to a CFG RD that does not carry a Retry Status. + Until such time that the timeout occurs and Retry + Status is received for a CFG RD, the Read CFG Read + will be resent. A value of 0 disables retries and + treats a CPL Retry as a CPL UR. */ + uint64_t ring_en : 1; /**< When '0' forces "relative Q position" received + from PKO to be zero, and replicates the back- + pressure indication for the first ring attached + to a PKO port across all the rings attached to a + PKO port. When '1' backpressure is on a per + port/ring. */ + uint64_t lnk_rst : 1; /**< Set when PCIe Core 0 request a link reset due to + link down state. This bit is only reset on raw + reset so it can be read for state to determine if + a reset occured. Bit is cleared when a '1' is + written to this field. */ + uint64_t arb : 1; /**< PCIe switch arbitration mode. '0' == fixed priority + NPEI, PCIe0, then PCIe1. '1' == round robin. */ + uint64_t pkt_bp : 4; /**< Unused */ + uint64_t host_mode : 1; /**< Host mode */ + uint64_t chip_rev : 8; /**< The chip revision. */ +#else + uint64_t chip_rev : 8; + uint64_t host_mode : 1; + uint64_t pkt_bp : 4; + uint64_t arb : 1; + uint64_t lnk_rst : 1; + uint64_t ring_en : 1; + uint64_t cfg_rtry : 16; + uint64_t p0_ntags : 6; + uint64_t p1_ntags : 6; + uint64_t reserved_44_63 : 20; +#endif + } s; + struct cvmx_npei_ctl_status_s cn52xx; + struct cvmx_npei_ctl_status_cn52xxp1 + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_44_63 : 20; + uint64_t p1_ntags : 6; /**< Number of tags avaiable for PCIe Port1. + In RC mode 1 tag is needed for each outbound TLP + that requires a CPL TLP. In Endpoint mode the + number of tags required for a TLP request is + 1 per 64-bytes of CPL data + 1. + This field should only be written as part of + reset sequence, before issuing any reads, CFGs, or + IO transactions from the core(s). */ + uint64_t p0_ntags : 6; /**< Number of tags avaiable for PCIe Port0. + In RC mode 1 tag is needed for each outbound TLP + that requires a CPL TLP. In Endpoint mode the + number of tags required for a TLP request is + 1 per 64-bytes of CPL data + 1. + This field should only be written as part of + reset sequence, before issuing any reads, CFGs, or + IO transactions from the core(s). */ + uint64_t cfg_rtry : 16; /**< The time x 0x10000 in core clocks to wait for a + CPL to a CFG RD that does not carry a Retry Status. + Until such time that the timeout occurs and Retry + Status is received for a CFG RD, the Read CFG Read + will be resent. A value of 0 disables retries and + treats a CPL Retry as a CPL UR. */ + uint64_t reserved_15_15 : 1; + uint64_t lnk_rst : 1; /**< Set when PCIe Core 0 request a link reset due to + link down state. This bit is only reset on raw + reset so it can be read for state to determine if + a reset occured. Bit is cleared when a '1' is + written to this field. */ + uint64_t arb : 1; /**< PCIe switch arbitration mode. '0' == fixed priority + NPEI, PCIe0, then PCIe1. '1' == round robin. */ + uint64_t reserved_9_12 : 4; + uint64_t host_mode : 1; /**< Host mode */ + uint64_t chip_rev : 8; /**< The chip revision. */ +#else + uint64_t chip_rev : 8; + uint64_t host_mode : 1; + uint64_t reserved_9_12 : 4; + uint64_t arb : 1; + uint64_t lnk_rst : 1; + uint64_t reserved_15_15 : 1; + uint64_t cfg_rtry : 16; + uint64_t p0_ntags : 6; + uint64_t p1_ntags : 6; + uint64_t reserved_44_63 : 20; +#endif + } cn52xxp1; + struct cvmx_npei_ctl_status_s cn56xx; + struct cvmx_npei_ctl_status_cn56xxp1 + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_15_63 : 49; + uint64_t lnk_rst : 1; /**< Set when PCIe Core 0 request a link reset due to + link down state. This bit is only reset on raw + reset so it can be read for state to determine if + a reset occured. Bit is cleared when a '1' is + written to this field. */ + uint64_t arb : 1; /**< PCIe switch arbitration mode. '0' == fixed priority + NPEI, PCIe0, then PCIe1. '1' == round robin. */ + uint64_t pkt_bp : 4; /**< Unused */ + uint64_t host_mode : 1; /**< Host mode */ + uint64_t chip_rev : 8; /**< The chip revision. */ +#else + uint64_t chip_rev : 8; + uint64_t host_mode : 1; + uint64_t pkt_bp : 4; + uint64_t arb : 1; + uint64_t lnk_rst : 1; + uint64_t reserved_15_63 : 49; +#endif + } cn56xxp1; +}; +typedef union cvmx_npei_ctl_status cvmx_npei_ctl_status_t; + +/** + * cvmx_npei_ctl_status2 + * + * NPEI_CTL_STATUS2 = NPEI's Control Status2 Register + * + * Contains control and status for NPEI. + * Writes to this register are not ordered with writes/reads to the PCI Memory space. + * To ensure that a write has completed the user must read the register before + * making an access(i.e. PCI memory space) that requires the value of this register to be updated. + */ +union cvmx_npei_ctl_status2 +{ + uint64_t u64; + struct cvmx_npei_ctl_status2_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_16_63 : 48; + uint64_t mps : 1; /**< Max Payload Size + 0 = 128B + 1 = 256B + Note: PCIE*_CFG030[MPS] must be set to the same + value for proper function. */ + uint64_t mrrs : 3; /**< Max Read Request Size + 0 = 128B + 1 = 256B + 2 = 512B + 3 = 1024B + 4 = 2048B + 5 = 4096B + Note: This field must not exceed the desired + max read request size. This means this field + should not exceed PCIE*_CFG030[MRRS]. */ + uint64_t c1_w_flt : 1; /**< When '1' enables the window filter for reads and + writes using the window registers. + PCIE-Port1. + Unfilter writes are: + MIO, SubId0 + MIO, SubId7 + NPEI, SubId0 + NPEI, SubId7 + POW, SubId7 + IPD, SubId7 + USBN0, SubId7 + Unfiltered Reads are: + MIO, SubId0 + MIO, SubId7 + NPEI, SubId0 + NPEI, SubId7 + POW, SubId1 + POW, SubId2 + POW, SubId3 + POW, SubId7 + IPD, SubId7 + USBN0, SubId7 */ + uint64_t c0_w_flt : 1; /**< When '1' enables the window filter for reads and + writes using the window registers. + PCIE-Port0. + Unfilter writes are: + MIO, SubId0 + MIO, SubId7 + NPEI, SubId0 + NPEI, SubId7 + POW, SubId7 + IPD, SubId7 + USBN0, SubId7 + Unfiltered Reads are: + MIO, SubId0 + MIO, SubId7 + NPEI, SubId0 + NPEI, SubId7 + POW, SubId1 + POW, SubId2 + POW, SubId3 + POW, SubId7 + IPD, SubId7 + USBN0, SubId7 */ + uint64_t c1_b1_s : 3; /**< Pcie-Port1, Bar1 Size. 1 == 64MB, 2 == 128MB, + 3 == 256MB, 4 == 512MB, 5 == 1024MB, 6 == 2048MB, + 0 and 7 are reserved. */ + uint64_t c0_b1_s : 3; /**< Pcie-Port0, Bar1 Size. 1 == 64MB, 2 == 128MB, + 3 == 256MB, 4 == 512MB, 5 == 1024MB, 6 == 2048MB, + 0 and 7 are reserved. */ + uint64_t c1_wi_d : 1; /**< When set '1' disables access to the Window + Registers from the PCIe-Port1. */ + uint64_t c1_b0_d : 1; /**< When set '1' disables access from PCIe-Port1 to + BAR-0 address offsets: Less Than 0x270, + Greater than 0x270 AND less than 0x0520, 0x3BC0, + 0x3CD0. */ + uint64_t c0_wi_d : 1; /**< When set '1' disables access to the Window + Registers from the PCIe-Port0. */ + uint64_t c0_b0_d : 1; /**< When set '1' disables access from PCIe-Port0 to + BAR-0 address offsets: Less Than 0x270, + Greater than 0x270 AND less than 0x0520, 0x3BC0, + 0x3CD0. */ +#else + uint64_t c0_b0_d : 1; + uint64_t c0_wi_d : 1; + uint64_t c1_b0_d : 1; + uint64_t c1_wi_d : 1; + uint64_t c0_b1_s : 3; + uint64_t c1_b1_s : 3; + uint64_t c0_w_flt : 1; + uint64_t c1_w_flt : 1; + uint64_t mrrs : 3; + uint64_t mps : 1; + uint64_t reserved_16_63 : 48; +#endif + } s; + struct cvmx_npei_ctl_status2_s cn52xx; + struct cvmx_npei_ctl_status2_s cn52xxp1; + struct cvmx_npei_ctl_status2_s cn56xx; + struct cvmx_npei_ctl_status2_s cn56xxp1; +}; +typedef union cvmx_npei_ctl_status2 cvmx_npei_ctl_status2_t; + +/** + * cvmx_npei_data_out_cnt + * + * NPEI_DATA_OUT_CNT = NPEI DATA OUT COUNT + * + * The EXEC data out fifo-count and the data unload counter. + */ +union cvmx_npei_data_out_cnt +{ + uint64_t u64; + struct cvmx_npei_data_out_cnt_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_44_63 : 20; + uint64_t p1_ucnt : 16; /**< PCIE-Port1 Fifo Unload Count. This counter is + incremented by '1' every time a word is removed + from the Data Out FIFO, whose count is shown in + P0_FCNT. */ + uint64_t p1_fcnt : 6; /**< PCIE-Port1 Data Out Fifo Count. Number of address + data words to be sent out the PCIe port presently + buffered in the FIFO. */ + uint64_t p0_ucnt : 16; /**< PCIE-Port0 Fifo Unload Count. This counter is + incremented by '1' every time a word is removed + from the Data Out FIFO, whose count is shown in + P0_FCNT. */ + uint64_t p0_fcnt : 6; /**< PCIE-Port0 Data Out Fifo Count. Number of address + data words to be sent out the PCIe port presently + buffered in the FIFO. */ +#else + uint64_t p0_fcnt : 6; + uint64_t p0_ucnt : 16; + uint64_t p1_fcnt : 6; + uint64_t p1_ucnt : 16; + uint64_t reserved_44_63 : 20; +#endif + } s; + struct cvmx_npei_data_out_cnt_s cn52xx; + struct cvmx_npei_data_out_cnt_s cn52xxp1; + struct cvmx_npei_data_out_cnt_s cn56xx; + struct cvmx_npei_data_out_cnt_s cn56xxp1; +}; +typedef union cvmx_npei_data_out_cnt cvmx_npei_data_out_cnt_t; + +/** + * cvmx_npei_dbg_data + * + * NPEI_DBG_DATA = NPEI Debug Data Register + * + * Value returned on the debug-data lines from the RSLs + */ +union cvmx_npei_dbg_data +{ + uint64_t u64; + struct cvmx_npei_dbg_data_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_28_63 : 36; + uint64_t qlm0_rev_lanes : 1; /**< Lane reversal for PCIe port 0 */ + uint64_t reserved_25_26 : 2; + uint64_t qlm1_spd : 2; /**< Sets the QLM1 frequency + 0=1.25 Gbaud + 1=2.5 Gbaud + 2=3.125 Gbaud + 3=3.75 Gbaud */ + uint64_t c_mul : 5; /**< PLL_MUL pins sampled at DCOK assertion + Core frequency = 50MHz*C_MUL */ + uint64_t dsel_ext : 1; /**< Allows changes in the external pins to set the + debug select value. */ + uint64_t data : 17; /**< Value on the debug data lines. */ +#else + uint64_t data : 17; + uint64_t dsel_ext : 1; + uint64_t c_mul : 5; + uint64_t qlm1_spd : 2; + uint64_t reserved_25_26 : 2; + uint64_t qlm0_rev_lanes : 1; + uint64_t reserved_28_63 : 36; +#endif + } s; + struct cvmx_npei_dbg_data_cn52xx + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_29_63 : 35; + uint64_t qlm0_link_width : 1; /**< Link width of PCIe port 0 + 0 = PCIe port 0 is 2 lanes, + 2 lane PCIe port 1 exists + 1 = PCIe port 0 is 4 lanes, + PCIe port 1 does not exist */ + uint64_t qlm0_rev_lanes : 1; /**< Lane reversal for PCIe port 0 */ + uint64_t qlm1_mode : 2; /**< Sets the QLM1 Mode + 0=Reserved + 1=XAUI + 2=SGMII + 3=PICMG */ + uint64_t qlm1_spd : 2; /**< Sets the QLM1 frequency + 0=1.25 Gbaud + 1=2.5 Gbaud + 2=3.125 Gbaud + 3=3.75 Gbaud */ + uint64_t c_mul : 5; /**< PLL_MUL pins sampled at DCOK assertion + Core frequency = 50MHz*C_MUL */ + uint64_t dsel_ext : 1; /**< Allows changes in the external pins to set the + debug select value. */ + uint64_t data : 17; /**< Value on the debug data lines. */ +#else + uint64_t data : 17; + uint64_t dsel_ext : 1; + uint64_t c_mul : 5; + uint64_t qlm1_spd : 2; + uint64_t qlm1_mode : 2; + uint64_t qlm0_rev_lanes : 1; + uint64_t qlm0_link_width : 1; + uint64_t reserved_29_63 : 35; +#endif + } cn52xx; + struct cvmx_npei_dbg_data_cn52xx cn52xxp1; + struct cvmx_npei_dbg_data_cn56xx + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_29_63 : 35; + uint64_t qlm2_rev_lanes : 1; /**< Lane reversal for PCIe port 1 */ + uint64_t qlm0_rev_lanes : 1; /**< Lane reversal for PCIe port 0 */ + uint64_t qlm3_spd : 2; /**< Sets the QLM3 frequency + 0=1.25 Gbaud + 1=2.5 Gbaud + 2=3.125 Gbaud + 3=3.75 Gbaud */ + uint64_t qlm1_spd : 2; /**< Sets the QLM1 frequency + 0=1.25 Gbaud + 1=2.5 Gbaud + 2=3.125 Gbaud + 3=3.75 Gbaud */ + uint64_t c_mul : 5; /**< PLL_MUL pins sampled at DCOK assertion + Core frequency = 50MHz*C_MUL */ + uint64_t dsel_ext : 1; /**< Allows changes in the external pins to set the + debug select value. */ + uint64_t data : 17; /**< Value on the debug data lines. */ +#else + uint64_t data : 17; + uint64_t dsel_ext : 1; + uint64_t c_mul : 5; + uint64_t qlm1_spd : 2; + uint64_t qlm3_spd : 2; + uint64_t qlm0_rev_lanes : 1; + uint64_t qlm2_rev_lanes : 1; + uint64_t reserved_29_63 : 35; +#endif + } cn56xx; + struct cvmx_npei_dbg_data_cn56xx cn56xxp1; +}; +typedef union cvmx_npei_dbg_data cvmx_npei_dbg_data_t; + +/** + * cvmx_npei_dbg_select + * + * NPEI_DBG_SELECT = Debug Select Register + * + * Contains the debug select value last written to the RSLs. + */ +union cvmx_npei_dbg_select +{ + uint64_t u64; + struct cvmx_npei_dbg_select_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_16_63 : 48; + uint64_t dbg_sel : 16; /**< When this register is written its value is sent to + all RSLs. */ +#else + uint64_t dbg_sel : 16; + uint64_t reserved_16_63 : 48; +#endif + } s; + struct cvmx_npei_dbg_select_s cn52xx; + struct cvmx_npei_dbg_select_s cn52xxp1; + struct cvmx_npei_dbg_select_s cn56xx; + struct cvmx_npei_dbg_select_s cn56xxp1; +}; +typedef union cvmx_npei_dbg_select cvmx_npei_dbg_select_t; + +/** + * cvmx_npei_dma#_counts + * + * NPEI_DMA[0..4]_COUNTS = DMA Instruction Counts + * + * Values for determing the number of instructions for DMA[0..4] in the NPEI. + */ +union cvmx_npei_dmax_counts +{ + uint64_t u64; + struct cvmx_npei_dmax_counts_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_39_63 : 25; + uint64_t fcnt : 7; /**< Number of words in the Instruction FIFO. */ + uint64_t dbell : 32; /**< Number of available words of Instructions to read. */ +#else + uint64_t dbell : 32; + uint64_t fcnt : 7; + uint64_t reserved_39_63 : 25; +#endif + } s; + struct cvmx_npei_dmax_counts_s cn52xx; + struct cvmx_npei_dmax_counts_s cn52xxp1; + struct cvmx_npei_dmax_counts_s cn56xx; + struct cvmx_npei_dmax_counts_s cn56xxp1; +}; +typedef union cvmx_npei_dmax_counts cvmx_npei_dmax_counts_t; + +/** + * cvmx_npei_dma#_dbell + * + * NPEI_DMA_DBELL[0..4] = DMA Door Bell + * + * The door bell register for DMA[0..4] queue. + */ +union cvmx_npei_dmax_dbell +{ + uint32_t u32; + struct cvmx_npei_dmax_dbell_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint32_t reserved_16_31 : 16; + uint32_t dbell : 16; /**< The value written to this register is added to the + number of 8byte words to be read and processes for + the low priority dma queue. */ +#else + uint32_t dbell : 16; + uint32_t reserved_16_31 : 16; +#endif + } s; + struct cvmx_npei_dmax_dbell_s cn52xx; + struct cvmx_npei_dmax_dbell_s cn52xxp1; + struct cvmx_npei_dmax_dbell_s cn56xx; + struct cvmx_npei_dmax_dbell_s cn56xxp1; +}; +typedef union cvmx_npei_dmax_dbell cvmx_npei_dmax_dbell_t; + +/** + * cvmx_npei_dma#_ibuff_saddr + * + * NPEI_DMA[0..4]_IBUFF_SADDR = DMA Instruction Buffer Starting Address + * + * The address to start reading Instructions from for DMA[0..4]. + */ +union cvmx_npei_dmax_ibuff_saddr +{ + uint64_t u64; + struct cvmx_npei_dmax_ibuff_saddr_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_37_63 : 27; + uint64_t idle : 1; /**< DMA Engine IDLE state */ + uint64_t saddr : 29; /**< The 128 byte aligned starting address to read the + first instruction. SADDR is address bit 35:7 of the + first instructions address. */ + uint64_t reserved_0_6 : 7; +#else + uint64_t reserved_0_6 : 7; + uint64_t saddr : 29; + uint64_t idle : 1; + uint64_t reserved_37_63 : 27; +#endif + } s; + struct cvmx_npei_dmax_ibuff_saddr_s cn52xx; + struct cvmx_npei_dmax_ibuff_saddr_cn52xxp1 + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_36_63 : 28; + uint64_t saddr : 29; /**< The 128 byte aligned starting address to read the + first instruction. SADDR is address bit 35:7 of the + first instructions address. */ + uint64_t reserved_0_6 : 7; +#else + uint64_t reserved_0_6 : 7; + uint64_t saddr : 29; + uint64_t reserved_36_63 : 28; +#endif + } cn52xxp1; + struct cvmx_npei_dmax_ibuff_saddr_s cn56xx; + struct cvmx_npei_dmax_ibuff_saddr_cn52xxp1 cn56xxp1; +}; +typedef union cvmx_npei_dmax_ibuff_saddr cvmx_npei_dmax_ibuff_saddr_t; + +/** + * cvmx_npei_dma#_naddr + * + * NPEI_DMA[0..4]_NADDR = DMA Next Ichunk Address + * + * Place NPEI will read the next Ichunk data from. This is valid when state is 0 + */ +union cvmx_npei_dmax_naddr +{ + uint64_t u64; + struct cvmx_npei_dmax_naddr_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_36_63 : 28; + uint64_t addr : 36; /**< The next L2C address to read DMA# instructions + from. */ +#else + uint64_t addr : 36; + uint64_t reserved_36_63 : 28; +#endif + } s; + struct cvmx_npei_dmax_naddr_s cn52xx; + struct cvmx_npei_dmax_naddr_s cn52xxp1; + struct cvmx_npei_dmax_naddr_s cn56xx; + struct cvmx_npei_dmax_naddr_s cn56xxp1; +}; +typedef union cvmx_npei_dmax_naddr cvmx_npei_dmax_naddr_t; + +/** + * cvmx_npei_dma0_int_level + * + * NPEI_DMA0_INT_LEVEL = NPEI DMA0 Interrupt Level + * + * Thresholds for DMA count and timer interrupts for DMA0. + */ +union cvmx_npei_dma0_int_level +{ + uint64_t u64; + struct cvmx_npei_dma0_int_level_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t time : 32; /**< Whenever the DMA_CNT0 timer exceeds + this value, NPEI_INT_SUM[DTIME0] is set. + The DMA_CNT0 timer increments every core clock + whenever NPEI_DMA_CNTS[DMA0]!=0, and is cleared + when NPEI_INT_SUM[DTIME0] is written with one. */ + uint64_t cnt : 32; /**< Whenever NPEI_DMA_CNTS[DMA0] exceeds this value, + NPEI_INT_SUM[DCNT0] is set. */ +#else + uint64_t cnt : 32; + uint64_t time : 32; +#endif + } s; + struct cvmx_npei_dma0_int_level_s cn52xx; + struct cvmx_npei_dma0_int_level_s cn52xxp1; + struct cvmx_npei_dma0_int_level_s cn56xx; + struct cvmx_npei_dma0_int_level_s cn56xxp1; +}; +typedef union cvmx_npei_dma0_int_level cvmx_npei_dma0_int_level_t; + +/** + * cvmx_npei_dma1_int_level + * + * NPEI_DMA1_INT_LEVEL = NPEI DMA1 Interrupt Level + * + * Thresholds for DMA count and timer interrupts for DMA1. + */ +union cvmx_npei_dma1_int_level +{ + uint64_t u64; + struct cvmx_npei_dma1_int_level_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t time : 32; /**< Whenever the DMA_CNT1 timer exceeds + this value, NPEI_INT_SUM[DTIME1] is set. + The DMA_CNT1 timer increments every core clock + whenever NPEI_DMA_CNTS[DMA1]!=0, and is cleared + when NPEI_INT_SUM[DTIME1] is written with one. */ + uint64_t cnt : 32; /**< Whenever NPEI_DMA_CNTS[DMA1] exceeds this value, + NPEI_INT_SUM[DCNT1] is set. */ +#else + uint64_t cnt : 32; + uint64_t time : 32; +#endif + } s; + struct cvmx_npei_dma1_int_level_s cn52xx; + struct cvmx_npei_dma1_int_level_s cn52xxp1; + struct cvmx_npei_dma1_int_level_s cn56xx; + struct cvmx_npei_dma1_int_level_s cn56xxp1; +}; +typedef union cvmx_npei_dma1_int_level cvmx_npei_dma1_int_level_t; + +/** + * cvmx_npei_dma_cnts + * + * NPEI_DMA_CNTS = NPEI DMA Count + * + * The DMA Count values for DMA0 and DMA1. + */ +union cvmx_npei_dma_cnts +{ + uint64_t u64; + struct cvmx_npei_dma_cnts_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t dma1 : 32; /**< The DMA counter 1. + Writing this field will cause the written value to + be subtracted from DMA1. SW should use a 4-byte + write to access this field so as not to change the + value of other fields in this register. + HW will optionally increment this field after + it completes an OUTBOUND or EXTERNAL-ONLY DMA + instruction. These increments may cause interrupts. + Refer to NPEI_DMA1_INT_LEVEL and + NPEI_INT_SUM[DCNT1,DTIME1]. */ + uint64_t dma0 : 32; /**< The DMA counter 0. + Writing this field will cause the written value to + be subtracted from DMA0. SW should use a 4-byte + write to access this field so as not to change the + value of other fields in this register. + HW will optionally increment this field after + it completes an OUTBOUND or EXTERNAL-ONLY DMA + instruction. These increments may cause interrupts. + Refer to NPEI_DMA0_INT_LEVEL and + NPEI_INT_SUM[DCNT0,DTIME0]. */ +#else + uint64_t dma0 : 32; + uint64_t dma1 : 32; +#endif + } s; + struct cvmx_npei_dma_cnts_s cn52xx; + struct cvmx_npei_dma_cnts_s cn52xxp1; + struct cvmx_npei_dma_cnts_s cn56xx; + struct cvmx_npei_dma_cnts_s cn56xxp1; +}; +typedef union cvmx_npei_dma_cnts cvmx_npei_dma_cnts_t; + +/** + * cvmx_npei_dma_control + * + * NPEI_DMA_CONTROL = DMA Control Register + * + * Controls operation of the DMA IN/OUT. + */ +union cvmx_npei_dma_control +{ + uint64_t u64; + struct cvmx_npei_dma_control_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_40_63 : 24; + uint64_t p_32b_m : 1; /**< DMA PCIE 32-bit word read disable bit + When 0, enable the feature */ + uint64_t dma4_enb : 1; /**< DMA# enable. Enables the operation of the DMA + engine. After being enabled a DMA engine should not + be dis-abled while processing instructions. */ + uint64_t dma3_enb : 1; /**< DMA# enable. Enables the operation of the DMA + engine. After being enabled a DMA engine should not + be dis-abled while processing instructions. */ + uint64_t dma2_enb : 1; /**< DMA# enable. Enables the operation of the DMA + engine. After being enabled a DMA engine should not + be dis-abled while processing instructions. */ + uint64_t dma1_enb : 1; /**< DMA# enable. Enables the operation of the DMA + engine. After being enabled a DMA engine should not + be dis-abled while processing instructions. */ + uint64_t dma0_enb : 1; /**< DMA# enable. Enables the operation of the DMA + engine. After being enabled a DMA engine should not + be dis-abled while processing instructions. */ + uint64_t b0_lend : 1; /**< When set '1' and the NPEI is in the mode to write + 0 to L2C memory when a DMA is done, the address + to be written to will be treated as a Little + Endian address. */ + uint64_t dwb_denb : 1; /**< When set '1' the NPEI will send a value in the DWB + field for a free page operation for the memory + that contained the data. */ + uint64_t dwb_ichk : 9; /**< When Instruction Chunks for DMA operations are freed + this value is used for the DWB field of the + operation. */ + uint64_t fpa_que : 3; /**< The FPA queue that the instruction-chunk page will + be returned to when used. */ + uint64_t o_add1 : 1; /**< When set '1' 1 will be added to the DMA counters, + if '0' then the number of bytes in the dma transfer + will be added to the count register. */ + uint64_t o_ro : 1; /**< Relaxed Ordering Mode for DMA. */ + uint64_t o_ns : 1; /**< Nosnoop For DMA. */ + uint64_t o_es : 2; /**< Endian Swap Mode for DMA. */ + uint64_t o_mode : 1; /**< Select PCI_POINTER MODE to be used. + '1' use pointer values for address and register + values for RO, ES, and NS, '0' use register + values for address and pointer values for + RO, ES, and NS. */ + uint64_t csize : 14; /**< The size in words of the DMA Instruction Chunk. + This value should only be written once. After + writing this value a new value will not be + recognized until the end of the DMA I-Chunk is + reached. */ +#else + uint64_t csize : 14; + uint64_t o_mode : 1; + uint64_t o_es : 2; + uint64_t o_ns : 1; + uint64_t o_ro : 1; + uint64_t o_add1 : 1; + uint64_t fpa_que : 3; + uint64_t dwb_ichk : 9; + uint64_t dwb_denb : 1; + uint64_t b0_lend : 1; + uint64_t dma0_enb : 1; + uint64_t dma1_enb : 1; + uint64_t dma2_enb : 1; + uint64_t dma3_enb : 1; + uint64_t dma4_enb : 1; + uint64_t p_32b_m : 1; + uint64_t reserved_40_63 : 24; +#endif + } s; + struct cvmx_npei_dma_control_s cn52xx; + struct cvmx_npei_dma_control_cn52xxp1 + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_38_63 : 26; + uint64_t dma3_enb : 1; /**< DMA# enable. Enables the operation of the DMA + engine. After being enabled a DMA engine should not + be dis-abled while processing instructions. */ + uint64_t dma2_enb : 1; /**< DMA# enable. Enables the operation of the DMA + engine. After being enabled a DMA engine should not + be dis-abled while processing instructions. */ + uint64_t dma1_enb : 1; /**< DMA# enable. Enables the operation of the DMA + engine. After being enabled a DMA engine should not + be dis-abled while processing instructions. */ + uint64_t dma0_enb : 1; /**< DMA# enable. Enables the operation of the DMA + engine. After being enabled a DMA engine should not + be dis-abled while processing instructions. */ + uint64_t b0_lend : 1; /**< When set '1' and the NPEI is in the mode to write + 0 to L2C memory when a DMA is done, the address + to be written to will be treated as a Little + Endian address. */ + uint64_t dwb_denb : 1; /**< When set '1' the NPEI will send a value in the DWB + field for a free page operation for the memory + that contained the data. */ + uint64_t dwb_ichk : 9; /**< When Instruction Chunks for DMA operations are freed + this value is used for the DWB field of the + operation. */ + uint64_t fpa_que : 3; /**< The FPA queue that the instruction-chunk page will + be returned to when used. */ + uint64_t o_add1 : 1; /**< When set '1' 1 will be added to the DMA counters, + if '0' then the number of bytes in the dma transfer + will be added to the count register. */ + uint64_t o_ro : 1; /**< Relaxed Ordering Mode for DMA. */ + uint64_t o_ns : 1; /**< Nosnoop For DMA. */ + uint64_t o_es : 2; /**< Endian Swap Mode for DMA. */ + uint64_t o_mode : 1; /**< Select PCI_POINTER MODE to be used. + '1' use pointer values for address and register + values for RO, ES, and NS, '0' use register + values for address and pointer values for + RO, ES, and NS. */ + uint64_t csize : 14; /**< The size in words of the DMA Instruction Chunk. + This value should only be written once. After + writing this value a new value will not be + recognized until the end of the DMA I-Chunk is + reached. */ +#else + uint64_t csize : 14; + uint64_t o_mode : 1; + uint64_t o_es : 2; + uint64_t o_ns : 1; + uint64_t o_ro : 1; + uint64_t o_add1 : 1; + uint64_t fpa_que : 3; + uint64_t dwb_ichk : 9; + uint64_t dwb_denb : 1; + uint64_t b0_lend : 1; + uint64_t dma0_enb : 1; + uint64_t dma1_enb : 1; + uint64_t dma2_enb : 1; + uint64_t dma3_enb : 1; + uint64_t reserved_38_63 : 26; +#endif + } cn52xxp1; + struct cvmx_npei_dma_control_s cn56xx; + struct cvmx_npei_dma_control_cn56xxp1 + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_39_63 : 25; + uint64_t dma4_enb : 1; /**< DMA# enable. Enables the operation of the DMA + engine. After being enabled a DMA engine should not + be dis-abled while processing instructions. */ + uint64_t dma3_enb : 1; /**< DMA# enable. Enables the operation of the DMA + engine. After being enabled a DMA engine should not + be dis-abled while processing instructions. */ + uint64_t dma2_enb : 1; /**< DMA# enable. Enables the operation of the DMA + engine. After being enabled a DMA engine should not + be dis-abled while processing instructions. */ + uint64_t dma1_enb : 1; /**< DMA# enable. Enables the operation of the DMA + engine. After being enabled a DMA engine should not + be dis-abled while processing instructions. */ + uint64_t dma0_enb : 1; /**< DMA# enable. Enables the operation of the DMA + engine. After being enabled a DMA engine should not + be dis-abled while processing instructions. */ + uint64_t b0_lend : 1; /**< When set '1' and the NPEI is in the mode to write + 0 to L2C memory when a DMA is done, the address + to be written to will be treated as a Little + Endian address. */ + uint64_t dwb_denb : 1; /**< When set '1' the NPEI will send a value in the DWB + field for a free page operation for the memory + that contained the data. */ + uint64_t dwb_ichk : 9; /**< When Instruction Chunks for DMA operations are freed + this value is used for the DWB field of the + operation. */ + uint64_t fpa_que : 3; /**< The FPA queue that the instruction-chunk page will + be returned to when used. */ + uint64_t o_add1 : 1; /**< When set '1' 1 will be added to the DMA counters, + if '0' then the number of bytes in the dma transfer + will be added to the count register. */ + uint64_t o_ro : 1; /**< Relaxed Ordering Mode for DMA. */ + uint64_t o_ns : 1; /**< Nosnoop For DMA. */ + uint64_t o_es : 2; /**< Endian Swap Mode for DMA. */ + uint64_t o_mode : 1; /**< Select PCI_POINTER MODE to be used. + '1' use pointer values for address and register + values for RO, ES, and NS, '0' use register + values for address and pointer values for + RO, ES, and NS. */ + uint64_t csize : 14; /**< The size in words of the DMA Instruction Chunk. + This value should only be written once. After + writing this value a new value will not be + recognized until the end of the DMA I-Chunk is + reached. */ +#else + uint64_t csize : 14; + uint64_t o_mode : 1; + uint64_t o_es : 2; + uint64_t o_ns : 1; + uint64_t o_ro : 1; + uint64_t o_add1 : 1; + uint64_t fpa_que : 3; + uint64_t dwb_ichk : 9; + uint64_t dwb_denb : 1; + uint64_t b0_lend : 1; + uint64_t dma0_enb : 1; + uint64_t dma1_enb : 1; + uint64_t dma2_enb : 1; + uint64_t dma3_enb : 1; + uint64_t dma4_enb : 1; + uint64_t reserved_39_63 : 25; +#endif + } cn56xxp1; +}; +typedef union cvmx_npei_dma_control cvmx_npei_dma_control_t; + +/** + * cvmx_npei_dma_pcie_req_num + * + * NPEI_DMA_PCIE_REQ_NUM = NPEI DMA PCIE Outstanding Read Request Number + * + * Outstanding PCIE read request number for DMAs and Packet, maximum number is 16 + */ +union cvmx_npei_dma_pcie_req_num +{ + uint64_t u64; + struct cvmx_npei_dma_pcie_req_num_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t dma_arb : 1; /**< DMA_PKT Read Request Arbitration + - 1: DMA0-4 and PKT are round robin. i.e. + DMA0-DMA1-DMA2-DMA3-DMA4-PKT... + - 0: DMA0-4 are round robin, pkt gets selected + half the time. i.e. + DMA0-PKT-DMA1-PKT-DMA2-PKT-DMA3-PKT-DMA4-PKT... */ + uint64_t reserved_53_62 : 10; + uint64_t pkt_cnt : 5; /**< PKT outstanding PCIE Read Request Number for each + PCIe port + When PKT_CNT=x, for each PCIe port, the number + of outstanding PCIe memory space reads by the PCIe + packet input/output will not exceed x. + Valid Number is between 1 and 16 */ + uint64_t reserved_45_47 : 3; + uint64_t dma4_cnt : 5; /**< DMA4 outstanding PCIE Read Request Number + When DMA4_CNT=x, the number of outstanding PCIe + memory space reads by the PCIe DMA engine 4 + will not exceed x. + Valid Number is between 1 and 16 */ + uint64_t reserved_37_39 : 3; + uint64_t dma3_cnt : 5; /**< DMA3 outstanding PCIE Read Request Number + When DMA3_CNT=x, the number of outstanding PCIe + memory space reads by the PCIe DMA engine 3 + will not exceed x. + Valid Number is between 1 and 16 */ + uint64_t reserved_29_31 : 3; + uint64_t dma2_cnt : 5; /**< DMA2 outstanding PCIE Read Request Number + When DMA2_CNT=x, the number of outstanding PCIe + memory space reads by the PCIe DMA engine 2 + will not exceed x. + Valid Number is between 1 and 16 */ + uint64_t reserved_21_23 : 3; + uint64_t dma1_cnt : 5; /**< DMA1 outstanding PCIE Read Request Number + When DMA1_CNT=x, the number of outstanding PCIe + memory space reads by the PCIe DMA engine 1 + will not exceed x. + Valid Number is between 1 and 16 */ + uint64_t reserved_13_15 : 3; + uint64_t dma0_cnt : 5; /**< DMA0 outstanding PCIE Read Request Number + When DMA0_CNT=x, the number of outstanding PCIe + memory space reads by the PCIe DMA engine 0 + will not exceed x. + Valid Number is between 1 and 16 */ + uint64_t reserved_5_7 : 3; + uint64_t dma_cnt : 5; /**< Total outstanding PCIE Read Request Number for each + PCIe port + When DMA_CNT=x, for each PCIe port, the total + number of outstanding PCIe memory space reads + by the PCIe DMA engines and packet input/output + will not exceed x. + Valid Number is between 1 and 16 */ +#else + uint64_t dma_cnt : 5; + uint64_t reserved_5_7 : 3; + uint64_t dma0_cnt : 5; + uint64_t reserved_13_15 : 3; + uint64_t dma1_cnt : 5; + uint64_t reserved_21_23 : 3; + uint64_t dma2_cnt : 5; + uint64_t reserved_29_31 : 3; + uint64_t dma3_cnt : 5; + uint64_t reserved_37_39 : 3; + uint64_t dma4_cnt : 5; + uint64_t reserved_45_47 : 3; + uint64_t pkt_cnt : 5; + uint64_t reserved_53_62 : 10; + uint64_t dma_arb : 1; +#endif + } s; + struct cvmx_npei_dma_pcie_req_num_s cn52xx; + struct cvmx_npei_dma_pcie_req_num_s cn56xx; +}; +typedef union cvmx_npei_dma_pcie_req_num cvmx_npei_dma_pcie_req_num_t; + +/** + * cvmx_npei_dma_state1 + * + * NPEI_DMA_STATE1 = NPI's DMA State 1 + * + * Results from DMA state register 1 + */ +union cvmx_npei_dma_state1 +{ + uint64_t u64; + struct cvmx_npei_dma_state1_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_40_63 : 24; + uint64_t d4_dwe : 8; /**< DMA4 PICe Write State */ + uint64_t d3_dwe : 8; /**< DMA3 PICe Write State */ + uint64_t d2_dwe : 8; /**< DMA2 PICe Write State */ + uint64_t d1_dwe : 8; /**< DMA1 PICe Write State */ + uint64_t d0_dwe : 8; /**< DMA0 PICe Write State */ +#else + uint64_t d0_dwe : 8; + uint64_t d1_dwe : 8; + uint64_t d2_dwe : 8; + uint64_t d3_dwe : 8; + uint64_t d4_dwe : 8; + uint64_t reserved_40_63 : 24; +#endif + } s; + struct cvmx_npei_dma_state1_s cn52xx; +}; +typedef union cvmx_npei_dma_state1 cvmx_npei_dma_state1_t; + +/** + * cvmx_npei_dma_state1_p1 + * + * NPEI_DMA_STATE1_P1 = NPEI DMA Request and Instruction State + * + * DMA engine Debug information. + */ +union cvmx_npei_dma_state1_p1 +{ + uint64_t u64; + struct cvmx_npei_dma_state1_p1_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_60_63 : 4; + uint64_t d0_difst : 7; /**< DMA engine 0 dif instruction read state */ + uint64_t d1_difst : 7; /**< DMA engine 1 dif instruction read state */ + uint64_t d2_difst : 7; /**< DMA engine 2 dif instruction read state */ + uint64_t d3_difst : 7; /**< DMA engine 3 dif instruction read state */ + uint64_t d4_difst : 7; /**< DMA engine 4 dif instruction read state */ + uint64_t d0_reqst : 5; /**< DMA engine 0 request data state */ + uint64_t d1_reqst : 5; /**< DMA engine 1 request data state */ + uint64_t d2_reqst : 5; /**< DMA engine 2 request data state */ + uint64_t d3_reqst : 5; /**< DMA engine 3 request data state */ + uint64_t d4_reqst : 5; /**< DMA engine 4 request data state */ +#else + uint64_t d4_reqst : 5; + uint64_t d3_reqst : 5; + uint64_t d2_reqst : 5; + uint64_t d1_reqst : 5; + uint64_t d0_reqst : 5; + uint64_t d4_difst : 7; + uint64_t d3_difst : 7; + uint64_t d2_difst : 7; + uint64_t d1_difst : 7; + uint64_t d0_difst : 7; + uint64_t reserved_60_63 : 4; +#endif + } s; + struct cvmx_npei_dma_state1_p1_cn52xxp1 + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_60_63 : 4; + uint64_t d0_difst : 7; /**< DMA engine 0 dif instruction read state */ + uint64_t d1_difst : 7; /**< DMA engine 1 dif instruction read state */ + uint64_t d2_difst : 7; /**< DMA engine 2 dif instruction read state */ + uint64_t d3_difst : 7; /**< DMA engine 3 dif instruction read state */ + uint64_t reserved_25_31 : 7; + uint64_t d0_reqst : 5; /**< DMA engine 0 request data state */ + uint64_t d1_reqst : 5; /**< DMA engine 1 request data state */ + uint64_t d2_reqst : 5; /**< DMA engine 2 request data state */ + uint64_t d3_reqst : 5; /**< DMA engine 3 request data state */ + uint64_t reserved_0_4 : 5; +#else + uint64_t reserved_0_4 : 5; + uint64_t d3_reqst : 5; + uint64_t d2_reqst : 5; + uint64_t d1_reqst : 5; + uint64_t d0_reqst : 5; + uint64_t reserved_25_31 : 7; + uint64_t d3_difst : 7; + uint64_t d2_difst : 7; + uint64_t d1_difst : 7; + uint64_t d0_difst : 7; + uint64_t reserved_60_63 : 4; +#endif + } cn52xxp1; + struct cvmx_npei_dma_state1_p1_s cn56xxp1; +}; +typedef union cvmx_npei_dma_state1_p1 cvmx_npei_dma_state1_p1_t; + +/** + * cvmx_npei_dma_state2 + * + * NPEI_DMA_STATE2 = NPI's DMA State 2 + * + * Results from DMA state register 2 + */ +union cvmx_npei_dma_state2 +{ + uint64_t u64; + struct cvmx_npei_dma_state2_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_28_63 : 36; + uint64_t ndwe : 4; /**< DMA L2C Write State */ + uint64_t reserved_21_23 : 3; + uint64_t ndre : 5; /**< DMA L2C Read State */ + uint64_t reserved_10_15 : 6; + uint64_t prd : 10; /**< DMA PICe Read State */ +#else + uint64_t prd : 10; + uint64_t reserved_10_15 : 6; + uint64_t ndre : 5; + uint64_t reserved_21_23 : 3; + uint64_t ndwe : 4; + uint64_t reserved_28_63 : 36; +#endif + } s; + struct cvmx_npei_dma_state2_s cn52xx; +}; +typedef union cvmx_npei_dma_state2 cvmx_npei_dma_state2_t; + +/** + * cvmx_npei_dma_state2_p1 + * + * NPEI_DMA_STATE2_P1 = NPEI DMA Instruction Fetch State + * + * DMA engine Debug information. + */ +union cvmx_npei_dma_state2_p1 +{ + uint64_t u64; + struct cvmx_npei_dma_state2_p1_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_45_63 : 19; + uint64_t d0_dffst : 9; /**< DMA engine 0 dif instruction fetch state */ + uint64_t d1_dffst : 9; /**< DMA engine 1 dif instruction fetch state */ + uint64_t d2_dffst : 9; /**< DMA engine 2 dif instruction fetch state */ + uint64_t d3_dffst : 9; /**< DMA engine 3 dif instruction fetch state */ + uint64_t d4_dffst : 9; /**< DMA engine 4 dif instruction fetch state */ +#else + uint64_t d4_dffst : 9; + uint64_t d3_dffst : 9; + uint64_t d2_dffst : 9; + uint64_t d1_dffst : 9; + uint64_t d0_dffst : 9; + uint64_t reserved_45_63 : 19; +#endif + } s; + struct cvmx_npei_dma_state2_p1_cn52xxp1 + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_45_63 : 19; + uint64_t d0_dffst : 9; /**< DMA engine 0 dif instruction fetch state */ + uint64_t d1_dffst : 9; /**< DMA engine 1 dif instruction fetch state */ + uint64_t d2_dffst : 9; /**< DMA engine 2 dif instruction fetch state */ + uint64_t d3_dffst : 9; /**< DMA engine 3 dif instruction fetch state */ + uint64_t reserved_0_8 : 9; +#else + uint64_t reserved_0_8 : 9; + uint64_t d3_dffst : 9; + uint64_t d2_dffst : 9; + uint64_t d1_dffst : 9; + uint64_t d0_dffst : 9; + uint64_t reserved_45_63 : 19; +#endif + } cn52xxp1; + struct cvmx_npei_dma_state2_p1_s cn56xxp1; +}; +typedef union cvmx_npei_dma_state2_p1 cvmx_npei_dma_state2_p1_t; + +/** + * cvmx_npei_dma_state3_p1 + * + * NPEI_DMA_STATE3_P1 = NPEI DMA DRE State + * + * DMA engine Debug information. + */ +union cvmx_npei_dma_state3_p1 +{ + uint64_t u64; + struct cvmx_npei_dma_state3_p1_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_60_63 : 4; + uint64_t d0_drest : 15; /**< DMA engine 0 dre state */ + uint64_t d1_drest : 15; /**< DMA engine 1 dre state */ + uint64_t d2_drest : 15; /**< DMA engine 2 dre state */ + uint64_t d3_drest : 15; /**< DMA engine 3 dre state */ +#else + uint64_t d3_drest : 15; + uint64_t d2_drest : 15; + uint64_t d1_drest : 15; + uint64_t d0_drest : 15; + uint64_t reserved_60_63 : 4; +#endif + } s; + struct cvmx_npei_dma_state3_p1_s cn52xxp1; + struct cvmx_npei_dma_state3_p1_s cn56xxp1; +}; +typedef union cvmx_npei_dma_state3_p1 cvmx_npei_dma_state3_p1_t; + +/** + * cvmx_npei_dma_state4_p1 + * + * NPEI_DMA_STATE4_P1 = NPEI DMA DWE State + * + * DMA engine Debug information. + */ +union cvmx_npei_dma_state4_p1 +{ + uint64_t u64; + struct cvmx_npei_dma_state4_p1_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_52_63 : 12; + uint64_t d0_dwest : 13; /**< DMA engine 0 dwe state */ + uint64_t d1_dwest : 13; /**< DMA engine 1 dwe state */ + uint64_t d2_dwest : 13; /**< DMA engine 2 dwe state */ + uint64_t d3_dwest : 13; /**< DMA engine 3 dwe state */ +#else + uint64_t d3_dwest : 13; + uint64_t d2_dwest : 13; + uint64_t d1_dwest : 13; + uint64_t d0_dwest : 13; + uint64_t reserved_52_63 : 12; +#endif + } s; + struct cvmx_npei_dma_state4_p1_s cn52xxp1; + struct cvmx_npei_dma_state4_p1_s cn56xxp1; +}; +typedef union cvmx_npei_dma_state4_p1 cvmx_npei_dma_state4_p1_t; + +/** + * cvmx_npei_dma_state5_p1 + * + * NPEI_DMA_STATE5_P1 = NPEI DMA DWE and DRE State + * + * DMA engine Debug information. + */ +union cvmx_npei_dma_state5_p1 +{ + uint64_t u64; + struct cvmx_npei_dma_state5_p1_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_28_63 : 36; + uint64_t d4_drest : 15; /**< DMA engine 4 dre state */ + uint64_t d4_dwest : 13; /**< DMA engine 4 dwe state */ +#else + uint64_t d4_dwest : 13; + uint64_t d4_drest : 15; + uint64_t reserved_28_63 : 36; +#endif + } s; + struct cvmx_npei_dma_state5_p1_s cn56xxp1; +}; +typedef union cvmx_npei_dma_state5_p1 cvmx_npei_dma_state5_p1_t; + +/** + * cvmx_npei_int_a_enb + * + * NPEI_INTERRUPT_A_ENB = NPI's Interrupt A Enable Register + * + * Used to allow the generation of interrupts (MSI/INTA) to the PCIe CoresUsed to enable the various interrupting conditions of NPEI + */ +union cvmx_npei_int_a_enb +{ + uint64_t u64; + struct cvmx_npei_int_a_enb_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_10_63 : 54; + uint64_t pout_err : 1; /**< Enables NPEI_INT_A_SUM[9] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t pin_bp : 1; /**< Enables NPEI_INT_A_SUM[8] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t p1_rdlk : 1; /**< Enables NPEI_INT_A_SUM[7] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t p0_rdlk : 1; /**< Enables NPEI_INT_A_SUM[6] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t pgl_err : 1; /**< Enables NPEI_INT_A_SUM[5] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t pdi_err : 1; /**< Enables NPEI_INT_A_SUM[4] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t pop_err : 1; /**< Enables NPEI_INT_A_SUM[3] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t pins_err : 1; /**< Enables NPEI_INT_A_SUM[2] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dma1_cpl : 1; /**< Enables NPEI_INT_A_SUM[1] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dma0_cpl : 1; /**< Enables NPEI_INT_A_SUM[0] to generate an + interrupt to the PCIE core for MSI/inta. */ +#else + uint64_t dma0_cpl : 1; + uint64_t dma1_cpl : 1; + uint64_t pins_err : 1; + uint64_t pop_err : 1; + uint64_t pdi_err : 1; + uint64_t pgl_err : 1; + uint64_t p0_rdlk : 1; + uint64_t p1_rdlk : 1; + uint64_t pin_bp : 1; + uint64_t pout_err : 1; + uint64_t reserved_10_63 : 54; +#endif + } s; + struct cvmx_npei_int_a_enb_s cn52xx; + struct cvmx_npei_int_a_enb_cn52xxp1 + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_2_63 : 62; + uint64_t dma1_cpl : 1; /**< Enables NPEI_INT_A_SUM[1] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dma0_cpl : 1; /**< Enables NPEI_INT_A_SUM[0] to generate an + interrupt to the PCIE core for MSI/inta. */ +#else + uint64_t dma0_cpl : 1; + uint64_t dma1_cpl : 1; + uint64_t reserved_2_63 : 62; +#endif + } cn52xxp1; + struct cvmx_npei_int_a_enb_s cn56xx; +}; +typedef union cvmx_npei_int_a_enb cvmx_npei_int_a_enb_t; + +/** + * cvmx_npei_int_a_enb2 + * + * NPEI_INTERRUPT_A_ENB2 = NPEI's Interrupt A Enable2 Register + * + * Used to enable the various interrupting conditions of NPEI + */ +union cvmx_npei_int_a_enb2 +{ + uint64_t u64; + struct cvmx_npei_int_a_enb2_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_10_63 : 54; + uint64_t pout_err : 1; /**< Enables NPEI_INT_A_SUM[9] to generate an + interrupt on the RSL. */ + uint64_t pin_bp : 1; /**< Enables NPEI_INT_A_SUM[8] to generate an + interrupt on the RSL. */ + uint64_t p1_rdlk : 1; /**< Enables NPEI_INT_A_SUM[7] to generate an + interrupt on the RSL. */ + uint64_t p0_rdlk : 1; /**< Enables NPEI_INT_A_SUM[6] to generate an + interrupt on the RSL. */ + uint64_t pgl_err : 1; /**< Enables NPEI_INT_A_SUM[5] to generate an + interrupt on the RSL. */ + uint64_t pdi_err : 1; /**< Enables NPEI_INT_A_SUM[4] to generate an + interrupt on the RSL. */ + uint64_t pop_err : 1; /**< Enables NPEI_INT_A_SUM[3] to generate an + interrupt on the RSL. */ + uint64_t pins_err : 1; /**< Enables NPEI_INT_A_SUM[2] to generate an + interrupt on the RSL. */ + uint64_t dma1_cpl : 1; /**< Enables NPEI_INT_A_SUM[1] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dma0_cpl : 1; /**< Enables NPEI_INT_A_SUM[0] to generate an + interrupt to the PCIE core for MSI/inta. */ +#else + uint64_t dma0_cpl : 1; + uint64_t dma1_cpl : 1; + uint64_t pins_err : 1; + uint64_t pop_err : 1; + uint64_t pdi_err : 1; + uint64_t pgl_err : 1; + uint64_t p0_rdlk : 1; + uint64_t p1_rdlk : 1; + uint64_t pin_bp : 1; + uint64_t pout_err : 1; + uint64_t reserved_10_63 : 54; +#endif + } s; + struct cvmx_npei_int_a_enb2_s cn52xx; + struct cvmx_npei_int_a_enb2_cn52xxp1 + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_2_63 : 62; + uint64_t dma1_cpl : 1; /**< Enables NPEI_INT_A_SUM[1] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dma0_cpl : 1; /**< Enables NPEI_INT_A_SUM[0] to generate an + interrupt to the PCIE core for MSI/inta. */ +#else + uint64_t dma0_cpl : 1; + uint64_t dma1_cpl : 1; + uint64_t reserved_2_63 : 62; +#endif + } cn52xxp1; + struct cvmx_npei_int_a_enb2_s cn56xx; +}; +typedef union cvmx_npei_int_a_enb2 cvmx_npei_int_a_enb2_t; + +/** + * cvmx_npei_int_a_sum + * + * NPEI_INTERRUPT_A_SUM = NPI Interrupt A Summary Register + * + * Set when an interrupt condition occurs, write '1' to clear. When an interrupt bitin this register is set and + * the cooresponding bit in the NPEI_INT_A_ENB register is set, then NPEI_INT_SUM[61] will be set. + */ +union cvmx_npei_int_a_sum +{ + uint64_t u64; + struct cvmx_npei_int_a_sum_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_10_63 : 54; + uint64_t pout_err : 1; /**< Set when PKO sends packet data with the error bit + set. */ + uint64_t pin_bp : 1; /**< Packet input count has exceeded the WMARK. + See NPEI_PKT_IN_BP */ + uint64_t p1_rdlk : 1; /**< PCIe port 1 received a read lock. */ + uint64_t p0_rdlk : 1; /**< PCIe port 0 received a read lock. */ + uint64_t pgl_err : 1; /**< When a read error occurs on a packet gather list + read this bit is set. */ + uint64_t pdi_err : 1; /**< When a read error occurs on a packet data read + this bit is set. */ + uint64_t pop_err : 1; /**< When a read error occurs on a packet scatter + pointer pair this bit is set. */ + uint64_t pins_err : 1; /**< When a read error occurs on a packet instruction + this bit is set. */ + uint64_t dma1_cpl : 1; /**< Set each time any PCIe DMA engine recieves a UR/CA + response from PCIe Port 1 */ + uint64_t dma0_cpl : 1; /**< Set each time any PCIe DMA engine recieves a UR/CA + response from PCIe Port 0 */ +#else + uint64_t dma0_cpl : 1; + uint64_t dma1_cpl : 1; + uint64_t pins_err : 1; + uint64_t pop_err : 1; + uint64_t pdi_err : 1; + uint64_t pgl_err : 1; + uint64_t p0_rdlk : 1; + uint64_t p1_rdlk : 1; + uint64_t pin_bp : 1; + uint64_t pout_err : 1; + uint64_t reserved_10_63 : 54; +#endif + } s; + struct cvmx_npei_int_a_sum_s cn52xx; + struct cvmx_npei_int_a_sum_cn52xxp1 + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_2_63 : 62; + uint64_t dma1_cpl : 1; /**< Set each time any PCIe DMA engine recieves a UR/CA + response from PCIe Port 1 */ + uint64_t dma0_cpl : 1; /**< Set each time any PCIe DMA engine recieves a UR/CA + response from PCIe Port 0 */ +#else + uint64_t dma0_cpl : 1; + uint64_t dma1_cpl : 1; + uint64_t reserved_2_63 : 62; +#endif + } cn52xxp1; + struct cvmx_npei_int_a_sum_s cn56xx; +}; +typedef union cvmx_npei_int_a_sum cvmx_npei_int_a_sum_t; + +/** + * cvmx_npei_int_enb + * + * NPEI_INTERRUPT_ENB = NPI's Interrupt Enable Register + * + * Used to allow the generation of interrupts (MSI/INTA) to the PCIe CoresUsed to enable the various interrupting conditions of NPI + */ +union cvmx_npei_int_enb +{ + uint64_t u64; + struct cvmx_npei_int_enb_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t mio_inta : 1; /**< Enables NPEI_INT_SUM[63] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t reserved_62_62 : 1; + uint64_t int_a : 1; /**< Enables NPEI_INT_SUM[61] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_ldwn : 1; /**< Enables NPEI_INT_SUM[60] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_ldwn : 1; /**< Enables NPEI_INT_SUM[59] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_exc : 1; /**< Enables NPEI_INT_SUM[58] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_exc : 1; /**< Enables NPEI_INT_SUM[57] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_up_wf : 1; /**< Enables NPEI_INT_SUM[56] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_up_wf : 1; /**< Enables NPEI_INT_SUM[55] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_un_wf : 1; /**< Enables NPEI_INT_SUM[54] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_un_wf : 1; /**< Enables NPEI_INT_SUM[53] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_un_bx : 1; /**< Enables NPEI_INT_SUM[52] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_un_wi : 1; /**< Enables NPEI_INT_SUM[51] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_un_b2 : 1; /**< Enables NPEI_INT_SUM[50] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_un_b1 : 1; /**< Enables NPEI_INT_SUM[49] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_un_b0 : 1; /**< Enables NPEI_INT_SUM[48] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_up_bx : 1; /**< Enables NPEI_INT_SUM[47] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_up_wi : 1; /**< Enables NPEI_INT_SUM[46] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_up_b2 : 1; /**< Enables NPEI_INT_SUM[45] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_up_b1 : 1; /**< Enables NPEI_INT_SUM[44] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_up_b0 : 1; /**< Enables NPEI_INT_SUM[43] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_un_bx : 1; /**< Enables NPEI_INT_SUM[42] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_un_wi : 1; /**< Enables NPEI_INT_SUM[41] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_un_b2 : 1; /**< Enables NPEI_INT_SUM[40] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_un_b1 : 1; /**< Enables NPEI_INT_SUM[39] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_un_b0 : 1; /**< Enables NPEI_INT_SUM[38] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_up_bx : 1; /**< Enables NPEI_INT_SUM[37] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_up_wi : 1; /**< Enables NPEI_INT_SUM[36] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_up_b2 : 1; /**< Enables NPEI_INT_SUM[35] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_up_b1 : 1; /**< Enables NPEI_INT_SUM[34] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_up_b0 : 1; /**< Enables NPEI_INT_SUM[33] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_hpint : 1; /**< Enables NPEI_INT_SUM[32] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_pmei : 1; /**< Enables NPEI_INT_SUM[31] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_wake : 1; /**< Enables NPEI_INT_SUM[30] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t crs1_dr : 1; /**< Enables NPEI_INT_SUM[29] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_se : 1; /**< Enables NPEI_INT_SUM[28] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t crs1_er : 1; /**< Enables NPEI_INT_SUM[27] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_aeri : 1; /**< Enables NPEI_INT_SUM[26] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_hpint : 1; /**< Enables NPEI_INT_SUM[25] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_pmei : 1; /**< Enables NPEI_INT_SUM[24] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_wake : 1; /**< Enables NPEI_INT_SUM[23] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t crs0_dr : 1; /**< Enables NPEI_INT_SUM[22] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_se : 1; /**< Enables NPEI_INT_SUM[21] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t crs0_er : 1; /**< Enables NPEI_INT_SUM[20] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_aeri : 1; /**< Enables NPEI_INT_SUM[19] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t ptime : 1; /**< Enables NPEI_INT_SUM[18] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t pcnt : 1; /**< Enables NPEI_INT_SUM[17] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t pidbof : 1; /**< Enables NPEI_INT_SUM[16] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t psldbof : 1; /**< Enables NPEI_INT_SUM[15] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dtime1 : 1; /**< Enables NPEI_INT_SUM[14] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dtime0 : 1; /**< Enables NPEI_INT_SUM[13] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dcnt1 : 1; /**< Enables NPEI_INT_SUM[12] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dcnt0 : 1; /**< Enables NPEI_INT_SUM[11] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dma1fi : 1; /**< Enables NPEI_INT_SUM[10] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dma0fi : 1; /**< Enables NPEI_INT_SUM[9] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dma4dbo : 1; /**< Enables NPEI_INT_SUM[8] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dma3dbo : 1; /**< Enables NPEI_INT_SUM[7] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dma2dbo : 1; /**< Enables NPEI_INT_SUM[6] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dma1dbo : 1; /**< Enables NPEI_INT_SUM[5] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dma0dbo : 1; /**< Enables NPEI_INT_SUM[4] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t iob2big : 1; /**< Enables NPEI_INT_SUM[3] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t bar0_to : 1; /**< Enables NPEI_INT_SUM[2] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t rml_wto : 1; /**< Enables NPEI_INT_SUM[1] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t rml_rto : 1; /**< Enables NPEI_INT_SUM[0] to generate an + interrupt to the PCIE core for MSI/inta. */ +#else + uint64_t rml_rto : 1; + uint64_t rml_wto : 1; + uint64_t bar0_to : 1; + uint64_t iob2big : 1; + uint64_t dma0dbo : 1; + uint64_t dma1dbo : 1; + uint64_t dma2dbo : 1; + uint64_t dma3dbo : 1; + uint64_t dma4dbo : 1; + uint64_t dma0fi : 1; + uint64_t dma1fi : 1; + uint64_t dcnt0 : 1; + uint64_t dcnt1 : 1; + uint64_t dtime0 : 1; + uint64_t dtime1 : 1; + uint64_t psldbof : 1; + uint64_t pidbof : 1; + uint64_t pcnt : 1; + uint64_t ptime : 1; + uint64_t c0_aeri : 1; + uint64_t crs0_er : 1; + uint64_t c0_se : 1; + uint64_t crs0_dr : 1; + uint64_t c0_wake : 1; + uint64_t c0_pmei : 1; + uint64_t c0_hpint : 1; + uint64_t c1_aeri : 1; + uint64_t crs1_er : 1; + uint64_t c1_se : 1; + uint64_t crs1_dr : 1; + uint64_t c1_wake : 1; + uint64_t c1_pmei : 1; + uint64_t c1_hpint : 1; + uint64_t c0_up_b0 : 1; + uint64_t c0_up_b1 : 1; + uint64_t c0_up_b2 : 1; + uint64_t c0_up_wi : 1; + uint64_t c0_up_bx : 1; + uint64_t c0_un_b0 : 1; + uint64_t c0_un_b1 : 1; + uint64_t c0_un_b2 : 1; + uint64_t c0_un_wi : 1; + uint64_t c0_un_bx : 1; + uint64_t c1_up_b0 : 1; + uint64_t c1_up_b1 : 1; + uint64_t c1_up_b2 : 1; + uint64_t c1_up_wi : 1; + uint64_t c1_up_bx : 1; + uint64_t c1_un_b0 : 1; + uint64_t c1_un_b1 : 1; + uint64_t c1_un_b2 : 1; + uint64_t c1_un_wi : 1; + uint64_t c1_un_bx : 1; + uint64_t c0_un_wf : 1; + uint64_t c1_un_wf : 1; + uint64_t c0_up_wf : 1; + uint64_t c1_up_wf : 1; + uint64_t c0_exc : 1; + uint64_t c1_exc : 1; + uint64_t c0_ldwn : 1; + uint64_t c1_ldwn : 1; + uint64_t int_a : 1; + uint64_t reserved_62_62 : 1; + uint64_t mio_inta : 1; +#endif + } s; + struct cvmx_npei_int_enb_s cn52xx; + struct cvmx_npei_int_enb_cn52xxp1 + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t mio_inta : 1; /**< Enables NPEI_INT_SUM[63] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t reserved_62_62 : 1; + uint64_t int_a : 1; /**< Enables NPEI_INT_SUM[61] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_ldwn : 1; /**< Enables NPEI_INT_SUM[60] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_ldwn : 1; /**< Enables NPEI_INT_SUM[59] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_exc : 1; /**< Enables NPEI_INT_SUM[58] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_exc : 1; /**< Enables NPEI_INT_SUM[57] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_up_wf : 1; /**< Enables NPEI_INT_SUM[56] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_up_wf : 1; /**< Enables NPEI_INT_SUM[55] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_un_wf : 1; /**< Enables NPEI_INT_SUM[54] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_un_wf : 1; /**< Enables NPEI_INT_SUM[53] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_un_bx : 1; /**< Enables NPEI_INT_SUM[52] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_un_wi : 1; /**< Enables NPEI_INT_SUM[51] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_un_b2 : 1; /**< Enables NPEI_INT_SUM[50] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_un_b1 : 1; /**< Enables NPEI_INT_SUM[49] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_un_b0 : 1; /**< Enables NPEI_INT_SUM[48] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_up_bx : 1; /**< Enables NPEI_INT_SUM[47] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_up_wi : 1; /**< Enables NPEI_INT_SUM[46] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_up_b2 : 1; /**< Enables NPEI_INT_SUM[45] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_up_b1 : 1; /**< Enables NPEI_INT_SUM[44] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_up_b0 : 1; /**< Enables NPEI_INT_SUM[43] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_un_bx : 1; /**< Enables NPEI_INT_SUM[42] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_un_wi : 1; /**< Enables NPEI_INT_SUM[41] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_un_b2 : 1; /**< Enables NPEI_INT_SUM[40] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_un_b1 : 1; /**< Enables NPEI_INT_SUM[39] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_un_b0 : 1; /**< Enables NPEI_INT_SUM[38] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_up_bx : 1; /**< Enables NPEI_INT_SUM[37] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_up_wi : 1; /**< Enables NPEI_INT_SUM[36] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_up_b2 : 1; /**< Enables NPEI_INT_SUM[35] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_up_b1 : 1; /**< Enables NPEI_INT_SUM[34] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_up_b0 : 1; /**< Enables NPEI_INT_SUM[33] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_hpint : 1; /**< Enables NPEI_INT_SUM[32] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_pmei : 1; /**< Enables NPEI_INT_SUM[31] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_wake : 1; /**< Enables NPEI_INT_SUM[30] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t crs1_dr : 1; /**< Enables NPEI_INT_SUM[29] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_se : 1; /**< Enables NPEI_INT_SUM[28] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t crs1_er : 1; /**< Enables NPEI_INT_SUM[27] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_aeri : 1; /**< Enables NPEI_INT_SUM[26] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_hpint : 1; /**< Enables NPEI_INT_SUM[25] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_pmei : 1; /**< Enables NPEI_INT_SUM[24] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_wake : 1; /**< Enables NPEI_INT_SUM[23] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t crs0_dr : 1; /**< Enables NPEI_INT_SUM[22] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_se : 1; /**< Enables NPEI_INT_SUM[21] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t crs0_er : 1; /**< Enables NPEI_INT_SUM[20] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_aeri : 1; /**< Enables NPEI_INT_SUM[19] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t ptime : 1; /**< Enables NPEI_INT_SUM[18] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t pcnt : 1; /**< Enables NPEI_INT_SUM[17] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t pidbof : 1; /**< Enables NPEI_INT_SUM[16] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t psldbof : 1; /**< Enables NPEI_INT_SUM[15] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dtime1 : 1; /**< Enables NPEI_INT_SUM[14] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dtime0 : 1; /**< Enables NPEI_INT_SUM[13] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dcnt1 : 1; /**< Enables NPEI_INT_SUM[12] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dcnt0 : 1; /**< Enables NPEI_INT_SUM[11] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dma1fi : 1; /**< Enables NPEI_INT_SUM[10] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dma0fi : 1; /**< Enables NPEI_INT_SUM[9] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t reserved_8_8 : 1; + uint64_t dma3dbo : 1; /**< Enables NPEI_INT_SUM[7] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dma2dbo : 1; /**< Enables NPEI_INT_SUM[6] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dma1dbo : 1; /**< Enables NPEI_INT_SUM[5] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dma0dbo : 1; /**< Enables NPEI_INT_SUM[4] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t iob2big : 1; /**< Enables NPEI_INT_SUM[3] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t bar0_to : 1; /**< Enables NPEI_INT_SUM[2] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t rml_wto : 1; /**< Enables NPEI_INT_SUM[1] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t rml_rto : 1; /**< Enables NPEI_INT_SUM[0] to generate an + interrupt to the PCIE core for MSI/inta. */ +#else + uint64_t rml_rto : 1; + uint64_t rml_wto : 1; + uint64_t bar0_to : 1; + uint64_t iob2big : 1; + uint64_t dma0dbo : 1; + uint64_t dma1dbo : 1; + uint64_t dma2dbo : 1; + uint64_t dma3dbo : 1; + uint64_t reserved_8_8 : 1; + uint64_t dma0fi : 1; + uint64_t dma1fi : 1; + uint64_t dcnt0 : 1; + uint64_t dcnt1 : 1; + uint64_t dtime0 : 1; + uint64_t dtime1 : 1; + uint64_t psldbof : 1; + uint64_t pidbof : 1; + uint64_t pcnt : 1; + uint64_t ptime : 1; + uint64_t c0_aeri : 1; + uint64_t crs0_er : 1; + uint64_t c0_se : 1; + uint64_t crs0_dr : 1; + uint64_t c0_wake : 1; + uint64_t c0_pmei : 1; + uint64_t c0_hpint : 1; + uint64_t c1_aeri : 1; + uint64_t crs1_er : 1; + uint64_t c1_se : 1; + uint64_t crs1_dr : 1; + uint64_t c1_wake : 1; + uint64_t c1_pmei : 1; + uint64_t c1_hpint : 1; + uint64_t c0_up_b0 : 1; + uint64_t c0_up_b1 : 1; + uint64_t c0_up_b2 : 1; + uint64_t c0_up_wi : 1; + uint64_t c0_up_bx : 1; + uint64_t c0_un_b0 : 1; + uint64_t c0_un_b1 : 1; + uint64_t c0_un_b2 : 1; + uint64_t c0_un_wi : 1; + uint64_t c0_un_bx : 1; + uint64_t c1_up_b0 : 1; + uint64_t c1_up_b1 : 1; + uint64_t c1_up_b2 : 1; + uint64_t c1_up_wi : 1; + uint64_t c1_up_bx : 1; + uint64_t c1_un_b0 : 1; + uint64_t c1_un_b1 : 1; + uint64_t c1_un_b2 : 1; + uint64_t c1_un_wi : 1; + uint64_t c1_un_bx : 1; + uint64_t c0_un_wf : 1; + uint64_t c1_un_wf : 1; + uint64_t c0_up_wf : 1; + uint64_t c1_up_wf : 1; + uint64_t c0_exc : 1; + uint64_t c1_exc : 1; + uint64_t c0_ldwn : 1; + uint64_t c1_ldwn : 1; + uint64_t int_a : 1; + uint64_t reserved_62_62 : 1; + uint64_t mio_inta : 1; +#endif + } cn52xxp1; + struct cvmx_npei_int_enb_s cn56xx; + struct cvmx_npei_int_enb_cn56xxp1 + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t mio_inta : 1; /**< Enables NPEI_INT_SUM[63] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t reserved_61_62 : 2; + uint64_t c1_ldwn : 1; /**< Enables NPEI_INT_SUM[60] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_ldwn : 1; /**< Enables NPEI_INT_SUM[59] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_exc : 1; /**< Enables NPEI_INT_SUM[58] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_exc : 1; /**< Enables NPEI_INT_SUM[57] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_up_wf : 1; /**< Enables NPEI_INT_SUM[56] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_up_wf : 1; /**< Enables NPEI_INT_SUM[55] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_un_wf : 1; /**< Enables NPEI_INT_SUM[54] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_un_wf : 1; /**< Enables NPEI_INT_SUM[53] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_un_bx : 1; /**< Enables NPEI_INT_SUM[52] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_un_wi : 1; /**< Enables NPEI_INT_SUM[51] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_un_b2 : 1; /**< Enables NPEI_INT_SUM[50] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_un_b1 : 1; /**< Enables NPEI_INT_SUM[49] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_un_b0 : 1; /**< Enables NPEI_INT_SUM[48] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_up_bx : 1; /**< Enables NPEI_INT_SUM[47] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_up_wi : 1; /**< Enables NPEI_INT_SUM[46] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_up_b2 : 1; /**< Enables NPEI_INT_SUM[45] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_up_b1 : 1; /**< Enables NPEI_INT_SUM[44] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_up_b0 : 1; /**< Enables NPEI_INT_SUM[43] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_un_bx : 1; /**< Enables NPEI_INT_SUM[42] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_un_wi : 1; /**< Enables NPEI_INT_SUM[41] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_un_b2 : 1; /**< Enables NPEI_INT_SUM[40] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_un_b1 : 1; /**< Enables NPEI_INT_SUM[39] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_un_b0 : 1; /**< Enables NPEI_INT_SUM[38] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_up_bx : 1; /**< Enables NPEI_INT_SUM[37] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_up_wi : 1; /**< Enables NPEI_INT_SUM[36] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_up_b2 : 1; /**< Enables NPEI_INT_SUM[35] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_up_b1 : 1; /**< Enables NPEI_INT_SUM[34] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_up_b0 : 1; /**< Enables NPEI_INT_SUM[33] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_hpint : 1; /**< Enables NPEI_INT_SUM[32] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_pmei : 1; /**< Enables NPEI_INT_SUM[31] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c1_wake : 1; /**< Enables NPEI_INT_SUM[30] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t reserved_29_29 : 1; + uint64_t c1_se : 1; /**< Enables NPEI_INT_SUM[28] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t reserved_27_27 : 1; + uint64_t c1_aeri : 1; /**< Enables NPEI_INT_SUM[26] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_hpint : 1; /**< Enables NPEI_INT_SUM[25] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_pmei : 1; /**< Enables NPEI_INT_SUM[24] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t c0_wake : 1; /**< Enables NPEI_INT_SUM[23] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t reserved_22_22 : 1; + uint64_t c0_se : 1; /**< Enables NPEI_INT_SUM[21] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t reserved_20_20 : 1; + uint64_t c0_aeri : 1; /**< Enables NPEI_INT_SUM[19] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t ptime : 1; /**< Enables NPEI_INT_SUM[18] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t pcnt : 1; /**< Enables NPEI_INT_SUM[17] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t pidbof : 1; /**< Enables NPEI_INT_SUM[16] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t psldbof : 1; /**< Enables NPEI_INT_SUM[15] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dtime1 : 1; /**< Enables NPEI_INT_SUM[14] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dtime0 : 1; /**< Enables NPEI_INT_SUM[13] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dcnt1 : 1; /**< Enables NPEI_INT_SUM[12] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dcnt0 : 1; /**< Enables NPEI_INT_SUM[11] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dma1fi : 1; /**< Enables NPEI_INT_SUM[10] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dma0fi : 1; /**< Enables NPEI_INT_SUM[9] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dma4dbo : 1; /**< Enables NPEI_INT_SUM[8] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dma3dbo : 1; /**< Enables NPEI_INT_SUM[7] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dma2dbo : 1; /**< Enables NPEI_INT_SUM[6] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dma1dbo : 1; /**< Enables NPEI_INT_SUM[5] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t dma0dbo : 1; /**< Enables NPEI_INT_SUM[4] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t iob2big : 1; /**< Enables NPEI_INT_SUM[3] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t bar0_to : 1; /**< Enables NPEI_INT_SUM[2] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t rml_wto : 1; /**< Enables NPEI_INT_SUM[1] to generate an + interrupt to the PCIE core for MSI/inta. */ + uint64_t rml_rto : 1; /**< Enables NPEI_INT_SUM[0] to generate an + interrupt to the PCIE core for MSI/inta. */ +#else + uint64_t rml_rto : 1; + uint64_t rml_wto : 1; + uint64_t bar0_to : 1; + uint64_t iob2big : 1; + uint64_t dma0dbo : 1; + uint64_t dma1dbo : 1; + uint64_t dma2dbo : 1; + uint64_t dma3dbo : 1; + uint64_t dma4dbo : 1; + uint64_t dma0fi : 1; + uint64_t dma1fi : 1; + uint64_t dcnt0 : 1; + uint64_t dcnt1 : 1; + uint64_t dtime0 : 1; + uint64_t dtime1 : 1; + uint64_t psldbof : 1; + uint64_t pidbof : 1; + uint64_t pcnt : 1; + uint64_t ptime : 1; + uint64_t c0_aeri : 1; + uint64_t reserved_20_20 : 1; + uint64_t c0_se : 1; + uint64_t reserved_22_22 : 1; + uint64_t c0_wake : 1; + uint64_t c0_pmei : 1; + uint64_t c0_hpint : 1; + uint64_t c1_aeri : 1; + uint64_t reserved_27_27 : 1; + uint64_t c1_se : 1; + uint64_t reserved_29_29 : 1; + uint64_t c1_wake : 1; + uint64_t c1_pmei : 1; + uint64_t c1_hpint : 1; + uint64_t c0_up_b0 : 1; + uint64_t c0_up_b1 : 1; + uint64_t c0_up_b2 : 1; + uint64_t c0_up_wi : 1; + uint64_t c0_up_bx : 1; + uint64_t c0_un_b0 : 1; + uint64_t c0_un_b1 : 1; + uint64_t c0_un_b2 : 1; + uint64_t c0_un_wi : 1; + uint64_t c0_un_bx : 1; + uint64_t c1_up_b0 : 1; + uint64_t c1_up_b1 : 1; + uint64_t c1_up_b2 : 1; + uint64_t c1_up_wi : 1; + uint64_t c1_up_bx : 1; + uint64_t c1_un_b0 : 1; + uint64_t c1_un_b1 : 1; + uint64_t c1_un_b2 : 1; + uint64_t c1_un_wi : 1; + uint64_t c1_un_bx : 1; + uint64_t c0_un_wf : 1; + uint64_t c1_un_wf : 1; + uint64_t c0_up_wf : 1; + uint64_t c1_up_wf : 1; + uint64_t c0_exc : 1; + uint64_t c1_exc : 1; + uint64_t c0_ldwn : 1; + uint64_t c1_ldwn : 1; + uint64_t reserved_61_62 : 2; + uint64_t mio_inta : 1; +#endif + } cn56xxp1; +}; +typedef union cvmx_npei_int_enb cvmx_npei_int_enb_t; + +/** + * cvmx_npei_int_enb2 + * + * NPEI_INTERRUPT_ENB2 = NPI's Interrupt Enable2 Register + * + * Used to enable the various interrupting conditions of NPI + */ +union cvmx_npei_int_enb2 +{ + uint64_t u64; + struct cvmx_npei_int_enb2_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_62_63 : 2; + uint64_t int_a : 1; /**< Enables NPEI_INT_SUM2[61] to generate an + interrupt on the RSL. */ + uint64_t c1_ldwn : 1; /**< Enables NPEI_INT_SUM[60] to generate an + interrupt on the RSL. */ + uint64_t c0_ldwn : 1; /**< Enables NPEI_INT_SUM[59] to generate an + interrupt on the RSL. */ + uint64_t c1_exc : 1; /**< Enables NPEI_INT_SUM[58] to generate an + interrupt on the RSL. */ + uint64_t c0_exc : 1; /**< Enables NPEI_INT_SUM[57] to generate an + interrupt on the RSL. */ + uint64_t c1_up_wf : 1; /**< Enables NPEI_INT_SUM[56] to generate an + interrupt on the RSL. */ + uint64_t c0_up_wf : 1; /**< Enables NPEI_INT_SUM[55] to generate an + interrupt on the RSL. */ + uint64_t c1_un_wf : 1; /**< Enables NPEI_INT_SUM[54] to generate an + interrupt on the RSL. */ + uint64_t c0_un_wf : 1; /**< Enables NPEI_INT_SUM[53] to generate an + interrupt on the RSL. */ + uint64_t c1_un_bx : 1; /**< Enables NPEI_INT_SUM[52] to generate an + interrupt on the RSL. */ + uint64_t c1_un_wi : 1; /**< Enables NPEI_INT_SUM[51] to generate an + interrupt on the RSL. */ + uint64_t c1_un_b2 : 1; /**< Enables NPEI_INT_SUM[50] to generate an + interrupt on the RSL. */ + uint64_t c1_un_b1 : 1; /**< Enables NPEI_INT_SUM[49] to generate an + interrupt on the RSL. */ + uint64_t c1_un_b0 : 1; /**< Enables NPEI_INT_SUM[48] to generate an + interrupt on the RSL. */ + uint64_t c1_up_bx : 1; /**< Enables NPEI_INT_SUM[47] to generate an + interrupt on the RSL. */ + uint64_t c1_up_wi : 1; /**< Enables NPEI_INT_SUM[46] to generate an + interrupt on the RSL. */ + uint64_t c1_up_b2 : 1; /**< Enables NPEI_INT_SUM[45] to generate an + interrupt on the RSL. */ + uint64_t c1_up_b1 : 1; /**< Enables NPEI_INT_SUM[44] to generate an + interrupt on the RSL. */ + uint64_t c1_up_b0 : 1; /**< Enables NPEI_INT_SUM[43] to generate an + interrupt on the RSL. */ + uint64_t c0_un_bx : 1; /**< Enables NPEI_INT_SUM[42] to generate an + interrupt on the RSL. */ + uint64_t c0_un_wi : 1; /**< Enables NPEI_INT_SUM[41] to generate an + interrupt on the RSL. */ + uint64_t c0_un_b2 : 1; /**< Enables NPEI_INT_SUM[40] to generate an + interrupt on the RSL. */ + uint64_t c0_un_b1 : 1; /**< Enables NPEI_INT_SUM[39] to generate an + interrupt on the RSL. */ + uint64_t c0_un_b0 : 1; /**< Enables NPEI_INT_SUM[38] to generate an + interrupt on the RSL. */ + uint64_t c0_up_bx : 1; /**< Enables NPEI_INT_SUM[37] to generate an + interrupt on the RSL. */ + uint64_t c0_up_wi : 1; /**< Enables NPEI_INT_SUM[36] to generate an + interrupt on the RSL. */ + uint64_t c0_up_b2 : 1; /**< Enables NPEI_INT_SUM[35] to generate an + interrupt on the RSL. */ + uint64_t c0_up_b1 : 1; /**< Enables NPEI_INT_SUM[34] to generate an + interrupt on the RSL. */ + uint64_t c0_up_b0 : 1; /**< Enables NPEI_INT_SUM[33] to generate an + interrupt on the RSL. */ + uint64_t c1_hpint : 1; /**< Enables NPEI_INT_SUM[32] to generate an + interrupt on the RSL. */ + uint64_t c1_pmei : 1; /**< Enables NPEI_INT_SUM[31] to generate an + interrupt on the RSL. */ + uint64_t c1_wake : 1; /**< Enables NPEI_INT_SUM[30] to generate an + interrupt on the RSL. */ + uint64_t crs1_dr : 1; /**< Enables NPEI_INT_SUM2[29] to generate an + interrupt on the RSL. */ + uint64_t c1_se : 1; /**< Enables NPEI_INT_SUM[28] to generate an + interrupt on the RSL. */ + uint64_t crs1_er : 1; /**< Enables NPEI_INT_SUM2[27] to generate an + interrupt on the RSL. */ + uint64_t c1_aeri : 1; /**< Enables NPEI_INT_SUM[26] to generate an + interrupt on the RSL. */ + uint64_t c0_hpint : 1; /**< Enables NPEI_INT_SUM[25] to generate an + interrupt on the RSL. */ + uint64_t c0_pmei : 1; /**< Enables NPEI_INT_SUM[24] to generate an + interrupt on the RSL. */ + uint64_t c0_wake : 1; /**< Enables NPEI_INT_SUM[23] to generate an + interrupt on the RSL. */ + uint64_t crs0_dr : 1; /**< Enables NPEI_INT_SUM2[22] to generate an + interrupt on the RSL. */ + uint64_t c0_se : 1; /**< Enables NPEI_INT_SUM[21] to generate an + interrupt on the RSL. */ + uint64_t crs0_er : 1; /**< Enables NPEI_INT_SUM2[20] to generate an + interrupt on the RSL. */ + uint64_t c0_aeri : 1; /**< Enables NPEI_INT_SUM[19] to generate an + interrupt on the RSL. */ + uint64_t ptime : 1; /**< Enables NPEI_INT_SUM[18] to generate an + interrupt on the RSL. */ + uint64_t pcnt : 1; /**< Enables NPEI_INT_SUM[17] to generate an + interrupt on the RSL. */ + uint64_t pidbof : 1; /**< Enables NPEI_INT_SUM[16] to generate an + interrupt on the RSL. */ + uint64_t psldbof : 1; /**< Enables NPEI_INT_SUM[15] to generate an + interrupt on the RSL. */ + uint64_t dtime1 : 1; /**< Enables NPEI_INT_SUM[14] to generate an + interrupt on the RSL. */ + uint64_t dtime0 : 1; /**< Enables NPEI_INT_SUM[13] to generate an + interrupt on the RSL. */ + uint64_t dcnt1 : 1; /**< Enables NPEI_INT_SUM[12] to generate an + interrupt on the RSL. */ + uint64_t dcnt0 : 1; /**< Enables NPEI_INT_SUM[11] to generate an + interrupt on the RSL. */ + uint64_t dma1fi : 1; /**< Enables NPEI_INT_SUM[10] to generate an + interrupt on the RSL. */ + uint64_t dma0fi : 1; /**< Enables NPEI_INT_SUM[9] to generate an + interrupt on the RSL. */ + uint64_t dma4dbo : 1; /**< Enables NPEI_INT_SUM[8] to generate an + interrupt on the RSL. */ + uint64_t dma3dbo : 1; /**< Enables NPEI_INT_SUM[7] to generate an + interrupt on the RSL. */ + uint64_t dma2dbo : 1; /**< Enables NPEI_INT_SUM[6] to generate an + interrupt on the RSL. */ + uint64_t dma1dbo : 1; /**< Enables NPEI_INT_SUM[5] to generate an + interrupt on the RSL. */ + uint64_t dma0dbo : 1; /**< Enables NPEI_INT_SUM[4] to generate an + interrupt on the RSL. */ + uint64_t iob2big : 1; /**< Enables NPEI_INT_SUM[3] to generate an + interrupt on the RSL. */ + uint64_t bar0_to : 1; /**< Enables NPEI_INT_SUM[2] to generate an + interrupt on the RSL. */ + uint64_t rml_wto : 1; /**< Enables NPEI_INT_SUM[1] to generate an + interrupt on the RSL. */ + uint64_t rml_rto : 1; /**< Enables NPEI_INT_UM[0] to generate an + interrupt on the RSL. */ +#else + uint64_t rml_rto : 1; + uint64_t rml_wto : 1; + uint64_t bar0_to : 1; + uint64_t iob2big : 1; + uint64_t dma0dbo : 1; + uint64_t dma1dbo : 1; + uint64_t dma2dbo : 1; + uint64_t dma3dbo : 1; + uint64_t dma4dbo : 1; + uint64_t dma0fi : 1; + uint64_t dma1fi : 1; + uint64_t dcnt0 : 1; + uint64_t dcnt1 : 1; + uint64_t dtime0 : 1; + uint64_t dtime1 : 1; + uint64_t psldbof : 1; + uint64_t pidbof : 1; + uint64_t pcnt : 1; + uint64_t ptime : 1; + uint64_t c0_aeri : 1; + uint64_t crs0_er : 1; + uint64_t c0_se : 1; + uint64_t crs0_dr : 1; + uint64_t c0_wake : 1; + uint64_t c0_pmei : 1; + uint64_t c0_hpint : 1; + uint64_t c1_aeri : 1; + uint64_t crs1_er : 1; + uint64_t c1_se : 1; + uint64_t crs1_dr : 1; + uint64_t c1_wake : 1; + uint64_t c1_pmei : 1; + uint64_t c1_hpint : 1; + uint64_t c0_up_b0 : 1; + uint64_t c0_up_b1 : 1; + uint64_t c0_up_b2 : 1; + uint64_t c0_up_wi : 1; + uint64_t c0_up_bx : 1; + uint64_t c0_un_b0 : 1; + uint64_t c0_un_b1 : 1; + uint64_t c0_un_b2 : 1; + uint64_t c0_un_wi : 1; + uint64_t c0_un_bx : 1; + uint64_t c1_up_b0 : 1; + uint64_t c1_up_b1 : 1; + uint64_t c1_up_b2 : 1; + uint64_t c1_up_wi : 1; + uint64_t c1_up_bx : 1; + uint64_t c1_un_b0 : 1; + uint64_t c1_un_b1 : 1; + uint64_t c1_un_b2 : 1; + uint64_t c1_un_wi : 1; + uint64_t c1_un_bx : 1; + uint64_t c0_un_wf : 1; + uint64_t c1_un_wf : 1; + uint64_t c0_up_wf : 1; + uint64_t c1_up_wf : 1; + uint64_t c0_exc : 1; + uint64_t c1_exc : 1; + uint64_t c0_ldwn : 1; + uint64_t c1_ldwn : 1; + uint64_t int_a : 1; + uint64_t reserved_62_63 : 2; +#endif + } s; + struct cvmx_npei_int_enb2_s cn52xx; + struct cvmx_npei_int_enb2_cn52xxp1 + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_62_63 : 2; + uint64_t int_a : 1; /**< Enables NPEI_INT_SUM2[61] to generate an + interrupt on the RSL. */ + uint64_t c1_ldwn : 1; /**< Enables NPEI_INT_SUM2[60] to generate an + interrupt on the RSL. */ + uint64_t c0_ldwn : 1; /**< Enables NPEI_INT_SUM2[59] to generate an + interrupt on the RSL. */ + uint64_t c1_exc : 1; /**< Enables NPEI_INT_SUM2[58] to generate an + interrupt on the RSL. */ + uint64_t c0_exc : 1; /**< Enables NPEI_INT_SUM2[57] to generate an + interrupt on the RSL. */ + uint64_t c1_up_wf : 1; /**< Enables NPEI_INT_SUM2[56] to generate an + interrupt on the RSL. */ + uint64_t c0_up_wf : 1; /**< Enables NPEI_INT_SUM2[55] to generate an + interrupt on the RSL. */ + uint64_t c1_un_wf : 1; /**< Enables NPEI_INT_SUM2[54] to generate an + interrupt on the RSL. */ + uint64_t c0_un_wf : 1; /**< Enables NPEI_INT_SUM2[53] to generate an + interrupt on the RSL. */ + uint64_t c1_un_bx : 1; /**< Enables NPEI_INT_SUM2[52] to generate an + interrupt on the RSL. */ + uint64_t c1_un_wi : 1; /**< Enables NPEI_INT_SUM2[51] to generate an + interrupt on the RSL. */ + uint64_t c1_un_b2 : 1; /**< Enables NPEI_INT_SUM2[50] to generate an + interrupt on the RSL. */ + uint64_t c1_un_b1 : 1; /**< Enables NPEI_INT_SUM2[49] to generate an + interrupt on the RSL. */ + uint64_t c1_un_b0 : 1; /**< Enables NPEI_INT_SUM2[48] to generate an + interrupt on the RSL. */ + uint64_t c1_up_bx : 1; /**< Enables NPEI_INT_SUM2[47] to generate an + interrupt on the RSL. */ + uint64_t c1_up_wi : 1; /**< Enables NPEI_INT_SUM2[46] to generate an + interrupt on the RSL. */ + uint64_t c1_up_b2 : 1; /**< Enables NPEI_INT_SUM2[45] to generate an + interrupt on the RSL. */ + uint64_t c1_up_b1 : 1; /**< Enables NPEI_INT_SUM2[44] to generate an + interrupt on the RSL. */ + uint64_t c1_up_b0 : 1; /**< Enables NPEI_INT_SUM2[43] to generate an + interrupt on the RSL. */ + uint64_t c0_un_bx : 1; /**< Enables NPEI_INT_SUM2[42] to generate an + interrupt on the RSL. */ + uint64_t c0_un_wi : 1; /**< Enables NPEI_INT_SUM2[41] to generate an + interrupt on the RSL. */ + uint64_t c0_un_b2 : 1; /**< Enables NPEI_INT_SUM2[40] to generate an + interrupt on the RSL. */ + uint64_t c0_un_b1 : 1; /**< Enables NPEI_INT_SUM2[39] to generate an + interrupt on the RSL. */ + uint64_t c0_un_b0 : 1; /**< Enables NPEI_INT_SUM2[38] to generate an + interrupt on the RSL. */ + uint64_t c0_up_bx : 1; /**< Enables NPEI_INT_SUM2[37] to generate an + interrupt on the RSL. */ + uint64_t c0_up_wi : 1; /**< Enables NPEI_INT_SUM2[36] to generate an + interrupt on the RSL. */ + uint64_t c0_up_b2 : 1; /**< Enables NPEI_INT_SUM2[35] to generate an + interrupt on the RSL. */ + uint64_t c0_up_b1 : 1; /**< Enables NPEI_INT_SUM2[34] to generate an + interrupt on the RSL. */ + uint64_t c0_up_b0 : 1; /**< Enables NPEI_INT_SUM2[33] to generate an + interrupt on the RSL. */ + uint64_t c1_hpint : 1; /**< Enables NPEI_INT_SUM2[32] to generate an + interrupt on the RSL. */ + uint64_t c1_pmei : 1; /**< Enables NPEI_INT_SUM2[31] to generate an + interrupt on the RSL. */ + uint64_t c1_wake : 1; /**< Enables NPEI_INT_SUM2[30] to generate an + interrupt on the RSL. */ + uint64_t crs1_dr : 1; /**< Enables NPEI_INT_SUM2[29] to generate an + interrupt on the RSL. */ + uint64_t c1_se : 1; /**< Enables NPEI_INT_SUM2[28] to generate an + interrupt on the RSL. */ + uint64_t crs1_er : 1; /**< Enables NPEI_INT_SUM2[27] to generate an + interrupt on the RSL. */ + uint64_t c1_aeri : 1; /**< Enables NPEI_INT_SUM2[26] to generate an + interrupt on the RSL. */ + uint64_t c0_hpint : 1; /**< Enables NPEI_INT_SUM2[25] to generate an + interrupt on the RSL. */ + uint64_t c0_pmei : 1; /**< Enables NPEI_INT_SUM2[24] to generate an + interrupt on the RSL. */ + uint64_t c0_wake : 1; /**< Enables NPEI_INT_SUM2[23] to generate an + interrupt on the RSL. */ + uint64_t crs0_dr : 1; /**< Enables NPEI_INT_SUM2[22] to generate an + interrupt on the RSL. */ + uint64_t c0_se : 1; /**< Enables NPEI_INT_SUM2[21] to generate an + interrupt on the RSL. */ + uint64_t crs0_er : 1; /**< Enables NPEI_INT_SUM2[20] to generate an + interrupt on the RSL. */ + uint64_t c0_aeri : 1; /**< Enables NPEI_INT_SUM2[19] to generate an + interrupt on the RSL. */ + uint64_t ptime : 1; /**< Enables NPEI_INT_SUM2[18] to generate an + interrupt on the RSL. */ + uint64_t pcnt : 1; /**< Enables NPEI_INT_SUM2[17] to generate an + interrupt on the RSL. */ + uint64_t pidbof : 1; /**< Enables NPEI_INT_SUM2[16] to generate an + interrupt on the RSL. */ + uint64_t psldbof : 1; /**< Enables NPEI_INT_SUM2[15] to generate an + interrupt on the RSL. */ + uint64_t dtime1 : 1; /**< Enables NPEI_INT_SUM2[14] to generate an + interrupt on the RSL. */ + uint64_t dtime0 : 1; /**< Enables NPEI_INT_SUM2[13] to generate an + interrupt on the RSL. */ + uint64_t dcnt1 : 1; /**< Enables NPEI_INT_SUM2[12] to generate an + interrupt on the RSL. */ + uint64_t dcnt0 : 1; /**< Enables NPEI_INT_SUM2[11] to generate an + interrupt on the RSL. */ + uint64_t dma1fi : 1; /**< Enables NPEI_INT_SUM2[10] to generate an + interrupt on the RSL. */ + uint64_t dma0fi : 1; /**< Enables NPEI_INT_SUM2[9] to generate an + interrupt on the RSL. */ + uint64_t reserved_8_8 : 1; + uint64_t dma3dbo : 1; /**< Enables NPEI_INT_SUM2[7] to generate an + interrupt on the RSL. */ + uint64_t dma2dbo : 1; /**< Enables NPEI_INT_SUM2[6] to generate an + interrupt on the RSL. */ + uint64_t dma1dbo : 1; /**< Enables NPEI_INT_SUM2[5] to generate an + interrupt on the RSL. */ + uint64_t dma0dbo : 1; /**< Enables NPEI_INT_SUM2[4] to generate an + interrupt on the RSL. */ + uint64_t iob2big : 1; /**< Enables NPEI_INT_SUM2[3] to generate an + interrupt on the RSL. */ + uint64_t bar0_to : 1; /**< Enables NPEI_INT_SUM2[2] to generate an + interrupt on the RSL. */ + uint64_t rml_wto : 1; /**< Enables NPEI_INT_SUM2[1] to generate an + interrupt on the RSL. */ + uint64_t rml_rto : 1; /**< Enables NPEI_INT_SUM2[0] to generate an + interrupt on the RSL. */ +#else + uint64_t rml_rto : 1; + uint64_t rml_wto : 1; + uint64_t bar0_to : 1; + uint64_t iob2big : 1; + uint64_t dma0dbo : 1; + uint64_t dma1dbo : 1; + uint64_t dma2dbo : 1; + uint64_t dma3dbo : 1; + uint64_t reserved_8_8 : 1; + uint64_t dma0fi : 1; + uint64_t dma1fi : 1; + uint64_t dcnt0 : 1; + uint64_t dcnt1 : 1; + uint64_t dtime0 : 1; + uint64_t dtime1 : 1; + uint64_t psldbof : 1; + uint64_t pidbof : 1; + uint64_t pcnt : 1; + uint64_t ptime : 1; + uint64_t c0_aeri : 1; + uint64_t crs0_er : 1; + uint64_t c0_se : 1; + uint64_t crs0_dr : 1; + uint64_t c0_wake : 1; + uint64_t c0_pmei : 1; + uint64_t c0_hpint : 1; + uint64_t c1_aeri : 1; + uint64_t crs1_er : 1; + uint64_t c1_se : 1; + uint64_t crs1_dr : 1; + uint64_t c1_wake : 1; + uint64_t c1_pmei : 1; + uint64_t c1_hpint : 1; + uint64_t c0_up_b0 : 1; + uint64_t c0_up_b1 : 1; + uint64_t c0_up_b2 : 1; + uint64_t c0_up_wi : 1; + uint64_t c0_up_bx : 1; + uint64_t c0_un_b0 : 1; + uint64_t c0_un_b1 : 1; + uint64_t c0_un_b2 : 1; + uint64_t c0_un_wi : 1; + uint64_t c0_un_bx : 1; + uint64_t c1_up_b0 : 1; + uint64_t c1_up_b1 : 1; + uint64_t c1_up_b2 : 1; + uint64_t c1_up_wi : 1; + uint64_t c1_up_bx : 1; + uint64_t c1_un_b0 : 1; + uint64_t c1_un_b1 : 1; + uint64_t c1_un_b2 : 1; + uint64_t c1_un_wi : 1; + uint64_t c1_un_bx : 1; + uint64_t c0_un_wf : 1; + uint64_t c1_un_wf : 1; + uint64_t c0_up_wf : 1; + uint64_t c1_up_wf : 1; + uint64_t c0_exc : 1; + uint64_t c1_exc : 1; + uint64_t c0_ldwn : 1; + uint64_t c1_ldwn : 1; + uint64_t int_a : 1; + uint64_t reserved_62_63 : 2; +#endif + } cn52xxp1; + struct cvmx_npei_int_enb2_s cn56xx; + struct cvmx_npei_int_enb2_cn56xxp1 + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_61_63 : 3; + uint64_t c1_ldwn : 1; /**< Enables NPEI_INT_SUM[60] to generate an + interrupt on the RSL. */ + uint64_t c0_ldwn : 1; /**< Enables NPEI_INT_SUM[59] to generate an + interrupt on the RSL. */ + uint64_t c1_exc : 1; /**< Enables NPEI_INT_SUM[58] to generate an + interrupt on the RSL. */ + uint64_t c0_exc : 1; /**< Enables NPEI_INT_SUM[57] to generate an + interrupt on the RSL. */ + uint64_t c1_up_wf : 1; /**< Enables NPEI_INT_SUM[56] to generate an + interrupt on the RSL. */ + uint64_t c0_up_wf : 1; /**< Enables NPEI_INT_SUM[55] to generate an + interrupt on the RSL. */ + uint64_t c1_un_wf : 1; /**< Enables NPEI_INT_SUM[54] to generate an + interrupt on the RSL. */ + uint64_t c0_un_wf : 1; /**< Enables NPEI_INT_SUM[53] to generate an + interrupt on the RSL. */ + uint64_t c1_un_bx : 1; /**< Enables NPEI_INT_SUM[52] to generate an + interrupt on the RSL. */ + uint64_t c1_un_wi : 1; /**< Enables NPEI_INT_SUM[51] to generate an + interrupt on the RSL. */ + uint64_t c1_un_b2 : 1; /**< Enables NPEI_INT_SUM[50] to generate an + interrupt on the RSL. */ + uint64_t c1_un_b1 : 1; /**< Enables NPEI_INT_SUM[49] to generate an + interrupt on the RSL. */ + uint64_t c1_un_b0 : 1; /**< Enables NPEI_INT_SUM[48] to generate an + interrupt on the RSL. */ + uint64_t c1_up_bx : 1; /**< Enables NPEI_INT_SUM[47] to generate an + interrupt on the RSL. */ + uint64_t c1_up_wi : 1; /**< Enables NPEI_INT_SUM[46] to generate an + interrupt on the RSL. */ + uint64_t c1_up_b2 : 1; /**< Enables NPEI_INT_SUM[45] to generate an + interrupt on the RSL. */ + uint64_t c1_up_b1 : 1; /**< Enables NPEI_INT_SUM[44] to generate an + interrupt on the RSL. */ + uint64_t c1_up_b0 : 1; /**< Enables NPEI_INT_SUM[43] to generate an + interrupt on the RSL. */ + uint64_t c0_un_bx : 1; /**< Enables NPEI_INT_SUM[42] to generate an + interrupt on the RSL. */ + uint64_t c0_un_wi : 1; /**< Enables NPEI_INT_SUM[41] to generate an + interrupt on the RSL. */ + uint64_t c0_un_b2 : 1; /**< Enables NPEI_INT_SUM[40] to generate an + interrupt on the RSL. */ + uint64_t c0_un_b1 : 1; /**< Enables NPEI_INT_SUM[39] to generate an + interrupt on the RSL. */ + uint64_t c0_un_b0 : 1; /**< Enables NPEI_INT_SUM[38] to generate an + interrupt on the RSL. */ + uint64_t c0_up_bx : 1; /**< Enables NPEI_INT_SUM[37] to generate an + interrupt on the RSL. */ + uint64_t c0_up_wi : 1; /**< Enables NPEI_INT_SUM[36] to generate an + interrupt on the RSL. */ + uint64_t c0_up_b2 : 1; /**< Enables NPEI_INT_SUM[35] to generate an + interrupt on the RSL. */ + uint64_t c0_up_b1 : 1; /**< Enables NPEI_INT_SUM[34] to generate an + interrupt on the RSL. */ + uint64_t c0_up_b0 : 1; /**< Enables NPEI_INT_SUM[33] to generate an + interrupt on the RSL. */ + uint64_t c1_hpint : 1; /**< Enables NPEI_INT_SUM[32] to generate an + interrupt on the RSL. */ + uint64_t c1_pmei : 1; /**< Enables NPEI_INT_SUM[31] to generate an + interrupt on the RSL. */ + uint64_t c1_wake : 1; /**< Enables NPEI_INT_SUM[30] to generate an + interrupt on the RSL. */ + uint64_t reserved_29_29 : 1; + uint64_t c1_se : 1; /**< Enables NPEI_INT_SUM[28] to generate an + interrupt on the RSL. */ + uint64_t reserved_27_27 : 1; + uint64_t c1_aeri : 1; /**< Enables NPEI_INT_SUM[26] to generate an + interrupt on the RSL. */ + uint64_t c0_hpint : 1; /**< Enables NPEI_INT_SUM[25] to generate an + interrupt on the RSL. */ + uint64_t c0_pmei : 1; /**< Enables NPEI_INT_SUM[24] to generate an + interrupt on the RSL. */ + uint64_t c0_wake : 1; /**< Enables NPEI_INT_SUM[23] to generate an + interrupt on the RSL. */ + uint64_t reserved_22_22 : 1; + uint64_t c0_se : 1; /**< Enables NPEI_INT_SUM[21] to generate an + interrupt on the RSL. */ + uint64_t reserved_20_20 : 1; + uint64_t c0_aeri : 1; /**< Enables NPEI_INT_SUM[19] to generate an + interrupt on the RSL. */ + uint64_t ptime : 1; /**< Enables NPEI_INT_SUM[18] to generate an + interrupt on the RSL. */ + uint64_t pcnt : 1; /**< Enables NPEI_INT_SUM[17] to generate an + interrupt on the RSL. */ + uint64_t pidbof : 1; /**< Enables NPEI_INT_SUM[16] to generate an + interrupt on the RSL. */ + uint64_t psldbof : 1; /**< Enables NPEI_INT_SUM[15] to generate an + interrupt on the RSL. */ + uint64_t dtime1 : 1; /**< Enables NPEI_INT_SUM[14] to generate an + interrupt on the RSL. */ + uint64_t dtime0 : 1; /**< Enables NPEI_INT_SUM[13] to generate an + interrupt on the RSL. */ + uint64_t dcnt1 : 1; /**< Enables NPEI_INT_SUM[12] to generate an + interrupt on the RSL. */ + uint64_t dcnt0 : 1; /**< Enables NPEI_INT_SUM[11] to generate an + interrupt on the RSL. */ + uint64_t dma1fi : 1; /**< Enables NPEI_INT_SUM[10] to generate an + interrupt on the RSL. */ + uint64_t dma0fi : 1; /**< Enables NPEI_INT_SUM[9] to generate an + interrupt on the RSL. */ + uint64_t dma4dbo : 1; /**< Enables NPEI_INT_SUM[8] to generate an + interrupt on the RSL. */ + uint64_t dma3dbo : 1; /**< Enables NPEI_INT_SUM[7] to generate an + interrupt on the RSL. */ + uint64_t dma2dbo : 1; /**< Enables NPEI_INT_SUM[6] to generate an + interrupt on the RSL. */ + uint64_t dma1dbo : 1; /**< Enables NPEI_INT_SUM[5] to generate an + interrupt on the RSL. */ + uint64_t dma0dbo : 1; /**< Enables NPEI_INT_SUM[4] to generate an + interrupt on the RSL. */ + uint64_t iob2big : 1; /**< Enables NPEI_INT_SUM[3] to generate an + interrupt on the RSL. */ + uint64_t bar0_to : 1; /**< Enables NPEI_INT_SUM[2] to generate an + interrupt on the RSL. */ + uint64_t rml_wto : 1; /**< Enables NPEI_INT_SUM[1] to generate an + interrupt on the RSL. */ + uint64_t rml_rto : 1; /**< Enables NPEI_INT_UM[0] to generate an + interrupt on the RSL. */ +#else + uint64_t rml_rto : 1; + uint64_t rml_wto : 1; + uint64_t bar0_to : 1; + uint64_t iob2big : 1; + uint64_t dma0dbo : 1; + uint64_t dma1dbo : 1; + uint64_t dma2dbo : 1; + uint64_t dma3dbo : 1; + uint64_t dma4dbo : 1; + uint64_t dma0fi : 1; + uint64_t dma1fi : 1; + uint64_t dcnt0 : 1; + uint64_t dcnt1 : 1; + uint64_t dtime0 : 1; + uint64_t dtime1 : 1; + uint64_t psldbof : 1; + uint64_t pidbof : 1; + uint64_t pcnt : 1; + uint64_t ptime : 1; + uint64_t c0_aeri : 1; + uint64_t reserved_20_20 : 1; + uint64_t c0_se : 1; + uint64_t reserved_22_22 : 1; + uint64_t c0_wake : 1; + uint64_t c0_pmei : 1; + uint64_t c0_hpint : 1; + uint64_t c1_aeri : 1; + uint64_t reserved_27_27 : 1; + uint64_t c1_se : 1; + uint64_t reserved_29_29 : 1; + uint64_t c1_wake : 1; + uint64_t c1_pmei : 1; + uint64_t c1_hpint : 1; + uint64_t c0_up_b0 : 1; + uint64_t c0_up_b1 : 1; + uint64_t c0_up_b2 : 1; + uint64_t c0_up_wi : 1; + uint64_t c0_up_bx : 1; + uint64_t c0_un_b0 : 1; + uint64_t c0_un_b1 : 1; + uint64_t c0_un_b2 : 1; + uint64_t c0_un_wi : 1; + uint64_t c0_un_bx : 1; + uint64_t c1_up_b0 : 1; + uint64_t c1_up_b1 : 1; + uint64_t c1_up_b2 : 1; + uint64_t c1_up_wi : 1; + uint64_t c1_up_bx : 1; + uint64_t c1_un_b0 : 1; + uint64_t c1_un_b1 : 1; + uint64_t c1_un_b2 : 1; + uint64_t c1_un_wi : 1; + uint64_t c1_un_bx : 1; + uint64_t c0_un_wf : 1; + uint64_t c1_un_wf : 1; + uint64_t c0_up_wf : 1; + uint64_t c1_up_wf : 1; + uint64_t c0_exc : 1; + uint64_t c1_exc : 1; + uint64_t c0_ldwn : 1; + uint64_t c1_ldwn : 1; + uint64_t reserved_61_63 : 3; +#endif + } cn56xxp1; +}; +typedef union cvmx_npei_int_enb2 cvmx_npei_int_enb2_t; + +/** + * cvmx_npei_int_info + * + * NPEI_INT_INFO = NPI Interrupt Information + * + * Contains information about some of the interrupt condition that can occur in the NPEI_INTERRUPT_SUM register. + */ +union cvmx_npei_int_info +{ + uint64_t u64; + struct cvmx_npei_int_info_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_12_63 : 52; + uint64_t pidbof : 6; /**< Field set when the NPEI_INTERRUPT_SUM[PIDBOF] bit + is set. This field when set will not change again + unitl NPEI_INTERRUPT_SUM[PIDBOF] is cleared. */ + uint64_t psldbof : 6; /**< Field set when the NPEI_INTERRUPT_SUM[PSLDBOF] bit + is set. This field when set will not change again + unitl NPEI_INTERRUPT_SUM[PSLDBOF] is cleared. */ +#else + uint64_t psldbof : 6; + uint64_t pidbof : 6; + uint64_t reserved_12_63 : 52; +#endif + } s; + struct cvmx_npei_int_info_s cn52xx; + struct cvmx_npei_int_info_s cn56xx; + struct cvmx_npei_int_info_s cn56xxp1; +}; +typedef union cvmx_npei_int_info cvmx_npei_int_info_t; + +/** + * cvmx_npei_int_sum + * + * NPEI_INTERRUPT_SUM = NPI Interrupt Summary Register + * + * Set when an interrupt condition occurs, write '1' to clear. + * + * HACK: These used to exist, how are TO handled? + * <3> PO0_2SML R/W1C 0x0 0 The packet being sent out on Port0 is smaller $R NS + * than the NPI_BUFF_SIZE_OUTPUT0[ISIZE] field. + * <7> I0_RTOUT R/W1C 0x0 0 Port-0 had a read timeout while attempting to $R NS + * read instructions. + * <15> P0_RTOUT R/W1C 0x0 0 Port-0 had a read timeout while attempting to $R NS + * read packet data. + * <23> G0_RTOUT R/W1C 0x0 0 Port-0 had a read timeout while attempting to $R NS + * read a gather list. + * <31> P0_PTOUT R/W1C 0x0 0 Port-0 output had a read timeout on a DATA/INFO $R NS + * pair. + */ +union cvmx_npei_int_sum +{ + uint64_t u64; + struct cvmx_npei_int_sum_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t mio_inta : 1; /**< Interrupt from MIO. */ + uint64_t reserved_62_62 : 1; + uint64_t int_a : 1; /**< Set when a bit in the NPEI_INT_A_SUM register and + the cooresponding bit in the NPEI_INT_A_ENB + register is set. */ + uint64_t c1_ldwn : 1; /**< Reset request due to link1 down status. */ + uint64_t c0_ldwn : 1; /**< Reset request due to link0 down status. */ + uint64_t c1_exc : 1; /**< Set when the PESC1_DBG_INFO register has a bit + set and its cooresponding PESC1_DBG_INFO_EN bit + is set. */ + uint64_t c0_exc : 1; /**< Set when the PESC0_DBG_INFO register has a bit + set and its cooresponding PESC0_DBG_INFO_EN bit + is set. */ + uint64_t c1_up_wf : 1; /**< Received Unsupported P-TLP for filtered window + register. Core1. */ + uint64_t c0_up_wf : 1; /**< Received Unsupported P-TLP for filtered window + register. Core0. */ + uint64_t c1_un_wf : 1; /**< Received Unsupported N-TLP for filtered window + register. Core1. */ + uint64_t c0_un_wf : 1; /**< Received Unsupported N-TLP for filtered window + register. Core0. */ + uint64_t c1_un_bx : 1; /**< Received Unsupported N-TLP for unknown Bar. + Core 1. */ + uint64_t c1_un_wi : 1; /**< Received Unsupported N-TLP for Window Register. + Core 1. */ + uint64_t c1_un_b2 : 1; /**< Received Unsupported N-TLP for Bar2. + Core 1. */ + uint64_t c1_un_b1 : 1; /**< Received Unsupported N-TLP for Bar1. + Core 1. */ + uint64_t c1_un_b0 : 1; /**< Received Unsupported N-TLP for Bar0. + Core 1. */ + uint64_t c1_up_bx : 1; /**< Received Unsupported P-TLP for unknown Bar. + Core 1. */ + uint64_t c1_up_wi : 1; /**< Received Unsupported P-TLP for Window Register. + Core 1. */ + uint64_t c1_up_b2 : 1; /**< Received Unsupported P-TLP for Bar2. + Core 1. */ + uint64_t c1_up_b1 : 1; /**< Received Unsupported P-TLP for Bar1. + Core 1. */ + uint64_t c1_up_b0 : 1; /**< Received Unsupported P-TLP for Bar0. + Core 1. */ + uint64_t c0_un_bx : 1; /**< Received Unsupported N-TLP for unknown Bar. + Core 0. */ + uint64_t c0_un_wi : 1; /**< Received Unsupported N-TLP for Window Register. + Core 0. */ + uint64_t c0_un_b2 : 1; /**< Received Unsupported N-TLP for Bar2. + Core 0. */ + uint64_t c0_un_b1 : 1; /**< Received Unsupported N-TLP for Bar1. + Core 0. */ + uint64_t c0_un_b0 : 1; /**< Received Unsupported N-TLP for Bar0. + Core 0. */ + uint64_t c0_up_bx : 1; /**< Received Unsupported P-TLP for unknown Bar. + Core 0. */ + uint64_t c0_up_wi : 1; /**< Received Unsupported P-TLP for Window Register. + Core 0. */ + uint64_t c0_up_b2 : 1; /**< Received Unsupported P-TLP for Bar2. + Core 0. */ + uint64_t c0_up_b1 : 1; /**< Received Unsupported P-TLP for Bar1. + Core 0. */ + uint64_t c0_up_b0 : 1; /**< Received Unsupported P-TLP for Bar0. + Core 0. */ + uint64_t c1_hpint : 1; /**< Hot-Plug Interrupt. + Pcie Core 1 (hp_int). + This interrupt will only be generated when + PCIERC1_CFG034[DLLS_C] is generated. Hot plug is + not supported. */ + uint64_t c1_pmei : 1; /**< PME Interrupt. + Pcie Core 1. (cfg_pme_int) */ + uint64_t c1_wake : 1; /**< Wake up from Power Management Unit. + Pcie Core 1. (wake_n) + Octeon will never generate this interrupt. */ + uint64_t crs1_dr : 1; /**< Had a CRS when Retries were disabled. */ + uint64_t c1_se : 1; /**< System Error, RC Mode Only. + Pcie Core 1. (cfg_sys_err_rc) */ + uint64_t crs1_er : 1; /**< Had a CRS Timeout when Retries were enabled. */ + uint64_t c1_aeri : 1; /**< Advanced Error Reporting Interrupt, RC Mode Only. + Pcie Core 1. */ + uint64_t c0_hpint : 1; /**< Hot-Plug Interrupt. + Pcie Core 0 (hp_int). + This interrupt will only be generated when + PCIERC0_CFG034[DLLS_C] is generated. Hot plug is + not supported. */ + uint64_t c0_pmei : 1; /**< PME Interrupt. + Pcie Core 0. (cfg_pme_int) */ + uint64_t c0_wake : 1; /**< Wake up from Power Management Unit. + Pcie Core 0. (wake_n) + Octeon will never generate this interrupt. */ + uint64_t crs0_dr : 1; /**< Had a CRS when Retries were disabled. */ + uint64_t c0_se : 1; /**< System Error, RC Mode Only. + Pcie Core 0. (cfg_sys_err_rc) */ + uint64_t crs0_er : 1; /**< Had a CRS Timeout when Retries were enabled. */ + uint64_t c0_aeri : 1; /**< Advanced Error Reporting Interrupt, RC Mode Only. + Pcie Core 0 (cfg_aer_rc_err_int). */ + uint64_t ptime : 1; /**< Packet Timer has an interrupt. Which rings can + be found in NPEI_PKT_TIME_INT. */ + uint64_t pcnt : 1; /**< Packet Counter has an interrupt. Which rings can + be found in NPEI_PKT_CNT_INT. */ + uint64_t pidbof : 1; /**< Packet Instruction Doorbell count overflowed. Which + doorbell can be found in NPEI_INT_INFO[PIDBOF] */ + uint64_t psldbof : 1; /**< Packet Scatterlist Doorbell count overflowed. Which + doorbell can be found in NPEI_INT_INFO[PSLDBOF] */ + uint64_t dtime1 : 1; /**< Whenever NPEI_DMA_CNTS[DMA1] is not 0, the + DMA_CNT1 timer increments every core clock. When + DMA_CNT1 timer exceeds NPEI_DMA1_INT_LEVEL[TIME], + this bit is set. Writing a '1' to this bit also + clears the DMA_CNT1 timer. */ + uint64_t dtime0 : 1; /**< Whenever NPEI_DMA_CNTS[DMA0] is not 0, the + DMA_CNT0 timer increments every core clock. When + DMA_CNT0 timer exceeds NPEI_DMA0_INT_LEVEL[TIME], + this bit is set. Writing a '1' to this bit also + clears the DMA_CNT0 timer. */ + uint64_t dcnt1 : 1; /**< This bit indicates that NPEI_DMA_CNTS[DMA1] was/is + greater than NPEI_DMA1_INT_LEVEL[CNT]. */ + uint64_t dcnt0 : 1; /**< This bit indicates that NPEI_DMA_CNTS[DMA0] was/is + greater than NPEI_DMA0_INT_LEVEL[CNT]. */ + uint64_t dma1fi : 1; /**< DMA0 set Forced Interrupt. */ + uint64_t dma0fi : 1; /**< DMA0 set Forced Interrupt. */ + uint64_t dma4dbo : 1; /**< DMA4 doorbell overflow. + Bit[32] of the doorbell count was set. */ + uint64_t dma3dbo : 1; /**< DMA3 doorbell overflow. + Bit[32] of the doorbell count was set. */ + uint64_t dma2dbo : 1; /**< DMA2 doorbell overflow. + Bit[32] of the doorbell count was set. */ + uint64_t dma1dbo : 1; /**< DMA1 doorbell overflow. + Bit[32] of the doorbell count was set. */ + uint64_t dma0dbo : 1; /**< DMA0 doorbell overflow. + Bit[32] of the doorbell count was set. */ + uint64_t iob2big : 1; /**< A requested IOBDMA is to large. */ + uint64_t bar0_to : 1; /**< BAR0 R/W to a NCB device did not receive + read-data/commit in 0xffff core clocks. */ + uint64_t rml_wto : 1; /**< RML write did not get commit in 0xffff core clocks. */ + uint64_t rml_rto : 1; /**< RML read did not return data in 0xffff core clocks. */ +#else + uint64_t rml_rto : 1; + uint64_t rml_wto : 1; + uint64_t bar0_to : 1; + uint64_t iob2big : 1; + uint64_t dma0dbo : 1; + uint64_t dma1dbo : 1; + uint64_t dma2dbo : 1; + uint64_t dma3dbo : 1; + uint64_t dma4dbo : 1; + uint64_t dma0fi : 1; + uint64_t dma1fi : 1; + uint64_t dcnt0 : 1; + uint64_t dcnt1 : 1; + uint64_t dtime0 : 1; + uint64_t dtime1 : 1; + uint64_t psldbof : 1; + uint64_t pidbof : 1; + uint64_t pcnt : 1; + uint64_t ptime : 1; + uint64_t c0_aeri : 1; + uint64_t crs0_er : 1; + uint64_t c0_se : 1; + uint64_t crs0_dr : 1; + uint64_t c0_wake : 1; + uint64_t c0_pmei : 1; + uint64_t c0_hpint : 1; + uint64_t c1_aeri : 1; + uint64_t crs1_er : 1; + uint64_t c1_se : 1; + uint64_t crs1_dr : 1; + uint64_t c1_wake : 1; + uint64_t c1_pmei : 1; + uint64_t c1_hpint : 1; + uint64_t c0_up_b0 : 1; + uint64_t c0_up_b1 : 1; + uint64_t c0_up_b2 : 1; + uint64_t c0_up_wi : 1; + uint64_t c0_up_bx : 1; + uint64_t c0_un_b0 : 1; + uint64_t c0_un_b1 : 1; + uint64_t c0_un_b2 : 1; + uint64_t c0_un_wi : 1; + uint64_t c0_un_bx : 1; + uint64_t c1_up_b0 : 1; + uint64_t c1_up_b1 : 1; + uint64_t c1_up_b2 : 1; + uint64_t c1_up_wi : 1; + uint64_t c1_up_bx : 1; + uint64_t c1_un_b0 : 1; + uint64_t c1_un_b1 : 1; + uint64_t c1_un_b2 : 1; + uint64_t c1_un_wi : 1; + uint64_t c1_un_bx : 1; + uint64_t c0_un_wf : 1; + uint64_t c1_un_wf : 1; + uint64_t c0_up_wf : 1; + uint64_t c1_up_wf : 1; + uint64_t c0_exc : 1; + uint64_t c1_exc : 1; + uint64_t c0_ldwn : 1; + uint64_t c1_ldwn : 1; + uint64_t int_a : 1; + uint64_t reserved_62_62 : 1; + uint64_t mio_inta : 1; +#endif + } s; + struct cvmx_npei_int_sum_s cn52xx; + struct cvmx_npei_int_sum_cn52xxp1 + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t mio_inta : 1; /**< Interrupt from MIO. */ + uint64_t reserved_62_62 : 1; + uint64_t int_a : 1; /**< Set when a bit in the NPEI_INT_A_SUM register and + the cooresponding bit in the NPEI_INT_A_ENB + register is set. */ + uint64_t c1_ldwn : 1; /**< Reset request due to link1 down status. */ + uint64_t c0_ldwn : 1; /**< Reset request due to link0 down status. */ + uint64_t c1_exc : 1; /**< Set when the PESC1_DBG_INFO register has a bit + set and its cooresponding PESC1_DBG_INFO_EN bit + is set. */ + uint64_t c0_exc : 1; /**< Set when the PESC0_DBG_INFO register has a bit + set and its cooresponding PESC0_DBG_INFO_EN bit + is set. */ + uint64_t c1_up_wf : 1; /**< Received Unsupported P-TLP for filtered window + register. Core1. */ + uint64_t c0_up_wf : 1; /**< Received Unsupported P-TLP for filtered window + register. Core0. */ + uint64_t c1_un_wf : 1; /**< Received Unsupported N-TLP for filtered window + register. Core1. */ + uint64_t c0_un_wf : 1; /**< Received Unsupported N-TLP for filtered window + register. Core0. */ + uint64_t c1_un_bx : 1; /**< Received Unsupported N-TLP for unknown Bar. + Core 1. */ + uint64_t c1_un_wi : 1; /**< Received Unsupported N-TLP for Window Register. + Core 1. */ + uint64_t c1_un_b2 : 1; /**< Received Unsupported N-TLP for Bar2. + Core 1. */ + uint64_t c1_un_b1 : 1; /**< Received Unsupported N-TLP for Bar1. + Core 1. */ + uint64_t c1_un_b0 : 1; /**< Received Unsupported N-TLP for Bar0. + Core 1. */ + uint64_t c1_up_bx : 1; /**< Received Unsupported P-TLP for unknown Bar. + Core 1. */ + uint64_t c1_up_wi : 1; /**< Received Unsupported P-TLP for Window Register. + Core 1. */ + uint64_t c1_up_b2 : 1; /**< Received Unsupported P-TLP for Bar2. + Core 1. */ + uint64_t c1_up_b1 : 1; /**< Received Unsupported P-TLP for Bar1. + Core 1. */ + uint64_t c1_up_b0 : 1; /**< Received Unsupported P-TLP for Bar0. + Core 1. */ + uint64_t c0_un_bx : 1; /**< Received Unsupported N-TLP for unknown Bar. + Core 0. */ + uint64_t c0_un_wi : 1; /**< Received Unsupported N-TLP for Window Register. + Core 0. */ + uint64_t c0_un_b2 : 1; /**< Received Unsupported N-TLP for Bar2. + Core 0. */ + uint64_t c0_un_b1 : 1; /**< Received Unsupported N-TLP for Bar1. + Core 0. */ + uint64_t c0_un_b0 : 1; /**< Received Unsupported N-TLP for Bar0. + Core 0. */ + uint64_t c0_up_bx : 1; /**< Received Unsupported P-TLP for unknown Bar. + Core 0. */ + uint64_t c0_up_wi : 1; /**< Received Unsupported P-TLP for Window Register. + Core 0. */ + uint64_t c0_up_b2 : 1; /**< Received Unsupported P-TLP for Bar2. + Core 0. */ + uint64_t c0_up_b1 : 1; /**< Received Unsupported P-TLP for Bar1. + Core 0. */ + uint64_t c0_up_b0 : 1; /**< Received Unsupported P-TLP for Bar0. + Core 0. */ + uint64_t c1_hpint : 1; /**< Hot-Plug Interrupt. + Pcie Core 1 (hp_int). + This interrupt will only be generated when + PCIERC1_CFG034[DLLS_C] is generated. Hot plug is + not supported. */ + uint64_t c1_pmei : 1; /**< PME Interrupt. + Pcie Core 1. (cfg_pme_int) */ + uint64_t c1_wake : 1; /**< Wake up from Power Management Unit. + Pcie Core 1. (wake_n) + Octeon will never generate this interrupt. */ + uint64_t crs1_dr : 1; /**< Had a CRS when Retries were disabled. */ + uint64_t c1_se : 1; /**< System Error, RC Mode Only. + Pcie Core 1. (cfg_sys_err_rc) */ + uint64_t crs1_er : 1; /**< Had a CRS Timeout when Retries were enabled. */ + uint64_t c1_aeri : 1; /**< Advanced Error Reporting Interrupt, RC Mode Only. + Pcie Core 1. */ + uint64_t c0_hpint : 1; /**< Hot-Plug Interrupt. + Pcie Core 0 (hp_int). + This interrupt will only be generated when + PCIERC0_CFG034[DLLS_C] is generated. Hot plug is + not supported. */ + uint64_t c0_pmei : 1; /**< PME Interrupt. + Pcie Core 0. (cfg_pme_int) */ + uint64_t c0_wake : 1; /**< Wake up from Power Management Unit. + Pcie Core 0. (wake_n) + Octeon will never generate this interrupt. */ + uint64_t crs0_dr : 1; /**< Had a CRS when Retries were disabled. */ + uint64_t c0_se : 1; /**< System Error, RC Mode Only. + Pcie Core 0. (cfg_sys_err_rc) */ + uint64_t crs0_er : 1; /**< Had a CRS Timeout when Retries were enabled. */ + uint64_t c0_aeri : 1; /**< Advanced Error Reporting Interrupt, RC Mode Only. + Pcie Core 0 (cfg_aer_rc_err_int). */ + uint64_t reserved_15_18 : 4; + uint64_t dtime1 : 1; /**< Whenever NPEI_DMA_CNTS[DMA1] is not 0, the + DMA_CNT1 timer increments every core clock. When + DMA_CNT1 timer exceeds NPEI_DMA1_INT_LEVEL[TIME], + this bit is set. Writing a '1' to this bit also + clears the DMA_CNT1 timer. */ + uint64_t dtime0 : 1; /**< Whenever NPEI_DMA_CNTS[DMA0] is not 0, the + DMA_CNT0 timer increments every core clock. When + DMA_CNT0 timer exceeds NPEI_DMA0_INT_LEVEL[TIME], + this bit is set. Writing a '1' to this bit also + clears the DMA_CNT0 timer. */ + uint64_t dcnt1 : 1; /**< This bit indicates that NPEI_DMA_CNTS[DMA1] was/is + greater than NPEI_DMA1_INT_LEVEL[CNT]. */ + uint64_t dcnt0 : 1; /**< This bit indicates that NPEI_DMA_CNTS[DMA0] was/is + greater than NPEI_DMA0_INT_LEVEL[CNT]. */ + uint64_t dma1fi : 1; /**< DMA0 set Forced Interrupt. */ + uint64_t dma0fi : 1; /**< DMA0 set Forced Interrupt. */ + uint64_t reserved_8_8 : 1; + uint64_t dma3dbo : 1; /**< DMA3 doorbell count overflow. + Bit[32] of the doorbell count was set. */ + uint64_t dma2dbo : 1; /**< DMA2 doorbell count overflow. + Bit[32] of the doorbell count was set. */ + uint64_t dma1dbo : 1; /**< DMA1 doorbell count overflow. + Bit[32] of the doorbell count was set. */ + uint64_t dma0dbo : 1; /**< DMA0 doorbell count overflow. + Bit[32] of the doorbell count was set. */ + uint64_t iob2big : 1; /**< A requested IOBDMA is to large. */ + uint64_t bar0_to : 1; /**< BAR0 R/W to a NCB device did not receive + read-data/commit in 0xffff core clocks. */ + uint64_t rml_wto : 1; /**< RML write did not get commit in 0xffff core clocks. */ + uint64_t rml_rto : 1; /**< RML read did not return data in 0xffff core clocks. */ +#else + uint64_t rml_rto : 1; + uint64_t rml_wto : 1; + uint64_t bar0_to : 1; + uint64_t iob2big : 1; + uint64_t dma0dbo : 1; + uint64_t dma1dbo : 1; + uint64_t dma2dbo : 1; + uint64_t dma3dbo : 1; + uint64_t reserved_8_8 : 1; + uint64_t dma0fi : 1; + uint64_t dma1fi : 1; + uint64_t dcnt0 : 1; + uint64_t dcnt1 : 1; + uint64_t dtime0 : 1; + uint64_t dtime1 : 1; + uint64_t reserved_15_18 : 4; + uint64_t c0_aeri : 1; + uint64_t crs0_er : 1; + uint64_t c0_se : 1; + uint64_t crs0_dr : 1; + uint64_t c0_wake : 1; + uint64_t c0_pmei : 1; + uint64_t c0_hpint : 1; + uint64_t c1_aeri : 1; + uint64_t crs1_er : 1; + uint64_t c1_se : 1; + uint64_t crs1_dr : 1; + uint64_t c1_wake : 1; + uint64_t c1_pmei : 1; + uint64_t c1_hpint : 1; + uint64_t c0_up_b0 : 1; + uint64_t c0_up_b1 : 1; + uint64_t c0_up_b2 : 1; + uint64_t c0_up_wi : 1; + uint64_t c0_up_bx : 1; + uint64_t c0_un_b0 : 1; + uint64_t c0_un_b1 : 1; + uint64_t c0_un_b2 : 1; + uint64_t c0_un_wi : 1; + uint64_t c0_un_bx : 1; + uint64_t c1_up_b0 : 1; + uint64_t c1_up_b1 : 1; + uint64_t c1_up_b2 : 1; + uint64_t c1_up_wi : 1; + uint64_t c1_up_bx : 1; + uint64_t c1_un_b0 : 1; + uint64_t c1_un_b1 : 1; + uint64_t c1_un_b2 : 1; + uint64_t c1_un_wi : 1; + uint64_t c1_un_bx : 1; + uint64_t c0_un_wf : 1; + uint64_t c1_un_wf : 1; + uint64_t c0_up_wf : 1; + uint64_t c1_up_wf : 1; + uint64_t c0_exc : 1; + uint64_t c1_exc : 1; + uint64_t c0_ldwn : 1; + uint64_t c1_ldwn : 1; + uint64_t int_a : 1; + uint64_t reserved_62_62 : 1; + uint64_t mio_inta : 1; +#endif + } cn52xxp1; + struct cvmx_npei_int_sum_s cn56xx; + struct cvmx_npei_int_sum_cn56xxp1 + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t mio_inta : 1; /**< Interrupt from MIO. */ + uint64_t reserved_61_62 : 2; + uint64_t c1_ldwn : 1; /**< Reset request due to link1 down status. */ + uint64_t c0_ldwn : 1; /**< Reset request due to link0 down status. */ + uint64_t c1_exc : 1; /**< Set when the PESC1_DBG_INFO register has a bit + set and its cooresponding PESC1_DBG_INFO_EN bit + is set. */ + uint64_t c0_exc : 1; /**< Set when the PESC0_DBG_INFO register has a bit + set and its cooresponding PESC0_DBG_INFO_EN bit + is set. */ + uint64_t c1_up_wf : 1; /**< Received Unsupported P-TLP for filtered window + register. Core1. */ + uint64_t c0_up_wf : 1; /**< Received Unsupported P-TLP for filtered window + register. Core0. */ + uint64_t c1_un_wf : 1; /**< Received Unsupported N-TLP for filtered window + register. Core1. */ + uint64_t c0_un_wf : 1; /**< Received Unsupported N-TLP for filtered window + register. Core0. */ + uint64_t c1_un_bx : 1; /**< Received Unsupported N-TLP for unknown Bar. + Core 1. */ + uint64_t c1_un_wi : 1; /**< Received Unsupported N-TLP for Window Register. + Core 1. */ + uint64_t c1_un_b2 : 1; /**< Received Unsupported N-TLP for Bar2. + Core 1. */ + uint64_t c1_un_b1 : 1; /**< Received Unsupported N-TLP for Bar1. + Core 1. */ + uint64_t c1_un_b0 : 1; /**< Received Unsupported N-TLP for Bar0. + Core 1. */ + uint64_t c1_up_bx : 1; /**< Received Unsupported P-TLP for unknown Bar. + Core 1. */ + uint64_t c1_up_wi : 1; /**< Received Unsupported P-TLP for Window Register. + Core 1. */ + uint64_t c1_up_b2 : 1; /**< Received Unsupported P-TLP for Bar2. + Core 1. */ + uint64_t c1_up_b1 : 1; /**< Received Unsupported P-TLP for Bar1. + Core 1. */ + uint64_t c1_up_b0 : 1; /**< Received Unsupported P-TLP for Bar0. + Core 1. */ + uint64_t c0_un_bx : 1; /**< Received Unsupported N-TLP for unknown Bar. + Core 0. */ + uint64_t c0_un_wi : 1; /**< Received Unsupported N-TLP for Window Register. + Core 0. */ + uint64_t c0_un_b2 : 1; /**< Received Unsupported N-TLP for Bar2. + Core 0. */ + uint64_t c0_un_b1 : 1; /**< Received Unsupported N-TLP for Bar1. + Core 0. */ + uint64_t c0_un_b0 : 1; /**< Received Unsupported N-TLP for Bar0. + Core 0. */ + uint64_t c0_up_bx : 1; /**< Received Unsupported P-TLP for unknown Bar. + Core 0. */ + uint64_t c0_up_wi : 1; /**< Received Unsupported P-TLP for Window Register. + Core 0. */ + uint64_t c0_up_b2 : 1; /**< Received Unsupported P-TLP for Bar2. + Core 0. */ + uint64_t c0_up_b1 : 1; /**< Received Unsupported P-TLP for Bar1. + Core 0. */ + uint64_t c0_up_b0 : 1; /**< Received Unsupported P-TLP for Bar0. + Core 0. */ + uint64_t c1_hpint : 1; /**< Hot-Plug Interrupt. + Pcie Core 1 (hp_int). + This interrupt will only be generated when + PCIERC1_CFG034[DLLS_C] is generated. Hot plug is + not supported. */ + uint64_t c1_pmei : 1; /**< PME Interrupt. + Pcie Core 1. (cfg_pme_int) */ + uint64_t c1_wake : 1; /**< Wake up from Power Management Unit. + Pcie Core 1. (wake_n) + Octeon will never generate this interrupt. */ + uint64_t reserved_29_29 : 1; + uint64_t c1_se : 1; /**< System Error, RC Mode Only. + Pcie Core 1. (cfg_sys_err_rc) */ + uint64_t reserved_27_27 : 1; + uint64_t c1_aeri : 1; /**< Advanced Error Reporting Interrupt, RC Mode Only. + Pcie Core 1. */ + uint64_t c0_hpint : 1; /**< Hot-Plug Interrupt. + Pcie Core 0 (hp_int). + This interrupt will only be generated when + PCIERC0_CFG034[DLLS_C] is generated. Hot plug is + not supported. */ + uint64_t c0_pmei : 1; /**< PME Interrupt. + Pcie Core 0. (cfg_pme_int) */ + uint64_t c0_wake : 1; /**< Wake up from Power Management Unit. + Pcie Core 0. (wake_n) + Octeon will never generate this interrupt. */ + uint64_t reserved_22_22 : 1; + uint64_t c0_se : 1; /**< System Error, RC Mode Only. + Pcie Core 0. (cfg_sys_err_rc) */ + uint64_t reserved_20_20 : 1; + uint64_t c0_aeri : 1; /**< Advanced Error Reporting Interrupt, RC Mode Only. + Pcie Core 0 (cfg_aer_rc_err_int). */ + uint64_t reserved_15_18 : 4; + uint64_t dtime1 : 1; /**< Whenever NPEI_DMA_CNTS[DMA1] is not 0, the + DMA_CNT1 timer increments every core clock. When + DMA_CNT1 timer exceeds NPEI_DMA1_INT_LEVEL[TIME], + this bit is set. Writing a '1' to this bit also + clears the DMA_CNT1 timer. */ + uint64_t dtime0 : 1; /**< Whenever NPEI_DMA_CNTS[DMA0] is not 0, the + DMA_CNT0 timer increments every core clock. When + DMA_CNT0 timer exceeds NPEI_DMA0_INT_LEVEL[TIME], + this bit is set. Writing a '1' to this bit also + clears the DMA_CNT0 timer. */ + uint64_t dcnt1 : 1; /**< This bit indicates that NPEI_DMA_CNTS[DMA1] was/is + greater than NPEI_DMA1_INT_LEVEL[CNT]. */ + uint64_t dcnt0 : 1; /**< This bit indicates that NPEI_DMA_CNTS[DMA0] was/is + greater than NPEI_DMA0_INT_LEVEL[CNT]. */ + uint64_t dma1fi : 1; /**< DMA0 set Forced Interrupt. */ + uint64_t dma0fi : 1; /**< DMA0 set Forced Interrupt. */ + uint64_t dma4dbo : 1; /**< DMA4 doorbell overflow. + Bit[32] of the doorbell count was set. */ + uint64_t dma3dbo : 1; /**< DMA3 doorbell overflow. + Bit[32] of the doorbell count was set. */ + uint64_t dma2dbo : 1; /**< DMA2 doorbell overflow. + Bit[32] of the doorbell count was set. */ + uint64_t dma1dbo : 1; /**< DMA1 doorbell overflow. + Bit[32] of the doorbell count was set. */ + uint64_t dma0dbo : 1; /**< DMA0 doorbell overflow. + Bit[32] of the doorbell count was set. */ + uint64_t iob2big : 1; /**< A requested IOBDMA is to large. */ + uint64_t bar0_to : 1; /**< BAR0 R/W to a NCB device did not receive + read-data/commit in 0xffff core clocks. */ + uint64_t rml_wto : 1; /**< RML write did not get commit in 0xffff core clocks. */ + uint64_t rml_rto : 1; /**< RML read did not return data in 0xffff core clocks. */ +#else + uint64_t rml_rto : 1; + uint64_t rml_wto : 1; + uint64_t bar0_to : 1; + uint64_t iob2big : 1; + uint64_t dma0dbo : 1; + uint64_t dma1dbo : 1; + uint64_t dma2dbo : 1; + uint64_t dma3dbo : 1; + uint64_t dma4dbo : 1; + uint64_t dma0fi : 1; + uint64_t dma1fi : 1; + uint64_t dcnt0 : 1; + uint64_t dcnt1 : 1; + uint64_t dtime0 : 1; + uint64_t dtime1 : 1; + uint64_t reserved_15_18 : 4; + uint64_t c0_aeri : 1; + uint64_t reserved_20_20 : 1; + uint64_t c0_se : 1; + uint64_t reserved_22_22 : 1; + uint64_t c0_wake : 1; + uint64_t c0_pmei : 1; + uint64_t c0_hpint : 1; + uint64_t c1_aeri : 1; + uint64_t reserved_27_27 : 1; + uint64_t c1_se : 1; + uint64_t reserved_29_29 : 1; + uint64_t c1_wake : 1; + uint64_t c1_pmei : 1; + uint64_t c1_hpint : 1; + uint64_t c0_up_b0 : 1; + uint64_t c0_up_b1 : 1; + uint64_t c0_up_b2 : 1; + uint64_t c0_up_wi : 1; + uint64_t c0_up_bx : 1; + uint64_t c0_un_b0 : 1; + uint64_t c0_un_b1 : 1; + uint64_t c0_un_b2 : 1; + uint64_t c0_un_wi : 1; + uint64_t c0_un_bx : 1; + uint64_t c1_up_b0 : 1; + uint64_t c1_up_b1 : 1; + uint64_t c1_up_b2 : 1; + uint64_t c1_up_wi : 1; + uint64_t c1_up_bx : 1; + uint64_t c1_un_b0 : 1; + uint64_t c1_un_b1 : 1; + uint64_t c1_un_b2 : 1; + uint64_t c1_un_wi : 1; + uint64_t c1_un_bx : 1; + uint64_t c0_un_wf : 1; + uint64_t c1_un_wf : 1; + uint64_t c0_up_wf : 1; + uint64_t c1_up_wf : 1; + uint64_t c0_exc : 1; + uint64_t c1_exc : 1; + uint64_t c0_ldwn : 1; + uint64_t c1_ldwn : 1; + uint64_t reserved_61_62 : 2; + uint64_t mio_inta : 1; +#endif + } cn56xxp1; +}; +typedef union cvmx_npei_int_sum cvmx_npei_int_sum_t; + +/** + * cvmx_npei_int_sum2 + * + * NPEI_INTERRUPT_SUM2 = NPI Interrupt Summary2 Register + * + * This is a read only copy of the NPEI_INTERRUPT_SUM register with bit variances. + */ +union cvmx_npei_int_sum2 +{ + uint64_t u64; + struct cvmx_npei_int_sum2_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t mio_inta : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t reserved_62_62 : 1; + uint64_t int_a : 1; /**< Set when a bit in the NPEI_INT_A_SUM register and + the cooresponding bit in the NPEI_INT_A_ENB2 + register is set. */ + uint64_t c1_ldwn : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c0_ldwn : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c1_exc : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c0_exc : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c1_up_wf : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c0_up_wf : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c1_un_wf : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c0_un_wf : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c1_un_bx : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c1_un_wi : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c1_un_b2 : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c1_un_b1 : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c1_un_b0 : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c1_up_bx : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c1_up_wi : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c1_up_b2 : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c1_up_b1 : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c1_up_b0 : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c0_un_bx : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c0_un_wi : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c0_un_b2 : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c0_un_b1 : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c0_un_b0 : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c0_up_bx : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c0_up_wi : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c0_up_b2 : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c0_up_b1 : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c0_up_b0 : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c1_hpint : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c1_pmei : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c1_wake : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t crs1_dr : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c1_se : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t crs1_er : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c1_aeri : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c0_hpint : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c0_pmei : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c0_wake : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t crs0_dr : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c0_se : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t crs0_er : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t c0_aeri : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t reserved_15_18 : 4; + uint64_t dtime1 : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t dtime0 : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t dcnt1 : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t dcnt0 : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t dma1fi : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t dma0fi : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t reserved_8_8 : 1; + uint64_t dma3dbo : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t dma2dbo : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t dma1dbo : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t dma0dbo : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t iob2big : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t bar0_to : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t rml_wto : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ + uint64_t rml_rto : 1; /**< Equal to the cooresponding bit if the + NPEI_INT_SUM register. */ +#else + uint64_t rml_rto : 1; + uint64_t rml_wto : 1; + uint64_t bar0_to : 1; + uint64_t iob2big : 1; + uint64_t dma0dbo : 1; + uint64_t dma1dbo : 1; + uint64_t dma2dbo : 1; + uint64_t dma3dbo : 1; + uint64_t reserved_8_8 : 1; + uint64_t dma0fi : 1; + uint64_t dma1fi : 1; + uint64_t dcnt0 : 1; + uint64_t dcnt1 : 1; + uint64_t dtime0 : 1; + uint64_t dtime1 : 1; + uint64_t reserved_15_18 : 4; + uint64_t c0_aeri : 1; + uint64_t crs0_er : 1; + uint64_t c0_se : 1; + uint64_t crs0_dr : 1; + uint64_t c0_wake : 1; + uint64_t c0_pmei : 1; + uint64_t c0_hpint : 1; + uint64_t c1_aeri : 1; + uint64_t crs1_er : 1; + uint64_t c1_se : 1; + uint64_t crs1_dr : 1; + uint64_t c1_wake : 1; + uint64_t c1_pmei : 1; + uint64_t c1_hpint : 1; + uint64_t c0_up_b0 : 1; + uint64_t c0_up_b1 : 1; + uint64_t c0_up_b2 : 1; + uint64_t c0_up_wi : 1; + uint64_t c0_up_bx : 1; + uint64_t c0_un_b0 : 1; + uint64_t c0_un_b1 : 1; + uint64_t c0_un_b2 : 1; + uint64_t c0_un_wi : 1; + uint64_t c0_un_bx : 1; + uint64_t c1_up_b0 : 1; + uint64_t c1_up_b1 : 1; + uint64_t c1_up_b2 : 1; + uint64_t c1_up_wi : 1; + uint64_t c1_up_bx : 1; + uint64_t c1_un_b0 : 1; + uint64_t c1_un_b1 : 1; + uint64_t c1_un_b2 : 1; + uint64_t c1_un_wi : 1; + uint64_t c1_un_bx : 1; + uint64_t c0_un_wf : 1; + uint64_t c1_un_wf : 1; + uint64_t c0_up_wf : 1; + uint64_t c1_up_wf : 1; + uint64_t c0_exc : 1; + uint64_t c1_exc : 1; + uint64_t c0_ldwn : 1; + uint64_t c1_ldwn : 1; + uint64_t int_a : 1; + uint64_t reserved_62_62 : 1; + uint64_t mio_inta : 1; +#endif + } s; + struct cvmx_npei_int_sum2_s cn52xx; + struct cvmx_npei_int_sum2_s cn52xxp1; + struct cvmx_npei_int_sum2_s cn56xx; +}; +typedef union cvmx_npei_int_sum2 cvmx_npei_int_sum2_t; + +/** + * cvmx_npei_last_win_rdata0 + * + * NPEI_LAST_WIN_RDATA0 = NPEI Last Window Read Data Port0 + * + * The data from the last initiated window read. + */ +union cvmx_npei_last_win_rdata0 +{ + uint64_t u64; + struct cvmx_npei_last_win_rdata0_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t data : 64; /**< Last window read data. */ +#else + uint64_t data : 64; +#endif + } s; + struct cvmx_npei_last_win_rdata0_s cn52xx; + struct cvmx_npei_last_win_rdata0_s cn52xxp1; + struct cvmx_npei_last_win_rdata0_s cn56xx; + struct cvmx_npei_last_win_rdata0_s cn56xxp1; +}; +typedef union cvmx_npei_last_win_rdata0 cvmx_npei_last_win_rdata0_t; + +/** + * cvmx_npei_last_win_rdata1 + * + * NPEI_LAST_WIN_RDATA1 = NPEI Last Window Read Data Port1 + * + * The data from the last initiated window read. + */ +union cvmx_npei_last_win_rdata1 +{ + uint64_t u64; + struct cvmx_npei_last_win_rdata1_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t data : 64; /**< Last window read data. */ +#else + uint64_t data : 64; +#endif + } s; + struct cvmx_npei_last_win_rdata1_s cn52xx; + struct cvmx_npei_last_win_rdata1_s cn52xxp1; + struct cvmx_npei_last_win_rdata1_s cn56xx; + struct cvmx_npei_last_win_rdata1_s cn56xxp1; +}; +typedef union cvmx_npei_last_win_rdata1 cvmx_npei_last_win_rdata1_t; + +/** + * cvmx_npei_mem_access_ctl + * + * NPEI_MEM_ACCESS_CTL = NPEI's Memory Access Control + * + * Contains control for access to the PCIe address space. + */ +union cvmx_npei_mem_access_ctl +{ + uint64_t u64; + struct cvmx_npei_mem_access_ctl_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_14_63 : 50; + uint64_t max_word : 4; /**< The maximum number of words to merge into a single + write operation from the PPs to the PCIe. Legal + values are 1 to 16, where a '0' is treated as 16. */ + uint64_t timer : 10; /**< When the NPEI starts a PP to PCIe write it waits + no longer than the value of TIMER in eclks to + merge additional writes from the PPs into 1 + large write. The values for this field is 1 to + 1024 where a value of '0' is treated as 1024. */ +#else + uint64_t timer : 10; + uint64_t max_word : 4; + uint64_t reserved_14_63 : 50; +#endif + } s; + struct cvmx_npei_mem_access_ctl_s cn52xx; + struct cvmx_npei_mem_access_ctl_s cn52xxp1; + struct cvmx_npei_mem_access_ctl_s cn56xx; + struct cvmx_npei_mem_access_ctl_s cn56xxp1; +}; +typedef union cvmx_npei_mem_access_ctl cvmx_npei_mem_access_ctl_t; + +/** + * cvmx_npei_mem_access_subid# + * + * NPEI_MEM_ACCESS_SUBIDX = NPEI Memory Access SubidX Register + * + * Contains address index and control bits for access to memory from Core PPs. + */ +union cvmx_npei_mem_access_subidx +{ + uint64_t u64; + struct cvmx_npei_mem_access_subidx_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_42_63 : 22; + uint64_t zero : 1; /**< Causes all byte reads to be zero length reads. + Returns to the EXEC a zero for all read data. */ + uint64_t port : 2; /**< Port the request is sent to. */ + uint64_t nmerge : 1; /**< No merging is allowed in this window. */ + uint64_t esr : 2; /**< Endian-swap for Reads. */ + uint64_t esw : 2; /**< Endian-swap for Writes. */ + uint64_t nsr : 1; /**< No Snoop for Reads. */ + uint64_t nsw : 1; /**< No Snoop for Writes. */ + uint64_t ror : 1; /**< Relaxed Ordering for Reads. */ + uint64_t row : 1; /**< Relaxed Ordering for Writes. */ + uint64_t ba : 30; /**< PCIe Adddress Bits <63:34>. */ +#else + uint64_t ba : 30; + uint64_t row : 1; + uint64_t ror : 1; + uint64_t nsw : 1; + uint64_t nsr : 1; + uint64_t esw : 2; + uint64_t esr : 2; + uint64_t nmerge : 1; + uint64_t port : 2; + uint64_t zero : 1; + uint64_t reserved_42_63 : 22; +#endif + } s; + struct cvmx_npei_mem_access_subidx_s cn52xx; + struct cvmx_npei_mem_access_subidx_s cn52xxp1; + struct cvmx_npei_mem_access_subidx_s cn56xx; + struct cvmx_npei_mem_access_subidx_s cn56xxp1; +}; +typedef union cvmx_npei_mem_access_subidx cvmx_npei_mem_access_subidx_t; + +/** + * cvmx_npei_msi_enb0 + * + * NPEI_MSI_ENB0 = NPEI MSI Enable0 + * + * Used to enable the interrupt generation for the bits in the NPEI_MSI_RCV0. + */ +union cvmx_npei_msi_enb0 +{ + uint64_t u64; + struct cvmx_npei_msi_enb0_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t enb : 64; /**< Enables bit [63:0] of NPEI_MSI_RCV0. */ +#else + uint64_t enb : 64; +#endif + } s; + struct cvmx_npei_msi_enb0_s cn52xx; + struct cvmx_npei_msi_enb0_s cn52xxp1; + struct cvmx_npei_msi_enb0_s cn56xx; + struct cvmx_npei_msi_enb0_s cn56xxp1; +}; +typedef union cvmx_npei_msi_enb0 cvmx_npei_msi_enb0_t; + +/** + * cvmx_npei_msi_enb1 + * + * NPEI_MSI_ENB1 = NPEI MSI Enable1 + * + * Used to enable the interrupt generation for the bits in the NPEI_MSI_RCV1. + */ +union cvmx_npei_msi_enb1 +{ + uint64_t u64; + struct cvmx_npei_msi_enb1_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t enb : 64; /**< Enables bit [63:0] of NPEI_MSI_RCV1. */ +#else + uint64_t enb : 64; +#endif + } s; + struct cvmx_npei_msi_enb1_s cn52xx; + struct cvmx_npei_msi_enb1_s cn52xxp1; + struct cvmx_npei_msi_enb1_s cn56xx; + struct cvmx_npei_msi_enb1_s cn56xxp1; +}; +typedef union cvmx_npei_msi_enb1 cvmx_npei_msi_enb1_t; + +/** + * cvmx_npei_msi_enb2 + * + * NPEI_MSI_ENB2 = NPEI MSI Enable2 + * + * Used to enable the interrupt generation for the bits in the NPEI_MSI_RCV2. + */ +union cvmx_npei_msi_enb2 +{ + uint64_t u64; + struct cvmx_npei_msi_enb2_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t enb : 64; /**< Enables bit [63:0] of NPEI_MSI_RCV2. */ +#else + uint64_t enb : 64; +#endif + } s; + struct cvmx_npei_msi_enb2_s cn52xx; + struct cvmx_npei_msi_enb2_s cn52xxp1; + struct cvmx_npei_msi_enb2_s cn56xx; + struct cvmx_npei_msi_enb2_s cn56xxp1; +}; +typedef union cvmx_npei_msi_enb2 cvmx_npei_msi_enb2_t; + +/** + * cvmx_npei_msi_enb3 + * + * NPEI_MSI_ENB3 = NPEI MSI Enable3 + * + * Used to enable the interrupt generation for the bits in the NPEI_MSI_RCV3. + */ +union cvmx_npei_msi_enb3 +{ + uint64_t u64; + struct cvmx_npei_msi_enb3_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t enb : 64; /**< Enables bit [63:0] of NPEI_MSI_RCV3. */ +#else + uint64_t enb : 64; +#endif + } s; + struct cvmx_npei_msi_enb3_s cn52xx; + struct cvmx_npei_msi_enb3_s cn52xxp1; + struct cvmx_npei_msi_enb3_s cn56xx; + struct cvmx_npei_msi_enb3_s cn56xxp1; +}; +typedef union cvmx_npei_msi_enb3 cvmx_npei_msi_enb3_t; + +/** + * cvmx_npei_msi_rcv0 + * + * NPEI_MSI_RCV0 = NPEI MSI Receive0 + * + * Contains bits [63:0] of the 256 bits oof MSI interrupts. + */ +union cvmx_npei_msi_rcv0 +{ + uint64_t u64; + struct cvmx_npei_msi_rcv0_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t intr : 64; /**< Bits 63-0 of the 256 bits of MSI interrupt. */ +#else + uint64_t intr : 64; +#endif + } s; + struct cvmx_npei_msi_rcv0_s cn52xx; + struct cvmx_npei_msi_rcv0_s cn52xxp1; + struct cvmx_npei_msi_rcv0_s cn56xx; + struct cvmx_npei_msi_rcv0_s cn56xxp1; +}; +typedef union cvmx_npei_msi_rcv0 cvmx_npei_msi_rcv0_t; + +/** + * cvmx_npei_msi_rcv1 + * + * NPEI_MSI_RCV1 = NPEI MSI Receive1 + * + * Contains bits [127:64] of the 256 bits oof MSI interrupts. + */ +union cvmx_npei_msi_rcv1 +{ + uint64_t u64; + struct cvmx_npei_msi_rcv1_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t intr : 64; /**< Bits 127-64 of the 256 bits of MSI interrupt. */ +#else + uint64_t intr : 64; +#endif + } s; + struct cvmx_npei_msi_rcv1_s cn52xx; + struct cvmx_npei_msi_rcv1_s cn52xxp1; + struct cvmx_npei_msi_rcv1_s cn56xx; + struct cvmx_npei_msi_rcv1_s cn56xxp1; +}; +typedef union cvmx_npei_msi_rcv1 cvmx_npei_msi_rcv1_t; + +/** + * cvmx_npei_msi_rcv2 + * + * NPEI_MSI_RCV2 = NPEI MSI Receive2 + * + * Contains bits [191:128] of the 256 bits oof MSI interrupts. + */ +union cvmx_npei_msi_rcv2 +{ + uint64_t u64; + struct cvmx_npei_msi_rcv2_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t intr : 64; /**< Bits 191-128 of the 256 bits of MSI interrupt. */ +#else + uint64_t intr : 64; +#endif + } s; + struct cvmx_npei_msi_rcv2_s cn52xx; + struct cvmx_npei_msi_rcv2_s cn52xxp1; + struct cvmx_npei_msi_rcv2_s cn56xx; + struct cvmx_npei_msi_rcv2_s cn56xxp1; +}; +typedef union cvmx_npei_msi_rcv2 cvmx_npei_msi_rcv2_t; + +/** + * cvmx_npei_msi_rcv3 + * + * NPEI_MSI_RCV3 = NPEI MSI Receive3 + * + * Contains bits [255:192] of the 256 bits oof MSI interrupts. + */ +union cvmx_npei_msi_rcv3 +{ + uint64_t u64; + struct cvmx_npei_msi_rcv3_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t intr : 64; /**< Bits 255-192 of the 256 bits of MSI interrupt. */ +#else + uint64_t intr : 64; +#endif + } s; + struct cvmx_npei_msi_rcv3_s cn52xx; + struct cvmx_npei_msi_rcv3_s cn52xxp1; + struct cvmx_npei_msi_rcv3_s cn56xx; + struct cvmx_npei_msi_rcv3_s cn56xxp1; +}; +typedef union cvmx_npei_msi_rcv3 cvmx_npei_msi_rcv3_t; + +/** + * cvmx_npei_msi_rd_map + * + * NPEI_MSI_RD_MAP = NPEI MSI Read MAP + * + * Used to read the mapping function of the NPEI_PCIE_MSI_RCV to NPEI_MSI_RCV registers. + */ +union cvmx_npei_msi_rd_map +{ + uint64_t u64; + struct cvmx_npei_msi_rd_map_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_16_63 : 48; + uint64_t rd_int : 8; /**< The value of the map at the location PREVIOUSLY + written to the MSI_INT field of this register. */ + uint64_t msi_int : 8; /**< Selects the value that would be received when the + NPEI_PCIE_MSI_RCV register is written. */ +#else + uint64_t msi_int : 8; + uint64_t rd_int : 8; + uint64_t reserved_16_63 : 48; +#endif + } s; + struct cvmx_npei_msi_rd_map_s cn52xx; + struct cvmx_npei_msi_rd_map_s cn52xxp1; + struct cvmx_npei_msi_rd_map_s cn56xx; + struct cvmx_npei_msi_rd_map_s cn56xxp1; +}; +typedef union cvmx_npei_msi_rd_map cvmx_npei_msi_rd_map_t; + +/** + * cvmx_npei_msi_w1c_enb0 + * + * NPEI_MSI_W1C_ENB0 = NPEI MSI Write 1 To Clear Enable0 + * + * Used to clear bits in NPEI_MSI_ENB0. This is a PASS2 register. + */ +union cvmx_npei_msi_w1c_enb0 +{ + uint64_t u64; + struct cvmx_npei_msi_w1c_enb0_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t clr : 64; /**< A write of '1' to a vector will clear the + cooresponding bit in NPEI_MSI_ENB0. + A read to this address will return 0. */ +#else + uint64_t clr : 64; +#endif + } s; + struct cvmx_npei_msi_w1c_enb0_s cn52xx; + struct cvmx_npei_msi_w1c_enb0_s cn56xx; +}; +typedef union cvmx_npei_msi_w1c_enb0 cvmx_npei_msi_w1c_enb0_t; + +/** + * cvmx_npei_msi_w1c_enb1 + * + * NPEI_MSI_W1C_ENB1 = NPEI MSI Write 1 To Clear Enable1 + * + * Used to clear bits in NPEI_MSI_ENB1. This is a PASS2 register. + */ +union cvmx_npei_msi_w1c_enb1 +{ + uint64_t u64; + struct cvmx_npei_msi_w1c_enb1_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t clr : 64; /**< A write of '1' to a vector will clear the + cooresponding bit in NPEI_MSI_ENB1. + A read to this address will return 0. */ +#else + uint64_t clr : 64; +#endif + } s; + struct cvmx_npei_msi_w1c_enb1_s cn52xx; + struct cvmx_npei_msi_w1c_enb1_s cn56xx; +}; +typedef union cvmx_npei_msi_w1c_enb1 cvmx_npei_msi_w1c_enb1_t; + +/** + * cvmx_npei_msi_w1c_enb2 + * + * NPEI_MSI_W1C_ENB2 = NPEI MSI Write 1 To Clear Enable2 + * + * Used to clear bits in NPEI_MSI_ENB2. This is a PASS2 register. + */ +union cvmx_npei_msi_w1c_enb2 +{ + uint64_t u64; + struct cvmx_npei_msi_w1c_enb2_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t clr : 64; /**< A write of '1' to a vector will clear the + cooresponding bit in NPEI_MSI_ENB2. + A read to this address will return 0. */ +#else + uint64_t clr : 64; +#endif + } s; + struct cvmx_npei_msi_w1c_enb2_s cn52xx; + struct cvmx_npei_msi_w1c_enb2_s cn56xx; +}; +typedef union cvmx_npei_msi_w1c_enb2 cvmx_npei_msi_w1c_enb2_t; + +/** + * cvmx_npei_msi_w1c_enb3 + * + * NPEI_MSI_W1C_ENB3 = NPEI MSI Write 1 To Clear Enable3 + * + * Used to clear bits in NPEI_MSI_ENB3. This is a PASS2 register. + */ +union cvmx_npei_msi_w1c_enb3 +{ + uint64_t u64; + struct cvmx_npei_msi_w1c_enb3_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t clr : 64; /**< A write of '1' to a vector will clear the + cooresponding bit in NPEI_MSI_ENB3. + A read to this address will return 0. */ +#else + uint64_t clr : 64; +#endif + } s; + struct cvmx_npei_msi_w1c_enb3_s cn52xx; + struct cvmx_npei_msi_w1c_enb3_s cn56xx; +}; +typedef union cvmx_npei_msi_w1c_enb3 cvmx_npei_msi_w1c_enb3_t; + +/** + * cvmx_npei_msi_w1s_enb0 + * + * NPEI_MSI_W1S_ENB0 = NPEI MSI Write 1 To Set Enable0 + * + * Used to set bits in NPEI_MSI_ENB0. This is a PASS2 register. + */ +union cvmx_npei_msi_w1s_enb0 +{ + uint64_t u64; + struct cvmx_npei_msi_w1s_enb0_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t set : 64; /**< A write of '1' to a vector will set the + cooresponding bit in NPEI_MSI_ENB0. + A read to this address will return 0. */ +#else + uint64_t set : 64; +#endif + } s; + struct cvmx_npei_msi_w1s_enb0_s cn52xx; + struct cvmx_npei_msi_w1s_enb0_s cn56xx; +}; +typedef union cvmx_npei_msi_w1s_enb0 cvmx_npei_msi_w1s_enb0_t; + +/** + * cvmx_npei_msi_w1s_enb1 + * + * NPEI_MSI_W1S_ENB0 = NPEI MSI Write 1 To Set Enable1 + * + * Used to set bits in NPEI_MSI_ENB1. This is a PASS2 register. + */ +union cvmx_npei_msi_w1s_enb1 +{ + uint64_t u64; + struct cvmx_npei_msi_w1s_enb1_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t set : 64; /**< A write of '1' to a vector will set the + cooresponding bit in NPEI_MSI_ENB1. + A read to this address will return 0. */ +#else + uint64_t set : 64; +#endif + } s; + struct cvmx_npei_msi_w1s_enb1_s cn52xx; + struct cvmx_npei_msi_w1s_enb1_s cn56xx; +}; +typedef union cvmx_npei_msi_w1s_enb1 cvmx_npei_msi_w1s_enb1_t; + +/** + * cvmx_npei_msi_w1s_enb2 + * + * NPEI_MSI_W1S_ENB2 = NPEI MSI Write 1 To Set Enable2 + * + * Used to set bits in NPEI_MSI_ENB2. This is a PASS2 register. + */ +union cvmx_npei_msi_w1s_enb2 +{ + uint64_t u64; + struct cvmx_npei_msi_w1s_enb2_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t set : 64; /**< A write of '1' to a vector will set the + cooresponding bit in NPEI_MSI_ENB2. + A read to this address will return 0. */ +#else + uint64_t set : 64; +#endif + } s; + struct cvmx_npei_msi_w1s_enb2_s cn52xx; + struct cvmx_npei_msi_w1s_enb2_s cn56xx; +}; +typedef union cvmx_npei_msi_w1s_enb2 cvmx_npei_msi_w1s_enb2_t; + +/** + * cvmx_npei_msi_w1s_enb3 + * + * NPEI_MSI_W1S_ENB3 = NPEI MSI Write 1 To Set Enable3 + * + * Used to set bits in NPEI_MSI_ENB3. This is a PASS2 register. + */ +union cvmx_npei_msi_w1s_enb3 +{ + uint64_t u64; + struct cvmx_npei_msi_w1s_enb3_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t set : 64; /**< A write of '1' to a vector will set the + cooresponding bit in NPEI_MSI_ENB3. + A read to this address will return 0. */ +#else + uint64_t set : 64; +#endif + } s; + struct cvmx_npei_msi_w1s_enb3_s cn52xx; + struct cvmx_npei_msi_w1s_enb3_s cn56xx; +}; +typedef union cvmx_npei_msi_w1s_enb3 cvmx_npei_msi_w1s_enb3_t; + +/** + * cvmx_npei_msi_wr_map + * + * NPEI_MSI_WR_MAP = NPEI MSI Write MAP + * + * Used to write the mapping function of the NPEI_PCIE_MSI_RCV to NPEI_MSI_RCV registers. + */ +union cvmx_npei_msi_wr_map +{ + uint64_t u64; + struct cvmx_npei_msi_wr_map_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_16_63 : 48; + uint64_t ciu_int : 8; /**< Selects which bit in the NPEI_MSI_RCV# (0-255) + will be set when the value specified in the + MSI_INT of this register is recevied during a + write to the NPEI_PCIE_MSI_RCV register. */ + uint64_t msi_int : 8; /**< Selects the value that would be received when the + NPEI_PCIE_MSI_RCV register is written. */ +#else + uint64_t msi_int : 8; + uint64_t ciu_int : 8; + uint64_t reserved_16_63 : 48; +#endif + } s; + struct cvmx_npei_msi_wr_map_s cn52xx; + struct cvmx_npei_msi_wr_map_s cn52xxp1; + struct cvmx_npei_msi_wr_map_s cn56xx; + struct cvmx_npei_msi_wr_map_s cn56xxp1; +}; +typedef union cvmx_npei_msi_wr_map cvmx_npei_msi_wr_map_t; + +/** + * cvmx_npei_pcie_credit_cnt + * + * NPEI_PCIE_CREDIT_CNT = NPEI PCIE Credit Count + * + * Contains the number of credits for the pcie port FIFOs used by the NPEI. This value needs to be set BEFORE PCIe traffic + * flow from NPEI to PCIE Ports starts. A write to this register will cause the credit counts in the NPEI for the two + * PCIE ports to be reset to the value in this register. + */ +union cvmx_npei_pcie_credit_cnt +{ + uint64_t u64; + struct cvmx_npei_pcie_credit_cnt_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_48_63 : 16; + uint64_t p1_ccnt : 8; /**< Port1 C-TLP FIFO Credits. + Legal values are 0x25 to 0x80. */ + uint64_t p1_ncnt : 8; /**< Port1 N-TLP FIFO Credits. + Legal values are 0x5 to 0x10. */ + uint64_t p1_pcnt : 8; /**< Port1 P-TLP FIFO Credits. + Legal values are 0x25 to 0x80. */ + uint64_t p0_ccnt : 8; /**< Port0 C-TLP FIFO Credits. + Legal values are 0x25 to 0x80. */ + uint64_t p0_ncnt : 8; /**< Port0 N-TLP FIFO Credits. + Legal values are 0x5 to 0x10. */ + uint64_t p0_pcnt : 8; /**< Port0 P-TLP FIFO Credits. + Legal values are 0x25 to 0x80. */ +#else + uint64_t p0_pcnt : 8; + uint64_t p0_ncnt : 8; + uint64_t p0_ccnt : 8; + uint64_t p1_pcnt : 8; + uint64_t p1_ncnt : 8; + uint64_t p1_ccnt : 8; + uint64_t reserved_48_63 : 16; +#endif + } s; + struct cvmx_npei_pcie_credit_cnt_s cn52xx; + struct cvmx_npei_pcie_credit_cnt_s cn56xx; +}; +typedef union cvmx_npei_pcie_credit_cnt cvmx_npei_pcie_credit_cnt_t; + +/** + * cvmx_npei_pcie_msi_rcv + * + * NPEI_PCIE_MSI_RCV = NPEI PCIe MSI Receive + * + * Register where MSI writes are directed from the PCIe. + */ +union cvmx_npei_pcie_msi_rcv +{ + uint64_t u64; + struct cvmx_npei_pcie_msi_rcv_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_8_63 : 56; + uint64_t intr : 8; /**< A write to this register will result in a bit in + one of the NPEI_MSI_RCV# registers being set. + Which bit is set is dependent on the previously + written using the NPEI_MSI_WR_MAP register or if + not previously written the reset value of the MAP. */ +#else + uint64_t intr : 8; + uint64_t reserved_8_63 : 56; +#endif + } s; + struct cvmx_npei_pcie_msi_rcv_s cn52xx; + struct cvmx_npei_pcie_msi_rcv_s cn52xxp1; + struct cvmx_npei_pcie_msi_rcv_s cn56xx; + struct cvmx_npei_pcie_msi_rcv_s cn56xxp1; +}; +typedef union cvmx_npei_pcie_msi_rcv cvmx_npei_pcie_msi_rcv_t; + +/** + * cvmx_npei_pcie_msi_rcv_b1 + * + * NPEI_PCIE_MSI_RCV_B1 = NPEI PCIe MSI Receive Byte 1 + * + * Register where MSI writes are directed from the PCIe. + */ +union cvmx_npei_pcie_msi_rcv_b1 +{ + uint64_t u64; + struct cvmx_npei_pcie_msi_rcv_b1_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_16_63 : 48; + uint64_t intr : 8; /**< A write to this register will result in a bit in + one of the NPEI_MSI_RCV# registers being set. + Which bit is set is dependent on the previously + written using the NPEI_MSI_WR_MAP register or if + not previously written the reset value of the MAP. */ + uint64_t reserved_0_7 : 8; +#else + uint64_t reserved_0_7 : 8; + uint64_t intr : 8; + uint64_t reserved_16_63 : 48; +#endif + } s; + struct cvmx_npei_pcie_msi_rcv_b1_s cn52xx; + struct cvmx_npei_pcie_msi_rcv_b1_s cn52xxp1; + struct cvmx_npei_pcie_msi_rcv_b1_s cn56xx; + struct cvmx_npei_pcie_msi_rcv_b1_s cn56xxp1; +}; +typedef union cvmx_npei_pcie_msi_rcv_b1 cvmx_npei_pcie_msi_rcv_b1_t; + +/** + * cvmx_npei_pcie_msi_rcv_b2 + * + * NPEI_PCIE_MSI_RCV_B2 = NPEI PCIe MSI Receive Byte 2 + * + * Register where MSI writes are directed from the PCIe. + */ +union cvmx_npei_pcie_msi_rcv_b2 +{ + uint64_t u64; + struct cvmx_npei_pcie_msi_rcv_b2_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_24_63 : 40; + uint64_t intr : 8; /**< A write to this register will result in a bit in + one of the NPEI_MSI_RCV# registers being set. + Which bit is set is dependent on the previously + written using the NPEI_MSI_WR_MAP register or if + not previously written the reset value of the MAP. */ + uint64_t reserved_0_15 : 16; +#else + uint64_t reserved_0_15 : 16; + uint64_t intr : 8; + uint64_t reserved_24_63 : 40; +#endif + } s; + struct cvmx_npei_pcie_msi_rcv_b2_s cn52xx; + struct cvmx_npei_pcie_msi_rcv_b2_s cn52xxp1; + struct cvmx_npei_pcie_msi_rcv_b2_s cn56xx; + struct cvmx_npei_pcie_msi_rcv_b2_s cn56xxp1; +}; +typedef union cvmx_npei_pcie_msi_rcv_b2 cvmx_npei_pcie_msi_rcv_b2_t; + +/** + * cvmx_npei_pcie_msi_rcv_b3 + * + * NPEI_PCIE_MSI_RCV_B3 = NPEI PCIe MSI Receive Byte 3 + * + * Register where MSI writes are directed from the PCIe. + */ +union cvmx_npei_pcie_msi_rcv_b3 +{ + uint64_t u64; + struct cvmx_npei_pcie_msi_rcv_b3_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_32_63 : 32; + uint64_t intr : 8; /**< A write to this register will result in a bit in + one of the NPEI_MSI_RCV# registers being set. + Which bit is set is dependent on the previously + written using the NPEI_MSI_WR_MAP register or if + not previously written the reset value of the MAP. */ + uint64_t reserved_0_23 : 24; +#else + uint64_t reserved_0_23 : 24; + uint64_t intr : 8; + uint64_t reserved_32_63 : 32; +#endif + } s; + struct cvmx_npei_pcie_msi_rcv_b3_s cn52xx; + struct cvmx_npei_pcie_msi_rcv_b3_s cn52xxp1; + struct cvmx_npei_pcie_msi_rcv_b3_s cn56xx; + struct cvmx_npei_pcie_msi_rcv_b3_s cn56xxp1; +}; +typedef union cvmx_npei_pcie_msi_rcv_b3 cvmx_npei_pcie_msi_rcv_b3_t; + +/** + * cvmx_npei_pkt#_cnts + * + * NPEI_PKT[0..31]_CNTS = NPEI Packet ring# Counts + * + * The counters for output rings. + */ +union cvmx_npei_pktx_cnts +{ + uint64_t u64; + struct cvmx_npei_pktx_cnts_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_54_63 : 10; + uint64_t timer : 22; /**< Timer incremented every 1024 core clocks + when NPEI_PKTS#_CNTS[CNT] is non zero. Field + cleared when NPEI_PKTS#_CNTS[CNT] goes to 0. + Field is also cleared when NPEI_PKT_TIME_INT is + cleared. + The first increment of this count can occur + between 0 to 1023 core clocks. */ + uint64_t cnt : 32; /**< ring counter. This field is incremented as + packets are sent out and decremented in response to + writes to this field. + When NPEI_PKT_OUT_BMODE is '0' a value of 1 is + added to the register for each packet, when '1' + and the info-pointer is NOT used the length of the + packet plus 8 is added, when '1' and info-pointer + mode IS used the packet length is added to this + field. */ +#else + uint64_t cnt : 32; + uint64_t timer : 22; + uint64_t reserved_54_63 : 10; +#endif + } s; + struct cvmx_npei_pktx_cnts_s cn52xx; + struct cvmx_npei_pktx_cnts_s cn56xx; +}; +typedef union cvmx_npei_pktx_cnts cvmx_npei_pktx_cnts_t; + +/** + * cvmx_npei_pkt#_in_bp + * + * NPEI_PKT[0..31]_IN_BP = NPEI Packet ring# Input Backpressure + * + * The counters and thresholds for input packets to apply backpressure to processing of the packets. + */ +union cvmx_npei_pktx_in_bp +{ + uint64_t u64; + struct cvmx_npei_pktx_in_bp_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t wmark : 32; /**< When CNT is greater than this threshold no more + packets will be processed for this ring. + When writing this field of the NPEI_PKT#_IN_BP + register, use a 4-byte write so as to not write + any other field of this register. */ + uint64_t cnt : 32; /**< ring counter. This field is incremented by one + whenever OCTEON receives, buffers, and creates a + work queue entry for a packet that arrives by the + cooresponding input ring. A write to this field + will be subtracted from the field value. + When writing this field of the NPEI_PKT#_IN_BP + register, use a 4-byte write so as to not write + any other field of this register. */ +#else + uint64_t cnt : 32; + uint64_t wmark : 32; +#endif + } s; + struct cvmx_npei_pktx_in_bp_s cn52xx; + struct cvmx_npei_pktx_in_bp_s cn56xx; +}; +typedef union cvmx_npei_pktx_in_bp cvmx_npei_pktx_in_bp_t; + +/** + * cvmx_npei_pkt#_instr_baddr + * + * NPEI_PKT[0..31]_INSTR_BADDR = NPEI Packet ring# Instruction Base Address + * + * Start of Instruction for input packets. + */ +union cvmx_npei_pktx_instr_baddr +{ + uint64_t u64; + struct cvmx_npei_pktx_instr_baddr_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t addr : 61; /**< Base address for Instructions. */ + uint64_t reserved_0_2 : 3; +#else + uint64_t reserved_0_2 : 3; + uint64_t addr : 61; +#endif + } s; + struct cvmx_npei_pktx_instr_baddr_s cn52xx; + struct cvmx_npei_pktx_instr_baddr_s cn56xx; +}; +typedef union cvmx_npei_pktx_instr_baddr cvmx_npei_pktx_instr_baddr_t; + +/** + * cvmx_npei_pkt#_instr_baoff_dbell + * + * NPEI_PKT[0..31]_INSTR_BAOFF_DBELL = NPEI Packet ring# Instruction Base Address Offset and Doorbell + * + * The doorbell and base address offset for next read. + */ +union cvmx_npei_pktx_instr_baoff_dbell +{ + uint64_t u64; + struct cvmx_npei_pktx_instr_baoff_dbell_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t aoff : 32; /**< The offset from the NPEI_PKT[0..31]_INSTR_BADDR + where the next instruction will be read. */ + uint64_t dbell : 32; /**< Instruction doorbell count. Writes to this field + will increment the value here. Reads will return + present value. A write of 0xffffffff will set the + DBELL and AOFF fields to '0'. */ +#else + uint64_t dbell : 32; + uint64_t aoff : 32; +#endif + } s; + struct cvmx_npei_pktx_instr_baoff_dbell_s cn52xx; + struct cvmx_npei_pktx_instr_baoff_dbell_s cn56xx; +}; +typedef union cvmx_npei_pktx_instr_baoff_dbell cvmx_npei_pktx_instr_baoff_dbell_t; + +/** + * cvmx_npei_pkt#_instr_fifo_rsize + * + * NPEI_PKT[0..31]_INSTR_FIFO_RSIZE = NPEI Packet ring# Instruction FIFO and Ring Size. + * + * Fifo field and ring size for Instructions. + */ +union cvmx_npei_pktx_instr_fifo_rsize +{ + uint64_t u64; + struct cvmx_npei_pktx_instr_fifo_rsize_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t max : 9; /**< Max Fifo Size. */ + uint64_t rrp : 9; /**< Fifo read pointer. */ + uint64_t wrp : 9; /**< Fifo write pointer. */ + uint64_t fcnt : 5; /**< Fifo count. */ + uint64_t rsize : 32; /**< Instruction ring size. */ +#else + uint64_t rsize : 32; + uint64_t fcnt : 5; + uint64_t wrp : 9; + uint64_t rrp : 9; + uint64_t max : 9; +#endif + } s; + struct cvmx_npei_pktx_instr_fifo_rsize_s cn52xx; + struct cvmx_npei_pktx_instr_fifo_rsize_s cn56xx; +}; +typedef union cvmx_npei_pktx_instr_fifo_rsize cvmx_npei_pktx_instr_fifo_rsize_t; + +/** + * cvmx_npei_pkt#_instr_header + * + * NPEI_PKT[0..31]_INSTR_HEADER = NPEI Packet ring# Instruction Header. + * + * VAlues used to build input packet header. + */ +union cvmx_npei_pktx_instr_header +{ + uint64_t u64; + struct cvmx_npei_pktx_instr_header_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_44_63 : 20; + uint64_t pbp : 1; /**< Enable Packet-by-packet mode. */ + uint64_t reserved_38_42 : 5; + uint64_t rparmode : 2; /**< Parse Mode. Used when packet is raw and PBP==0. */ + uint64_t reserved_35_35 : 1; + uint64_t rskp_len : 7; /**< Skip Length. Used when packet is raw and PBP==0. */ + uint64_t reserved_22_27 : 6; + uint64_t use_ihdr : 1; /**< When set '1' the instruction header will be sent + as part of the packet data, regardless of the + value of bit [63] of the instruction header. + USE_IHDR must be set whenever PBP is set. */ + uint64_t reserved_16_20 : 5; + uint64_t par_mode : 2; /**< Parse Mode. Used when USE_IHDR is set and packet + is not raw and PBP is not set. */ + uint64_t reserved_13_13 : 1; + uint64_t skp_len : 7; /**< Skip Length. Used when USE_IHDR is set and packet + is not raw and PBP is not set. */ + uint64_t reserved_0_5 : 6; +#else + uint64_t reserved_0_5 : 6; + uint64_t skp_len : 7; + uint64_t reserved_13_13 : 1; + uint64_t par_mode : 2; + uint64_t reserved_16_20 : 5; + uint64_t use_ihdr : 1; + uint64_t reserved_22_27 : 6; + uint64_t rskp_len : 7; + uint64_t reserved_35_35 : 1; + uint64_t rparmode : 2; + uint64_t reserved_38_42 : 5; + uint64_t pbp : 1; + uint64_t reserved_44_63 : 20; +#endif + } s; + struct cvmx_npei_pktx_instr_header_s cn52xx; + struct cvmx_npei_pktx_instr_header_s cn56xx; +}; +typedef union cvmx_npei_pktx_instr_header cvmx_npei_pktx_instr_header_t; + +/** + * cvmx_npei_pkt#_slist_baddr + * + * NPEI_PKT[0..31]_SLIST_BADDR = NPEI Packet ring# Scatter List Base Address + * + * Start of Scatter List for output packet pointers - MUST be 16 byte alligned + */ +union cvmx_npei_pktx_slist_baddr +{ + uint64_t u64; + struct cvmx_npei_pktx_slist_baddr_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t addr : 60; /**< Base address for scatter list pointers. */ + uint64_t reserved_0_3 : 4; +#else + uint64_t reserved_0_3 : 4; + uint64_t addr : 60; +#endif + } s; + struct cvmx_npei_pktx_slist_baddr_s cn52xx; + struct cvmx_npei_pktx_slist_baddr_s cn56xx; +}; +typedef union cvmx_npei_pktx_slist_baddr cvmx_npei_pktx_slist_baddr_t; + +/** + * cvmx_npei_pkt#_slist_baoff_dbell + * + * NPEI_PKT[0..31]_SLIST_BAOFF_DBELL = NPEI Packet ring# Scatter List Base Address Offset and Doorbell + * + * The doorbell and base address offset for next read. + */ +union cvmx_npei_pktx_slist_baoff_dbell +{ + uint64_t u64; + struct cvmx_npei_pktx_slist_baoff_dbell_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t aoff : 32; /**< The offset from the NPEI_PKT[0..31]_SLIST_BADDR + where the next SList pointer will be read. + A write of 0xFFFFFFFF to the DBELL field will + clear DBELL and AOFF */ + uint64_t dbell : 32; /**< Scatter list doorbell count. Writes to this field + will increment the value here. Reads will return + present value. The value of this field is + decremented as read operations are ISSUED for + scatter pointers. + A write of 0xFFFFFFFF will clear DBELL and AOFF */ +#else + uint64_t dbell : 32; + uint64_t aoff : 32; +#endif + } s; + struct cvmx_npei_pktx_slist_baoff_dbell_s cn52xx; + struct cvmx_npei_pktx_slist_baoff_dbell_s cn56xx; +}; +typedef union cvmx_npei_pktx_slist_baoff_dbell cvmx_npei_pktx_slist_baoff_dbell_t; + +/** + * cvmx_npei_pkt#_slist_fifo_rsize + * + * NPEI_PKT[0..31]_SLIST_FIFO_RSIZE = NPEI Packet ring# Scatter List FIFO and Ring Size. + * + * The number of scatter pointer pairs in the scatter list. + */ +union cvmx_npei_pktx_slist_fifo_rsize +{ + uint64_t u64; + struct cvmx_npei_pktx_slist_fifo_rsize_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_32_63 : 32; + uint64_t rsize : 32; /**< The number of scatter pointer pairs contained in + the scatter list ring. */ +#else + uint64_t rsize : 32; + uint64_t reserved_32_63 : 32; +#endif + } s; + struct cvmx_npei_pktx_slist_fifo_rsize_s cn52xx; + struct cvmx_npei_pktx_slist_fifo_rsize_s cn56xx; +}; +typedef union cvmx_npei_pktx_slist_fifo_rsize cvmx_npei_pktx_slist_fifo_rsize_t; + +/** + * cvmx_npei_pkt_cnt_int + * + * NPEI_PKT_CNT_INT = NPI Packet Counter Interrupt + * + * The packets rings that are interrupting because of Packet Counters. + */ +union cvmx_npei_pkt_cnt_int +{ + uint64_t u64; + struct cvmx_npei_pkt_cnt_int_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_32_63 : 32; + uint64_t port : 32; /**< Bit vector cooresponding to ring number is set when + NPEI_PKT#_CNTS[CNT] is greater + than NPEI_PKT_INT_LEVELS[CNT]. */ +#else + uint64_t port : 32; + uint64_t reserved_32_63 : 32; +#endif + } s; + struct cvmx_npei_pkt_cnt_int_s cn52xx; + struct cvmx_npei_pkt_cnt_int_s cn56xx; +}; +typedef union cvmx_npei_pkt_cnt_int cvmx_npei_pkt_cnt_int_t; + +/** + * cvmx_npei_pkt_cnt_int_enb + * + * NPEI_PKT_CNT_INT_ENB = NPI Packet Counter Interrupt Enable + * + * Enable for the packets rings that are interrupting because of Packet Counters. + */ +union cvmx_npei_pkt_cnt_int_enb +{ + uint64_t u64; + struct cvmx_npei_pkt_cnt_int_enb_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_32_63 : 32; + uint64_t port : 32; /**< Bit vector cooresponding to ring number when set + allows NPEI_PKT_CNT_INT to generate an interrupt. */ +#else + uint64_t port : 32; + uint64_t reserved_32_63 : 32; +#endif + } s; + struct cvmx_npei_pkt_cnt_int_enb_s cn52xx; + struct cvmx_npei_pkt_cnt_int_enb_s cn56xx; +}; +typedef union cvmx_npei_pkt_cnt_int_enb cvmx_npei_pkt_cnt_int_enb_t; + +/** + * cvmx_npei_pkt_data_out_es + * + * NPEI_PKT_DATA_OUT_ES = NPEI's Packet Data Out Endian Swap + * + * The Endian Swap for writing Data Out. + */ +union cvmx_npei_pkt_data_out_es +{ + uint64_t u64; + struct cvmx_npei_pkt_data_out_es_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t es : 64; /**< The endian swap mode for Packet rings 0 through 31. + Two bits are used per ring (i.e. ring 0 [1:0], + ring 1 [3:2], ....). */ +#else + uint64_t es : 64; +#endif + } s; + struct cvmx_npei_pkt_data_out_es_s cn52xx; + struct cvmx_npei_pkt_data_out_es_s cn56xx; +}; +typedef union cvmx_npei_pkt_data_out_es cvmx_npei_pkt_data_out_es_t; + +/** + * cvmx_npei_pkt_data_out_ns + * + * NPEI_PKT_DATA_OUT_NS = NPEI's Packet Data Out No Snoop + * + * The NS field for the TLP when writing packet data. + */ +union cvmx_npei_pkt_data_out_ns +{ + uint64_t u64; + struct cvmx_npei_pkt_data_out_ns_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_32_63 : 32; + uint64_t nsr : 32; /**< When asserted '1' the vector bit cooresponding + to the Packet-ring will enable NS in TLP header. */ +#else + uint64_t nsr : 32; + uint64_t reserved_32_63 : 32; +#endif + } s; + struct cvmx_npei_pkt_data_out_ns_s cn52xx; + struct cvmx_npei_pkt_data_out_ns_s cn56xx; +}; +typedef union cvmx_npei_pkt_data_out_ns cvmx_npei_pkt_data_out_ns_t; + +/** + * cvmx_npei_pkt_data_out_ror + * + * NPEI_PKT_DATA_OUT_ROR = NPEI's Packet Data Out Relaxed Ordering + * + * The ROR field for the TLP when writing Packet Data. + */ +union cvmx_npei_pkt_data_out_ror +{ + uint64_t u64; + struct cvmx_npei_pkt_data_out_ror_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_32_63 : 32; + uint64_t ror : 32; /**< When asserted '1' the vector bit cooresponding + to the Packet-ring will enable ROR in TLP header. */ +#else + uint64_t ror : 32; + uint64_t reserved_32_63 : 32; +#endif + } s; + struct cvmx_npei_pkt_data_out_ror_s cn52xx; + struct cvmx_npei_pkt_data_out_ror_s cn56xx; +}; +typedef union cvmx_npei_pkt_data_out_ror cvmx_npei_pkt_data_out_ror_t; + +/** + * cvmx_npei_pkt_dpaddr + * + * NPEI_PKT_DPADDR = NPEI's Packet Data Pointer Addr + * + * Used to detemine address and attributes for packet data writes. + */ +union cvmx_npei_pkt_dpaddr +{ + uint64_t u64; + struct cvmx_npei_pkt_dpaddr_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_32_63 : 32; + uint64_t dptr : 32; /**< When asserted '1' the vector bit cooresponding + to the Packet-ring will use: + the address[63:60] to write packet data + comes from the DPTR[63:60] in the scatter-list + pair and the RO, NS, ES values come from the O0_ES, + O0_NS, O0_RO. When '0' the RO == DPTR[60], + NS == DPTR[61], ES == DPTR[63:62], the address the + packet will be written to is ADDR[63:60] == + O0_ES[1:0], O0_NS, O0_RO. */ +#else + uint64_t dptr : 32; + uint64_t reserved_32_63 : 32; +#endif + } s; + struct cvmx_npei_pkt_dpaddr_s cn52xx; + struct cvmx_npei_pkt_dpaddr_s cn56xx; +}; +typedef union cvmx_npei_pkt_dpaddr cvmx_npei_pkt_dpaddr_t; + +/** + * cvmx_npei_pkt_in_bp + * + * NPEI_PKT_IN_BP = NPEI Packet Input Backpressure + * + * Which input rings have backpressure applied. + */ +union cvmx_npei_pkt_in_bp +{ + uint64_t u64; + struct cvmx_npei_pkt_in_bp_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_32_63 : 32; + uint64_t bp : 32; /**< A packet input ring that has its count greater + than its WMARK will have backpressure applied. + Each of the 32 bits coorespond to an input ring. + When '1' that ring has backpressure applied an + will fetch no more instructions, but will process + any previously fetched instructions. */ +#else + uint64_t bp : 32; + uint64_t reserved_32_63 : 32; +#endif + } s; + struct cvmx_npei_pkt_in_bp_s cn52xx; + struct cvmx_npei_pkt_in_bp_s cn56xx; +}; +typedef union cvmx_npei_pkt_in_bp cvmx_npei_pkt_in_bp_t; + +/** + * cvmx_npei_pkt_in_done#_cnts + * + * NPEI_PKT_IN_DONE[0..31]_CNTS = NPEI Instruction Done ring# Counts + * + * Counters for instructions completed on Input rings. + */ +union cvmx_npei_pkt_in_donex_cnts +{ + uint64_t u64; + struct cvmx_npei_pkt_in_donex_cnts_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_32_63 : 32; + uint64_t cnt : 32; /**< This field is incrmented by '1' when an instruction + is completed. This field is incremented as the + last of the data is read from the PCIe. */ +#else + uint64_t cnt : 32; + uint64_t reserved_32_63 : 32; +#endif + } s; + struct cvmx_npei_pkt_in_donex_cnts_s cn52xx; + struct cvmx_npei_pkt_in_donex_cnts_s cn56xx; +}; +typedef union cvmx_npei_pkt_in_donex_cnts cvmx_npei_pkt_in_donex_cnts_t; + +/** + * cvmx_npei_pkt_in_instr_counts + * + * NPEI_PKT_IN_INSTR_COUNTS = NPEI Packet Input Instrutction Counts + * + * Keeps track of the number of instructions read into the FIFO and Packets sent to IPD. + */ +union cvmx_npei_pkt_in_instr_counts +{ + uint64_t u64; + struct cvmx_npei_pkt_in_instr_counts_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t wr_cnt : 32; /**< Shows the number of packets sent to the IPD. */ + uint64_t rd_cnt : 32; /**< Shows the value of instructions that have had reads + issued for them. + to the Packet-ring is in reset. */ +#else + uint64_t rd_cnt : 32; + uint64_t wr_cnt : 32; +#endif + } s; + struct cvmx_npei_pkt_in_instr_counts_s cn52xx; + struct cvmx_npei_pkt_in_instr_counts_s cn56xx; +}; +typedef union cvmx_npei_pkt_in_instr_counts cvmx_npei_pkt_in_instr_counts_t; + +/** + * cvmx_npei_pkt_in_pcie_port + * + * NPEI_PKT_IN_PCIE_PORT = NPEI's Packet In To PCIe Port Assignment + * + * Assigns Packet Input rings to PCIe ports. + */ +union cvmx_npei_pkt_in_pcie_port +{ + uint64_t u64; + struct cvmx_npei_pkt_in_pcie_port_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t pp : 64; /**< The PCIe port that the Packet ring number is + assigned. Two bits are used per ring (i.e. ring 0 + [1:0], ring 1 [3:2], ....). A value of '0 means + that the Packetring is assign to PCIe Port 0, a '1' + PCIe Port 1, '2' and '3' are reserved. */ +#else + uint64_t pp : 64; +#endif + } s; + struct cvmx_npei_pkt_in_pcie_port_s cn52xx; + struct cvmx_npei_pkt_in_pcie_port_s cn56xx; +}; +typedef union cvmx_npei_pkt_in_pcie_port cvmx_npei_pkt_in_pcie_port_t; + +/** + * cvmx_npei_pkt_input_control + * + * NPEI_PKT_INPUT_CONTROL = NPEI's Packet Input Control + * + * Control for reads for gather list and instructions. + */ +union cvmx_npei_pkt_input_control +{ + uint64_t u64; + struct cvmx_npei_pkt_input_control_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_23_63 : 41; + uint64_t pkt_rr : 1; /**< When set '1' the input packet selection will be + made with a Round Robin arbitration. When '0' + the input packet ring is fixed in priority, + where the lower ring number has higher priority. */ + uint64_t pbp_dhi : 13; /**< Field when in [PBP] is set to be used in + calculating a DPTR. */ + uint64_t d_nsr : 1; /**< Enables '1' NoSnoop for reading of + gather data. */ + uint64_t d_esr : 2; /**< The Endian-Swap-Mode for reading of + gather data. */ + uint64_t d_ror : 1; /**< Enables '1' Relaxed Ordering for reading of + gather data. */ + uint64_t use_csr : 1; /**< When set '1' the csr value will be used for + ROR, ESR, and NSR. When clear '0' the value in + DPTR will be used. In turn the bits not used for + ROR, ESR, and NSR, will be used for bits [63:60] + of the address used to fetch packet data. */ + uint64_t nsr : 1; /**< Enables '1' NoSnoop for reading of + gather list and gather instruction. */ + uint64_t esr : 2; /**< The Endian-Swap-Mode for reading of + gather list and gather instruction. */ + uint64_t ror : 1; /**< Enables '1' Relaxed Ordering for reading of + gather list and gather instruction. */ +#else + uint64_t ror : 1; + uint64_t esr : 2; + uint64_t nsr : 1; + uint64_t use_csr : 1; + uint64_t d_ror : 1; + uint64_t d_esr : 2; + uint64_t d_nsr : 1; + uint64_t pbp_dhi : 13; + uint64_t pkt_rr : 1; + uint64_t reserved_23_63 : 41; +#endif + } s; + struct cvmx_npei_pkt_input_control_s cn52xx; + struct cvmx_npei_pkt_input_control_s cn56xx; +}; +typedef union cvmx_npei_pkt_input_control cvmx_npei_pkt_input_control_t; + +/** + * cvmx_npei_pkt_instr_enb + * + * NPEI_PKT_INSTR_ENB = NPEI's Packet Instruction Enable + * + * Enables the instruction fetch for a Packet-ring. + */ +union cvmx_npei_pkt_instr_enb +{ + uint64_t u64; + struct cvmx_npei_pkt_instr_enb_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_32_63 : 32; + uint64_t enb : 32; /**< When asserted '1' the vector bit cooresponding + to the Packet-ring is enabled. */ +#else + uint64_t enb : 32; + uint64_t reserved_32_63 : 32; +#endif + } s; + struct cvmx_npei_pkt_instr_enb_s cn52xx; + struct cvmx_npei_pkt_instr_enb_s cn56xx; +}; +typedef union cvmx_npei_pkt_instr_enb cvmx_npei_pkt_instr_enb_t; + +/** + * cvmx_npei_pkt_instr_rd_size + * + * NPEI_PKT_INSTR_RD_SIZE = NPEI Instruction Read Size + * + * The number of instruction allowed to be read at one time. + */ +union cvmx_npei_pkt_instr_rd_size +{ + uint64_t u64; + struct cvmx_npei_pkt_instr_rd_size_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t rdsize : 64; /**< Number of instructions to be read in one PCIe read + request for the 4 PKOport - 8 rings. Every two bits + (i.e. 1:0, 3:2, 5:4..) are assign to the port/ring + combinations. + - 15:0 PKOPort0,Ring 7..0 31:16 PKOPort1,Ring 7..0 + - 47:32 PKOPort2,Ring 7..0 63:48 PKOPort3,Ring 7..0 + Two bit value are: + 0 - 1 Instruction + 1 - 2 Instructions + 2 - 3 Instructions + 3 - 4 Instructions */ +#else + uint64_t rdsize : 64; +#endif + } s; + struct cvmx_npei_pkt_instr_rd_size_s cn52xx; + struct cvmx_npei_pkt_instr_rd_size_s cn56xx; +}; +typedef union cvmx_npei_pkt_instr_rd_size cvmx_npei_pkt_instr_rd_size_t; + +/** + * cvmx_npei_pkt_instr_size + * + * NPEI_PKT_INSTR_SIZE = NPEI's Packet Instruction Size + * + * Determines if instructions are 64 or 32 byte in size for a Packet-ring. + */ +union cvmx_npei_pkt_instr_size +{ + uint64_t u64; + struct cvmx_npei_pkt_instr_size_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_32_63 : 32; + uint64_t is_64b : 32; /**< When asserted '1' the vector bit cooresponding + to the Packet-ring is a 64-byte instruction. */ +#else + uint64_t is_64b : 32; + uint64_t reserved_32_63 : 32; +#endif + } s; + struct cvmx_npei_pkt_instr_size_s cn52xx; + struct cvmx_npei_pkt_instr_size_s cn56xx; +}; +typedef union cvmx_npei_pkt_instr_size cvmx_npei_pkt_instr_size_t; + +/** + * cvmx_npei_pkt_int_levels + * + * 0x90F0 reserved NPEI_PKT_PCIE_PORT2 + * + * + * NPEI_PKT_INT_LEVELS = NPEI's Packet Interrupt Levels + * + * Output packet interrupt levels. + */ +union cvmx_npei_pkt_int_levels +{ + uint64_t u64; + struct cvmx_npei_pkt_int_levels_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_54_63 : 10; + uint64_t time : 22; /**< When NPEI_PKT#_CNTS[TIMER] is greater than this + value an interrupt is generated. */ + uint64_t cnt : 32; /**< When NPEI_PKT#_CNTS[CNT] becomes + greater than this value an interrupt is generated. */ +#else + uint64_t cnt : 32; + uint64_t time : 22; + uint64_t reserved_54_63 : 10; +#endif + } s; + struct cvmx_npei_pkt_int_levels_s cn52xx; + struct cvmx_npei_pkt_int_levels_s cn56xx; +}; +typedef union cvmx_npei_pkt_int_levels cvmx_npei_pkt_int_levels_t; + +/** + * cvmx_npei_pkt_iptr + * + * NPEI_PKT_IPTR = NPEI's Packet Info Poitner + * + * Controls using the Info-Pointer to store length and data. + */ +union cvmx_npei_pkt_iptr +{ + uint64_t u64; + struct cvmx_npei_pkt_iptr_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_32_63 : 32; + uint64_t iptr : 32; /**< When asserted '1' the vector bit cooresponding + to the Packet-ring will use the Info-Pointer to + store length and data. */ +#else + uint64_t iptr : 32; + uint64_t reserved_32_63 : 32; +#endif + } s; + struct cvmx_npei_pkt_iptr_s cn52xx; + struct cvmx_npei_pkt_iptr_s cn56xx; +}; +typedef union cvmx_npei_pkt_iptr cvmx_npei_pkt_iptr_t; + +/** + * cvmx_npei_pkt_out_bmode + * + * NPEI_PKT_OUT_BMODE = NPEI's Packet Out Byte Mode + * + * Control the updating of the NPEI_PKT#_CNT register. + */ +union cvmx_npei_pkt_out_bmode +{ + uint64_t u64; + struct cvmx_npei_pkt_out_bmode_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_32_63 : 32; + uint64_t bmode : 32; /**< When asserted '1' the vector bit cooresponding + to the Packet-ring will have its NPEI_PKT#_CNT + register updated with the number of bytes in the + packet sent, when '0' the register will have a + value of '1' added. */ +#else + uint64_t bmode : 32; + uint64_t reserved_32_63 : 32; +#endif + } s; + struct cvmx_npei_pkt_out_bmode_s cn52xx; + struct cvmx_npei_pkt_out_bmode_s cn56xx; +}; +typedef union cvmx_npei_pkt_out_bmode cvmx_npei_pkt_out_bmode_t; + +/** + * cvmx_npei_pkt_out_enb + * + * NPEI_PKT_OUT_ENB = NPEI's Packet Output Enable + * + * Enables the output packet engines. + */ +union cvmx_npei_pkt_out_enb +{ + uint64_t u64; + struct cvmx_npei_pkt_out_enb_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_32_63 : 32; + uint64_t enb : 32; /**< When asserted '1' the vector bit cooresponding + to the Packet-ring is enabled. + If an error occurs on reading pointers for an + output ring, the ring will be disabled by clearing + the bit associated with the ring to '0'. */ +#else + uint64_t enb : 32; + uint64_t reserved_32_63 : 32; +#endif + } s; + struct cvmx_npei_pkt_out_enb_s cn52xx; + struct cvmx_npei_pkt_out_enb_s cn56xx; +}; +typedef union cvmx_npei_pkt_out_enb cvmx_npei_pkt_out_enb_t; + +/** + * cvmx_npei_pkt_output_wmark + * + * NPEI_PKT_OUTPUT_WMARK = NPEI's Packet Output Water Mark + * + * Value that when the NPEI_PKT#_SLIST_BAOFF_DBELL[DBELL] value is less then that backpressure for the rings will be applied. + */ +union cvmx_npei_pkt_output_wmark +{ + uint64_t u64; + struct cvmx_npei_pkt_output_wmark_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_32_63 : 32; + uint64_t wmark : 32; /**< Value when DBELL count drops below backpressure + for the ring will be applied to the PKO. */ +#else + uint64_t wmark : 32; + uint64_t reserved_32_63 : 32; +#endif + } s; + struct cvmx_npei_pkt_output_wmark_s cn52xx; + struct cvmx_npei_pkt_output_wmark_s cn56xx; +}; +typedef union cvmx_npei_pkt_output_wmark cvmx_npei_pkt_output_wmark_t; + +/** + * cvmx_npei_pkt_pcie_port + * + * NPEI_PKT_PCIE_PORT = NPEI's Packet To PCIe Port Assignment + * + * Assigns Packet Ports to PCIe ports. + */ +union cvmx_npei_pkt_pcie_port +{ + uint64_t u64; + struct cvmx_npei_pkt_pcie_port_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t pp : 64; /**< The PCIe port that the Packet ring number is + assigned. Two bits are used per ring (i.e. ring 0 + [1:0], ring 1 [3:2], ....). A value of '0 means + that the Packetring is assign to PCIe Port 0, a '1' + PCIe Port 1, '2' and '3' are reserved. */ +#else + uint64_t pp : 64; +#endif + } s; + struct cvmx_npei_pkt_pcie_port_s cn52xx; + struct cvmx_npei_pkt_pcie_port_s cn56xx; +}; +typedef union cvmx_npei_pkt_pcie_port cvmx_npei_pkt_pcie_port_t; + +/** + * cvmx_npei_pkt_port_in_rst + * + * NPEI_PKT_PORT_IN_RST = NPEI Packet Port In Reset + * + * Vector bits related to ring-port for ones that are reset. + */ +union cvmx_npei_pkt_port_in_rst +{ + uint64_t u64; + struct cvmx_npei_pkt_port_in_rst_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t in_rst : 32; /**< When asserted '1' the vector bit cooresponding + to the inbound Packet-ring is in reset. */ + uint64_t out_rst : 32; /**< When asserted '1' the vector bit cooresponding + to the outbound Packet-ring is in reset. */ +#else + uint64_t out_rst : 32; + uint64_t in_rst : 32; +#endif + } s; + struct cvmx_npei_pkt_port_in_rst_s cn52xx; + struct cvmx_npei_pkt_port_in_rst_s cn56xx; +}; +typedef union cvmx_npei_pkt_port_in_rst cvmx_npei_pkt_port_in_rst_t; + +/** + * cvmx_npei_pkt_slist_es + * + * NPEI_PKT_SLIST_ES = NPEI's Packet Scatter List Endian Swap + * + * The Endian Swap for Scatter List Read. + */ +union cvmx_npei_pkt_slist_es +{ + uint64_t u64; + struct cvmx_npei_pkt_slist_es_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t es : 64; /**< The endian swap mode for Packet rings 0 through 31. + Two bits are used per ring (i.e. ring 0 [1:0], + ring 1 [3:2], ....). */ +#else + uint64_t es : 64; +#endif + } s; + struct cvmx_npei_pkt_slist_es_s cn52xx; + struct cvmx_npei_pkt_slist_es_s cn56xx; +}; +typedef union cvmx_npei_pkt_slist_es cvmx_npei_pkt_slist_es_t; + +/** + * cvmx_npei_pkt_slist_id_size + * + * NPEI_PKT_SLIST_ID_SIZE = NPEI Packet Scatter List Info and Data Size + * + * The Size of the information and data fields pointed to by Scatter List pointers. + */ +union cvmx_npei_pkt_slist_id_size +{ + uint64_t u64; + struct cvmx_npei_pkt_slist_id_size_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_23_63 : 41; + uint64_t isize : 7; /**< Information size. Legal sizes are 0 to 120. */ + uint64_t bsize : 16; /**< Data size. */ +#else + uint64_t bsize : 16; + uint64_t isize : 7; + uint64_t reserved_23_63 : 41; +#endif + } s; + struct cvmx_npei_pkt_slist_id_size_s cn52xx; + struct cvmx_npei_pkt_slist_id_size_s cn56xx; +}; +typedef union cvmx_npei_pkt_slist_id_size cvmx_npei_pkt_slist_id_size_t; + +/** + * cvmx_npei_pkt_slist_ns + * + * NPEI_PKT_SLIST_NS = NPEI's Packet Scatter List No Snoop + * + * The NS field for the TLP when fetching Scatter List. + */ +union cvmx_npei_pkt_slist_ns +{ + uint64_t u64; + struct cvmx_npei_pkt_slist_ns_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_32_63 : 32; + uint64_t nsr : 32; /**< When asserted '1' the vector bit cooresponding + to the Packet-ring will enable NS in TLP header. */ +#else + uint64_t nsr : 32; + uint64_t reserved_32_63 : 32; +#endif + } s; + struct cvmx_npei_pkt_slist_ns_s cn52xx; + struct cvmx_npei_pkt_slist_ns_s cn56xx; +}; +typedef union cvmx_npei_pkt_slist_ns cvmx_npei_pkt_slist_ns_t; + +/** + * cvmx_npei_pkt_slist_ror + * + * NPEI_PKT_SLIST_ROR = NPEI's Packet Scatter List Relaxed Ordering + * + * The ROR field for the TLP when fetching Scatter List. + */ +union cvmx_npei_pkt_slist_ror +{ + uint64_t u64; + struct cvmx_npei_pkt_slist_ror_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_32_63 : 32; + uint64_t ror : 32; /**< When asserted '1' the vector bit cooresponding + to the Packet-ring will enable ROR in TLP header. */ +#else + uint64_t ror : 32; + uint64_t reserved_32_63 : 32; +#endif + } s; + struct cvmx_npei_pkt_slist_ror_s cn52xx; + struct cvmx_npei_pkt_slist_ror_s cn56xx; +}; +typedef union cvmx_npei_pkt_slist_ror cvmx_npei_pkt_slist_ror_t; + +/** + * cvmx_npei_pkt_time_int + * + * NPEI_PKT_TIME_INT = NPEI Packet Timer Interrupt + * + * The packets rings that are interrupting because of Packet Timers. + */ +union cvmx_npei_pkt_time_int +{ + uint64_t u64; + struct cvmx_npei_pkt_time_int_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_32_63 : 32; + uint64_t port : 32; /**< Bit vector cooresponding to ring number is set when + NPEI_PKT#_CNTS[TIMER] is greater than + NPEI_PKT_INT_LEVELS[TIME]. */ +#else + uint64_t port : 32; + uint64_t reserved_32_63 : 32; +#endif + } s; + struct cvmx_npei_pkt_time_int_s cn52xx; + struct cvmx_npei_pkt_time_int_s cn56xx; +}; +typedef union cvmx_npei_pkt_time_int cvmx_npei_pkt_time_int_t; + +/** + * cvmx_npei_pkt_time_int_enb + * + * NPEI_PKT_TIME_INT_ENB = NPEI Packet Timer Interrupt Enable + * + * The packets rings that are interrupting because of Packet Timers. + */ +union cvmx_npei_pkt_time_int_enb +{ + uint64_t u64; + struct cvmx_npei_pkt_time_int_enb_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_32_63 : 32; + uint64_t port : 32; /**< Bit vector cooresponding to ring number when set + allows NPEI_PKT_TIME_INT to generate an interrupt. */ +#else + uint64_t port : 32; + uint64_t reserved_32_63 : 32; +#endif + } s; + struct cvmx_npei_pkt_time_int_enb_s cn52xx; + struct cvmx_npei_pkt_time_int_enb_s cn56xx; +}; +typedef union cvmx_npei_pkt_time_int_enb cvmx_npei_pkt_time_int_enb_t; + +/** + * cvmx_npei_rsl_int_blocks + * + * NPEI_RSL_INT_BLOCKS = NPEI RSL Interrupt Blocks Register + * + * Reading this register will return a vector with a bit set '1' for a corresponding RSL block + * that presently has an interrupt pending. The Field Description below supplies the name of the + * register that software should read to find out why that intterupt bit is set. + */ +union cvmx_npei_rsl_int_blocks +{ + uint64_t u64; + struct cvmx_npei_rsl_int_blocks_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_31_63 : 33; + uint64_t iob : 1; /**< IOB_INT_SUM */ + uint64_t lmc1 : 1; /**< LMC1_MEM_CFG0 */ + uint64_t agl : 1; /**< AGL_GMX_RX0_INT_REG & AGL_GMX_TX_INT_REG */ + uint64_t reserved_24_27 : 4; + uint64_t asxpcs1 : 1; /**< PCS1_INT*_REG */ + uint64_t asxpcs0 : 1; /**< PCS0_INT*_REG */ + uint64_t reserved_21_21 : 1; + uint64_t pip : 1; /**< PIP_INT_REG. */ + uint64_t spx1 : 1; /**< Always reads as zero */ + uint64_t spx0 : 1; /**< Always reads as zero */ + uint64_t lmc0 : 1; /**< LMC0_MEM_CFG0 */ + uint64_t l2c : 1; /**< L2C_INT_STAT */ + uint64_t usb1 : 1; /**< Always reads as zero */ + uint64_t rad : 1; /**< RAD_REG_ERROR */ + uint64_t usb : 1; /**< USBN0_INT_SUM */ + uint64_t pow : 1; /**< POW_ECC_ERR */ + uint64_t tim : 1; /**< TIM_REG_ERROR */ + uint64_t pko : 1; /**< PKO_REG_ERROR */ + uint64_t ipd : 1; /**< IPD_INT_SUM */ + uint64_t reserved_8_8 : 1; + uint64_t zip : 1; /**< ZIP_ERROR */ + uint64_t dfa : 1; /**< Always reads as zero */ + uint64_t fpa : 1; /**< FPA_INT_SUM */ + uint64_t key : 1; /**< KEY_INT_SUM */ + uint64_t npei : 1; /**< NPEI_INT_SUM */ + uint64_t gmx1 : 1; /**< GMX1_RX*_INT_REG & GMX1_TX_INT_REG */ + uint64_t gmx0 : 1; /**< GMX0_RX*_INT_REG & GMX0_TX_INT_REG */ + uint64_t mio : 1; /**< MIO_BOOT_ERR */ +#else + uint64_t mio : 1; + uint64_t gmx0 : 1; + uint64_t gmx1 : 1; + uint64_t npei : 1; + uint64_t key : 1; + uint64_t fpa : 1; + uint64_t dfa : 1; + uint64_t zip : 1; + uint64_t reserved_8_8 : 1; + uint64_t ipd : 1; + uint64_t pko : 1; + uint64_t tim : 1; + uint64_t pow : 1; + uint64_t usb : 1; + uint64_t rad : 1; + uint64_t usb1 : 1; + uint64_t l2c : 1; + uint64_t lmc0 : 1; + uint64_t spx0 : 1; + uint64_t spx1 : 1; + uint64_t pip : 1; + uint64_t reserved_21_21 : 1; + uint64_t asxpcs0 : 1; + uint64_t asxpcs1 : 1; + uint64_t reserved_24_27 : 4; + uint64_t agl : 1; + uint64_t lmc1 : 1; + uint64_t iob : 1; + uint64_t reserved_31_63 : 33; +#endif + } s; + struct cvmx_npei_rsl_int_blocks_s cn52xx; + struct cvmx_npei_rsl_int_blocks_s cn52xxp1; + struct cvmx_npei_rsl_int_blocks_s cn56xx; + struct cvmx_npei_rsl_int_blocks_s cn56xxp1; +}; +typedef union cvmx_npei_rsl_int_blocks cvmx_npei_rsl_int_blocks_t; + +/** + * cvmx_npei_scratch_1 + * + * NPEI_SCRATCH_1 = NPEI's Scratch 1 + * + * A general purpose 64 bit register for SW use. + */ +union cvmx_npei_scratch_1 +{ + uint64_t u64; + struct cvmx_npei_scratch_1_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t data : 64; /**< The value in this register is totaly SW dependent. */ +#else + uint64_t data : 64; +#endif + } s; + struct cvmx_npei_scratch_1_s cn52xx; + struct cvmx_npei_scratch_1_s cn52xxp1; + struct cvmx_npei_scratch_1_s cn56xx; + struct cvmx_npei_scratch_1_s cn56xxp1; +}; +typedef union cvmx_npei_scratch_1 cvmx_npei_scratch_1_t; + +/** + * cvmx_npei_state1 + * + * NPEI_STATE1 = NPEI State 1 + * + * State machines in NPEI. For debug. + */ +union cvmx_npei_state1 +{ + uint64_t u64; + struct cvmx_npei_state1_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t cpl1 : 12; /**< CPL1 State */ + uint64_t cpl0 : 12; /**< CPL0 State */ + uint64_t arb : 1; /**< ARB State */ + uint64_t csr : 39; /**< CSR State */ +#else + uint64_t csr : 39; + uint64_t arb : 1; + uint64_t cpl0 : 12; + uint64_t cpl1 : 12; +#endif + } s; + struct cvmx_npei_state1_s cn52xx; + struct cvmx_npei_state1_s cn52xxp1; + struct cvmx_npei_state1_s cn56xx; + struct cvmx_npei_state1_s cn56xxp1; +}; +typedef union cvmx_npei_state1 cvmx_npei_state1_t; + +/** + * cvmx_npei_state2 + * + * NPEI_STATE2 = NPEI State 2 + * + * State machines in NPEI. For debug. + */ +union cvmx_npei_state2 +{ + uint64_t u64; + struct cvmx_npei_state2_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_48_63 : 16; + uint64_t npei : 1; /**< NPEI State */ + uint64_t rac : 1; /**< RAC State */ + uint64_t csm1 : 15; /**< CSM1 State */ + uint64_t csm0 : 15; /**< CSM0 State */ + uint64_t nnp0 : 8; /**< NNP0 State */ + uint64_t nnd : 8; /**< NND State */ +#else + uint64_t nnd : 8; + uint64_t nnp0 : 8; + uint64_t csm0 : 15; + uint64_t csm1 : 15; + uint64_t rac : 1; + uint64_t npei : 1; + uint64_t reserved_48_63 : 16; +#endif + } s; + struct cvmx_npei_state2_s cn52xx; + struct cvmx_npei_state2_s cn52xxp1; + struct cvmx_npei_state2_s cn56xx; + struct cvmx_npei_state2_s cn56xxp1; +}; +typedef union cvmx_npei_state2 cvmx_npei_state2_t; + +/** + * cvmx_npei_state3 + * + * NPEI_STATE3 = NPEI State 3 + * + * State machines in NPEI. For debug. + */ +union cvmx_npei_state3 +{ + uint64_t u64; + struct cvmx_npei_state3_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_56_63 : 8; + uint64_t psm1 : 15; /**< PSM1 State */ + uint64_t psm0 : 15; /**< PSM0 State */ + uint64_t nsm1 : 13; /**< NSM1 State */ + uint64_t nsm0 : 13; /**< NSM0 State */ +#else + uint64_t nsm0 : 13; + uint64_t nsm1 : 13; + uint64_t psm0 : 15; + uint64_t psm1 : 15; + uint64_t reserved_56_63 : 8; +#endif + } s; + struct cvmx_npei_state3_s cn52xx; + struct cvmx_npei_state3_s cn52xxp1; + struct cvmx_npei_state3_s cn56xx; + struct cvmx_npei_state3_s cn56xxp1; +}; +typedef union cvmx_npei_state3 cvmx_npei_state3_t; + +/** + * cvmx_npei_win_rd_addr + * + * NPEI_WIN_RD_ADDR = NPEI Window Read Address Register + * + * The address to be read when the NPEI_WIN_RD_DATA register is read. + */ +union cvmx_npei_win_rd_addr +{ + uint64_t u64; + struct cvmx_npei_win_rd_addr_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_51_63 : 13; + uint64_t ld_cmd : 2; /**< The load command sent wit hthe read. + 0x0 == Load 8-bytes, 0x1 == Load 4-bytes, + 0x2 == Load 2-bytes, 0x3 == Load 1-bytes, */ + uint64_t iobit : 1; /**< A 1 or 0 can be written here but this will always + read as '0'. */ + uint64_t rd_addr : 48; /**< The address to be read from. Whenever the LSB of + this register is written, the Read Operation will + take place. + [47:40] = NCB_ID + [39:0] = Address + When [47:43] == NPI & [42:0] == 0 bits [39:0] are: + [39:32] == x, Not Used + [31:27] == RSL_ID + [12:0] == RSL Register Offset */ +#else + uint64_t rd_addr : 48; + uint64_t iobit : 1; + uint64_t ld_cmd : 2; + uint64_t reserved_51_63 : 13; +#endif + } s; + struct cvmx_npei_win_rd_addr_s cn52xx; + struct cvmx_npei_win_rd_addr_s cn52xxp1; + struct cvmx_npei_win_rd_addr_s cn56xx; + struct cvmx_npei_win_rd_addr_s cn56xxp1; +}; +typedef union cvmx_npei_win_rd_addr cvmx_npei_win_rd_addr_t; + +/** + * cvmx_npei_win_rd_data + * + * NPEI_WIN_RD_DATA = NPEI Window Read Data Register + * + * Reading this register causes a window read operation to take place. Address read is taht contained in the NPEI_WIN_RD_ADDR + * register. + */ +union cvmx_npei_win_rd_data +{ + uint64_t u64; + struct cvmx_npei_win_rd_data_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t rd_data : 64; /**< The read data. */ +#else + uint64_t rd_data : 64; +#endif + } s; + struct cvmx_npei_win_rd_data_s cn52xx; + struct cvmx_npei_win_rd_data_s cn52xxp1; + struct cvmx_npei_win_rd_data_s cn56xx; + struct cvmx_npei_win_rd_data_s cn56xxp1; +}; +typedef union cvmx_npei_win_rd_data cvmx_npei_win_rd_data_t; + +/** + * cvmx_npei_win_wr_addr + * + * NPEI_WIN_WR_ADDR = NPEI Window Write Address Register + * + * Contains the address to be writen to when a write operation is started by writing the + * NPEI_WIN_WR_DATA register (see below). + * + * Notes: + * Even though address bit [2] can be set, it should always be kept to '0'. + * + */ +union cvmx_npei_win_wr_addr +{ + uint64_t u64; + struct cvmx_npei_win_wr_addr_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_49_63 : 15; + uint64_t iobit : 1; /**< A 1 or 0 can be written here but this will always + read as '0'. */ + uint64_t wr_addr : 46; /**< The address that will be written to when the + NPEI_WIN_WR_DATA register is written. + [47:40] = NCB_ID + [39:3] = Address + When [47:43] == NPI & [42:0] == 0 bits [39:0] are: + [39:32] == x, Not Used + [31:27] == RSL_ID + [12:2] == RSL Register Offset + [1:0] == x, Not Used */ + uint64_t reserved_0_1 : 2; +#else + uint64_t reserved_0_1 : 2; + uint64_t wr_addr : 46; + uint64_t iobit : 1; + uint64_t reserved_49_63 : 15; +#endif + } s; + struct cvmx_npei_win_wr_addr_s cn52xx; + struct cvmx_npei_win_wr_addr_s cn52xxp1; + struct cvmx_npei_win_wr_addr_s cn56xx; + struct cvmx_npei_win_wr_addr_s cn56xxp1; +}; +typedef union cvmx_npei_win_wr_addr cvmx_npei_win_wr_addr_t; + +/** + * cvmx_npei_win_wr_data + * + * NPEI_WIN_WR_DATA = NPEI Window Write Data Register + * + * Contains the data to write to the address located in the NPEI_WIN_WR_ADDR Register. + * Writing the least-significant-byte of this register will cause a write operation to take place. + */ +union cvmx_npei_win_wr_data +{ + uint64_t u64; + struct cvmx_npei_win_wr_data_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t wr_data : 64; /**< The data to be written. Whenever the LSB of this + register is written, the Window Write will take + place. */ +#else + uint64_t wr_data : 64; +#endif + } s; + struct cvmx_npei_win_wr_data_s cn52xx; + struct cvmx_npei_win_wr_data_s cn52xxp1; + struct cvmx_npei_win_wr_data_s cn56xx; + struct cvmx_npei_win_wr_data_s cn56xxp1; +}; +typedef union cvmx_npei_win_wr_data cvmx_npei_win_wr_data_t; + +/** + * cvmx_npei_win_wr_mask + * + * NPEI_WIN_WR_MASK = NPEI Window Write Mask Register + * + * Contains the mask for the data in the NPEI_WIN_WR_DATA Register. + */ +union cvmx_npei_win_wr_mask +{ + uint64_t u64; + struct cvmx_npei_win_wr_mask_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_8_63 : 56; + uint64_t wr_mask : 8; /**< The data to be written. When a bit is '0' + the corresponding byte will be written. */ +#else + uint64_t wr_mask : 8; + uint64_t reserved_8_63 : 56; +#endif + } s; + struct cvmx_npei_win_wr_mask_s cn52xx; + struct cvmx_npei_win_wr_mask_s cn52xxp1; + struct cvmx_npei_win_wr_mask_s cn56xx; + struct cvmx_npei_win_wr_mask_s cn56xxp1; +}; +typedef union cvmx_npei_win_wr_mask cvmx_npei_win_wr_mask_t; + +/** + * cvmx_npei_window_ctl + * + * NPEI_WINDOW_CTL = NPEI's Window Control + * + * The name of this register is misleading. The timeout value is used for BAR0 access from PCIE0 and PCIE1. + * Any access to the regigisters on the RML will timeout as 0xFFFF clock cycle. At time of timeout the next + * RML access will start, and interrupt will be set, and in the case of reads no data will be returned. + * + * The value of this register should be set to a minimum of 0x200000 to ensure that a timeout to an RML register + * occurs on the RML 0xFFFF timer before the timeout for a BAR0 access from the PCIE#. + */ +union cvmx_npei_window_ctl +{ + uint64_t u64; + struct cvmx_npei_window_ctl_s + { +#if __BYTE_ORDER == __BIG_ENDIAN + uint64_t reserved_32_63 : 32; + uint64_t time : 32; /**< Time to wait in core clocks to wait for a + BAR0 access to completeon the NCB + before timing out. A value of 0 will cause no + timeouts. A minimum value of 0x200000 should be + used when this register is not set to 0x0. */ +#else + uint64_t time : 32; + uint64_t reserved_32_63 : 32; +#endif + } s; + struct cvmx_npei_window_ctl_s cn52xx; + struct cvmx_npei_window_ctl_s cn52xxp1; + struct cvmx_npei_window_ctl_s cn56xx; + struct cvmx_npei_window_ctl_s cn56xxp1; +}; +typedef union cvmx_npei_window_ctl cvmx_npei_window_ctl_t; + +#endif |