summaryrefslogtreecommitdiff
path: root/arch
diff options
context:
space:
mode:
authorAaron Williams <awilliams@marvell.com>2020-12-11 17:05:54 +0100
committerDaniel Schwierzeck <daniel.schwierzeck@gmail.com>2021-04-23 21:02:12 +0200
commitdeb8b23bc06c12cb5d0d5e7c9919dabdb3d6b8e2 (patch)
tree9e0d5fb1ced62d916266c4e63ffd42cbc1de3b2f /arch
parent693a776d4d8c568e4b83936bb3a5fa6f83cec92f (diff)
downloadu-boot-deb8b23bc06c12cb5d0d5e7c9919dabdb3d6b8e2.tar.gz
mips: octeon: Add cvmx-sso-defs.h header file
Import cvmx-sso-defs.h header file from 2013 U-Boot. It will be used by the later added drivers to support PCIe and networking on the MIPS Octeon II / III platforms. Signed-off-by: Aaron Williams <awilliams@marvell.com> Signed-off-by: Stefan Roese <sr@denx.de>
Diffstat (limited to 'arch')
-rw-r--r--arch/mips/mach-octeon/include/mach/cvmx-sso-defs.h2904
1 files changed, 2904 insertions, 0 deletions
diff --git a/arch/mips/mach-octeon/include/mach/cvmx-sso-defs.h b/arch/mips/mach-octeon/include/mach/cvmx-sso-defs.h
new file mode 100644
index 0000000000..4fc69079ac
--- /dev/null
+++ b/arch/mips/mach-octeon/include/mach/cvmx-sso-defs.h
@@ -0,0 +1,2904 @@
+/* SPDX-License-Identifier: GPL-2.0 */
+/*
+ * Copyright (C) 2020 Marvell International Ltd.
+ *
+ * Configuration and status register (CSR) type definitions for
+ * Octeon sso.
+ */
+
+#ifndef __CVMX_SSO_DEFS_H__
+#define __CVMX_SSO_DEFS_H__
+
+#define CVMX_SSO_ACTIVE_CYCLES (0x00016700000010E8ull)
+#define CVMX_SSO_ACTIVE_CYCLESX(offset) (0x0001670000001100ull + ((offset) & 3) * 8)
+#define CVMX_SSO_AW_ADD (0x0001670000002080ull)
+#define CVMX_SSO_AW_CFG (0x00016700000010F0ull)
+#define CVMX_SSO_AW_ECO (0x0001670000001030ull)
+#define CVMX_SSO_AW_READ_ARB (0x0001670000002090ull)
+#define CVMX_SSO_AW_STATUS (0x00016700000010E0ull)
+#define CVMX_SSO_AW_TAG_LATENCY_PC (0x00016700000020A8ull)
+#define CVMX_SSO_AW_TAG_REQ_PC (0x00016700000020A0ull)
+#define CVMX_SSO_AW_WE (0x0001670000001080ull)
+#define CVMX_SSO_BIST_STAT (0x0001670000001078ull)
+#define CVMX_SSO_BIST_STATUS0 (0x0001670000001200ull)
+#define CVMX_SSO_BIST_STATUS1 (0x0001670000001208ull)
+#define CVMX_SSO_BIST_STATUS2 (0x0001670000001210ull)
+#define CVMX_SSO_CFG (0x0001670000001088ull)
+#define CVMX_SSO_DS_PC (0x0001670000001070ull)
+#define CVMX_SSO_ECC_CTL0 (0x0001670000001280ull)
+#define CVMX_SSO_ECC_CTL1 (0x0001670000001288ull)
+#define CVMX_SSO_ECC_CTL2 (0x0001670000001290ull)
+#define CVMX_SSO_ERR (0x0001670000001038ull)
+#define CVMX_SSO_ERR0 (0x0001670000001240ull)
+#define CVMX_SSO_ERR1 (0x0001670000001248ull)
+#define CVMX_SSO_ERR2 (0x0001670000001250ull)
+#define CVMX_SSO_ERR_ENB (0x0001670000001030ull)
+#define CVMX_SSO_FIDX_ECC_CTL (0x00016700000010D0ull)
+#define CVMX_SSO_FIDX_ECC_ST (0x00016700000010D8ull)
+#define CVMX_SSO_FPAGE_CNT (0x0001670000001090ull)
+#define CVMX_SSO_GRPX_AQ_CNT(offset) (0x0001670020000700ull + ((offset) & 255) * 0x10000ull)
+#define CVMX_SSO_GRPX_AQ_THR(offset) (0x0001670020000800ull + ((offset) & 255) * 0x10000ull)
+#define CVMX_SSO_GRPX_DS_PC(offset) (0x0001670020001400ull + ((offset) & 255) * 0x10000ull)
+#define CVMX_SSO_GRPX_EXT_PC(offset) (0x0001670020001100ull + ((offset) & 255) * 0x10000ull)
+#define CVMX_SSO_GRPX_IAQ_THR(offset) (0x0001670020000000ull + ((offset) & 255) * 0x10000ull)
+#define CVMX_SSO_GRPX_INT(offset) (0x0001670020000400ull + ((offset) & 255) * 0x10000ull)
+#define CVMX_SSO_GRPX_INT_CNT(offset) (0x0001670020000600ull + ((offset) & 255) * 0x10000ull)
+#define CVMX_SSO_GRPX_INT_THR(offset) (0x0001670020000500ull + ((offset) & 255) * 0x10000ull)
+#define CVMX_SSO_GRPX_PRI(offset) (0x0001670020000200ull + ((offset) & 255) * 0x10000ull)
+#define CVMX_SSO_GRPX_TAQ_THR(offset) (0x0001670020000100ull + ((offset) & 255) * 0x10000ull)
+#define CVMX_SSO_GRPX_TS_PC(offset) (0x0001670020001300ull + ((offset) & 255) * 0x10000ull)
+#define CVMX_SSO_GRPX_WA_PC(offset) (0x0001670020001200ull + ((offset) & 255) * 0x10000ull)
+#define CVMX_SSO_GRPX_WS_PC(offset) (0x0001670020001000ull + ((offset) & 255) * 0x10000ull)
+#define CVMX_SSO_GWE_CFG (0x0001670000001098ull)
+#define CVMX_SSO_GWE_RANDOM (0x00016700000010B0ull)
+#define CVMX_SSO_GW_ECO (0x0001670000001038ull)
+#define CVMX_SSO_IDX_ECC_CTL (0x00016700000010C0ull)
+#define CVMX_SSO_IDX_ECC_ST (0x00016700000010C8ull)
+#define CVMX_SSO_IENTX_LINKS(offset) (0x00016700A0060000ull + ((offset) & 4095) * 8)
+#define CVMX_SSO_IENTX_PENDTAG(offset) (0x00016700A0040000ull + ((offset) & 4095) * 8)
+#define CVMX_SSO_IENTX_QLINKS(offset) (0x00016700A0080000ull + ((offset) & 4095) * 8)
+#define CVMX_SSO_IENTX_TAG(offset) (0x00016700A0000000ull + ((offset) & 4095) * 8)
+#define CVMX_SSO_IENTX_WQPGRP(offset) (0x00016700A0020000ull + ((offset) & 4095) * 8)
+#define CVMX_SSO_IPL_CONFX(offset) (0x0001670080080000ull + ((offset) & 255) * 8)
+#define CVMX_SSO_IPL_DESCHEDX(offset) (0x0001670080060000ull + ((offset) & 255) * 8)
+#define CVMX_SSO_IPL_FREEX(offset) (0x0001670080000000ull + ((offset) & 7) * 8)
+#define CVMX_SSO_IPL_IAQX(offset) (0x0001670080040000ull + ((offset) & 255) * 8)
+#define CVMX_SSO_IQ_CNTX(offset) (0x0001670000009000ull + ((offset) & 7) * 8)
+#define CVMX_SSO_IQ_COM_CNT (0x0001670000001058ull)
+#define CVMX_SSO_IQ_INT (0x0001670000001048ull)
+#define CVMX_SSO_IQ_INT_EN (0x0001670000001050ull)
+#define CVMX_SSO_IQ_THRX(offset) (0x000167000000A000ull + ((offset) & 7) * 8)
+#define CVMX_SSO_NOS_CNT (0x0001670000001040ull)
+#define CVMX_SSO_NW_TIM (0x0001670000001028ull)
+#define CVMX_SSO_OTH_ECC_CTL (0x00016700000010B0ull)
+#define CVMX_SSO_OTH_ECC_ST (0x00016700000010B8ull)
+#define CVMX_SSO_PAGE_CNT (0x0001670000001090ull)
+#define CVMX_SSO_PND_ECC_CTL (0x00016700000010A0ull)
+#define CVMX_SSO_PND_ECC_ST (0x00016700000010A8ull)
+#define CVMX_SSO_PPX_ARB(offset) (0x0001670040000000ull + ((offset) & 63) * 0x10000ull)
+#define CVMX_SSO_PPX_GRP_MSK(offset) (0x0001670000006000ull + ((offset) & 31) * 8)
+#define CVMX_SSO_PPX_QOS_PRI(offset) (0x0001670000003000ull + ((offset) & 31) * 8)
+#define CVMX_SSO_PPX_SX_GRPMSKX(a, b, c) \
+ (0x0001670040001000ull + ((a) << 16) + ((b) << 5) + ((c) << 3))
+#define CVMX_SSO_PP_STRICT (0x00016700000010E0ull)
+#define CVMX_SSO_QOSX_RND(offset) (0x0001670000002000ull + ((offset) & 7) * 8)
+#define CVMX_SSO_QOS_THRX(offset) (0x000167000000B000ull + ((offset) & 7) * 8)
+#define CVMX_SSO_QOS_WE (0x0001670000001080ull)
+#define CVMX_SSO_RESET CVMX_SSO_RESET_FUNC()
+static inline u64 CVMX_SSO_RESET_FUNC(void)
+{
+ switch (cvmx_get_octeon_family()) {
+ case OCTEON_CNF75XX & OCTEON_FAMILY_MASK:
+ case OCTEON_CN78XX & OCTEON_FAMILY_MASK:
+ if (OCTEON_IS_MODEL(OCTEON_CN78XX_PASS1_X))
+ return 0x00016700000010F8ull;
+ if (OCTEON_IS_MODEL(OCTEON_CN78XX))
+ return 0x00016700000010F8ull;
+ case OCTEON_CN73XX & OCTEON_FAMILY_MASK:
+ return 0x00016700000010F8ull;
+ case OCTEON_CN68XX & OCTEON_FAMILY_MASK:
+ return 0x00016700000010F0ull;
+ }
+ return 0x00016700000010F8ull;
+}
+
+#define CVMX_SSO_RWQ_HEAD_PTRX(offset) (0x000167000000C000ull + ((offset) & 7) * 8)
+#define CVMX_SSO_RWQ_POP_FPTR (0x000167000000C408ull)
+#define CVMX_SSO_RWQ_PSH_FPTR (0x000167000000C400ull)
+#define CVMX_SSO_RWQ_TAIL_PTRX(offset) (0x000167000000C200ull + ((offset) & 7) * 8)
+#define CVMX_SSO_SL_PPX_LINKS(offset) (0x0001670060000040ull + ((offset) & 63) * 0x10000ull)
+#define CVMX_SSO_SL_PPX_PENDTAG(offset) (0x0001670060000000ull + ((offset) & 63) * 0x10000ull)
+#define CVMX_SSO_SL_PPX_PENDWQP(offset) (0x0001670060000010ull + ((offset) & 63) * 0x10000ull)
+#define CVMX_SSO_SL_PPX_TAG(offset) (0x0001670060000020ull + ((offset) & 63) * 0x10000ull)
+#define CVMX_SSO_SL_PPX_WQP(offset) (0x0001670060000030ull + ((offset) & 63) * 0x10000ull)
+#define CVMX_SSO_TAQX_LINK(offset) (0x00016700C0000000ull + ((offset) & 2047) * 4096)
+#define CVMX_SSO_TAQX_WAEX_TAG(offset, block_id) \
+ (0x00016700D0000000ull + (((offset) & 15) + ((block_id) & 2047) * 0x100ull) * 16)
+#define CVMX_SSO_TAQX_WAEX_WQP(offset, block_id) \
+ (0x00016700D0000008ull + (((offset) & 15) + ((block_id) & 2047) * 0x100ull) * 16)
+#define CVMX_SSO_TAQ_ADD (0x00016700000020E0ull)
+#define CVMX_SSO_TAQ_CNT (0x00016700000020C0ull)
+#define CVMX_SSO_TIAQX_STATUS(offset) (0x00016700000C0000ull + ((offset) & 255) * 8)
+#define CVMX_SSO_TOAQX_STATUS(offset) (0x00016700000D0000ull + ((offset) & 255) * 8)
+#define CVMX_SSO_TS_PC (0x0001670000001068ull)
+#define CVMX_SSO_WA_COM_PC (0x0001670000001060ull)
+#define CVMX_SSO_WA_PCX(offset) (0x0001670000005000ull + ((offset) & 7) * 8)
+#define CVMX_SSO_WQ_INT (0x0001670000001000ull)
+#define CVMX_SSO_WQ_INT_CNTX(offset) (0x0001670000008000ull + ((offset) & 63) * 8)
+#define CVMX_SSO_WQ_INT_PC (0x0001670000001020ull)
+#define CVMX_SSO_WQ_INT_THRX(offset) (0x0001670000007000ull + ((offset) & 63) * 8)
+#define CVMX_SSO_WQ_IQ_DIS (0x0001670000001010ull)
+#define CVMX_SSO_WS_CFG (0x0001670000001088ull)
+#define CVMX_SSO_WS_ECO (0x0001670000001048ull)
+#define CVMX_SSO_WS_PCX(offset) (0x0001670000004000ull + ((offset) & 63) * 8)
+#define CVMX_SSO_XAQX_HEAD_NEXT(offset) (0x00016700000A0000ull + ((offset) & 255) * 8)
+#define CVMX_SSO_XAQX_HEAD_PTR(offset) (0x0001670000080000ull + ((offset) & 255) * 8)
+#define CVMX_SSO_XAQX_TAIL_NEXT(offset) (0x00016700000B0000ull + ((offset) & 255) * 8)
+#define CVMX_SSO_XAQX_TAIL_PTR(offset) (0x0001670000090000ull + ((offset) & 255) * 8)
+#define CVMX_SSO_XAQ_AURA (0x0001670000002100ull)
+#define CVMX_SSO_XAQ_LATENCY_PC (0x00016700000020B8ull)
+#define CVMX_SSO_XAQ_REQ_PC (0x00016700000020B0ull)
+
+/**
+ * cvmx_sso_active_cycles
+ *
+ * SSO_ACTIVE_CYCLES = SSO cycles SSO active
+ *
+ * This register counts every sclk cycle that the SSO clocks are active.
+ * **NOTE: Added in pass 2.0
+ */
+union cvmx_sso_active_cycles {
+ u64 u64;
+ struct cvmx_sso_active_cycles_s {
+ u64 act_cyc : 64;
+ } s;
+ struct cvmx_sso_active_cycles_s cn68xx;
+};
+
+typedef union cvmx_sso_active_cycles cvmx_sso_active_cycles_t;
+
+/**
+ * cvmx_sso_active_cycles#
+ *
+ * This register counts every coprocessor clock (SCLK) cycle that the SSO clocks are active.
+ *
+ */
+union cvmx_sso_active_cyclesx {
+ u64 u64;
+ struct cvmx_sso_active_cyclesx_s {
+ u64 act_cyc : 64;
+ } s;
+ struct cvmx_sso_active_cyclesx_s cn73xx;
+ struct cvmx_sso_active_cyclesx_s cn78xx;
+ struct cvmx_sso_active_cyclesx_s cn78xxp1;
+ struct cvmx_sso_active_cyclesx_s cnf75xx;
+};
+
+typedef union cvmx_sso_active_cyclesx cvmx_sso_active_cyclesx_t;
+
+/**
+ * cvmx_sso_aw_add
+ */
+union cvmx_sso_aw_add {
+ u64 u64;
+ struct cvmx_sso_aw_add_s {
+ u64 reserved_30_63 : 34;
+ u64 rsvd_free : 14;
+ u64 reserved_0_15 : 16;
+ } s;
+ struct cvmx_sso_aw_add_s cn73xx;
+ struct cvmx_sso_aw_add_s cn78xx;
+ struct cvmx_sso_aw_add_s cn78xxp1;
+ struct cvmx_sso_aw_add_s cnf75xx;
+};
+
+typedef union cvmx_sso_aw_add cvmx_sso_aw_add_t;
+
+/**
+ * cvmx_sso_aw_cfg
+ *
+ * This register controls the operation of the add-work block (AW).
+ *
+ */
+union cvmx_sso_aw_cfg {
+ u64 u64;
+ struct cvmx_sso_aw_cfg_s {
+ u64 reserved_9_63 : 55;
+ u64 ldt_short : 1;
+ u64 lol : 1;
+ u64 xaq_alloc_dis : 1;
+ u64 ocla_bp : 1;
+ u64 xaq_byp_dis : 1;
+ u64 stt : 1;
+ u64 ldt : 1;
+ u64 ldwb : 1;
+ u64 rwen : 1;
+ } s;
+ struct cvmx_sso_aw_cfg_s cn73xx;
+ struct cvmx_sso_aw_cfg_s cn78xx;
+ struct cvmx_sso_aw_cfg_s cn78xxp1;
+ struct cvmx_sso_aw_cfg_s cnf75xx;
+};
+
+typedef union cvmx_sso_aw_cfg cvmx_sso_aw_cfg_t;
+
+/**
+ * cvmx_sso_aw_eco
+ */
+union cvmx_sso_aw_eco {
+ u64 u64;
+ struct cvmx_sso_aw_eco_s {
+ u64 reserved_8_63 : 56;
+ u64 eco_rw : 8;
+ } s;
+ struct cvmx_sso_aw_eco_s cn73xx;
+ struct cvmx_sso_aw_eco_s cnf75xx;
+};
+
+typedef union cvmx_sso_aw_eco cvmx_sso_aw_eco_t;
+
+/**
+ * cvmx_sso_aw_read_arb
+ *
+ * This register fine tunes the AW read arbiter and is for diagnostic use.
+ *
+ */
+union cvmx_sso_aw_read_arb {
+ u64 u64;
+ struct cvmx_sso_aw_read_arb_s {
+ u64 reserved_30_63 : 34;
+ u64 xaq_lev : 6;
+ u64 reserved_21_23 : 3;
+ u64 xaq_min : 5;
+ u64 reserved_14_15 : 2;
+ u64 aw_tag_lev : 6;
+ u64 reserved_5_7 : 3;
+ u64 aw_tag_min : 5;
+ } s;
+ struct cvmx_sso_aw_read_arb_s cn73xx;
+ struct cvmx_sso_aw_read_arb_s cn78xx;
+ struct cvmx_sso_aw_read_arb_s cn78xxp1;
+ struct cvmx_sso_aw_read_arb_s cnf75xx;
+};
+
+typedef union cvmx_sso_aw_read_arb cvmx_sso_aw_read_arb_t;
+
+/**
+ * cvmx_sso_aw_status
+ *
+ * This register indicates the status of the add-work block (AW).
+ *
+ */
+union cvmx_sso_aw_status {
+ u64 u64;
+ struct cvmx_sso_aw_status_s {
+ u64 reserved_6_63 : 58;
+ u64 xaq_buf_cached : 6;
+ } s;
+ struct cvmx_sso_aw_status_s cn73xx;
+ struct cvmx_sso_aw_status_s cn78xx;
+ struct cvmx_sso_aw_status_s cn78xxp1;
+ struct cvmx_sso_aw_status_s cnf75xx;
+};
+
+typedef union cvmx_sso_aw_status cvmx_sso_aw_status_t;
+
+/**
+ * cvmx_sso_aw_tag_latency_pc
+ */
+union cvmx_sso_aw_tag_latency_pc {
+ u64 u64;
+ struct cvmx_sso_aw_tag_latency_pc_s {
+ u64 count : 64;
+ } s;
+ struct cvmx_sso_aw_tag_latency_pc_s cn73xx;
+ struct cvmx_sso_aw_tag_latency_pc_s cn78xx;
+ struct cvmx_sso_aw_tag_latency_pc_s cn78xxp1;
+ struct cvmx_sso_aw_tag_latency_pc_s cnf75xx;
+};
+
+typedef union cvmx_sso_aw_tag_latency_pc cvmx_sso_aw_tag_latency_pc_t;
+
+/**
+ * cvmx_sso_aw_tag_req_pc
+ */
+union cvmx_sso_aw_tag_req_pc {
+ u64 u64;
+ struct cvmx_sso_aw_tag_req_pc_s {
+ u64 count : 64;
+ } s;
+ struct cvmx_sso_aw_tag_req_pc_s cn73xx;
+ struct cvmx_sso_aw_tag_req_pc_s cn78xx;
+ struct cvmx_sso_aw_tag_req_pc_s cn78xxp1;
+ struct cvmx_sso_aw_tag_req_pc_s cnf75xx;
+};
+
+typedef union cvmx_sso_aw_tag_req_pc cvmx_sso_aw_tag_req_pc_t;
+
+/**
+ * cvmx_sso_aw_we
+ */
+union cvmx_sso_aw_we {
+ u64 u64;
+ struct cvmx_sso_aw_we_s {
+ u64 reserved_29_63 : 35;
+ u64 rsvd_free : 13;
+ u64 reserved_13_15 : 3;
+ u64 free_cnt : 13;
+ } s;
+ struct cvmx_sso_aw_we_s cn73xx;
+ struct cvmx_sso_aw_we_s cn78xx;
+ struct cvmx_sso_aw_we_s cn78xxp1;
+ struct cvmx_sso_aw_we_s cnf75xx;
+};
+
+typedef union cvmx_sso_aw_we cvmx_sso_aw_we_t;
+
+/**
+ * cvmx_sso_bist_stat
+ *
+ * SSO_BIST_STAT = SSO BIST Status Register
+ *
+ * Contains the BIST status for the SSO memories ('0' = pass, '1' = fail).
+ * Note that PP BIST status is not reported here as it was in previous designs.
+ *
+ * There may be more for DDR interface buffers.
+ * It's possible that a RAM will be used for SSO_PP_QOS_RND.
+ */
+union cvmx_sso_bist_stat {
+ u64 u64;
+ struct cvmx_sso_bist_stat_s {
+ u64 reserved_62_63 : 2;
+ u64 odu_pref : 2;
+ u64 reserved_54_59 : 6;
+ u64 fptr : 2;
+ u64 reserved_45_51 : 7;
+ u64 rwo_dat : 1;
+ u64 rwo : 2;
+ u64 reserved_35_41 : 7;
+ u64 rwi_dat : 1;
+ u64 reserved_32_33 : 2;
+ u64 soc : 1;
+ u64 reserved_28_30 : 3;
+ u64 ncbo : 4;
+ u64 reserved_21_23 : 3;
+ u64 index : 1;
+ u64 reserved_17_19 : 3;
+ u64 fidx : 1;
+ u64 reserved_10_15 : 6;
+ u64 pend : 2;
+ u64 reserved_2_7 : 6;
+ u64 oth : 2;
+ } s;
+ struct cvmx_sso_bist_stat_s cn68xx;
+ struct cvmx_sso_bist_stat_cn68xxp1 {
+ u64 reserved_54_63 : 10;
+ u64 fptr : 2;
+ u64 reserved_45_51 : 7;
+ u64 rwo_dat : 1;
+ u64 rwo : 2;
+ u64 reserved_35_41 : 7;
+ u64 rwi_dat : 1;
+ u64 reserved_32_33 : 2;
+ u64 soc : 1;
+ u64 reserved_28_30 : 3;
+ u64 ncbo : 4;
+ u64 reserved_21_23 : 3;
+ u64 index : 1;
+ u64 reserved_17_19 : 3;
+ u64 fidx : 1;
+ u64 reserved_10_15 : 6;
+ u64 pend : 2;
+ u64 reserved_2_7 : 6;
+ u64 oth : 2;
+ } cn68xxp1;
+};
+
+typedef union cvmx_sso_bist_stat cvmx_sso_bist_stat_t;
+
+/**
+ * cvmx_sso_bist_status0
+ *
+ * Contains the BIST status for the SSO memories.
+ *
+ */
+union cvmx_sso_bist_status0 {
+ u64 u64;
+ struct cvmx_sso_bist_status0_s {
+ u64 reserved_10_63 : 54;
+ u64 bist : 10;
+ } s;
+ struct cvmx_sso_bist_status0_s cn73xx;
+ struct cvmx_sso_bist_status0_s cn78xx;
+ struct cvmx_sso_bist_status0_s cn78xxp1;
+ struct cvmx_sso_bist_status0_s cnf75xx;
+};
+
+typedef union cvmx_sso_bist_status0 cvmx_sso_bist_status0_t;
+
+/**
+ * cvmx_sso_bist_status1
+ *
+ * Contains the BIST status for the SSO memories.
+ *
+ */
+union cvmx_sso_bist_status1 {
+ u64 u64;
+ struct cvmx_sso_bist_status1_s {
+ u64 reserved_7_63 : 57;
+ u64 bist : 7;
+ } s;
+ struct cvmx_sso_bist_status1_s cn73xx;
+ struct cvmx_sso_bist_status1_s cn78xx;
+ struct cvmx_sso_bist_status1_s cn78xxp1;
+ struct cvmx_sso_bist_status1_s cnf75xx;
+};
+
+typedef union cvmx_sso_bist_status1 cvmx_sso_bist_status1_t;
+
+/**
+ * cvmx_sso_bist_status2
+ *
+ * Contains the BIST status for the SSO memories.
+ *
+ */
+union cvmx_sso_bist_status2 {
+ u64 u64;
+ struct cvmx_sso_bist_status2_s {
+ u64 reserved_9_63 : 55;
+ u64 bist : 9;
+ } s;
+ struct cvmx_sso_bist_status2_s cn73xx;
+ struct cvmx_sso_bist_status2_s cn78xx;
+ struct cvmx_sso_bist_status2_s cn78xxp1;
+ struct cvmx_sso_bist_status2_s cnf75xx;
+};
+
+typedef union cvmx_sso_bist_status2 cvmx_sso_bist_status2_t;
+
+/**
+ * cvmx_sso_cfg
+ *
+ * SSO_CFG = SSO Config
+ *
+ * This register is an assortment of various SSO configuration bits.
+ */
+union cvmx_sso_cfg {
+ u64 u64;
+ struct cvmx_sso_cfg_s {
+ u64 reserved_16_63 : 48;
+ u64 qck_gw_rsp_adj : 3;
+ u64 qck_gw_rsp_dis : 1;
+ u64 qck_sw_dis : 1;
+ u64 rwq_alloc_dis : 1;
+ u64 soc_ccam_dis : 1;
+ u64 sso_cclk_dis : 1;
+ u64 rwo_flush : 1;
+ u64 wfe_thr : 1;
+ u64 rwio_byp_dis : 1;
+ u64 rwq_byp_dis : 1;
+ u64 stt : 1;
+ u64 ldt : 1;
+ u64 dwb : 1;
+ u64 rwen : 1;
+ } s;
+ struct cvmx_sso_cfg_s cn68xx;
+ struct cvmx_sso_cfg_cn68xxp1 {
+ u64 reserved_8_63 : 56;
+ u64 rwo_flush : 1;
+ u64 wfe_thr : 1;
+ u64 rwio_byp_dis : 1;
+ u64 rwq_byp_dis : 1;
+ u64 stt : 1;
+ u64 ldt : 1;
+ u64 dwb : 1;
+ u64 rwen : 1;
+ } cn68xxp1;
+};
+
+typedef union cvmx_sso_cfg cvmx_sso_cfg_t;
+
+/**
+ * cvmx_sso_ds_pc
+ *
+ * SSO_DS_PC = SSO De-Schedule Performance Counter
+ *
+ * Counts the number of de-schedule requests.
+ * Counter rolls over through zero when max value exceeded.
+ */
+union cvmx_sso_ds_pc {
+ u64 u64;
+ struct cvmx_sso_ds_pc_s {
+ u64 ds_pc : 64;
+ } s;
+ struct cvmx_sso_ds_pc_s cn68xx;
+ struct cvmx_sso_ds_pc_s cn68xxp1;
+};
+
+typedef union cvmx_sso_ds_pc cvmx_sso_ds_pc_t;
+
+/**
+ * cvmx_sso_ecc_ctl0
+ */
+union cvmx_sso_ecc_ctl0 {
+ u64 u64;
+ struct cvmx_sso_ecc_ctl0_s {
+ u64 reserved_30_63 : 34;
+ u64 toaqt_flip : 2;
+ u64 toaqt_cdis : 1;
+ u64 toaqh_flip : 2;
+ u64 toaqh_cdis : 1;
+ u64 tiaqt_flip : 2;
+ u64 tiaqt_cdis : 1;
+ u64 tiaqh_flip : 2;
+ u64 tiaqh_cdis : 1;
+ u64 llm_flip : 2;
+ u64 llm_cdis : 1;
+ u64 inp_flip : 2;
+ u64 inp_cdis : 1;
+ u64 qtc_flip : 2;
+ u64 qtc_cdis : 1;
+ u64 xaq_flip : 2;
+ u64 xaq_cdis : 1;
+ u64 fff_flip : 2;
+ u64 fff_cdis : 1;
+ u64 wes_flip : 2;
+ u64 wes_cdis : 1;
+ } s;
+ struct cvmx_sso_ecc_ctl0_s cn73xx;
+ struct cvmx_sso_ecc_ctl0_s cn78xx;
+ struct cvmx_sso_ecc_ctl0_s cn78xxp1;
+ struct cvmx_sso_ecc_ctl0_s cnf75xx;
+};
+
+typedef union cvmx_sso_ecc_ctl0 cvmx_sso_ecc_ctl0_t;
+
+/**
+ * cvmx_sso_ecc_ctl1
+ */
+union cvmx_sso_ecc_ctl1 {
+ u64 u64;
+ struct cvmx_sso_ecc_ctl1_s {
+ u64 reserved_21_63 : 43;
+ u64 thrint_flip : 2;
+ u64 thrint_cdis : 1;
+ u64 mask_flip : 2;
+ u64 mask_cdis : 1;
+ u64 gdw_flip : 2;
+ u64 gdw_cdis : 1;
+ u64 qidx_flip : 2;
+ u64 qidx_cdis : 1;
+ u64 tptr_flip : 2;
+ u64 tptr_cdis : 1;
+ u64 hptr_flip : 2;
+ u64 hptr_cdis : 1;
+ u64 cntr_flip : 2;
+ u64 cntr_cdis : 1;
+ } s;
+ struct cvmx_sso_ecc_ctl1_s cn73xx;
+ struct cvmx_sso_ecc_ctl1_s cn78xx;
+ struct cvmx_sso_ecc_ctl1_s cn78xxp1;
+ struct cvmx_sso_ecc_ctl1_s cnf75xx;
+};
+
+typedef union cvmx_sso_ecc_ctl1 cvmx_sso_ecc_ctl1_t;
+
+/**
+ * cvmx_sso_ecc_ctl2
+ */
+union cvmx_sso_ecc_ctl2 {
+ u64 u64;
+ struct cvmx_sso_ecc_ctl2_s {
+ u64 reserved_15_63 : 49;
+ u64 ncbo_flip : 2;
+ u64 ncbo_cdis : 1;
+ u64 pnd_flip : 2;
+ u64 pnd_cdis : 1;
+ u64 oth_flip : 2;
+ u64 oth_cdis : 1;
+ u64 nidx_flip : 2;
+ u64 nidx_cdis : 1;
+ u64 pidx_flip : 2;
+ u64 pidx_cdis : 1;
+ } s;
+ struct cvmx_sso_ecc_ctl2_s cn73xx;
+ struct cvmx_sso_ecc_ctl2_s cn78xx;
+ struct cvmx_sso_ecc_ctl2_s cn78xxp1;
+ struct cvmx_sso_ecc_ctl2_s cnf75xx;
+};
+
+typedef union cvmx_sso_ecc_ctl2 cvmx_sso_ecc_ctl2_t;
+
+/**
+ * cvmx_sso_err
+ *
+ * SSO_ERR = SSO Error Register
+ *
+ * Contains ECC and other misc error bits.
+ *
+ * <45> The free page error bit will assert when SSO_FPAGE_CNT <= 16 and
+ * SSO_CFG[RWEN] is 1. Software will want to disable the interrupt
+ * associated with this error when recovering SSO pointers from the
+ * FPA and SSO.
+ *
+ * This register also contains the illegal operation error bits:
+ *
+ * <42> Received ADDWQ with tag specified as EMPTY
+ * <41> Received illegal opcode
+ * <40> Received SWTAG/SWTAG_FULL/SWTAG_DESCH/DESCH/UPD_WQP/GET_WORK/ALLOC_WE
+ * from WS with CLR_NSCHED pending
+ * <39> Received CLR_NSCHED
+ * from WS with SWTAG_DESCH/DESCH/CLR_NSCHED pending
+ * <38> Received SWTAG/SWTAG_FULL/SWTAG_DESCH/DESCH/UPD_WQP/GET_WORK/ALLOC_WE
+ * from WS with ALLOC_WE pending
+ * <37> Received SWTAG/SWTAG_FULL/SWTAG_DESCH/DESCH/UPD_WQP/GET_WORK/ALLOC_WE/CLR_NSCHED
+ * from WS with GET_WORK pending
+ * <36> Received SWTAG_FULL/SWTAG_DESCH
+ * with tag specified as UNSCHEDULED
+ * <35> Received SWTAG/SWTAG_FULL/SWTAG_DESCH
+ * with tag specified as EMPTY
+ * <34> Received SWTAG/SWTAG_FULL/SWTAG_DESCH/GET_WORK
+ * from WS with pending tag switch to ORDERED or ATOMIC
+ * <33> Received SWTAG/SWTAG_DESCH/DESCH/UPD_WQP
+ * from WS in UNSCHEDULED state
+ * <32> Received SWTAG/SWTAG_FULL/SWTAG_DESCH/DESCH/UPD_WQP
+ * from WS in EMPTY state
+ */
+union cvmx_sso_err {
+ u64 u64;
+ struct cvmx_sso_err_s {
+ u64 reserved_48_63 : 16;
+ u64 bfp : 1;
+ u64 awe : 1;
+ u64 fpe : 1;
+ u64 reserved_43_44 : 2;
+ u64 iop : 11;
+ u64 reserved_12_31 : 20;
+ u64 pnd_dbe0 : 1;
+ u64 pnd_sbe0 : 1;
+ u64 pnd_dbe1 : 1;
+ u64 pnd_sbe1 : 1;
+ u64 oth_dbe0 : 1;
+ u64 oth_sbe0 : 1;
+ u64 oth_dbe1 : 1;
+ u64 oth_sbe1 : 1;
+ u64 idx_dbe : 1;
+ u64 idx_sbe : 1;
+ u64 fidx_dbe : 1;
+ u64 fidx_sbe : 1;
+ } s;
+ struct cvmx_sso_err_s cn68xx;
+ struct cvmx_sso_err_s cn68xxp1;
+};
+
+typedef union cvmx_sso_err cvmx_sso_err_t;
+
+/**
+ * cvmx_sso_err0
+ *
+ * This register contains ECC and other miscellaneous error bits.
+ *
+ */
+union cvmx_sso_err0 {
+ u64 u64;
+ struct cvmx_sso_err0_s {
+ u64 reserved_52_63 : 12;
+ u64 toaqt_dbe : 1;
+ u64 toaqt_sbe : 1;
+ u64 toaqh_dbe : 1;
+ u64 toaqh_sbe : 1;
+ u64 tiaqt_dbe : 1;
+ u64 tiaqt_sbe : 1;
+ u64 tiaqh_dbe : 1;
+ u64 tiaqh_sbe : 1;
+ u64 llm_dbe : 1;
+ u64 llm_sbe : 1;
+ u64 inp_dbe : 1;
+ u64 inp_sbe : 1;
+ u64 qtc_dbe : 1;
+ u64 qtc_sbe : 1;
+ u64 xaq_dbe : 1;
+ u64 xaq_sbe : 1;
+ u64 fff_dbe : 1;
+ u64 fff_sbe : 1;
+ u64 wes_dbe : 1;
+ u64 wes_sbe : 1;
+ u64 reserved_6_31 : 26;
+ u64 addwq_dropped : 1;
+ u64 awempty : 1;
+ u64 grpdis : 1;
+ u64 bfp : 1;
+ u64 awe : 1;
+ u64 fpe : 1;
+ } s;
+ struct cvmx_sso_err0_s cn73xx;
+ struct cvmx_sso_err0_s cn78xx;
+ struct cvmx_sso_err0_s cn78xxp1;
+ struct cvmx_sso_err0_s cnf75xx;
+};
+
+typedef union cvmx_sso_err0 cvmx_sso_err0_t;
+
+/**
+ * cvmx_sso_err1
+ *
+ * This register contains ECC and other miscellaneous error bits.
+ *
+ */
+union cvmx_sso_err1 {
+ u64 u64;
+ struct cvmx_sso_err1_s {
+ u64 reserved_14_63 : 50;
+ u64 thrint_dbe : 1;
+ u64 thrint_sbe : 1;
+ u64 mask_dbe : 1;
+ u64 mask_sbe : 1;
+ u64 gdw_dbe : 1;
+ u64 gdw_sbe : 1;
+ u64 qidx_dbe : 1;
+ u64 qidx_sbe : 1;
+ u64 tptr_dbe : 1;
+ u64 tptr_sbe : 1;
+ u64 hptr_dbe : 1;
+ u64 hptr_sbe : 1;
+ u64 cntr_dbe : 1;
+ u64 cntr_sbe : 1;
+ } s;
+ struct cvmx_sso_err1_s cn73xx;
+ struct cvmx_sso_err1_s cn78xx;
+ struct cvmx_sso_err1_s cn78xxp1;
+ struct cvmx_sso_err1_s cnf75xx;
+};
+
+typedef union cvmx_sso_err1 cvmx_sso_err1_t;
+
+/**
+ * cvmx_sso_err2
+ *
+ * This register contains ECC and other miscellaneous error bits.
+ *
+ */
+union cvmx_sso_err2 {
+ u64 u64;
+ struct cvmx_sso_err2_s {
+ u64 reserved_42_63 : 22;
+ u64 ncbo_dbe : 1;
+ u64 ncbo_sbe : 1;
+ u64 pnd_dbe : 1;
+ u64 pnd_sbe : 1;
+ u64 oth_dbe : 1;
+ u64 oth_sbe : 1;
+ u64 nidx_dbe : 1;
+ u64 nidx_sbe : 1;
+ u64 pidx_dbe : 1;
+ u64 pidx_sbe : 1;
+ u64 reserved_13_31 : 19;
+ u64 iop : 13;
+ } s;
+ struct cvmx_sso_err2_s cn73xx;
+ struct cvmx_sso_err2_s cn78xx;
+ struct cvmx_sso_err2_s cn78xxp1;
+ struct cvmx_sso_err2_s cnf75xx;
+};
+
+typedef union cvmx_sso_err2 cvmx_sso_err2_t;
+
+/**
+ * cvmx_sso_err_enb
+ *
+ * SSO_ERR_ENB = SSO Error Enable Register
+ *
+ * Contains the interrupt enables corresponding to SSO_ERR.
+ */
+union cvmx_sso_err_enb {
+ u64 u64;
+ struct cvmx_sso_err_enb_s {
+ u64 reserved_48_63 : 16;
+ u64 bfp_ie : 1;
+ u64 awe_ie : 1;
+ u64 fpe_ie : 1;
+ u64 reserved_43_44 : 2;
+ u64 iop_ie : 11;
+ u64 reserved_12_31 : 20;
+ u64 pnd_dbe0_ie : 1;
+ u64 pnd_sbe0_ie : 1;
+ u64 pnd_dbe1_ie : 1;
+ u64 pnd_sbe1_ie : 1;
+ u64 oth_dbe0_ie : 1;
+ u64 oth_sbe0_ie : 1;
+ u64 oth_dbe1_ie : 1;
+ u64 oth_sbe1_ie : 1;
+ u64 idx_dbe_ie : 1;
+ u64 idx_sbe_ie : 1;
+ u64 fidx_dbe_ie : 1;
+ u64 fidx_sbe_ie : 1;
+ } s;
+ struct cvmx_sso_err_enb_s cn68xx;
+ struct cvmx_sso_err_enb_s cn68xxp1;
+};
+
+typedef union cvmx_sso_err_enb cvmx_sso_err_enb_t;
+
+/**
+ * cvmx_sso_fidx_ecc_ctl
+ *
+ * SSO_FIDX_ECC_CTL = SSO FIDX ECC Control
+ *
+ */
+union cvmx_sso_fidx_ecc_ctl {
+ u64 u64;
+ struct cvmx_sso_fidx_ecc_ctl_s {
+ u64 reserved_3_63 : 61;
+ u64 flip_synd : 2;
+ u64 ecc_ena : 1;
+ } s;
+ struct cvmx_sso_fidx_ecc_ctl_s cn68xx;
+ struct cvmx_sso_fidx_ecc_ctl_s cn68xxp1;
+};
+
+typedef union cvmx_sso_fidx_ecc_ctl cvmx_sso_fidx_ecc_ctl_t;
+
+/**
+ * cvmx_sso_fidx_ecc_st
+ *
+ * SSO_FIDX_ECC_ST = SSO FIDX ECC Status
+ *
+ */
+union cvmx_sso_fidx_ecc_st {
+ u64 u64;
+ struct cvmx_sso_fidx_ecc_st_s {
+ u64 reserved_27_63 : 37;
+ u64 addr : 11;
+ u64 reserved_9_15 : 7;
+ u64 syndrom : 5;
+ u64 reserved_0_3 : 4;
+ } s;
+ struct cvmx_sso_fidx_ecc_st_s cn68xx;
+ struct cvmx_sso_fidx_ecc_st_s cn68xxp1;
+};
+
+typedef union cvmx_sso_fidx_ecc_st cvmx_sso_fidx_ecc_st_t;
+
+/**
+ * cvmx_sso_fpage_cnt
+ *
+ * SSO_FPAGE_CNT = SSO Free Page Cnt
+ *
+ * This register keeps track of the number of free pages pointers available for use in external memory.
+ */
+union cvmx_sso_fpage_cnt {
+ u64 u64;
+ struct cvmx_sso_fpage_cnt_s {
+ u64 reserved_32_63 : 32;
+ u64 fpage_cnt : 32;
+ } s;
+ struct cvmx_sso_fpage_cnt_s cn68xx;
+ struct cvmx_sso_fpage_cnt_s cn68xxp1;
+};
+
+typedef union cvmx_sso_fpage_cnt cvmx_sso_fpage_cnt_t;
+
+/**
+ * cvmx_sso_grp#_aq_cnt
+ */
+union cvmx_sso_grpx_aq_cnt {
+ u64 u64;
+ struct cvmx_sso_grpx_aq_cnt_s {
+ u64 reserved_33_63 : 31;
+ u64 aq_cnt : 33;
+ } s;
+ struct cvmx_sso_grpx_aq_cnt_s cn73xx;
+ struct cvmx_sso_grpx_aq_cnt_s cn78xx;
+ struct cvmx_sso_grpx_aq_cnt_s cn78xxp1;
+ struct cvmx_sso_grpx_aq_cnt_s cnf75xx;
+};
+
+typedef union cvmx_sso_grpx_aq_cnt cvmx_sso_grpx_aq_cnt_t;
+
+/**
+ * cvmx_sso_grp#_aq_thr
+ */
+union cvmx_sso_grpx_aq_thr {
+ u64 u64;
+ struct cvmx_sso_grpx_aq_thr_s {
+ u64 reserved_33_63 : 31;
+ u64 aq_thr : 33;
+ } s;
+ struct cvmx_sso_grpx_aq_thr_s cn73xx;
+ struct cvmx_sso_grpx_aq_thr_s cn78xx;
+ struct cvmx_sso_grpx_aq_thr_s cn78xxp1;
+ struct cvmx_sso_grpx_aq_thr_s cnf75xx;
+};
+
+typedef union cvmx_sso_grpx_aq_thr cvmx_sso_grpx_aq_thr_t;
+
+/**
+ * cvmx_sso_grp#_ds_pc
+ *
+ * Counts the number of deschedule requests for each group. Counter rolls over through zero when
+ * max value exceeded.
+ */
+union cvmx_sso_grpx_ds_pc {
+ u64 u64;
+ struct cvmx_sso_grpx_ds_pc_s {
+ u64 cnt : 64;
+ } s;
+ struct cvmx_sso_grpx_ds_pc_s cn73xx;
+ struct cvmx_sso_grpx_ds_pc_s cn78xx;
+ struct cvmx_sso_grpx_ds_pc_s cn78xxp1;
+ struct cvmx_sso_grpx_ds_pc_s cnf75xx;
+};
+
+typedef union cvmx_sso_grpx_ds_pc cvmx_sso_grpx_ds_pc_t;
+
+/**
+ * cvmx_sso_grp#_ext_pc
+ *
+ * Counts the number of cache lines of WAEs sent to L2/DDR. Counter rolls over through zero when
+ * max value exceeded.
+ */
+union cvmx_sso_grpx_ext_pc {
+ u64 u64;
+ struct cvmx_sso_grpx_ext_pc_s {
+ u64 cnt : 64;
+ } s;
+ struct cvmx_sso_grpx_ext_pc_s cn73xx;
+ struct cvmx_sso_grpx_ext_pc_s cn78xx;
+ struct cvmx_sso_grpx_ext_pc_s cn78xxp1;
+ struct cvmx_sso_grpx_ext_pc_s cnf75xx;
+};
+
+typedef union cvmx_sso_grpx_ext_pc cvmx_sso_grpx_ext_pc_t;
+
+/**
+ * cvmx_sso_grp#_iaq_thr
+ *
+ * These registers contain the thresholds for allocating SSO in-unit admission queue entries, see
+ * In-Unit Thresholds.
+ */
+union cvmx_sso_grpx_iaq_thr {
+ u64 u64;
+ struct cvmx_sso_grpx_iaq_thr_s {
+ u64 reserved_61_63 : 3;
+ u64 grp_cnt : 13;
+ u64 reserved_45_47 : 3;
+ u64 max_thr : 13;
+ u64 reserved_13_31 : 19;
+ u64 rsvd_thr : 13;
+ } s;
+ struct cvmx_sso_grpx_iaq_thr_s cn73xx;
+ struct cvmx_sso_grpx_iaq_thr_s cn78xx;
+ struct cvmx_sso_grpx_iaq_thr_s cn78xxp1;
+ struct cvmx_sso_grpx_iaq_thr_s cnf75xx;
+};
+
+typedef union cvmx_sso_grpx_iaq_thr cvmx_sso_grpx_iaq_thr_t;
+
+/**
+ * cvmx_sso_grp#_int
+ *
+ * Contains the per-group interrupts and are used to clear these interrupts. For more information
+ * on this register, refer to Interrupts.
+ */
+union cvmx_sso_grpx_int {
+ u64 u64;
+ struct cvmx_sso_grpx_int_s {
+ u64 exe_dis : 1;
+ u64 reserved_2_62 : 61;
+ u64 exe_int : 1;
+ u64 aq_int : 1;
+ } s;
+ struct cvmx_sso_grpx_int_s cn73xx;
+ struct cvmx_sso_grpx_int_s cn78xx;
+ struct cvmx_sso_grpx_int_s cn78xxp1;
+ struct cvmx_sso_grpx_int_s cnf75xx;
+};
+
+typedef union cvmx_sso_grpx_int cvmx_sso_grpx_int_t;
+
+/**
+ * cvmx_sso_grp#_int_cnt
+ *
+ * These registers contain a read-only copy of the counts used to trigger work-queue interrupts
+ * (one per group). For more information on this register, refer to Interrupts.
+ */
+union cvmx_sso_grpx_int_cnt {
+ u64 u64;
+ struct cvmx_sso_grpx_int_cnt_s {
+ u64 reserved_61_63 : 3;
+ u64 tc_cnt : 13;
+ u64 reserved_45_47 : 3;
+ u64 cq_cnt : 13;
+ u64 reserved_29_31 : 3;
+ u64 ds_cnt : 13;
+ u64 reserved_13_15 : 3;
+ u64 iaq_cnt : 13;
+ } s;
+ struct cvmx_sso_grpx_int_cnt_s cn73xx;
+ struct cvmx_sso_grpx_int_cnt_s cn78xx;
+ struct cvmx_sso_grpx_int_cnt_s cn78xxp1;
+ struct cvmx_sso_grpx_int_cnt_s cnf75xx;
+};
+
+typedef union cvmx_sso_grpx_int_cnt cvmx_sso_grpx_int_cnt_t;
+
+/**
+ * cvmx_sso_grp#_int_thr
+ *
+ * These registers contain the thresholds for enabling and setting work-queue interrupts (one per
+ * group). For more information on this register, refer to Interrupts.
+ */
+union cvmx_sso_grpx_int_thr {
+ u64 u64;
+ struct cvmx_sso_grpx_int_thr_s {
+ u64 tc_en : 1;
+ u64 reserved_61_62 : 2;
+ u64 tc_thr : 13;
+ u64 reserved_45_47 : 3;
+ u64 cq_thr : 13;
+ u64 reserved_29_31 : 3;
+ u64 ds_thr : 13;
+ u64 reserved_13_15 : 3;
+ u64 iaq_thr : 13;
+ } s;
+ struct cvmx_sso_grpx_int_thr_s cn73xx;
+ struct cvmx_sso_grpx_int_thr_s cn78xx;
+ struct cvmx_sso_grpx_int_thr_s cn78xxp1;
+ struct cvmx_sso_grpx_int_thr_s cnf75xx;
+};
+
+typedef union cvmx_sso_grpx_int_thr cvmx_sso_grpx_int_thr_t;
+
+/**
+ * cvmx_sso_grp#_pri
+ *
+ * Controls the priority and group affinity arbitration for each group.
+ *
+ */
+union cvmx_sso_grpx_pri {
+ u64 u64;
+ struct cvmx_sso_grpx_pri_s {
+ u64 reserved_30_63 : 34;
+ u64 wgt_left : 6;
+ u64 reserved_22_23 : 2;
+ u64 weight : 6;
+ u64 reserved_12_15 : 4;
+ u64 affinity : 4;
+ u64 reserved_3_7 : 5;
+ u64 pri : 3;
+ } s;
+ struct cvmx_sso_grpx_pri_s cn73xx;
+ struct cvmx_sso_grpx_pri_s cn78xx;
+ struct cvmx_sso_grpx_pri_s cn78xxp1;
+ struct cvmx_sso_grpx_pri_s cnf75xx;
+};
+
+typedef union cvmx_sso_grpx_pri cvmx_sso_grpx_pri_t;
+
+/**
+ * cvmx_sso_grp#_taq_thr
+ *
+ * These registers contain the thresholds for allocating SSO transitory admission queue storage
+ * buffers, see Transitory-Admission Thresholds.
+ */
+union cvmx_sso_grpx_taq_thr {
+ u64 u64;
+ struct cvmx_sso_grpx_taq_thr_s {
+ u64 reserved_59_63 : 5;
+ u64 grp_cnt : 11;
+ u64 reserved_43_47 : 5;
+ u64 max_thr : 11;
+ u64 reserved_11_31 : 21;
+ u64 rsvd_thr : 11;
+ } s;
+ struct cvmx_sso_grpx_taq_thr_s cn73xx;
+ struct cvmx_sso_grpx_taq_thr_s cn78xx;
+ struct cvmx_sso_grpx_taq_thr_s cn78xxp1;
+ struct cvmx_sso_grpx_taq_thr_s cnf75xx;
+};
+
+typedef union cvmx_sso_grpx_taq_thr cvmx_sso_grpx_taq_thr_t;
+
+/**
+ * cvmx_sso_grp#_ts_pc
+ *
+ * Counts the number of tag switch requests for each group being switched to. Counter rolls over
+ * through zero when max value exceeded.
+ */
+union cvmx_sso_grpx_ts_pc {
+ u64 u64;
+ struct cvmx_sso_grpx_ts_pc_s {
+ u64 cnt : 64;
+ } s;
+ struct cvmx_sso_grpx_ts_pc_s cn73xx;
+ struct cvmx_sso_grpx_ts_pc_s cn78xx;
+ struct cvmx_sso_grpx_ts_pc_s cn78xxp1;
+ struct cvmx_sso_grpx_ts_pc_s cnf75xx;
+};
+
+typedef union cvmx_sso_grpx_ts_pc cvmx_sso_grpx_ts_pc_t;
+
+/**
+ * cvmx_sso_grp#_wa_pc
+ *
+ * Counts the number of add new work requests for each group. The counter rolls over through zero
+ * when the max value exceeded.
+ */
+union cvmx_sso_grpx_wa_pc {
+ u64 u64;
+ struct cvmx_sso_grpx_wa_pc_s {
+ u64 cnt : 64;
+ } s;
+ struct cvmx_sso_grpx_wa_pc_s cn73xx;
+ struct cvmx_sso_grpx_wa_pc_s cn78xx;
+ struct cvmx_sso_grpx_wa_pc_s cn78xxp1;
+ struct cvmx_sso_grpx_wa_pc_s cnf75xx;
+};
+
+typedef union cvmx_sso_grpx_wa_pc cvmx_sso_grpx_wa_pc_t;
+
+/**
+ * cvmx_sso_grp#_ws_pc
+ *
+ * Counts the number of work schedules for each group. The counter rolls over through zero when
+ * the maximum value is exceeded.
+ */
+union cvmx_sso_grpx_ws_pc {
+ u64 u64;
+ struct cvmx_sso_grpx_ws_pc_s {
+ u64 cnt : 64;
+ } s;
+ struct cvmx_sso_grpx_ws_pc_s cn73xx;
+ struct cvmx_sso_grpx_ws_pc_s cn78xx;
+ struct cvmx_sso_grpx_ws_pc_s cn78xxp1;
+ struct cvmx_sso_grpx_ws_pc_s cnf75xx;
+};
+
+typedef union cvmx_sso_grpx_ws_pc cvmx_sso_grpx_ws_pc_t;
+
+/**
+ * cvmx_sso_gw_eco
+ */
+union cvmx_sso_gw_eco {
+ u64 u64;
+ struct cvmx_sso_gw_eco_s {
+ u64 reserved_8_63 : 56;
+ u64 eco_rw : 8;
+ } s;
+ struct cvmx_sso_gw_eco_s cn73xx;
+ struct cvmx_sso_gw_eco_s cnf75xx;
+};
+
+typedef union cvmx_sso_gw_eco cvmx_sso_gw_eco_t;
+
+/**
+ * cvmx_sso_gwe_cfg
+ *
+ * This register controls the operation of the get-work examiner (GWE).
+ *
+ */
+union cvmx_sso_gwe_cfg {
+ u64 u64;
+ struct cvmx_sso_gwe_cfg_s {
+ u64 reserved_12_63 : 52;
+ u64 odu_ffpgw_dis : 1;
+ u64 gwe_rfpgw_dis : 1;
+ u64 odu_prf_dis : 1;
+ u64 reserved_0_8 : 9;
+ } s;
+ struct cvmx_sso_gwe_cfg_cn68xx {
+ u64 reserved_12_63 : 52;
+ u64 odu_ffpgw_dis : 1;
+ u64 gwe_rfpgw_dis : 1;
+ u64 odu_prf_dis : 1;
+ u64 odu_bmp_dis : 1;
+ u64 reserved_5_7 : 3;
+ u64 gwe_hvy_dis : 1;
+ u64 gwe_poe : 1;
+ u64 gwe_fpor : 1;
+ u64 gwe_rah : 1;
+ u64 gwe_dis : 1;
+ } cn68xx;
+ struct cvmx_sso_gwe_cfg_cn68xxp1 {
+ u64 reserved_4_63 : 60;
+ u64 gwe_poe : 1;
+ u64 gwe_fpor : 1;
+ u64 gwe_rah : 1;
+ u64 gwe_dis : 1;
+ } cn68xxp1;
+ struct cvmx_sso_gwe_cfg_cn73xx {
+ u64 reserved_9_63 : 55;
+ u64 dis_wgt_credit : 1;
+ u64 ws_retries : 8;
+ } cn73xx;
+ struct cvmx_sso_gwe_cfg_cn73xx cn78xx;
+ struct cvmx_sso_gwe_cfg_cn73xx cn78xxp1;
+ struct cvmx_sso_gwe_cfg_cn73xx cnf75xx;
+};
+
+typedef union cvmx_sso_gwe_cfg cvmx_sso_gwe_cfg_t;
+
+/**
+ * cvmx_sso_gwe_random
+ *
+ * This register contains the random search start position for the get-work examiner (GWE).
+ *
+ */
+union cvmx_sso_gwe_random {
+ u64 u64;
+ struct cvmx_sso_gwe_random_s {
+ u64 reserved_16_63 : 48;
+ u64 rnd : 16;
+ } s;
+ struct cvmx_sso_gwe_random_s cn73xx;
+ struct cvmx_sso_gwe_random_s cn78xx;
+ struct cvmx_sso_gwe_random_s cn78xxp1;
+ struct cvmx_sso_gwe_random_s cnf75xx;
+};
+
+typedef union cvmx_sso_gwe_random cvmx_sso_gwe_random_t;
+
+/**
+ * cvmx_sso_idx_ecc_ctl
+ *
+ * SSO_IDX_ECC_CTL = SSO IDX ECC Control
+ *
+ */
+union cvmx_sso_idx_ecc_ctl {
+ u64 u64;
+ struct cvmx_sso_idx_ecc_ctl_s {
+ u64 reserved_3_63 : 61;
+ u64 flip_synd : 2;
+ u64 ecc_ena : 1;
+ } s;
+ struct cvmx_sso_idx_ecc_ctl_s cn68xx;
+ struct cvmx_sso_idx_ecc_ctl_s cn68xxp1;
+};
+
+typedef union cvmx_sso_idx_ecc_ctl cvmx_sso_idx_ecc_ctl_t;
+
+/**
+ * cvmx_sso_idx_ecc_st
+ *
+ * SSO_IDX_ECC_ST = SSO IDX ECC Status
+ *
+ */
+union cvmx_sso_idx_ecc_st {
+ u64 u64;
+ struct cvmx_sso_idx_ecc_st_s {
+ u64 reserved_27_63 : 37;
+ u64 addr : 11;
+ u64 reserved_9_15 : 7;
+ u64 syndrom : 5;
+ u64 reserved_0_3 : 4;
+ } s;
+ struct cvmx_sso_idx_ecc_st_s cn68xx;
+ struct cvmx_sso_idx_ecc_st_s cn68xxp1;
+};
+
+typedef union cvmx_sso_idx_ecc_st cvmx_sso_idx_ecc_st_t;
+
+/**
+ * cvmx_sso_ient#_links
+ *
+ * Returns unit memory status for an index.
+ *
+ */
+union cvmx_sso_ientx_links {
+ u64 u64;
+ struct cvmx_sso_ientx_links_s {
+ u64 reserved_28_63 : 36;
+ u64 prev_index : 12;
+ u64 reserved_0_15 : 16;
+ } s;
+ struct cvmx_sso_ientx_links_cn73xx {
+ u64 reserved_26_63 : 38;
+ u64 prev_index : 10;
+ u64 reserved_11_15 : 5;
+ u64 next_index_vld : 1;
+ u64 next_index : 10;
+ } cn73xx;
+ struct cvmx_sso_ientx_links_cn78xx {
+ u64 reserved_28_63 : 36;
+ u64 prev_index : 12;
+ u64 reserved_13_15 : 3;
+ u64 next_index_vld : 1;
+ u64 next_index : 12;
+ } cn78xx;
+ struct cvmx_sso_ientx_links_cn78xx cn78xxp1;
+ struct cvmx_sso_ientx_links_cn73xx cnf75xx;
+};
+
+typedef union cvmx_sso_ientx_links cvmx_sso_ientx_links_t;
+
+/**
+ * cvmx_sso_ient#_pendtag
+ *
+ * Returns unit memory status for an index.
+ *
+ */
+union cvmx_sso_ientx_pendtag {
+ u64 u64;
+ struct cvmx_sso_ientx_pendtag_s {
+ u64 reserved_38_63 : 26;
+ u64 pend_switch : 1;
+ u64 reserved_34_36 : 3;
+ u64 pend_tt : 2;
+ u64 pend_tag : 32;
+ } s;
+ struct cvmx_sso_ientx_pendtag_s cn73xx;
+ struct cvmx_sso_ientx_pendtag_s cn78xx;
+ struct cvmx_sso_ientx_pendtag_s cn78xxp1;
+ struct cvmx_sso_ientx_pendtag_s cnf75xx;
+};
+
+typedef union cvmx_sso_ientx_pendtag cvmx_sso_ientx_pendtag_t;
+
+/**
+ * cvmx_sso_ient#_qlinks
+ *
+ * Returns unit memory status for an index.
+ *
+ */
+union cvmx_sso_ientx_qlinks {
+ u64 u64;
+ struct cvmx_sso_ientx_qlinks_s {
+ u64 reserved_12_63 : 52;
+ u64 next_index : 12;
+ } s;
+ struct cvmx_sso_ientx_qlinks_s cn73xx;
+ struct cvmx_sso_ientx_qlinks_s cn78xx;
+ struct cvmx_sso_ientx_qlinks_s cn78xxp1;
+ struct cvmx_sso_ientx_qlinks_s cnf75xx;
+};
+
+typedef union cvmx_sso_ientx_qlinks cvmx_sso_ientx_qlinks_t;
+
+/**
+ * cvmx_sso_ient#_tag
+ *
+ * Returns unit memory status for an index.
+ *
+ */
+union cvmx_sso_ientx_tag {
+ u64 u64;
+ struct cvmx_sso_ientx_tag_s {
+ u64 reserved_39_63 : 25;
+ u64 tailc : 1;
+ u64 tail : 1;
+ u64 reserved_34_36 : 3;
+ u64 tt : 2;
+ u64 tag : 32;
+ } s;
+ struct cvmx_sso_ientx_tag_s cn73xx;
+ struct cvmx_sso_ientx_tag_s cn78xx;
+ struct cvmx_sso_ientx_tag_s cn78xxp1;
+ struct cvmx_sso_ientx_tag_s cnf75xx;
+};
+
+typedef union cvmx_sso_ientx_tag cvmx_sso_ientx_tag_t;
+
+/**
+ * cvmx_sso_ient#_wqpgrp
+ *
+ * Returns unit memory status for an index.
+ *
+ */
+union cvmx_sso_ientx_wqpgrp {
+ u64 u64;
+ struct cvmx_sso_ientx_wqpgrp_s {
+ u64 reserved_62_63 : 2;
+ u64 head : 1;
+ u64 nosched : 1;
+ u64 reserved_58_59 : 2;
+ u64 grp : 10;
+ u64 reserved_42_47 : 6;
+ u64 wqp : 42;
+ } s;
+ struct cvmx_sso_ientx_wqpgrp_cn73xx {
+ u64 reserved_62_63 : 2;
+ u64 head : 1;
+ u64 nosched : 1;
+ u64 reserved_56_59 : 4;
+ u64 grp : 8;
+ u64 reserved_42_47 : 6;
+ u64 wqp : 42;
+ } cn73xx;
+ struct cvmx_sso_ientx_wqpgrp_s cn78xx;
+ struct cvmx_sso_ientx_wqpgrp_s cn78xxp1;
+ struct cvmx_sso_ientx_wqpgrp_cn73xx cnf75xx;
+};
+
+typedef union cvmx_sso_ientx_wqpgrp cvmx_sso_ientx_wqpgrp_t;
+
+/**
+ * cvmx_sso_ipl_conf#
+ *
+ * Returns list status for the conflicted list indexed by group. Register
+ * fields are identical to those in SSO_IPL_IAQ() above.
+ */
+union cvmx_sso_ipl_confx {
+ u64 u64;
+ struct cvmx_sso_ipl_confx_s {
+ u64 reserved_28_63 : 36;
+ u64 queue_val : 1;
+ u64 queue_one : 1;
+ u64 reserved_25_25 : 1;
+ u64 queue_head : 12;
+ u64 reserved_12_12 : 1;
+ u64 queue_tail : 12;
+ } s;
+ struct cvmx_sso_ipl_confx_s cn73xx;
+ struct cvmx_sso_ipl_confx_s cn78xx;
+ struct cvmx_sso_ipl_confx_s cn78xxp1;
+ struct cvmx_sso_ipl_confx_s cnf75xx;
+};
+
+typedef union cvmx_sso_ipl_confx cvmx_sso_ipl_confx_t;
+
+/**
+ * cvmx_sso_ipl_desched#
+ *
+ * Returns list status for the deschedule list indexed by group. Register
+ * fields are identical to those in SSO_IPL_IAQ() above.
+ */
+union cvmx_sso_ipl_deschedx {
+ u64 u64;
+ struct cvmx_sso_ipl_deschedx_s {
+ u64 reserved_28_63 : 36;
+ u64 queue_val : 1;
+ u64 queue_one : 1;
+ u64 reserved_25_25 : 1;
+ u64 queue_head : 12;
+ u64 reserved_12_12 : 1;
+ u64 queue_tail : 12;
+ } s;
+ struct cvmx_sso_ipl_deschedx_s cn73xx;
+ struct cvmx_sso_ipl_deschedx_s cn78xx;
+ struct cvmx_sso_ipl_deschedx_s cn78xxp1;
+ struct cvmx_sso_ipl_deschedx_s cnf75xx;
+};
+
+typedef union cvmx_sso_ipl_deschedx cvmx_sso_ipl_deschedx_t;
+
+/**
+ * cvmx_sso_ipl_free#
+ *
+ * Returns list status.
+ *
+ */
+union cvmx_sso_ipl_freex {
+ u64 u64;
+ struct cvmx_sso_ipl_freex_s {
+ u64 reserved_62_63 : 2;
+ u64 qnum_head : 3;
+ u64 qnum_tail : 3;
+ u64 reserved_28_55 : 28;
+ u64 queue_val : 1;
+ u64 reserved_25_26 : 2;
+ u64 queue_head : 12;
+ u64 reserved_12_12 : 1;
+ u64 queue_tail : 12;
+ } s;
+ struct cvmx_sso_ipl_freex_cn73xx {
+ u64 reserved_62_63 : 2;
+ u64 qnum_head : 3;
+ u64 qnum_tail : 3;
+ u64 reserved_28_55 : 28;
+ u64 queue_val : 1;
+ u64 reserved_23_26 : 4;
+ u64 queue_head : 10;
+ u64 reserved_10_12 : 3;
+ u64 queue_tail : 10;
+ } cn73xx;
+ struct cvmx_sso_ipl_freex_s cn78xx;
+ struct cvmx_sso_ipl_freex_s cn78xxp1;
+ struct cvmx_sso_ipl_freex_cn73xx cnf75xx;
+};
+
+typedef union cvmx_sso_ipl_freex cvmx_sso_ipl_freex_t;
+
+/**
+ * cvmx_sso_ipl_iaq#
+ *
+ * Returns list status for the internal admission queue indexed by group.
+ *
+ */
+union cvmx_sso_ipl_iaqx {
+ u64 u64;
+ struct cvmx_sso_ipl_iaqx_s {
+ u64 reserved_28_63 : 36;
+ u64 queue_val : 1;
+ u64 queue_one : 1;
+ u64 reserved_25_25 : 1;
+ u64 queue_head : 12;
+ u64 reserved_12_12 : 1;
+ u64 queue_tail : 12;
+ } s;
+ struct cvmx_sso_ipl_iaqx_s cn73xx;
+ struct cvmx_sso_ipl_iaqx_s cn78xx;
+ struct cvmx_sso_ipl_iaqx_s cn78xxp1;
+ struct cvmx_sso_ipl_iaqx_s cnf75xx;
+};
+
+typedef union cvmx_sso_ipl_iaqx cvmx_sso_ipl_iaqx_t;
+
+/**
+ * cvmx_sso_iq_cnt#
+ *
+ * CSR reserved addresses: (64): 0x8200..0x83f8
+ * CSR align addresses: ===========================================================================================================
+ * SSO_IQ_CNTX = SSO Input Queue Count Register
+ * (one per QOS level)
+ *
+ * Contains a read-only count of the number of work queue entries for each QOS
+ * level. Counts both in-unit and in-memory entries.
+ */
+union cvmx_sso_iq_cntx {
+ u64 u64;
+ struct cvmx_sso_iq_cntx_s {
+ u64 reserved_32_63 : 32;
+ u64 iq_cnt : 32;
+ } s;
+ struct cvmx_sso_iq_cntx_s cn68xx;
+ struct cvmx_sso_iq_cntx_s cn68xxp1;
+};
+
+typedef union cvmx_sso_iq_cntx cvmx_sso_iq_cntx_t;
+
+/**
+ * cvmx_sso_iq_com_cnt
+ *
+ * SSO_IQ_COM_CNT = SSO Input Queue Combined Count Register
+ *
+ * Contains a read-only count of the total number of work queue entries in all
+ * QOS levels. Counts both in-unit and in-memory entries.
+ */
+union cvmx_sso_iq_com_cnt {
+ u64 u64;
+ struct cvmx_sso_iq_com_cnt_s {
+ u64 reserved_32_63 : 32;
+ u64 iq_cnt : 32;
+ } s;
+ struct cvmx_sso_iq_com_cnt_s cn68xx;
+ struct cvmx_sso_iq_com_cnt_s cn68xxp1;
+};
+
+typedef union cvmx_sso_iq_com_cnt cvmx_sso_iq_com_cnt_t;
+
+/**
+ * cvmx_sso_iq_int
+ *
+ * SSO_IQ_INT = SSO Input Queue Interrupt Register
+ *
+ * Contains the bits (one per QOS level) that can trigger the input queue
+ * interrupt. An IQ_INT bit will be set if SSO_IQ_CNT#QOS# changes and the
+ * resulting value is equal to SSO_IQ_THR#QOS#.
+ */
+union cvmx_sso_iq_int {
+ u64 u64;
+ struct cvmx_sso_iq_int_s {
+ u64 reserved_8_63 : 56;
+ u64 iq_int : 8;
+ } s;
+ struct cvmx_sso_iq_int_s cn68xx;
+ struct cvmx_sso_iq_int_s cn68xxp1;
+};
+
+typedef union cvmx_sso_iq_int cvmx_sso_iq_int_t;
+
+/**
+ * cvmx_sso_iq_int_en
+ *
+ * SSO_IQ_INT_EN = SSO Input Queue Interrupt Enable Register
+ *
+ * Contains the bits (one per QOS level) that enable the input queue interrupt.
+ */
+union cvmx_sso_iq_int_en {
+ u64 u64;
+ struct cvmx_sso_iq_int_en_s {
+ u64 reserved_8_63 : 56;
+ u64 int_en : 8;
+ } s;
+ struct cvmx_sso_iq_int_en_s cn68xx;
+ struct cvmx_sso_iq_int_en_s cn68xxp1;
+};
+
+typedef union cvmx_sso_iq_int_en cvmx_sso_iq_int_en_t;
+
+/**
+ * cvmx_sso_iq_thr#
+ *
+ * CSR reserved addresses: (24): 0x9040..0x90f8
+ * CSR align addresses: ===========================================================================================================
+ * SSO_IQ_THRX = SSO Input Queue Threshold Register
+ * (one per QOS level)
+ *
+ * Threshold value for triggering input queue interrupts.
+ */
+union cvmx_sso_iq_thrx {
+ u64 u64;
+ struct cvmx_sso_iq_thrx_s {
+ u64 reserved_32_63 : 32;
+ u64 iq_thr : 32;
+ } s;
+ struct cvmx_sso_iq_thrx_s cn68xx;
+ struct cvmx_sso_iq_thrx_s cn68xxp1;
+};
+
+typedef union cvmx_sso_iq_thrx cvmx_sso_iq_thrx_t;
+
+/**
+ * cvmx_sso_nos_cnt
+ *
+ * Contains the number of work-queue entries on the no-schedule list.
+ *
+ */
+union cvmx_sso_nos_cnt {
+ u64 u64;
+ struct cvmx_sso_nos_cnt_s {
+ u64 reserved_13_63 : 51;
+ u64 nos_cnt : 13;
+ } s;
+ struct cvmx_sso_nos_cnt_cn68xx {
+ u64 reserved_12_63 : 52;
+ u64 nos_cnt : 12;
+ } cn68xx;
+ struct cvmx_sso_nos_cnt_cn68xx cn68xxp1;
+ struct cvmx_sso_nos_cnt_s cn73xx;
+ struct cvmx_sso_nos_cnt_s cn78xx;
+ struct cvmx_sso_nos_cnt_s cn78xxp1;
+ struct cvmx_sso_nos_cnt_s cnf75xx;
+};
+
+typedef union cvmx_sso_nos_cnt cvmx_sso_nos_cnt_t;
+
+/**
+ * cvmx_sso_nw_tim
+ *
+ * Sets the minimum period for a new-work-request timeout. The period is specified in n-1
+ * notation, with the increment value of 1024 clock cycles. Thus, a value of 0x0 in this register
+ * translates to 1024 cycles, 0x1 translates to 2048 cycles, 0x2 translates to 3072 cycles, etc.
+ */
+union cvmx_sso_nw_tim {
+ u64 u64;
+ struct cvmx_sso_nw_tim_s {
+ u64 reserved_10_63 : 54;
+ u64 nw_tim : 10;
+ } s;
+ struct cvmx_sso_nw_tim_s cn68xx;
+ struct cvmx_sso_nw_tim_s cn68xxp1;
+ struct cvmx_sso_nw_tim_s cn73xx;
+ struct cvmx_sso_nw_tim_s cn78xx;
+ struct cvmx_sso_nw_tim_s cn78xxp1;
+ struct cvmx_sso_nw_tim_s cnf75xx;
+};
+
+typedef union cvmx_sso_nw_tim cvmx_sso_nw_tim_t;
+
+/**
+ * cvmx_sso_oth_ecc_ctl
+ *
+ * SSO_OTH_ECC_CTL = SSO OTH ECC Control
+ *
+ */
+union cvmx_sso_oth_ecc_ctl {
+ u64 u64;
+ struct cvmx_sso_oth_ecc_ctl_s {
+ u64 reserved_6_63 : 58;
+ u64 flip_synd1 : 2;
+ u64 ecc_ena1 : 1;
+ u64 flip_synd0 : 2;
+ u64 ecc_ena0 : 1;
+ } s;
+ struct cvmx_sso_oth_ecc_ctl_s cn68xx;
+ struct cvmx_sso_oth_ecc_ctl_s cn68xxp1;
+};
+
+typedef union cvmx_sso_oth_ecc_ctl cvmx_sso_oth_ecc_ctl_t;
+
+/**
+ * cvmx_sso_oth_ecc_st
+ *
+ * SSO_OTH_ECC_ST = SSO OTH ECC Status
+ *
+ */
+union cvmx_sso_oth_ecc_st {
+ u64 u64;
+ struct cvmx_sso_oth_ecc_st_s {
+ u64 reserved_59_63 : 5;
+ u64 addr1 : 11;
+ u64 reserved_43_47 : 5;
+ u64 syndrom1 : 7;
+ u64 reserved_27_35 : 9;
+ u64 addr0 : 11;
+ u64 reserved_11_15 : 5;
+ u64 syndrom0 : 7;
+ u64 reserved_0_3 : 4;
+ } s;
+ struct cvmx_sso_oth_ecc_st_s cn68xx;
+ struct cvmx_sso_oth_ecc_st_s cn68xxp1;
+};
+
+typedef union cvmx_sso_oth_ecc_st cvmx_sso_oth_ecc_st_t;
+
+/**
+ * cvmx_sso_page_cnt
+ */
+union cvmx_sso_page_cnt {
+ u64 u64;
+ struct cvmx_sso_page_cnt_s {
+ u64 reserved_32_63 : 32;
+ u64 cnt : 32;
+ } s;
+ struct cvmx_sso_page_cnt_s cn73xx;
+ struct cvmx_sso_page_cnt_s cn78xx;
+ struct cvmx_sso_page_cnt_s cn78xxp1;
+ struct cvmx_sso_page_cnt_s cnf75xx;
+};
+
+typedef union cvmx_sso_page_cnt cvmx_sso_page_cnt_t;
+
+/**
+ * cvmx_sso_pnd_ecc_ctl
+ *
+ * SSO_PND_ECC_CTL = SSO PND ECC Control
+ *
+ */
+union cvmx_sso_pnd_ecc_ctl {
+ u64 u64;
+ struct cvmx_sso_pnd_ecc_ctl_s {
+ u64 reserved_6_63 : 58;
+ u64 flip_synd1 : 2;
+ u64 ecc_ena1 : 1;
+ u64 flip_synd0 : 2;
+ u64 ecc_ena0 : 1;
+ } s;
+ struct cvmx_sso_pnd_ecc_ctl_s cn68xx;
+ struct cvmx_sso_pnd_ecc_ctl_s cn68xxp1;
+};
+
+typedef union cvmx_sso_pnd_ecc_ctl cvmx_sso_pnd_ecc_ctl_t;
+
+/**
+ * cvmx_sso_pnd_ecc_st
+ *
+ * SSO_PND_ECC_ST = SSO PND ECC Status
+ *
+ */
+union cvmx_sso_pnd_ecc_st {
+ u64 u64;
+ struct cvmx_sso_pnd_ecc_st_s {
+ u64 reserved_59_63 : 5;
+ u64 addr1 : 11;
+ u64 reserved_43_47 : 5;
+ u64 syndrom1 : 7;
+ u64 reserved_27_35 : 9;
+ u64 addr0 : 11;
+ u64 reserved_11_15 : 5;
+ u64 syndrom0 : 7;
+ u64 reserved_0_3 : 4;
+ } s;
+ struct cvmx_sso_pnd_ecc_st_s cn68xx;
+ struct cvmx_sso_pnd_ecc_st_s cn68xxp1;
+};
+
+typedef union cvmx_sso_pnd_ecc_st cvmx_sso_pnd_ecc_st_t;
+
+/**
+ * cvmx_sso_pp#_arb
+ *
+ * For diagnostic use, returns the group affinity arbitration state for each core.
+ *
+ */
+union cvmx_sso_ppx_arb {
+ u64 u64;
+ struct cvmx_sso_ppx_arb_s {
+ u64 reserved_20_63 : 44;
+ u64 aff_left : 4;
+ u64 reserved_8_15 : 8;
+ u64 last_grp : 8;
+ } s;
+ struct cvmx_sso_ppx_arb_s cn73xx;
+ struct cvmx_sso_ppx_arb_s cn78xx;
+ struct cvmx_sso_ppx_arb_s cn78xxp1;
+ struct cvmx_sso_ppx_arb_s cnf75xx;
+};
+
+typedef union cvmx_sso_ppx_arb cvmx_sso_ppx_arb_t;
+
+/**
+ * cvmx_sso_pp#_grp_msk
+ *
+ * CSR reserved addresses: (24): 0x5040..0x50f8
+ * CSR align addresses: ===========================================================================================================
+ * SSO_PPX_GRP_MSK = SSO PP Group Mask Register
+ * (one bit per group per PP)
+ *
+ * Selects which group(s) a PP belongs to. A '1' in any bit position sets the
+ * PP's membership in the corresponding group. A value of 0x0 will prevent the
+ * PP from receiving new work.
+ *
+ * Note that these do not contain QOS level priorities for each PP. This is a
+ * change from previous POW designs.
+ */
+union cvmx_sso_ppx_grp_msk {
+ u64 u64;
+ struct cvmx_sso_ppx_grp_msk_s {
+ u64 grp_msk : 64;
+ } s;
+ struct cvmx_sso_ppx_grp_msk_s cn68xx;
+ struct cvmx_sso_ppx_grp_msk_s cn68xxp1;
+};
+
+typedef union cvmx_sso_ppx_grp_msk cvmx_sso_ppx_grp_msk_t;
+
+/**
+ * cvmx_sso_pp#_qos_pri
+ *
+ * CSR reserved addresses: (56): 0x2040..0x21f8
+ * CSR align addresses: ===========================================================================================================
+ * SSO_PP(0..31)_QOS_PRI = SSO PP QOS Priority Register
+ * (one field per IQ per PP)
+ *
+ * Contains the QOS level priorities for each PP.
+ * 0x0 is the highest priority
+ * 0x7 is the lowest priority
+ * 0xf prevents the PP from receiving work from that QOS level
+ * 0x8-0xe Reserved
+ *
+ * For a given PP, priorities should begin at 0x0, and remain contiguous
+ * throughout the range. Failure to do so may result in severe
+ * performance degradation.
+ *
+ *
+ * Priorities for IQs 0..7
+ */
+union cvmx_sso_ppx_qos_pri {
+ u64 u64;
+ struct cvmx_sso_ppx_qos_pri_s {
+ u64 reserved_60_63 : 4;
+ u64 qos7_pri : 4;
+ u64 reserved_52_55 : 4;
+ u64 qos6_pri : 4;
+ u64 reserved_44_47 : 4;
+ u64 qos5_pri : 4;
+ u64 reserved_36_39 : 4;
+ u64 qos4_pri : 4;
+ u64 reserved_28_31 : 4;
+ u64 qos3_pri : 4;
+ u64 reserved_20_23 : 4;
+ u64 qos2_pri : 4;
+ u64 reserved_12_15 : 4;
+ u64 qos1_pri : 4;
+ u64 reserved_4_7 : 4;
+ u64 qos0_pri : 4;
+ } s;
+ struct cvmx_sso_ppx_qos_pri_s cn68xx;
+ struct cvmx_sso_ppx_qos_pri_s cn68xxp1;
+};
+
+typedef union cvmx_sso_ppx_qos_pri cvmx_sso_ppx_qos_pri_t;
+
+/**
+ * cvmx_sso_pp#_s#_grpmsk#
+ *
+ * These registers select which group or groups a core belongs to. There are 2 sets of masks per
+ * core, each with 1 register corresponding to 64 groups.
+ */
+union cvmx_sso_ppx_sx_grpmskx {
+ u64 u64;
+ struct cvmx_sso_ppx_sx_grpmskx_s {
+ u64 grp_msk : 64;
+ } s;
+ struct cvmx_sso_ppx_sx_grpmskx_s cn73xx;
+ struct cvmx_sso_ppx_sx_grpmskx_s cn78xx;
+ struct cvmx_sso_ppx_sx_grpmskx_s cn78xxp1;
+ struct cvmx_sso_ppx_sx_grpmskx_s cnf75xx;
+};
+
+typedef union cvmx_sso_ppx_sx_grpmskx cvmx_sso_ppx_sx_grpmskx_t;
+
+/**
+ * cvmx_sso_pp_strict
+ *
+ * SSO_PP_STRICT = SSO Strict Priority
+ *
+ * This register controls getting work from the input queues. If the bit
+ * corresponding to a PP is set, that PP will not take work off the input
+ * queues until it is known that there is no higher-priority work available.
+ *
+ * Setting SSO_PP_STRICT may incur a performance penalty if highest-priority
+ * work is not found early.
+ *
+ * It is possible to starve a PP of work with SSO_PP_STRICT. If the
+ * SSO_PPX_GRP_MSK for a PP masks-out much of the work added to the input
+ * queues that are higher-priority for that PP, and if there is a constant
+ * stream of work through one or more of those higher-priority input queues,
+ * then that PP may not accept work from lower-priority input queues. This can
+ * be alleviated by ensuring that most or all the work added to the
+ * higher-priority input queues for a PP with SSO_PP_STRICT set are in a group
+ * acceptable to that PP.
+ *
+ * It is also possible to neglect work in an input queue if SSO_PP_STRICT is
+ * used. If an input queue is a lower-priority queue for all PPs, and if all
+ * the PPs have their corresponding bit in SSO_PP_STRICT set, then work may
+ * never be taken (or be seldom taken) from that queue. This can be alleviated
+ * by ensuring that work in all input queues can be serviced by one or more PPs
+ * that do not have SSO_PP_STRICT set, or that the input queue is the
+ * highest-priority input queue for one or more PPs that do have SSO_PP_STRICT
+ * set.
+ */
+union cvmx_sso_pp_strict {
+ u64 u64;
+ struct cvmx_sso_pp_strict_s {
+ u64 reserved_32_63 : 32;
+ u64 pp_strict : 32;
+ } s;
+ struct cvmx_sso_pp_strict_s cn68xx;
+ struct cvmx_sso_pp_strict_s cn68xxp1;
+};
+
+typedef union cvmx_sso_pp_strict cvmx_sso_pp_strict_t;
+
+/**
+ * cvmx_sso_qos#_rnd
+ *
+ * CSR align addresses: ===========================================================================================================
+ * SSO_QOS(0..7)_RND = SSO QOS Issue Round Register
+ * (one per IQ)
+ *
+ * The number of arbitration rounds each QOS level participates in.
+ */
+union cvmx_sso_qosx_rnd {
+ u64 u64;
+ struct cvmx_sso_qosx_rnd_s {
+ u64 reserved_8_63 : 56;
+ u64 rnds_qos : 8;
+ } s;
+ struct cvmx_sso_qosx_rnd_s cn68xx;
+ struct cvmx_sso_qosx_rnd_s cn68xxp1;
+};
+
+typedef union cvmx_sso_qosx_rnd cvmx_sso_qosx_rnd_t;
+
+/**
+ * cvmx_sso_qos_thr#
+ *
+ * CSR reserved addresses: (24): 0xa040..0xa0f8
+ * CSR align addresses: ===========================================================================================================
+ * SSO_QOS_THRX = SSO QOS Threshold Register
+ * (one per QOS level)
+ *
+ * Contains the thresholds for allocating SSO internal storage buffers. If the
+ * number of remaining free buffers drops below the minimum threshold (MIN_THR)
+ * or the number of allocated buffers for this QOS level rises above the
+ * maximum threshold (MAX_THR), future incoming work queue entries will be
+ * buffered externally rather than internally. This register also contains the
+ * number of internal buffers currently allocated to this QOS level (BUF_CNT).
+ */
+union cvmx_sso_qos_thrx {
+ u64 u64;
+ struct cvmx_sso_qos_thrx_s {
+ u64 reserved_40_63 : 24;
+ u64 buf_cnt : 12;
+ u64 reserved_26_27 : 2;
+ u64 max_thr : 12;
+ u64 reserved_12_13 : 2;
+ u64 min_thr : 12;
+ } s;
+ struct cvmx_sso_qos_thrx_s cn68xx;
+ struct cvmx_sso_qos_thrx_s cn68xxp1;
+};
+
+typedef union cvmx_sso_qos_thrx cvmx_sso_qos_thrx_t;
+
+/**
+ * cvmx_sso_qos_we
+ *
+ * SSO_QOS_WE = SSO WE Buffers
+ *
+ * This register contains a read-only count of the current number of free
+ * buffers (FREE_CNT) and the total number of tag chain heads on the de-schedule list
+ * (DES_CNT) (which is not the same as the total number of entries on all of the descheduled
+ * tag chains.)
+ */
+union cvmx_sso_qos_we {
+ u64 u64;
+ struct cvmx_sso_qos_we_s {
+ u64 reserved_26_63 : 38;
+ u64 des_cnt : 12;
+ u64 reserved_12_13 : 2;
+ u64 free_cnt : 12;
+ } s;
+ struct cvmx_sso_qos_we_s cn68xx;
+ struct cvmx_sso_qos_we_s cn68xxp1;
+};
+
+typedef union cvmx_sso_qos_we cvmx_sso_qos_we_t;
+
+/**
+ * cvmx_sso_reset
+ *
+ * Writing a 1 to SSO_RESET[RESET] resets the SSO. After receiving a store to this CSR, the SSO
+ * must not be sent any other operations for 2500 coprocessor (SCLK) cycles. Note that the
+ * contents of this register are reset along with the rest of the SSO.
+ */
+union cvmx_sso_reset {
+ u64 u64;
+ struct cvmx_sso_reset_s {
+ u64 busy : 1;
+ u64 reserved_1_62 : 62;
+ u64 reset : 1;
+ } s;
+ struct cvmx_sso_reset_cn68xx {
+ u64 reserved_1_63 : 63;
+ u64 reset : 1;
+ } cn68xx;
+ struct cvmx_sso_reset_s cn73xx;
+ struct cvmx_sso_reset_s cn78xx;
+ struct cvmx_sso_reset_s cn78xxp1;
+ struct cvmx_sso_reset_s cnf75xx;
+};
+
+typedef union cvmx_sso_reset cvmx_sso_reset_t;
+
+/**
+ * cvmx_sso_rwq_head_ptr#
+ *
+ * CSR reserved addresses: (24): 0xb040..0xb0f8
+ * CSR align addresses: ===========================================================================================================
+ * SSO_RWQ_HEAD_PTRX = SSO Remote Queue Head Register
+ * (one per QOS level)
+ * Contains the ptr to the first entry of the remote linked list(s) for a particular
+ * QoS level. SW should initialize the remote linked list(s) by programming
+ * SSO_RWQ_HEAD_PTRX and SSO_RWQ_TAIL_PTRX to identical values.
+ */
+union cvmx_sso_rwq_head_ptrx {
+ u64 u64;
+ struct cvmx_sso_rwq_head_ptrx_s {
+ u64 reserved_38_63 : 26;
+ u64 ptr : 31;
+ u64 reserved_5_6 : 2;
+ u64 rctr : 5;
+ } s;
+ struct cvmx_sso_rwq_head_ptrx_s cn68xx;
+ struct cvmx_sso_rwq_head_ptrx_s cn68xxp1;
+};
+
+typedef union cvmx_sso_rwq_head_ptrx cvmx_sso_rwq_head_ptrx_t;
+
+/**
+ * cvmx_sso_rwq_pop_fptr
+ *
+ * SSO_RWQ_POP_FPTR = SSO Pop Free Pointer
+ *
+ * This register is used by SW to remove pointers for buffer-reallocation and diagnostics, and
+ * should only be used when SSO is idle.
+ *
+ * To remove ALL pointers, software must insure that there are modulus 16
+ * pointers in the FPA. To do this, SSO_CFG.RWQ_BYP_DIS must be set, the FPA
+ * pointer count read, and enough fake buffers pushed via SSO_RWQ_PSH_FPTR to
+ * bring the FPA pointer count up to mod 16.
+ */
+union cvmx_sso_rwq_pop_fptr {
+ u64 u64;
+ struct cvmx_sso_rwq_pop_fptr_s {
+ u64 val : 1;
+ u64 reserved_38_62 : 25;
+ u64 fptr : 31;
+ u64 reserved_0_6 : 7;
+ } s;
+ struct cvmx_sso_rwq_pop_fptr_s cn68xx;
+ struct cvmx_sso_rwq_pop_fptr_s cn68xxp1;
+};
+
+typedef union cvmx_sso_rwq_pop_fptr cvmx_sso_rwq_pop_fptr_t;
+
+/**
+ * cvmx_sso_rwq_psh_fptr
+ *
+ * CSR reserved addresses: (56): 0xc240..0xc3f8
+ * SSO_RWQ_PSH_FPTR = SSO Free Pointer FIFO
+ *
+ * This register is used by SW to initialize the SSO with a pool of free
+ * pointers by writing the FPTR field whenever FULL = 0. Free pointers are
+ * fetched/released from/to the pool when accessing WQE entries stored remotely
+ * (in remote linked lists). Free pointers should be 128 byte aligned, each of
+ * 256 bytes. This register should only be used when SSO is idle.
+ *
+ * Software needs to set aside buffering for
+ * 8 + 48 + ROUNDUP(N/26)
+ *
+ * where as many as N DRAM work queue entries may be used. The first 8 buffers
+ * are used to setup the SSO_RWQ_HEAD_PTR and SSO_RWQ_TAIL_PTRs, and the
+ * remainder are pushed via this register.
+ *
+ * IMPLEMENTATION NOTES--NOT FOR SPEC:
+ * 48 avoids false out of buffer error due to (16) FPA and in-sso FPA buffering (32)
+ * 26 is number of WAE's per 256B buffer
+ */
+union cvmx_sso_rwq_psh_fptr {
+ u64 u64;
+ struct cvmx_sso_rwq_psh_fptr_s {
+ u64 full : 1;
+ u64 reserved_38_62 : 25;
+ u64 fptr : 31;
+ u64 reserved_0_6 : 7;
+ } s;
+ struct cvmx_sso_rwq_psh_fptr_s cn68xx;
+ struct cvmx_sso_rwq_psh_fptr_s cn68xxp1;
+};
+
+typedef union cvmx_sso_rwq_psh_fptr cvmx_sso_rwq_psh_fptr_t;
+
+/**
+ * cvmx_sso_rwq_tail_ptr#
+ *
+ * CSR reserved addresses: (56): 0xc040..0xc1f8
+ * SSO_RWQ_TAIL_PTRX = SSO Remote Queue Tail Register
+ * (one per QOS level)
+ * Contains the ptr to the last entry of the remote linked list(s) for a particular
+ * QoS level. SW must initialize the remote linked list(s) by programming
+ * SSO_RWQ_HEAD_PTRX and SSO_RWQ_TAIL_PTRX to identical values.
+ */
+union cvmx_sso_rwq_tail_ptrx {
+ u64 u64;
+ struct cvmx_sso_rwq_tail_ptrx_s {
+ u64 reserved_38_63 : 26;
+ u64 ptr : 31;
+ u64 reserved_5_6 : 2;
+ u64 rctr : 5;
+ } s;
+ struct cvmx_sso_rwq_tail_ptrx_s cn68xx;
+ struct cvmx_sso_rwq_tail_ptrx_s cn68xxp1;
+};
+
+typedef union cvmx_sso_rwq_tail_ptrx cvmx_sso_rwq_tail_ptrx_t;
+
+/**
+ * cvmx_sso_sl_pp#_links
+ *
+ * Returns status of each core.
+ *
+ */
+union cvmx_sso_sl_ppx_links {
+ u64 u64;
+ struct cvmx_sso_sl_ppx_links_s {
+ u64 tailc : 1;
+ u64 reserved_60_62 : 3;
+ u64 index : 12;
+ u64 reserved_38_47 : 10;
+ u64 grp : 10;
+ u64 head : 1;
+ u64 tail : 1;
+ u64 reserved_0_25 : 26;
+ } s;
+ struct cvmx_sso_sl_ppx_links_cn73xx {
+ u64 tailc : 1;
+ u64 reserved_58_62 : 5;
+ u64 index : 10;
+ u64 reserved_36_47 : 12;
+ u64 grp : 8;
+ u64 head : 1;
+ u64 tail : 1;
+ u64 reserved_21_25 : 5;
+ u64 revlink_index : 10;
+ u64 link_index_vld : 1;
+ u64 link_index : 10;
+ } cn73xx;
+ struct cvmx_sso_sl_ppx_links_cn78xx {
+ u64 tailc : 1;
+ u64 reserved_60_62 : 3;
+ u64 index : 12;
+ u64 reserved_38_47 : 10;
+ u64 grp : 10;
+ u64 head : 1;
+ u64 tail : 1;
+ u64 reserved_25_25 : 1;
+ u64 revlink_index : 12;
+ u64 link_index_vld : 1;
+ u64 link_index : 12;
+ } cn78xx;
+ struct cvmx_sso_sl_ppx_links_cn78xx cn78xxp1;
+ struct cvmx_sso_sl_ppx_links_cn73xx cnf75xx;
+};
+
+typedef union cvmx_sso_sl_ppx_links cvmx_sso_sl_ppx_links_t;
+
+/**
+ * cvmx_sso_sl_pp#_pendtag
+ *
+ * Returns status of each core.
+ *
+ */
+union cvmx_sso_sl_ppx_pendtag {
+ u64 u64;
+ struct cvmx_sso_sl_ppx_pendtag_s {
+ u64 pend_switch : 1;
+ u64 pend_get_work : 1;
+ u64 pend_get_work_wait : 1;
+ u64 pend_nosched : 1;
+ u64 pend_nosched_clr : 1;
+ u64 pend_desched : 1;
+ u64 pend_alloc_we : 1;
+ u64 pend_gw_insert : 1;
+ u64 reserved_34_55 : 22;
+ u64 pend_tt : 2;
+ u64 pend_tag : 32;
+ } s;
+ struct cvmx_sso_sl_ppx_pendtag_s cn73xx;
+ struct cvmx_sso_sl_ppx_pendtag_s cn78xx;
+ struct cvmx_sso_sl_ppx_pendtag_s cn78xxp1;
+ struct cvmx_sso_sl_ppx_pendtag_s cnf75xx;
+};
+
+typedef union cvmx_sso_sl_ppx_pendtag cvmx_sso_sl_ppx_pendtag_t;
+
+/**
+ * cvmx_sso_sl_pp#_pendwqp
+ *
+ * Returns status of each core.
+ *
+ */
+union cvmx_sso_sl_ppx_pendwqp {
+ u64 u64;
+ struct cvmx_sso_sl_ppx_pendwqp_s {
+ u64 pend_switch : 1;
+ u64 pend_get_work : 1;
+ u64 pend_get_work_wait : 1;
+ u64 pend_nosched : 1;
+ u64 pend_nosched_clr : 1;
+ u64 pend_desched : 1;
+ u64 pend_alloc_we : 1;
+ u64 reserved_56_56 : 1;
+ u64 pend_index : 12;
+ u64 reserved_42_43 : 2;
+ u64 pend_wqp : 42;
+ } s;
+ struct cvmx_sso_sl_ppx_pendwqp_cn73xx {
+ u64 pend_switch : 1;
+ u64 pend_get_work : 1;
+ u64 pend_get_work_wait : 1;
+ u64 pend_nosched : 1;
+ u64 pend_nosched_clr : 1;
+ u64 pend_desched : 1;
+ u64 pend_alloc_we : 1;
+ u64 reserved_54_56 : 3;
+ u64 pend_index : 10;
+ u64 reserved_42_43 : 2;
+ u64 pend_wqp : 42;
+ } cn73xx;
+ struct cvmx_sso_sl_ppx_pendwqp_s cn78xx;
+ struct cvmx_sso_sl_ppx_pendwqp_s cn78xxp1;
+ struct cvmx_sso_sl_ppx_pendwqp_cn73xx cnf75xx;
+};
+
+typedef union cvmx_sso_sl_ppx_pendwqp cvmx_sso_sl_ppx_pendwqp_t;
+
+/**
+ * cvmx_sso_sl_pp#_tag
+ *
+ * Returns status of each core.
+ *
+ */
+union cvmx_sso_sl_ppx_tag {
+ u64 u64;
+ struct cvmx_sso_sl_ppx_tag_s {
+ u64 tailc : 1;
+ u64 reserved_60_62 : 3;
+ u64 index : 12;
+ u64 reserved_46_47 : 2;
+ u64 grp : 10;
+ u64 head : 1;
+ u64 tail : 1;
+ u64 tt : 2;
+ u64 tag : 32;
+ } s;
+ struct cvmx_sso_sl_ppx_tag_cn73xx {
+ u64 tailc : 1;
+ u64 reserved_58_62 : 5;
+ u64 index : 10;
+ u64 reserved_44_47 : 4;
+ u64 grp : 8;
+ u64 head : 1;
+ u64 tail : 1;
+ u64 tt : 2;
+ u64 tag : 32;
+ } cn73xx;
+ struct cvmx_sso_sl_ppx_tag_s cn78xx;
+ struct cvmx_sso_sl_ppx_tag_s cn78xxp1;
+ struct cvmx_sso_sl_ppx_tag_cn73xx cnf75xx;
+};
+
+typedef union cvmx_sso_sl_ppx_tag cvmx_sso_sl_ppx_tag_t;
+
+/**
+ * cvmx_sso_sl_pp#_wqp
+ *
+ * Returns status of each core.
+ *
+ */
+union cvmx_sso_sl_ppx_wqp {
+ u64 u64;
+ struct cvmx_sso_sl_ppx_wqp_s {
+ u64 reserved_58_63 : 6;
+ u64 grp : 10;
+ u64 reserved_42_47 : 6;
+ u64 wqp : 42;
+ } s;
+ struct cvmx_sso_sl_ppx_wqp_cn73xx {
+ u64 reserved_56_63 : 8;
+ u64 grp : 8;
+ u64 reserved_42_47 : 6;
+ u64 wqp : 42;
+ } cn73xx;
+ struct cvmx_sso_sl_ppx_wqp_s cn78xx;
+ struct cvmx_sso_sl_ppx_wqp_s cn78xxp1;
+ struct cvmx_sso_sl_ppx_wqp_cn73xx cnf75xx;
+};
+
+typedef union cvmx_sso_sl_ppx_wqp cvmx_sso_sl_ppx_wqp_t;
+
+/**
+ * cvmx_sso_taq#_link
+ *
+ * Returns TAQ status for a given line.
+ *
+ */
+union cvmx_sso_taqx_link {
+ u64 u64;
+ struct cvmx_sso_taqx_link_s {
+ u64 reserved_11_63 : 53;
+ u64 next : 11;
+ } s;
+ struct cvmx_sso_taqx_link_s cn73xx;
+ struct cvmx_sso_taqx_link_s cn78xx;
+ struct cvmx_sso_taqx_link_s cn78xxp1;
+ struct cvmx_sso_taqx_link_s cnf75xx;
+};
+
+typedef union cvmx_sso_taqx_link cvmx_sso_taqx_link_t;
+
+/**
+ * cvmx_sso_taq#_wae#_tag
+ *
+ * Returns TAQ status for a given line and WAE within that line.
+ *
+ */
+union cvmx_sso_taqx_waex_tag {
+ u64 u64;
+ struct cvmx_sso_taqx_waex_tag_s {
+ u64 reserved_34_63 : 30;
+ u64 tt : 2;
+ u64 tag : 32;
+ } s;
+ struct cvmx_sso_taqx_waex_tag_s cn73xx;
+ struct cvmx_sso_taqx_waex_tag_s cn78xx;
+ struct cvmx_sso_taqx_waex_tag_s cn78xxp1;
+ struct cvmx_sso_taqx_waex_tag_s cnf75xx;
+};
+
+typedef union cvmx_sso_taqx_waex_tag cvmx_sso_taqx_waex_tag_t;
+
+/**
+ * cvmx_sso_taq#_wae#_wqp
+ *
+ * Returns TAQ status for a given line and WAE within that line.
+ *
+ */
+union cvmx_sso_taqx_waex_wqp {
+ u64 u64;
+ struct cvmx_sso_taqx_waex_wqp_s {
+ u64 reserved_42_63 : 22;
+ u64 wqp : 42;
+ } s;
+ struct cvmx_sso_taqx_waex_wqp_s cn73xx;
+ struct cvmx_sso_taqx_waex_wqp_s cn78xx;
+ struct cvmx_sso_taqx_waex_wqp_s cn78xxp1;
+ struct cvmx_sso_taqx_waex_wqp_s cnf75xx;
+};
+
+typedef union cvmx_sso_taqx_waex_wqp cvmx_sso_taqx_waex_wqp_t;
+
+/**
+ * cvmx_sso_taq_add
+ */
+union cvmx_sso_taq_add {
+ u64 u64;
+ struct cvmx_sso_taq_add_s {
+ u64 reserved_29_63 : 35;
+ u64 rsvd_free : 13;
+ u64 reserved_0_15 : 16;
+ } s;
+ struct cvmx_sso_taq_add_s cn73xx;
+ struct cvmx_sso_taq_add_s cn78xx;
+ struct cvmx_sso_taq_add_s cn78xxp1;
+ struct cvmx_sso_taq_add_s cnf75xx;
+};
+
+typedef union cvmx_sso_taq_add cvmx_sso_taq_add_t;
+
+/**
+ * cvmx_sso_taq_cnt
+ */
+union cvmx_sso_taq_cnt {
+ u64 u64;
+ struct cvmx_sso_taq_cnt_s {
+ u64 reserved_27_63 : 37;
+ u64 rsvd_free : 11;
+ u64 reserved_11_15 : 5;
+ u64 free_cnt : 11;
+ } s;
+ struct cvmx_sso_taq_cnt_s cn73xx;
+ struct cvmx_sso_taq_cnt_s cn78xx;
+ struct cvmx_sso_taq_cnt_s cn78xxp1;
+ struct cvmx_sso_taq_cnt_s cnf75xx;
+};
+
+typedef union cvmx_sso_taq_cnt cvmx_sso_taq_cnt_t;
+
+/**
+ * cvmx_sso_tiaq#_status
+ *
+ * Returns TAQ inbound status indexed by group.
+ *
+ */
+union cvmx_sso_tiaqx_status {
+ u64 u64;
+ struct cvmx_sso_tiaqx_status_s {
+ u64 wae_head : 4;
+ u64 wae_tail : 4;
+ u64 reserved_47_55 : 9;
+ u64 wae_used : 15;
+ u64 reserved_23_31 : 9;
+ u64 ent_head : 11;
+ u64 reserved_11_11 : 1;
+ u64 ent_tail : 11;
+ } s;
+ struct cvmx_sso_tiaqx_status_s cn73xx;
+ struct cvmx_sso_tiaqx_status_s cn78xx;
+ struct cvmx_sso_tiaqx_status_s cn78xxp1;
+ struct cvmx_sso_tiaqx_status_s cnf75xx;
+};
+
+typedef union cvmx_sso_tiaqx_status cvmx_sso_tiaqx_status_t;
+
+/**
+ * cvmx_sso_toaq#_status
+ *
+ * Returns TAQ outbound status indexed by group.
+ *
+ */
+union cvmx_sso_toaqx_status {
+ u64 u64;
+ struct cvmx_sso_toaqx_status_s {
+ u64 reserved_62_63 : 2;
+ u64 ext_vld : 1;
+ u64 partial : 1;
+ u64 wae_tail : 4;
+ u64 reserved_43_55 : 13;
+ u64 cl_used : 11;
+ u64 reserved_23_31 : 9;
+ u64 ent_head : 11;
+ u64 reserved_11_11 : 1;
+ u64 ent_tail : 11;
+ } s;
+ struct cvmx_sso_toaqx_status_s cn73xx;
+ struct cvmx_sso_toaqx_status_s cn78xx;
+ struct cvmx_sso_toaqx_status_s cn78xxp1;
+ struct cvmx_sso_toaqx_status_s cnf75xx;
+};
+
+typedef union cvmx_sso_toaqx_status cvmx_sso_toaqx_status_t;
+
+/**
+ * cvmx_sso_ts_pc
+ *
+ * SSO_TS_PC = SSO Tag Switch Performance Counter
+ *
+ * Counts the number of tag switch requests.
+ * Counter rolls over through zero when max value exceeded.
+ */
+union cvmx_sso_ts_pc {
+ u64 u64;
+ struct cvmx_sso_ts_pc_s {
+ u64 ts_pc : 64;
+ } s;
+ struct cvmx_sso_ts_pc_s cn68xx;
+ struct cvmx_sso_ts_pc_s cn68xxp1;
+};
+
+typedef union cvmx_sso_ts_pc cvmx_sso_ts_pc_t;
+
+/**
+ * cvmx_sso_wa_com_pc
+ *
+ * SSO_WA_COM_PC = SSO Work Add Combined Performance Counter
+ *
+ * Counts the number of add new work requests for all QOS levels.
+ * Counter rolls over through zero when max value exceeded.
+ */
+union cvmx_sso_wa_com_pc {
+ u64 u64;
+ struct cvmx_sso_wa_com_pc_s {
+ u64 wa_pc : 64;
+ } s;
+ struct cvmx_sso_wa_com_pc_s cn68xx;
+ struct cvmx_sso_wa_com_pc_s cn68xxp1;
+};
+
+typedef union cvmx_sso_wa_com_pc cvmx_sso_wa_com_pc_t;
+
+/**
+ * cvmx_sso_wa_pc#
+ *
+ * CSR reserved addresses: (64): 0x4200..0x43f8
+ * CSR align addresses: ===========================================================================================================
+ * SSO_WA_PCX = SSO Work Add Performance Counter
+ * (one per QOS level)
+ *
+ * Counts the number of add new work requests for each QOS level.
+ * Counter rolls over through zero when max value exceeded.
+ */
+union cvmx_sso_wa_pcx {
+ u64 u64;
+ struct cvmx_sso_wa_pcx_s {
+ u64 wa_pc : 64;
+ } s;
+ struct cvmx_sso_wa_pcx_s cn68xx;
+ struct cvmx_sso_wa_pcx_s cn68xxp1;
+};
+
+typedef union cvmx_sso_wa_pcx cvmx_sso_wa_pcx_t;
+
+/**
+ * cvmx_sso_wq_int
+ *
+ * Note, the old POW offsets ran from 0x0 to 0x3f8, leaving the next available slot at 0x400.
+ * To ensure no overlap, start on 4k boundary: 0x1000.
+ * SSO_WQ_INT = SSO Work Queue Interrupt Register
+ *
+ * Contains the bits (one per group) that set work queue interrupts and are
+ * used to clear these interrupts. For more information regarding this
+ * register, see the interrupt section of the SSO spec.
+ */
+union cvmx_sso_wq_int {
+ u64 u64;
+ struct cvmx_sso_wq_int_s {
+ u64 wq_int : 64;
+ } s;
+ struct cvmx_sso_wq_int_s cn68xx;
+ struct cvmx_sso_wq_int_s cn68xxp1;
+};
+
+typedef union cvmx_sso_wq_int cvmx_sso_wq_int_t;
+
+/**
+ * cvmx_sso_wq_int_cnt#
+ *
+ * CSR reserved addresses: (64): 0x7200..0x73f8
+ * CSR align addresses: ===========================================================================================================
+ * SSO_WQ_INT_CNTX = SSO Work Queue Interrupt Count Register
+ * (one per group)
+ *
+ * Contains a read-only copy of the counts used to trigger work queue
+ * interrupts. For more information regarding this register, see the interrupt
+ * section.
+ */
+union cvmx_sso_wq_int_cntx {
+ u64 u64;
+ struct cvmx_sso_wq_int_cntx_s {
+ u64 reserved_32_63 : 32;
+ u64 tc_cnt : 4;
+ u64 reserved_26_27 : 2;
+ u64 ds_cnt : 12;
+ u64 reserved_12_13 : 2;
+ u64 iq_cnt : 12;
+ } s;
+ struct cvmx_sso_wq_int_cntx_s cn68xx;
+ struct cvmx_sso_wq_int_cntx_s cn68xxp1;
+};
+
+typedef union cvmx_sso_wq_int_cntx cvmx_sso_wq_int_cntx_t;
+
+/**
+ * cvmx_sso_wq_int_pc
+ *
+ * Contains the threshold value for the work-executable interrupt periodic counter and also a
+ * read-only copy of the periodic counter. For more information on this register, refer to
+ * Interrupts.
+ */
+union cvmx_sso_wq_int_pc {
+ u64 u64;
+ struct cvmx_sso_wq_int_pc_s {
+ u64 reserved_60_63 : 4;
+ u64 pc : 28;
+ u64 reserved_28_31 : 4;
+ u64 pc_thr : 20;
+ u64 reserved_0_7 : 8;
+ } s;
+ struct cvmx_sso_wq_int_pc_s cn68xx;
+ struct cvmx_sso_wq_int_pc_s cn68xxp1;
+ struct cvmx_sso_wq_int_pc_s cn73xx;
+ struct cvmx_sso_wq_int_pc_s cn78xx;
+ struct cvmx_sso_wq_int_pc_s cn78xxp1;
+ struct cvmx_sso_wq_int_pc_s cnf75xx;
+};
+
+typedef union cvmx_sso_wq_int_pc cvmx_sso_wq_int_pc_t;
+
+/**
+ * cvmx_sso_wq_int_thr#
+ *
+ * CSR reserved addresses: (96): 0x6100..0x63f8
+ * CSR align addresses: ===========================================================================================================
+ * SSO_WQ_INT_THR(0..63) = SSO Work Queue Interrupt Threshold Registers
+ * (one per group)
+ *
+ * Contains the thresholds for enabling and setting work queue interrupts. For
+ * more information, see the interrupt section.
+ *
+ * Note: Up to 16 of the SSO's internal storage buffers can be allocated
+ * for hardware use and are therefore not available for incoming work queue
+ * entries. Additionally, any WS that is not in the EMPTY state consumes a
+ * buffer. Thus in a 32 PP system, it is not advisable to set either IQ_THR or
+ * DS_THR to greater than 2048 - 16 - 32*2 = 1968. Doing so may prevent the
+ * interrupt from ever triggering.
+ *
+ * Priorities for QOS levels 0..7
+ */
+union cvmx_sso_wq_int_thrx {
+ u64 u64;
+ struct cvmx_sso_wq_int_thrx_s {
+ u64 reserved_33_63 : 31;
+ u64 tc_en : 1;
+ u64 tc_thr : 4;
+ u64 reserved_26_27 : 2;
+ u64 ds_thr : 12;
+ u64 reserved_12_13 : 2;
+ u64 iq_thr : 12;
+ } s;
+ struct cvmx_sso_wq_int_thrx_s cn68xx;
+ struct cvmx_sso_wq_int_thrx_s cn68xxp1;
+};
+
+typedef union cvmx_sso_wq_int_thrx cvmx_sso_wq_int_thrx_t;
+
+/**
+ * cvmx_sso_wq_iq_dis
+ *
+ * CSR reserved addresses: (1): 0x1008..0x1008
+ * SSO_WQ_IQ_DIS = SSO Input Queue Interrupt Temporary Disable Mask
+ *
+ * Contains the input queue interrupt temporary disable bits (one per group).
+ * For more information regarding this register, see the interrupt section.
+ */
+union cvmx_sso_wq_iq_dis {
+ u64 u64;
+ struct cvmx_sso_wq_iq_dis_s {
+ u64 iq_dis : 64;
+ } s;
+ struct cvmx_sso_wq_iq_dis_s cn68xx;
+ struct cvmx_sso_wq_iq_dis_s cn68xxp1;
+};
+
+typedef union cvmx_sso_wq_iq_dis cvmx_sso_wq_iq_dis_t;
+
+/**
+ * cvmx_sso_ws_cfg
+ *
+ * This register contains various SSO work-slot configuration bits.
+ *
+ */
+union cvmx_sso_ws_cfg {
+ u64 u64;
+ struct cvmx_sso_ws_cfg_s {
+ u64 reserved_56_63 : 8;
+ u64 ocla_bp : 8;
+ u64 reserved_7_47 : 41;
+ u64 aw_clk_dis : 1;
+ u64 gw_clk_dis : 1;
+ u64 disable_pw : 1;
+ u64 arbc_step_en : 1;
+ u64 ncbo_step_en : 1;
+ u64 soc_ccam_dis : 1;
+ u64 sso_cclk_dis : 1;
+ } s;
+ struct cvmx_sso_ws_cfg_s cn73xx;
+ struct cvmx_sso_ws_cfg_cn78xx {
+ u64 reserved_56_63 : 8;
+ u64 ocla_bp : 8;
+ u64 reserved_5_47 : 43;
+ u64 disable_pw : 1;
+ u64 arbc_step_en : 1;
+ u64 ncbo_step_en : 1;
+ u64 soc_ccam_dis : 1;
+ u64 sso_cclk_dis : 1;
+ } cn78xx;
+ struct cvmx_sso_ws_cfg_cn78xx cn78xxp1;
+ struct cvmx_sso_ws_cfg_s cnf75xx;
+};
+
+typedef union cvmx_sso_ws_cfg cvmx_sso_ws_cfg_t;
+
+/**
+ * cvmx_sso_ws_eco
+ */
+union cvmx_sso_ws_eco {
+ u64 u64;
+ struct cvmx_sso_ws_eco_s {
+ u64 reserved_8_63 : 56;
+ u64 eco_rw : 8;
+ } s;
+ struct cvmx_sso_ws_eco_s cn73xx;
+ struct cvmx_sso_ws_eco_s cnf75xx;
+};
+
+typedef union cvmx_sso_ws_eco cvmx_sso_ws_eco_t;
+
+/**
+ * cvmx_sso_ws_pc#
+ *
+ * CSR reserved addresses: (225): 0x3100..0x3800
+ * CSR align addresses: ===========================================================================================================
+ * SSO_WS_PCX = SSO Work Schedule Performance Counter
+ * (one per group)
+ *
+ * Counts the number of work schedules for each group.
+ * Counter rolls over through zero when max value exceeded.
+ */
+union cvmx_sso_ws_pcx {
+ u64 u64;
+ struct cvmx_sso_ws_pcx_s {
+ u64 ws_pc : 64;
+ } s;
+ struct cvmx_sso_ws_pcx_s cn68xx;
+ struct cvmx_sso_ws_pcx_s cn68xxp1;
+};
+
+typedef union cvmx_sso_ws_pcx cvmx_sso_ws_pcx_t;
+
+/**
+ * cvmx_sso_xaq#_head_next
+ *
+ * These registers contain the pointer to the next buffer to become the head when the final cache
+ * line in this buffer is read.
+ */
+union cvmx_sso_xaqx_head_next {
+ u64 u64;
+ struct cvmx_sso_xaqx_head_next_s {
+ u64 reserved_42_63 : 22;
+ u64 ptr : 35;
+ u64 reserved_0_6 : 7;
+ } s;
+ struct cvmx_sso_xaqx_head_next_s cn73xx;
+ struct cvmx_sso_xaqx_head_next_s cn78xx;
+ struct cvmx_sso_xaqx_head_next_s cn78xxp1;
+ struct cvmx_sso_xaqx_head_next_s cnf75xx;
+};
+
+typedef union cvmx_sso_xaqx_head_next cvmx_sso_xaqx_head_next_t;
+
+/**
+ * cvmx_sso_xaq#_head_ptr
+ *
+ * These registers contain the pointer to the first entry of the external linked list(s) for a
+ * particular group. Software must initialize the external linked list(s) by programming
+ * SSO_XAQ()_HEAD_PTR, SSO_XAQ()_HEAD_NEXT, SSO_XAQ()_TAIL_PTR and
+ * SSO_XAQ()_TAIL_NEXT to identical values.
+ */
+union cvmx_sso_xaqx_head_ptr {
+ u64 u64;
+ struct cvmx_sso_xaqx_head_ptr_s {
+ u64 reserved_42_63 : 22;
+ u64 ptr : 35;
+ u64 reserved_5_6 : 2;
+ u64 cl : 5;
+ } s;
+ struct cvmx_sso_xaqx_head_ptr_s cn73xx;
+ struct cvmx_sso_xaqx_head_ptr_s cn78xx;
+ struct cvmx_sso_xaqx_head_ptr_s cn78xxp1;
+ struct cvmx_sso_xaqx_head_ptr_s cnf75xx;
+};
+
+typedef union cvmx_sso_xaqx_head_ptr cvmx_sso_xaqx_head_ptr_t;
+
+/**
+ * cvmx_sso_xaq#_tail_next
+ *
+ * These registers contain the pointer to the next buffer to become the tail when the final cache
+ * line in this buffer is written. Register fields are identical to those in
+ * SSO_XAQ()_HEAD_NEXT above.
+ */
+union cvmx_sso_xaqx_tail_next {
+ u64 u64;
+ struct cvmx_sso_xaqx_tail_next_s {
+ u64 reserved_42_63 : 22;
+ u64 ptr : 35;
+ u64 reserved_0_6 : 7;
+ } s;
+ struct cvmx_sso_xaqx_tail_next_s cn73xx;
+ struct cvmx_sso_xaqx_tail_next_s cn78xx;
+ struct cvmx_sso_xaqx_tail_next_s cn78xxp1;
+ struct cvmx_sso_xaqx_tail_next_s cnf75xx;
+};
+
+typedef union cvmx_sso_xaqx_tail_next cvmx_sso_xaqx_tail_next_t;
+
+/**
+ * cvmx_sso_xaq#_tail_ptr
+ *
+ * These registers contain the pointer to the last entry of the external linked list(s) for a
+ * particular group. Register fields are identical to those in SSO_XAQ()_HEAD_PTR above.
+ * Software must initialize the external linked list(s) by programming
+ * SSO_XAQ()_HEAD_PTR, SSO_XAQ()_HEAD_NEXT, SSO_XAQ()_TAIL_PTR and
+ * SSO_XAQ()_TAIL_NEXT to identical values.
+ */
+union cvmx_sso_xaqx_tail_ptr {
+ u64 u64;
+ struct cvmx_sso_xaqx_tail_ptr_s {
+ u64 reserved_42_63 : 22;
+ u64 ptr : 35;
+ u64 reserved_5_6 : 2;
+ u64 cl : 5;
+ } s;
+ struct cvmx_sso_xaqx_tail_ptr_s cn73xx;
+ struct cvmx_sso_xaqx_tail_ptr_s cn78xx;
+ struct cvmx_sso_xaqx_tail_ptr_s cn78xxp1;
+ struct cvmx_sso_xaqx_tail_ptr_s cnf75xx;
+};
+
+typedef union cvmx_sso_xaqx_tail_ptr cvmx_sso_xaqx_tail_ptr_t;
+
+/**
+ * cvmx_sso_xaq_aura
+ */
+union cvmx_sso_xaq_aura {
+ u64 u64;
+ struct cvmx_sso_xaq_aura_s {
+ u64 reserved_12_63 : 52;
+ u64 node : 2;
+ u64 laura : 10;
+ } s;
+ struct cvmx_sso_xaq_aura_s cn73xx;
+ struct cvmx_sso_xaq_aura_s cn78xx;
+ struct cvmx_sso_xaq_aura_s cn78xxp1;
+ struct cvmx_sso_xaq_aura_s cnf75xx;
+};
+
+typedef union cvmx_sso_xaq_aura cvmx_sso_xaq_aura_t;
+
+/**
+ * cvmx_sso_xaq_latency_pc
+ */
+union cvmx_sso_xaq_latency_pc {
+ u64 u64;
+ struct cvmx_sso_xaq_latency_pc_s {
+ u64 count : 64;
+ } s;
+ struct cvmx_sso_xaq_latency_pc_s cn73xx;
+ struct cvmx_sso_xaq_latency_pc_s cn78xx;
+ struct cvmx_sso_xaq_latency_pc_s cn78xxp1;
+ struct cvmx_sso_xaq_latency_pc_s cnf75xx;
+};
+
+typedef union cvmx_sso_xaq_latency_pc cvmx_sso_xaq_latency_pc_t;
+
+/**
+ * cvmx_sso_xaq_req_pc
+ */
+union cvmx_sso_xaq_req_pc {
+ u64 u64;
+ struct cvmx_sso_xaq_req_pc_s {
+ u64 count : 64;
+ } s;
+ struct cvmx_sso_xaq_req_pc_s cn73xx;
+ struct cvmx_sso_xaq_req_pc_s cn78xx;
+ struct cvmx_sso_xaq_req_pc_s cn78xxp1;
+ struct cvmx_sso_xaq_req_pc_s cnf75xx;
+};
+
+typedef union cvmx_sso_xaq_req_pc cvmx_sso_xaq_req_pc_t;
+
+#endif