summaryrefslogtreecommitdiffstats
path: root/sys/contrib/octeon-sdk/cvmx-helper.c
diff options
context:
space:
mode:
Diffstat (limited to 'sys/contrib/octeon-sdk/cvmx-helper.c')
-rw-r--r--sys/contrib/octeon-sdk/cvmx-helper.c1999
1 files changed, 1999 insertions, 0 deletions
diff --git a/sys/contrib/octeon-sdk/cvmx-helper.c b/sys/contrib/octeon-sdk/cvmx-helper.c
new file mode 100644
index 0000000..26582f0
--- /dev/null
+++ b/sys/contrib/octeon-sdk/cvmx-helper.c
@@ -0,0 +1,1999 @@
+/***********************license start***************
+ * Copyright (c) 2003-2010 Cavium Inc. (support@cavium.com). All rights
+ * reserved.
+ *
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions are
+ * met:
+ *
+ * * Redistributions of source code must retain the above copyright
+ * notice, this list of conditions and the following disclaimer.
+ *
+ * * Redistributions in binary form must reproduce the above
+ * copyright notice, this list of conditions and the following
+ * disclaimer in the documentation and/or other materials provided
+ * with the distribution.
+
+ * * Neither the name of Cavium Inc. nor the names of
+ * its contributors may be used to endorse or promote products
+ * derived from this software without specific prior written
+ * permission.
+
+ * This Software, including technical data, may be subject to U.S. export control
+ * laws, including the U.S. Export Administration Act and its associated
+ * regulations, and may be subject to export or import regulations in other
+ * countries.
+
+ * TO THE MAXIMUM EXTENT PERMITTED BY LAW, THE SOFTWARE IS PROVIDED "AS IS"
+ * AND WITH ALL FAULTS AND CAVIUM INC. MAKES NO PROMISES, REPRESENTATIONS OR
+ * WARRANTIES, EITHER EXPRESS, IMPLIED, STATUTORY, OR OTHERWISE, WITH RESPECT TO
+ * THE SOFTWARE, INCLUDING ITS CONDITION, ITS CONFORMITY TO ANY REPRESENTATION OR
+ * DESCRIPTION, OR THE EXISTENCE OF ANY LATENT OR PATENT DEFECTS, AND CAVIUM
+ * SPECIFICALLY DISCLAIMS ALL IMPLIED (IF ANY) WARRANTIES OF TITLE,
+ * MERCHANTABILITY, NONINFRINGEMENT, FITNESS FOR A PARTICULAR PURPOSE, LACK OF
+ * VIRUSES, ACCURACY OR COMPLETENESS, QUIET ENJOYMENT, QUIET POSSESSION OR
+ * CORRESPONDENCE TO DESCRIPTION. THE ENTIRE RISK ARISING OUT OF USE OR
+ * PERFORMANCE OF THE SOFTWARE LIES WITH YOU.
+ ***********************license end**************************************/
+
+
+
+
+
+
+
+/**
+ * @file
+ *
+ * Helper functions for common, but complicated tasks.
+ *
+ * <hr>$Revision: 70030 $<hr>
+ */
+#ifdef CVMX_BUILD_FOR_LINUX_KERNEL
+#include <linux/module.h>
+#include <asm/octeon/cvmx.h>
+#include <asm/octeon/cvmx-config.h>
+#include <asm/octeon/cvmx-bootmem.h>
+#include <asm/octeon/cvmx-sriox-defs.h>
+#include <asm/octeon/cvmx-npi-defs.h>
+#include <asm/octeon/cvmx-mio-defs.h>
+#include <asm/octeon/cvmx-pexp-defs.h>
+#include <asm/octeon/cvmx-pip-defs.h>
+#include <asm/octeon/cvmx-asxx-defs.h>
+#include <asm/octeon/cvmx-gmxx-defs.h>
+#include <asm/octeon/cvmx-smix-defs.h>
+#include <asm/octeon/cvmx-dbg-defs.h>
+#include <asm/octeon/cvmx-sso-defs.h>
+
+#include <asm/octeon/cvmx-gmx.h>
+#include <asm/octeon/cvmx-fpa.h>
+#include <asm/octeon/cvmx-pip.h>
+#include <asm/octeon/cvmx-pko.h>
+#include <asm/octeon/cvmx-ipd.h>
+#include <asm/octeon/cvmx-spi.h>
+#include <asm/octeon/cvmx-clock.h>
+#include <asm/octeon/cvmx-helper.h>
+#include <asm/octeon/cvmx-helper-board.h>
+#include <asm/octeon/cvmx-helper-errata.h>
+#include <asm/octeon/cvmx-helper-cfg.h>
+#else
+#if !defined(__FreeBSD__) || !defined(_KERNEL)
+#include "executive-config.h"
+#endif
+#include "cvmx.h"
+#include "cvmx-sysinfo.h"
+#include "cvmx-bootmem.h"
+#include "cvmx-version.h"
+#include "cvmx-helper-check-defines.h"
+#include "cvmx-gmx.h"
+#if !defined(__FreeBSD__) || !defined(_KERNEL)
+#include "cvmx-error.h"
+#include "cvmx-config.h"
+#endif
+
+#include "cvmx-fpa.h"
+#include "cvmx-pip.h"
+#include "cvmx-pko.h"
+#include "cvmx-ipd.h"
+#include "cvmx-spi.h"
+#include "cvmx-helper.h"
+#include "cvmx-helper-board.h"
+#include "cvmx-helper-errata.h"
+#include "cvmx-helper-cfg.h"
+#endif
+
+
+#ifdef CVMX_ENABLE_PKO_FUNCTIONS
+
+/**
+ * cvmx_override_pko_queue_priority(int pko_port, uint64_t
+ * priorities[16]) is a function pointer. It is meant to allow
+ * customization of the PKO queue priorities based on the port
+ * number. Users should set this pointer to a function before
+ * calling any cvmx-helper operations.
+ */
+CVMX_SHARED void (*cvmx_override_pko_queue_priority)(int ipd_port,
+ uint64_t *priorities) = NULL;
+#ifdef CVMX_BUILD_FOR_LINUX_KERNEL
+EXPORT_SYMBOL(cvmx_override_pko_queue_priority);
+#endif
+
+/**
+ * cvmx_override_ipd_port_setup(int ipd_port) is a function
+ * pointer. It is meant to allow customization of the IPD
+ * port/port kind setup before packet input/output comes online.
+ * It is called after cvmx-helper does the default IPD configuration,
+ * but before IPD is enabled. Users should set this pointer to a
+ * function before calling any cvmx-helper operations.
+ */
+CVMX_SHARED void (*cvmx_override_ipd_port_setup)(int ipd_port) = NULL;
+
+/**
+ * Return the number of interfaces the chip has. Each interface
+ * may have multiple ports. Most chips support two interfaces,
+ * but the CNX0XX and CNX1XX are exceptions. These only support
+ * one interface.
+ *
+ * @return Number of interfaces on chip
+ */
+int cvmx_helper_get_number_of_interfaces(void)
+{
+ switch (cvmx_sysinfo_get()->board_type) {
+#if defined(OCTEON_VENDOR_LANNER)
+ case CVMX_BOARD_TYPE_CUST_LANNER_MR955:
+ return 2;
+ case CVMX_BOARD_TYPE_CUST_LANNER_MR730:
+ return 1;
+#endif
+#if defined(OCTEON_VENDOR_RADISYS)
+ case CVMX_BOARD_TYPE_CUST_RADISYS_RSYS4GBE:
+ return 2;
+#endif
+ default:
+ break;
+ }
+
+ if (OCTEON_IS_MODEL(OCTEON_CN68XX))
+ return 9;
+ else if (OCTEON_IS_MODEL(OCTEON_CN66XX))
+ if (OCTEON_IS_MODEL(OCTEON_CN66XX_PASS1_0))
+ return 7;
+ else
+ return 8;
+ else if (OCTEON_IS_MODEL(OCTEON_CN63XX))
+ return 6;
+ else if (OCTEON_IS_MODEL(OCTEON_CN56XX) || OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN61XX))
+ return 4;
+ else
+ return 3;
+}
+#ifdef CVMX_BUILD_FOR_LINUX_KERNEL
+EXPORT_SYMBOL(cvmx_helper_get_number_of_interfaces);
+#endif
+
+
+/**
+ * Return the number of ports on an interface. Depending on the
+ * chip and configuration, this can be 1-16. A value of 0
+ * specifies that the interface doesn't exist or isn't usable.
+ *
+ * @param interface Interface to get the port count for
+ *
+ * @return Number of ports on interface. Can be Zero.
+ */
+int cvmx_helper_ports_on_interface(int interface)
+{
+ if (octeon_has_feature(OCTEON_FEATURE_PKND))
+ return cvmx_helper_interface_enumerate(interface);
+ else
+ return __cvmx_helper_get_num_ipd_ports(interface);
+}
+#ifdef CVMX_BUILD_FOR_LINUX_KERNEL
+EXPORT_SYMBOL(cvmx_helper_ports_on_interface);
+#endif
+
+
+/**
+ * Get the operating mode of an interface. Depending on the Octeon
+ * chip and configuration, this function returns an enumeration
+ * of the type of packet I/O supported by an interface.
+ *
+ * @param interface Interface to probe
+ *
+ * @return Mode of the interface. Unknown or unsupported interfaces return
+ * DISABLED.
+ */
+cvmx_helper_interface_mode_t cvmx_helper_interface_get_mode(int interface)
+{
+ cvmx_gmxx_inf_mode_t mode;
+
+ if (OCTEON_IS_MODEL(OCTEON_CN68XX))
+ {
+ cvmx_mio_qlmx_cfg_t qlm_cfg;
+ switch(interface)
+ {
+ case 0:
+ qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(0));
+ /* QLM is disabled when QLM SPD is 15. */
+ if (qlm_cfg.s.qlm_spd == 15)
+ return CVMX_HELPER_INTERFACE_MODE_DISABLED;
+
+ if (qlm_cfg.s.qlm_cfg == 7)
+ return CVMX_HELPER_INTERFACE_MODE_RXAUI;
+ else if (qlm_cfg.s.qlm_cfg == 2)
+ return CVMX_HELPER_INTERFACE_MODE_SGMII;
+ else if (qlm_cfg.s.qlm_cfg == 3)
+ return CVMX_HELPER_INTERFACE_MODE_XAUI;
+ else
+ return CVMX_HELPER_INTERFACE_MODE_DISABLED;
+ break;
+ case 1:
+ qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(0));
+ /* QLM is disabled when QLM SPD is 15. */
+ if (qlm_cfg.s.qlm_spd == 15)
+ return CVMX_HELPER_INTERFACE_MODE_DISABLED;
+
+ if (qlm_cfg.s.qlm_cfg == 7)
+ return CVMX_HELPER_INTERFACE_MODE_RXAUI;
+ else
+ return CVMX_HELPER_INTERFACE_MODE_DISABLED;
+ break;
+ case 2:
+ case 3:
+ case 4:
+ qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(interface));
+ /* QLM is disabled when QLM SPD is 15. */
+ if (qlm_cfg.s.qlm_spd == 15)
+ return CVMX_HELPER_INTERFACE_MODE_DISABLED;
+
+ if (qlm_cfg.s.qlm_cfg == 2)
+ return CVMX_HELPER_INTERFACE_MODE_SGMII;
+ else if (qlm_cfg.s.qlm_cfg == 3)
+ return CVMX_HELPER_INTERFACE_MODE_XAUI;
+ else
+ return CVMX_HELPER_INTERFACE_MODE_DISABLED;
+ break;
+ case 5:
+ case 6:
+ qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(interface - 4));
+ /* QLM is disabled when QLM SPD is 15. */
+ if (qlm_cfg.s.qlm_spd == 15)
+ return CVMX_HELPER_INTERFACE_MODE_DISABLED;
+
+ if (qlm_cfg.s.qlm_cfg == 1)
+ {
+ return CVMX_HELPER_INTERFACE_MODE_ILK;
+ }
+ else
+ return CVMX_HELPER_INTERFACE_MODE_DISABLED;
+ break;
+ case 7:
+ qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(3));
+ /* QLM is disabled when QLM SPD is 15. */
+ if (qlm_cfg.s.qlm_spd == 15)
+ return CVMX_HELPER_INTERFACE_MODE_DISABLED;
+ else if (qlm_cfg.s.qlm_cfg != 0)
+ {
+ qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(1));
+ if (qlm_cfg.s.qlm_cfg != 0)
+ return CVMX_HELPER_INTERFACE_MODE_DISABLED;
+ }
+ return CVMX_HELPER_INTERFACE_MODE_NPI;
+ break;
+ case 8:
+ return CVMX_HELPER_INTERFACE_MODE_LOOP;
+ break;
+ default:
+ return CVMX_HELPER_INTERFACE_MODE_DISABLED;
+ break;
+ }
+ }
+
+ if (interface == 2)
+ return CVMX_HELPER_INTERFACE_MODE_NPI;
+
+ if (interface == 3)
+ {
+ if (OCTEON_IS_MODEL(OCTEON_CN56XX)
+ || OCTEON_IS_MODEL(OCTEON_CN52XX)
+ || OCTEON_IS_MODEL(OCTEON_CN6XXX)
+ || OCTEON_IS_MODEL(OCTEON_CNF7XXX))
+ return CVMX_HELPER_INTERFACE_MODE_LOOP;
+ else
+ return CVMX_HELPER_INTERFACE_MODE_DISABLED;
+ }
+
+ /* Only present in CN63XX & CN66XX Octeon model */
+ if ((OCTEON_IS_MODEL(OCTEON_CN63XX) && (interface == 4 || interface == 5))
+ || (OCTEON_IS_MODEL(OCTEON_CN66XX) && interface >= 4 && interface <= 7))
+ {
+ cvmx_sriox_status_reg_t sriox_status_reg;
+
+ /* cn66xx pass1.0 has only 2 SRIO interfaces. */
+ if ((interface == 5 || interface == 7) && OCTEON_IS_MODEL(OCTEON_CN66XX_PASS1_0))
+ return CVMX_HELPER_INTERFACE_MODE_DISABLED;
+
+ sriox_status_reg.u64 = cvmx_read_csr(CVMX_SRIOX_STATUS_REG(interface-4));
+ if (sriox_status_reg.s.srio)
+ return CVMX_HELPER_INTERFACE_MODE_SRIO;
+ else
+ return CVMX_HELPER_INTERFACE_MODE_DISABLED;
+ }
+
+ /* Interface 5 always disabled in CN66XX */
+ if (OCTEON_IS_MODEL(OCTEON_CN66XX))
+ {
+ cvmx_mio_qlmx_cfg_t mio_qlm_cfg;
+
+ /* QLM2 is SGMII0 and QLM1 is SGMII1 */
+ if (interface == 0)
+ mio_qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(2));
+ else if (interface == 1)
+ mio_qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(1));
+ else
+ return CVMX_HELPER_INTERFACE_MODE_DISABLED;
+
+ if (mio_qlm_cfg.s.qlm_spd == 15)
+ return CVMX_HELPER_INTERFACE_MODE_DISABLED;
+
+ if (mio_qlm_cfg.s.qlm_cfg == 9)
+ return CVMX_HELPER_INTERFACE_MODE_SGMII;
+ else if (mio_qlm_cfg.s.qlm_cfg == 11)
+ return CVMX_HELPER_INTERFACE_MODE_XAUI;
+ else
+ return CVMX_HELPER_INTERFACE_MODE_DISABLED;
+ }
+ else if (OCTEON_IS_MODEL(OCTEON_CN61XX))
+ {
+ cvmx_mio_qlmx_cfg_t qlm_cfg;
+
+ if (interface == 0)
+ {
+ qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(2));
+ if (qlm_cfg.s.qlm_cfg == 2)
+ return CVMX_HELPER_INTERFACE_MODE_SGMII;
+ else if (qlm_cfg.s.qlm_cfg == 3)
+ return CVMX_HELPER_INTERFACE_MODE_XAUI;
+ else
+ return CVMX_HELPER_INTERFACE_MODE_DISABLED;
+ }
+ else if (interface == 1)
+ {
+ /* If QLM 1 is PEV0/PEM1 mode, them QLM0 cannot be SGMII/XAUI */
+ qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(1));
+ if (qlm_cfg.s.qlm_cfg == 1)
+ return CVMX_HELPER_INTERFACE_MODE_DISABLED;
+
+ qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(0));
+ if (qlm_cfg.s.qlm_cfg == 2)
+ return CVMX_HELPER_INTERFACE_MODE_SGMII;
+ else if (qlm_cfg.s.qlm_cfg == 3)
+ return CVMX_HELPER_INTERFACE_MODE_XAUI;
+ else
+ return CVMX_HELPER_INTERFACE_MODE_DISABLED;
+ }
+ }
+
+ if (interface == 0 && cvmx_sysinfo_get()->board_type == CVMX_BOARD_TYPE_CN3005_EVB_HS5 && cvmx_sysinfo_get()->board_rev_major == 1)
+ {
+ /* Lie about interface type of CN3005 board. This board has a switch on port 1 like
+ ** the other evaluation boards, but it is connected over RGMII instead of GMII. Report
+ ** GMII mode so that the speed is forced to 1 Gbit full duplex. Other than some initial configuration
+ ** (which does not use the output of this function) there is no difference in setup between GMII and RGMII modes.
+ */
+ return CVMX_HELPER_INTERFACE_MODE_GMII;
+ }
+
+ /* Interface 1 is always disabled on CN31XX and CN30XX */
+ if ((interface == 1)
+ && (OCTEON_IS_MODEL(OCTEON_CN31XX)
+ || OCTEON_IS_MODEL(OCTEON_CN30XX)
+ || OCTEON_IS_MODEL(OCTEON_CN50XX)
+ || OCTEON_IS_MODEL(OCTEON_CN52XX)
+ || OCTEON_IS_MODEL(OCTEON_CN63XX)
+ || OCTEON_IS_MODEL(OCTEON_CNF71XX)))
+ return CVMX_HELPER_INTERFACE_MODE_DISABLED;
+
+ mode.u64 = cvmx_read_csr(CVMX_GMXX_INF_MODE(interface));
+
+ if (OCTEON_IS_MODEL(OCTEON_CN56XX) || OCTEON_IS_MODEL(OCTEON_CN52XX))
+ {
+ switch(mode.cn56xx.mode)
+ {
+ case 0: return CVMX_HELPER_INTERFACE_MODE_DISABLED;
+ case 1: return CVMX_HELPER_INTERFACE_MODE_XAUI;
+ case 2: return CVMX_HELPER_INTERFACE_MODE_SGMII;
+ case 3: return CVMX_HELPER_INTERFACE_MODE_PICMG;
+ default:return CVMX_HELPER_INTERFACE_MODE_DISABLED;
+ }
+ }
+ else if (OCTEON_IS_MODEL(OCTEON_CN63XX) || OCTEON_IS_MODEL(OCTEON_CNF71XX))
+ {
+ switch(mode.cn63xx.mode)
+ {
+ case 0: return CVMX_HELPER_INTERFACE_MODE_SGMII;
+ case 1: return CVMX_HELPER_INTERFACE_MODE_XAUI;
+ default: return CVMX_HELPER_INTERFACE_MODE_DISABLED;
+ }
+ }
+ else
+ {
+ if (!mode.s.en)
+ return CVMX_HELPER_INTERFACE_MODE_DISABLED;
+
+ if (mode.s.type)
+ {
+ if (OCTEON_IS_MODEL(OCTEON_CN38XX) || OCTEON_IS_MODEL(OCTEON_CN58XX))
+ return CVMX_HELPER_INTERFACE_MODE_SPI;
+ else
+ return CVMX_HELPER_INTERFACE_MODE_GMII;
+ }
+ else
+ return CVMX_HELPER_INTERFACE_MODE_RGMII;
+ }
+}
+#ifdef CVMX_BUILD_FOR_LINUX_KERNEL
+EXPORT_SYMBOL(cvmx_helper_interface_get_mode);
+#endif
+
+/**
+ * @INTERNAL
+ * Configure the IPD/PIP tagging and QoS options for a specific
+ * port. This function determines the POW work queue entry
+ * contents for a port. The setup performed here is controlled by
+ * the defines in executive-config.h.
+ *
+ * @param ipd_port Port/Port kind to configure. This follows the IPD numbering,
+ * not the per interface numbering
+ *
+ * @return Zero on success, negative on failure
+ */
+static int __cvmx_helper_port_setup_ipd(int ipd_port)
+{
+ cvmx_pip_prt_cfgx_t port_config;
+ cvmx_pip_prt_tagx_t tag_config;
+
+ if (octeon_has_feature(OCTEON_FEATURE_PKND))
+ {
+ int interface, index, pknd;
+ cvmx_pip_prt_cfgbx_t prt_cfgbx;
+
+ interface = cvmx_helper_get_interface_num(ipd_port);
+ index = cvmx_helper_get_interface_index_num(ipd_port);
+ pknd = cvmx_helper_get_pknd(interface, index);
+
+ port_config.u64 = cvmx_read_csr(CVMX_PIP_PRT_CFGX(pknd));
+ tag_config.u64 = cvmx_read_csr(CVMX_PIP_PRT_TAGX(pknd));
+
+ port_config.s.qos = pknd & 0x7;
+
+ /* Default BPID to use for packets on this port-kind */
+ prt_cfgbx.u64 = cvmx_read_csr(CVMX_PIP_PRT_CFGBX(pknd));
+ prt_cfgbx.s.bpid = pknd;
+ cvmx_write_csr(CVMX_PIP_PRT_CFGBX(pknd), prt_cfgbx.u64);
+ }
+ else
+ {
+ port_config.u64 = cvmx_read_csr(CVMX_PIP_PRT_CFGX(ipd_port));
+ tag_config.u64 = cvmx_read_csr(CVMX_PIP_PRT_TAGX(ipd_port));
+
+ /* Have each port go to a different POW queue */
+ port_config.s.qos = ipd_port & 0x7;
+ }
+
+ /* Process the headers and place the IP header in the work queue */
+ port_config.s.mode = CVMX_HELPER_INPUT_PORT_SKIP_MODE;
+
+ tag_config.s.ip6_src_flag = CVMX_HELPER_INPUT_TAG_IPV6_SRC_IP;
+ tag_config.s.ip6_dst_flag = CVMX_HELPER_INPUT_TAG_IPV6_DST_IP;
+ tag_config.s.ip6_sprt_flag = CVMX_HELPER_INPUT_TAG_IPV6_SRC_PORT;
+ tag_config.s.ip6_dprt_flag = CVMX_HELPER_INPUT_TAG_IPV6_DST_PORT;
+ tag_config.s.ip6_nxth_flag = CVMX_HELPER_INPUT_TAG_IPV6_NEXT_HEADER;
+ tag_config.s.ip4_src_flag = CVMX_HELPER_INPUT_TAG_IPV4_SRC_IP;
+ tag_config.s.ip4_dst_flag = CVMX_HELPER_INPUT_TAG_IPV4_DST_IP;
+ tag_config.s.ip4_sprt_flag = CVMX_HELPER_INPUT_TAG_IPV4_SRC_PORT;
+ tag_config.s.ip4_dprt_flag = CVMX_HELPER_INPUT_TAG_IPV4_DST_PORT;
+ tag_config.s.ip4_pctl_flag = CVMX_HELPER_INPUT_TAG_IPV4_PROTOCOL;
+ tag_config.s.inc_prt_flag = CVMX_HELPER_INPUT_TAG_INPUT_PORT;
+ tag_config.s.tcp6_tag_type = CVMX_HELPER_INPUT_TAG_TYPE;
+ tag_config.s.tcp4_tag_type = CVMX_HELPER_INPUT_TAG_TYPE;
+ tag_config.s.ip6_tag_type = CVMX_HELPER_INPUT_TAG_TYPE;
+ tag_config.s.ip4_tag_type = CVMX_HELPER_INPUT_TAG_TYPE;
+ tag_config.s.non_tag_type = CVMX_HELPER_INPUT_TAG_TYPE;
+ /* Put all packets in group 0. Other groups can be used by the app */
+ tag_config.s.grp = 0;
+
+ cvmx_pip_config_port(ipd_port, port_config, tag_config);
+
+ /* Give the user a chance to override our setting for each port */
+ if (cvmx_override_ipd_port_setup)
+ cvmx_override_ipd_port_setup(ipd_port);
+
+ return 0;
+}
+
+/**
+ * Enable or disable FCS stripping for all the ports on an interface.
+ *
+ * @param interface
+ * @param nports number of ports
+ * @param has_fcs 0 for disable and !0 for enable
+ */
+static int cvmx_helper_fcs_op(int interface, int nports, int has_fcs)
+{
+ uint64_t port_bit;
+ int index;
+ int pknd;
+ cvmx_pip_sub_pkind_fcsx_t pkind_fcsx;
+ cvmx_pip_prt_cfgx_t port_cfg;
+
+ if (!octeon_has_feature(OCTEON_FEATURE_PKND))
+ return 0;
+
+ port_bit = 0;
+ for (index = 0; index < nports; index++)
+ port_bit |= ((uint64_t)1 << cvmx_helper_get_pknd(interface, index));
+
+ pkind_fcsx.u64 = cvmx_read_csr(CVMX_PIP_SUB_PKIND_FCSX(0));
+ if (has_fcs)
+ pkind_fcsx.s.port_bit |= port_bit;
+ else
+ pkind_fcsx.s.port_bit &= ~port_bit;
+ cvmx_write_csr(CVMX_PIP_SUB_PKIND_FCSX(0), pkind_fcsx.u64);
+
+ for (pknd = 0; pknd < 64; pknd++)
+ {
+ if ((1ull << pknd) & port_bit)
+ {
+ port_cfg.u64 = cvmx_read_csr(CVMX_PIP_PRT_CFGX(pknd));
+ port_cfg.s.crc_en = (has_fcs) ? 1 : 0;
+ cvmx_write_csr(CVMX_PIP_PRT_CFGX(pknd), port_cfg.u64);
+ }
+ }
+
+ return 0;
+}
+
+/**
+ * Determine the actual number of hardware ports connected to an
+ * interface. It doesn't setup the ports or enable them.
+ *
+ * @param interface Interface to enumerate
+ *
+ * @return The number of ports on the interface, negative on failure
+ */
+int cvmx_helper_interface_enumerate(int interface)
+{
+ switch (cvmx_helper_interface_get_mode(interface)) {
+ /* XAUI is a single high speed port */
+ case CVMX_HELPER_INTERFACE_MODE_XAUI:
+ case CVMX_HELPER_INTERFACE_MODE_RXAUI:
+ return __cvmx_helper_xaui_enumerate(interface);
+ /* RGMII/GMII/MII are all treated about the same. Most functions
+ refer to these ports as RGMII */
+ case CVMX_HELPER_INTERFACE_MODE_RGMII:
+ case CVMX_HELPER_INTERFACE_MODE_GMII:
+ return __cvmx_helper_rgmii_enumerate(interface);
+ /* SPI4 can have 1-16 ports depending on the device at the other end */
+ case CVMX_HELPER_INTERFACE_MODE_SPI:
+ return __cvmx_helper_spi_enumerate(interface);
+ /* SGMII can have 1-4 ports depending on how many are hooked up */
+ case CVMX_HELPER_INTERFACE_MODE_SGMII:
+ case CVMX_HELPER_INTERFACE_MODE_PICMG:
+ return __cvmx_helper_sgmii_enumerate(interface);
+ /* PCI target Network Packet Interface */
+ case CVMX_HELPER_INTERFACE_MODE_NPI:
+ return __cvmx_helper_npi_enumerate(interface);
+ /* Special loopback only ports. These are not the same
+ * as other ports in loopback mode */
+ case CVMX_HELPER_INTERFACE_MODE_LOOP:
+ return __cvmx_helper_loop_enumerate(interface);
+ /* SRIO has 2^N ports, where N is number of interfaces */
+ case CVMX_HELPER_INTERFACE_MODE_SRIO:
+ return __cvmx_helper_srio_enumerate(interface);
+
+ case CVMX_HELPER_INTERFACE_MODE_ILK:
+ return __cvmx_helper_ilk_enumerate(interface);
+ /* These types don't support ports to IPD/PKO */
+ case CVMX_HELPER_INTERFACE_MODE_DISABLED:
+ case CVMX_HELPER_INTERFACE_MODE_PCIE:
+ default:
+ return 0;
+ }
+}
+
+/**
+ * This function probes an interface to determine the actual number of
+ * hardware ports connected to it. It does some setup the ports but
+ * doesn't enable them. The main goal here is to set the global
+ * interface_port_count[interface] correctly. Final hardware setup of
+ * the ports will be performed later.
+ *
+ * @param interface Interface to probe
+ *
+ * @return Zero on success, negative on failure
+ */
+int cvmx_helper_interface_probe(int interface)
+{
+ /* At this stage in the game we don't want packets to be moving yet.
+ The following probe calls should perform hardware setup
+ needed to determine port counts. Receive must still be disabled */
+ int nports;
+ int has_fcs;
+ enum cvmx_pko_padding padding = CVMX_PKO_PADDING_NONE;
+
+ nports = -1;
+ has_fcs = 0;
+ switch (cvmx_helper_interface_get_mode(interface))
+ {
+ /* These types don't support ports to IPD/PKO */
+ case CVMX_HELPER_INTERFACE_MODE_DISABLED:
+ case CVMX_HELPER_INTERFACE_MODE_PCIE:
+ nports = 0;
+ break;
+ /* XAUI is a single high speed port */
+ case CVMX_HELPER_INTERFACE_MODE_XAUI:
+ case CVMX_HELPER_INTERFACE_MODE_RXAUI:
+ nports = __cvmx_helper_xaui_probe(interface);
+ has_fcs = 1;
+ padding = CVMX_PKO_PADDING_60;
+ break;
+ /* RGMII/GMII/MII are all treated about the same. Most functions
+ refer to these ports as RGMII */
+ case CVMX_HELPER_INTERFACE_MODE_RGMII:
+ case CVMX_HELPER_INTERFACE_MODE_GMII:
+ nports = __cvmx_helper_rgmii_probe(interface);
+ padding = CVMX_PKO_PADDING_60;
+ break;
+ /* SPI4 can have 1-16 ports depending on the device at the other end */
+ case CVMX_HELPER_INTERFACE_MODE_SPI:
+ nports = __cvmx_helper_spi_probe(interface);
+ padding = CVMX_PKO_PADDING_60;
+ break;
+ /* SGMII can have 1-4 ports depending on how many are hooked up */
+ case CVMX_HELPER_INTERFACE_MODE_SGMII:
+ padding = CVMX_PKO_PADDING_60;
+ case CVMX_HELPER_INTERFACE_MODE_PICMG:
+ nports = __cvmx_helper_sgmii_probe(interface);
+ has_fcs = 1;
+ break;
+ /* PCI target Network Packet Interface */
+ case CVMX_HELPER_INTERFACE_MODE_NPI:
+ nports = __cvmx_helper_npi_probe(interface);
+ break;
+ /* Special loopback only ports. These are not the same as other ports
+ in loopback mode */
+ case CVMX_HELPER_INTERFACE_MODE_LOOP:
+ nports = __cvmx_helper_loop_probe(interface);
+ break;
+ /* SRIO has 2^N ports, where N is number of interfaces */
+ case CVMX_HELPER_INTERFACE_MODE_SRIO:
+ nports = __cvmx_helper_srio_probe(interface);
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_ILK:
+ nports = __cvmx_helper_ilk_probe(interface);
+ has_fcs = 1;
+ padding = CVMX_PKO_PADDING_60;
+ break;
+ }
+
+ if (nports == -1)
+ return -1;
+
+ if (!octeon_has_feature(OCTEON_FEATURE_PKND))
+ has_fcs = 0;
+
+ nports = __cvmx_helper_board_interface_probe(interface, nports);
+ __cvmx_helper_init_interface(interface, nports, has_fcs, padding);
+ cvmx_helper_fcs_op(interface, nports, has_fcs);
+
+ /* Make sure all global variables propagate to other cores */
+ CVMX_SYNCWS;
+
+ return 0;
+}
+
+
+/**
+ * @INTERNAL
+ * Setup the IPD/PIP for the ports on an interface. Packet
+ * classification and tagging are set for every port on the
+ * interface. The number of ports on the interface must already
+ * have been probed.
+ *
+ * @param interface Interface to setup IPD/PIP for
+ *
+ * @return Zero on success, negative on failure
+ */
+static int __cvmx_helper_interface_setup_ipd(int interface)
+{
+
+ cvmx_helper_interface_mode_t mode;
+ int ipd_port = cvmx_helper_get_ipd_port(interface, 0);
+ int num_ports = cvmx_helper_ports_on_interface(interface);
+ int delta;
+
+ if (num_ports == CVMX_HELPER_CFG_INVALID_VALUE)
+ return 0;
+
+ mode = cvmx_helper_interface_get_mode(interface);
+
+ if (mode == CVMX_HELPER_INTERFACE_MODE_LOOP)
+ __cvmx_helper_loop_enable(interface);
+
+ delta = 1;
+ if (octeon_has_feature(OCTEON_FEATURE_PKND))
+ {
+ if (mode == CVMX_HELPER_INTERFACE_MODE_SGMII)
+ delta = 16;
+ }
+
+ while (num_ports--)
+ {
+ __cvmx_helper_port_setup_ipd(ipd_port);
+ ipd_port += delta;
+ }
+
+ return 0;
+}
+
+
+/**
+ * @INTERNAL
+ * Setup global setting for IPD/PIP not related to a specific
+ * interface or port. This must be called before IPD is enabled.
+ *
+ * @return Zero on success, negative on failure.
+ */
+static int __cvmx_helper_global_setup_ipd(void)
+{
+#ifndef CVMX_HELPER_IPD_DRAM_MODE
+#define CVMX_HELPER_IPD_DRAM_MODE CVMX_IPD_OPC_MODE_STT
+#endif
+ /* Setup the global packet input options */
+ cvmx_ipd_config(CVMX_FPA_PACKET_POOL_SIZE/8,
+ CVMX_HELPER_FIRST_MBUFF_SKIP/8,
+ CVMX_HELPER_NOT_FIRST_MBUFF_SKIP/8,
+ (CVMX_HELPER_FIRST_MBUFF_SKIP+8) / 128, /* The +8 is to account for the next ptr */
+ (CVMX_HELPER_NOT_FIRST_MBUFF_SKIP+8) / 128, /* The +8 is to account for the next ptr */
+ CVMX_FPA_WQE_POOL,
+ CVMX_HELPER_IPD_DRAM_MODE,
+ 1);
+ return 0;
+}
+
+
+/**
+ * @INTERNAL
+ * Setup the PKO for the ports on an interface. The number of
+ * queues per port and the priority of each PKO output queue
+ * is set here. PKO must be disabled when this function is called.
+ *
+ * @param interface Interface to setup PKO for
+ *
+ * @return Zero on success, negative on failure
+ */
+static int __cvmx_helper_interface_setup_pko(int interface)
+{
+ /* Each packet output queue has an associated priority. The higher the
+ priority, the more often it can send a packet. A priority of 8 means
+ it can send in all 8 rounds of contention. We're going to make each
+ queue one less than the last.
+ The vector of priorities has been extended to support CN5xxx CPUs,
+ where up to 16 queues can be associated to a port.
+ To keep backward compatibility we don't change the initial 8
+ priorities and replicate them in the second half.
+ With per-core PKO queues (PKO lockless operation) all queues have
+ the same priority. */
+ /* uint64_t priorities[16] = {8,7,6,5,4,3,2,1,8,7,6,5,4,3,2,1}; */
+ uint64_t priorities[16] = {[0 ... 15] = 8};
+
+ /* Setup the IPD/PIP and PKO for the ports discovered above. Here packet
+ classification, tagging and output priorities are set */
+ int ipd_port = cvmx_helper_get_ipd_port(interface, 0);
+ int num_ports = cvmx_helper_ports_on_interface(interface);
+ while (num_ports--)
+ {
+ /* Give the user a chance to override the per queue priorities */
+ if (cvmx_override_pko_queue_priority)
+ cvmx_override_pko_queue_priority(ipd_port, priorities);
+
+ cvmx_pko_config_port(ipd_port, cvmx_pko_get_base_queue_per_core(ipd_port, 0),
+ cvmx_pko_get_num_queues(ipd_port), priorities);
+ ipd_port++;
+ }
+ return 0;
+}
+
+
+/**
+ * @INTERNAL
+ * Setup global setting for PKO not related to a specific
+ * interface or port. This must be called before PKO is enabled.
+ *
+ * @return Zero on success, negative on failure.
+ */
+static int __cvmx_helper_global_setup_pko(void)
+{
+ /* Disable tagwait FAU timeout. This needs to be done before anyone might
+ start packet output using tags */
+ cvmx_iob_fau_timeout_t fau_to;
+ fau_to.u64 = 0;
+ fau_to.s.tout_val = 0xfff;
+ fau_to.s.tout_enb = 0;
+ cvmx_write_csr(CVMX_IOB_FAU_TIMEOUT, fau_to.u64);
+
+ if (OCTEON_IS_MODEL(OCTEON_CN68XX)) {
+ cvmx_pko_reg_min_pkt_t min_pkt;
+
+ min_pkt.u64 = 0;
+ min_pkt.s.size1 = 59;
+ min_pkt.s.size2 = 59;
+ min_pkt.s.size3 = 59;
+ min_pkt.s.size4 = 59;
+ min_pkt.s.size5 = 59;
+ min_pkt.s.size6 = 59;
+ min_pkt.s.size7 = 59;
+ cvmx_write_csr(CVMX_PKO_REG_MIN_PKT, min_pkt.u64);
+ }
+
+ return 0;
+}
+
+
+/**
+ * @INTERNAL
+ * Setup global backpressure setting.
+ *
+ * @return Zero on success, negative on failure
+ */
+static int __cvmx_helper_global_setup_backpressure(void)
+{
+#if CVMX_HELPER_DISABLE_RGMII_BACKPRESSURE
+ /* Disable backpressure if configured to do so */
+ /* Disable backpressure (pause frame) generation */
+ int num_interfaces = cvmx_helper_get_number_of_interfaces();
+ int interface;
+ for (interface=0; interface<num_interfaces; interface++)
+ {
+ switch (cvmx_helper_interface_get_mode(interface))
+ {
+ case CVMX_HELPER_INTERFACE_MODE_DISABLED:
+ case CVMX_HELPER_INTERFACE_MODE_PCIE:
+ case CVMX_HELPER_INTERFACE_MODE_SRIO:
+ case CVMX_HELPER_INTERFACE_MODE_ILK:
+ case CVMX_HELPER_INTERFACE_MODE_NPI:
+ case CVMX_HELPER_INTERFACE_MODE_LOOP:
+ case CVMX_HELPER_INTERFACE_MODE_XAUI:
+ case CVMX_HELPER_INTERFACE_MODE_RXAUI:
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_RGMII:
+ case CVMX_HELPER_INTERFACE_MODE_GMII:
+ case CVMX_HELPER_INTERFACE_MODE_SPI:
+ case CVMX_HELPER_INTERFACE_MODE_SGMII:
+ case CVMX_HELPER_INTERFACE_MODE_PICMG:
+ cvmx_gmx_set_backpressure_override(interface, 0xf);
+ break;
+ }
+ }
+ //cvmx_dprintf("Disabling backpressure\n");
+#endif
+
+ return 0;
+}
+
+/**
+ * @INTERNAL
+ * Verify the per port IPD backpressure is aligned properly.
+ * @return Zero if working, non zero if misaligned
+ */
+static int __cvmx_helper_backpressure_is_misaligned(void)
+{
+ uint64_t ipd_int_enb;
+ cvmx_ipd_ctl_status_t ipd_reg;
+ uint64_t bp_status0;
+ uint64_t bp_status1;
+ const int port0 = 0;
+ const int port1 = 16;
+ cvmx_helper_interface_mode_t mode0 = cvmx_helper_interface_get_mode(0);
+ cvmx_helper_interface_mode_t mode1 = cvmx_helper_interface_get_mode(1);
+
+ /* Disable error interrupts while we check backpressure */
+ ipd_int_enb = cvmx_read_csr(CVMX_IPD_INT_ENB);
+ cvmx_write_csr(CVMX_IPD_INT_ENB, 0);
+
+ /* Enable per port backpressure */
+ ipd_reg.u64 = cvmx_read_csr(CVMX_IPD_CTL_STATUS);
+ ipd_reg.s.pbp_en = 1;
+ cvmx_write_csr(CVMX_IPD_CTL_STATUS, ipd_reg.u64);
+
+ if (mode0 != CVMX_HELPER_INTERFACE_MODE_DISABLED)
+ {
+ /* Enable backpressure for port with a zero threshold */
+ cvmx_write_csr(CVMX_IPD_PORTX_BP_PAGE_CNT(port0), 1<<17);
+ /* Add 1000 to the page count to simulate packets coming in */
+ cvmx_write_csr(CVMX_IPD_SUB_PORT_BP_PAGE_CNT, (port0<<25) | 1000);
+ }
+
+ if (mode1 != CVMX_HELPER_INTERFACE_MODE_DISABLED)
+ {
+ /* Enable backpressure for port with a zero threshold */
+ cvmx_write_csr(CVMX_IPD_PORTX_BP_PAGE_CNT(port1), 1<<17);
+ /* Add 1000 to the page count to simulate packets coming in */
+ cvmx_write_csr(CVMX_IPD_SUB_PORT_BP_PAGE_CNT, (port1<<25) | 1000);
+ }
+
+ /* Wait 500 cycles for the BP to update */
+ cvmx_wait(500);
+
+ /* Read the BP state from the debug select register */
+ switch (mode0)
+ {
+ case CVMX_HELPER_INTERFACE_MODE_SPI:
+ cvmx_write_csr(CVMX_NPI_DBG_SELECT, 0x9004);
+ bp_status0 = cvmx_read_csr(CVMX_DBG_DATA);
+ bp_status0 = 0xffff & ~bp_status0;
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_RGMII:
+ case CVMX_HELPER_INTERFACE_MODE_GMII:
+ cvmx_write_csr(CVMX_NPI_DBG_SELECT, 0x0e00);
+ bp_status0 = 0xffff & cvmx_read_csr(CVMX_DBG_DATA);
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_XAUI:
+ case CVMX_HELPER_INTERFACE_MODE_SGMII:
+ case CVMX_HELPER_INTERFACE_MODE_PICMG:
+ cvmx_write_csr(CVMX_PEXP_NPEI_DBG_SELECT, 0x0e00);
+ bp_status0 = 0xffff & cvmx_read_csr(CVMX_PEXP_NPEI_DBG_DATA);
+ break;
+ default:
+ bp_status0 = 1<<port0;
+ break;
+ }
+
+ /* Read the BP state from the debug select register */
+ switch (mode1)
+ {
+ case CVMX_HELPER_INTERFACE_MODE_SPI:
+ cvmx_write_csr(CVMX_NPI_DBG_SELECT, 0x9804);
+ bp_status1 = cvmx_read_csr(CVMX_DBG_DATA);
+ bp_status1 = 0xffff & ~bp_status1;
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_RGMII:
+ case CVMX_HELPER_INTERFACE_MODE_GMII:
+ cvmx_write_csr(CVMX_NPI_DBG_SELECT, 0x1600);
+ bp_status1 = 0xffff & cvmx_read_csr(CVMX_DBG_DATA);
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_XAUI:
+ case CVMX_HELPER_INTERFACE_MODE_SGMII:
+ case CVMX_HELPER_INTERFACE_MODE_PICMG:
+ cvmx_write_csr(CVMX_PEXP_NPEI_DBG_SELECT, 0x1600);
+ bp_status1 = 0xffff & cvmx_read_csr(CVMX_PEXP_NPEI_DBG_DATA);
+ break;
+ default:
+ bp_status1 = 1<<(port1-16);
+ break;
+ }
+
+ if (mode0 != CVMX_HELPER_INTERFACE_MODE_DISABLED)
+ {
+ /* Shutdown BP */
+ cvmx_write_csr(CVMX_IPD_SUB_PORT_BP_PAGE_CNT, (port0<<25) | (0x1ffffff & -1000));
+ cvmx_write_csr(CVMX_IPD_PORTX_BP_PAGE_CNT(port0), 0);
+ }
+
+ if (mode1 != CVMX_HELPER_INTERFACE_MODE_DISABLED)
+ {
+ /* Shutdown BP */
+ cvmx_write_csr(CVMX_IPD_SUB_PORT_BP_PAGE_CNT, (port1<<25) | (0x1ffffff & -1000));
+ cvmx_write_csr(CVMX_IPD_PORTX_BP_PAGE_CNT(port1), 0);
+ }
+
+ /* Clear any error interrupts that might have been set */
+ cvmx_write_csr(CVMX_IPD_INT_SUM, 0x1f);
+ cvmx_write_csr(CVMX_IPD_INT_ENB, ipd_int_enb);
+
+ return ((bp_status0 != 1ull<<port0) || (bp_status1 != 1ull<<(port1-16)));
+}
+
+
+/**
+ * @INTERNAL
+ * Enable packet input/output from the hardware. This function is
+ * called after all internal setup is complete and IPD is enabled.
+ * After this function completes, packets will be accepted from the
+ * hardware ports. PKO should still be disabled to make sure packets
+ * aren't sent out partially setup hardware.
+ *
+ * @param interface Interface to enable
+ *
+ * @return Zero on success, negative on failure
+ */
+static int __cvmx_helper_packet_hardware_enable(int interface)
+{
+ int result = 0;
+ switch (cvmx_helper_interface_get_mode(interface))
+ {
+ /* These types don't support ports to IPD/PKO */
+ case CVMX_HELPER_INTERFACE_MODE_DISABLED:
+ case CVMX_HELPER_INTERFACE_MODE_PCIE:
+ case CVMX_HELPER_INTERFACE_MODE_LOOP:
+ /* Nothing to do */
+ break;
+ /* XAUI is a single high speed port */
+ case CVMX_HELPER_INTERFACE_MODE_XAUI:
+ case CVMX_HELPER_INTERFACE_MODE_RXAUI:
+ result = __cvmx_helper_xaui_enable(interface);
+ break;
+ /* RGMII/GMII/MII are all treated about the same. Most functions
+ refer to these ports as RGMII */
+ case CVMX_HELPER_INTERFACE_MODE_RGMII:
+ case CVMX_HELPER_INTERFACE_MODE_GMII:
+ result = __cvmx_helper_rgmii_enable(interface);
+ break;
+ /* SPI4 can have 1-16 ports depending on the device at the other end */
+ case CVMX_HELPER_INTERFACE_MODE_SPI:
+ result = __cvmx_helper_spi_enable(interface);
+ break;
+ /* SGMII can have 1-4 ports depending on how many are hooked up */
+ case CVMX_HELPER_INTERFACE_MODE_SGMII:
+ case CVMX_HELPER_INTERFACE_MODE_PICMG:
+ result = __cvmx_helper_sgmii_enable(interface);
+ break;
+ /* PCI target Network Packet Interface */
+ case CVMX_HELPER_INTERFACE_MODE_NPI:
+ result = __cvmx_helper_npi_enable(interface);
+ break;
+ /* SRIO has 2^N ports, where N is number of interfaces */
+ case CVMX_HELPER_INTERFACE_MODE_SRIO:
+ result = __cvmx_helper_srio_enable(interface);
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_ILK:
+ result = __cvmx_helper_ilk_enable(interface);
+ break;
+ }
+ result |= __cvmx_helper_board_hardware_enable(interface);
+ return result;
+}
+
+
+/**
+ * Called after all internal packet IO paths are setup. This
+ * function enables IPD/PIP and begins packet input and output.
+ *
+ * @return Zero on success, negative on failure
+ */
+int cvmx_helper_ipd_and_packet_input_enable(void)
+{
+ int num_interfaces;
+ int interface;
+
+ /* Enable IPD */
+ cvmx_ipd_enable();
+
+ /* Time to enable hardware ports packet input and output. Note that at this
+ point IPD/PIP must be fully functional and PKO must be disabled */
+ num_interfaces = cvmx_helper_get_number_of_interfaces();
+ for (interface=0; interface<num_interfaces; interface++)
+ {
+ if (cvmx_helper_ports_on_interface(interface) > 0)
+ {
+ //cvmx_dprintf("Enabling packet I/O on interface %d\n", interface);
+ __cvmx_helper_packet_hardware_enable(interface);
+ }
+ }
+
+ /* Finally enable PKO now that the entire path is up and running */
+ cvmx_pko_enable();
+
+ if ((OCTEON_IS_MODEL(OCTEON_CN31XX_PASS1) || OCTEON_IS_MODEL(OCTEON_CN30XX_PASS1)) &&
+ (cvmx_sysinfo_get()->board_type != CVMX_BOARD_TYPE_SIM))
+ __cvmx_helper_errata_fix_ipd_ptr_alignment();
+ return 0;
+}
+#ifdef CVMX_BUILD_FOR_LINUX_KERNEL
+EXPORT_SYMBOL(cvmx_helper_ipd_and_packet_input_enable);
+#endif
+
+#define __CVMX_SSO_RWQ_SIZE 256
+
+int cvmx_helper_initialize_sso(int wqe_entries)
+{
+ int cvm_oct_sso_number_rwq_bufs;
+ char *mem;
+ int i;
+ cvmx_sso_cfg_t sso_cfg;
+ cvmx_fpa_fpfx_marks_t fpa_marks;
+
+ if (!OCTEON_IS_MODEL(OCTEON_CN68XX))
+ return 0;
+
+ /*
+ * CN68XX-P1 may reset with the wrong values, put in
+ * the correct values.
+ */
+ fpa_marks.u64 = 0;
+ fpa_marks.s.fpf_wr = 0xa4;
+ fpa_marks.s.fpf_rd = 0x40;
+ cvmx_write_csr(CVMX_FPA_FPF8_MARKS, fpa_marks.u64);
+
+ cvm_oct_sso_number_rwq_bufs = ((wqe_entries - 1) / 26) + 1 + 48 + 8;
+
+ mem = cvmx_bootmem_alloc(__CVMX_SSO_RWQ_SIZE * cvm_oct_sso_number_rwq_bufs, CVMX_CACHE_LINE_SIZE);
+ if (mem == NULL) {
+ cvmx_dprintf("Out of memory initializing sso pool\n");
+ return -1;
+ }
+ /* Make sure RWI/RWO is disabled. */
+ sso_cfg.u64 = cvmx_read_csr(CVMX_SSO_CFG);
+ sso_cfg.s.rwen = 0;
+ cvmx_write_csr(CVMX_SSO_CFG, sso_cfg.u64);
+
+ for (i = cvm_oct_sso_number_rwq_bufs - 8; i > 0; i--) {
+ cvmx_sso_rwq_psh_fptr_t fptr;
+
+ for (;;) {
+ fptr.u64 = cvmx_read_csr(CVMX_SSO_RWQ_PSH_FPTR);
+ if (!fptr.s.full)
+ break;
+ cvmx_wait(1000);
+ }
+ fptr.s.fptr = cvmx_ptr_to_phys(mem) >> 7;
+ cvmx_write_csr(CVMX_SSO_RWQ_PSH_FPTR, fptr.u64);
+ mem = mem + __CVMX_SSO_RWQ_SIZE;
+ }
+
+ for (i = 0; i < 8; i++) {
+ cvmx_sso_rwq_head_ptrx_t head_ptr;
+ cvmx_sso_rwq_tail_ptrx_t tail_ptr;
+
+ head_ptr.u64 = 0;
+ tail_ptr.u64 = 0;
+ head_ptr.s.ptr = cvmx_ptr_to_phys(mem) >> 7;
+ tail_ptr.s.ptr = head_ptr.s.ptr;
+ cvmx_write_csr(CVMX_SSO_RWQ_HEAD_PTRX(i), head_ptr.u64);
+ cvmx_write_csr(CVMX_SSO_RWQ_TAIL_PTRX(i), tail_ptr.u64);
+ mem = mem + __CVMX_SSO_RWQ_SIZE;
+ }
+
+ sso_cfg.u64 = cvmx_read_csr(CVMX_SSO_CFG);
+ sso_cfg.s.rwen = 1;
+ sso_cfg.s.dwb = cvmx_helper_cfg_opt_get(CVMX_HELPER_CFG_OPT_USE_DWB);
+ sso_cfg.s.rwq_byp_dis = 0;
+ sso_cfg.s.rwio_byp_dis = 0;
+ cvmx_write_csr(CVMX_SSO_CFG, sso_cfg.u64);
+
+ return 0;
+}
+
+int cvmx_helper_uninitialize_sso(void)
+{
+ cvmx_fpa_quex_available_t queue_available;
+ cvmx_sso_cfg_t sso_cfg;
+ cvmx_sso_rwq_pop_fptr_t pop_fptr;
+ cvmx_sso_rwq_psh_fptr_t fptr;
+ cvmx_sso_fpage_cnt_t fpage_cnt;
+ int num_to_transfer, i;
+ char *mem;
+
+ if (!OCTEON_IS_MODEL(OCTEON_CN68XX))
+ return 0;
+
+ sso_cfg.u64 = cvmx_read_csr(CVMX_SSO_CFG);
+ sso_cfg.s.rwen = 0;
+ sso_cfg.s.rwq_byp_dis = 1;
+ cvmx_write_csr(CVMX_SSO_CFG, sso_cfg.u64);
+ cvmx_read_csr(CVMX_SSO_CFG);
+ queue_available.u64 = cvmx_read_csr(CVMX_FPA_QUEX_AVAILABLE(8));
+
+ /* Make CVMX_FPA_QUEX_AVAILABLE(8) % 16 == 0*/
+ for (num_to_transfer = (16 - queue_available.s.que_siz) % 16;
+ num_to_transfer > 0; num_to_transfer--) {
+ do {
+ pop_fptr.u64 = cvmx_read_csr(CVMX_SSO_RWQ_POP_FPTR);
+ } while (!pop_fptr.s.val);
+ for (;;) {
+ fptr.u64 = cvmx_read_csr(CVMX_SSO_RWQ_PSH_FPTR);
+ if (!fptr.s.full)
+ break;
+ cvmx_wait(1000);
+ }
+ fptr.s.fptr = pop_fptr.s.fptr;
+ cvmx_write_csr(CVMX_SSO_RWQ_PSH_FPTR, fptr.u64);
+ }
+ cvmx_read_csr(CVMX_SSO_CFG);
+
+ do {
+ queue_available.u64 = cvmx_read_csr(CVMX_FPA_QUEX_AVAILABLE(8));
+ } while (queue_available.s.que_siz % 16);
+
+ sso_cfg.s.rwen = 1;
+ sso_cfg.s.rwq_byp_dis = 0;
+ cvmx_write_csr(CVMX_SSO_CFG, sso_cfg.u64);
+
+ for (i = 0; i < 8; i++) {
+ cvmx_sso_rwq_head_ptrx_t head_ptr;
+ cvmx_sso_rwq_tail_ptrx_t tail_ptr;
+
+ head_ptr.u64 = cvmx_read_csr(CVMX_SSO_RWQ_HEAD_PTRX(i));
+ tail_ptr.u64 = cvmx_read_csr(CVMX_SSO_RWQ_TAIL_PTRX(i));
+ if (head_ptr.s.ptr != tail_ptr.s.ptr) {
+ cvmx_dprintf("head_ptr.s.ptr != tail_ptr.s.ptr, idx: %d\n", i);
+ }
+
+ mem = cvmx_phys_to_ptr(((uint64_t)head_ptr.s.ptr) << 7);
+ /* Leak the memory */
+ }
+
+ do {
+ do {
+ pop_fptr.u64 = cvmx_read_csr(CVMX_SSO_RWQ_POP_FPTR);
+ if (pop_fptr.s.val) {
+ mem = cvmx_phys_to_ptr(((uint64_t)pop_fptr.s.fptr) << 7);
+ /* Leak the memory */
+ }
+ } while (pop_fptr.s.val);
+ fpage_cnt.u64 = cvmx_read_csr(CVMX_SSO_FPAGE_CNT);
+ } while (fpage_cnt.s.fpage_cnt);
+
+ sso_cfg.s.rwen = 0;
+ sso_cfg.s.rwq_byp_dis = 0;
+ cvmx_write_csr(CVMX_SSO_CFG, sso_cfg.u64);
+
+ return 0;
+}
+
+/**
+ * Initialize the PIP, IPD, and PKO hardware to support
+ * simple priority based queues for the ethernet ports. Each
+ * port is configured with a number of priority queues based
+ * on CVMX_PKO_QUEUES_PER_PORT_* where each queue is lower
+ * priority than the previous.
+ *
+ * @return Zero on success, non-zero on failure
+ */
+int cvmx_helper_initialize_packet_io_global(void)
+{
+ int result = 0;
+ int interface;
+ cvmx_l2c_cfg_t l2c_cfg;
+ cvmx_smix_en_t smix_en;
+ const int num_interfaces = cvmx_helper_get_number_of_interfaces();
+
+ /* CN52XX pass 1: Due to a bug in 2nd order CDR, it needs to be disabled */
+ if (OCTEON_IS_MODEL(OCTEON_CN52XX_PASS1_0))
+ __cvmx_helper_errata_qlm_disable_2nd_order_cdr(1);
+
+ /* Tell L2 to give the IOB statically higher priority compared to the
+ cores. This avoids conditions where IO blocks might be starved under
+ very high L2 loads */
+ if (OCTEON_IS_MODEL(OCTEON_CN6XXX) || OCTEON_IS_MODEL(OCTEON_CNF7XXX))
+ {
+ cvmx_l2c_ctl_t l2c_ctl;
+ l2c_ctl.u64 = cvmx_read_csr(CVMX_L2C_CTL);
+ l2c_ctl.s.rsp_arb_mode = 1;
+ l2c_ctl.s.xmc_arb_mode = 0;
+ cvmx_write_csr(CVMX_L2C_CTL, l2c_ctl.u64);
+ }
+ else
+ {
+ l2c_cfg.u64 = cvmx_read_csr(CVMX_L2C_CFG);
+ l2c_cfg.s.lrf_arb_mode = 0;
+ l2c_cfg.s.rfb_arb_mode = 0;
+ cvmx_write_csr(CVMX_L2C_CFG, l2c_cfg.u64);
+ }
+
+ if (cvmx_sysinfo_get()->board_type != CVMX_BOARD_TYPE_SIM)
+ {
+ int smi_inf = 1;
+ int i;
+
+ /* Newer chips have more than one SMI/MDIO interface */
+ if (OCTEON_IS_MODEL(OCTEON_CN68XX))
+ smi_inf = 4;
+ else if (!OCTEON_IS_MODEL(OCTEON_CN3XXX)
+ && !OCTEON_IS_MODEL(OCTEON_CN58XX)
+ && !OCTEON_IS_MODEL(OCTEON_CN50XX))
+ smi_inf = 2;
+
+ for (i = 0; i < smi_inf; i++)
+ {
+ /* Make sure SMI/MDIO is enabled so we can query PHYs */
+ smix_en.u64 = cvmx_read_csr(CVMX_SMIX_EN(i));
+ if (!smix_en.s.en)
+ {
+ smix_en.s.en = 1;
+ cvmx_write_csr(CVMX_SMIX_EN(i), smix_en.u64);
+ }
+ }
+ }
+
+ __cvmx_helper_cfg_init();
+
+ for (interface=0; interface<num_interfaces; interface++)
+ result |= cvmx_helper_interface_probe(interface);
+
+ cvmx_pko_initialize_global();
+ for (interface=0; interface<num_interfaces; interface++)
+ {
+ if (cvmx_helper_ports_on_interface(interface) > 0)
+ cvmx_dprintf("Interface %d has %d ports (%s)\n",
+ interface, cvmx_helper_ports_on_interface(interface),
+ cvmx_helper_interface_mode_to_string(cvmx_helper_interface_get_mode(interface)));
+ result |= __cvmx_helper_interface_setup_ipd(interface);
+ if (!OCTEON_IS_MODEL(OCTEON_CN68XX))
+ result |= __cvmx_helper_interface_setup_pko(interface);
+ }
+
+ result |= __cvmx_helper_global_setup_ipd();
+ result |= __cvmx_helper_global_setup_pko();
+
+ /* Enable any flow control and backpressure */
+ result |= __cvmx_helper_global_setup_backpressure();
+
+#if CVMX_HELPER_ENABLE_IPD
+ result |= cvmx_helper_ipd_and_packet_input_enable();
+#endif
+ return result;
+}
+#ifdef CVMX_BUILD_FOR_LINUX_KERNEL
+EXPORT_SYMBOL(cvmx_helper_initialize_packet_io_global);
+#endif
+
+
+/**
+ * Does core local initialization for packet io
+ *
+ * @return Zero on success, non-zero on failure
+ */
+int cvmx_helper_initialize_packet_io_local(void)
+{
+ return cvmx_pko_initialize_local();
+}
+
+/**
+ * wait for the pko queue to drain
+ *
+ * @param queue a valid pko queue
+ * @return count is the length of the queue after calling this
+ * function
+ */
+static int cvmx_helper_wait_pko_queue_drain(int queue)
+{
+ const int timeout = 5; /* Wait up to 5 seconds for timeouts */
+ int count;
+ uint64_t start_cycle, stop_cycle;
+
+ count = cvmx_cmd_queue_length(CVMX_CMD_QUEUE_PKO(queue));
+ start_cycle = cvmx_get_cycle();
+ stop_cycle = start_cycle + cvmx_clock_get_rate(CVMX_CLOCK_CORE) * timeout;
+ while (count && (cvmx_get_cycle() < stop_cycle))
+ {
+ cvmx_wait(10000);
+ count = cvmx_cmd_queue_length(CVMX_CMD_QUEUE_PKO(queue));
+ }
+
+ return count;
+}
+
+struct cvmx_buffer_list {
+ struct cvmx_buffer_list *next;
+};
+
+/**
+ * Undo the initialization performed in
+ * cvmx_helper_initialize_packet_io_global(). After calling this routine and the
+ * local version on each core, packet IO for Octeon will be disabled and placed
+ * in the initial reset state. It will then be safe to call the initialize
+ * later on. Note that this routine does not empty the FPA pools. It frees all
+ * buffers used by the packet IO hardware to the FPA so a function emptying the
+ * FPA after shutdown should find all packet buffers in the FPA.
+ *
+ * @return Zero on success, negative on failure.
+ */
+int cvmx_helper_shutdown_packet_io_global(void)
+{
+ const int timeout = 5; /* Wait up to 5 seconds for timeouts */
+ int result = 0;
+ int num_interfaces;
+ int interface;
+ int num_ports;
+ int index;
+ struct cvmx_buffer_list *pool0_buffers;
+ struct cvmx_buffer_list *pool0_buffers_tail;
+ cvmx_wqe_t *work;
+
+ /* Step 1: Disable all backpressure */
+ for (interface=0; interface<CVMX_HELPER_MAX_GMX; interface++)
+ if (cvmx_helper_interface_get_mode(interface) != CVMX_HELPER_INTERFACE_MODE_DISABLED)
+ cvmx_gmx_set_backpressure_override(interface, 0xf);
+
+step2:
+ /* Step 2: Wait for the PKO queues to drain */
+ if (octeon_has_feature(OCTEON_FEATURE_PKND))
+ {
+ int queue, max_queue;
+
+ max_queue = __cvmx_helper_cfg_pko_max_queue();
+ for (queue = 0; queue < max_queue; queue++)
+ {
+ if (cvmx_helper_wait_pko_queue_drain(queue))
+ {
+ result = -1;
+ goto step3;
+ }
+ }
+ }
+ else
+ {
+ num_interfaces = cvmx_helper_get_number_of_interfaces();
+ for (interface=0; interface<num_interfaces; interface++)
+ {
+ num_ports = cvmx_helper_ports_on_interface(interface);
+ for (index=0; index<num_ports; index++)
+ {
+ int pko_port = cvmx_helper_get_ipd_port(interface, index);
+ int queue = cvmx_pko_get_base_queue(pko_port);
+ int max_queue = queue + cvmx_pko_get_num_queues(pko_port);
+ while (queue < max_queue)
+ {
+ if (cvmx_helper_wait_pko_queue_drain(queue))
+ {
+ result = -1;
+ goto step3;
+ }
+ queue++;
+ }
+ }
+ }
+ }
+
+step3:
+ /* Step 3: Disable TX and RX on all ports */
+ for (interface=0; interface<CVMX_HELPER_MAX_GMX; interface++)
+ {
+ switch (cvmx_helper_interface_get_mode(interface))
+ {
+ case CVMX_HELPER_INTERFACE_MODE_DISABLED:
+ case CVMX_HELPER_INTERFACE_MODE_PCIE:
+ /* Not a packet interface */
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_NPI:
+ case CVMX_HELPER_INTERFACE_MODE_SRIO:
+ case CVMX_HELPER_INTERFACE_MODE_ILK:
+ /* We don't handle the NPI/NPEI/SRIO packet engines. The caller
+ must know these are idle */
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_LOOP:
+ /* Nothing needed. Once PKO is idle, the loopback devices
+ must be idle */
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_SPI:
+ /* SPI cannot be disabled from Octeon. It is the responsibility
+ of the caller to make sure SPI is idle before doing
+ shutdown */
+ /* Fall through and do the same processing as RGMII/GMII */
+ case CVMX_HELPER_INTERFACE_MODE_GMII:
+ case CVMX_HELPER_INTERFACE_MODE_RGMII:
+ /* Disable outermost RX at the ASX block */
+ cvmx_write_csr(CVMX_ASXX_RX_PRT_EN(interface), 0);
+ num_ports = cvmx_helper_ports_on_interface(interface);
+ if (num_ports > 4)
+ num_ports = 4;
+ for (index=0; index<num_ports; index++)
+ {
+ cvmx_gmxx_prtx_cfg_t gmx_cfg;
+ gmx_cfg.u64 = cvmx_read_csr(CVMX_GMXX_PRTX_CFG(index, interface));
+ gmx_cfg.s.en = 0;
+ cvmx_write_csr(CVMX_GMXX_PRTX_CFG(index, interface), gmx_cfg.u64);
+ /* Poll the GMX state machine waiting for it to become idle */
+ cvmx_write_csr(CVMX_NPI_DBG_SELECT, interface*0x800 + index*0x100 + 0x880);
+ if (CVMX_WAIT_FOR_FIELD64(CVMX_DBG_DATA, cvmx_dbg_data_t, data&7, ==, 0, timeout*1000000))
+ {
+ cvmx_dprintf("GMX RX path timeout waiting for idle\n");
+ result = -1;
+ }
+ if (CVMX_WAIT_FOR_FIELD64(CVMX_DBG_DATA, cvmx_dbg_data_t, data&0xf, ==, 0, timeout*1000000))
+ {
+ cvmx_dprintf("GMX TX path timeout waiting for idle\n");
+ result = -1;
+ }
+ }
+ /* Disable outermost TX at the ASX block */
+ cvmx_write_csr(CVMX_ASXX_TX_PRT_EN(interface), 0);
+ /* Disable interrupts for interface */
+ cvmx_write_csr(CVMX_ASXX_INT_EN(interface), 0);
+ cvmx_write_csr(CVMX_GMXX_TX_INT_EN(interface), 0);
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_XAUI:
+ case CVMX_HELPER_INTERFACE_MODE_RXAUI:
+ case CVMX_HELPER_INTERFACE_MODE_SGMII:
+ case CVMX_HELPER_INTERFACE_MODE_PICMG:
+ num_ports = cvmx_helper_ports_on_interface(interface);
+ if (num_ports > 4)
+ num_ports = 4;
+ for (index=0; index<num_ports; index++)
+ {
+ cvmx_gmxx_prtx_cfg_t gmx_cfg;
+ gmx_cfg.u64 = cvmx_read_csr(CVMX_GMXX_PRTX_CFG(index, interface));
+ gmx_cfg.s.en = 0;
+ cvmx_write_csr(CVMX_GMXX_PRTX_CFG(index, interface), gmx_cfg.u64);
+ if (CVMX_WAIT_FOR_FIELD64(CVMX_GMXX_PRTX_CFG(index, interface), cvmx_gmxx_prtx_cfg_t, rx_idle, ==, 1, timeout*1000000))
+ {
+ cvmx_dprintf("GMX RX path timeout waiting for idle\n");
+ result = -1;
+ }
+ if (CVMX_WAIT_FOR_FIELD64(CVMX_GMXX_PRTX_CFG(index, interface), cvmx_gmxx_prtx_cfg_t, tx_idle, ==, 1, timeout*1000000))
+ {
+ cvmx_dprintf("GMX TX path timeout waiting for idle\n");
+ result = -1;
+ }
+ }
+ break;
+ }
+ }
+
+ /* Step 4: Retrieve all packets from the POW and free them */
+ while ((work = cvmx_pow_work_request_sync(CVMX_POW_WAIT)))
+ {
+ cvmx_helper_free_packet_data(work);
+ cvmx_fpa_free(work, CVMX_FPA_WQE_POOL, 0);
+ }
+
+ /* Step 4b: Special workaround for pass 2 errata */
+ if (OCTEON_IS_MODEL(OCTEON_CN38XX_PASS2))
+ {
+ cvmx_ipd_ptr_count_t ipd_cnt;
+ int to_add;
+ ipd_cnt.u64 = cvmx_read_csr(CVMX_IPD_PTR_COUNT);
+ to_add = (ipd_cnt.s.wqev_cnt + ipd_cnt.s.wqe_pcnt) & 0x7;
+ if (to_add)
+ {
+ int port = -1;
+ cvmx_dprintf("Aligning CN38XX pass 2 IPD counters\n");
+ if (cvmx_helper_interface_get_mode(0) == CVMX_HELPER_INTERFACE_MODE_RGMII)
+ port = 0;
+ else if (cvmx_helper_interface_get_mode(1) == CVMX_HELPER_INTERFACE_MODE_RGMII)
+ port = 16;
+
+ if (port != -1)
+ {
+ char *buffer = cvmx_fpa_alloc(CVMX_FPA_PACKET_POOL);
+ if (buffer)
+ {
+ int queue = cvmx_pko_get_base_queue(port);
+ cvmx_pko_command_word0_t pko_command;
+ cvmx_buf_ptr_t packet;
+ uint64_t start_cycle;
+ uint64_t stop_cycle;
+
+ /* Populate a minimal packet */
+ memset(buffer, 0xff, 6);
+ memset(buffer+6, 0, 54);
+ pko_command.u64 = 0;
+ pko_command.s.dontfree = 1;
+ pko_command.s.total_bytes = 60;
+ pko_command.s.segs = 1;
+ packet.u64 = 0;
+ packet.s.addr = cvmx_ptr_to_phys(buffer);
+ packet.s.size = CVMX_FPA_PACKET_POOL_SIZE;
+ __cvmx_helper_rgmii_configure_loopback(port, 1, 0);
+ while (to_add--)
+ {
+ cvmx_pko_send_packet_prepare(port, queue, CVMX_PKO_LOCK_CMD_QUEUE);
+ if (cvmx_pko_send_packet_finish(port, queue, pko_command, packet, CVMX_PKO_LOCK_CMD_QUEUE))
+ {
+ cvmx_dprintf("ERROR: Unable to align IPD counters (PKO failed)\n");
+ break;
+ }
+ }
+ cvmx_fpa_free(buffer, CVMX_FPA_PACKET_POOL, 0);
+
+ /* Wait for the packets to loop back */
+ start_cycle = cvmx_get_cycle();
+ stop_cycle = start_cycle + cvmx_clock_get_rate(CVMX_CLOCK_CORE) * timeout;
+ while (cvmx_cmd_queue_length(CVMX_CMD_QUEUE_PKO(queue)) &&
+ (cvmx_get_cycle() < stop_cycle))
+ {
+ cvmx_wait(1000);
+ }
+ cvmx_wait(1000);
+ __cvmx_helper_rgmii_configure_loopback(port, 0, 0);
+ if (to_add == -1)
+ goto step2;
+ }
+ else
+ cvmx_dprintf("ERROR: Unable to align IPD counters (Packet pool empty)\n");
+ }
+ else
+ cvmx_dprintf("ERROR: Unable to align IPD counters\n");
+ }
+ }
+
+ /* Step 5 */
+ cvmx_ipd_disable();
+
+ /* Step 6: Drain all prefetched buffers from IPD/PIP. Note that IPD/PIP
+ have not been reset yet */
+ __cvmx_ipd_free_ptr();
+
+ /* Step 7: Free the PKO command buffers and put PKO in reset */
+ cvmx_pko_shutdown();
+
+ /* Step 8: Disable MAC address filtering */
+ for (interface=0; interface<CVMX_HELPER_MAX_GMX; interface++)
+ {
+ switch (cvmx_helper_interface_get_mode(interface))
+ {
+ case CVMX_HELPER_INTERFACE_MODE_DISABLED:
+ case CVMX_HELPER_INTERFACE_MODE_PCIE:
+ case CVMX_HELPER_INTERFACE_MODE_SRIO:
+ case CVMX_HELPER_INTERFACE_MODE_ILK:
+ case CVMX_HELPER_INTERFACE_MODE_NPI:
+ case CVMX_HELPER_INTERFACE_MODE_LOOP:
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_XAUI:
+ case CVMX_HELPER_INTERFACE_MODE_RXAUI:
+ case CVMX_HELPER_INTERFACE_MODE_GMII:
+ case CVMX_HELPER_INTERFACE_MODE_RGMII:
+ case CVMX_HELPER_INTERFACE_MODE_SPI:
+ case CVMX_HELPER_INTERFACE_MODE_SGMII:
+ case CVMX_HELPER_INTERFACE_MODE_PICMG:
+ num_ports = cvmx_helper_ports_on_interface(interface);
+ if (num_ports > 4)
+ num_ports = 4;
+ for (index=0; index<num_ports; index++)
+ {
+ cvmx_write_csr(CVMX_GMXX_RXX_ADR_CTL(index, interface), 1);
+ cvmx_write_csr(CVMX_GMXX_RXX_ADR_CAM_EN(index, interface), 0);
+ cvmx_write_csr(CVMX_GMXX_RXX_ADR_CAM0(index, interface), 0);
+ cvmx_write_csr(CVMX_GMXX_RXX_ADR_CAM1(index, interface), 0);
+ cvmx_write_csr(CVMX_GMXX_RXX_ADR_CAM2(index, interface), 0);
+ cvmx_write_csr(CVMX_GMXX_RXX_ADR_CAM3(index, interface), 0);
+ cvmx_write_csr(CVMX_GMXX_RXX_ADR_CAM4(index, interface), 0);
+ cvmx_write_csr(CVMX_GMXX_RXX_ADR_CAM5(index, interface), 0);
+ }
+ break;
+ }
+ }
+
+ /* Step 9: Drain all FPA buffers out of pool 0 before we reset
+ * IPD/PIP. This is needed to keep IPD_QUE0_FREE_PAGE_CNT in
+ * sync. We temporarily keep the buffers in the pool0_buffers
+ * list.
+ */
+ pool0_buffers = NULL;
+ pool0_buffers_tail = NULL;
+ while (1)
+ {
+ struct cvmx_buffer_list *buffer = cvmx_fpa_alloc(0);
+ if (buffer) {
+ buffer->next = NULL;
+
+ if (pool0_buffers == NULL)
+ pool0_buffers = buffer;
+ else
+ pool0_buffers_tail->next = buffer;
+
+ pool0_buffers_tail = buffer;
+ }
+ else
+ break;
+ }
+
+ /* Step 10: Reset IPD and PIP */
+ {
+ cvmx_ipd_ctl_status_t ipd_ctl_status;
+ ipd_ctl_status.u64 = cvmx_read_csr(CVMX_IPD_CTL_STATUS);
+ ipd_ctl_status.s.reset = 1;
+ cvmx_write_csr(CVMX_IPD_CTL_STATUS, ipd_ctl_status.u64);
+
+ if ((cvmx_sysinfo_get()->board_type != CVMX_BOARD_TYPE_SIM) &&
+ (OCTEON_IS_MODEL(OCTEON_CN3XXX) || OCTEON_IS_MODEL(OCTEON_CN5XXX)))
+ {
+ /* only try 1000 times. Normally if this works it will happen in
+ ** the first 50 loops. */
+ int max_loops = 1000;
+ int loop = 0;
+ /* Per port backpressure counters can get misaligned after an
+ IPD reset. This code realigns them by performing repeated
+ resets. See IPD-13473 */
+ cvmx_wait(100);
+ if (__cvmx_helper_backpressure_is_misaligned())
+ {
+ cvmx_dprintf("Starting to align per port backpressure counters.\n");
+ while (__cvmx_helper_backpressure_is_misaligned() && (loop++ < max_loops))
+ {
+ cvmx_write_csr(CVMX_IPD_CTL_STATUS, ipd_ctl_status.u64);
+ cvmx_wait(123);
+ }
+ if (loop < max_loops)
+ cvmx_dprintf("Completed aligning per port backpressure counters (%d loops).\n", loop);
+ else
+ {
+ cvmx_dprintf("ERROR: unable to align per port backpressure counters.\n");
+ /* For now, don't hang.... */
+ }
+ }
+ }
+
+ /* PIP_SFT_RST not present in CN38XXp{1,2} */
+ if (!OCTEON_IS_MODEL(OCTEON_CN38XX_PASS2))
+ {
+ cvmx_pip_sft_rst_t pip_sft_rst;
+ pip_sft_rst.u64 = cvmx_read_csr(CVMX_PIP_SFT_RST);
+ pip_sft_rst.s.rst = 1;
+ cvmx_write_csr(CVMX_PIP_SFT_RST, pip_sft_rst.u64);
+ }
+
+ /* Make sure IPD has finished reset. */
+ if (OCTEON_IS_MODEL(OCTEON_CN6XXX))
+ {
+ if (CVMX_WAIT_FOR_FIELD64(CVMX_IPD_CTL_STATUS, cvmx_ipd_ctl_status_t, rst_done, ==, 0, 1000))
+ {
+ cvmx_dprintf("IPD reset timeout waiting for idle\n");
+ result = -1;
+ }
+ }
+ }
+
+ /* Step 11: Restore the FPA buffers into pool 0 */
+ while (pool0_buffers) {
+ struct cvmx_buffer_list *n = pool0_buffers->next;
+ cvmx_fpa_free(pool0_buffers, 0, 0);
+ pool0_buffers = n;
+ }
+
+ /* Step 12: Release interface structures */
+ __cvmx_helper_shutdown_interfaces();
+
+ return result;
+}
+#ifdef CVMX_BUILD_FOR_LINUX_KERNEL
+EXPORT_SYMBOL(cvmx_helper_shutdown_packet_io_global);
+#endif
+
+
+/**
+ * Does core local shutdown of packet io
+ *
+ * @return Zero on success, non-zero on failure
+ */
+int cvmx_helper_shutdown_packet_io_local(void)
+{
+ /* Currently there is nothing to do per core. This may change in
+ the future */
+ return 0;
+}
+
+
+
+/**
+ * Auto configure an IPD/PKO port link state and speed. This
+ * function basically does the equivalent of:
+ * cvmx_helper_link_set(ipd_port, cvmx_helper_link_get(ipd_port));
+ *
+ * @param ipd_port IPD/PKO port to auto configure
+ *
+ * @return Link state after configure
+ */
+cvmx_helper_link_info_t cvmx_helper_link_autoconf(int ipd_port)
+{
+ cvmx_helper_link_info_t link_info;
+ int interface = cvmx_helper_get_interface_num(ipd_port);
+ int index = cvmx_helper_get_interface_index_num(ipd_port);
+
+ if (index >= cvmx_helper_ports_on_interface(interface))
+ {
+ link_info.u64 = 0;
+ return link_info;
+ }
+
+ link_info = cvmx_helper_link_get(ipd_port);
+ if (link_info.u64 == (__cvmx_helper_get_link_info(interface, index)).u64)
+ return link_info;
+
+#if !defined(CVMX_BUILD_FOR_LINUX_KERNEL) && !defined(CVMX_BUILD_FOR_FREEBSD_KERNEL)
+ if (!link_info.s.link_up)
+ cvmx_error_disable_group(CVMX_ERROR_GROUP_ETHERNET, ipd_port);
+#endif
+
+ /* If we fail to set the link speed, port_link_info will not change */
+ cvmx_helper_link_set(ipd_port, link_info);
+
+#if !defined(CVMX_BUILD_FOR_LINUX_KERNEL) && !defined(CVMX_BUILD_FOR_FREEBSD_KERNEL)
+ if (link_info.s.link_up)
+ cvmx_error_enable_group(CVMX_ERROR_GROUP_ETHERNET, ipd_port);
+#endif
+
+ return link_info;
+}
+#ifdef CVMX_BUILD_FOR_LINUX_KERNEL
+EXPORT_SYMBOL(cvmx_helper_link_autoconf);
+#endif
+
+/**
+ * Return the link state of an IPD/PKO port as returned by
+ * auto negotiation. The result of this function may not match
+ * Octeon's link config if auto negotiation has changed since
+ * the last call to cvmx_helper_link_set().
+ *
+ * @param ipd_port IPD/PKO port to query
+ *
+ * @return Link state
+ */
+cvmx_helper_link_info_t cvmx_helper_link_get(int ipd_port)
+{
+ cvmx_helper_link_info_t result;
+ int interface = cvmx_helper_get_interface_num(ipd_port);
+ int index = cvmx_helper_get_interface_index_num(ipd_port);
+
+ /* The default result will be a down link unless the code below
+ changes it */
+ result.u64 = 0;
+
+ if (index >= cvmx_helper_ports_on_interface(interface))
+ return result;
+
+ switch (cvmx_helper_interface_get_mode(interface))
+ {
+ case CVMX_HELPER_INTERFACE_MODE_DISABLED:
+ case CVMX_HELPER_INTERFACE_MODE_PCIE:
+ /* Network links are not supported */
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_XAUI:
+ case CVMX_HELPER_INTERFACE_MODE_RXAUI:
+ result = __cvmx_helper_xaui_link_get(ipd_port);
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_GMII:
+ if (index == 0)
+ result = __cvmx_helper_rgmii_link_get(ipd_port);
+ else
+ {
+ result.s.full_duplex = 1;
+ result.s.link_up = 1;
+ result.s.speed = 1000;
+ }
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_RGMII:
+ result = __cvmx_helper_rgmii_link_get(ipd_port);
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_SPI:
+ result = __cvmx_helper_spi_link_get(ipd_port);
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_SGMII:
+ case CVMX_HELPER_INTERFACE_MODE_PICMG:
+ result = __cvmx_helper_sgmii_link_get(ipd_port);
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_SRIO:
+ result = __cvmx_helper_srio_link_get(ipd_port);
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_ILK:
+ result = __cvmx_helper_ilk_link_get(ipd_port);
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_NPI:
+ case CVMX_HELPER_INTERFACE_MODE_LOOP:
+ /* Network links are not supported */
+ break;
+ }
+ return result;
+}
+#ifdef CVMX_BUILD_FOR_LINUX_KERNEL
+EXPORT_SYMBOL(cvmx_helper_link_get);
+#endif
+
+
+/**
+ * Configure an IPD/PKO port for the specified link state. This
+ * function does not influence auto negotiation at the PHY level.
+ * The passed link state must always match the link state returned
+ * by cvmx_helper_link_get(). It is normally best to use
+ * cvmx_helper_link_autoconf() instead.
+ *
+ * @param ipd_port IPD/PKO port to configure
+ * @param link_info The new link state
+ *
+ * @return Zero on success, negative on failure
+ */
+int cvmx_helper_link_set(int ipd_port, cvmx_helper_link_info_t link_info)
+{
+ int result = -1;
+ int interface = cvmx_helper_get_interface_num(ipd_port);
+ int index = cvmx_helper_get_interface_index_num(ipd_port);
+
+ if (index >= cvmx_helper_ports_on_interface(interface))
+ return -1;
+
+ switch (cvmx_helper_interface_get_mode(interface))
+ {
+ case CVMX_HELPER_INTERFACE_MODE_DISABLED:
+ case CVMX_HELPER_INTERFACE_MODE_PCIE:
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_XAUI:
+ case CVMX_HELPER_INTERFACE_MODE_RXAUI:
+ result = __cvmx_helper_xaui_link_set(ipd_port, link_info);
+ break;
+ /* RGMII/GMII/MII are all treated about the same. Most functions
+ refer to these ports as RGMII */
+ case CVMX_HELPER_INTERFACE_MODE_RGMII:
+ case CVMX_HELPER_INTERFACE_MODE_GMII:
+ result = __cvmx_helper_rgmii_link_set(ipd_port, link_info);
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_SPI:
+ result = __cvmx_helper_spi_link_set(ipd_port, link_info);
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_SGMII:
+ case CVMX_HELPER_INTERFACE_MODE_PICMG:
+ result = __cvmx_helper_sgmii_link_set(ipd_port, link_info);
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_SRIO:
+ result = __cvmx_helper_srio_link_set(ipd_port, link_info);
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_ILK:
+ result = __cvmx_helper_ilk_link_set(ipd_port, link_info);
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_NPI:
+ case CVMX_HELPER_INTERFACE_MODE_LOOP:
+ break;
+ }
+ /* Set the port_link_info here so that the link status is updated
+ no matter how cvmx_helper_link_set is called. We don't change
+ the value if link_set failed */
+ if (result == 0)
+ __cvmx_helper_set_link_info(interface, index, link_info);
+ return result;
+}
+#ifdef CVMX_BUILD_FOR_LINUX_KERNEL
+EXPORT_SYMBOL(cvmx_helper_link_set);
+#endif
+
+
+/**
+ * Configure a port for internal and/or external loopback. Internal loopback
+ * causes packets sent by the port to be received by Octeon. External loopback
+ * causes packets received from the wire to sent out again.
+ *
+ * @param ipd_port IPD/PKO port to loopback.
+ * @param enable_internal
+ * Non zero if you want internal loopback
+ * @param enable_external
+ * Non zero if you want external loopback
+ *
+ * @return Zero on success, negative on failure.
+ */
+int cvmx_helper_configure_loopback(int ipd_port, int enable_internal, int enable_external)
+{
+ int result = -1;
+ int interface = cvmx_helper_get_interface_num(ipd_port);
+ int index = cvmx_helper_get_interface_index_num(ipd_port);
+
+ if (index >= cvmx_helper_ports_on_interface(interface))
+ return -1;
+
+ switch (cvmx_helper_interface_get_mode(interface))
+ {
+ case CVMX_HELPER_INTERFACE_MODE_DISABLED:
+ case CVMX_HELPER_INTERFACE_MODE_PCIE:
+ case CVMX_HELPER_INTERFACE_MODE_SRIO:
+ case CVMX_HELPER_INTERFACE_MODE_ILK:
+ case CVMX_HELPER_INTERFACE_MODE_SPI:
+ case CVMX_HELPER_INTERFACE_MODE_NPI:
+ case CVMX_HELPER_INTERFACE_MODE_LOOP:
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_XAUI:
+ case CVMX_HELPER_INTERFACE_MODE_RXAUI:
+ result = __cvmx_helper_xaui_configure_loopback(ipd_port, enable_internal, enable_external);
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_RGMII:
+ case CVMX_HELPER_INTERFACE_MODE_GMII:
+ result = __cvmx_helper_rgmii_configure_loopback(ipd_port, enable_internal, enable_external);
+ break;
+ case CVMX_HELPER_INTERFACE_MODE_SGMII:
+ case CVMX_HELPER_INTERFACE_MODE_PICMG:
+ result = __cvmx_helper_sgmii_configure_loopback(ipd_port, enable_internal, enable_external);
+ break;
+ }
+ return result;
+}
+
+#endif /* CVMX_ENABLE_PKO_FUNCTIONS */
OpenPOWER on IntegriCloud