diff options
Diffstat (limited to 'sys/contrib/octeon-sdk/cvmx-helper.c')
-rw-r--r-- | sys/contrib/octeon-sdk/cvmx-helper.c | 1999 |
1 files changed, 1999 insertions, 0 deletions
diff --git a/sys/contrib/octeon-sdk/cvmx-helper.c b/sys/contrib/octeon-sdk/cvmx-helper.c new file mode 100644 index 0000000..26582f0 --- /dev/null +++ b/sys/contrib/octeon-sdk/cvmx-helper.c @@ -0,0 +1,1999 @@ +/***********************license start*************** + * Copyright (c) 2003-2010 Cavium Inc. (support@cavium.com). All rights + * reserved. + * + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions are + * met: + * + * * Redistributions of source code must retain the above copyright + * notice, this list of conditions and the following disclaimer. + * + * * Redistributions in binary form must reproduce the above + * copyright notice, this list of conditions and the following + * disclaimer in the documentation and/or other materials provided + * with the distribution. + + * * Neither the name of Cavium Inc. nor the names of + * its contributors may be used to endorse or promote products + * derived from this software without specific prior written + * permission. + + * This Software, including technical data, may be subject to U.S. export control + * laws, including the U.S. Export Administration Act and its associated + * regulations, and may be subject to export or import regulations in other + * countries. + + * TO THE MAXIMUM EXTENT PERMITTED BY LAW, THE SOFTWARE IS PROVIDED "AS IS" + * AND WITH ALL FAULTS AND CAVIUM INC. MAKES NO PROMISES, REPRESENTATIONS OR + * WARRANTIES, EITHER EXPRESS, IMPLIED, STATUTORY, OR OTHERWISE, WITH RESPECT TO + * THE SOFTWARE, INCLUDING ITS CONDITION, ITS CONFORMITY TO ANY REPRESENTATION OR + * DESCRIPTION, OR THE EXISTENCE OF ANY LATENT OR PATENT DEFECTS, AND CAVIUM + * SPECIFICALLY DISCLAIMS ALL IMPLIED (IF ANY) WARRANTIES OF TITLE, + * MERCHANTABILITY, NONINFRINGEMENT, FITNESS FOR A PARTICULAR PURPOSE, LACK OF + * VIRUSES, ACCURACY OR COMPLETENESS, QUIET ENJOYMENT, QUIET POSSESSION OR + * CORRESPONDENCE TO DESCRIPTION. THE ENTIRE RISK ARISING OUT OF USE OR + * PERFORMANCE OF THE SOFTWARE LIES WITH YOU. + ***********************license end**************************************/ + + + + + + + +/** + * @file + * + * Helper functions for common, but complicated tasks. + * + * <hr>$Revision: 70030 $<hr> + */ +#ifdef CVMX_BUILD_FOR_LINUX_KERNEL +#include <linux/module.h> +#include <asm/octeon/cvmx.h> +#include <asm/octeon/cvmx-config.h> +#include <asm/octeon/cvmx-bootmem.h> +#include <asm/octeon/cvmx-sriox-defs.h> +#include <asm/octeon/cvmx-npi-defs.h> +#include <asm/octeon/cvmx-mio-defs.h> +#include <asm/octeon/cvmx-pexp-defs.h> +#include <asm/octeon/cvmx-pip-defs.h> +#include <asm/octeon/cvmx-asxx-defs.h> +#include <asm/octeon/cvmx-gmxx-defs.h> +#include <asm/octeon/cvmx-smix-defs.h> +#include <asm/octeon/cvmx-dbg-defs.h> +#include <asm/octeon/cvmx-sso-defs.h> + +#include <asm/octeon/cvmx-gmx.h> +#include <asm/octeon/cvmx-fpa.h> +#include <asm/octeon/cvmx-pip.h> +#include <asm/octeon/cvmx-pko.h> +#include <asm/octeon/cvmx-ipd.h> +#include <asm/octeon/cvmx-spi.h> +#include <asm/octeon/cvmx-clock.h> +#include <asm/octeon/cvmx-helper.h> +#include <asm/octeon/cvmx-helper-board.h> +#include <asm/octeon/cvmx-helper-errata.h> +#include <asm/octeon/cvmx-helper-cfg.h> +#else +#if !defined(__FreeBSD__) || !defined(_KERNEL) +#include "executive-config.h" +#endif +#include "cvmx.h" +#include "cvmx-sysinfo.h" +#include "cvmx-bootmem.h" +#include "cvmx-version.h" +#include "cvmx-helper-check-defines.h" +#include "cvmx-gmx.h" +#if !defined(__FreeBSD__) || !defined(_KERNEL) +#include "cvmx-error.h" +#include "cvmx-config.h" +#endif + +#include "cvmx-fpa.h" +#include "cvmx-pip.h" +#include "cvmx-pko.h" +#include "cvmx-ipd.h" +#include "cvmx-spi.h" +#include "cvmx-helper.h" +#include "cvmx-helper-board.h" +#include "cvmx-helper-errata.h" +#include "cvmx-helper-cfg.h" +#endif + + +#ifdef CVMX_ENABLE_PKO_FUNCTIONS + +/** + * cvmx_override_pko_queue_priority(int pko_port, uint64_t + * priorities[16]) is a function pointer. It is meant to allow + * customization of the PKO queue priorities based on the port + * number. Users should set this pointer to a function before + * calling any cvmx-helper operations. + */ +CVMX_SHARED void (*cvmx_override_pko_queue_priority)(int ipd_port, + uint64_t *priorities) = NULL; +#ifdef CVMX_BUILD_FOR_LINUX_KERNEL +EXPORT_SYMBOL(cvmx_override_pko_queue_priority); +#endif + +/** + * cvmx_override_ipd_port_setup(int ipd_port) is a function + * pointer. It is meant to allow customization of the IPD + * port/port kind setup before packet input/output comes online. + * It is called after cvmx-helper does the default IPD configuration, + * but before IPD is enabled. Users should set this pointer to a + * function before calling any cvmx-helper operations. + */ +CVMX_SHARED void (*cvmx_override_ipd_port_setup)(int ipd_port) = NULL; + +/** + * Return the number of interfaces the chip has. Each interface + * may have multiple ports. Most chips support two interfaces, + * but the CNX0XX and CNX1XX are exceptions. These only support + * one interface. + * + * @return Number of interfaces on chip + */ +int cvmx_helper_get_number_of_interfaces(void) +{ + switch (cvmx_sysinfo_get()->board_type) { +#if defined(OCTEON_VENDOR_LANNER) + case CVMX_BOARD_TYPE_CUST_LANNER_MR955: + return 2; + case CVMX_BOARD_TYPE_CUST_LANNER_MR730: + return 1; +#endif +#if defined(OCTEON_VENDOR_RADISYS) + case CVMX_BOARD_TYPE_CUST_RADISYS_RSYS4GBE: + return 2; +#endif + default: + break; + } + + if (OCTEON_IS_MODEL(OCTEON_CN68XX)) + return 9; + else if (OCTEON_IS_MODEL(OCTEON_CN66XX)) + if (OCTEON_IS_MODEL(OCTEON_CN66XX_PASS1_0)) + return 7; + else + return 8; + else if (OCTEON_IS_MODEL(OCTEON_CN63XX)) + return 6; + else if (OCTEON_IS_MODEL(OCTEON_CN56XX) || OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN61XX)) + return 4; + else + return 3; +} +#ifdef CVMX_BUILD_FOR_LINUX_KERNEL +EXPORT_SYMBOL(cvmx_helper_get_number_of_interfaces); +#endif + + +/** + * Return the number of ports on an interface. Depending on the + * chip and configuration, this can be 1-16. A value of 0 + * specifies that the interface doesn't exist or isn't usable. + * + * @param interface Interface to get the port count for + * + * @return Number of ports on interface. Can be Zero. + */ +int cvmx_helper_ports_on_interface(int interface) +{ + if (octeon_has_feature(OCTEON_FEATURE_PKND)) + return cvmx_helper_interface_enumerate(interface); + else + return __cvmx_helper_get_num_ipd_ports(interface); +} +#ifdef CVMX_BUILD_FOR_LINUX_KERNEL +EXPORT_SYMBOL(cvmx_helper_ports_on_interface); +#endif + + +/** + * Get the operating mode of an interface. Depending on the Octeon + * chip and configuration, this function returns an enumeration + * of the type of packet I/O supported by an interface. + * + * @param interface Interface to probe + * + * @return Mode of the interface. Unknown or unsupported interfaces return + * DISABLED. + */ +cvmx_helper_interface_mode_t cvmx_helper_interface_get_mode(int interface) +{ + cvmx_gmxx_inf_mode_t mode; + + if (OCTEON_IS_MODEL(OCTEON_CN68XX)) + { + cvmx_mio_qlmx_cfg_t qlm_cfg; + switch(interface) + { + case 0: + qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(0)); + /* QLM is disabled when QLM SPD is 15. */ + if (qlm_cfg.s.qlm_spd == 15) + return CVMX_HELPER_INTERFACE_MODE_DISABLED; + + if (qlm_cfg.s.qlm_cfg == 7) + return CVMX_HELPER_INTERFACE_MODE_RXAUI; + else if (qlm_cfg.s.qlm_cfg == 2) + return CVMX_HELPER_INTERFACE_MODE_SGMII; + else if (qlm_cfg.s.qlm_cfg == 3) + return CVMX_HELPER_INTERFACE_MODE_XAUI; + else + return CVMX_HELPER_INTERFACE_MODE_DISABLED; + break; + case 1: + qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(0)); + /* QLM is disabled when QLM SPD is 15. */ + if (qlm_cfg.s.qlm_spd == 15) + return CVMX_HELPER_INTERFACE_MODE_DISABLED; + + if (qlm_cfg.s.qlm_cfg == 7) + return CVMX_HELPER_INTERFACE_MODE_RXAUI; + else + return CVMX_HELPER_INTERFACE_MODE_DISABLED; + break; + case 2: + case 3: + case 4: + qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(interface)); + /* QLM is disabled when QLM SPD is 15. */ + if (qlm_cfg.s.qlm_spd == 15) + return CVMX_HELPER_INTERFACE_MODE_DISABLED; + + if (qlm_cfg.s.qlm_cfg == 2) + return CVMX_HELPER_INTERFACE_MODE_SGMII; + else if (qlm_cfg.s.qlm_cfg == 3) + return CVMX_HELPER_INTERFACE_MODE_XAUI; + else + return CVMX_HELPER_INTERFACE_MODE_DISABLED; + break; + case 5: + case 6: + qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(interface - 4)); + /* QLM is disabled when QLM SPD is 15. */ + if (qlm_cfg.s.qlm_spd == 15) + return CVMX_HELPER_INTERFACE_MODE_DISABLED; + + if (qlm_cfg.s.qlm_cfg == 1) + { + return CVMX_HELPER_INTERFACE_MODE_ILK; + } + else + return CVMX_HELPER_INTERFACE_MODE_DISABLED; + break; + case 7: + qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(3)); + /* QLM is disabled when QLM SPD is 15. */ + if (qlm_cfg.s.qlm_spd == 15) + return CVMX_HELPER_INTERFACE_MODE_DISABLED; + else if (qlm_cfg.s.qlm_cfg != 0) + { + qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(1)); + if (qlm_cfg.s.qlm_cfg != 0) + return CVMX_HELPER_INTERFACE_MODE_DISABLED; + } + return CVMX_HELPER_INTERFACE_MODE_NPI; + break; + case 8: + return CVMX_HELPER_INTERFACE_MODE_LOOP; + break; + default: + return CVMX_HELPER_INTERFACE_MODE_DISABLED; + break; + } + } + + if (interface == 2) + return CVMX_HELPER_INTERFACE_MODE_NPI; + + if (interface == 3) + { + if (OCTEON_IS_MODEL(OCTEON_CN56XX) + || OCTEON_IS_MODEL(OCTEON_CN52XX) + || OCTEON_IS_MODEL(OCTEON_CN6XXX) + || OCTEON_IS_MODEL(OCTEON_CNF7XXX)) + return CVMX_HELPER_INTERFACE_MODE_LOOP; + else + return CVMX_HELPER_INTERFACE_MODE_DISABLED; + } + + /* Only present in CN63XX & CN66XX Octeon model */ + if ((OCTEON_IS_MODEL(OCTEON_CN63XX) && (interface == 4 || interface == 5)) + || (OCTEON_IS_MODEL(OCTEON_CN66XX) && interface >= 4 && interface <= 7)) + { + cvmx_sriox_status_reg_t sriox_status_reg; + + /* cn66xx pass1.0 has only 2 SRIO interfaces. */ + if ((interface == 5 || interface == 7) && OCTEON_IS_MODEL(OCTEON_CN66XX_PASS1_0)) + return CVMX_HELPER_INTERFACE_MODE_DISABLED; + + sriox_status_reg.u64 = cvmx_read_csr(CVMX_SRIOX_STATUS_REG(interface-4)); + if (sriox_status_reg.s.srio) + return CVMX_HELPER_INTERFACE_MODE_SRIO; + else + return CVMX_HELPER_INTERFACE_MODE_DISABLED; + } + + /* Interface 5 always disabled in CN66XX */ + if (OCTEON_IS_MODEL(OCTEON_CN66XX)) + { + cvmx_mio_qlmx_cfg_t mio_qlm_cfg; + + /* QLM2 is SGMII0 and QLM1 is SGMII1 */ + if (interface == 0) + mio_qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(2)); + else if (interface == 1) + mio_qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(1)); + else + return CVMX_HELPER_INTERFACE_MODE_DISABLED; + + if (mio_qlm_cfg.s.qlm_spd == 15) + return CVMX_HELPER_INTERFACE_MODE_DISABLED; + + if (mio_qlm_cfg.s.qlm_cfg == 9) + return CVMX_HELPER_INTERFACE_MODE_SGMII; + else if (mio_qlm_cfg.s.qlm_cfg == 11) + return CVMX_HELPER_INTERFACE_MODE_XAUI; + else + return CVMX_HELPER_INTERFACE_MODE_DISABLED; + } + else if (OCTEON_IS_MODEL(OCTEON_CN61XX)) + { + cvmx_mio_qlmx_cfg_t qlm_cfg; + + if (interface == 0) + { + qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(2)); + if (qlm_cfg.s.qlm_cfg == 2) + return CVMX_HELPER_INTERFACE_MODE_SGMII; + else if (qlm_cfg.s.qlm_cfg == 3) + return CVMX_HELPER_INTERFACE_MODE_XAUI; + else + return CVMX_HELPER_INTERFACE_MODE_DISABLED; + } + else if (interface == 1) + { + /* If QLM 1 is PEV0/PEM1 mode, them QLM0 cannot be SGMII/XAUI */ + qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(1)); + if (qlm_cfg.s.qlm_cfg == 1) + return CVMX_HELPER_INTERFACE_MODE_DISABLED; + + qlm_cfg.u64 = cvmx_read_csr(CVMX_MIO_QLMX_CFG(0)); + if (qlm_cfg.s.qlm_cfg == 2) + return CVMX_HELPER_INTERFACE_MODE_SGMII; + else if (qlm_cfg.s.qlm_cfg == 3) + return CVMX_HELPER_INTERFACE_MODE_XAUI; + else + return CVMX_HELPER_INTERFACE_MODE_DISABLED; + } + } + + if (interface == 0 && cvmx_sysinfo_get()->board_type == CVMX_BOARD_TYPE_CN3005_EVB_HS5 && cvmx_sysinfo_get()->board_rev_major == 1) + { + /* Lie about interface type of CN3005 board. This board has a switch on port 1 like + ** the other evaluation boards, but it is connected over RGMII instead of GMII. Report + ** GMII mode so that the speed is forced to 1 Gbit full duplex. Other than some initial configuration + ** (which does not use the output of this function) there is no difference in setup between GMII and RGMII modes. + */ + return CVMX_HELPER_INTERFACE_MODE_GMII; + } + + /* Interface 1 is always disabled on CN31XX and CN30XX */ + if ((interface == 1) + && (OCTEON_IS_MODEL(OCTEON_CN31XX) + || OCTEON_IS_MODEL(OCTEON_CN30XX) + || OCTEON_IS_MODEL(OCTEON_CN50XX) + || OCTEON_IS_MODEL(OCTEON_CN52XX) + || OCTEON_IS_MODEL(OCTEON_CN63XX) + || OCTEON_IS_MODEL(OCTEON_CNF71XX))) + return CVMX_HELPER_INTERFACE_MODE_DISABLED; + + mode.u64 = cvmx_read_csr(CVMX_GMXX_INF_MODE(interface)); + + if (OCTEON_IS_MODEL(OCTEON_CN56XX) || OCTEON_IS_MODEL(OCTEON_CN52XX)) + { + switch(mode.cn56xx.mode) + { + case 0: return CVMX_HELPER_INTERFACE_MODE_DISABLED; + case 1: return CVMX_HELPER_INTERFACE_MODE_XAUI; + case 2: return CVMX_HELPER_INTERFACE_MODE_SGMII; + case 3: return CVMX_HELPER_INTERFACE_MODE_PICMG; + default:return CVMX_HELPER_INTERFACE_MODE_DISABLED; + } + } + else if (OCTEON_IS_MODEL(OCTEON_CN63XX) || OCTEON_IS_MODEL(OCTEON_CNF71XX)) + { + switch(mode.cn63xx.mode) + { + case 0: return CVMX_HELPER_INTERFACE_MODE_SGMII; + case 1: return CVMX_HELPER_INTERFACE_MODE_XAUI; + default: return CVMX_HELPER_INTERFACE_MODE_DISABLED; + } + } + else + { + if (!mode.s.en) + return CVMX_HELPER_INTERFACE_MODE_DISABLED; + + if (mode.s.type) + { + if (OCTEON_IS_MODEL(OCTEON_CN38XX) || OCTEON_IS_MODEL(OCTEON_CN58XX)) + return CVMX_HELPER_INTERFACE_MODE_SPI; + else + return CVMX_HELPER_INTERFACE_MODE_GMII; + } + else + return CVMX_HELPER_INTERFACE_MODE_RGMII; + } +} +#ifdef CVMX_BUILD_FOR_LINUX_KERNEL +EXPORT_SYMBOL(cvmx_helper_interface_get_mode); +#endif + +/** + * @INTERNAL + * Configure the IPD/PIP tagging and QoS options for a specific + * port. This function determines the POW work queue entry + * contents for a port. The setup performed here is controlled by + * the defines in executive-config.h. + * + * @param ipd_port Port/Port kind to configure. This follows the IPD numbering, + * not the per interface numbering + * + * @return Zero on success, negative on failure + */ +static int __cvmx_helper_port_setup_ipd(int ipd_port) +{ + cvmx_pip_prt_cfgx_t port_config; + cvmx_pip_prt_tagx_t tag_config; + + if (octeon_has_feature(OCTEON_FEATURE_PKND)) + { + int interface, index, pknd; + cvmx_pip_prt_cfgbx_t prt_cfgbx; + + interface = cvmx_helper_get_interface_num(ipd_port); + index = cvmx_helper_get_interface_index_num(ipd_port); + pknd = cvmx_helper_get_pknd(interface, index); + + port_config.u64 = cvmx_read_csr(CVMX_PIP_PRT_CFGX(pknd)); + tag_config.u64 = cvmx_read_csr(CVMX_PIP_PRT_TAGX(pknd)); + + port_config.s.qos = pknd & 0x7; + + /* Default BPID to use for packets on this port-kind */ + prt_cfgbx.u64 = cvmx_read_csr(CVMX_PIP_PRT_CFGBX(pknd)); + prt_cfgbx.s.bpid = pknd; + cvmx_write_csr(CVMX_PIP_PRT_CFGBX(pknd), prt_cfgbx.u64); + } + else + { + port_config.u64 = cvmx_read_csr(CVMX_PIP_PRT_CFGX(ipd_port)); + tag_config.u64 = cvmx_read_csr(CVMX_PIP_PRT_TAGX(ipd_port)); + + /* Have each port go to a different POW queue */ + port_config.s.qos = ipd_port & 0x7; + } + + /* Process the headers and place the IP header in the work queue */ + port_config.s.mode = CVMX_HELPER_INPUT_PORT_SKIP_MODE; + + tag_config.s.ip6_src_flag = CVMX_HELPER_INPUT_TAG_IPV6_SRC_IP; + tag_config.s.ip6_dst_flag = CVMX_HELPER_INPUT_TAG_IPV6_DST_IP; + tag_config.s.ip6_sprt_flag = CVMX_HELPER_INPUT_TAG_IPV6_SRC_PORT; + tag_config.s.ip6_dprt_flag = CVMX_HELPER_INPUT_TAG_IPV6_DST_PORT; + tag_config.s.ip6_nxth_flag = CVMX_HELPER_INPUT_TAG_IPV6_NEXT_HEADER; + tag_config.s.ip4_src_flag = CVMX_HELPER_INPUT_TAG_IPV4_SRC_IP; + tag_config.s.ip4_dst_flag = CVMX_HELPER_INPUT_TAG_IPV4_DST_IP; + tag_config.s.ip4_sprt_flag = CVMX_HELPER_INPUT_TAG_IPV4_SRC_PORT; + tag_config.s.ip4_dprt_flag = CVMX_HELPER_INPUT_TAG_IPV4_DST_PORT; + tag_config.s.ip4_pctl_flag = CVMX_HELPER_INPUT_TAG_IPV4_PROTOCOL; + tag_config.s.inc_prt_flag = CVMX_HELPER_INPUT_TAG_INPUT_PORT; + tag_config.s.tcp6_tag_type = CVMX_HELPER_INPUT_TAG_TYPE; + tag_config.s.tcp4_tag_type = CVMX_HELPER_INPUT_TAG_TYPE; + tag_config.s.ip6_tag_type = CVMX_HELPER_INPUT_TAG_TYPE; + tag_config.s.ip4_tag_type = CVMX_HELPER_INPUT_TAG_TYPE; + tag_config.s.non_tag_type = CVMX_HELPER_INPUT_TAG_TYPE; + /* Put all packets in group 0. Other groups can be used by the app */ + tag_config.s.grp = 0; + + cvmx_pip_config_port(ipd_port, port_config, tag_config); + + /* Give the user a chance to override our setting for each port */ + if (cvmx_override_ipd_port_setup) + cvmx_override_ipd_port_setup(ipd_port); + + return 0; +} + +/** + * Enable or disable FCS stripping for all the ports on an interface. + * + * @param interface + * @param nports number of ports + * @param has_fcs 0 for disable and !0 for enable + */ +static int cvmx_helper_fcs_op(int interface, int nports, int has_fcs) +{ + uint64_t port_bit; + int index; + int pknd; + cvmx_pip_sub_pkind_fcsx_t pkind_fcsx; + cvmx_pip_prt_cfgx_t port_cfg; + + if (!octeon_has_feature(OCTEON_FEATURE_PKND)) + return 0; + + port_bit = 0; + for (index = 0; index < nports; index++) + port_bit |= ((uint64_t)1 << cvmx_helper_get_pknd(interface, index)); + + pkind_fcsx.u64 = cvmx_read_csr(CVMX_PIP_SUB_PKIND_FCSX(0)); + if (has_fcs) + pkind_fcsx.s.port_bit |= port_bit; + else + pkind_fcsx.s.port_bit &= ~port_bit; + cvmx_write_csr(CVMX_PIP_SUB_PKIND_FCSX(0), pkind_fcsx.u64); + + for (pknd = 0; pknd < 64; pknd++) + { + if ((1ull << pknd) & port_bit) + { + port_cfg.u64 = cvmx_read_csr(CVMX_PIP_PRT_CFGX(pknd)); + port_cfg.s.crc_en = (has_fcs) ? 1 : 0; + cvmx_write_csr(CVMX_PIP_PRT_CFGX(pknd), port_cfg.u64); + } + } + + return 0; +} + +/** + * Determine the actual number of hardware ports connected to an + * interface. It doesn't setup the ports or enable them. + * + * @param interface Interface to enumerate + * + * @return The number of ports on the interface, negative on failure + */ +int cvmx_helper_interface_enumerate(int interface) +{ + switch (cvmx_helper_interface_get_mode(interface)) { + /* XAUI is a single high speed port */ + case CVMX_HELPER_INTERFACE_MODE_XAUI: + case CVMX_HELPER_INTERFACE_MODE_RXAUI: + return __cvmx_helper_xaui_enumerate(interface); + /* RGMII/GMII/MII are all treated about the same. Most functions + refer to these ports as RGMII */ + case CVMX_HELPER_INTERFACE_MODE_RGMII: + case CVMX_HELPER_INTERFACE_MODE_GMII: + return __cvmx_helper_rgmii_enumerate(interface); + /* SPI4 can have 1-16 ports depending on the device at the other end */ + case CVMX_HELPER_INTERFACE_MODE_SPI: + return __cvmx_helper_spi_enumerate(interface); + /* SGMII can have 1-4 ports depending on how many are hooked up */ + case CVMX_HELPER_INTERFACE_MODE_SGMII: + case CVMX_HELPER_INTERFACE_MODE_PICMG: + return __cvmx_helper_sgmii_enumerate(interface); + /* PCI target Network Packet Interface */ + case CVMX_HELPER_INTERFACE_MODE_NPI: + return __cvmx_helper_npi_enumerate(interface); + /* Special loopback only ports. These are not the same + * as other ports in loopback mode */ + case CVMX_HELPER_INTERFACE_MODE_LOOP: + return __cvmx_helper_loop_enumerate(interface); + /* SRIO has 2^N ports, where N is number of interfaces */ + case CVMX_HELPER_INTERFACE_MODE_SRIO: + return __cvmx_helper_srio_enumerate(interface); + + case CVMX_HELPER_INTERFACE_MODE_ILK: + return __cvmx_helper_ilk_enumerate(interface); + /* These types don't support ports to IPD/PKO */ + case CVMX_HELPER_INTERFACE_MODE_DISABLED: + case CVMX_HELPER_INTERFACE_MODE_PCIE: + default: + return 0; + } +} + +/** + * This function probes an interface to determine the actual number of + * hardware ports connected to it. It does some setup the ports but + * doesn't enable them. The main goal here is to set the global + * interface_port_count[interface] correctly. Final hardware setup of + * the ports will be performed later. + * + * @param interface Interface to probe + * + * @return Zero on success, negative on failure + */ +int cvmx_helper_interface_probe(int interface) +{ + /* At this stage in the game we don't want packets to be moving yet. + The following probe calls should perform hardware setup + needed to determine port counts. Receive must still be disabled */ + int nports; + int has_fcs; + enum cvmx_pko_padding padding = CVMX_PKO_PADDING_NONE; + + nports = -1; + has_fcs = 0; + switch (cvmx_helper_interface_get_mode(interface)) + { + /* These types don't support ports to IPD/PKO */ + case CVMX_HELPER_INTERFACE_MODE_DISABLED: + case CVMX_HELPER_INTERFACE_MODE_PCIE: + nports = 0; + break; + /* XAUI is a single high speed port */ + case CVMX_HELPER_INTERFACE_MODE_XAUI: + case CVMX_HELPER_INTERFACE_MODE_RXAUI: + nports = __cvmx_helper_xaui_probe(interface); + has_fcs = 1; + padding = CVMX_PKO_PADDING_60; + break; + /* RGMII/GMII/MII are all treated about the same. Most functions + refer to these ports as RGMII */ + case CVMX_HELPER_INTERFACE_MODE_RGMII: + case CVMX_HELPER_INTERFACE_MODE_GMII: + nports = __cvmx_helper_rgmii_probe(interface); + padding = CVMX_PKO_PADDING_60; + break; + /* SPI4 can have 1-16 ports depending on the device at the other end */ + case CVMX_HELPER_INTERFACE_MODE_SPI: + nports = __cvmx_helper_spi_probe(interface); + padding = CVMX_PKO_PADDING_60; + break; + /* SGMII can have 1-4 ports depending on how many are hooked up */ + case CVMX_HELPER_INTERFACE_MODE_SGMII: + padding = CVMX_PKO_PADDING_60; + case CVMX_HELPER_INTERFACE_MODE_PICMG: + nports = __cvmx_helper_sgmii_probe(interface); + has_fcs = 1; + break; + /* PCI target Network Packet Interface */ + case CVMX_HELPER_INTERFACE_MODE_NPI: + nports = __cvmx_helper_npi_probe(interface); + break; + /* Special loopback only ports. These are not the same as other ports + in loopback mode */ + case CVMX_HELPER_INTERFACE_MODE_LOOP: + nports = __cvmx_helper_loop_probe(interface); + break; + /* SRIO has 2^N ports, where N is number of interfaces */ + case CVMX_HELPER_INTERFACE_MODE_SRIO: + nports = __cvmx_helper_srio_probe(interface); + break; + case CVMX_HELPER_INTERFACE_MODE_ILK: + nports = __cvmx_helper_ilk_probe(interface); + has_fcs = 1; + padding = CVMX_PKO_PADDING_60; + break; + } + + if (nports == -1) + return -1; + + if (!octeon_has_feature(OCTEON_FEATURE_PKND)) + has_fcs = 0; + + nports = __cvmx_helper_board_interface_probe(interface, nports); + __cvmx_helper_init_interface(interface, nports, has_fcs, padding); + cvmx_helper_fcs_op(interface, nports, has_fcs); + + /* Make sure all global variables propagate to other cores */ + CVMX_SYNCWS; + + return 0; +} + + +/** + * @INTERNAL + * Setup the IPD/PIP for the ports on an interface. Packet + * classification and tagging are set for every port on the + * interface. The number of ports on the interface must already + * have been probed. + * + * @param interface Interface to setup IPD/PIP for + * + * @return Zero on success, negative on failure + */ +static int __cvmx_helper_interface_setup_ipd(int interface) +{ + + cvmx_helper_interface_mode_t mode; + int ipd_port = cvmx_helper_get_ipd_port(interface, 0); + int num_ports = cvmx_helper_ports_on_interface(interface); + int delta; + + if (num_ports == CVMX_HELPER_CFG_INVALID_VALUE) + return 0; + + mode = cvmx_helper_interface_get_mode(interface); + + if (mode == CVMX_HELPER_INTERFACE_MODE_LOOP) + __cvmx_helper_loop_enable(interface); + + delta = 1; + if (octeon_has_feature(OCTEON_FEATURE_PKND)) + { + if (mode == CVMX_HELPER_INTERFACE_MODE_SGMII) + delta = 16; + } + + while (num_ports--) + { + __cvmx_helper_port_setup_ipd(ipd_port); + ipd_port += delta; + } + + return 0; +} + + +/** + * @INTERNAL + * Setup global setting for IPD/PIP not related to a specific + * interface or port. This must be called before IPD is enabled. + * + * @return Zero on success, negative on failure. + */ +static int __cvmx_helper_global_setup_ipd(void) +{ +#ifndef CVMX_HELPER_IPD_DRAM_MODE +#define CVMX_HELPER_IPD_DRAM_MODE CVMX_IPD_OPC_MODE_STT +#endif + /* Setup the global packet input options */ + cvmx_ipd_config(CVMX_FPA_PACKET_POOL_SIZE/8, + CVMX_HELPER_FIRST_MBUFF_SKIP/8, + CVMX_HELPER_NOT_FIRST_MBUFF_SKIP/8, + (CVMX_HELPER_FIRST_MBUFF_SKIP+8) / 128, /* The +8 is to account for the next ptr */ + (CVMX_HELPER_NOT_FIRST_MBUFF_SKIP+8) / 128, /* The +8 is to account for the next ptr */ + CVMX_FPA_WQE_POOL, + CVMX_HELPER_IPD_DRAM_MODE, + 1); + return 0; +} + + +/** + * @INTERNAL + * Setup the PKO for the ports on an interface. The number of + * queues per port and the priority of each PKO output queue + * is set here. PKO must be disabled when this function is called. + * + * @param interface Interface to setup PKO for + * + * @return Zero on success, negative on failure + */ +static int __cvmx_helper_interface_setup_pko(int interface) +{ + /* Each packet output queue has an associated priority. The higher the + priority, the more often it can send a packet. A priority of 8 means + it can send in all 8 rounds of contention. We're going to make each + queue one less than the last. + The vector of priorities has been extended to support CN5xxx CPUs, + where up to 16 queues can be associated to a port. + To keep backward compatibility we don't change the initial 8 + priorities and replicate them in the second half. + With per-core PKO queues (PKO lockless operation) all queues have + the same priority. */ + /* uint64_t priorities[16] = {8,7,6,5,4,3,2,1,8,7,6,5,4,3,2,1}; */ + uint64_t priorities[16] = {[0 ... 15] = 8}; + + /* Setup the IPD/PIP and PKO for the ports discovered above. Here packet + classification, tagging and output priorities are set */ + int ipd_port = cvmx_helper_get_ipd_port(interface, 0); + int num_ports = cvmx_helper_ports_on_interface(interface); + while (num_ports--) + { + /* Give the user a chance to override the per queue priorities */ + if (cvmx_override_pko_queue_priority) + cvmx_override_pko_queue_priority(ipd_port, priorities); + + cvmx_pko_config_port(ipd_port, cvmx_pko_get_base_queue_per_core(ipd_port, 0), + cvmx_pko_get_num_queues(ipd_port), priorities); + ipd_port++; + } + return 0; +} + + +/** + * @INTERNAL + * Setup global setting for PKO not related to a specific + * interface or port. This must be called before PKO is enabled. + * + * @return Zero on success, negative on failure. + */ +static int __cvmx_helper_global_setup_pko(void) +{ + /* Disable tagwait FAU timeout. This needs to be done before anyone might + start packet output using tags */ + cvmx_iob_fau_timeout_t fau_to; + fau_to.u64 = 0; + fau_to.s.tout_val = 0xfff; + fau_to.s.tout_enb = 0; + cvmx_write_csr(CVMX_IOB_FAU_TIMEOUT, fau_to.u64); + + if (OCTEON_IS_MODEL(OCTEON_CN68XX)) { + cvmx_pko_reg_min_pkt_t min_pkt; + + min_pkt.u64 = 0; + min_pkt.s.size1 = 59; + min_pkt.s.size2 = 59; + min_pkt.s.size3 = 59; + min_pkt.s.size4 = 59; + min_pkt.s.size5 = 59; + min_pkt.s.size6 = 59; + min_pkt.s.size7 = 59; + cvmx_write_csr(CVMX_PKO_REG_MIN_PKT, min_pkt.u64); + } + + return 0; +} + + +/** + * @INTERNAL + * Setup global backpressure setting. + * + * @return Zero on success, negative on failure + */ +static int __cvmx_helper_global_setup_backpressure(void) +{ +#if CVMX_HELPER_DISABLE_RGMII_BACKPRESSURE + /* Disable backpressure if configured to do so */ + /* Disable backpressure (pause frame) generation */ + int num_interfaces = cvmx_helper_get_number_of_interfaces(); + int interface; + for (interface=0; interface<num_interfaces; interface++) + { + switch (cvmx_helper_interface_get_mode(interface)) + { + case CVMX_HELPER_INTERFACE_MODE_DISABLED: + case CVMX_HELPER_INTERFACE_MODE_PCIE: + case CVMX_HELPER_INTERFACE_MODE_SRIO: + case CVMX_HELPER_INTERFACE_MODE_ILK: + case CVMX_HELPER_INTERFACE_MODE_NPI: + case CVMX_HELPER_INTERFACE_MODE_LOOP: + case CVMX_HELPER_INTERFACE_MODE_XAUI: + case CVMX_HELPER_INTERFACE_MODE_RXAUI: + break; + case CVMX_HELPER_INTERFACE_MODE_RGMII: + case CVMX_HELPER_INTERFACE_MODE_GMII: + case CVMX_HELPER_INTERFACE_MODE_SPI: + case CVMX_HELPER_INTERFACE_MODE_SGMII: + case CVMX_HELPER_INTERFACE_MODE_PICMG: + cvmx_gmx_set_backpressure_override(interface, 0xf); + break; + } + } + //cvmx_dprintf("Disabling backpressure\n"); +#endif + + return 0; +} + +/** + * @INTERNAL + * Verify the per port IPD backpressure is aligned properly. + * @return Zero if working, non zero if misaligned + */ +static int __cvmx_helper_backpressure_is_misaligned(void) +{ + uint64_t ipd_int_enb; + cvmx_ipd_ctl_status_t ipd_reg; + uint64_t bp_status0; + uint64_t bp_status1; + const int port0 = 0; + const int port1 = 16; + cvmx_helper_interface_mode_t mode0 = cvmx_helper_interface_get_mode(0); + cvmx_helper_interface_mode_t mode1 = cvmx_helper_interface_get_mode(1); + + /* Disable error interrupts while we check backpressure */ + ipd_int_enb = cvmx_read_csr(CVMX_IPD_INT_ENB); + cvmx_write_csr(CVMX_IPD_INT_ENB, 0); + + /* Enable per port backpressure */ + ipd_reg.u64 = cvmx_read_csr(CVMX_IPD_CTL_STATUS); + ipd_reg.s.pbp_en = 1; + cvmx_write_csr(CVMX_IPD_CTL_STATUS, ipd_reg.u64); + + if (mode0 != CVMX_HELPER_INTERFACE_MODE_DISABLED) + { + /* Enable backpressure for port with a zero threshold */ + cvmx_write_csr(CVMX_IPD_PORTX_BP_PAGE_CNT(port0), 1<<17); + /* Add 1000 to the page count to simulate packets coming in */ + cvmx_write_csr(CVMX_IPD_SUB_PORT_BP_PAGE_CNT, (port0<<25) | 1000); + } + + if (mode1 != CVMX_HELPER_INTERFACE_MODE_DISABLED) + { + /* Enable backpressure for port with a zero threshold */ + cvmx_write_csr(CVMX_IPD_PORTX_BP_PAGE_CNT(port1), 1<<17); + /* Add 1000 to the page count to simulate packets coming in */ + cvmx_write_csr(CVMX_IPD_SUB_PORT_BP_PAGE_CNT, (port1<<25) | 1000); + } + + /* Wait 500 cycles for the BP to update */ + cvmx_wait(500); + + /* Read the BP state from the debug select register */ + switch (mode0) + { + case CVMX_HELPER_INTERFACE_MODE_SPI: + cvmx_write_csr(CVMX_NPI_DBG_SELECT, 0x9004); + bp_status0 = cvmx_read_csr(CVMX_DBG_DATA); + bp_status0 = 0xffff & ~bp_status0; + break; + case CVMX_HELPER_INTERFACE_MODE_RGMII: + case CVMX_HELPER_INTERFACE_MODE_GMII: + cvmx_write_csr(CVMX_NPI_DBG_SELECT, 0x0e00); + bp_status0 = 0xffff & cvmx_read_csr(CVMX_DBG_DATA); + break; + case CVMX_HELPER_INTERFACE_MODE_XAUI: + case CVMX_HELPER_INTERFACE_MODE_SGMII: + case CVMX_HELPER_INTERFACE_MODE_PICMG: + cvmx_write_csr(CVMX_PEXP_NPEI_DBG_SELECT, 0x0e00); + bp_status0 = 0xffff & cvmx_read_csr(CVMX_PEXP_NPEI_DBG_DATA); + break; + default: + bp_status0 = 1<<port0; + break; + } + + /* Read the BP state from the debug select register */ + switch (mode1) + { + case CVMX_HELPER_INTERFACE_MODE_SPI: + cvmx_write_csr(CVMX_NPI_DBG_SELECT, 0x9804); + bp_status1 = cvmx_read_csr(CVMX_DBG_DATA); + bp_status1 = 0xffff & ~bp_status1; + break; + case CVMX_HELPER_INTERFACE_MODE_RGMII: + case CVMX_HELPER_INTERFACE_MODE_GMII: + cvmx_write_csr(CVMX_NPI_DBG_SELECT, 0x1600); + bp_status1 = 0xffff & cvmx_read_csr(CVMX_DBG_DATA); + break; + case CVMX_HELPER_INTERFACE_MODE_XAUI: + case CVMX_HELPER_INTERFACE_MODE_SGMII: + case CVMX_HELPER_INTERFACE_MODE_PICMG: + cvmx_write_csr(CVMX_PEXP_NPEI_DBG_SELECT, 0x1600); + bp_status1 = 0xffff & cvmx_read_csr(CVMX_PEXP_NPEI_DBG_DATA); + break; + default: + bp_status1 = 1<<(port1-16); + break; + } + + if (mode0 != CVMX_HELPER_INTERFACE_MODE_DISABLED) + { + /* Shutdown BP */ + cvmx_write_csr(CVMX_IPD_SUB_PORT_BP_PAGE_CNT, (port0<<25) | (0x1ffffff & -1000)); + cvmx_write_csr(CVMX_IPD_PORTX_BP_PAGE_CNT(port0), 0); + } + + if (mode1 != CVMX_HELPER_INTERFACE_MODE_DISABLED) + { + /* Shutdown BP */ + cvmx_write_csr(CVMX_IPD_SUB_PORT_BP_PAGE_CNT, (port1<<25) | (0x1ffffff & -1000)); + cvmx_write_csr(CVMX_IPD_PORTX_BP_PAGE_CNT(port1), 0); + } + + /* Clear any error interrupts that might have been set */ + cvmx_write_csr(CVMX_IPD_INT_SUM, 0x1f); + cvmx_write_csr(CVMX_IPD_INT_ENB, ipd_int_enb); + + return ((bp_status0 != 1ull<<port0) || (bp_status1 != 1ull<<(port1-16))); +} + + +/** + * @INTERNAL + * Enable packet input/output from the hardware. This function is + * called after all internal setup is complete and IPD is enabled. + * After this function completes, packets will be accepted from the + * hardware ports. PKO should still be disabled to make sure packets + * aren't sent out partially setup hardware. + * + * @param interface Interface to enable + * + * @return Zero on success, negative on failure + */ +static int __cvmx_helper_packet_hardware_enable(int interface) +{ + int result = 0; + switch (cvmx_helper_interface_get_mode(interface)) + { + /* These types don't support ports to IPD/PKO */ + case CVMX_HELPER_INTERFACE_MODE_DISABLED: + case CVMX_HELPER_INTERFACE_MODE_PCIE: + case CVMX_HELPER_INTERFACE_MODE_LOOP: + /* Nothing to do */ + break; + /* XAUI is a single high speed port */ + case CVMX_HELPER_INTERFACE_MODE_XAUI: + case CVMX_HELPER_INTERFACE_MODE_RXAUI: + result = __cvmx_helper_xaui_enable(interface); + break; + /* RGMII/GMII/MII are all treated about the same. Most functions + refer to these ports as RGMII */ + case CVMX_HELPER_INTERFACE_MODE_RGMII: + case CVMX_HELPER_INTERFACE_MODE_GMII: + result = __cvmx_helper_rgmii_enable(interface); + break; + /* SPI4 can have 1-16 ports depending on the device at the other end */ + case CVMX_HELPER_INTERFACE_MODE_SPI: + result = __cvmx_helper_spi_enable(interface); + break; + /* SGMII can have 1-4 ports depending on how many are hooked up */ + case CVMX_HELPER_INTERFACE_MODE_SGMII: + case CVMX_HELPER_INTERFACE_MODE_PICMG: + result = __cvmx_helper_sgmii_enable(interface); + break; + /* PCI target Network Packet Interface */ + case CVMX_HELPER_INTERFACE_MODE_NPI: + result = __cvmx_helper_npi_enable(interface); + break; + /* SRIO has 2^N ports, where N is number of interfaces */ + case CVMX_HELPER_INTERFACE_MODE_SRIO: + result = __cvmx_helper_srio_enable(interface); + break; + case CVMX_HELPER_INTERFACE_MODE_ILK: + result = __cvmx_helper_ilk_enable(interface); + break; + } + result |= __cvmx_helper_board_hardware_enable(interface); + return result; +} + + +/** + * Called after all internal packet IO paths are setup. This + * function enables IPD/PIP and begins packet input and output. + * + * @return Zero on success, negative on failure + */ +int cvmx_helper_ipd_and_packet_input_enable(void) +{ + int num_interfaces; + int interface; + + /* Enable IPD */ + cvmx_ipd_enable(); + + /* Time to enable hardware ports packet input and output. Note that at this + point IPD/PIP must be fully functional and PKO must be disabled */ + num_interfaces = cvmx_helper_get_number_of_interfaces(); + for (interface=0; interface<num_interfaces; interface++) + { + if (cvmx_helper_ports_on_interface(interface) > 0) + { + //cvmx_dprintf("Enabling packet I/O on interface %d\n", interface); + __cvmx_helper_packet_hardware_enable(interface); + } + } + + /* Finally enable PKO now that the entire path is up and running */ + cvmx_pko_enable(); + + if ((OCTEON_IS_MODEL(OCTEON_CN31XX_PASS1) || OCTEON_IS_MODEL(OCTEON_CN30XX_PASS1)) && + (cvmx_sysinfo_get()->board_type != CVMX_BOARD_TYPE_SIM)) + __cvmx_helper_errata_fix_ipd_ptr_alignment(); + return 0; +} +#ifdef CVMX_BUILD_FOR_LINUX_KERNEL +EXPORT_SYMBOL(cvmx_helper_ipd_and_packet_input_enable); +#endif + +#define __CVMX_SSO_RWQ_SIZE 256 + +int cvmx_helper_initialize_sso(int wqe_entries) +{ + int cvm_oct_sso_number_rwq_bufs; + char *mem; + int i; + cvmx_sso_cfg_t sso_cfg; + cvmx_fpa_fpfx_marks_t fpa_marks; + + if (!OCTEON_IS_MODEL(OCTEON_CN68XX)) + return 0; + + /* + * CN68XX-P1 may reset with the wrong values, put in + * the correct values. + */ + fpa_marks.u64 = 0; + fpa_marks.s.fpf_wr = 0xa4; + fpa_marks.s.fpf_rd = 0x40; + cvmx_write_csr(CVMX_FPA_FPF8_MARKS, fpa_marks.u64); + + cvm_oct_sso_number_rwq_bufs = ((wqe_entries - 1) / 26) + 1 + 48 + 8; + + mem = cvmx_bootmem_alloc(__CVMX_SSO_RWQ_SIZE * cvm_oct_sso_number_rwq_bufs, CVMX_CACHE_LINE_SIZE); + if (mem == NULL) { + cvmx_dprintf("Out of memory initializing sso pool\n"); + return -1; + } + /* Make sure RWI/RWO is disabled. */ + sso_cfg.u64 = cvmx_read_csr(CVMX_SSO_CFG); + sso_cfg.s.rwen = 0; + cvmx_write_csr(CVMX_SSO_CFG, sso_cfg.u64); + + for (i = cvm_oct_sso_number_rwq_bufs - 8; i > 0; i--) { + cvmx_sso_rwq_psh_fptr_t fptr; + + for (;;) { + fptr.u64 = cvmx_read_csr(CVMX_SSO_RWQ_PSH_FPTR); + if (!fptr.s.full) + break; + cvmx_wait(1000); + } + fptr.s.fptr = cvmx_ptr_to_phys(mem) >> 7; + cvmx_write_csr(CVMX_SSO_RWQ_PSH_FPTR, fptr.u64); + mem = mem + __CVMX_SSO_RWQ_SIZE; + } + + for (i = 0; i < 8; i++) { + cvmx_sso_rwq_head_ptrx_t head_ptr; + cvmx_sso_rwq_tail_ptrx_t tail_ptr; + + head_ptr.u64 = 0; + tail_ptr.u64 = 0; + head_ptr.s.ptr = cvmx_ptr_to_phys(mem) >> 7; + tail_ptr.s.ptr = head_ptr.s.ptr; + cvmx_write_csr(CVMX_SSO_RWQ_HEAD_PTRX(i), head_ptr.u64); + cvmx_write_csr(CVMX_SSO_RWQ_TAIL_PTRX(i), tail_ptr.u64); + mem = mem + __CVMX_SSO_RWQ_SIZE; + } + + sso_cfg.u64 = cvmx_read_csr(CVMX_SSO_CFG); + sso_cfg.s.rwen = 1; + sso_cfg.s.dwb = cvmx_helper_cfg_opt_get(CVMX_HELPER_CFG_OPT_USE_DWB); + sso_cfg.s.rwq_byp_dis = 0; + sso_cfg.s.rwio_byp_dis = 0; + cvmx_write_csr(CVMX_SSO_CFG, sso_cfg.u64); + + return 0; +} + +int cvmx_helper_uninitialize_sso(void) +{ + cvmx_fpa_quex_available_t queue_available; + cvmx_sso_cfg_t sso_cfg; + cvmx_sso_rwq_pop_fptr_t pop_fptr; + cvmx_sso_rwq_psh_fptr_t fptr; + cvmx_sso_fpage_cnt_t fpage_cnt; + int num_to_transfer, i; + char *mem; + + if (!OCTEON_IS_MODEL(OCTEON_CN68XX)) + return 0; + + sso_cfg.u64 = cvmx_read_csr(CVMX_SSO_CFG); + sso_cfg.s.rwen = 0; + sso_cfg.s.rwq_byp_dis = 1; + cvmx_write_csr(CVMX_SSO_CFG, sso_cfg.u64); + cvmx_read_csr(CVMX_SSO_CFG); + queue_available.u64 = cvmx_read_csr(CVMX_FPA_QUEX_AVAILABLE(8)); + + /* Make CVMX_FPA_QUEX_AVAILABLE(8) % 16 == 0*/ + for (num_to_transfer = (16 - queue_available.s.que_siz) % 16; + num_to_transfer > 0; num_to_transfer--) { + do { + pop_fptr.u64 = cvmx_read_csr(CVMX_SSO_RWQ_POP_FPTR); + } while (!pop_fptr.s.val); + for (;;) { + fptr.u64 = cvmx_read_csr(CVMX_SSO_RWQ_PSH_FPTR); + if (!fptr.s.full) + break; + cvmx_wait(1000); + } + fptr.s.fptr = pop_fptr.s.fptr; + cvmx_write_csr(CVMX_SSO_RWQ_PSH_FPTR, fptr.u64); + } + cvmx_read_csr(CVMX_SSO_CFG); + + do { + queue_available.u64 = cvmx_read_csr(CVMX_FPA_QUEX_AVAILABLE(8)); + } while (queue_available.s.que_siz % 16); + + sso_cfg.s.rwen = 1; + sso_cfg.s.rwq_byp_dis = 0; + cvmx_write_csr(CVMX_SSO_CFG, sso_cfg.u64); + + for (i = 0; i < 8; i++) { + cvmx_sso_rwq_head_ptrx_t head_ptr; + cvmx_sso_rwq_tail_ptrx_t tail_ptr; + + head_ptr.u64 = cvmx_read_csr(CVMX_SSO_RWQ_HEAD_PTRX(i)); + tail_ptr.u64 = cvmx_read_csr(CVMX_SSO_RWQ_TAIL_PTRX(i)); + if (head_ptr.s.ptr != tail_ptr.s.ptr) { + cvmx_dprintf("head_ptr.s.ptr != tail_ptr.s.ptr, idx: %d\n", i); + } + + mem = cvmx_phys_to_ptr(((uint64_t)head_ptr.s.ptr) << 7); + /* Leak the memory */ + } + + do { + do { + pop_fptr.u64 = cvmx_read_csr(CVMX_SSO_RWQ_POP_FPTR); + if (pop_fptr.s.val) { + mem = cvmx_phys_to_ptr(((uint64_t)pop_fptr.s.fptr) << 7); + /* Leak the memory */ + } + } while (pop_fptr.s.val); + fpage_cnt.u64 = cvmx_read_csr(CVMX_SSO_FPAGE_CNT); + } while (fpage_cnt.s.fpage_cnt); + + sso_cfg.s.rwen = 0; + sso_cfg.s.rwq_byp_dis = 0; + cvmx_write_csr(CVMX_SSO_CFG, sso_cfg.u64); + + return 0; +} + +/** + * Initialize the PIP, IPD, and PKO hardware to support + * simple priority based queues for the ethernet ports. Each + * port is configured with a number of priority queues based + * on CVMX_PKO_QUEUES_PER_PORT_* where each queue is lower + * priority than the previous. + * + * @return Zero on success, non-zero on failure + */ +int cvmx_helper_initialize_packet_io_global(void) +{ + int result = 0; + int interface; + cvmx_l2c_cfg_t l2c_cfg; + cvmx_smix_en_t smix_en; + const int num_interfaces = cvmx_helper_get_number_of_interfaces(); + + /* CN52XX pass 1: Due to a bug in 2nd order CDR, it needs to be disabled */ + if (OCTEON_IS_MODEL(OCTEON_CN52XX_PASS1_0)) + __cvmx_helper_errata_qlm_disable_2nd_order_cdr(1); + + /* Tell L2 to give the IOB statically higher priority compared to the + cores. This avoids conditions where IO blocks might be starved under + very high L2 loads */ + if (OCTEON_IS_MODEL(OCTEON_CN6XXX) || OCTEON_IS_MODEL(OCTEON_CNF7XXX)) + { + cvmx_l2c_ctl_t l2c_ctl; + l2c_ctl.u64 = cvmx_read_csr(CVMX_L2C_CTL); + l2c_ctl.s.rsp_arb_mode = 1; + l2c_ctl.s.xmc_arb_mode = 0; + cvmx_write_csr(CVMX_L2C_CTL, l2c_ctl.u64); + } + else + { + l2c_cfg.u64 = cvmx_read_csr(CVMX_L2C_CFG); + l2c_cfg.s.lrf_arb_mode = 0; + l2c_cfg.s.rfb_arb_mode = 0; + cvmx_write_csr(CVMX_L2C_CFG, l2c_cfg.u64); + } + + if (cvmx_sysinfo_get()->board_type != CVMX_BOARD_TYPE_SIM) + { + int smi_inf = 1; + int i; + + /* Newer chips have more than one SMI/MDIO interface */ + if (OCTEON_IS_MODEL(OCTEON_CN68XX)) + smi_inf = 4; + else if (!OCTEON_IS_MODEL(OCTEON_CN3XXX) + && !OCTEON_IS_MODEL(OCTEON_CN58XX) + && !OCTEON_IS_MODEL(OCTEON_CN50XX)) + smi_inf = 2; + + for (i = 0; i < smi_inf; i++) + { + /* Make sure SMI/MDIO is enabled so we can query PHYs */ + smix_en.u64 = cvmx_read_csr(CVMX_SMIX_EN(i)); + if (!smix_en.s.en) + { + smix_en.s.en = 1; + cvmx_write_csr(CVMX_SMIX_EN(i), smix_en.u64); + } + } + } + + __cvmx_helper_cfg_init(); + + for (interface=0; interface<num_interfaces; interface++) + result |= cvmx_helper_interface_probe(interface); + + cvmx_pko_initialize_global(); + for (interface=0; interface<num_interfaces; interface++) + { + if (cvmx_helper_ports_on_interface(interface) > 0) + cvmx_dprintf("Interface %d has %d ports (%s)\n", + interface, cvmx_helper_ports_on_interface(interface), + cvmx_helper_interface_mode_to_string(cvmx_helper_interface_get_mode(interface))); + result |= __cvmx_helper_interface_setup_ipd(interface); + if (!OCTEON_IS_MODEL(OCTEON_CN68XX)) + result |= __cvmx_helper_interface_setup_pko(interface); + } + + result |= __cvmx_helper_global_setup_ipd(); + result |= __cvmx_helper_global_setup_pko(); + + /* Enable any flow control and backpressure */ + result |= __cvmx_helper_global_setup_backpressure(); + +#if CVMX_HELPER_ENABLE_IPD + result |= cvmx_helper_ipd_and_packet_input_enable(); +#endif + return result; +} +#ifdef CVMX_BUILD_FOR_LINUX_KERNEL +EXPORT_SYMBOL(cvmx_helper_initialize_packet_io_global); +#endif + + +/** + * Does core local initialization for packet io + * + * @return Zero on success, non-zero on failure + */ +int cvmx_helper_initialize_packet_io_local(void) +{ + return cvmx_pko_initialize_local(); +} + +/** + * wait for the pko queue to drain + * + * @param queue a valid pko queue + * @return count is the length of the queue after calling this + * function + */ +static int cvmx_helper_wait_pko_queue_drain(int queue) +{ + const int timeout = 5; /* Wait up to 5 seconds for timeouts */ + int count; + uint64_t start_cycle, stop_cycle; + + count = cvmx_cmd_queue_length(CVMX_CMD_QUEUE_PKO(queue)); + start_cycle = cvmx_get_cycle(); + stop_cycle = start_cycle + cvmx_clock_get_rate(CVMX_CLOCK_CORE) * timeout; + while (count && (cvmx_get_cycle() < stop_cycle)) + { + cvmx_wait(10000); + count = cvmx_cmd_queue_length(CVMX_CMD_QUEUE_PKO(queue)); + } + + return count; +} + +struct cvmx_buffer_list { + struct cvmx_buffer_list *next; +}; + +/** + * Undo the initialization performed in + * cvmx_helper_initialize_packet_io_global(). After calling this routine and the + * local version on each core, packet IO for Octeon will be disabled and placed + * in the initial reset state. It will then be safe to call the initialize + * later on. Note that this routine does not empty the FPA pools. It frees all + * buffers used by the packet IO hardware to the FPA so a function emptying the + * FPA after shutdown should find all packet buffers in the FPA. + * + * @return Zero on success, negative on failure. + */ +int cvmx_helper_shutdown_packet_io_global(void) +{ + const int timeout = 5; /* Wait up to 5 seconds for timeouts */ + int result = 0; + int num_interfaces; + int interface; + int num_ports; + int index; + struct cvmx_buffer_list *pool0_buffers; + struct cvmx_buffer_list *pool0_buffers_tail; + cvmx_wqe_t *work; + + /* Step 1: Disable all backpressure */ + for (interface=0; interface<CVMX_HELPER_MAX_GMX; interface++) + if (cvmx_helper_interface_get_mode(interface) != CVMX_HELPER_INTERFACE_MODE_DISABLED) + cvmx_gmx_set_backpressure_override(interface, 0xf); + +step2: + /* Step 2: Wait for the PKO queues to drain */ + if (octeon_has_feature(OCTEON_FEATURE_PKND)) + { + int queue, max_queue; + + max_queue = __cvmx_helper_cfg_pko_max_queue(); + for (queue = 0; queue < max_queue; queue++) + { + if (cvmx_helper_wait_pko_queue_drain(queue)) + { + result = -1; + goto step3; + } + } + } + else + { + num_interfaces = cvmx_helper_get_number_of_interfaces(); + for (interface=0; interface<num_interfaces; interface++) + { + num_ports = cvmx_helper_ports_on_interface(interface); + for (index=0; index<num_ports; index++) + { + int pko_port = cvmx_helper_get_ipd_port(interface, index); + int queue = cvmx_pko_get_base_queue(pko_port); + int max_queue = queue + cvmx_pko_get_num_queues(pko_port); + while (queue < max_queue) + { + if (cvmx_helper_wait_pko_queue_drain(queue)) + { + result = -1; + goto step3; + } + queue++; + } + } + } + } + +step3: + /* Step 3: Disable TX and RX on all ports */ + for (interface=0; interface<CVMX_HELPER_MAX_GMX; interface++) + { + switch (cvmx_helper_interface_get_mode(interface)) + { + case CVMX_HELPER_INTERFACE_MODE_DISABLED: + case CVMX_HELPER_INTERFACE_MODE_PCIE: + /* Not a packet interface */ + break; + case CVMX_HELPER_INTERFACE_MODE_NPI: + case CVMX_HELPER_INTERFACE_MODE_SRIO: + case CVMX_HELPER_INTERFACE_MODE_ILK: + /* We don't handle the NPI/NPEI/SRIO packet engines. The caller + must know these are idle */ + break; + case CVMX_HELPER_INTERFACE_MODE_LOOP: + /* Nothing needed. Once PKO is idle, the loopback devices + must be idle */ + break; + case CVMX_HELPER_INTERFACE_MODE_SPI: + /* SPI cannot be disabled from Octeon. It is the responsibility + of the caller to make sure SPI is idle before doing + shutdown */ + /* Fall through and do the same processing as RGMII/GMII */ + case CVMX_HELPER_INTERFACE_MODE_GMII: + case CVMX_HELPER_INTERFACE_MODE_RGMII: + /* Disable outermost RX at the ASX block */ + cvmx_write_csr(CVMX_ASXX_RX_PRT_EN(interface), 0); + num_ports = cvmx_helper_ports_on_interface(interface); + if (num_ports > 4) + num_ports = 4; + for (index=0; index<num_ports; index++) + { + cvmx_gmxx_prtx_cfg_t gmx_cfg; + gmx_cfg.u64 = cvmx_read_csr(CVMX_GMXX_PRTX_CFG(index, interface)); + gmx_cfg.s.en = 0; + cvmx_write_csr(CVMX_GMXX_PRTX_CFG(index, interface), gmx_cfg.u64); + /* Poll the GMX state machine waiting for it to become idle */ + cvmx_write_csr(CVMX_NPI_DBG_SELECT, interface*0x800 + index*0x100 + 0x880); + if (CVMX_WAIT_FOR_FIELD64(CVMX_DBG_DATA, cvmx_dbg_data_t, data&7, ==, 0, timeout*1000000)) + { + cvmx_dprintf("GMX RX path timeout waiting for idle\n"); + result = -1; + } + if (CVMX_WAIT_FOR_FIELD64(CVMX_DBG_DATA, cvmx_dbg_data_t, data&0xf, ==, 0, timeout*1000000)) + { + cvmx_dprintf("GMX TX path timeout waiting for idle\n"); + result = -1; + } + } + /* Disable outermost TX at the ASX block */ + cvmx_write_csr(CVMX_ASXX_TX_PRT_EN(interface), 0); + /* Disable interrupts for interface */ + cvmx_write_csr(CVMX_ASXX_INT_EN(interface), 0); + cvmx_write_csr(CVMX_GMXX_TX_INT_EN(interface), 0); + break; + case CVMX_HELPER_INTERFACE_MODE_XAUI: + case CVMX_HELPER_INTERFACE_MODE_RXAUI: + case CVMX_HELPER_INTERFACE_MODE_SGMII: + case CVMX_HELPER_INTERFACE_MODE_PICMG: + num_ports = cvmx_helper_ports_on_interface(interface); + if (num_ports > 4) + num_ports = 4; + for (index=0; index<num_ports; index++) + { + cvmx_gmxx_prtx_cfg_t gmx_cfg; + gmx_cfg.u64 = cvmx_read_csr(CVMX_GMXX_PRTX_CFG(index, interface)); + gmx_cfg.s.en = 0; + cvmx_write_csr(CVMX_GMXX_PRTX_CFG(index, interface), gmx_cfg.u64); + if (CVMX_WAIT_FOR_FIELD64(CVMX_GMXX_PRTX_CFG(index, interface), cvmx_gmxx_prtx_cfg_t, rx_idle, ==, 1, timeout*1000000)) + { + cvmx_dprintf("GMX RX path timeout waiting for idle\n"); + result = -1; + } + if (CVMX_WAIT_FOR_FIELD64(CVMX_GMXX_PRTX_CFG(index, interface), cvmx_gmxx_prtx_cfg_t, tx_idle, ==, 1, timeout*1000000)) + { + cvmx_dprintf("GMX TX path timeout waiting for idle\n"); + result = -1; + } + } + break; + } + } + + /* Step 4: Retrieve all packets from the POW and free them */ + while ((work = cvmx_pow_work_request_sync(CVMX_POW_WAIT))) + { + cvmx_helper_free_packet_data(work); + cvmx_fpa_free(work, CVMX_FPA_WQE_POOL, 0); + } + + /* Step 4b: Special workaround for pass 2 errata */ + if (OCTEON_IS_MODEL(OCTEON_CN38XX_PASS2)) + { + cvmx_ipd_ptr_count_t ipd_cnt; + int to_add; + ipd_cnt.u64 = cvmx_read_csr(CVMX_IPD_PTR_COUNT); + to_add = (ipd_cnt.s.wqev_cnt + ipd_cnt.s.wqe_pcnt) & 0x7; + if (to_add) + { + int port = -1; + cvmx_dprintf("Aligning CN38XX pass 2 IPD counters\n"); + if (cvmx_helper_interface_get_mode(0) == CVMX_HELPER_INTERFACE_MODE_RGMII) + port = 0; + else if (cvmx_helper_interface_get_mode(1) == CVMX_HELPER_INTERFACE_MODE_RGMII) + port = 16; + + if (port != -1) + { + char *buffer = cvmx_fpa_alloc(CVMX_FPA_PACKET_POOL); + if (buffer) + { + int queue = cvmx_pko_get_base_queue(port); + cvmx_pko_command_word0_t pko_command; + cvmx_buf_ptr_t packet; + uint64_t start_cycle; + uint64_t stop_cycle; + + /* Populate a minimal packet */ + memset(buffer, 0xff, 6); + memset(buffer+6, 0, 54); + pko_command.u64 = 0; + pko_command.s.dontfree = 1; + pko_command.s.total_bytes = 60; + pko_command.s.segs = 1; + packet.u64 = 0; + packet.s.addr = cvmx_ptr_to_phys(buffer); + packet.s.size = CVMX_FPA_PACKET_POOL_SIZE; + __cvmx_helper_rgmii_configure_loopback(port, 1, 0); + while (to_add--) + { + cvmx_pko_send_packet_prepare(port, queue, CVMX_PKO_LOCK_CMD_QUEUE); + if (cvmx_pko_send_packet_finish(port, queue, pko_command, packet, CVMX_PKO_LOCK_CMD_QUEUE)) + { + cvmx_dprintf("ERROR: Unable to align IPD counters (PKO failed)\n"); + break; + } + } + cvmx_fpa_free(buffer, CVMX_FPA_PACKET_POOL, 0); + + /* Wait for the packets to loop back */ + start_cycle = cvmx_get_cycle(); + stop_cycle = start_cycle + cvmx_clock_get_rate(CVMX_CLOCK_CORE) * timeout; + while (cvmx_cmd_queue_length(CVMX_CMD_QUEUE_PKO(queue)) && + (cvmx_get_cycle() < stop_cycle)) + { + cvmx_wait(1000); + } + cvmx_wait(1000); + __cvmx_helper_rgmii_configure_loopback(port, 0, 0); + if (to_add == -1) + goto step2; + } + else + cvmx_dprintf("ERROR: Unable to align IPD counters (Packet pool empty)\n"); + } + else + cvmx_dprintf("ERROR: Unable to align IPD counters\n"); + } + } + + /* Step 5 */ + cvmx_ipd_disable(); + + /* Step 6: Drain all prefetched buffers from IPD/PIP. Note that IPD/PIP + have not been reset yet */ + __cvmx_ipd_free_ptr(); + + /* Step 7: Free the PKO command buffers and put PKO in reset */ + cvmx_pko_shutdown(); + + /* Step 8: Disable MAC address filtering */ + for (interface=0; interface<CVMX_HELPER_MAX_GMX; interface++) + { + switch (cvmx_helper_interface_get_mode(interface)) + { + case CVMX_HELPER_INTERFACE_MODE_DISABLED: + case CVMX_HELPER_INTERFACE_MODE_PCIE: + case CVMX_HELPER_INTERFACE_MODE_SRIO: + case CVMX_HELPER_INTERFACE_MODE_ILK: + case CVMX_HELPER_INTERFACE_MODE_NPI: + case CVMX_HELPER_INTERFACE_MODE_LOOP: + break; + case CVMX_HELPER_INTERFACE_MODE_XAUI: + case CVMX_HELPER_INTERFACE_MODE_RXAUI: + case CVMX_HELPER_INTERFACE_MODE_GMII: + case CVMX_HELPER_INTERFACE_MODE_RGMII: + case CVMX_HELPER_INTERFACE_MODE_SPI: + case CVMX_HELPER_INTERFACE_MODE_SGMII: + case CVMX_HELPER_INTERFACE_MODE_PICMG: + num_ports = cvmx_helper_ports_on_interface(interface); + if (num_ports > 4) + num_ports = 4; + for (index=0; index<num_ports; index++) + { + cvmx_write_csr(CVMX_GMXX_RXX_ADR_CTL(index, interface), 1); + cvmx_write_csr(CVMX_GMXX_RXX_ADR_CAM_EN(index, interface), 0); + cvmx_write_csr(CVMX_GMXX_RXX_ADR_CAM0(index, interface), 0); + cvmx_write_csr(CVMX_GMXX_RXX_ADR_CAM1(index, interface), 0); + cvmx_write_csr(CVMX_GMXX_RXX_ADR_CAM2(index, interface), 0); + cvmx_write_csr(CVMX_GMXX_RXX_ADR_CAM3(index, interface), 0); + cvmx_write_csr(CVMX_GMXX_RXX_ADR_CAM4(index, interface), 0); + cvmx_write_csr(CVMX_GMXX_RXX_ADR_CAM5(index, interface), 0); + } + break; + } + } + + /* Step 9: Drain all FPA buffers out of pool 0 before we reset + * IPD/PIP. This is needed to keep IPD_QUE0_FREE_PAGE_CNT in + * sync. We temporarily keep the buffers in the pool0_buffers + * list. + */ + pool0_buffers = NULL; + pool0_buffers_tail = NULL; + while (1) + { + struct cvmx_buffer_list *buffer = cvmx_fpa_alloc(0); + if (buffer) { + buffer->next = NULL; + + if (pool0_buffers == NULL) + pool0_buffers = buffer; + else + pool0_buffers_tail->next = buffer; + + pool0_buffers_tail = buffer; + } + else + break; + } + + /* Step 10: Reset IPD and PIP */ + { + cvmx_ipd_ctl_status_t ipd_ctl_status; + ipd_ctl_status.u64 = cvmx_read_csr(CVMX_IPD_CTL_STATUS); + ipd_ctl_status.s.reset = 1; + cvmx_write_csr(CVMX_IPD_CTL_STATUS, ipd_ctl_status.u64); + + if ((cvmx_sysinfo_get()->board_type != CVMX_BOARD_TYPE_SIM) && + (OCTEON_IS_MODEL(OCTEON_CN3XXX) || OCTEON_IS_MODEL(OCTEON_CN5XXX))) + { + /* only try 1000 times. Normally if this works it will happen in + ** the first 50 loops. */ + int max_loops = 1000; + int loop = 0; + /* Per port backpressure counters can get misaligned after an + IPD reset. This code realigns them by performing repeated + resets. See IPD-13473 */ + cvmx_wait(100); + if (__cvmx_helper_backpressure_is_misaligned()) + { + cvmx_dprintf("Starting to align per port backpressure counters.\n"); + while (__cvmx_helper_backpressure_is_misaligned() && (loop++ < max_loops)) + { + cvmx_write_csr(CVMX_IPD_CTL_STATUS, ipd_ctl_status.u64); + cvmx_wait(123); + } + if (loop < max_loops) + cvmx_dprintf("Completed aligning per port backpressure counters (%d loops).\n", loop); + else + { + cvmx_dprintf("ERROR: unable to align per port backpressure counters.\n"); + /* For now, don't hang.... */ + } + } + } + + /* PIP_SFT_RST not present in CN38XXp{1,2} */ + if (!OCTEON_IS_MODEL(OCTEON_CN38XX_PASS2)) + { + cvmx_pip_sft_rst_t pip_sft_rst; + pip_sft_rst.u64 = cvmx_read_csr(CVMX_PIP_SFT_RST); + pip_sft_rst.s.rst = 1; + cvmx_write_csr(CVMX_PIP_SFT_RST, pip_sft_rst.u64); + } + + /* Make sure IPD has finished reset. */ + if (OCTEON_IS_MODEL(OCTEON_CN6XXX)) + { + if (CVMX_WAIT_FOR_FIELD64(CVMX_IPD_CTL_STATUS, cvmx_ipd_ctl_status_t, rst_done, ==, 0, 1000)) + { + cvmx_dprintf("IPD reset timeout waiting for idle\n"); + result = -1; + } + } + } + + /* Step 11: Restore the FPA buffers into pool 0 */ + while (pool0_buffers) { + struct cvmx_buffer_list *n = pool0_buffers->next; + cvmx_fpa_free(pool0_buffers, 0, 0); + pool0_buffers = n; + } + + /* Step 12: Release interface structures */ + __cvmx_helper_shutdown_interfaces(); + + return result; +} +#ifdef CVMX_BUILD_FOR_LINUX_KERNEL +EXPORT_SYMBOL(cvmx_helper_shutdown_packet_io_global); +#endif + + +/** + * Does core local shutdown of packet io + * + * @return Zero on success, non-zero on failure + */ +int cvmx_helper_shutdown_packet_io_local(void) +{ + /* Currently there is nothing to do per core. This may change in + the future */ + return 0; +} + + + +/** + * Auto configure an IPD/PKO port link state and speed. This + * function basically does the equivalent of: + * cvmx_helper_link_set(ipd_port, cvmx_helper_link_get(ipd_port)); + * + * @param ipd_port IPD/PKO port to auto configure + * + * @return Link state after configure + */ +cvmx_helper_link_info_t cvmx_helper_link_autoconf(int ipd_port) +{ + cvmx_helper_link_info_t link_info; + int interface = cvmx_helper_get_interface_num(ipd_port); + int index = cvmx_helper_get_interface_index_num(ipd_port); + + if (index >= cvmx_helper_ports_on_interface(interface)) + { + link_info.u64 = 0; + return link_info; + } + + link_info = cvmx_helper_link_get(ipd_port); + if (link_info.u64 == (__cvmx_helper_get_link_info(interface, index)).u64) + return link_info; + +#if !defined(CVMX_BUILD_FOR_LINUX_KERNEL) && !defined(CVMX_BUILD_FOR_FREEBSD_KERNEL) + if (!link_info.s.link_up) + cvmx_error_disable_group(CVMX_ERROR_GROUP_ETHERNET, ipd_port); +#endif + + /* If we fail to set the link speed, port_link_info will not change */ + cvmx_helper_link_set(ipd_port, link_info); + +#if !defined(CVMX_BUILD_FOR_LINUX_KERNEL) && !defined(CVMX_BUILD_FOR_FREEBSD_KERNEL) + if (link_info.s.link_up) + cvmx_error_enable_group(CVMX_ERROR_GROUP_ETHERNET, ipd_port); +#endif + + return link_info; +} +#ifdef CVMX_BUILD_FOR_LINUX_KERNEL +EXPORT_SYMBOL(cvmx_helper_link_autoconf); +#endif + +/** + * Return the link state of an IPD/PKO port as returned by + * auto negotiation. The result of this function may not match + * Octeon's link config if auto negotiation has changed since + * the last call to cvmx_helper_link_set(). + * + * @param ipd_port IPD/PKO port to query + * + * @return Link state + */ +cvmx_helper_link_info_t cvmx_helper_link_get(int ipd_port) +{ + cvmx_helper_link_info_t result; + int interface = cvmx_helper_get_interface_num(ipd_port); + int index = cvmx_helper_get_interface_index_num(ipd_port); + + /* The default result will be a down link unless the code below + changes it */ + result.u64 = 0; + + if (index >= cvmx_helper_ports_on_interface(interface)) + return result; + + switch (cvmx_helper_interface_get_mode(interface)) + { + case CVMX_HELPER_INTERFACE_MODE_DISABLED: + case CVMX_HELPER_INTERFACE_MODE_PCIE: + /* Network links are not supported */ + break; + case CVMX_HELPER_INTERFACE_MODE_XAUI: + case CVMX_HELPER_INTERFACE_MODE_RXAUI: + result = __cvmx_helper_xaui_link_get(ipd_port); + break; + case CVMX_HELPER_INTERFACE_MODE_GMII: + if (index == 0) + result = __cvmx_helper_rgmii_link_get(ipd_port); + else + { + result.s.full_duplex = 1; + result.s.link_up = 1; + result.s.speed = 1000; + } + break; + case CVMX_HELPER_INTERFACE_MODE_RGMII: + result = __cvmx_helper_rgmii_link_get(ipd_port); + break; + case CVMX_HELPER_INTERFACE_MODE_SPI: + result = __cvmx_helper_spi_link_get(ipd_port); + break; + case CVMX_HELPER_INTERFACE_MODE_SGMII: + case CVMX_HELPER_INTERFACE_MODE_PICMG: + result = __cvmx_helper_sgmii_link_get(ipd_port); + break; + case CVMX_HELPER_INTERFACE_MODE_SRIO: + result = __cvmx_helper_srio_link_get(ipd_port); + break; + case CVMX_HELPER_INTERFACE_MODE_ILK: + result = __cvmx_helper_ilk_link_get(ipd_port); + break; + case CVMX_HELPER_INTERFACE_MODE_NPI: + case CVMX_HELPER_INTERFACE_MODE_LOOP: + /* Network links are not supported */ + break; + } + return result; +} +#ifdef CVMX_BUILD_FOR_LINUX_KERNEL +EXPORT_SYMBOL(cvmx_helper_link_get); +#endif + + +/** + * Configure an IPD/PKO port for the specified link state. This + * function does not influence auto negotiation at the PHY level. + * The passed link state must always match the link state returned + * by cvmx_helper_link_get(). It is normally best to use + * cvmx_helper_link_autoconf() instead. + * + * @param ipd_port IPD/PKO port to configure + * @param link_info The new link state + * + * @return Zero on success, negative on failure + */ +int cvmx_helper_link_set(int ipd_port, cvmx_helper_link_info_t link_info) +{ + int result = -1; + int interface = cvmx_helper_get_interface_num(ipd_port); + int index = cvmx_helper_get_interface_index_num(ipd_port); + + if (index >= cvmx_helper_ports_on_interface(interface)) + return -1; + + switch (cvmx_helper_interface_get_mode(interface)) + { + case CVMX_HELPER_INTERFACE_MODE_DISABLED: + case CVMX_HELPER_INTERFACE_MODE_PCIE: + break; + case CVMX_HELPER_INTERFACE_MODE_XAUI: + case CVMX_HELPER_INTERFACE_MODE_RXAUI: + result = __cvmx_helper_xaui_link_set(ipd_port, link_info); + break; + /* RGMII/GMII/MII are all treated about the same. Most functions + refer to these ports as RGMII */ + case CVMX_HELPER_INTERFACE_MODE_RGMII: + case CVMX_HELPER_INTERFACE_MODE_GMII: + result = __cvmx_helper_rgmii_link_set(ipd_port, link_info); + break; + case CVMX_HELPER_INTERFACE_MODE_SPI: + result = __cvmx_helper_spi_link_set(ipd_port, link_info); + break; + case CVMX_HELPER_INTERFACE_MODE_SGMII: + case CVMX_HELPER_INTERFACE_MODE_PICMG: + result = __cvmx_helper_sgmii_link_set(ipd_port, link_info); + break; + case CVMX_HELPER_INTERFACE_MODE_SRIO: + result = __cvmx_helper_srio_link_set(ipd_port, link_info); + break; + case CVMX_HELPER_INTERFACE_MODE_ILK: + result = __cvmx_helper_ilk_link_set(ipd_port, link_info); + break; + case CVMX_HELPER_INTERFACE_MODE_NPI: + case CVMX_HELPER_INTERFACE_MODE_LOOP: + break; + } + /* Set the port_link_info here so that the link status is updated + no matter how cvmx_helper_link_set is called. We don't change + the value if link_set failed */ + if (result == 0) + __cvmx_helper_set_link_info(interface, index, link_info); + return result; +} +#ifdef CVMX_BUILD_FOR_LINUX_KERNEL +EXPORT_SYMBOL(cvmx_helper_link_set); +#endif + + +/** + * Configure a port for internal and/or external loopback. Internal loopback + * causes packets sent by the port to be received by Octeon. External loopback + * causes packets received from the wire to sent out again. + * + * @param ipd_port IPD/PKO port to loopback. + * @param enable_internal + * Non zero if you want internal loopback + * @param enable_external + * Non zero if you want external loopback + * + * @return Zero on success, negative on failure. + */ +int cvmx_helper_configure_loopback(int ipd_port, int enable_internal, int enable_external) +{ + int result = -1; + int interface = cvmx_helper_get_interface_num(ipd_port); + int index = cvmx_helper_get_interface_index_num(ipd_port); + + if (index >= cvmx_helper_ports_on_interface(interface)) + return -1; + + switch (cvmx_helper_interface_get_mode(interface)) + { + case CVMX_HELPER_INTERFACE_MODE_DISABLED: + case CVMX_HELPER_INTERFACE_MODE_PCIE: + case CVMX_HELPER_INTERFACE_MODE_SRIO: + case CVMX_HELPER_INTERFACE_MODE_ILK: + case CVMX_HELPER_INTERFACE_MODE_SPI: + case CVMX_HELPER_INTERFACE_MODE_NPI: + case CVMX_HELPER_INTERFACE_MODE_LOOP: + break; + case CVMX_HELPER_INTERFACE_MODE_XAUI: + case CVMX_HELPER_INTERFACE_MODE_RXAUI: + result = __cvmx_helper_xaui_configure_loopback(ipd_port, enable_internal, enable_external); + break; + case CVMX_HELPER_INTERFACE_MODE_RGMII: + case CVMX_HELPER_INTERFACE_MODE_GMII: + result = __cvmx_helper_rgmii_configure_loopback(ipd_port, enable_internal, enable_external); + break; + case CVMX_HELPER_INTERFACE_MODE_SGMII: + case CVMX_HELPER_INTERFACE_MODE_PICMG: + result = __cvmx_helper_sgmii_configure_loopback(ipd_port, enable_internal, enable_external); + break; + } + return result; +} + +#endif /* CVMX_ENABLE_PKO_FUNCTIONS */ |