summaryrefslogtreecommitdiffstats
path: root/sys/contrib/octeon-sdk/cvmx-tim.c
diff options
context:
space:
mode:
Diffstat (limited to 'sys/contrib/octeon-sdk/cvmx-tim.c')
-rw-r--r--sys/contrib/octeon-sdk/cvmx-tim.c313
1 files changed, 313 insertions, 0 deletions
diff --git a/sys/contrib/octeon-sdk/cvmx-tim.c b/sys/contrib/octeon-sdk/cvmx-tim.c
new file mode 100644
index 0000000..5d2d3e1
--- /dev/null
+++ b/sys/contrib/octeon-sdk/cvmx-tim.c
@@ -0,0 +1,313 @@
+/***********************license start***************
+ * Copyright (c) 2003-2010 Cavium Inc. (support@cavium.com). All rights
+ * reserved.
+ *
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions are
+ * met:
+ *
+ * * Redistributions of source code must retain the above copyright
+ * notice, this list of conditions and the following disclaimer.
+ *
+ * * Redistributions in binary form must reproduce the above
+ * copyright notice, this list of conditions and the following
+ * disclaimer in the documentation and/or other materials provided
+ * with the distribution.
+
+ * * Neither the name of Cavium Inc. nor the names of
+ * its contributors may be used to endorse or promote products
+ * derived from this software without specific prior written
+ * permission.
+
+ * This Software, including technical data, may be subject to U.S. export control
+ * laws, including the U.S. Export Administration Act and its associated
+ * regulations, and may be subject to export or import regulations in other
+ * countries.
+
+ * TO THE MAXIMUM EXTENT PERMITTED BY LAW, THE SOFTWARE IS PROVIDED "AS IS"
+ * AND WITH ALL FAULTS AND CAVIUM INC. MAKES NO PROMISES, REPRESENTATIONS OR
+ * WARRANTIES, EITHER EXPRESS, IMPLIED, STATUTORY, OR OTHERWISE, WITH RESPECT TO
+ * THE SOFTWARE, INCLUDING ITS CONDITION, ITS CONFORMITY TO ANY REPRESENTATION OR
+ * DESCRIPTION, OR THE EXISTENCE OF ANY LATENT OR PATENT DEFECTS, AND CAVIUM
+ * SPECIFICALLY DISCLAIMS ALL IMPLIED (IF ANY) WARRANTIES OF TITLE,
+ * MERCHANTABILITY, NONINFRINGEMENT, FITNESS FOR A PARTICULAR PURPOSE, LACK OF
+ * VIRUSES, ACCURACY OR COMPLETENESS, QUIET ENJOYMENT, QUIET POSSESSION OR
+ * CORRESPONDENCE TO DESCRIPTION. THE ENTIRE RISK ARISING OUT OF USE OR
+ * PERFORMANCE OF THE SOFTWARE LIES WITH YOU.
+ ***********************license end**************************************/
+
+
+
+
+
+
+
+/**
+ * @file
+ *
+ * Support library for the hardware work queue timers.
+ *
+ * <hr>$Revision: 70030 $<hr>
+ */
+#include "executive-config.h"
+#include "cvmx-config.h"
+#include "cvmx.h"
+#include "cvmx-sysinfo.h"
+#include "cvmx-tim.h"
+#include "cvmx-bootmem.h"
+
+/* CSR typedefs have been moved to cvmx-tim-defs.h */
+
+/**
+ * Global structure holding the state of all timers.
+ */
+CVMX_SHARED cvmx_tim_t cvmx_tim;
+
+
+#ifdef CVMX_ENABLE_TIMER_FUNCTIONS
+/**
+ * Setup a timer for use. Must be called before the timer
+ * can be used.
+ *
+ * @param tick Time between each bucket in microseconds. This must not be
+ * smaller than 1024/(clock frequency in MHz).
+ * @param max_ticks The maximum number of ticks the timer must be able
+ * to schedule in the future. There are guaranteed to be enough
+ * timer buckets such that:
+ * number of buckets >= max_ticks.
+ * @return Zero on success. Negative on error. Failures are possible
+ * if the number of buckets needed is too large or memory
+ * allocation fails for creating the buckets.
+ */
+int cvmx_tim_setup(uint64_t tick, uint64_t max_ticks)
+{
+ uint64_t timer_id;
+ int error = -1;
+ uint64_t tim_clock_hz = cvmx_clock_get_rate(CVMX_CLOCK_TIM);
+ uint64_t hw_tick_ns;
+ uint64_t hw_tick_ns_allowed;
+ uint64_t tick_ns = 1000 * tick;
+ int i;
+ uint32_t temp;
+ int timer_thr = 1024;
+
+ /* for the simulator */
+ if (tim_clock_hz == 0)
+ tim_clock_hz = 800000000;
+
+ if (OCTEON_IS_MODEL(OCTEON_CN68XX))
+ {
+ cvmx_tim_fr_rn_tt_t fr_tt;
+ fr_tt.u64 = cvmx_read_csr(CVMX_TIM_FR_RN_TT);
+ timer_thr = fr_tt.s.fr_rn_tt;
+ }
+
+ hw_tick_ns = timer_thr * 1000000000ull / tim_clock_hz;
+ /*
+ * Double the minimal allowed tick to 2 * HW tick. tick between
+ * (hw_tick_ns, 2*hw_tick_ns) will set config_ring1.s.interval
+ * to zero, or 1024 cycles. This is not enough time for the timer unit
+ * to fetch the bucket data, Resulting in timer ring error interrupt
+ * be always generated. Avoid such setting in software.
+ */
+ hw_tick_ns_allowed = hw_tick_ns * 2;
+
+ /* Make sure the timers are stopped */
+ cvmx_tim_stop();
+
+ /* Reinitialize out timer state */
+ memset(&cvmx_tim, 0, sizeof(cvmx_tim));
+
+ if (tick_ns < hw_tick_ns_allowed)
+ {
+ cvmx_dprintf("ERROR: cvmx_tim_setup: Requested tick %lu(ns) is smaller than"
+ " the minimal ticks allowed by hardware %lu(ns)\n",
+ tick_ns, hw_tick_ns_allowed);
+ return error;
+ }
+ else if (tick_ns > 4194304 * hw_tick_ns)
+ {
+ cvmx_dprintf("ERROR: cvmx_tim_setup: Requested tick %lu(ns) is greater than"
+ " the max ticks %lu(ns)\n", tick_ns, hw_tick_ns);
+ return error;
+ }
+
+ for (i=2; i<20; i++)
+ {
+ if (tick_ns < (hw_tick_ns << i))
+ break;
+ }
+
+ cvmx_tim.max_ticks = (uint32_t)max_ticks;
+ cvmx_tim.bucket_shift = (uint32_t)(i - 1 + 10);
+ cvmx_tim.tick_cycles = tick * tim_clock_hz / 1000000;
+
+ temp = (max_ticks * cvmx_tim.tick_cycles) >> cvmx_tim.bucket_shift;
+
+ /* round up to nearest power of 2 */
+ temp -= 1;
+ temp = temp | (temp >> 1);
+ temp = temp | (temp >> 2);
+ temp = temp | (temp >> 4);
+ temp = temp | (temp >> 8);
+ temp = temp | (temp >> 16);
+ cvmx_tim.num_buckets = temp + 1;
+
+ /* ensure input params fall into permitted ranges */
+ if ((cvmx_tim.num_buckets < 3) || cvmx_tim.num_buckets > 1048576)
+ {
+ cvmx_dprintf("ERROR: cvmx_tim_setup: num_buckets out of range\n");
+ return error;
+ }
+
+ /* Allocate the timer buckets from hardware addressable memory */
+ cvmx_tim.bucket = cvmx_bootmem_alloc(CVMX_TIM_NUM_TIMERS * cvmx_tim.num_buckets
+ * sizeof(cvmx_tim_bucket_entry_t), CVMX_CACHE_LINE_SIZE);
+ if (cvmx_tim.bucket == NULL)
+ {
+ cvmx_dprintf("ERROR: cvmx_tim_setup: allocation problem\n");
+ return error;
+ }
+ memset(cvmx_tim.bucket, 0, CVMX_TIM_NUM_TIMERS * cvmx_tim.num_buckets * sizeof(cvmx_tim_bucket_entry_t));
+
+ cvmx_tim.start_time = 0;
+
+ /* Loop through all timers */
+ for (timer_id = 0; timer_id<CVMX_TIM_NUM_TIMERS; timer_id++)
+ {
+ int interval = ((1 << (cvmx_tim.bucket_shift - 10)) - 1);
+ cvmx_tim_bucket_entry_t *bucket = cvmx_tim.bucket + timer_id * cvmx_tim.num_buckets;
+ if (OCTEON_IS_MODEL(OCTEON_CN68XX))
+ {
+ cvmx_tim_ringx_ctl0_t ring_ctl0;
+ cvmx_tim_ringx_ctl1_t ring_ctl1;
+ cvmx_tim_ringx_ctl2_t ring_ctl2;
+ cvmx_tim_reg_flags_t reg_flags;
+
+ /* Tell the hardware where about the bucket array */
+ ring_ctl2.u64 = 0;
+ ring_ctl2.s.csize = CVMX_FPA_TIMER_POOL_SIZE / 8;
+ ring_ctl2.s.base = cvmx_ptr_to_phys(bucket) >> 5;
+ cvmx_write_csr(CVMX_TIM_RINGX_CTL2(timer_id), ring_ctl2.u64);
+
+ reg_flags.u64 = cvmx_read_csr(CVMX_TIM_REG_FLAGS);
+ ring_ctl1.u64 = 0;
+ ring_ctl1.s.cpool = ((reg_flags.s.ena_dfb == 0) ? CVMX_FPA_TIMER_POOL : 0);
+ ring_ctl1.s.bsize = cvmx_tim.num_buckets - 1;
+ cvmx_write_csr(CVMX_TIM_RINGX_CTL1(timer_id), ring_ctl1.u64);
+
+ ring_ctl0.u64 = 0;
+ ring_ctl0.s.timercount = interval + timer_id * interval / CVMX_TIM_NUM_TIMERS;
+ cvmx_write_csr(CVMX_TIM_RINGX_CTL0(timer_id), ring_ctl0.u64);
+
+ ring_ctl0.u64 = cvmx_read_csr(CVMX_TIM_RINGX_CTL0(timer_id));
+ ring_ctl0.s.ena = 1;
+ ring_ctl0.s.interval = interval;
+ cvmx_write_csr(CVMX_TIM_RINGX_CTL0(timer_id), ring_ctl0.u64);
+ ring_ctl0.u64 = cvmx_read_csr(CVMX_TIM_RINGX_CTL0(timer_id));
+ }
+ else
+ {
+ cvmx_tim_mem_ring0_t config_ring0;
+ cvmx_tim_mem_ring1_t config_ring1;
+ /* Tell the hardware where about the bucket array */
+ config_ring0.u64 = 0;
+ config_ring0.s.first_bucket = cvmx_ptr_to_phys(bucket) >> 5;
+ config_ring0.s.num_buckets = cvmx_tim.num_buckets - 1;
+ config_ring0.s.ring = timer_id;
+ cvmx_write_csr(CVMX_TIM_MEM_RING0, config_ring0.u64);
+
+ /* Tell the hardware the size of each chunk block in pointers */
+ config_ring1.u64 = 0;
+ config_ring1.s.enable = 1;
+ config_ring1.s.pool = CVMX_FPA_TIMER_POOL;
+ config_ring1.s.words_per_chunk = CVMX_FPA_TIMER_POOL_SIZE / 8;
+ config_ring1.s.interval = interval;
+ config_ring1.s.ring = timer_id;
+ cvmx_write_csr(CVMX_TIM_MEM_RING1, config_ring1.u64);
+ }
+ }
+
+ return 0;
+}
+#endif
+
+/**
+ * Start the hardware timer processing
+ */
+void cvmx_tim_start(void)
+{
+ cvmx_tim_control_t control;
+
+ control.u64 = cvmx_read_csr(CVMX_TIM_REG_FLAGS);
+ control.s.enable_dwb = 1;
+ control.s.enable_timers = 1;
+
+ /* Remember when we started the timers */
+ cvmx_tim.start_time = cvmx_clock_get_count(CVMX_CLOCK_TIM);
+ cvmx_write_csr(CVMX_TIM_REG_FLAGS, control.u64);
+}
+
+
+/**
+ * Stop the hardware timer processing. Timers stay configured.
+ */
+void cvmx_tim_stop(void)
+{
+ cvmx_tim_control_t control;
+ control.u64 = cvmx_read_csr(CVMX_TIM_REG_FLAGS);
+ control.s.enable_dwb = 0;
+ control.s.enable_timers = 0;
+ cvmx_write_csr(CVMX_TIM_REG_FLAGS, control.u64);
+}
+
+
+/**
+ * Stop the timer. After this the timer must be setup again
+ * before use.
+ */
+#ifdef CVMX_ENABLE_TIMER_FUNCTIONS
+void cvmx_tim_shutdown(void)
+{
+ uint32_t bucket;
+ uint64_t timer_id;
+ uint64_t entries_per_chunk;
+
+ /* Make sure the timers are stopped */
+ cvmx_tim_stop();
+
+ entries_per_chunk = CVMX_FPA_TIMER_POOL_SIZE/8 - 1;
+
+ /* Now walk all buckets freeing the chunks */
+ for (timer_id = 0; timer_id<CVMX_TIM_NUM_TIMERS; timer_id++)
+ {
+ for (bucket=0; bucket<cvmx_tim.num_buckets; bucket++)
+ {
+ uint64_t chunk_addr;
+ uint64_t next_chunk_addr;
+ cvmx_tim_bucket_entry_t *bucket_ptr = cvmx_tim.bucket + timer_id * cvmx_tim.num_buckets + bucket;
+ CVMX_PREFETCH128(CAST64(bucket_ptr)); /* prefetch the next cacheline for future buckets */
+
+ /* Each bucket contains a list of chunks */
+ chunk_addr = bucket_ptr->first_chunk_addr;
+ while (bucket_ptr->num_entries)
+ {
+#ifdef DEBUG
+ cvmx_dprintf("Freeing Timer Chunk 0x%llx\n", CAST64(chunk_addr));
+#endif
+ /* Read next chunk pointer from end of the current chunk */
+ next_chunk_addr = cvmx_read_csr(CVMX_ADD_SEG(CVMX_MIPS_SPACE_XKPHYS, chunk_addr + CVMX_FPA_TIMER_POOL_SIZE - 8));
+
+ cvmx_fpa_free(cvmx_phys_to_ptr(chunk_addr), CVMX_FPA_TIMER_POOL, 0);
+ chunk_addr = next_chunk_addr;
+ if (bucket_ptr->num_entries > entries_per_chunk)
+ bucket_ptr->num_entries -= entries_per_chunk;
+ else
+ bucket_ptr->num_entries = 0;
+ }
+ }
+ }
+}
+
+#endif
OpenPOWER on IntegriCloud