summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--sys/amd64/conf/GENERIC2
-rw-r--r--sys/conf/NOTES9
-rw-r--r--sys/conf/files3
-rw-r--r--sys/dev/iir/iir.c2018
-rw-r--r--sys/dev/iir/iir.h712
-rw-r--r--sys/dev/iir/iir_ctrl.c371
-rw-r--r--sys/dev/iir/iir_pci.c476
-rw-r--r--sys/i386/conf/GENERIC2
-rw-r--r--sys/i386/conf/NOTES9
-rw-r--r--sys/modules/Makefile1
-rw-r--r--sys/modules/iir/Makefile9
11 files changed, 3612 insertions, 0 deletions
diff --git a/sys/amd64/conf/GENERIC b/sys/amd64/conf/GENERIC
index f062d27..dec9a76 100644
--- a/sys/amd64/conf/GENERIC
+++ b/sys/amd64/conf/GENERIC
@@ -101,7 +101,9 @@ device stg # TMC 18C30/18C50
# RAID controllers interfaced to the SCSI subsystem
device asr # DPT SmartRAID V, VI and Adaptec SCSI RAID
+device ciss # Compaq Smart RAID 5*
device dpt # DPT Smartcache III, IV - See NOTES for options!
+device iir # Intel Integrated RAID
device mly # Mylex AcceleRAID/eXtremeRAID
# SCSI peripherals
diff --git a/sys/conf/NOTES b/sys/conf/NOTES
index ad17181..1311973 100644
--- a/sys/conf/NOTES
+++ b/sys/conf/NOTES
@@ -1546,6 +1546,15 @@ options DPT_ALLOW_MEMIO
device ciss
#
+# Intel Integrated RAID controllers.
+# This driver was developed and is maintained by Intel. Contacts
+# at Intel for this driver are
+# "Kannanthanam, Boji T" <boji.t.kannanthanam@intel.com> and
+# "Leubner, Achim" <achim.leubner@intel.com>.
+#
+device iir
+
+#
# Mylex AcceleRAID and eXtremeRAID controllers with v6 and later
# firmware. These controllers have a SCSI-like interface, and require
# the CAM infrastructure.
diff --git a/sys/conf/files b/sys/conf/files
index d435ae0..a70072a 100644
--- a/sys/conf/files
+++ b/sys/conf/files
@@ -369,6 +369,9 @@ dev/iicbus/iicbus.c optional iicbus
dev/iicbus/iiconf.c optional iicbus
dev/iicbus/iicsmb.c optional iicsmb \
dependency "iicbus_if.h"
+dev/iir/iir.c optional iir
+dev/iir/iir_ctrl.c optional iir
+dev/iir/iir_pci.c optional iir pci
dev/isp/isp.c optional isp
dev/isp/isp_freebsd.c optional isp
dev/isp/isp_target.c optional isp
diff --git a/sys/dev/iir/iir.c b/sys/dev/iir/iir.c
new file mode 100644
index 0000000..4e07233
--- /dev/null
+++ b/sys/dev/iir/iir.c
@@ -0,0 +1,2018 @@
+/* $FreeBSD$ */
+/*
+ * Copyright (c) 2000-01 Intel Corporation
+ * All Rights Reserved
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions
+ * are met:
+ * 1. Redistributions of source code must retain the above copyright
+ * notice, this list of conditions, and the following disclaimer,
+ * without modification, immediately at the beginning of the file.
+ * 2. Redistributions in binary form must reproduce the above copyright
+ * notice, this list of conditions and the following disclaimer in the
+ * documentation and/or other materials provided with the distribution.
+ * 3. The name of the author may not be used to endorse or promote products
+ * derived from this software without specific prior written permission.
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
+ * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
+ * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE FOR
+ * ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
+ * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
+ * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
+ * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
+ * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
+ * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
+ * SUCH DAMAGE.
+ */
+
+/*
+ * iir.c: SCSI dependant code for the Intel Integrated RAID Controller driver
+ *
+ * Written by: Achim Leubner <achim.leubner@intel.com>
+ * Fixes/Additions: Boji Tony Kannanthanam <boji.t.kannanthanam@intel.com>
+ *
+ * credits: Niklas Hallqvist; OpenBSD driver for the ICP Controllers.
+ * Mike Smith; Some driver source code.
+ * FreeBSD.ORG; Great O/S to work on and for.
+ *
+ * TODO:
+ */
+
+#ident "$Id: iir.c 1.2 2001/06/21 20:28:32 achim Exp $"
+
+#define _IIR_C_
+
+/* #include "opt_iir.h" */
+#include <sys/param.h>
+#include <sys/systm.h>
+#include <sys/eventhandler.h>
+#include <sys/malloc.h>
+#include <sys/kernel.h>
+#include <sys/bus.h>
+
+#include <stddef.h> /* For offsetof */
+
+#include <machine/bus_memio.h>
+#include <machine/bus_pio.h>
+#include <machine/bus.h>
+#include <machine/clock.h>
+#include <machine/stdarg.h>
+
+#include <cam/cam.h>
+#include <cam/cam_ccb.h>
+#include <cam/cam_sim.h>
+#include <cam/cam_xpt_sim.h>
+#include <cam/cam_debug.h>
+#include <cam/scsi/scsi_all.h>
+#include <cam/scsi/scsi_message.h>
+
+#include <vm/vm.h>
+#include <vm/pmap.h>
+
+#include <dev/iir/iir.h>
+
+struct gdt_softc *gdt_wait_gdt;
+int gdt_wait_index;
+
+#ifdef GDT_DEBUG
+int gdt_debug = GDT_DEBUG;
+#ifdef __SERIAL__
+#define MAX_SERBUF 160
+static void ser_init(void);
+static void ser_puts(char *str);
+static void ser_putc(int c);
+static char strbuf[MAX_SERBUF+1];
+#ifdef __COM2__
+#define COM_BASE 0x2f8
+#else
+#define COM_BASE 0x3f8
+#endif
+static void ser_init()
+{
+ unsigned port=COM_BASE;
+
+ outb(port+3, 0x80);
+ outb(port+1, 0);
+ /* 19200 Baud, if 9600: outb(12,port) */
+ outb(port, 6);
+ outb(port+3, 3);
+ outb(port+1, 0);
+}
+
+static void ser_puts(char *str)
+{
+ char *ptr;
+
+ ser_init();
+ for (ptr=str;*ptr;++ptr)
+ ser_putc((int)(*ptr));
+}
+
+static void ser_putc(int c)
+{
+ unsigned port=COM_BASE;
+
+ while ((inb(port+5) & 0x20)==0);
+ outb(port, c);
+ if (c==0x0a)
+ {
+ while ((inb(port+5) & 0x20)==0);
+ outb(port, 0x0d);
+ }
+}
+
+int ser_printf(const char *fmt, ...)
+{
+ va_list args;
+ int i;
+
+ va_start(args,fmt);
+ i = vsprintf(strbuf,fmt,args);
+ ser_puts(strbuf);
+ va_end(args);
+ return i;
+}
+#endif
+#endif
+
+/* The linked list of softc structures */
+struct gdt_softc_list gdt_softcs = TAILQ_HEAD_INITIALIZER(gdt_softcs);
+/* controller cnt. */
+int gdt_cnt = 0;
+/* event buffer */
+static gdt_evt_str ebuffer[GDT_MAX_EVENTS];
+static int elastidx, eoldidx;
+/* statistics */
+gdt_statist_t gdt_stat;
+
+/* Definitions for our use of the SIM private CCB area */
+#define ccb_sim_ptr spriv_ptr0
+#define ccb_priority spriv_field1
+
+static void iir_action(struct cam_sim *sim, union ccb *ccb);
+static void iir_poll(struct cam_sim *sim);
+static void iir_shutdown(void *arg, int howto);
+static void iir_timeout(void *arg);
+static void iir_watchdog(void *arg);
+
+static void gdt_eval_mapping(u_int32_t size, int *cyls, int *heads,
+ int *secs);
+static int gdt_internal_cmd(struct gdt_softc *gdt, struct gdt_ccb *gccb,
+ u_int8_t service, u_int16_t opcode,
+ u_int32_t arg1, u_int32_t arg2, u_int32_t arg3);
+static int gdt_wait(struct gdt_softc *gdt, struct gdt_ccb *ccb,
+ int timeout);
+
+static struct gdt_ccb *gdt_get_ccb(struct gdt_softc *gdt);
+static u_int32_t gdt_ccb_vtop(struct gdt_softc *gdt,
+ struct gdt_ccb *gccb);
+
+static int gdt_sync_event(struct gdt_softc *gdt, int service,
+ u_int8_t index, struct gdt_ccb *gccb);
+static int gdt_async_event(struct gdt_softc *gdt, int service);
+static struct gdt_ccb *gdt_raw_cmd(struct gdt_softc *gdt,
+ union ccb *ccb, int *lock);
+static struct gdt_ccb *gdt_cache_cmd(struct gdt_softc *gdt,
+ union ccb *ccb, int *lock);
+static struct gdt_ccb *gdt_ioctl_cmd(struct gdt_softc *gdt,
+ gdt_ucmd_t *ucmd, int *lock);
+static void gdt_internal_cache_cmd(struct gdt_softc *gdt,union ccb *ccb);
+
+static void gdtmapmem(void *arg, bus_dma_segment_t *dm_segs,
+ int nseg, int error);
+static void gdtexecuteccb(void *arg, bus_dma_segment_t *dm_segs,
+ int nseg, int error);
+
+int
+iir_init(struct gdt_softc *gdt)
+{
+ u_int16_t cdev_cnt;
+ int i, id, drv_cyls, drv_hds, drv_secs;
+ struct gdt_ccb *gccb;
+
+ GDT_DPRINTF(GDT_D_DEBUG, ("iir_init()\n"));
+
+ gdt->sc_state = GDT_POLLING;
+ gdt_clear_events();
+ bzero(&gdt_stat, sizeof(gdt_statist_t));
+
+ SLIST_INIT(&gdt->sc_free_gccb);
+ SLIST_INIT(&gdt->sc_pending_gccb);
+ TAILQ_INIT(&gdt->sc_ccb_queue);
+ TAILQ_INIT(&gdt->sc_ucmd_queue);
+ TAILQ_INSERT_TAIL(&gdt_softcs, gdt, links);
+
+ /* DMA tag for mapping buffers into device visible space. */
+ if (bus_dma_tag_create(gdt->sc_parent_dmat, /*alignment*/1, /*boundary*/0,
+ /*lowaddr*/BUS_SPACE_MAXADDR,
+ /*highaddr*/BUS_SPACE_MAXADDR,
+ /*filter*/NULL, /*filterarg*/NULL,
+ /*maxsize*/MAXBSIZE, /*nsegments*/GDT_MAXSG,
+ /*maxsegsz*/BUS_SPACE_MAXSIZE_32BIT,
+ /*flags*/BUS_DMA_ALLOCNOW,
+ &gdt->sc_buffer_dmat) != 0) {
+ printf("iir%d: bus_dma_tag_create(...,gdt->sc_buffer_dmat) failed\n",
+ gdt->sc_hanum);
+ return (1);
+ }
+ gdt->sc_init_level++;
+
+ /* DMA tag for our ccb structures */
+ if (bus_dma_tag_create(gdt->sc_parent_dmat, /*alignment*/1, /*boundary*/0,
+ /*lowaddr*/BUS_SPACE_MAXADDR,
+ /*highaddr*/BUS_SPACE_MAXADDR,
+ /*filter*/NULL, /*filterarg*/NULL,
+ GDT_MAXCMDS * sizeof(struct gdt_ccb),
+ /*nsegments*/1,
+ /*maxsegsz*/BUS_SPACE_MAXSIZE_32BIT,
+ /*flags*/0, &gdt->sc_gccb_dmat) != 0) {
+ printf("iir%d: bus_dma_tag_create(...,gdt->sc_gccb_dmat) failed\n",
+ gdt->sc_hanum);
+ return (1);
+ }
+ gdt->sc_init_level++;
+
+ /* Allocation for our ccbs */
+ if (bus_dmamem_alloc(gdt->sc_gccb_dmat, (void **)&gdt->sc_gccbs,
+ BUS_DMA_NOWAIT, &gdt->sc_gccb_dmamap) != 0) {
+ printf("iir%d: bus_dmamem_alloc(...,&gdt->sc_gccbs,...) failed\n",
+ gdt->sc_hanum);
+ return (1);
+ }
+ gdt->sc_init_level++;
+
+ /* And permanently map them */
+ bus_dmamap_load(gdt->sc_gccb_dmat, gdt->sc_gccb_dmamap,
+ gdt->sc_gccbs, GDT_MAXCMDS * sizeof(struct gdt_ccb),
+ gdtmapmem, &gdt->sc_gccb_busbase, /*flags*/0);
+ gdt->sc_init_level++;
+
+ /* Clear them out. */
+ bzero(gdt->sc_gccbs, GDT_MAXCMDS * sizeof(struct gdt_ccb));
+
+ /* Initialize the ccbs */
+ for (i = GDT_MAXCMDS-1; i >= 0; i--) {
+ gdt->sc_gccbs[i].gc_cmd_index = i + 2;
+ gdt->sc_gccbs[i].gc_flags = GDT_GCF_UNUSED;
+ gdt->sc_gccbs[i].gc_map_flag = FALSE;
+ if (bus_dmamap_create(gdt->sc_buffer_dmat, /*flags*/0,
+ &gdt->sc_gccbs[i].gc_dmamap) != 0)
+ return(1);
+ gdt->sc_gccbs[i].gc_map_flag = TRUE;
+ SLIST_INSERT_HEAD(&gdt->sc_free_gccb, &gdt->sc_gccbs[i], sle);
+ }
+ gdt->sc_init_level++;
+
+ /* create the control device */
+ gdt->sc_dev = gdt_make_dev(gdt->sc_hanum);
+
+ /* allocate ccb for gdt_internal_cmd() */
+ gccb = gdt_get_ccb(gdt);
+ if (gccb == NULL) {
+ printf("iir%d: No free command index found\n",
+ gdt->sc_hanum);
+ return (1);
+ }
+
+ if (!gdt_internal_cmd(gdt, gccb, GDT_SCREENSERVICE, GDT_INIT,
+ 0, 0, 0)) {
+ printf("iir%d: Screen service initialization error %d\n",
+ gdt->sc_hanum, gdt->sc_status);
+ gdt_free_ccb(gdt, gccb);
+ return (1);
+ }
+
+ if (!gdt_internal_cmd(gdt, gccb, GDT_CACHESERVICE, GDT_INIT,
+ GDT_LINUX_OS, 0, 0)) {
+ printf("iir%d: Cache service initialization error %d\n",
+ gdt->sc_hanum, gdt->sc_status);
+ gdt_free_ccb(gdt, gccb);
+ return (1);
+ }
+ gdt_internal_cmd(gdt, gccb, GDT_CACHESERVICE, GDT_UNFREEZE_IO,
+ 0, 0, 0);
+
+ if (!gdt_internal_cmd(gdt, gccb, GDT_CACHESERVICE, GDT_MOUNT,
+ 0xffff, 1, 0)) {
+ printf("iir%d: Cache service mount error %d\n",
+ gdt->sc_hanum, gdt->sc_status);
+ gdt_free_ccb(gdt, gccb);
+ return (1);
+ }
+
+ if (!gdt_internal_cmd(gdt, gccb, GDT_CACHESERVICE, GDT_INIT,
+ GDT_LINUX_OS, 0, 0)) {
+ printf("iir%d: Cache service post-mount initialization error %d\n",
+ gdt->sc_hanum, gdt->sc_status);
+ gdt_free_ccb(gdt, gccb);
+ return (1);
+ }
+ cdev_cnt = (u_int16_t)gdt->sc_info;
+ gdt->sc_fw_vers = gdt->sc_service;
+
+ /* Detect number of buses */
+ gdt_enc32(gccb->gc_scratch + GDT_IOC_VERSION, GDT_IOC_NEWEST);
+ gccb->gc_scratch[GDT_IOC_LIST_ENTRIES] = GDT_MAXBUS;
+ gccb->gc_scratch[GDT_IOC_FIRST_CHAN] = 0;
+ gccb->gc_scratch[GDT_IOC_LAST_CHAN] = GDT_MAXBUS - 1;
+ gdt_enc32(gccb->gc_scratch + GDT_IOC_LIST_OFFSET, GDT_IOC_HDR_SZ);
+ if (gdt_internal_cmd(gdt, gccb, GDT_CACHESERVICE, GDT_IOCTL,
+ GDT_IOCHAN_RAW_DESC, GDT_INVALID_CHANNEL,
+ GDT_IOC_HDR_SZ + GDT_MAXBUS * GDT_RAWIOC_SZ)) {
+ gdt->sc_bus_cnt = gccb->gc_scratch[GDT_IOC_CHAN_COUNT];
+ for (i = 0; i < gdt->sc_bus_cnt; i++) {
+ id = gccb->gc_scratch[GDT_IOC_HDR_SZ +
+ i * GDT_RAWIOC_SZ + GDT_RAWIOC_PROC_ID];
+ gdt->sc_bus_id[i] = id < GDT_MAXID_FC ? id : 0xff;
+ }
+ } else {
+ /* New method failed, use fallback. */
+ for (i = 0; i < GDT_MAXBUS; i++) {
+ gdt_enc32(gccb->gc_scratch + GDT_GETCH_CHANNEL_NO, i);
+ if (!gdt_internal_cmd(gdt, gccb, GDT_CACHESERVICE, GDT_IOCTL,
+ GDT_SCSI_CHAN_CNT | GDT_L_CTRL_PATTERN,
+ GDT_IO_CHANNEL | GDT_INVALID_CHANNEL,
+ GDT_GETCH_SZ)) {
+ if (i == 0) {
+ printf("iir%d: Cannot get channel count, "
+ "error %d\n", gdt->sc_hanum, gdt->sc_status);
+ gdt_free_ccb(gdt, gccb);
+ return (1);
+ }
+ break;
+ }
+ gdt->sc_bus_id[i] =
+ (gccb->gc_scratch[GDT_GETCH_SIOP_ID] < GDT_MAXID_FC) ?
+ gccb->gc_scratch[GDT_GETCH_SIOP_ID] : 0xff;
+ }
+ gdt->sc_bus_cnt = i;
+ }
+ /* add one "virtual" channel for the host drives */
+ gdt->sc_virt_bus = gdt->sc_bus_cnt;
+ gdt->sc_bus_cnt++;
+
+ if (!gdt_internal_cmd(gdt, gccb, GDT_SCSIRAWSERVICE, GDT_INIT,
+ 0, 0, 0)) {
+ printf("iir%d: Raw service initialization error %d\n",
+ gdt->sc_hanum, gdt->sc_status);
+ gdt_free_ccb(gdt, gccb);
+ return (1);
+ }
+
+ /* Set/get features raw service (scatter/gather) */
+ gdt->sc_raw_feat = 0;
+ if (gdt_internal_cmd(gdt, gccb, GDT_SCSIRAWSERVICE, GDT_SET_FEAT,
+ GDT_SCATTER_GATHER, 0, 0)) {
+ if (gdt_internal_cmd(gdt, gccb, GDT_SCSIRAWSERVICE, GDT_GET_FEAT,
+ 0, 0, 0)) {
+ gdt->sc_raw_feat = gdt->sc_info;
+ if (!(gdt->sc_info & GDT_SCATTER_GATHER)) {
+ panic("iir%d: Scatter/Gather Raw Service "
+ "required but not supported!\n", gdt->sc_hanum);
+ gdt_free_ccb(gdt, gccb);
+ return (1);
+ }
+ }
+ }
+
+ /* Set/get features cache service (scatter/gather) */
+ gdt->sc_cache_feat = 0;
+ if (gdt_internal_cmd(gdt, gccb, GDT_CACHESERVICE, GDT_SET_FEAT,
+ 0, GDT_SCATTER_GATHER, 0)) {
+ if (gdt_internal_cmd(gdt, gccb, GDT_CACHESERVICE, GDT_GET_FEAT,
+ 0, 0, 0)) {
+ gdt->sc_cache_feat = gdt->sc_info;
+ if (!(gdt->sc_info & GDT_SCATTER_GATHER)) {
+ panic("iir%d: Scatter/Gather Cache Service "
+ "required but not supported!\n", gdt->sc_hanum);
+ gdt_free_ccb(gdt, gccb);
+ return (1);
+ }
+ }
+ }
+
+ /* Scan for cache devices */
+ for (i = 0; i < cdev_cnt && i < GDT_MAX_HDRIVES; i++) {
+ if (gdt_internal_cmd(gdt, gccb, GDT_CACHESERVICE, GDT_INFO,
+ i, 0, 0)) {
+ gdt->sc_hdr[i].hd_present = 1;
+ gdt->sc_hdr[i].hd_size = gdt->sc_info;
+
+ /*
+ * Evaluate mapping (sectors per head, heads per cyl)
+ */
+ gdt->sc_hdr[i].hd_size &= ~GDT_SECS32;
+ if (gdt->sc_info2 == 0)
+ gdt_eval_mapping(gdt->sc_hdr[i].hd_size,
+ &drv_cyls, &drv_hds, &drv_secs);
+ else {
+ drv_hds = gdt->sc_info2 & 0xff;
+ drv_secs = (gdt->sc_info2 >> 8) & 0xff;
+ drv_cyls = gdt->sc_hdr[i].hd_size / drv_hds /
+ drv_secs;
+ }
+ gdt->sc_hdr[i].hd_heads = drv_hds;
+ gdt->sc_hdr[i].hd_secs = drv_secs;
+ /* Round the size */
+ gdt->sc_hdr[i].hd_size = drv_cyls * drv_hds * drv_secs;
+
+ if (gdt_internal_cmd(gdt, gccb, GDT_CACHESERVICE,
+ GDT_DEVTYPE, i, 0, 0))
+ gdt->sc_hdr[i].hd_devtype = gdt->sc_info;
+ }
+ }
+
+ GDT_DPRINTF(GDT_D_INIT, ("dpmem %x %d-bus %d cache device%s\n",
+ gdt->sc_dpmembase,
+ gdt->sc_bus_cnt, cdev_cnt,
+ cdev_cnt == 1 ? "" : "s"));
+ gdt_free_ccb(gdt, gccb);
+
+ gdt_cnt++;
+ return (0);
+}
+
+void
+iir_free(struct gdt_softc *gdt)
+{
+ int i;
+
+ GDT_DPRINTF(GDT_D_INIT, ("iir_free()\n"));
+
+ switch (gdt->sc_init_level) {
+ default:
+ gdt_destroy_dev(gdt->sc_dev);
+ case 5:
+ for (i = GDT_MAXCMDS-1; i >= 0; i--)
+ if (gdt->sc_gccbs[i].gc_map_flag)
+ bus_dmamap_destroy(gdt->sc_buffer_dmat,
+ gdt->sc_gccbs[i].gc_dmamap);
+ bus_dmamap_unload(gdt->sc_gccb_dmat, gdt->sc_gccb_dmamap);
+ case 4:
+ bus_dmamem_free(gdt->sc_gccb_dmat, gdt->sc_gccbs, gdt->sc_gccb_dmamap);
+ case 3:
+ bus_dma_tag_destroy(gdt->sc_gccb_dmat);
+ case 2:
+ bus_dma_tag_destroy(gdt->sc_buffer_dmat);
+ case 1:
+ bus_dma_tag_destroy(gdt->sc_parent_dmat);
+ case 0:
+ break;
+ }
+ TAILQ_REMOVE(&gdt_softcs, gdt, links);
+}
+
+void
+iir_attach(struct gdt_softc *gdt)
+{
+ struct cam_devq *devq;
+ int i;
+
+ GDT_DPRINTF(GDT_D_INIT, ("iir_attach()\n"));
+
+ /*
+ * Create the device queue for our SIM.
+ */
+ devq = cam_simq_alloc(GDT_MAXCMDS);
+ if (devq == NULL)
+ return;
+
+ for (i = 0; i < gdt->sc_bus_cnt; i++) {
+ /*
+ * Construct our SIM entry
+ */
+ gdt->sims[i] = cam_sim_alloc(iir_action, iir_poll, "iir",
+ gdt, gdt->sc_hanum, /*untagged*/2,
+ /*tagged*/GDT_MAXCMDS, devq);
+ if (xpt_bus_register(gdt->sims[i], i) != CAM_SUCCESS) {
+ cam_sim_free(gdt->sims[i], /*free_devq*/i == 0);
+ break;
+ }
+
+ if (xpt_create_path(&gdt->paths[i], /*periph*/NULL,
+ cam_sim_path(gdt->sims[i]),
+ CAM_TARGET_WILDCARD,
+ CAM_LUN_WILDCARD) != CAM_REQ_CMP) {
+ xpt_bus_deregister(cam_sim_path(gdt->sims[i]));
+ cam_sim_free(gdt->sims[i], /*free_devq*/i == 0);
+ break;
+ }
+ }
+ if (i > 0)
+ EVENTHANDLER_REGISTER(shutdown_final, iir_shutdown,
+ gdt, SHUTDOWN_PRI_DEFAULT);
+ /* iir_watchdog(gdt); */
+ gdt->sc_state = GDT_NORMAL;
+}
+
+static void
+gdt_eval_mapping(u_int32_t size, int *cyls, int *heads, int *secs)
+{
+ *cyls = size / GDT_HEADS / GDT_SECS;
+ if (*cyls < GDT_MAXCYLS) {
+ *heads = GDT_HEADS;
+ *secs = GDT_SECS;
+ } else {
+ /* Too high for 64 * 32 */
+ *cyls = size / GDT_MEDHEADS / GDT_MEDSECS;
+ if (*cyls < GDT_MAXCYLS) {
+ *heads = GDT_MEDHEADS;
+ *secs = GDT_MEDSECS;
+ } else {
+ /* Too high for 127 * 63 */
+ *cyls = size / GDT_BIGHEADS / GDT_BIGSECS;
+ *heads = GDT_BIGHEADS;
+ *secs = GDT_BIGSECS;
+ }
+ }
+}
+
+static int
+gdt_wait(struct gdt_softc *gdt, struct gdt_ccb *gccb,
+ int timeout)
+{
+ int rv = 0;
+
+ GDT_DPRINTF(GDT_D_INIT,
+ ("gdt_wait(%p, %p, %d)\n", gdt, gccb, timeout));
+
+ gdt->sc_state |= GDT_POLL_WAIT;
+ do {
+ iir_intr(gdt);
+ if (gdt == gdt_wait_gdt &&
+ gccb->gc_cmd_index == gdt_wait_index) {
+ rv = 1;
+ break;
+ }
+ DELAY(1);
+ } while (--timeout);
+ gdt->sc_state &= ~GDT_POLL_WAIT;
+
+ while (gdt->sc_test_busy(gdt))
+ DELAY(1); /* XXX correct? */
+
+ return (rv);
+}
+
+static int
+gdt_internal_cmd(struct gdt_softc *gdt, struct gdt_ccb *gccb,
+ u_int8_t service, u_int16_t opcode,
+ u_int32_t arg1, u_int32_t arg2, u_int32_t arg3)
+{
+ int retries;
+
+ GDT_DPRINTF(GDT_D_CMD, ("gdt_internal_cmd(%p, %d, %d, %d, %d, %d)\n",
+ gdt, service, opcode, arg1, arg2, arg3));
+
+ bzero(gdt->sc_cmd, GDT_CMD_SZ);
+
+ for (retries = GDT_RETRIES; ; ) {
+ gccb->gc_service = service;
+ gccb->gc_flags = GDT_GCF_INTERNAL;
+
+ gdt->sc_set_sema0(gdt);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_COMMANDINDEX,
+ gccb->gc_cmd_index);
+ gdt_enc16(gdt->sc_cmd + GDT_CMD_OPCODE, opcode);
+
+ switch (service) {
+ case GDT_CACHESERVICE:
+ if (opcode == GDT_IOCTL) {
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION +
+ GDT_IOCTL_SUBFUNC, arg1);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION +
+ GDT_IOCTL_CHANNEL, arg2);
+ gdt_enc16(gdt->sc_cmd + GDT_CMD_UNION +
+ GDT_IOCTL_PARAM_SIZE, (u_int16_t)arg3);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_IOCTL_P_PARAM,
+ gdt_ccb_vtop(gdt, gccb) +
+ offsetof(struct gdt_ccb, gc_scratch[0]));
+ } else {
+ gdt_enc16(gdt->sc_cmd + GDT_CMD_UNION +
+ GDT_CACHE_DEVICENO, (u_int16_t)arg1);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION +
+ GDT_CACHE_BLOCKNO, arg2);
+ }
+ break;
+
+ case GDT_SCSIRAWSERVICE:
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION +
+ GDT_RAW_DIRECTION, arg1);
+ gdt->sc_cmd[GDT_CMD_UNION + GDT_RAW_BUS] =
+ (u_int8_t)arg2;
+ gdt->sc_cmd[GDT_CMD_UNION + GDT_RAW_TARGET] =
+ (u_int8_t)arg3;
+ gdt->sc_cmd[GDT_CMD_UNION + GDT_RAW_LUN] =
+ (u_int8_t)(arg3 >> 8);
+ }
+
+ gdt->sc_cmd_len = GDT_CMD_SZ;
+ gdt->sc_cmd_off = 0;
+ gdt->sc_cmd_cnt = 0;
+ gdt->sc_copy_cmd(gdt, gccb);
+ gdt->sc_release_event(gdt);
+ DELAY(20);
+ if (!gdt_wait(gdt, gccb, GDT_POLL_TIMEOUT))
+ return (0);
+ if (gdt->sc_status != GDT_S_BSY || --retries == 0)
+ break;
+ DELAY(1);
+ }
+ return (gdt->sc_status == GDT_S_OK);
+}
+
+static struct gdt_ccb *
+gdt_get_ccb(struct gdt_softc *gdt)
+{
+ struct gdt_ccb *gccb;
+ int lock;
+
+ GDT_DPRINTF(GDT_D_QUEUE, ("gdt_get_ccb(%p)\n", gdt));
+
+ lock = splcam();
+ gccb = SLIST_FIRST(&gdt->sc_free_gccb);
+ if (gccb != NULL) {
+ SLIST_REMOVE_HEAD(&gdt->sc_free_gccb, sle);
+ SLIST_INSERT_HEAD(&gdt->sc_pending_gccb, gccb, sle);
+ ++gdt_stat.cmd_index_act;
+ if (gdt_stat.cmd_index_act > gdt_stat.cmd_index_max)
+ gdt_stat.cmd_index_max = gdt_stat.cmd_index_act;
+ }
+ splx(lock);
+ return (gccb);
+}
+
+void
+gdt_free_ccb(struct gdt_softc *gdt, struct gdt_ccb *gccb)
+{
+ int lock;
+
+ GDT_DPRINTF(GDT_D_QUEUE, ("gdt_free_ccb(%p, %p)\n", gdt, gccb));
+
+ lock = splcam();
+ gccb->gc_flags = GDT_GCF_UNUSED;
+ SLIST_REMOVE(&gdt->sc_pending_gccb, gccb, gdt_ccb, sle);
+ SLIST_INSERT_HEAD(&gdt->sc_free_gccb, gccb, sle);
+ --gdt_stat.cmd_index_act;
+ splx(lock);
+ if (gdt->sc_state & GDT_SHUTDOWN)
+ wakeup(gccb);
+}
+
+static u_int32_t
+gdt_ccb_vtop(struct gdt_softc *gdt, struct gdt_ccb *gccb)
+{
+ return (gdt->sc_gccb_busbase
+ + (u_int32_t)((caddr_t)gccb - (caddr_t)gdt->sc_gccbs));
+}
+
+void
+gdt_next(struct gdt_softc *gdt)
+{
+ int lock;
+ union ccb *ccb;
+ gdt_ucmd_t *ucmd;
+ struct cam_sim *sim;
+ int bus, target, lun;
+ int next_cmd;
+
+ struct ccb_scsiio *csio;
+ struct ccb_hdr *ccbh;
+ struct gdt_ccb *gccb = NULL;
+ u_int8_t cmd;
+
+ GDT_DPRINTF(GDT_D_QUEUE, ("gdt_next(%p)\n", gdt));
+
+ lock = splcam();
+ if (gdt->sc_test_busy(gdt)) {
+ if (!(gdt->sc_state & GDT_POLLING)) {
+ splx(lock);
+ return;
+ }
+ while (gdt->sc_test_busy(gdt))
+ DELAY(1);
+ }
+
+ gdt->sc_cmd_cnt = gdt->sc_cmd_off = 0;
+ next_cmd = TRUE;
+ for (;;) {
+ /* I/Os in queue? controller ready? */
+ if (!TAILQ_FIRST(&gdt->sc_ucmd_queue) &&
+ !TAILQ_FIRST(&gdt->sc_ccb_queue))
+ break;
+
+ /* 1.: I/Os without ccb (IOCTLs) */
+ ucmd = TAILQ_FIRST(&gdt->sc_ucmd_queue);
+ if (ucmd != NULL) {
+ TAILQ_REMOVE(&gdt->sc_ucmd_queue, ucmd, links);
+ if ((gccb = gdt_ioctl_cmd(gdt, ucmd, &lock)) == NULL) {
+ TAILQ_INSERT_HEAD(&gdt->sc_ucmd_queue, ucmd, links);
+ break;
+ }
+ break;
+ /* wenn mehrere Kdos. zulassen: if (!gdt_polling) continue; */
+ }
+
+ /* 2.: I/Os with ccb */
+ ccb = (union ccb *)TAILQ_FIRST(&gdt->sc_ccb_queue);
+ /* ist dann immer != NULL, da oben getestet */
+ sim = (struct cam_sim *)ccb->ccb_h.ccb_sim_ptr;
+ bus = cam_sim_bus(sim);
+ target = ccb->ccb_h.target_id;
+ lun = ccb->ccb_h.target_lun;
+
+ TAILQ_REMOVE(&gdt->sc_ccb_queue, &ccb->ccb_h, sim_links.tqe);
+ --gdt_stat.req_queue_act;
+ /* ccb->ccb_h.func_code is XPT_SCSI_IO */
+ GDT_DPRINTF(GDT_D_QUEUE, ("XPT_SCSI_IO flags 0x%x)\n",
+ ccb->ccb_h.flags));
+ csio = &ccb->csio;
+ ccbh = &ccb->ccb_h;
+ cmd = csio->cdb_io.cdb_bytes[0];
+ /* Max CDB length is 12 bytes */
+ if (csio->cdb_len > 12) {
+ ccbh->status = CAM_REQ_INVALID;
+ --gdt_stat.io_count_act;
+ xpt_done(ccb);
+ } else if (bus != gdt->sc_virt_bus) {
+ /* raw service command */
+ if ((gccb = gdt_raw_cmd(gdt, ccb, &lock)) == NULL) {
+ TAILQ_INSERT_HEAD(&gdt->sc_ccb_queue, &ccb->ccb_h,
+ sim_links.tqe);
+ ++gdt_stat.req_queue_act;
+ if (gdt_stat.req_queue_act > gdt_stat.req_queue_max)
+ gdt_stat.req_queue_max = gdt_stat.req_queue_act;
+ next_cmd = FALSE;
+ }
+ } else if (target >= GDT_MAX_HDRIVES ||
+ !gdt->sc_hdr[target].hd_present || lun != 0) {
+ ccbh->status = CAM_SEL_TIMEOUT;
+ --gdt_stat.io_count_act;
+ xpt_done(ccb);
+ } else {
+ /* cache service command */
+ if (cmd == READ_6 || cmd == WRITE_6 ||
+ cmd == READ_10 || cmd == WRITE_10) {
+ if ((gccb = gdt_cache_cmd(gdt, ccb, &lock)) == NULL) {
+ TAILQ_INSERT_HEAD(&gdt->sc_ccb_queue, &ccb->ccb_h,
+ sim_links.tqe);
+ ++gdt_stat.req_queue_act;
+ if (gdt_stat.req_queue_act > gdt_stat.req_queue_max)
+ gdt_stat.req_queue_max = gdt_stat.req_queue_act;
+ next_cmd = FALSE;
+ }
+ } else {
+ splx(lock);
+ gdt_internal_cache_cmd(gdt, ccb);
+ lock = splcam();
+ }
+ }
+ if ((gdt->sc_state & GDT_POLLING) || !next_cmd)
+ break;
+ }
+ if (gdt->sc_cmd_cnt > 0)
+ gdt->sc_release_event(gdt);
+
+ splx(lock);
+
+ if ((gdt->sc_state & GDT_POLLING) && gdt->sc_cmd_cnt > 0) {
+ gdt_wait(gdt, gccb, GDT_POLL_TIMEOUT);
+ }
+}
+
+static struct gdt_ccb *
+gdt_raw_cmd(struct gdt_softc *gdt, union ccb *ccb, int *lock)
+{
+ struct gdt_ccb *gccb;
+ struct cam_sim *sim;
+
+ GDT_DPRINTF(GDT_D_CMD, ("gdt_raw_cmd(%p, %p)\n", gdt, ccb));
+
+ if (roundup(GDT_CMD_UNION + GDT_RAW_SZ, sizeof(u_int32_t)) +
+ gdt->sc_cmd_off + GDT_DPMEM_COMMAND_OFFSET >
+ gdt->sc_ic_all_size) {
+ GDT_DPRINTF(GDT_D_INVALID, ("iir%d: gdt_raw_cmd(): DPMEM overflow\n",
+ gdt->sc_hanum));
+ return (NULL);
+ }
+
+ bzero(gdt->sc_cmd, GDT_CMD_SZ);
+
+ gccb = gdt_get_ccb(gdt);
+ if (gccb == NULL) {
+ GDT_DPRINTF(GDT_D_INVALID, ("iir%d: No free command index found\n",
+ gdt->sc_hanum));
+ return (gccb);
+ }
+ sim = (struct cam_sim *)ccb->ccb_h.ccb_sim_ptr;
+ gccb->gc_ccb = ccb;
+ gccb->gc_service = GDT_SCSIRAWSERVICE;
+ gccb->gc_flags = GDT_GCF_SCSI;
+
+ if (gdt->sc_cmd_cnt == 0)
+ gdt->sc_set_sema0(gdt);
+ splx(*lock);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_COMMANDINDEX,
+ gccb->gc_cmd_index);
+ gdt_enc16(gdt->sc_cmd + GDT_CMD_OPCODE, GDT_WRITE);
+
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_RAW_DIRECTION,
+ (ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN ?
+ GDT_DATA_IN : GDT_DATA_OUT);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_RAW_SDLEN,
+ ccb->csio.dxfer_len);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_RAW_CLEN,
+ ccb->csio.cdb_len);
+ bcopy(ccb->csio.cdb_io.cdb_bytes, gdt->sc_cmd + GDT_CMD_UNION + GDT_RAW_CMD,
+ ccb->csio.cdb_len);
+ gdt->sc_cmd[GDT_CMD_UNION + GDT_RAW_TARGET] =
+ ccb->ccb_h.target_id;
+ gdt->sc_cmd[GDT_CMD_UNION + GDT_RAW_LUN] =
+ ccb->ccb_h.target_lun;
+ gdt->sc_cmd[GDT_CMD_UNION + GDT_RAW_BUS] =
+ cam_sim_bus(sim);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_RAW_SENSE_LEN,
+ sizeof(struct scsi_sense_data));
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_RAW_SENSE_DATA,
+ gdt_ccb_vtop(gdt, gccb) +
+ offsetof(struct gdt_ccb, gc_scratch[0]));
+
+ /*
+ * If we have any data to send with this command,
+ * map it into bus space.
+ */
+ /* Only use S/G if there is a transfer */
+ if ((ccb->ccb_h.flags & CAM_DIR_MASK) != CAM_DIR_NONE) {
+ if ((ccb->ccb_h.flags & CAM_SCATTER_VALID) == 0) {
+ if ((ccb->ccb_h.flags & CAM_DATA_PHYS) == 0) {
+ int s;
+ int error;
+
+ /* vorher unlock von splcam() ??? */
+ s = splsoftvm();
+ error =
+ bus_dmamap_load(gdt->sc_buffer_dmat,
+ gccb->gc_dmamap,
+ ccb->csio.data_ptr,
+ ccb->csio.dxfer_len,
+ gdtexecuteccb,
+ gccb, /*flags*/0);
+ if (error == EINPROGRESS) {
+ xpt_freeze_simq(sim, 1);
+ gccb->gc_state |= CAM_RELEASE_SIMQ;
+ }
+ splx(s);
+ } else {
+ struct bus_dma_segment seg;
+
+ /* Pointer to physical buffer */
+ seg.ds_addr =
+ (bus_addr_t)ccb->csio.data_ptr;
+ seg.ds_len = ccb->csio.dxfer_len;
+ gdtexecuteccb(gccb, &seg, 1, 0);
+ }
+ } else {
+ struct bus_dma_segment *segs;
+
+ if ((ccb->ccb_h.flags & CAM_DATA_PHYS) != 0)
+ panic("iir%d: iir_action - Physical "
+ "segment pointers unsupported", gdt->sc_hanum);
+
+ if ((ccb->ccb_h.flags & CAM_SG_LIST_PHYS)==0)
+ panic("iir%d: iir_action - Virtual "
+ "segment addresses unsupported", gdt->sc_hanum);
+
+ /* Just use the segments provided */
+ segs = (struct bus_dma_segment *)ccb->csio.data_ptr;
+ gdtexecuteccb(gccb, segs, ccb->csio.sglist_cnt, 0);
+ }
+ } else {
+ gdtexecuteccb(gccb, NULL, 0, 0);
+ }
+
+ *lock = splcam();
+ return (gccb);
+}
+
+static struct gdt_ccb *
+gdt_cache_cmd(struct gdt_softc *gdt, union ccb *ccb, int *lock)
+{
+ struct gdt_ccb *gccb;
+ struct cam_sim *sim;
+ u_int8_t *cmdp;
+ u_int16_t opcode;
+ u_int32_t blockno, blockcnt;
+
+ GDT_DPRINTF(GDT_D_CMD, ("gdt_cache_cmd(%p, %p)\n", gdt, ccb));
+
+ if (roundup(GDT_CMD_UNION + GDT_CACHE_SZ, sizeof(u_int32_t)) +
+ gdt->sc_cmd_off + GDT_DPMEM_COMMAND_OFFSET >
+ gdt->sc_ic_all_size) {
+ GDT_DPRINTF(GDT_D_INVALID, ("iir%d: gdt_cache_cmd(): DPMEM overflow\n",
+ gdt->sc_hanum));
+ return (NULL);
+ }
+
+ bzero(gdt->sc_cmd, GDT_CMD_SZ);
+
+ gccb = gdt_get_ccb(gdt);
+ if (gccb == NULL) {
+ GDT_DPRINTF(GDT_D_DEBUG, ("iir%d: No free command index found\n",
+ gdt->sc_hanum));
+ return (gccb);
+ }
+ sim = (struct cam_sim *)ccb->ccb_h.ccb_sim_ptr;
+ gccb->gc_ccb = ccb;
+ gccb->gc_service = GDT_CACHESERVICE;
+ gccb->gc_flags = GDT_GCF_SCSI;
+
+ if (gdt->sc_cmd_cnt == 0)
+ gdt->sc_set_sema0(gdt);
+ splx(*lock);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_COMMANDINDEX,
+ gccb->gc_cmd_index);
+ cmdp = ccb->csio.cdb_io.cdb_bytes;
+ opcode = (*cmdp == WRITE_6 || *cmdp == WRITE_10) ? GDT_WRITE : GDT_READ;
+ if ((gdt->sc_state & GDT_SHUTDOWN) && opcode == GDT_WRITE)
+ opcode = GDT_WRITE_THR;
+ gdt_enc16(gdt->sc_cmd + GDT_CMD_OPCODE, opcode);
+
+ gdt_enc16(gdt->sc_cmd + GDT_CMD_UNION + GDT_CACHE_DEVICENO,
+ ccb->ccb_h.target_id);
+ if (ccb->csio.cdb_len == 6) {
+ struct scsi_rw_6 *rw = (struct scsi_rw_6 *)cmdp;
+ blockno = scsi_3btoul(rw->addr) & ((SRW_TOPADDR<<16) | 0xffff);
+ blockcnt = rw->length ? rw->length : 0x100;
+ } else {
+ struct scsi_rw_10 *rw = (struct scsi_rw_10 *)cmdp;
+ blockno = scsi_4btoul(rw->addr);
+ blockcnt = scsi_2btoul(rw->length);
+ }
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_CACHE_BLOCKNO,
+ blockno);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_CACHE_BLOCKCNT,
+ blockcnt);
+
+ /*
+ * If we have any data to send with this command,
+ * map it into bus space.
+ */
+ /* Only use S/G if there is a transfer */
+ if ((ccb->ccb_h.flags & CAM_SCATTER_VALID) == 0) {
+ if ((ccb->ccb_h.flags & CAM_DATA_PHYS) == 0) {
+ int s;
+ int error;
+
+ /* vorher unlock von splcam() ??? */
+ s = splsoftvm();
+ error =
+ bus_dmamap_load(gdt->sc_buffer_dmat,
+ gccb->gc_dmamap,
+ ccb->csio.data_ptr,
+ ccb->csio.dxfer_len,
+ gdtexecuteccb,
+ gccb, /*flags*/0);
+ if (error == EINPROGRESS) {
+ xpt_freeze_simq(sim, 1);
+ gccb->gc_state |= CAM_RELEASE_SIMQ;
+ }
+ splx(s);
+ } else {
+ struct bus_dma_segment seg;
+
+ /* Pointer to physical buffer */
+ seg.ds_addr =
+ (bus_addr_t)ccb->csio.data_ptr;
+ seg.ds_len = ccb->csio.dxfer_len;
+ gdtexecuteccb(gccb, &seg, 1, 0);
+ }
+ } else {
+ struct bus_dma_segment *segs;
+
+ if ((ccb->ccb_h.flags & CAM_DATA_PHYS) != 0)
+ panic("iir%d: iir_action - Physical "
+ "segment pointers unsupported", gdt->sc_hanum);
+
+ if ((ccb->ccb_h.flags & CAM_SG_LIST_PHYS)==0)
+ panic("iir%d: iir_action - Virtual "
+ "segment addresses unsupported", gdt->sc_hanum);
+
+ /* Just use the segments provided */
+ segs = (struct bus_dma_segment *)ccb->csio.data_ptr;
+ gdtexecuteccb(gccb, segs, ccb->csio.sglist_cnt, 0);
+ }
+
+ *lock = splcam();
+ return (gccb);
+}
+
+static struct gdt_ccb *
+gdt_ioctl_cmd(struct gdt_softc *gdt, gdt_ucmd_t *ucmd, int *lock)
+{
+ struct gdt_ccb *gccb;
+ u_int32_t cnt;
+
+ GDT_DPRINTF(GDT_D_DEBUG, ("gdt_ioctl_cmd(%p, %p)\n", gdt, ucmd));
+
+ bzero(gdt->sc_cmd, GDT_CMD_SZ);
+
+ gccb = gdt_get_ccb(gdt);
+ if (gccb == NULL) {
+ GDT_DPRINTF(GDT_D_DEBUG, ("iir%d: No free command index found\n",
+ gdt->sc_hanum));
+ return (gccb);
+ }
+ gccb->gc_ucmd = ucmd;
+ gccb->gc_service = ucmd->service;
+ gccb->gc_flags = GDT_GCF_IOCTL;
+
+ /* check DPMEM space, copy data buffer from user space */
+ if (ucmd->service == GDT_CACHESERVICE) {
+ if (ucmd->OpCode == GDT_IOCTL) {
+ gdt->sc_cmd_len = roundup(GDT_CMD_UNION + GDT_IOCTL_SZ,
+ sizeof(u_int32_t));
+ cnt = ucmd->u.ioctl.param_size;
+ if (cnt > GDT_SCRATCH_SZ) {
+ printf("iir%d: Scratch buffer too small (%d/%d)\n",
+ gdt->sc_hanum, GDT_SCRATCH_SZ, cnt);
+ gdt_free_ccb(gdt, gccb);
+ return (NULL);
+ }
+ } else {
+ gdt->sc_cmd_len = roundup(GDT_CMD_UNION + GDT_CACHE_SG_LST +
+ GDT_SG_SZ, sizeof(u_int32_t));
+ cnt = ucmd->u.cache.BlockCnt * GDT_SECTOR_SIZE;
+ if (cnt > GDT_SCRATCH_SZ) {
+ printf("iir%d: Scratch buffer too small (%d/%d)\n",
+ gdt->sc_hanum, GDT_SCRATCH_SZ, cnt);
+ gdt_free_ccb(gdt, gccb);
+ return (NULL);
+ }
+ }
+ } else {
+ gdt->sc_cmd_len = roundup(GDT_CMD_UNION + GDT_RAW_SG_LST +
+ GDT_SG_SZ, sizeof(u_int32_t));
+ cnt = ucmd->u.raw.sdlen;
+ if (cnt + ucmd->u.raw.sense_len > GDT_SCRATCH_SZ) {
+ printf("iir%d: Scratch buffer too small (%d/%d)\n",
+ gdt->sc_hanum, GDT_SCRATCH_SZ, cnt + ucmd->u.raw.sense_len);
+ gdt_free_ccb(gdt, gccb);
+ return (NULL);
+ }
+ }
+ if (cnt != 0)
+ bcopy(ucmd->data, gccb->gc_scratch, cnt);
+
+ if (gdt->sc_cmd_off + gdt->sc_cmd_len + GDT_DPMEM_COMMAND_OFFSET >
+ gdt->sc_ic_all_size) {
+ GDT_DPRINTF(GDT_D_INVALID, ("iir%d: gdt_ioctl_cmd(): DPMEM overflow\n",
+ gdt->sc_hanum));
+ gdt_free_ccb(gdt, gccb);
+ return (NULL);
+ }
+
+ if (gdt->sc_cmd_cnt == 0)
+ gdt->sc_set_sema0(gdt);
+ splx(*lock);
+
+ /* fill cmd structure */
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_COMMANDINDEX,
+ gccb->gc_cmd_index);
+ gdt_enc16(gdt->sc_cmd + GDT_CMD_OPCODE,
+ ucmd->OpCode);
+
+ if (ucmd->service == GDT_CACHESERVICE) {
+ if (ucmd->OpCode == GDT_IOCTL) {
+ /* IOCTL */
+ gdt_enc16(gdt->sc_cmd + GDT_CMD_UNION + GDT_IOCTL_PARAM_SIZE,
+ ucmd->u.ioctl.param_size);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_IOCTL_SUBFUNC,
+ ucmd->u.ioctl.subfunc);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_IOCTL_CHANNEL,
+ ucmd->u.ioctl.channel);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_IOCTL_P_PARAM,
+ gdt_ccb_vtop(gdt, gccb) +
+ offsetof(struct gdt_ccb, gc_scratch[0]));
+ } else {
+ /* cache service command */
+ gdt_enc16(gdt->sc_cmd + GDT_CMD_UNION + GDT_CACHE_DEVICENO,
+ ucmd->u.cache.DeviceNo);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_CACHE_BLOCKNO,
+ ucmd->u.cache.BlockNo);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_CACHE_BLOCKCNT,
+ ucmd->u.cache.BlockCnt);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_CACHE_DESTADDR,
+ 0xffffffffUL);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_CACHE_SG_CANZ,
+ 1);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_CACHE_SG_LST +
+ GDT_SG_PTR, gdt_ccb_vtop(gdt, gccb) +
+ offsetof(struct gdt_ccb, gc_scratch[0]));
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_CACHE_SG_LST +
+ GDT_SG_LEN, ucmd->u.cache.BlockCnt * GDT_SECTOR_SIZE);
+ }
+ } else {
+ /* raw service command */
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_RAW_DIRECTION,
+ ucmd->u.raw.direction);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_RAW_SDATA,
+ 0xffffffffUL);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_RAW_SDLEN,
+ ucmd->u.raw.sdlen);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_RAW_CLEN,
+ ucmd->u.raw.clen);
+ bcopy(ucmd->u.raw.cmd, gdt->sc_cmd + GDT_CMD_UNION + GDT_RAW_CMD,
+ 12);
+ gdt->sc_cmd[GDT_CMD_UNION + GDT_RAW_TARGET] =
+ ucmd->u.raw.target;
+ gdt->sc_cmd[GDT_CMD_UNION + GDT_RAW_LUN] =
+ ucmd->u.raw.lun;
+ gdt->sc_cmd[GDT_CMD_UNION + GDT_RAW_BUS] =
+ ucmd->u.raw.bus;
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_RAW_SENSE_LEN,
+ ucmd->u.raw.sense_len);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_RAW_SENSE_DATA,
+ gdt_ccb_vtop(gdt, gccb) +
+ offsetof(struct gdt_ccb, gc_scratch[ucmd->u.raw.sdlen]));
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_RAW_SG_RANZ,
+ 1);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_RAW_SG_LST +
+ GDT_SG_PTR, gdt_ccb_vtop(gdt, gccb) +
+ offsetof(struct gdt_ccb, gc_scratch[0]));
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_RAW_SG_LST +
+ GDT_SG_LEN, ucmd->u.raw.sdlen);
+ }
+
+ *lock = splcam();
+ gdt_stat.sg_count_act = 1;
+ gdt->sc_copy_cmd(gdt, gccb);
+ return (gccb);
+}
+
+static void
+gdt_internal_cache_cmd(struct gdt_softc *gdt,union ccb *ccb)
+{
+ int t;
+
+ t = ccb->ccb_h.target_id;
+ GDT_DPRINTF(GDT_D_CMD, ("gdt_internal_cache_cmd(%p, %p, 0x%x, %d)\n",
+ gdt, ccb, ccb->csio.cdb_io.cdb_bytes[0], t));
+
+ switch (ccb->csio.cdb_io.cdb_bytes[0]) {
+ case TEST_UNIT_READY:
+ case START_STOP:
+ break;
+ case REQUEST_SENSE:
+ GDT_DPRINTF(GDT_D_MISC, ("REQUEST_SENSE\n"));
+ break;
+ case INQUIRY:
+ {
+ struct scsi_inquiry_data *inq;
+
+ inq = (struct scsi_inquiry_data *)ccb->csio.data_ptr;
+ bzero(inq, sizeof(struct scsi_inquiry_data));
+ inq->device = (gdt->sc_hdr[t].hd_devtype & 4) ?
+ T_CDROM : T_DIRECT;
+ inq->dev_qual2 = (gdt->sc_hdr[t].hd_devtype & 1) ? 0x80 : 0;
+ inq->version = SCSI_REV_2;
+ inq->response_format = 2;
+ inq->additional_length = 32;
+ inq->flags = SID_CmdQue | SID_Sync;
+ strcpy(inq->vendor, "IIR ");
+ sprintf(inq->product, "Host Drive #%02d", t);
+ strcpy(inq->revision, " ");
+ break;
+ }
+ case MODE_SENSE_6:
+ {
+ struct mpd_data {
+ struct scsi_mode_hdr_6 hd;
+ struct scsi_mode_block_descr bd;
+ struct scsi_control_page cp;
+ } *mpd;
+ u_int8_t page;
+
+ mpd = (struct mpd_data *)ccb->csio.data_ptr;
+ bzero(mpd, sizeof(struct mpd_data));
+ mpd->hd.datalen = sizeof(struct scsi_mode_hdr_6) +
+ sizeof(struct scsi_mode_block_descr);
+ mpd->hd.dev_specific = (gdt->sc_hdr[t].hd_devtype & 2) ? 0x80 : 0;
+ mpd->hd.block_descr_len = sizeof(struct scsi_mode_block_descr);
+ mpd->bd.block_len[0] = (GDT_SECTOR_SIZE & 0x00ff0000) >> 16;
+ mpd->bd.block_len[1] = (GDT_SECTOR_SIZE & 0x0000ff00) >> 8;
+ mpd->bd.block_len[2] = (GDT_SECTOR_SIZE & 0x000000ff);
+ page=((struct scsi_mode_sense_6 *)ccb->csio.cdb_io.cdb_bytes)->page;
+ switch (page) {
+ default:
+ GDT_DPRINTF(GDT_D_MISC, ("MODE_SENSE_6: page 0x%x\n", page));
+ break;
+ }
+ break;
+ }
+ case READ_CAPACITY:
+ {
+ struct scsi_read_capacity_data *rcd;
+
+ rcd = (struct scsi_read_capacity_data *)ccb->csio.data_ptr;
+ bzero(rcd, sizeof(struct scsi_read_capacity_data));
+ scsi_ulto4b(gdt->sc_hdr[t].hd_size - 1, rcd->addr);
+ scsi_ulto4b(GDT_SECTOR_SIZE, rcd->length);
+ break;
+ }
+ default:
+ GDT_DPRINTF(GDT_D_MISC, ("gdt_internal_cache_cmd(%d) unknown\n",
+ ccb->csio.cdb_io.cdb_bytes[0]));
+ break;
+ }
+ ccb->ccb_h.status = CAM_REQ_CMP;
+ --gdt_stat.io_count_act;
+ xpt_done(ccb);
+}
+
+static void
+gdtmapmem(void *arg, bus_dma_segment_t *dm_segs, int nseg, int error)
+{
+ bus_addr_t *busaddrp;
+
+ busaddrp = (bus_addr_t *)arg;
+ *busaddrp = dm_segs->ds_addr;
+}
+
+static void
+gdtexecuteccb(void *arg, bus_dma_segment_t *dm_segs, int nseg, int error)
+{
+ struct gdt_ccb *gccb;
+ union ccb *ccb;
+ struct gdt_softc *gdt;
+ int i, lock;
+
+ lock = splcam();
+
+ gccb = (struct gdt_ccb *)arg;
+ ccb = gccb->gc_ccb;
+ gdt = cam_sim_softc((struct cam_sim *)ccb->ccb_h.ccb_sim_ptr);
+
+ GDT_DPRINTF(GDT_D_CMD, ("gdtexecuteccb(%p, %p, %p, %d, %d)\n",
+ gdt, gccb, dm_segs, nseg, error));
+ gdt_stat.sg_count_act = nseg;
+ if (nseg > gdt_stat.sg_count_max)
+ gdt_stat.sg_count_max = nseg;
+
+ /* Copy the segments into our SG list */
+ if (gccb->gc_service == GDT_CACHESERVICE) {
+ for (i = 0; i < nseg; ++i) {
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_CACHE_SG_LST +
+ i * GDT_SG_SZ + GDT_SG_PTR, dm_segs->ds_addr);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_CACHE_SG_LST +
+ i * GDT_SG_SZ + GDT_SG_LEN, dm_segs->ds_len);
+ dm_segs++;
+ }
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_CACHE_SG_CANZ,
+ nseg);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_CACHE_DESTADDR,
+ 0xffffffffUL);
+
+ gdt->sc_cmd_len = roundup(GDT_CMD_UNION + GDT_CACHE_SG_LST +
+ nseg * GDT_SG_SZ, sizeof(u_int32_t));
+ } else {
+ for (i = 0; i < nseg; ++i) {
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_RAW_SG_LST +
+ i * GDT_SG_SZ + GDT_SG_PTR, dm_segs->ds_addr);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_RAW_SG_LST +
+ i * GDT_SG_SZ + GDT_SG_LEN, dm_segs->ds_len);
+ dm_segs++;
+ }
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_RAW_SG_RANZ,
+ nseg);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_RAW_SDATA,
+ 0xffffffffUL);
+
+ gdt->sc_cmd_len = roundup(GDT_CMD_UNION + GDT_RAW_SG_LST +
+ nseg * GDT_SG_SZ, sizeof(u_int32_t));
+ }
+
+ if (nseg != 0) {
+ bus_dmasync_op_t op;
+
+ if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN)
+ op = BUS_DMASYNC_PREREAD;
+ else
+ op = BUS_DMASYNC_PREWRITE;
+ bus_dmamap_sync(gdt->sc_buffer_dmat, gccb->gc_dmamap, op);
+ }
+
+ /* We must NOT abort the command here if CAM_REQ_INPROG is not set,
+ * because command semaphore is already set!
+ */
+
+ ccb->ccb_h.status |= CAM_SIM_QUEUED;
+ /* timeout handling */
+ ccb->ccb_h.timeout_ch =
+ timeout(iir_timeout, (caddr_t)gccb,
+ (ccb->ccb_h.timeout * hz) / 1000);
+
+ gdt->sc_copy_cmd(gdt, gccb);
+ splx(lock);
+}
+
+
+static void
+iir_action( struct cam_sim *sim, union ccb *ccb )
+{
+ struct gdt_softc *gdt;
+ int lock, bus, target, lun;
+
+ gdt = (struct gdt_softc *)cam_sim_softc( sim );
+ ccb->ccb_h.ccb_sim_ptr = sim;
+ bus = cam_sim_bus(sim);
+ target = ccb->ccb_h.target_id;
+ lun = ccb->ccb_h.target_lun;
+ GDT_DPRINTF(GDT_D_CMD,
+ ("iir_action(%p) func 0x%x cmd 0x%x bus %d target %d lun %d\n",
+ gdt, ccb->ccb_h.func_code, ccb->csio.cdb_io.cdb_bytes[0],
+ bus, target, lun));
+ ++gdt_stat.io_count_act;
+ if (gdt_stat.io_count_act > gdt_stat.io_count_max)
+ gdt_stat.io_count_max = gdt_stat.io_count_act;
+
+ switch (ccb->ccb_h.func_code) {
+ case XPT_SCSI_IO:
+ lock = splcam();
+ TAILQ_INSERT_TAIL(&gdt->sc_ccb_queue, &ccb->ccb_h, sim_links.tqe);
+ ++gdt_stat.req_queue_act;
+ if (gdt_stat.req_queue_act > gdt_stat.req_queue_max)
+ gdt_stat.req_queue_max = gdt_stat.req_queue_act;
+ splx(lock);
+ gdt_next(gdt);
+ break;
+ case XPT_RESET_DEV: /* Bus Device Reset the specified SCSI device */
+ case XPT_ABORT: /* Abort the specified CCB */
+ /* XXX Implement */
+ ccb->ccb_h.status = CAM_REQ_INVALID;
+ --gdt_stat.io_count_act;
+ xpt_done(ccb);
+ break;
+ case XPT_SET_TRAN_SETTINGS:
+ ccb->ccb_h.status = CAM_FUNC_NOTAVAIL;
+ --gdt_stat.io_count_act;
+ xpt_done(ccb);
+ break;
+ case XPT_GET_TRAN_SETTINGS:
+ /* Get default/user set transfer settings for the target */
+ {
+ struct ccb_trans_settings *cts;
+ u_int target_mask;
+
+ cts = &ccb->cts;
+ target_mask = 0x01 << target;
+ if ((cts->flags & CCB_TRANS_USER_SETTINGS) != 0) {
+ cts->flags = CCB_TRANS_DISC_ENB|CCB_TRANS_TAG_ENB;
+ cts->bus_width = MSG_EXT_WDTR_BUS_16_BIT;
+ cts->sync_period = 25; /* 10MHz */
+ if (cts->sync_period != 0)
+ cts->sync_offset = 15;
+
+ cts->valid = CCB_TRANS_SYNC_RATE_VALID
+ | CCB_TRANS_SYNC_OFFSET_VALID
+ | CCB_TRANS_BUS_WIDTH_VALID
+ | CCB_TRANS_DISC_VALID
+ | CCB_TRANS_TQ_VALID;
+ ccb->ccb_h.status = CAM_REQ_CMP;
+ } else {
+ ccb->ccb_h.status = CAM_FUNC_NOTAVAIL;
+ }
+ --gdt_stat.io_count_act;
+ xpt_done(ccb);
+ break;
+ }
+ case XPT_CALC_GEOMETRY:
+ {
+ struct ccb_calc_geometry *ccg;
+ u_int32_t secs_per_cylinder;
+
+ ccg = &ccb->ccg;
+ ccg->heads = gdt->sc_hdr[target].hd_heads;
+ ccg->secs_per_track = gdt->sc_hdr[target].hd_secs;
+ secs_per_cylinder = ccg->heads * ccg->secs_per_track;
+ ccg->cylinders = ccg->volume_size / secs_per_cylinder;
+ ccb->ccb_h.status = CAM_REQ_CMP;
+ --gdt_stat.io_count_act;
+ xpt_done(ccb);
+ break;
+ }
+ case XPT_RESET_BUS: /* Reset the specified SCSI bus */
+ {
+ /* XXX Implement */
+ ccb->ccb_h.status = CAM_REQ_CMP;
+ --gdt_stat.io_count_act;
+ xpt_done(ccb);
+ break;
+ }
+ case XPT_TERM_IO: /* Terminate the I/O process */
+ /* XXX Implement */
+ ccb->ccb_h.status = CAM_REQ_INVALID;
+ --gdt_stat.io_count_act;
+ xpt_done(ccb);
+ break;
+ case XPT_PATH_INQ: /* Path routing inquiry */
+ {
+ struct ccb_pathinq *cpi = &ccb->cpi;
+
+ cpi->version_num = 1;
+ cpi->hba_inquiry = PI_SDTR_ABLE|PI_TAG_ABLE;
+ cpi->hba_inquiry |= PI_WIDE_16;
+ cpi->target_sprt = 1;
+ cpi->hba_misc = 0;
+ cpi->hba_eng_cnt = 0;
+ if (bus == gdt->sc_virt_bus)
+ cpi->max_target = GDT_MAX_HDRIVES - 1;
+ else if (gdt->sc_class & GDT_FC)
+ cpi->max_target = GDT_MAXID_FC - 1;
+ else
+ cpi->max_target = GDT_MAXID - 1;
+ cpi->max_lun = 7;
+ cpi->unit_number = cam_sim_unit(sim);
+ cpi->bus_id = bus;
+ cpi->initiator_id =
+ (bus == gdt->sc_virt_bus ? 127 : gdt->sc_bus_id[bus]);
+ cpi->base_transfer_speed = 3300;
+ strncpy(cpi->sim_vid, "FreeBSD", SIM_IDLEN);
+ strncpy(cpi->hba_vid, "Intel Corp.", HBA_IDLEN);
+ strncpy(cpi->dev_name, cam_sim_name(sim), DEV_IDLEN);
+ cpi->ccb_h.status = CAM_REQ_CMP;
+ --gdt_stat.io_count_act;
+ xpt_done(ccb);
+ break;
+ }
+ default:
+ GDT_DPRINTF(GDT_D_INVALID, ("gdt_next(%p) cmd 0x%x invalid\n",
+ gdt, ccb->ccb_h.func_code));
+ ccb->ccb_h.status = CAM_REQ_INVALID;
+ --gdt_stat.io_count_act;
+ xpt_done(ccb);
+ break;
+ }
+}
+
+static void
+iir_poll( struct cam_sim *sim )
+{
+ struct gdt_softc *gdt;
+
+ gdt = (struct gdt_softc *)cam_sim_softc( sim );
+ GDT_DPRINTF(GDT_D_CMD, ("iir_poll sim %p gdt %p\n", sim, gdt));
+ iir_intr(gdt);
+}
+
+static void
+iir_timeout(void *arg)
+{
+ GDT_DPRINTF(GDT_D_TIMEOUT, ("iir_timeout(%p)\n", arg));
+}
+
+static void
+iir_watchdog(void *arg)
+{
+ struct gdt_softc *gdt;
+
+ gdt = (struct gdt_softc *)arg;
+ GDT_DPRINTF(GDT_D_DEBUG, ("iir_watchdog(%p)\n", gdt));
+
+ {
+ int ccbs = 0, ucmds = 0, frees = 0, pends = 0;
+ struct gdt_ccb *p;
+ struct ccb_hdr *h;
+ struct gdt_ucmd *u;
+
+ for (h = TAILQ_FIRST(&gdt->sc_ccb_queue); h != NULL;
+ h = TAILQ_NEXT(h, sim_links.tqe))
+ ccbs++;
+ for (u = TAILQ_FIRST(&gdt->sc_ucmd_queue); u != NULL;
+ u = TAILQ_NEXT(u, links))
+ ucmds++;
+ for (p = SLIST_FIRST(&gdt->sc_free_gccb); p != NULL;
+ p = SLIST_NEXT(p, sle))
+ frees++;
+ for (p = SLIST_FIRST(&gdt->sc_pending_gccb); p != NULL;
+ p = SLIST_NEXT(p, sle))
+ pends++;
+
+ GDT_DPRINTF(GDT_D_TIMEOUT, ("ccbs %d ucmds %d frees %d pends %d\n",
+ ccbs, ucmds, frees, pends));
+ }
+
+ timeout(iir_watchdog, (caddr_t)gdt, hz * 15);
+}
+
+static void
+iir_shutdown( void *arg, int howto )
+{
+ struct gdt_softc *gdt;
+ struct gdt_ccb *gccb;
+ gdt_ucmd_t *ucmd;
+ int lock, i;
+
+ gdt = (struct gdt_softc *)arg;
+ GDT_DPRINTF(GDT_D_CMD, ("iir_shutdown(%p, %d)\n", gdt, howto));
+
+ printf("iir%d: Flushing all Host Drives. Please wait ... ",
+ gdt->sc_hanum);
+
+ /* allocate ucmd buffer */
+ ucmd = malloc(sizeof(gdt_ucmd_t), M_DEVBUF, M_NOWAIT);
+ if (ucmd == NULL) {
+ printf("iir%d: iir_shutdown(): Cannot allocate resource\n",
+ gdt->sc_hanum);
+ return;
+ }
+ bzero(ucmd, sizeof(gdt_ucmd_t));
+
+ /* wait for pending IOs */
+ lock = splcam();
+ gdt->sc_state = GDT_SHUTDOWN;
+ splx(lock);
+ if ((gccb = SLIST_FIRST(&gdt->sc_pending_gccb)) != NULL)
+ (void) tsleep((void *)gccb, PCATCH | PRIBIO, "iirshw", 100 * hz);
+
+ /* flush */
+ for (i = 0; i < GDT_MAX_HDRIVES; ++i) {
+ if (gdt->sc_hdr[i].hd_present) {
+ ucmd->service = GDT_CACHESERVICE;
+ ucmd->OpCode = GDT_FLUSH;
+ ucmd->u.cache.DeviceNo = i;
+ lock = splcam();
+ TAILQ_INSERT_TAIL(&gdt->sc_ucmd_queue, ucmd, links);
+ ucmd->complete_flag = FALSE;
+ splx(lock);
+ gdt_next(gdt);
+ if (!ucmd->complete_flag)
+ (void) tsleep((void *)ucmd, PCATCH|PRIBIO, "iirshw", 10*hz);
+ }
+ }
+
+ free(ucmd, M_DEVBUF);
+ printf("Done.\n");
+}
+
+void
+iir_intr(void *arg)
+{
+ struct gdt_softc *gdt = arg;
+ struct gdt_intr_ctx ctx;
+ int lock = 0;
+ struct gdt_ccb *gccb;
+ gdt_ucmd_t *ucmd;
+ u_int32_t cnt;
+
+ GDT_DPRINTF(GDT_D_INTR, ("gdt_intr(%p)\n", gdt));
+
+ /* If polling and we were not called from gdt_wait, just return */
+ if ((gdt->sc_state & GDT_POLLING) &&
+ !(gdt->sc_state & GDT_POLL_WAIT))
+ return;
+
+ if (!(gdt->sc_state & GDT_POLLING))
+ lock = splcam();
+ gdt_wait_index = 0;
+
+ ctx.istatus = gdt->sc_get_status(gdt);
+ if (!ctx.istatus) {
+ if (!(gdt->sc_state & GDT_POLLING))
+ splx(lock);
+ gdt->sc_status = GDT_S_NO_STATUS;
+ return;
+ }
+
+ gdt->sc_intr(gdt, &ctx);
+
+ gdt->sc_status = ctx.cmd_status;
+ gdt->sc_service = ctx.service;
+ gdt->sc_info = ctx.info;
+ gdt->sc_info2 = ctx.info2;
+
+ if (gdt->sc_state & GDT_POLL_WAIT) {
+ gdt_wait_gdt = gdt;
+ gdt_wait_index = ctx.istatus;
+ }
+
+ if (ctx.istatus == GDT_ASYNCINDEX) {
+ gdt_async_event(gdt, ctx.service);
+ if (!(gdt->sc_state & GDT_POLLING))
+ splx(lock);
+ return;
+ }
+ if (ctx.istatus == GDT_SPEZINDEX) {
+ GDT_DPRINTF(GDT_D_INVALID,
+ ("iir%d: Service unknown or not initialized!\n",
+ gdt->sc_hanum));
+ gdt->sc_dvr.size = sizeof(gdt->sc_dvr.eu.driver);
+ gdt->sc_dvr.eu.driver.ionode = gdt->sc_hanum;
+ gdt_store_event(GDT_ES_DRIVER, 4, &gdt->sc_dvr);
+ if (!(gdt->sc_state & GDT_POLLING))
+ splx(lock);
+ return;
+ }
+
+ gccb = &gdt->sc_gccbs[ctx.istatus - 2];
+ ctx.service = gccb->gc_service;
+
+ switch (gccb->gc_flags) {
+ case GDT_GCF_UNUSED:
+ GDT_DPRINTF(GDT_D_INVALID, ("iir%d: Index (%d) to unused command!\n",
+ gdt->sc_hanum, ctx.istatus));
+ gdt->sc_dvr.size = sizeof(gdt->sc_dvr.eu.driver);
+ gdt->sc_dvr.eu.driver.ionode = gdt->sc_hanum;
+ gdt->sc_dvr.eu.driver.index = ctx.istatus;
+ gdt_store_event(GDT_ES_DRIVER, 1, &gdt->sc_dvr);
+ gdt_free_ccb(gdt, gccb);
+ /* fallthrough */
+
+ case GDT_GCF_INTERNAL:
+ if (!(gdt->sc_state & GDT_POLLING))
+ splx(lock);
+ break;
+
+ case GDT_GCF_IOCTL:
+ ucmd = gccb->gc_ucmd;
+ if (gdt->sc_status == GDT_S_BSY) {
+ GDT_DPRINTF(GDT_D_DEBUG, ("iir_intr(%p) ioctl: gccb %p busy\n",
+ gdt, gccb));
+ TAILQ_INSERT_HEAD(&gdt->sc_ucmd_queue, ucmd, links);
+ if (!(gdt->sc_state & GDT_POLLING))
+ splx(lock);
+ } else {
+ ucmd->status = gdt->sc_status;
+ ucmd->info = gdt->sc_info;
+ ucmd->complete_flag = TRUE;
+ if (ucmd->service == GDT_CACHESERVICE) {
+ if (ucmd->OpCode == GDT_IOCTL) {
+ cnt = ucmd->u.ioctl.param_size;
+ if (cnt != 0)
+ bcopy(gccb->gc_scratch, ucmd->data, cnt);
+ } else {
+ cnt = ucmd->u.cache.BlockCnt * GDT_SECTOR_SIZE;
+ if (cnt != 0)
+ bcopy(gccb->gc_scratch, ucmd->data, cnt);
+ }
+ } else {
+ cnt = ucmd->u.raw.sdlen;
+ if (cnt != 0)
+ bcopy(gccb->gc_scratch, ucmd->data, cnt);
+ if (ucmd->u.raw.sense_len != 0)
+ bcopy(gccb->gc_scratch, ucmd->data, cnt);
+ }
+ gdt_free_ccb(gdt, gccb);
+ if (!(gdt->sc_state & GDT_POLLING))
+ splx(lock);
+ /* wakeup */
+ wakeup(ucmd);
+ }
+ gdt_next(gdt);
+ break;
+
+ default:
+ gdt_free_ccb(gdt, gccb);
+ gdt_sync_event(gdt, ctx.service, ctx.istatus, gccb);
+ if (!(gdt->sc_state & GDT_POLLING))
+ splx(lock);
+ gdt_next(gdt);
+ break;
+ }
+}
+
+int
+gdt_async_event(struct gdt_softc *gdt, int service)
+{
+ struct gdt_ccb *gccb;
+
+ GDT_DPRINTF(GDT_D_INTR, ("gdt_async_event(%p, %d)\n", gdt, service));
+
+ if (service == GDT_SCREENSERVICE) {
+ if (gdt->sc_status == GDT_MSG_REQUEST) {
+ while (gdt->sc_test_busy(gdt))
+ DELAY(1);
+ bzero(gdt->sc_cmd, GDT_CMD_SZ);
+ gccb = gdt_get_ccb(gdt);
+ if (gccb == NULL) {
+ printf("iir%d: No free command index found\n",
+ gdt->sc_hanum);
+ return (1);
+ }
+ gccb->gc_service = service;
+ gccb->gc_flags = GDT_GCF_SCREEN;
+ gdt->sc_set_sema0(gdt);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_COMMANDINDEX,
+ gccb->gc_cmd_index);
+ gdt_enc16(gdt->sc_cmd + GDT_CMD_OPCODE, GDT_READ);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_SCREEN_MSG_HANDLE,
+ GDT_MSG_INV_HANDLE);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_SCREEN_MSG_ADDR,
+ gdt_ccb_vtop(gdt, gccb) +
+ offsetof(struct gdt_ccb, gc_scratch[0]));
+ gdt->sc_cmd_off = 0;
+ gdt->sc_cmd_len = roundup(GDT_CMD_UNION + GDT_SCREEN_SZ,
+ sizeof(u_int32_t));
+ gdt->sc_cmd_cnt = 0;
+ gdt->sc_copy_cmd(gdt, gccb);
+ printf("iir%d: [PCI %d/%d] ",
+ gdt->sc_hanum,gdt->sc_bus,gdt->sc_slot);
+ gdt->sc_release_event(gdt);
+ }
+
+ } else {
+ if ((gdt->sc_fw_vers & 0xff) >= 0x1a) {
+ gdt->sc_dvr.size = 0;
+ gdt->sc_dvr.eu.async.ionode = gdt->sc_hanum;
+ gdt->sc_dvr.eu.async.status = gdt->sc_status;
+ /* severity and event_string already set! */
+ } else {
+ gdt->sc_dvr.size = sizeof(gdt->sc_dvr.eu.async);
+ gdt->sc_dvr.eu.async.ionode = gdt->sc_hanum;
+ gdt->sc_dvr.eu.async.service = service;
+ gdt->sc_dvr.eu.async.status = gdt->sc_status;
+ gdt->sc_dvr.eu.async.info = gdt->sc_info;
+ *(u_int32_t *)gdt->sc_dvr.eu.async.scsi_coord = gdt->sc_info2;
+ }
+ gdt_store_event(GDT_ES_ASYNC, service, &gdt->sc_dvr);
+ printf("iir%d: %s\n", gdt->sc_hanum, gdt->sc_dvr.event_string);
+ }
+
+ return (0);
+}
+
+int
+gdt_sync_event(struct gdt_softc *gdt, int service,
+ u_int8_t index, struct gdt_ccb *gccb)
+{
+ union ccb *ccb;
+ bus_dmasync_op_t op;
+
+ GDT_DPRINTF(GDT_D_INTR,
+ ("gdt_sync_event(%p, %d, %d, %p)\n", gdt,service,index,gccb));
+
+ ccb = gccb->gc_ccb;
+
+ if (service == GDT_SCREENSERVICE) {
+ u_int32_t msg_len;
+
+ msg_len = gdt_dec32(gccb->gc_scratch + GDT_SCR_MSG_LEN);
+ if (msg_len)
+ if (!(gccb->gc_scratch[GDT_SCR_MSG_ANSWER] &&
+ gccb->gc_scratch[GDT_SCR_MSG_EXT])) {
+ gccb->gc_scratch[GDT_SCR_MSG_TEXT + msg_len] = '\0';
+ printf("%s",&gccb->gc_scratch[GDT_SCR_MSG_TEXT]);
+ }
+
+ if (gccb->gc_scratch[GDT_SCR_MSG_EXT] &&
+ !gccb->gc_scratch[GDT_SCR_MSG_ANSWER]) {
+ while (gdt->sc_test_busy(gdt))
+ DELAY(1);
+ bzero(gdt->sc_cmd, GDT_CMD_SZ);
+ gccb = gdt_get_ccb(gdt);
+ if (gccb == NULL) {
+ printf("iir%d: No free command index found\n",
+ gdt->sc_hanum);
+ return (1);
+ }
+ gccb->gc_service = service;
+ gccb->gc_flags = GDT_GCF_SCREEN;
+ gdt->sc_set_sema0(gdt);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_COMMANDINDEX,
+ gccb->gc_cmd_index);
+ gdt_enc16(gdt->sc_cmd + GDT_CMD_OPCODE, GDT_READ);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_SCREEN_MSG_HANDLE,
+ gccb->gc_scratch[GDT_SCR_MSG_HANDLE]);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_SCREEN_MSG_ADDR,
+ gdt_ccb_vtop(gdt, gccb) +
+ offsetof(struct gdt_ccb, gc_scratch[0]));
+ gdt->sc_cmd_off = 0;
+ gdt->sc_cmd_len = roundup(GDT_CMD_UNION + GDT_SCREEN_SZ,
+ sizeof(u_int32_t));
+ gdt->sc_cmd_cnt = 0;
+ gdt->sc_copy_cmd(gdt, gccb);
+ gdt->sc_release_event(gdt);
+ return (0);
+ }
+
+ if (gccb->gc_scratch[GDT_SCR_MSG_ANSWER] &&
+ gdt_dec32(gccb->gc_scratch + GDT_SCR_MSG_ALEN)) {
+ /* default answers (getchar() not possible) */
+ if (gdt_dec32(gccb->gc_scratch + GDT_SCR_MSG_ALEN) == 1) {
+ gdt_enc32(gccb->gc_scratch + GDT_SCR_MSG_ALEN, 0);
+ gdt_enc32(gccb->gc_scratch + GDT_SCR_MSG_LEN, 1);
+ gccb->gc_scratch[GDT_SCR_MSG_TEXT] = 0;
+ } else {
+ gdt_enc32(gccb->gc_scratch + GDT_SCR_MSG_ALEN,
+ gdt_dec32(gccb->gc_scratch + GDT_SCR_MSG_ALEN) - 2);
+ gdt_enc32(gccb->gc_scratch + GDT_SCR_MSG_LEN, 2);
+ gccb->gc_scratch[GDT_SCR_MSG_TEXT] = 1;
+ gccb->gc_scratch[GDT_SCR_MSG_TEXT + 1] = 0;
+ }
+ gccb->gc_scratch[GDT_SCR_MSG_EXT] = 0;
+ gccb->gc_scratch[GDT_SCR_MSG_ANSWER] = 0;
+ while (gdt->sc_test_busy(gdt))
+ DELAY(1);
+ bzero(gdt->sc_cmd, GDT_CMD_SZ);
+ gccb = gdt_get_ccb(gdt);
+ if (gccb == NULL) {
+ printf("iir%d: No free command index found\n",
+ gdt->sc_hanum);
+ return (1);
+ }
+ gccb->gc_service = service;
+ gccb->gc_flags = GDT_GCF_SCREEN;
+ gdt->sc_set_sema0(gdt);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_COMMANDINDEX,
+ gccb->gc_cmd_index);
+ gdt_enc16(gdt->sc_cmd + GDT_CMD_OPCODE, GDT_WRITE);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_SCREEN_MSG_HANDLE,
+ gccb->gc_scratch[GDT_SCR_MSG_HANDLE]);
+ gdt_enc32(gdt->sc_cmd + GDT_CMD_UNION + GDT_SCREEN_MSG_ADDR,
+ gdt_ccb_vtop(gdt, gccb) +
+ offsetof(struct gdt_ccb, gc_scratch[0]));
+ gdt->sc_cmd_off = 0;
+ gdt->sc_cmd_len = roundup(GDT_CMD_UNION + GDT_SCREEN_SZ,
+ sizeof(u_int32_t));
+ gdt->sc_cmd_cnt = 0;
+ gdt->sc_copy_cmd(gdt, gccb);
+ gdt->sc_release_event(gdt);
+ return (0);
+ }
+ printf("\n");
+ return (0);
+ } else {
+ untimeout(iir_timeout, gccb, ccb->ccb_h.timeout_ch);
+ if (gdt->sc_status == GDT_S_BSY) {
+ GDT_DPRINTF(GDT_D_DEBUG, ("gdt_sync_event(%p) gccb %p busy\n",
+ gdt, gccb));
+ TAILQ_INSERT_HEAD(&gdt->sc_ccb_queue, &ccb->ccb_h, sim_links.tqe);
+ ++gdt_stat.req_queue_act;
+ if (gdt_stat.req_queue_act > gdt_stat.req_queue_max)
+ gdt_stat.req_queue_max = gdt_stat.req_queue_act;
+ return (2);
+ }
+
+ if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN)
+ op = BUS_DMASYNC_POSTREAD;
+ else
+ op = BUS_DMASYNC_POSTWRITE;
+ bus_dmamap_sync(gdt->sc_buffer_dmat, gccb->gc_dmamap, op);
+
+ ccb->csio.resid = 0;
+ if (gdt->sc_status == GDT_S_OK) {
+ ccb->ccb_h.status = CAM_REQ_CMP;
+ } else {
+ /* error */
+ if (gccb->gc_service == GDT_CACHESERVICE) {
+ ccb->ccb_h.status = CAM_SCSI_STATUS_ERROR | CAM_AUTOSNS_VALID;
+ ccb->csio.scsi_status = SCSI_STATUS_CHECK_COND;
+ bzero(&ccb->csio.sense_data, ccb->csio.sense_len);
+ ccb->csio.sense_data.error_code =
+ SSD_CURRENT_ERROR | SSD_ERRCODE_VALID;
+ ccb->csio.sense_data.flags = SSD_KEY_NOT_READY;
+
+ gdt->sc_dvr.size = sizeof(gdt->sc_dvr.eu.sync);
+ gdt->sc_dvr.eu.sync.ionode = gdt->sc_hanum;
+ gdt->sc_dvr.eu.sync.service = service;
+ gdt->sc_dvr.eu.sync.status = gdt->sc_status;
+ gdt->sc_dvr.eu.sync.info = gdt->sc_info;
+ gdt->sc_dvr.eu.sync.hostdrive = ccb->ccb_h.target_id;
+ if (gdt->sc_status >= 0x8000)
+ gdt_store_event(GDT_ES_SYNC, 0, &gdt->sc_dvr);
+ else
+ gdt_store_event(GDT_ES_SYNC, service, &gdt->sc_dvr);
+ } else {
+ /* raw service */
+ if (gdt->sc_status != GDT_S_RAW_SCSI || gdt->sc_info >= 0x100) {
+ ccb->ccb_h.status = CAM_SEL_TIMEOUT;
+ } else {
+ ccb->ccb_h.status = CAM_SCSI_STATUS_ERROR|CAM_AUTOSNS_VALID;
+ ccb->csio.scsi_status = gdt->sc_info;
+ bcopy(gccb->gc_scratch, &ccb->csio.sense_data,
+ ccb->csio.sense_len);
+ }
+ }
+ }
+ --gdt_stat.io_count_act;
+ xpt_done(ccb);
+ }
+ return (0);
+}
+
+/* Controller event handling functions */
+gdt_evt_str *gdt_store_event(u_int16_t source, u_int16_t idx,
+ gdt_evt_data *evt)
+{
+ gdt_evt_str *e;
+ struct timeval tv;
+
+ GDT_DPRINTF(GDT_D_MISC, ("gdt_store_event(%d, %d)\n", source, idx));
+ if (source == 0) /* no source -> no event */
+ return 0;
+
+ if (ebuffer[elastidx].event_source == source &&
+ ebuffer[elastidx].event_idx == idx &&
+ ((evt->size != 0 && ebuffer[elastidx].event_data.size != 0 &&
+ !memcmp((char *)&ebuffer[elastidx].event_data.eu,
+ (char *)&evt->eu, evt->size)) ||
+ (evt->size == 0 && ebuffer[elastidx].event_data.size == 0 &&
+ !strcmp((char *)&ebuffer[elastidx].event_data.event_string,
+ (char *)&evt->event_string)))) {
+ e = &ebuffer[elastidx];
+ getmicrotime(&tv);
+ e->last_stamp = tv.tv_sec;
+ ++e->same_count;
+ } else {
+ if (ebuffer[elastidx].event_source != 0) { /* entry not free ? */
+ ++elastidx;
+ if (elastidx == GDT_MAX_EVENTS)
+ elastidx = 0;
+ if (elastidx == eoldidx) { /* reached mark ? */
+ ++eoldidx;
+ if (eoldidx == GDT_MAX_EVENTS)
+ eoldidx = 0;
+ }
+ }
+ e = &ebuffer[elastidx];
+ e->event_source = source;
+ e->event_idx = idx;
+ getmicrotime(&tv);
+ e->first_stamp = e->last_stamp = tv.tv_sec;
+ e->same_count = 1;
+ e->event_data = *evt;
+ e->application = 0;
+ }
+ return e;
+}
+
+int gdt_read_event(int handle, gdt_evt_str *estr)
+{
+ gdt_evt_str *e;
+ int eindex, lock;
+
+ GDT_DPRINTF(GDT_D_MISC, ("gdt_read_event(%d)\n", handle));
+ lock = splcam();
+ if (handle == -1)
+ eindex = eoldidx;
+ else
+ eindex = handle;
+ estr->event_source = 0;
+
+ if (eindex >= GDT_MAX_EVENTS) {
+ splx(lock);
+ return eindex;
+ }
+ e = &ebuffer[eindex];
+ if (e->event_source != 0) {
+ if (eindex != elastidx) {
+ if (++eindex == GDT_MAX_EVENTS)
+ eindex = 0;
+ } else {
+ eindex = -1;
+ }
+ memcpy(estr, e, sizeof(gdt_evt_str));
+ }
+ splx(lock);
+ return eindex;
+}
+
+void gdt_readapp_event(u_int8_t application, gdt_evt_str *estr)
+{
+ gdt_evt_str *e;
+ int found = FALSE;
+ int eindex, lock;
+
+ GDT_DPRINTF(GDT_D_MISC, ("gdt_readapp_event(%d)\n", application));
+ lock = splcam();
+ eindex = eoldidx;
+ for (;;) {
+ e = &ebuffer[eindex];
+ if (e->event_source == 0)
+ break;
+ if ((e->application & application) == 0) {
+ e->application |= application;
+ found = TRUE;
+ break;
+ }
+ if (eindex == elastidx)
+ break;
+ if (++eindex == GDT_MAX_EVENTS)
+ eindex = 0;
+ }
+ if (found)
+ memcpy(estr, e, sizeof(gdt_evt_str));
+ else
+ estr->event_source = 0;
+ splx(lock);
+}
+
+void gdt_clear_events()
+{
+ GDT_DPRINTF(GDT_D_MISC, ("gdt_clear_events\n"));
+
+ eoldidx = elastidx = 0;
+ ebuffer[0].event_source = 0;
+}
diff --git a/sys/dev/iir/iir.h b/sys/dev/iir/iir.h
new file mode 100644
index 0000000..b165c84
--- /dev/null
+++ b/sys/dev/iir/iir.h
@@ -0,0 +1,712 @@
+/* $FreeBSD$ */
+/*
+ * Copyright (c) 2000-01 Intel Corporation
+ * All Rights Reserved
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions
+ * are met:
+ * 1. Redistributions of source code must retain the above copyright
+ * notice, this list of conditions, and the following disclaimer,
+ * without modification, immediately at the beginning of the file.
+ * 2. Redistributions in binary form must reproduce the above copyright
+ * notice, this list of conditions and the following disclaimer in the
+ * documentation and/or other materials provided with the distribution.
+ * 3. The name of the author may not be used to endorse or promote products
+ * derived from this software without specific prior written permission.
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
+ * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
+ * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE FOR
+ * ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
+ * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
+ * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
+ * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
+ * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
+ * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
+ * SUCH DAMAGE.
+ *
+ */
+
+/*
+ *
+ * iir.h: Definitions/Constants used by the Intel Integrated RAID driver
+ *
+ * Written by: Achim Leubner <achim.leubner@intel.com>
+ * Fixes/Additions: Boji Tony Kannanthanam <boji.t.kannanthanam@intel.com>
+ *
+ * credits: Niklas Hallqvist; OpenBSD driver for the ICP Controllers.
+ * FreeBSD.ORG; Great O/S to work on and for.
+ */
+
+
+#ident "$Id: iir.h 1.3 2001/07/03 11:28:57 achim Exp $"
+
+#ifndef _IIR_H
+#define _IIR_H
+
+#define IIR_DRIVER_VERSION 1
+#define IIR_DRIVER_SUBVERSION 1
+
+#define IIR_CDEV_MAJOR 164
+
+#define GDT_VENDOR_ID 0x1119
+#define GDT_DEVICE_ID_MIN 0x100
+#define GDT_DEVICE_ID_MAX 0x2ff
+#define GDT_DEVICE_ID_NEWRX 0x300
+
+#define INTEL_VENDOR_ID 0x8086
+#define INTEL_DEVICE_ID_IIR 0x600
+
+#define GDT_MAXBUS 6 /* XXX Why not 5? */
+#define GDT_MAX_HDRIVES 100 /* max 100 host drives */
+#define GDT_MAXID_FC 127 /* Fibre-channel IDs */
+#define GDT_MAXID 16 /* SCSI IDs */
+#define GDT_MAXOFFSETS 128
+#define GDT_MAXSG 32 /* Max. s/g elements */
+#define GDT_PROTOCOL_VERSION 1
+#define GDT_LINUX_OS 8 /* Used for cache optimization */
+#define GDT_SCATTER_GATHER 1 /* s/g feature */
+#define GDT_SECS32 0x1f /* round capacity */
+#define GDT_LOCALBOARD 0 /* Board node always 0 */
+#define GDT_MAXCMDS 124
+#define GDT_SECTOR_SIZE 0x200 /* Always 512 bytes for cache devs */
+#define GDT_MAX_EVENTS 0x100 /* event buffer */
+
+/* DPMEM constants */
+#define GDT_MPR_MAGIC 0xc0ffee11
+#define GDT_IC_HEADER_BYTES 48
+#define GDT_IC_QUEUE_BYTES 4
+#define GDT_DPMEM_COMMAND_OFFSET \
+ (GDT_IC_HEADER_BYTES + GDT_IC_QUEUE_BYTES * GDT_MAXOFFSETS)
+
+/* geometry constants */
+#define GDT_MAXCYLS 1024
+#define GDT_HEADS 64
+#define GDT_SECS 32 /* mapping 64*32 */
+#define GDT_MEDHEADS 127
+#define GDT_MEDSECS 63 /* mapping 127*63 */
+#define GDT_BIGHEADS 255
+#define GDT_BIGSECS 63 /* mapping 255*63 */
+
+/* data direction raw service */
+#define GDT_DATA_IN 0x01000000L
+#define GDT_DATA_OUT 0x00000000L
+
+/* Cache/raw service commands */
+#define GDT_INIT 0 /* service initialization */
+#define GDT_READ 1 /* read command */
+#define GDT_WRITE 2 /* write command */
+#define GDT_INFO 3 /* information about devices */
+#define GDT_FLUSH 4 /* flush dirty cache buffers */
+#define GDT_IOCTL 5 /* ioctl command */
+#define GDT_DEVTYPE 9 /* additional information */
+#define GDT_MOUNT 10 /* mount cache device */
+#define GDT_UNMOUNT 11 /* unmount cache device */
+#define GDT_SET_FEAT 12 /* set features (scatter/gather) */
+#define GDT_GET_FEAT 13 /* get features */
+#define GDT_WRITE_THR 16 /* write through */
+#define GDT_READ_THR 17 /* read through */
+#define GDT_EXT_INFO 18 /* extended info */
+#define GDT_RESET 19 /* controller reset */
+#define GDT_FREEZE_IO 25 /* freeze all IOs */
+#define GDT_UNFREEZE_IO 26 /* unfreeze all IOs */
+
+/* Additional raw service commands */
+#define GDT_RESERVE 14 /* reserve device to raw service */
+#define GDT_RELEASE 15 /* release device */
+#define GDT_RESERVE_ALL 16 /* reserve all devices */
+#define GDT_RELEASE_ALL 17 /* release all devices */
+#define GDT_RESET_BUS 18 /* reset bus */
+#define GDT_SCAN_START 19 /* start device scan */
+#define GDT_SCAN_END 20 /* stop device scan */
+
+/* IOCTL command defines */
+#define GDT_SCSI_DR_INFO 0x00 /* SCSI drive info */
+#define GDT_SCSI_CHAN_CNT 0x05 /* SCSI channel count */
+#define GDT_SCSI_DR_LIST 0x06 /* SCSI drive list */
+#define GDT_SCSI_DEF_CNT 0x15 /* grown/primary defects */
+#define GDT_DSK_STATISTICS 0x4b /* SCSI disk statistics */
+#define GDT_IOCHAN_DESC 0x5d /* description of IO channel */
+#define GDT_IOCHAN_RAW_DESC 0x5e /* description of raw IO channel */
+
+#define GDT_L_CTRL_PATTERN 0x20000000 /* SCSI IOCTL mask */
+#define GDT_ARRAY_INFO 0x12 /* array drive info */
+#define GDT_ARRAY_DRV_LIST 0x0f /* array drive list */
+#define GDT_LA_CTRL_PATTERN 0x10000000 /* array IOCTL mask */
+#define GDT_CACHE_DRV_CNT 0x01 /* cache drive count */
+#define GDT_CACHE_DRV_LIST 0x02 /* cache drive list */
+#define GDT_CACHE_INFO 0x04 /* cache info */
+#define GDT_CACHE_CONFIG 0x05 /* cache configuration */
+#define GDT_CACHE_DRV_INFO 0x07 /* cache drive info */
+#define GDT_BOARD_FEATURES 0x15 /* controller features */
+#define GDT_BOARD_INFO 0x28 /* controller info */
+#define GDT_HOST_GET 0x10001 /* get host drive list */
+#define GDT_IO_CHANNEL 0x20000 /* default IO channel */
+#define GDT_INVALID_CHANNEL 0xffff /* invalid channel */
+
+/* IOCTLs */
+#define GDT_IOCTL_GENERAL _IOWR('J', 0, gdt_ucmd_t) /* general IOCTL */
+#define GDT_IOCTL_DRVERS _IOWR('J', 1, int) /* get driver version */
+#define GDT_IOCTL_CTRTYPE _IOR('J', 2, gdt_ctrt_t) /* get ctr. type */
+#define GDT_IOCTL_OSVERS _IOR('J', 3, gdt_osv_t) /* get OS version */
+#define GDT_IOCTL_CTRCNT _IOR('J', 5, int) /* get ctr. count */
+#define GDT_IOCTL_EVENT _IOWR('J', 8, gdt_event_t) /* get event */
+#define GDT_IOCTL_STATIST _IOR('J', 9, gdt_statist_t) /* get statistics */
+
+/* Service errors */
+#define GDT_S_OK 1 /* no error */
+#define GDT_S_BSY 7 /* controller busy */
+#define GDT_S_RAW_SCSI 12 /* raw service: target error */
+#define GDT_S_RAW_ILL 0xff /* raw service: illegal */
+#define GDT_S_NO_STATUS 0x1000 /* got no status (driver-generated) */
+
+/* Controller services */
+#define GDT_SCSIRAWSERVICE 3
+#define GDT_CACHESERVICE 9
+#define GDT_SCREENSERVICE 11
+
+/* Scatter/gather element */
+#define GDT_SG_PTR 0x00 /* u_int32_t, address */
+#define GDT_SG_LEN 0x04 /* u_int32_t, length */
+#define GDT_SG_SZ 0x08
+
+/* Cache service command */
+#define GDT_CACHE_DEVICENO 0x00 /* u_int16_t, number of cache drive */
+#define GDT_CACHE_BLOCKNO 0x02 /* u_int32_t, block number */
+#define GDT_CACHE_BLOCKCNT 0x06 /* u_int32_t, block count */
+#define GDT_CACHE_DESTADDR 0x0a /* u_int32_t, dest. addr. (-1: s/g) */
+#define GDT_CACHE_SG_CANZ 0x0e /* u_int32_t, s/g element count */
+#define GDT_CACHE_SG_LST 0x12 /* [GDT_MAXSG], s/g list */
+#define GDT_CACHE_SZ (0x12 + GDT_MAXSG * GDT_SG_SZ)
+
+/* Ioctl command */
+#define GDT_IOCTL_PARAM_SIZE 0x00 /* u_int16_t, size of buffer */
+#define GDT_IOCTL_SUBFUNC 0x02 /* u_int32_t, ioctl function */
+#define GDT_IOCTL_CHANNEL 0x06 /* u_int32_t, device */
+#define GDT_IOCTL_P_PARAM 0x0a /* u_int32_t, buffer */
+#define GDT_IOCTL_SZ 0x0e
+
+/* Screen service defines */
+#define GDT_MSG_INV_HANDLE -1 /* special message handle */
+#define GDT_MSGLEN 16 /* size of message text */
+#define GDT_MSG_SIZE 34 /* size of message structure */
+#define GDT_MSG_REQUEST 0 /* async. event. message */
+
+/* Screen service command */
+#define GDT_SCREEN_MSG_HANDLE 0x02 /* u_int32_t, message handle */
+#define GDT_SCREEN_MSG_ADDR 0x06 /* u_int32_t, message buffer address */
+#define GDT_SCREEN_SZ 0x0a
+
+/* Screen service message */
+#define GDT_SCR_MSG_HANDLE 0x00 /* u_int32_t, message handle */
+#define GDT_SCR_MSG_LEN 0x04 /* u_int32_t, size of message */
+#define GDT_SCR_MSG_ALEN 0x08 /* u_int32_t, answer length */
+#define GDT_SCR_MSG_ANSWER 0x0c /* u_int8_t, answer flag */
+#define GDT_SCR_MSG_EXT 0x0d /* u_int8_t, more messages? */
+#define GDT_SCR_MSG_RES 0x0e /* u_int16_t, reserved */
+#define GDT_SCR_MSG_TEXT 0x10 /* GDT_MSGLEN+2, message text */
+#define GDT_SCR_MSG_SZ (0x12 + GDT_MSGLEN)
+
+/* Raw service command */
+#define GDT_RAW_DIRECTION 0x02 /* u_int32_t, data direction */
+#define GDT_RAW_MDISC_TIME 0x06 /* u_int32_t, disc. time (0: none) */
+#define GDT_RAW_MCON_TIME 0x0a /* u_int32_t, conn. time (0: none) */
+#define GDT_RAW_SDATA 0x0e /* u_int32_t, dest. addr. (-1: s/g) */
+#define GDT_RAW_SDLEN 0x12 /* u_int32_t, data length */
+#define GDT_RAW_CLEN 0x16 /* u_int32_t, SCSI cmd len (6/10/12) */
+#define GDT_RAW_CMD 0x1a /* u_int8_t [12], SCSI command */
+#define GDT_RAW_TARGET 0x26 /* u_int8_t, target ID */
+#define GDT_RAW_LUN 0x27 /* u_int8_t, LUN */
+#define GDT_RAW_BUS 0x28 /* u_int8_t, SCSI bus number */
+#define GDT_RAW_PRIORITY 0x29 /* u_int8_t, only 0 used */
+#define GDT_RAW_SENSE_LEN 0x2a /* u_int32_t, sense data length */
+#define GDT_RAW_SENSE_DATA 0x2e /* u_int32_t, sense data address */
+#define GDT_RAW_SG_RANZ 0x36 /* u_int32_t, s/g element count */
+#define GDT_RAW_SG_LST 0x3a /* [GDT_MAXSG], s/g list */
+#define GDT_RAW_SZ (0x3a + GDT_MAXSG * GDT_SG_SZ)
+
+/* Command structure */
+#define GDT_CMD_BOARDNODE 0x00 /* u_int32_t, board node (always 0) */
+#define GDT_CMD_COMMANDINDEX 0x04 /* u_int32_t, command number */
+#define GDT_CMD_OPCODE 0x08 /* u_int16_t, opcode (READ, ...) */
+#define GDT_CMD_UNION 0x0a /* cache/screen/raw service command */
+#define GDT_CMD_UNION_SZ GDT_RAW_SZ
+#define GDT_CMD_SZ (0x0a + GDT_CMD_UNION_SZ)
+
+/* Command queue entries */
+#define GDT_OFFSET 0x00 /* u_int16_t, command offset in the DP RAM */
+#define GDT_SERV_ID 0x02 /* u_int16_t, service */
+#define GDT_COMM_Q_SZ 0x04
+
+/* Interface area */
+#define GDT_S_CMD_INDX 0x00 /* u_int8_t, special command */
+#define GDT_S_STATUS 0x01 /* volatile u_int8_t, status special command */
+#define GDT_S_INFO 0x04 /* u_int32_t [4], add. info special command */
+#define GDT_SEMA0 0x14 /* volatile u_int8_t, command semaphore */
+#define GDT_CMD_INDEX 0x18 /* u_int8_t, command number */
+#define GDT_STATUS 0x1c /* volatile u_int16_t, command status */
+#define GDT_SERVICE 0x1e /* u_int16_t, service (for asynch. events) */
+#define GDT_DPR_INFO 0x20 /* u_int32_t [2], additional info */
+#define GDT_COMM_QUEUE 0x28 /* command queue */
+#define GDT_DPR_CMD (0x30 + GDT_MAXOFFSETS * GDT_COMM_Q_SZ)
+ /* u_int8_t [], commands */
+
+/* I/O channel header */
+#define GDT_IOC_VERSION 0x00 /* u_int32_t, version (~0: newest) */
+#define GDT_IOC_LIST_ENTRIES 0x04 /* u_int8_t, list entry count */
+#define GDT_IOC_FIRST_CHAN 0x05 /* u_int8_t, first channel number */
+#define GDT_IOC_LAST_CHAN 0x06 /* u_int8_t, last channel number */
+#define GDT_IOC_CHAN_COUNT 0x07 /* u_int8_t, (R) channel count */
+#define GDT_IOC_LIST_OFFSET 0x08 /* u_int32_t, offset of list[0] */
+#define GDT_IOC_HDR_SZ 0x0c
+
+#define GDT_IOC_NEWEST 0xffffffff /* goes into GDT_IOC_VERSION */
+
+/* Get I/O channel description */
+#define GDT_IOC_ADDRESS 0x00 /* u_int32_t, channel address */
+#define GDT_IOC_TYPE 0x04 /* u_int8_t, type (SCSI/FCSL) */
+#define GDT_IOC_LOCAL_NO 0x05 /* u_int8_t, local number */
+#define GDT_IOC_FEATURES 0x06 /* u_int16_t, channel features */
+#define GDT_IOC_SZ 0x08
+
+/* Get raw I/O channel description */
+#define GDT_RAWIOC_PROC_ID 0x00 /* u_int8_t, processor id */
+#define GDT_RAWIOC_PROC_DEFECT 0x01 /* u_int8_t, defect? */
+#define GDT_RAWIOC_SZ 0x04
+
+/* Get SCSI channel count */
+#define GDT_GETCH_CHANNEL_NO 0x00 /* u_int32_t, channel number */
+#define GDT_GETCH_DRIVE_CNT 0x04 /* u_int32_t, drive count */
+#define GDT_GETCH_SIOP_ID 0x08 /* u_int8_t, SCSI processor ID */
+#define GDT_GETCH_SIOP_STATE 0x09 /* u_int8_t, SCSI processor state */
+#define GDT_GETCH_SZ 0x0a
+
+/* Cache info/config IOCTL structures */
+#define GDT_CPAR_VERSION 0x00 /* u_int32_t, firmware version */
+#define GDT_CPAR_STATE 0x04 /* u_int16_t, cache state (on/off) */
+#define GDT_CPAR_STRATEGY 0x06 /* u_int16_t, cache strategy */
+#define GDT_CPAR_WRITE_BACK 0x08 /* u_int16_t, write back (on/off) */
+#define GDT_CPAR_BLOCK_SIZE 0x0a /* u_int16_t, cache block size */
+#define GDT_CPAR_SZ 0x0c
+
+#define GDT_CSTAT_CSIZE 0x00 /* u_int32_t, cache size */
+#define GDT_CSTAT_READ_CNT 0x04 /* u_int32_t, read counter */
+#define GDT_CSTAT_WRITE_CNT 0x08 /* u_int32_t, write counter */
+#define GDT_CSTAT_TR_HITS 0x0c /* u_int32_t, track hits */
+#define GDT_CSTAT_SEC_HITS 0x10 /* u_int32_t, sector hits */
+#define GDT_CSTAT_SEC_MISS 0x14 /* u_int32_t, sector misses */
+#define GDT_CSTAT_SZ 0x18
+
+/* Get cache info */
+#define GDT_CINFO_CPAR 0x00
+#define GDT_CINFO_CSTAT GDT_CPAR_SZ
+#define GDT_CINFO_SZ (GDT_CPAR_SZ + GDT_CSTAT_SZ)
+
+/* Get board info */
+#define GDT_BINFO_SER_NO 0x00 /* u_int32_t, serial number */
+#define GDT_BINFO_OEM_ID 0x04 /* u_int8_t [2], OEM ID */
+#define GDT_BINFO_EP_FLAGS 0x06 /* u_int16_t, eprom flags */
+#define GDT_BINFO_PROC_ID 0x08 /* u_int32_t, processor ID */
+#define GDT_BINFO_MEMSIZE 0x0c /* u_int32_t, memory size (bytes) */
+#define GDT_BINFO_MEM_BANKS 0x10 /* u_int8_t, memory banks */
+#define GDT_BINFO_CHAN_TYPE 0x11 /* u_int8_t, channel type */
+#define GDT_BINFO_CHAN_COUNT 0x12 /* u_int8_t, channel count */
+#define GDT_BINFO_RDONGLE_PRES 0x13 /* u_int8_t, dongle present */
+#define GDT_BINFO_EPR_FW_VER 0x14 /* u_int32_t, (eprom) firmware ver */
+#define GDT_BINFO_UPD_FW_VER 0x18 /* u_int32_t, (update) firmware ver */
+#define GDT_BINFO_UPD_REVISION 0x1c /* u_int32_t, update revision */
+#define GDT_BINFO_TYPE_STRING 0x20 /* char [16], controller name */
+#define GDT_BINFO_RAID_STRING 0x30 /* char [16], RAID firmware name */
+#define GDT_BINFO_UPDATE_PRES 0x40 /* u_int8_t, update present? */
+#define GDT_BINFO_XOR_PRES 0x41 /* u_int8_t, XOR engine present */
+#define GDT_BINFO_PROM_TYPE 0x42 /* u_int8_t, ROM type (eprom/flash) */
+#define GDT_BINFO_PROM_COUNT 0x43 /* u_int8_t, number of ROM devices */
+#define GDT_BINFO_DUP_PRES 0x44 /* u_int32_t, duplexing module pres? */
+#define GDT_BINFO_CHAN_PRES 0x48 /* u_int32_t, # of exp. channels */
+#define GDT_BINFO_MEM_PRES 0x4c /* u_int32_t, memory expansion inst? */
+#define GDT_BINFO_FT_BUS_SYSTEM 0x50 /* u_int8_t, fault bus supported? */
+#define GDT_BINFO_SUBTYPE_VALID 0x51 /* u_int8_t, board_subtype valid */
+#define GDT_BINFO_BOARD_SUBTYPE 0x52 /* u_int8_t, subtype/hardware level */
+#define GDT_BINFO_RAMPAR_PRES 0x53 /* u_int8_t, RAM parity check hw? */
+#define GDT_BINFO_SZ 0x54
+
+/* Get board features */
+#define GDT_BFEAT_CHAINING 0x00 /* u_int8_t, chaining supported */
+#define GDT_BFEAT_STRIPING 0x01 /* u_int8_t, striping (RAID-0) supp. */
+#define GDT_BFEAT_MIRRORING 0x02 /* u_int8_t, mirroring (RAID-1) supp */
+#define GDT_BFEAT_RAID 0x03 /* u_int8_t, RAID-4/5/10 supported */
+#define GDT_BFEAT_SZ 0x04
+
+/* Other defines */
+#define GDT_ASYNCINDEX 0 /* command index asynchronous event */
+#define GDT_SPEZINDEX 1 /* command index unknown service */
+
+/* Debugging */
+#ifdef GDT_DEBUG
+#define GDT_D_INTR 0x01
+#define GDT_D_MISC 0x02
+#define GDT_D_CMD 0x04
+#define GDT_D_QUEUE 0x08
+#define GDT_D_TIMEOUT 0x10
+#define GDT_D_INIT 0x20
+#define GDT_D_INVALID 0x40
+#define GDT_D_DEBUG 0x80
+extern int gdt_debug;
+#ifdef __SERIAL__
+extern int ser_printf(const char *fmt, ...);
+#define GDT_DPRINTF(mask, args) if (gdt_debug & (mask)) ser_printf args
+#else
+#define GDT_DPRINTF(mask, args) if (gdt_debug & (mask)) printf args
+#endif
+#else
+#define GDT_DPRINTF(mask, args)
+#endif
+
+/* Miscellaneous constants */
+#define GDT_RETRIES 100000000 /* 100000 * 1us = 100s */
+#define GDT_TIMEOUT 100000000 /* 100000 * 1us = 100s */
+#define GDT_POLL_TIMEOUT 10000000 /* 10000 * 1us = 10s */
+#define GDT_WATCH_TIMEOUT 10000000 /* 10000 * 1us = 10s */
+#define GDT_SCRATCH_SZ 3072 /* 3KB scratch buffer */
+
+/* macros */
+#define htole32(v) (v)
+#define htole16(v) (v)
+#define letoh32(v) (v)
+#define letoh16(v) (v)
+
+/* Map minor numbers to device identity */
+#define LUN_MASK 0x0007
+#define TARGET_MASK 0x03f8
+#define BUS_MASK 0x1c00
+#define HBA_MASK 0xe000
+
+#define minor2lun(minor) ( minor & LUN_MASK )
+#define minor2target(minor) ( (minor & TARGET_MASK) >> 3 )
+#define minor2bus(minor) ( (minor & BUS_MASK) >> 10 )
+#define minor2hba(minor) ( (minor & HBA_MASK) >> 13 )
+#define hba2minor(hba) ( (hba << 13) & HBA_MASK )
+
+
+/* struct for GDT_IOCTL_GENERAL */
+#pragma pack(1)
+typedef struct gdt_ucmd {
+ u_int16_t io_node;
+ u_int16_t service;
+ u_int32_t timeout;
+ u_int16_t status;
+ u_int32_t info;
+
+ u_int32_t BoardNode; /* board node (always 0) */
+ u_int32_t CommandIndex; /* command number */
+ u_int16_t OpCode; /* the command (READ,..) */
+ union {
+ struct {
+ u_int16_t DeviceNo; /* number of cache drive */
+ u_int32_t BlockNo; /* block number */
+ u_int32_t BlockCnt; /* block count */
+ void *DestAddr; /* data */
+ } cache; /* cache service cmd. str. */
+ struct {
+ u_int16_t param_size; /* size of p_param buffer */
+ u_int32_t subfunc; /* IOCTL function */
+ u_int32_t channel; /* device */
+ void *p_param; /* data */
+ } ioctl; /* IOCTL command structure */
+ struct {
+ u_int16_t reserved;
+ u_int32_t direction; /* data direction */
+ u_int32_t mdisc_time; /* disc. time (0: no timeout)*/
+ u_int32_t mcon_time; /* connect time(0: no to.) */
+ void *sdata; /* dest. addr. (if s/g: -1) */
+ u_int32_t sdlen; /* data length (bytes) */
+ u_int32_t clen; /* SCSI cmd. length(6,10,12) */
+ u_int8_t cmd[12]; /* SCSI command */
+ u_int8_t target; /* target ID */
+ u_int8_t lun; /* LUN */
+ u_int8_t bus; /* SCSI bus number */
+ u_int8_t priority; /* only 0 used */
+ u_int32_t sense_len; /* sense data length */
+ void *sense_data; /* sense data addr. */
+ u_int32_t link_p; /* linked cmds (not supp.) */
+ } raw; /* raw service cmd. struct. */
+ } u;
+ u_int8_t data[GDT_SCRATCH_SZ];
+ int complete_flag;
+ TAILQ_ENTRY(gdt_ucmd) links;
+} gdt_ucmd_t;
+
+/* struct for GDT_IOCTL_CTRTYPE */
+typedef struct gdt_ctrt {
+ u_int16_t io_node;
+ u_int16_t oem_id;
+ u_int16_t type;
+ u_int32_t info;
+ u_int8_t access;
+ u_int8_t remote;
+ u_int16_t ext_type;
+ u_int16_t device_id;
+ u_int16_t sub_device_id;
+} gdt_ctrt_t;
+
+/* struct for GDT_IOCTL_OSVERS */
+typedef struct gdt_osv {
+ u_int8_t oscode;
+ u_int8_t version;
+ u_int8_t subversion;
+ u_int16_t revision;
+ char name[64];
+} gdt_osv_t;
+
+/* controller event structure */
+#define GDT_ES_ASYNC 1
+#define GDT_ES_DRIVER 2
+#define GDT_ES_TEST 3
+#define GDT_ES_SYNC 4
+typedef struct {
+ u_int16_t size; /* size of structure */
+ union {
+ char stream[16];
+ struct {
+ u_int16_t ionode;
+ u_int16_t service;
+ u_int32_t index;
+ } driver;
+ struct {
+ u_int16_t ionode;
+ u_int16_t service;
+ u_int16_t status;
+ u_int32_t info;
+ u_int8_t scsi_coord[3];
+ } async;
+ struct {
+ u_int16_t ionode;
+ u_int16_t service;
+ u_int16_t status;
+ u_int32_t info;
+ u_int16_t hostdrive;
+ u_int8_t scsi_coord[3];
+ u_int8_t sense_key;
+ } sync;
+ struct {
+ u_int32_t l1, l2, l3, l4;
+ } test;
+ } eu;
+ u_int32_t severity;
+ u_int8_t event_string[256];
+} gdt_evt_data;
+
+/* dvrevt structure */
+typedef struct {
+ u_int32_t first_stamp;
+ u_int32_t last_stamp;
+ u_int16_t same_count;
+ u_int16_t event_source;
+ u_int16_t event_idx;
+ u_int8_t application;
+ u_int8_t reserved;
+ gdt_evt_data event_data;
+} gdt_evt_str;
+
+/* struct for GDT_IOCTL_EVENT */
+typedef struct gdt_event {
+ int erase;
+ int handle;
+ gdt_evt_str dvr;
+} gdt_event_t;
+
+/* struct for GDT_IOCTL_STATIST */
+typedef struct gdt_statist {
+ u_int16_t io_count_act;
+ u_int16_t io_count_max;
+ u_int16_t req_queue_act;
+ u_int16_t req_queue_max;
+ u_int16_t cmd_index_act;
+ u_int16_t cmd_index_max;
+ u_int16_t sg_count_act;
+ u_int16_t sg_count_max;
+} gdt_statist_t;
+
+#pragma pack()
+
+/* Context structure for interrupt services */
+struct gdt_intr_ctx {
+ u_int32_t info, info2;
+ u_int16_t cmd_status, service;
+ u_int8_t istatus;
+};
+
+/* softc structure */
+struct gdt_softc {
+ int sc_hanum;
+ int sc_class; /* Controller class */
+#define GDT_MPR 0x05
+#define GDT_CLASS_MASK 0x07
+#define GDT_FC 0x10
+#define GDT_CLASS(gdt) ((gdt)->sc_class & GDT_CLASS_MASK)
+ int sc_bus, sc_slot;
+ u_int16_t sc_device, sc_subdevice;
+ u_int16_t sc_fw_vers;
+ int sc_init_level;
+ int sc_state;
+#define GDT_NORMAL 0x00
+#define GDT_POLLING 0x01
+#define GDT_SHUTDOWN 0x02
+#define GDT_POLL_WAIT 0x80
+ dev_t sc_dev;
+ bus_space_tag_t sc_dpmemt;
+ bus_space_handle_t sc_dpmemh;
+ bus_addr_t sc_dpmembase;
+ bus_dma_tag_t sc_parent_dmat;
+ bus_dma_tag_t sc_buffer_dmat;
+ bus_dma_tag_t sc_gccb_dmat;
+ bus_dmamap_t sc_gccb_dmamap;
+ bus_addr_t sc_gccb_busbase;
+
+ struct gdt_ccb *sc_gccbs;
+ SLIST_HEAD(, gdt_ccb) sc_free_gccb, sc_pending_gccb;
+ TAILQ_HEAD(, ccb_hdr) sc_ccb_queue;
+ TAILQ_HEAD(, gdt_ucmd) sc_ucmd_queue;
+
+ u_int16_t sc_ic_all_size;
+ u_int16_t sc_cmd_len;
+ u_int16_t sc_cmd_off;
+ u_int16_t sc_cmd_cnt;
+ u_int8_t sc_cmd[GDT_CMD_SZ];
+
+ u_int32_t sc_info;
+ u_int32_t sc_info2;
+ u_int16_t sc_status;
+ u_int16_t sc_service;
+
+ u_int8_t sc_bus_cnt;
+ u_int8_t sc_virt_bus;
+ u_int8_t sc_bus_id[GDT_MAXBUS];
+ u_int8_t sc_more_proc;
+
+ struct {
+ u_int8_t hd_present;
+ u_int8_t hd_is_logdrv;
+ u_int8_t hd_is_arraydrv;
+ u_int8_t hd_is_master;
+ u_int8_t hd_is_parity;
+ u_int8_t hd_is_hotfix;
+ u_int8_t hd_master_no;
+ u_int8_t hd_lock;
+ u_int8_t hd_heads;
+ u_int8_t hd_secs;
+ u_int16_t hd_devtype;
+ u_int32_t hd_size;
+ u_int8_t hd_ldr_no;
+ u_int8_t hd_rw_attribs;
+ u_int32_t hd_start_sec;
+ } sc_hdr[GDT_MAX_HDRIVES];
+
+ u_int16_t sc_raw_feat;
+ u_int16_t sc_cache_feat;
+
+ gdt_evt_data sc_dvr;
+
+ struct cam_sim *sims[GDT_MAXBUS];
+ struct cam_path *paths[GDT_MAXBUS];
+
+ void (*sc_copy_cmd) __P((struct gdt_softc *, struct gdt_ccb *));
+ u_int8_t (*sc_get_status) __P((struct gdt_softc *));
+ void (*sc_intr) __P((struct gdt_softc *, struct gdt_intr_ctx *));
+ void (*sc_release_event) __P((struct gdt_softc *));
+ void (*sc_set_sema0) __P((struct gdt_softc *));
+ int (*sc_test_busy) __P((struct gdt_softc *));
+
+ TAILQ_ENTRY(gdt_softc) links;
+};
+
+/*
+ * A command control block, one for each corresponding command index of the
+ * controller.
+ */
+struct gdt_ccb {
+ u_int8_t gc_scratch[GDT_SCRATCH_SZ];
+ union ccb *gc_ccb;
+ gdt_ucmd_t *gc_ucmd;
+ bus_dmamap_t gc_dmamap;
+ int gc_map_flag;
+ int gc_timeout;
+ int gc_state;
+ u_int8_t gc_service;
+ u_int8_t gc_cmd_index;
+ u_int8_t gc_flags;
+#define GDT_GCF_UNUSED 0
+#define GDT_GCF_INTERNAL 1
+#define GDT_GCF_SCREEN 2
+#define GDT_GCF_SCSI 3
+#define GDT_GCF_IOCTL 4
+ SLIST_ENTRY(gdt_ccb) sle;
+};
+
+
+int iir_init __P((struct gdt_softc *));
+void iir_free __P((struct gdt_softc *));
+void iir_attach __P((struct gdt_softc *));
+void iir_intr __P((void *arg));
+
+#ifdef __GNUC__
+/* These all require correctly aligned buffers */
+static __inline__ void gdt_enc16 __P((u_int8_t *, u_int16_t));
+static __inline__ void gdt_enc32 __P((u_int8_t *, u_int32_t));
+static __inline__ u_int16_t gdt_dec16 __P((u_int8_t *));
+static __inline__ u_int32_t gdt_dec32 __P((u_int8_t *));
+
+static __inline__ void
+gdt_enc16(addr, value)
+ u_int8_t *addr;
+ u_int16_t value;
+{
+ *(u_int16_t *)addr = htole16(value);
+}
+
+static __inline__ void
+gdt_enc32(addr, value)
+ u_int8_t *addr;
+ u_int32_t value;
+{
+ *(u_int32_t *)addr = htole32(value);
+}
+
+static __inline__ u_int16_t
+gdt_dec16(addr)
+ u_int8_t *addr;
+{
+ return letoh16(*(u_int16_t *)addr);
+}
+
+static __inline__ u_int32_t
+gdt_dec32(addr)
+ u_int8_t *addr;
+{
+ return letoh32(*(u_int32_t *)addr);
+}
+#endif
+
+#if defined(__alpha__)
+/* XXX XXX NEED REAL DMA MAPPING SUPPORT XXX XXX */
+#undef vtophys
+#define vtophys(va) alpha_XXX_dmamap((vm_offset_t)(va))
+#endif
+
+extern TAILQ_HEAD(gdt_softc_list, gdt_softc) gdt_softcs;
+extern u_int8_t gdt_polling;
+
+dev_t gdt_make_dev(int unit);
+void gdt_destroy_dev(dev_t dev);
+void gdt_next(struct gdt_softc *gdt);
+void gdt_free_ccb(struct gdt_softc *gdt, struct gdt_ccb *gccb);
+
+gdt_evt_str *gdt_store_event(u_int16_t source, u_int16_t idx,
+ gdt_evt_data *evt);
+int gdt_read_event(int handle, gdt_evt_str *estr);
+void gdt_readapp_event(u_int8_t app, gdt_evt_str *estr);
+void gdt_clear_events(void);
+
+#endif
diff --git a/sys/dev/iir/iir_ctrl.c b/sys/dev/iir/iir_ctrl.c
new file mode 100644
index 0000000..ca3fffe
--- /dev/null
+++ b/sys/dev/iir/iir_ctrl.c
@@ -0,0 +1,371 @@
+/* $FreeBSD$ */
+/*
+ * Copyright (c) 2000-01 Intel Corporation
+ * All Rights Reserved
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions
+ * are met:
+ * 1. Redistributions of source code must retain the above copyright
+ * notice, this list of conditions, and the following disclaimer,
+ * without modification, immediately at the beginning of the file.
+ * 2. Redistributions in binary form must reproduce the above copyright
+ * notice, this list of conditions and the following disclaimer in the
+ * documentation and/or other materials provided with the distribution.
+ * 3. The name of the author may not be used to endorse or promote products
+ * derived from this software without specific prior written permission.
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
+ * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
+ * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE FOR
+ * ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
+ * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
+ * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
+ * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
+ * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
+ * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
+ * SUCH DAMAGE.
+ */
+
+/*
+ * iir_ctrl.c: Control functions and /dev entry points for /dev/iir*
+ *
+ * Written by: Achim Leubner <achim.leubner@intel.com>
+ * Fixes/Additions: Boji Tony Kannanthanam <boji.t.kannanthanam@intel.com>
+ *
+ * TODO:
+ */
+
+#ident "$Id: iir_ctrl.c 1.2 2001/07/18 11:17:22 achim Exp $"
+
+#include <sys/param.h>
+#include <sys/systm.h>
+#include <sys/malloc.h>
+#include <sys/kernel.h>
+#include <sys/uio.h>
+#include <sys/conf.h>
+#include <sys/stat.h>
+#include <sys/disklabel.h>
+#include <machine/bus.h>
+#include <vm/vm.h>
+#include <vm/vm_kern.h>
+#include <vm/vm_extern.h>
+#include <vm/pmap.h>
+
+#include <dev/iir/iir.h>
+
+/* Entry points and other prototypes */
+static struct gdt_softc *gdt_minor2softc(int minor_no);
+
+static d_open_t iir_open;
+static d_close_t iir_close;
+static d_write_t iir_write;
+static d_read_t iir_read;
+static d_ioctl_t iir_ioctl;
+
+#define CDEV_MAJOR IIR_CDEV_MAJOR
+
+/* Normally, this is a static structure. But we need it in pci/iir_pci.c */
+static struct cdevsw iir_cdevsw = {
+ /* open */ iir_open,
+ /* close */ iir_close,
+ /* read */ iir_read,
+ /* write */ iir_write,
+ /* ioctl */ iir_ioctl,
+ /* poll */ nopoll,
+ /* mmap */ nommap,
+ /* strategy */ nostrategy,
+ /* name */ "iir",
+ /* maj */ CDEV_MAJOR,
+ /* dump */ nodump,
+ /* psize */ nopsize,
+ /* flags */ 0,
+ /* kq */ nokqfilter
+};
+
+static int iir_devsw_installed = 0;
+#ifndef SDEV_PER_HBA
+static int sdev_made = 0;
+#endif
+extern int gdt_cnt;
+extern char ostype[];
+extern char osrelease[];
+extern gdt_statist_t gdt_stat;
+
+/*
+ * Given a controller number,
+ * make a special device and return the dev_t
+ */
+dev_t
+gdt_make_dev(int unit)
+{
+ dev_t dev;
+
+#ifdef SDEV_PER_HBA
+ dev = make_dev(&iir_cdevsw, hba2minor(unit), UID_ROOT, GID_OPERATOR,
+ S_IRUSR | S_IWUSR | S_IRGRP | S_IROTH, "iir%d", unit);
+#else
+ if (sdev_made)
+ return (0);
+ dev = make_dev(&iir_cdevsw, 0, UID_ROOT, GID_OPERATOR,
+ S_IRUSR | S_IWUSR | S_IRGRP | S_IROTH, "iir");
+ sdev_made = 1;
+#endif
+ return (dev);
+}
+
+void
+gdt_destroy_dev(dev_t dev)
+{
+ if (dev != NULL)
+ destroy_dev(dev);
+}
+
+/*
+ * Given a minor device number,
+ * return the pointer to its softc structure
+ */
+static struct gdt_softc *
+gdt_minor2softc(int minor_no)
+{
+ struct gdt_softc *gdt;
+ int hanum;
+
+#ifdef SDEV_PER_HBA
+ hanum = minor2hba(minor_no);
+#else
+ hanum = minor_no;
+#endif
+
+ for (gdt = TAILQ_FIRST(&gdt_softcs);
+ gdt != NULL && gdt->sc_hanum != hanum;
+ gdt = TAILQ_NEXT(gdt, links));
+
+ return (gdt);
+}
+
+static int
+iir_open(dev_t dev, int flags, int fmt, d_thread_t * p)
+{
+ GDT_DPRINTF(GDT_D_DEBUG, ("iir_open()\n"));
+
+#ifdef SDEV_PER_HBA
+ int minor_no;
+ struct gdt_softc *gdt;
+
+ minor_no = minor(dev);
+ gdt = gdt_minor2softc(minor_no);
+ if (gdt == NULL)
+ return (ENXIO);
+#endif
+
+ return (0);
+}
+
+static int
+iir_close(dev_t dev, int flags, int fmt, d_thread_t * p)
+{
+ GDT_DPRINTF(GDT_D_DEBUG, ("iir_close()\n"));
+
+#ifdef SDEV_PER_HBA
+ int minor_no;
+ struct gdt_softc *gdt;
+
+ minor_no = minor(dev);
+ gdt = gdt_minor2softc(minor_no);
+ if (gdt == NULL)
+ return (ENXIO);
+#endif
+
+ return (0);
+}
+
+static int
+iir_write(dev_t dev, struct uio * uio, int ioflag)
+{
+ GDT_DPRINTF(GDT_D_DEBUG, ("iir_write()\n"));
+
+#ifdef SDEV_PER_HBA
+ int minor_no;
+ struct gdt_softc *gdt;
+
+ minor_no = minor(dev);
+ gdt = gdt_minor2softc(minor_no);
+ if (gdt == NULL)
+ return (ENXIO);
+#endif
+
+ return (0);
+}
+
+static int
+iir_read(dev_t dev, struct uio * uio, int ioflag)
+{
+ GDT_DPRINTF(GDT_D_DEBUG, ("iir_read()\n"));
+
+#ifdef SDEV_PER_HBA
+ int minor_no;
+ struct gdt_softc *gdt;
+
+ minor_no = minor(dev);
+ gdt = gdt_minor2softc(minor_no);
+ if (gdt == NULL)
+ return (ENXIO);
+#endif
+
+ return (0);
+}
+
+/**
+ * This is the control syscall interface.
+ * It should be binary compatible with UnixWare,
+ * if not totally syntatically so.
+ */
+
+static int
+iir_ioctl(dev_t dev, u_long cmd, caddr_t cmdarg, int flags, d_thread_t * p)
+{
+ GDT_DPRINTF(GDT_D_DEBUG, ("iir_ioctl() cmd 0x%lx\n",cmd));
+
+#ifdef SDEV_PER_HBA
+ int minor_no;
+ struct gdt_softc *gdt;
+
+ minor_no = minor(dev);
+ gdt = gdt_minor2softc(minor_no);
+ if (gdt == NULL)
+ return (ENXIO);
+#endif
+ ++gdt_stat.io_count_act;
+ if (gdt_stat.io_count_act > gdt_stat.io_count_max)
+ gdt_stat.io_count_max = gdt_stat.io_count_act;
+
+ switch (cmd) {
+ case GDT_IOCTL_GENERAL:
+ {
+ gdt_ucmd_t *ucmd;
+ struct gdt_softc *gdt;
+ int lock;
+
+ ucmd = (gdt_ucmd_t *)cmdarg;
+ gdt = gdt_minor2softc(ucmd->io_node);
+ if (gdt == NULL)
+ return (ENXIO);
+ lock = splcam();
+ TAILQ_INSERT_TAIL(&gdt->sc_ucmd_queue, ucmd, links);
+ ucmd->complete_flag = FALSE;
+ splx(lock);
+ gdt_next(gdt);
+ if (!ucmd->complete_flag)
+ (void) tsleep((void *)ucmd, PCATCH | PRIBIO, "iirucw", 0);
+ break;
+ }
+
+ case GDT_IOCTL_DRVERS:
+ *(int *)cmdarg =
+ (IIR_DRIVER_VERSION << 8) | IIR_DRIVER_SUBVERSION;
+ break;
+
+ case GDT_IOCTL_CTRTYPE:
+ {
+ gdt_ctrt_t *p;
+ struct gdt_softc *gdt;
+
+ p = (gdt_ctrt_t *)cmdarg;
+ gdt = gdt_minor2softc(p->io_node);
+ if (gdt == NULL)
+ return (ENXIO);
+ p->oem_id = 0x8000;
+ p->type = 0xfd;
+ p->info = (gdt->sc_bus << 8) | (gdt->sc_slot << 3);
+ p->ext_type = 0x6000 | gdt->sc_subdevice;
+ p->device_id = gdt->sc_device;
+ p->sub_device_id = gdt->sc_subdevice;
+ break;
+ }
+
+ case GDT_IOCTL_OSVERS:
+ {
+ gdt_osv_t *p;
+
+ p = (gdt_osv_t *)cmdarg;
+ p->oscode = 10;
+ p->version = osrelease[0] - '0';
+ if (osrelease[1] == '.')
+ p->subversion = osrelease[2] - '0';
+ else
+ p->subversion = 0;
+ if (osrelease[3] == '.')
+ p->revision = osrelease[4] - '0';
+ else
+ p->revision = 0;
+ strcpy(p->name, ostype);
+ break;
+ }
+
+ case GDT_IOCTL_CTRCNT:
+ *(int *)cmdarg = gdt_cnt;
+ break;
+
+ case GDT_IOCTL_EVENT:
+ {
+ gdt_event_t *p;
+ int lock;
+
+ p = (gdt_event_t *)cmdarg;
+ if (p->erase == 0xff) {
+ if (p->dvr.event_source == GDT_ES_TEST)
+ p->dvr.event_data.size = sizeof(p->dvr.event_data.eu.test);
+ else if (p->dvr.event_source == GDT_ES_DRIVER)
+ p->dvr.event_data.size= sizeof(p->dvr.event_data.eu.driver);
+ else if (p->dvr.event_source == GDT_ES_SYNC)
+ p->dvr.event_data.size = sizeof(p->dvr.event_data.eu.sync);
+ else
+ p->dvr.event_data.size = sizeof(p->dvr.event_data.eu.async);
+ lock = splcam();
+ gdt_store_event(p->dvr.event_source, p->dvr.event_idx,
+ &p->dvr.event_data);
+ splx(lock);
+ } else if (p->erase == 0xfe) {
+ lock = splcam();
+ gdt_clear_events();
+ splx(lock);
+ } else if (p->erase == 0) {
+ p->handle = gdt_read_event(p->handle, &p->dvr);
+ } else {
+ gdt_readapp_event((u_int8_t)p->erase, &p->dvr);
+ }
+ break;
+ }
+
+ case GDT_IOCTL_STATIST:
+ {
+ gdt_statist_t *p;
+
+ p = (gdt_statist_t *)cmdarg;
+ bcopy(&gdt_stat, p, sizeof(gdt_statist_t));
+ break;
+ }
+
+ default:
+ break;
+ }
+
+ --gdt_stat.io_count_act;
+ return (0);
+}
+
+static void
+iir_drvinit(void *unused)
+{
+ GDT_DPRINTF(GDT_D_DEBUG, ("iir_drvinit()\n"));
+
+ if (!iir_devsw_installed) {
+ /* Add the I/O (data) channel */
+ cdevsw_add(&iir_cdevsw);
+ iir_devsw_installed = 1;
+ }
+}
+
+SYSINIT(iir_dev, SI_SUB_DRIVERS, SI_ORDER_MIDDLE + CDEV_MAJOR, iir_drvinit, NULL)
diff --git a/sys/dev/iir/iir_pci.c b/sys/dev/iir/iir_pci.c
new file mode 100644
index 0000000..d751e3d
--- /dev/null
+++ b/sys/dev/iir/iir_pci.c
@@ -0,0 +1,476 @@
+/* $FreeBSD$ */
+/*
+ * Copyright (c) 2000-01 Intel Corporation
+ * All Rights Reserved
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions
+ * are met:
+ * 1. Redistributions of source code must retain the above copyright
+ * notice, this list of conditions, and the following disclaimer,
+ * without modification, immediately at the beginning of the file.
+ * 2. Redistributions in binary form must reproduce the above copyright
+ * notice, this list of conditions and the following disclaimer in the
+ * documentation and/or other materials provided with the distribution.
+ * 3. The name of the author may not be used to endorse or promote products
+ * derived from this software without specific prior written permission.
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
+ * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
+ * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE FOR
+ * ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
+ * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
+ * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
+ * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
+ * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
+ * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
+ * SUCH DAMAGE.
+ *
+ */
+
+/*
+ * iir_pci.c: PCI Bus Attachment for Intel Integrated RAID Controller driver
+ *
+ * Written by: Achim Leubner <achim.leubner@intel.com>
+ * Fixes/Additions: Boji Tony Kannanthanam <boji.t.kannanthanam@intel.com>
+ *
+ * TODO:
+ */
+
+#ident "$Id: iir_pci.c 1.1 2001/05/22 20:14:12 achim Exp $"
+
+/* #include "opt_iir.h" */
+
+#include <sys/param.h>
+#include <sys/systm.h>
+#include <sys/kernel.h>
+#include <sys/module.h>
+#include <sys/bus.h>
+
+#include <machine/bus_memio.h>
+#include <machine/bus_pio.h>
+#include <machine/bus.h>
+#include <machine/resource.h>
+#include <machine/clock.h>
+#include <sys/rman.h>
+
+#include <pci/pcireg.h>
+#include <pci/pcivar.h>
+
+#include <cam/scsi/scsi_all.h>
+
+#include <dev/iir/iir.h>
+
+/* Mapping registers for various areas */
+#define PCI_DPMEM PCIR_MAPS
+
+/* Product numbers for Fibre-Channel are greater than or equal to 0x200 */
+#define GDT_PCI_PRODUCT_FC 0x200
+
+/* PCI SRAM structure */
+#define GDT_MAGIC 0x00 /* u_int32_t, controller ID from BIOS */
+#define GDT_NEED_DEINIT 0x04 /* u_int16_t, switch between BIOS/driver */
+#define GDT_SWITCH_SUPPORT 0x06 /* u_int8_t, see GDT_NEED_DEINIT */
+#define GDT_OS_USED 0x10 /* u_int8_t [16], OS code per service */
+#define GDT_FW_MAGIC 0x3c /* u_int8_t, controller ID from firmware */
+#define GDT_SRAM_SZ 0x40
+
+/* DPRAM PCI controllers */
+#define GDT_DPR_IF 0x00 /* interface area */
+#define GDT_6SR (0xff0 - GDT_SRAM_SZ)
+#define GDT_SEMA1 0xff1 /* volatile u_int8_t, command semaphore */
+#define GDT_IRQEN 0xff5 /* u_int8_t, board interrupts enable */
+#define GDT_EVENT 0xff8 /* u_int8_t, release event */
+#define GDT_IRQDEL 0xffc /* u_int8_t, acknowledge board interrupt */
+#define GDT_DPRAM_SZ 0x1000
+
+/* PLX register structure (new PCI controllers) */
+#define GDT_CFG_REG 0x00 /* u_int8_t, DPRAM cfg. (2: < 1MB, 0: any) */
+#define GDT_SEMA0_REG 0x40 /* volatile u_int8_t, command semaphore */
+#define GDT_SEMA1_REG 0x41 /* volatile u_int8_t, status semaphore */
+#define GDT_PLX_STATUS 0x44 /* volatile u_int16_t, command status */
+#define GDT_PLX_SERVICE 0x46 /* u_int16_t, service */
+#define GDT_PLX_INFO 0x48 /* u_int32_t [2], additional info */
+#define GDT_LDOOR_REG 0x60 /* u_int8_t, PCI to local doorbell */
+#define GDT_EDOOR_REG 0x64 /* volatile u_int8_t, local to PCI doorbell */
+#define GDT_CONTROL0 0x68 /* u_int8_t, control0 register (unused) */
+#define GDT_CONTROL1 0x69 /* u_int8_t, board interrupts enable */
+#define GDT_PLX_SZ 0x80
+
+/* DPRAM new PCI controllers */
+#define GDT_IC 0x00 /* interface */
+#define GDT_PCINEW_6SR (0x4000 - GDT_SRAM_SZ)
+ /* SRAM structure */
+#define GDT_PCINEW_SZ 0x4000
+
+/* i960 register structure (PCI MPR controllers) */
+#define GDT_MPR_SEMA0 0x10 /* volatile u_int8_t, command semaphore */
+#define GDT_MPR_SEMA1 0x12 /* volatile u_int8_t, status semaphore */
+#define GDT_MPR_STATUS 0x14 /* volatile u_int16_t, command status */
+#define GDT_MPR_SERVICE 0x16 /* u_int16_t, service */
+#define GDT_MPR_INFO 0x18 /* u_int32_t [2], additional info */
+#define GDT_MPR_LDOOR 0x20 /* u_int8_t, PCI to local doorbell */
+#define GDT_MPR_EDOOR 0x2c /* volatile u_int8_t, locl to PCI doorbell */
+#define GDT_EDOOR_EN 0x34 /* u_int8_t, board interrupts enable */
+#define GDT_SEVERITY 0xefc /* u_int8_t, event severity */
+#define GDT_EVT_BUF 0xf00 /* u_int8_t [256], event buffer */
+#define GDT_I960_SZ 0x1000
+
+/* DPRAM PCI MPR controllers */
+#define GDT_I960R 0x00 /* 4KB i960 registers */
+#define GDT_MPR_IC GDT_I960_SZ
+ /* i960 register area */
+#define GDT_MPR_6SR (GDT_I960_SZ + 0x3000 - GDT_SRAM_SZ)
+ /* DPRAM struct. */
+#define GDT_MPR_SZ (0x3000 - GDT_SRAM_SZ)
+
+static int iir_pci_probe __P((device_t dev));
+static int iir_pci_attach __P((device_t dev));
+
+void gdt_pci_enable_intr __P((struct gdt_softc *));
+
+void gdt_mpr_copy_cmd __P((struct gdt_softc *, struct gdt_ccb *));
+u_int8_t gdt_mpr_get_status __P((struct gdt_softc *));
+void gdt_mpr_intr __P((struct gdt_softc *, struct gdt_intr_ctx *));
+void gdt_mpr_release_event __P((struct gdt_softc *));
+void gdt_mpr_set_sema0 __P((struct gdt_softc *));
+int gdt_mpr_test_busy __P((struct gdt_softc *));
+
+static device_method_t iir_pci_methods[] = {
+ /* Device interface */
+ DEVMETHOD(device_probe, iir_pci_probe),
+ DEVMETHOD(device_attach, iir_pci_attach),
+ { 0, 0}
+};
+
+
+static driver_t iir_pci_driver =
+{
+ "iir",
+ iir_pci_methods,
+ sizeof(struct gdt_softc)
+};
+
+static devclass_t iir_devclass;
+
+DRIVER_MODULE(iir, pci, iir_pci_driver, iir_devclass, 0, 0);
+
+static int
+iir_pci_probe(device_t dev)
+{
+ if (pci_get_vendor(dev) == INTEL_VENDOR_ID &&
+ pci_get_device(dev) == INTEL_DEVICE_ID_IIR) {
+ device_set_desc(dev, "Intel Integrated RAID Controller");
+ return (0);
+ }
+ if (pci_get_vendor(dev) == GDT_VENDOR_ID &&
+ ((pci_get_device(dev) >= GDT_DEVICE_ID_MIN &&
+ pci_get_device(dev) <= GDT_DEVICE_ID_MAX) ||
+ pci_get_device(dev) == GDT_DEVICE_ID_NEWRX)) {
+ device_set_desc(dev, "ICP Disk Array Controller");
+ return (0);
+ }
+ return (ENXIO);
+}
+
+
+static int
+iir_pci_attach(device_t dev)
+{
+ struct gdt_softc *gdt;
+ struct resource *io = NULL, *irq = NULL;
+ int retries, rid, error = 0;
+ void *ih;
+ u_int8_t protocol;
+
+ /* map DPMEM */
+ rid = PCI_DPMEM;
+ io = bus_alloc_resource(dev, SYS_RES_MEMORY, &rid, 0, ~0, 1, RF_ACTIVE);
+ if (io == NULL) {
+ device_printf(dev, "can't allocate register resources\n");
+ error = ENOMEM;
+ goto err;
+ }
+
+ /* get IRQ */
+ rid = 0;
+ irq = bus_alloc_resource(dev, SYS_RES_IRQ, &rid, 0, ~0, 1,
+ RF_ACTIVE | RF_SHAREABLE);
+ if (io == NULL) {
+ device_printf(dev, "can't find IRQ value\n");
+ error = ENOMEM;
+ goto err;
+ }
+
+ gdt = device_get_softc(dev);
+ bzero(gdt, sizeof(struct gdt_softc));
+ gdt->sc_init_level = 0;
+ gdt->sc_dpmemt = rman_get_bustag(io);
+ gdt->sc_dpmemh = rman_get_bushandle(io);
+ gdt->sc_dpmembase = rman_get_start(io);
+ gdt->sc_hanum = device_get_unit(dev);
+ gdt->sc_bus = pci_get_bus(dev);
+ gdt->sc_slot = pci_get_slot(dev);
+ gdt->sc_device = pci_get_device(dev);
+ gdt->sc_subdevice = pci_get_subdevice(dev);
+ gdt->sc_class = GDT_MPR;
+/* no FC ctr.
+ if (gdt->sc_device >= GDT_PCI_PRODUCT_FC)
+ gdt->sc_class |= GDT_FC;
+*/
+
+ /* initialize RP controller */
+ /* check and reset interface area */
+ bus_space_write_4(gdt->sc_dpmemt, gdt->sc_dpmemh, GDT_MPR_IC,
+ htole32(GDT_MPR_MAGIC));
+ if (bus_space_read_4(gdt->sc_dpmemt, gdt->sc_dpmemh, GDT_MPR_IC) !=
+ htole32(GDT_MPR_MAGIC)) {
+ printf("cannot access DPMEM at 0x%x (shadowed?)\n",
+ gdt->sc_dpmembase);
+ error = ENXIO;
+ goto err;
+ }
+ bus_space_set_region_4(gdt->sc_dpmemt, gdt->sc_dpmemh, GDT_I960_SZ, htole32(0),
+ GDT_MPR_SZ >> 2);
+
+ /* Disable everything */
+ bus_space_write_1(gdt->sc_dpmemt, gdt->sc_dpmemh, GDT_EDOOR_EN,
+ bus_space_read_1(gdt->sc_dpmemt, gdt->sc_dpmemh,
+ GDT_EDOOR_EN) | 4);
+ bus_space_write_1(gdt->sc_dpmemt, gdt->sc_dpmemh, GDT_MPR_EDOOR, 0xff);
+ bus_space_write_1(gdt->sc_dpmemt, gdt->sc_dpmemh, GDT_MPR_IC + GDT_S_STATUS,
+ 0);
+ bus_space_write_1(gdt->sc_dpmemt, gdt->sc_dpmemh, GDT_MPR_IC + GDT_CMD_INDEX,
+ 0);
+
+ bus_space_write_4(gdt->sc_dpmemt, gdt->sc_dpmemh, GDT_MPR_IC + GDT_S_INFO,
+ htole32(gdt->sc_dpmembase));
+ bus_space_write_1(gdt->sc_dpmemt, gdt->sc_dpmemh, GDT_MPR_IC + GDT_S_CMD_INDX,
+ 0xff);
+ bus_space_write_1(gdt->sc_dpmemt, gdt->sc_dpmemh, GDT_MPR_LDOOR, 1);
+
+ DELAY(20);
+ retries = GDT_RETRIES;
+ while (bus_space_read_1(gdt->sc_dpmemt, gdt->sc_dpmemh,
+ GDT_MPR_IC + GDT_S_STATUS) != 0xff) {
+ if (--retries == 0) {
+ printf("DEINIT failed\n");
+ error = ENXIO;
+ goto err;
+ }
+ DELAY(1);
+ }
+
+ protocol = (u_int8_t)letoh32(bus_space_read_4(gdt->sc_dpmemt, gdt->sc_dpmemh,
+ GDT_MPR_IC + GDT_S_INFO));
+ bus_space_write_1(gdt->sc_dpmemt, gdt->sc_dpmemh, GDT_MPR_IC + GDT_S_STATUS,
+ 0);
+ if (protocol != GDT_PROTOCOL_VERSION) {
+ printf("unsupported protocol %d\n", protocol);
+ error = ENXIO;
+ goto err;
+ }
+
+ /* special commnd to controller BIOS */
+ bus_space_write_4(gdt->sc_dpmemt, gdt->sc_dpmemh, GDT_MPR_IC + GDT_S_INFO,
+ htole32(0));
+ bus_space_write_4(gdt->sc_dpmemt, gdt->sc_dpmemh,
+ GDT_MPR_IC + GDT_S_INFO + sizeof (u_int32_t), htole32(0));
+ bus_space_write_4(gdt->sc_dpmemt, gdt->sc_dpmemh,
+ GDT_MPR_IC + GDT_S_INFO + 2 * sizeof (u_int32_t),
+ htole32(1));
+ bus_space_write_4(gdt->sc_dpmemt, gdt->sc_dpmemh,
+ GDT_MPR_IC + GDT_S_INFO + 3 * sizeof (u_int32_t),
+ htole32(0));
+ bus_space_write_1(gdt->sc_dpmemt, gdt->sc_dpmemh, GDT_MPR_IC + GDT_S_CMD_INDX,
+ 0xfe);
+ bus_space_write_1(gdt->sc_dpmemt, gdt->sc_dpmemh, GDT_MPR_LDOOR, 1);
+
+ DELAY(20);
+ retries = GDT_RETRIES;
+ while (bus_space_read_1(gdt->sc_dpmemt, gdt->sc_dpmemh,
+ GDT_MPR_IC + GDT_S_STATUS) != 0xfe) {
+ if (--retries == 0) {
+ printf("initialization error\n");
+ error = ENXIO;
+ goto err;
+ }
+ DELAY(1);
+ }
+
+ bus_space_write_1(gdt->sc_dpmemt, gdt->sc_dpmemh, GDT_MPR_IC + GDT_S_STATUS,
+ 0);
+
+ gdt->sc_ic_all_size = GDT_MPR_SZ;
+
+ gdt->sc_copy_cmd = gdt_mpr_copy_cmd;
+ gdt->sc_get_status = gdt_mpr_get_status;
+ gdt->sc_intr = gdt_mpr_intr;
+ gdt->sc_release_event = gdt_mpr_release_event;
+ gdt->sc_set_sema0 = gdt_mpr_set_sema0;
+ gdt->sc_test_busy = gdt_mpr_test_busy;
+
+ /* Allocate a dmatag representing the capabilities of this attachment */
+ /* XXX Should be a child of the PCI bus dma tag */
+ if (bus_dma_tag_create(/*parent*/NULL, /*alignemnt*/1, /*boundary*/0,
+ /*lowaddr*/BUS_SPACE_MAXADDR_32BIT,
+ /*highaddr*/BUS_SPACE_MAXADDR,
+ /*filter*/NULL, /*filterarg*/NULL,
+ /*maxsize*/BUS_SPACE_MAXSIZE_32BIT,
+ /*nsegments*/GDT_MAXSG,
+ /*maxsegsz*/BUS_SPACE_MAXSIZE_32BIT,
+ /*flags*/0, &gdt->sc_parent_dmat) != 0) {
+ error = ENXIO;
+ goto err;
+ }
+ gdt->sc_init_level++;
+
+ if (iir_init(gdt) != 0) {
+ iir_free(gdt);
+ error = ENXIO;
+ goto err;
+ }
+
+ /* Register with the XPT */
+ iir_attach(gdt);
+
+ /* associate interrupt handler */
+ if (bus_setup_intr( dev, irq, INTR_TYPE_CAM,
+ iir_intr, gdt, &ih )) {
+ device_printf(dev, "Unable to register interrupt handler\n");
+ error = ENXIO;
+ goto err;
+ }
+
+ gdt_pci_enable_intr(gdt);
+ return (0);
+
+err:
+ if (irq)
+ bus_release_resource( dev, SYS_RES_IRQ, 0, irq );
+/*
+ if (io)
+ bus_release_resource( dev, SYS_RES_MEMORY, rid, io );
+*/
+ return (error);
+}
+
+
+/* Enable interrupts */
+void
+gdt_pci_enable_intr(struct gdt_softc *gdt)
+{
+ GDT_DPRINTF(GDT_D_INTR, ("gdt_pci_enable_intr(%p) ", gdt));
+
+ switch(GDT_CLASS(gdt)) {
+ case GDT_MPR:
+ bus_space_write_1(gdt->sc_dpmemt, gdt->sc_dpmemh,
+ GDT_MPR_EDOOR, 0xff);
+ bus_space_write_1(gdt->sc_dpmemt, gdt->sc_dpmemh, GDT_EDOOR_EN,
+ bus_space_read_1(gdt->sc_dpmemt, gdt->sc_dpmemh,
+ GDT_EDOOR_EN) & ~4);
+ break;
+ }
+}
+
+
+/*
+ * MPR PCI controller-specific functions
+ */
+
+void
+gdt_mpr_copy_cmd(struct gdt_softc *gdt, struct gdt_ccb *ccb)
+{
+ u_int16_t cp_count = roundup(gdt->sc_cmd_len, sizeof (u_int32_t));
+ u_int16_t dp_offset = gdt->sc_cmd_off;
+ u_int16_t cmd_no = gdt->sc_cmd_cnt++;
+
+ GDT_DPRINTF(GDT_D_CMD, ("gdt_mpr_copy_cmd(%p) ", gdt));
+
+ gdt->sc_cmd_off += cp_count;
+
+ bus_space_write_2(gdt->sc_dpmemt, gdt->sc_dpmemh,
+ GDT_MPR_IC + GDT_COMM_QUEUE + cmd_no * GDT_COMM_Q_SZ + GDT_OFFSET,
+ htole16(GDT_DPMEM_COMMAND_OFFSET + dp_offset));
+ bus_space_write_2(gdt->sc_dpmemt, gdt->sc_dpmemh,
+ GDT_MPR_IC + GDT_COMM_QUEUE + cmd_no * GDT_COMM_Q_SZ + GDT_SERV_ID,
+ htole16(ccb->gc_service));
+ bus_space_write_region_4(gdt->sc_dpmemt, gdt->sc_dpmemh,
+ GDT_MPR_IC + GDT_DPR_CMD + dp_offset,
+ (u_int32_t *)gdt->sc_cmd, cp_count >> 2);
+}
+
+u_int8_t
+gdt_mpr_get_status(struct gdt_softc *gdt)
+{
+ GDT_DPRINTF(GDT_D_MISC, ("gdt_mpr_get_status(%p) ", gdt));
+
+ return bus_space_read_1(gdt->sc_dpmemt, gdt->sc_dpmemh, GDT_MPR_EDOOR);
+}
+
+void
+gdt_mpr_intr(struct gdt_softc *gdt, struct gdt_intr_ctx *ctx)
+{
+ int i;
+
+ GDT_DPRINTF(GDT_D_INTR, ("gdt_mpr_intr(%p) ", gdt));
+
+ if (ctx->istatus & 0x80) { /* error flag */
+ ctx->istatus &= ~0x80;
+ ctx->cmd_status = bus_space_read_2(gdt->sc_dpmemt,
+ gdt->sc_dpmemh, GDT_MPR_STATUS);
+ } else /* no error */
+ ctx->cmd_status = GDT_S_OK;
+
+ ctx->info =
+ bus_space_read_4(gdt->sc_dpmemt, gdt->sc_dpmemh, GDT_MPR_INFO);
+ ctx->service =
+ bus_space_read_2(gdt->sc_dpmemt, gdt->sc_dpmemh, GDT_MPR_SERVICE);
+ ctx->info2 =
+ bus_space_read_4(gdt->sc_dpmemt, gdt->sc_dpmemh,
+ GDT_MPR_INFO + sizeof (u_int32_t));
+
+ /* event string */
+ if (ctx->istatus == GDT_ASYNCINDEX) {
+ if (ctx->service != GDT_SCREENSERVICE &&
+ (gdt->sc_fw_vers & 0xff) >= 0x1a) {
+ gdt->sc_dvr.severity =
+ bus_space_read_1(gdt->sc_dpmemt,gdt->sc_dpmemh, GDT_SEVERITY);
+ for (i = 0; i < 256; ++i) {
+ gdt->sc_dvr.event_string[i] =
+ bus_space_read_1(gdt->sc_dpmemt, gdt->sc_dpmemh,
+ GDT_EVT_BUF + i);
+ if (gdt->sc_dvr.event_string[i] == 0)
+ break;
+ }
+ }
+ }
+ bus_space_write_1(gdt->sc_dpmemt, gdt->sc_dpmemh, GDT_MPR_EDOOR, 0xff);
+ bus_space_write_1(gdt->sc_dpmemt, gdt->sc_dpmemh, GDT_MPR_SEMA1, 0);
+}
+
+void
+gdt_mpr_release_event(struct gdt_softc *gdt)
+{
+ GDT_DPRINTF(GDT_D_MISC, ("gdt_mpr_release_event(%p) ", gdt));
+
+ bus_space_write_1(gdt->sc_dpmemt, gdt->sc_dpmemh, GDT_MPR_LDOOR, 1);
+}
+
+void
+gdt_mpr_set_sema0(struct gdt_softc *gdt)
+{
+ GDT_DPRINTF(GDT_D_MISC, ("gdt_mpr_set_sema0(%p) ", gdt));
+
+ bus_space_write_1(gdt->sc_dpmemt, gdt->sc_dpmemh, GDT_MPR_SEMA0, 1);
+}
+
+int
+gdt_mpr_test_busy(struct gdt_softc *gdt)
+{
+ GDT_DPRINTF(GDT_D_MISC, ("gdt_mpr_test_busy(%p) ", gdt));
+
+ return (bus_space_read_1(gdt->sc_dpmemt, gdt->sc_dpmemh,
+ GDT_MPR_SEMA0) & 1);
+}
diff --git a/sys/i386/conf/GENERIC b/sys/i386/conf/GENERIC
index f062d27..dec9a76 100644
--- a/sys/i386/conf/GENERIC
+++ b/sys/i386/conf/GENERIC
@@ -101,7 +101,9 @@ device stg # TMC 18C30/18C50
# RAID controllers interfaced to the SCSI subsystem
device asr # DPT SmartRAID V, VI and Adaptec SCSI RAID
+device ciss # Compaq Smart RAID 5*
device dpt # DPT Smartcache III, IV - See NOTES for options!
+device iir # Intel Integrated RAID
device mly # Mylex AcceleRAID/eXtremeRAID
# SCSI peripherals
diff --git a/sys/i386/conf/NOTES b/sys/i386/conf/NOTES
index ad17181..1311973 100644
--- a/sys/i386/conf/NOTES
+++ b/sys/i386/conf/NOTES
@@ -1546,6 +1546,15 @@ options DPT_ALLOW_MEMIO
device ciss
#
+# Intel Integrated RAID controllers.
+# This driver was developed and is maintained by Intel. Contacts
+# at Intel for this driver are
+# "Kannanthanam, Boji T" <boji.t.kannanthanam@intel.com> and
+# "Leubner, Achim" <achim.leubner@intel.com>.
+#
+device iir
+
+#
# Mylex AcceleRAID and eXtremeRAID controllers with v6 and later
# firmware. These controllers have a SCSI-like interface, and require
# the CAM infrastructure.
diff --git a/sys/modules/Makefile b/sys/modules/Makefile
index 55427dd..355f33c 100644
--- a/sys/modules/Makefile
+++ b/sys/modules/Makefile
@@ -131,6 +131,7 @@ SUBDIR+=aac \
fpu \
gnufpu \
ibcs2 \
+ iir \
linux \
ncv \
nsp \
diff --git a/sys/modules/iir/Makefile b/sys/modules/iir/Makefile
new file mode 100644
index 0000000..591fef8
--- /dev/null
+++ b/sys/modules/iir/Makefile
@@ -0,0 +1,9 @@
+# $FreeBSD$
+
+KMOD = iir
+.PATH: ${.CURDIR}/../../dev/${KMOD}
+SRCS = iir.c iir_ctrl.c iir_pci.c
+SRCS += opt_scsi.h opt_cam.h
+SRCS += device_if.h bus_if.h pci_if.h
+
+.include <bsd.kmod.mk>
OpenPOWER on IntegriCloud