summaryrefslogtreecommitdiffstats
path: root/sys/dev/ata/ata-disk.c
diff options
context:
space:
mode:
Diffstat (limited to 'sys/dev/ata/ata-disk.c')
-rw-r--r--sys/dev/ata/ata-disk.c895
1 files changed, 895 insertions, 0 deletions
diff --git a/sys/dev/ata/ata-disk.c b/sys/dev/ata/ata-disk.c
new file mode 100644
index 0000000..b7b274d
--- /dev/null
+++ b/sys/dev/ata/ata-disk.c
@@ -0,0 +1,895 @@
+/*-
+ * Copyright (c) 1998 - 2003 Søren Schmidt <sos@FreeBSD.org>
+ * All rights reserved.
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions
+ * are met:
+ * 1. Redistributions of source code must retain the above copyright
+ * notice, this list of conditions and the following disclaimer,
+ * without modification, immediately at the beginning of the file.
+ * 2. Redistributions in binary form must reproduce the above copyright
+ * notice, this list of conditions and the following disclaimer in the
+ * documentation and/or other materials provided with the distribution.
+ * 3. The name of the author may not be used to endorse or promote products
+ * derived from this software without specific prior written permission.
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
+ * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
+ * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
+ * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
+ * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
+ * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
+ * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
+ * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
+ * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
+ *
+ * $FreeBSD$
+ */
+
+#include "opt_ata.h"
+#include <sys/param.h>
+#include <sys/systm.h>
+#include <sys/ata.h>
+#include <sys/kernel.h>
+#include <sys/malloc.h>
+#include <sys/bio.h>
+#include <sys/bus.h>
+#include <sys/conf.h>
+#include <sys/disk.h>
+#include <sys/cons.h>
+#include <sys/sysctl.h>
+#include <vm/vm.h>
+#include <vm/pmap.h>
+#include <machine/md_var.h>
+#include <machine/bus.h>
+#include <sys/rman.h>
+#include <geom/geom_disk.h>
+#include <dev/ata/ata-all.h>
+#include <dev/ata/ata-pci.h>
+#include <dev/ata/ata-disk.h>
+#include <dev/ata/ata-raid.h>
+
+/* prototypes */
+static disk_open_t adopen;
+static disk_strategy_t adstrategy;
+static dumper_t addump;
+static void ad_invalidatequeue(struct ad_softc *, struct ad_request *);
+static int ad_tagsupported(struct ad_softc *);
+static void ad_timeout(struct ad_request *);
+static void ad_free(struct ad_request *);
+static int ad_version(u_int16_t);
+
+/* misc defines */
+#define AD_MAX_RETRIES 3
+
+/* internal vars */
+static u_int32_t adp_lun_map = 0;
+static int ata_dma = 1;
+static int ata_wc = 1;
+static int ata_tags = 0;
+TUNABLE_INT("hw.ata.ata_dma", &ata_dma);
+TUNABLE_INT("hw.ata.wc", &ata_wc);
+TUNABLE_INT("hw.ata.tags", &ata_tags);
+static MALLOC_DEFINE(M_AD, "AD driver", "ATA disk driver");
+
+/* sysctl vars */
+SYSCTL_DECL(_hw_ata);
+SYSCTL_INT(_hw_ata, OID_AUTO, ata_dma, CTLFLAG_RD, &ata_dma, 0,
+ "ATA disk DMA mode control");
+SYSCTL_INT(_hw_ata, OID_AUTO, wc, CTLFLAG_RD, &ata_wc, 0,
+ "ATA disk write caching");
+SYSCTL_INT(_hw_ata, OID_AUTO, tags, CTLFLAG_RD, &ata_tags, 0,
+ "ATA disk tagged queuing support");
+
+void
+ad_attach(struct ata_device *atadev)
+{
+ struct ad_softc *adp;
+ u_int32_t lbasize;
+ u_int64_t lbasize48;
+
+ if (!(adp = malloc(sizeof(struct ad_softc), M_AD, M_NOWAIT | M_ZERO))) {
+ ata_prtdev(atadev, "failed to allocate driver storage\n");
+ return;
+ }
+ adp->device = atadev;
+#ifdef ATA_STATIC_ID
+ adp->lun = (device_get_unit(atadev->channel->dev)<<1)+ATA_DEV(atadev->unit);
+#else
+ adp->lun = ata_get_lun(&adp_lun_map);
+#endif
+ ata_set_name(atadev, "ad", adp->lun);
+ adp->heads = atadev->param->heads;
+ adp->sectors = atadev->param->sectors;
+ adp->total_secs = atadev->param->cylinders * adp->heads * adp->sectors;
+ adp->max_iosize = 256 * DEV_BSIZE;
+ if (adp->device->channel->flags & ATA_USE_PC98GEOM &&
+ adp->total_secs < 17 * 8 * 65536) {
+ adp->sectors = 17;
+ adp->heads = 8;
+ }
+ bioq_init(&adp->queue);
+
+ lbasize = (u_int32_t)atadev->param->lba_size_1 |
+ ((u_int32_t)atadev->param->lba_size_2 << 16);
+
+ /* does this device need oldstyle CHS addressing */
+ if (!ad_version(atadev->param->version_major) ||
+ !(atadev->param->atavalid & ATA_FLAG_54_58) || !lbasize)
+ adp->flags |= AD_F_CHS_USED;
+
+ /* use the 28bit LBA size if valid or bigger than the CHS mapping */
+ if (atadev->param->cylinders == 16383 || adp->total_secs < lbasize)
+ adp->total_secs = lbasize;
+
+ lbasize48 = ((u_int64_t)atadev->param->lba_size48_1) |
+ ((u_int64_t)atadev->param->lba_size48_2 << 16) |
+ ((u_int64_t)atadev->param->lba_size48_3 << 32) |
+ ((u_int64_t)atadev->param->lba_size48_4 << 48);
+
+ /* use the 48bit LBA size if valid */
+ if (atadev->param->support.address48 && lbasize48 > 268435455)
+ adp->total_secs = lbasize48;
+
+ ATA_SLEEPLOCK_CH(atadev->channel, ATA_CONTROL);
+
+ /* use multiple sectors/interrupt if device supports it */
+ adp->transfersize = DEV_BSIZE;
+ if (ad_version(atadev->param->version_major)) {
+ int secsperint = max(1, min(atadev->param->sectors_intr, 16));
+
+ if (!ata_command(atadev, ATA_C_SET_MULTI, 0, secsperint,
+ 0, ATA_WAIT_INTR) && !ata_wait(atadev, 0))
+ adp->transfersize *= secsperint;
+ }
+
+ /* enable read caching if not default on device */
+ if (ata_command(atadev, ATA_C_SETFEATURES,
+ 0, 0, ATA_C_F_ENAB_RCACHE, ATA_WAIT_INTR))
+ ata_prtdev(atadev, "enabling readahead cache failed\n");
+
+ /* enable write caching if allowed and not default on device */
+ if (ata_wc || (ata_tags && ad_tagsupported(adp))) {
+ if (ata_command(atadev, ATA_C_SETFEATURES,
+ 0, 0, ATA_C_F_ENAB_WCACHE, ATA_WAIT_INTR))
+ ata_prtdev(atadev, "enabling write cache failed\n");
+ }
+ else {
+ if (ata_command(atadev, ATA_C_SETFEATURES,
+ 0, 0, ATA_C_F_DIS_WCACHE, ATA_WAIT_INTR))
+ ata_prtdev(atadev, "disabling write cache failed\n");
+ }
+
+ /* use DMA if allowed and if drive/controller supports it */
+ if (ata_dma && atadev->channel->dma)
+ atadev->setmode(atadev, ATA_DMA_MAX);
+ else
+ atadev->setmode(atadev, ATA_PIO_MAX);
+
+ /* use tagged queueing if allowed and supported */
+#if 0 /* disable tags for now */
+ if (ata_tags && ad_tagsupported(adp)) {
+ adp->num_tags = atadev->param->queuelen;
+ adp->flags |= AD_F_TAG_ENABLED;
+ adp->device->channel->flags |= ATA_QUEUED;
+ if (ata_command(atadev, ATA_C_SETFEATURES,
+ 0, 0, ATA_C_F_DIS_RELIRQ, ATA_WAIT_INTR))
+ ata_prtdev(atadev, "disabling release interrupt failed\n");
+ if (ata_command(atadev, ATA_C_SETFEATURES,
+ 0, 0, ATA_C_F_DIS_SRVIRQ, ATA_WAIT_INTR))
+ ata_prtdev(atadev, "disabling service interrupt failed\n");
+ }
+#endif
+ ATA_UNLOCK_CH(atadev->channel);
+
+ adp->disk.d_open = adopen;
+ adp->disk.d_strategy = adstrategy;
+ adp->disk.d_dump = addump;
+ adp->disk.d_name = "ad";
+ adp->disk.d_drv1 = adp;
+ adp->disk.d_maxsize = adp->max_iosize;
+ adp->disk.d_sectorsize = DEV_BSIZE;
+ adp->disk.d_mediasize = DEV_BSIZE * (off_t)adp->total_secs;
+ adp->disk.d_fwsectors = adp->sectors;
+ adp->disk.d_fwheads = adp->heads;
+ disk_create(adp->lun, &adp->disk, 0, NULL, NULL);
+
+ atadev->driver = adp;
+ atadev->flags = 0;
+
+ ata_enclosure_print(atadev);
+ if (atadev->driver)
+ ad_print(adp);
+ ata_raiddisk_attach(adp);
+}
+
+void
+ad_detach(struct ata_device *atadev)
+{
+ struct ad_softc *adp = atadev->driver;
+ struct ad_request *request;
+
+ atadev->flags |= ATA_D_DETACHING;
+ ata_prtdev(atadev, "removed from configuration\n");
+ ad_invalidatequeue(adp, NULL);
+ TAILQ_FOREACH(request, &atadev->channel->ata_queue, chain) {
+ if (request->softc != adp)
+ continue;
+ TAILQ_REMOVE(&atadev->channel->ata_queue, request, chain);
+ biofinish(request->bp, NULL, ENXIO);
+ ad_free(request);
+ }
+ bioq_flush(&adp->queue, NULL, ENXIO);
+ disk_destroy(&adp->disk);
+
+ if (adp->flags & AD_F_RAID_SUBDISK)
+ ata_raiddisk_detach(adp);
+
+ ata_free_name(atadev);
+ ata_free_lun(&adp_lun_map, adp->lun);
+ atadev->driver = NULL;
+ atadev->flags = 0;
+ free(adp, M_AD);
+}
+
+static int
+adopen(struct disk *dp)
+{
+ struct ad_softc *adp = dp->d_drv1;
+
+ if (adp->flags & AD_F_RAID_SUBDISK)
+ return EPERM;
+
+ /* hold off access until we are fully attached */
+ while (ata_delayed_attach)
+ tsleep(&ata_delayed_attach, PRIBIO, "adopn", 1);
+ return 0;
+}
+
+static void
+adstrategy(struct bio *bp)
+{
+ struct ad_softc *adp = bp->bio_disk->d_drv1;
+ int s;
+
+ if (adp->device->flags & ATA_D_DETACHING) {
+ biofinish(bp, NULL, ENXIO);
+ return;
+ }
+ s = splbio();
+ bioq_disksort(&adp->queue, bp);
+ splx(s);
+ ata_start(adp->device->channel);
+}
+
+static int
+addump(void *arg, void *virtual, vm_offset_t physical,
+ off_t offset, size_t length)
+{
+ struct ad_softc *adp;
+ struct ad_request request;
+ static int once;
+ struct disk *dp;
+
+ dp = arg;
+ adp = dp->d_drv1;
+ if (!adp)
+ return ENXIO;
+
+ if (!once) {
+ /* force PIO mode for dumps */
+ adp->device->mode = ATA_PIO;
+ adp->device->channel->locking(adp->device->channel, ATA_LF_LOCK);
+ ata_reinit(adp->device->channel);
+ adp->device->channel->locking(adp->device->channel, ATA_LF_UNLOCK);
+ once = 1;
+ }
+
+ if (length > 0) {
+ bzero(&request, sizeof(struct ad_request));
+ request.softc = adp;
+ request.blockaddr = offset / DEV_BSIZE;
+ request.bytecount = length;
+ request.data = virtual;
+
+ while (request.bytecount > 0) {
+ ad_transfer(&request);
+ if (request.flags & ADR_F_ERROR)
+ return EIO;
+ request.donecount += request.currentsize;
+ request.bytecount -= request.currentsize;
+ DELAY(20);
+ }
+ } else {
+ if (ata_wait(adp->device, ATA_S_READY | ATA_S_DSC) < 0)
+ ata_prtdev(adp->device, "timeout waiting for final ready\n");
+ }
+ return 0;
+}
+
+void
+ad_start(struct ata_device *atadev)
+{
+ struct ad_softc *adp = atadev->driver;
+ struct bio *bp = bioq_first(&adp->queue);
+ struct ad_request *request;
+ int tag = 0;
+
+ if (!bp)
+ return;
+
+ /* if tagged queueing enabled get next free tag */
+ if (adp->flags & AD_F_TAG_ENABLED) {
+ while (tag <= adp->num_tags && adp->tags[tag])
+ tag++;
+ if (tag > adp->num_tags )
+ return;
+ }
+
+ /* remove request from drive queue */
+ bioq_remove(&adp->queue, bp);
+
+ if (!(request = malloc(sizeof(struct ad_request), M_AD, M_NOWAIT|M_ZERO))) {
+ ata_prtdev(atadev, "out of memory in start\n");
+ biofinish(bp, NULL, ENOMEM);
+ return;
+ }
+
+ /* setup request */
+ request->softc = adp;
+ request->bp = bp;
+ request->blockaddr = bp->bio_pblkno;
+ request->bytecount = bp->bio_bcount;
+ request->data = bp->bio_data;
+ request->tag = tag;
+ if (bp->bio_cmd == BIO_READ)
+ request->flags |= ADR_F_READ;
+
+ if (adp->device->mode >= ATA_DMA && !atadev->channel->dma)
+ adp->device->mode = ATA_PIO;
+
+ /* insert in tag array */
+ adp->tags[tag] = request;
+
+ /* link onto controller queue */
+ TAILQ_INSERT_TAIL(&atadev->channel->ata_queue, request, chain);
+}
+
+int
+ad_transfer(struct ad_request *request)
+{
+ struct ad_softc *adp;
+ u_int64_t lba;
+ u_int32_t count;
+ u_int8_t cmd;
+ int flags = ATA_IMMEDIATE;
+
+ /* get request params */
+ adp = request->softc;
+
+ /* calculate transfer details */
+ lba = request->blockaddr + (request->donecount / DEV_BSIZE);
+
+ /* start timeout for this transfer */
+ if (!request->timeout_handle.callout && !dumping)
+ request->timeout_handle =
+ timeout((timeout_t*)ad_timeout, request, 10 * hz);
+
+ if (request->donecount == 0) {
+
+ /* check & setup transfer parameters */
+ if (request->bytecount > adp->max_iosize) {
+ ata_prtdev(adp->device,
+ "%d byte transfers not supported\n", request->bytecount);
+ count = howmany(adp->max_iosize, DEV_BSIZE);
+ }
+ else
+ count = howmany(request->bytecount, DEV_BSIZE);
+
+ if (count > (adp->device->param->support.address48 ? 65536 : 256)) {
+ ata_prtdev(adp->device,
+ "%d block transfers not supported\n", count);
+ count = adp->device->param->support.address48 ? 65536 : 256;
+ }
+
+ if (adp->flags & AD_F_CHS_USED) {
+ int sector = (lba % adp->sectors) + 1;
+ int cylinder = lba / (adp->sectors * adp->heads);
+ int head = (lba % (adp->sectors * adp->heads)) / adp->sectors;
+
+ lba = (sector&0xff) | ((cylinder&0xffff)<<8) | ((head&0xf)<<24);
+ adp->device->flags |= ATA_D_USE_CHS;
+ }
+
+ /* does this drive & transfer work with DMA ? */
+ request->flags &= ~ADR_F_DMA_USED;
+ if (adp->device->mode >= ATA_DMA &&
+ !adp->device->channel->dma->setup(adp->device, request->data, request->bytecount)) {
+ request->flags |= ADR_F_DMA_USED;
+ request->currentsize = request->bytecount;
+
+ /* do we have tags enabled ? */
+ if (adp->flags & AD_F_TAG_ENABLED) {
+ cmd = (request->flags & ADR_F_READ) ?
+ ATA_C_READ_DMA_QUEUED : ATA_C_WRITE_DMA_QUEUED;
+
+ if (ata_command(adp->device, cmd, lba,
+ request->tag << 3, count, flags)) {
+ ata_prtdev(adp->device, "error executing command");
+ goto transfer_failed;
+ }
+ if (ata_wait(adp->device, ATA_S_READY)) {
+ ata_prtdev(adp->device, "timeout waiting for READY\n");
+ goto transfer_failed;
+ }
+ adp->outstanding++;
+
+ /* if ATA bus RELEASE check for SERVICE */
+ if (adp->flags & AD_F_TAG_ENABLED &&
+ ATA_IDX_INB(adp->device->channel, ATA_IREASON) & ATA_I_RELEASE)
+ return ad_service(adp, 1);
+ }
+ else {
+ cmd = (request->flags & ADR_F_READ) ?
+ ATA_C_READ_DMA : ATA_C_WRITE_DMA;
+
+ if (ata_command(adp->device, cmd, lba, count, 0, flags)) {
+ ata_prtdev(adp->device, "error executing command");
+ goto transfer_failed;
+ }
+#if 0
+ /*
+ * wait for data transfer phase
+ *
+ * well this should be here acording to specs, but older
+ * promise controllers doesn't like it, they lockup!
+ */
+ if (ata_wait(adp->device, ATA_S_READY | ATA_S_DRQ)) {
+ ata_prtdev(adp->device, "timeout waiting for data phase\n");
+ goto transfer_failed;
+ }
+#endif
+ }
+
+ /* start transfer, return and wait for interrupt */
+ adp->device->channel->dma->start(adp->device->channel, request->data, request->bytecount,
+ request->flags & ADR_F_READ);
+ return ATA_OP_CONTINUES;
+ }
+
+ /* does this drive support multi sector transfers ? */
+ if (adp->transfersize > DEV_BSIZE)
+ cmd = request->flags&ADR_F_READ ? ATA_C_READ_MUL : ATA_C_WRITE_MUL;
+
+ /* just plain old single sector transfer */
+ else
+ cmd = request->flags&ADR_F_READ ? ATA_C_READ : ATA_C_WRITE;
+
+ if (ata_command(adp->device, cmd, lba, count, 0, flags)){
+ ata_prtdev(adp->device, "error executing command");
+ goto transfer_failed;
+ }
+ }
+
+ /* calculate this transfer length */
+ request->currentsize = min(request->bytecount, adp->transfersize);
+
+ /* if this is a PIO read operation, return and wait for interrupt */
+ if (request->flags & ADR_F_READ)
+ return ATA_OP_CONTINUES;
+
+ /* ready to write PIO data ? */
+ if (ata_wait(adp->device, (ATA_S_READY | ATA_S_DSC | ATA_S_DRQ)) < 0) {
+ ata_prtdev(adp->device, "timeout waiting for DRQ");
+ goto transfer_failed;
+ }
+
+ /* output the data */
+ if (adp->device->channel->flags & ATA_USE_16BIT)
+ ATA_IDX_OUTSW_STRM(adp->device->channel, ATA_DATA,
+ (void *)((uintptr_t)request->data + request->donecount),
+ request->currentsize / sizeof(int16_t));
+ else
+ ATA_IDX_OUTSL_STRM(adp->device->channel, ATA_DATA,
+ (void *)((uintptr_t)request->data + request->donecount),
+ request->currentsize / sizeof(int32_t));
+ return ATA_OP_CONTINUES;
+
+transfer_failed:
+ untimeout((timeout_t *)ad_timeout, request, request->timeout_handle);
+ ad_invalidatequeue(adp, request);
+
+ /* if retries still permit, reinject this request */
+ if (request->retries++ < AD_MAX_RETRIES)
+ TAILQ_INSERT_HEAD(&adp->device->channel->ata_queue, request, chain);
+ else {
+ /* retries all used up, return error */
+ request->bp->bio_error = EIO;
+ request->bp->bio_flags |= BIO_ERROR;
+ request->bp->bio_resid = request->bytecount;
+ biodone(request->bp);
+ ad_free(request);
+ }
+ ata_reinit(adp->device->channel);
+ return ATA_OP_CONTINUES;
+}
+
+int
+ad_interrupt(struct ad_request *request)
+{
+ struct ad_softc *adp = request->softc;
+ int dma_stat = 0;
+
+ /* finish DMA transfer */
+ if (request->flags & ADR_F_DMA_USED)
+ dma_stat = adp->device->channel->dma->stop(adp->device->channel);
+
+ /* do we have a corrected soft error ? */
+ if (adp->device->channel->status & ATA_S_CORR)
+ disk_err(request->bp, "soft error (ECC corrected)",
+ request->donecount / DEV_BSIZE, 1);
+
+ /* did any real errors happen ? */
+ if ((adp->device->channel->status & ATA_S_ERROR) ||
+ (request->flags & ADR_F_DMA_USED && dma_stat & ATA_BMSTAT_ERROR)) {
+ adp->device->channel->error =
+ ATA_IDX_INB(adp->device->channel, ATA_ERROR);
+ disk_err(request->bp, (adp->device->channel->error & ATA_E_ICRC) ?
+ "UDMA ICRC error" : "hard error",
+ request->donecount / DEV_BSIZE, 0);
+
+ /* if this is a UDMA CRC error, reinject request */
+ if (request->flags & ADR_F_DMA_USED &&
+ adp->device->channel->error & ATA_E_ICRC) {
+ untimeout((timeout_t *)ad_timeout, request,request->timeout_handle);
+ ad_invalidatequeue(adp, request);
+
+ if (request->retries++ < AD_MAX_RETRIES)
+ printf(" retrying\n");
+ else {
+ adp->device->setmode(adp->device, ATA_PIO_MAX);
+ printf(" falling back to PIO mode\n");
+ }
+ TAILQ_INSERT_HEAD(&adp->device->channel->ata_queue, request, chain);
+ return ATA_OP_FINISHED;
+ }
+#if 0 /* XXX*/
+ /* if using DMA, try once again in PIO mode */
+ if (request->flags & ADR_F_DMA_USED) {
+ untimeout((timeout_t *)ad_timeout, request,request->timeout_handle);
+ ad_invalidatequeue(adp, request);
+ adp->device->setmode(adp->device, ATA_PIO_MAX);
+ request->flags |= ADR_F_FORCE_PIO;
+ printf(" trying PIO mode\n");
+ TAILQ_INSERT_HEAD(&adp->device->channel->ata_queue, request, chain);
+ return ATA_OP_FINISHED;
+ }
+#endif
+ request->flags |= ADR_F_ERROR;
+ printf(" status=%02x error=%02x\n",
+ adp->device->channel->status, adp->device->channel->error);
+ }
+
+ /* if we arrived here with forced PIO mode, DMA doesn't work right */
+ if (request->flags & ADR_F_FORCE_PIO && !(request->flags & ADR_F_ERROR))
+ ata_prtdev(adp->device, "DMA problem fallback to PIO mode\n");
+
+ /* if this was a PIO read operation, get the data */
+ if (!(request->flags & ADR_F_DMA_USED) &&
+ (request->flags & (ADR_F_READ | ADR_F_ERROR)) == ADR_F_READ) {
+
+ /* ready to receive data? */
+ if ((adp->device->channel->status & (ATA_S_READY|ATA_S_DSC|ATA_S_DRQ))
+ != (ATA_S_READY|ATA_S_DSC|ATA_S_DRQ))
+ ata_prtdev(adp->device, "read interrupt arrived early");
+
+ if (ata_wait(adp->device, (ATA_S_READY | ATA_S_DSC | ATA_S_DRQ)) != 0) {
+ ata_prtdev(adp->device, "read error detected (too) late");
+ request->flags |= ADR_F_ERROR;
+ }
+ else {
+ /* data ready, read in */
+ if (adp->device->channel->flags & ATA_USE_16BIT)
+ ATA_IDX_INSW_STRM(adp->device->channel, ATA_DATA,
+ (void*)((uintptr_t)request->data +
+ request->donecount), request->currentsize /
+ sizeof(int16_t));
+ else
+ ATA_IDX_INSL_STRM(adp->device->channel, ATA_DATA,
+ (void*)((uintptr_t)request->data +
+ request->donecount), request->currentsize /
+ sizeof(int32_t));
+ }
+ }
+
+ /* finish up transfer */
+ if (request->flags & ADR_F_ERROR) {
+ request->bp->bio_error = EIO;
+ request->bp->bio_flags |= BIO_ERROR;
+ }
+ else {
+ request->bytecount -= request->currentsize;
+ request->donecount += request->currentsize;
+ if (!(request->flags & ADR_F_DMA_USED) && request->bytecount > 0) {
+ ad_transfer(request);
+ return ATA_OP_CONTINUES;
+ }
+ }
+
+ /* disarm timeout for this transfer */
+ untimeout((timeout_t *)ad_timeout, request, request->timeout_handle);
+
+ request->bp->bio_resid = request->bytecount;
+
+ biodone(request->bp);
+ ad_free(request);
+ adp->outstanding--;
+
+ /* check for SERVICE */
+ return ad_service(adp, 1);
+}
+
+int
+ad_service(struct ad_softc *adp, int change)
+{
+ /* do we have to check the other device on this channel ? */
+ if (adp->device->channel->flags & ATA_QUEUED && change) {
+ int device = adp->device->unit;
+
+ if (adp->device->unit == ATA_MASTER) {
+ if ((adp->device->channel->devices & ATA_ATA_SLAVE) &&
+ (adp->device->channel->device[SLAVE].driver) &&
+ ((struct ad_softc *) (adp->device->channel->
+ device[SLAVE].driver))->flags & AD_F_TAG_ENABLED)
+ device = ATA_SLAVE;
+ }
+ else {
+ if ((adp->device->channel->devices & ATA_ATA_MASTER) &&
+ (adp->device->channel->device[MASTER].driver) &&
+ ((struct ad_softc *) (adp->device->channel->
+ device[MASTER].driver))->flags & AD_F_TAG_ENABLED)
+ device = ATA_MASTER;
+ }
+ if (device != adp->device->unit &&
+ ((struct ad_softc *)
+ (adp->device->channel->
+ device[ATA_DEV(device)].driver))->outstanding > 0) {
+ ATA_IDX_OUTB(adp->device->channel, ATA_DRIVE, ATA_D_IBM | device);
+ adp = adp->device->channel->device[ATA_DEV(device)].driver;
+ DELAY(1);
+ }
+ }
+ adp->device->channel->status =
+ ATA_IDX_INB(adp->device->channel, ATA_ALTSTAT);
+
+ /* do we have a SERVICE request from the drive ? */
+ if (adp->flags & AD_F_TAG_ENABLED &&
+ adp->outstanding > 0 &&
+ adp->device->channel->status & ATA_S_SERVICE) {
+ struct ad_request *request;
+ int tag;
+
+ /* check for error */
+ if (adp->device->channel->status & ATA_S_ERROR) {
+ ata_prtdev(adp->device, "Oops! controller says s=0x%02x e=0x%02x\n",
+ adp->device->channel->status,
+ adp->device->channel->error);
+ ad_invalidatequeue(adp, NULL);
+ return ATA_OP_FINISHED;
+ }
+
+ /* issue SERVICE cmd */
+ if (ata_command(adp->device, ATA_C_SERVICE, 0, 0, 0, ATA_IMMEDIATE)) {
+ ata_prtdev(adp->device, "problem executing SERVICE cmd\n");
+ ad_invalidatequeue(adp, NULL);
+ return ATA_OP_FINISHED;
+ }
+
+ /* setup the transfer environment when ready */
+ if (ata_wait(adp->device, ATA_S_READY)) {
+ ata_prtdev(adp->device, "SERVICE timeout tag=%d s=%02x e=%02x\n",
+ ATA_IDX_INB(adp->device->channel, ATA_COUNT) >> 3,
+ adp->device->channel->status,
+ adp->device->channel->error);
+ ad_invalidatequeue(adp, NULL);
+ return ATA_OP_FINISHED;
+ }
+ tag = ATA_IDX_INB(adp->device->channel, ATA_COUNT) >> 3;
+ if (!(request = adp->tags[tag])) {
+ ata_prtdev(adp->device, "no request for tag=%d\n", tag);
+ ad_invalidatequeue(adp, NULL);
+ return ATA_OP_FINISHED;
+ }
+ ATA_FORCELOCK_CH(adp->device->channel, ATA_ACTIVE_ATA);
+ adp->device->channel->running = request;
+ request->serv++;
+
+ /* start DMA transfer when ready */
+ if (ata_wait(adp->device, ATA_S_READY | ATA_S_DRQ)) {
+ ata_prtdev(adp->device, "timeout starting DMA s=%02x e=%02x\n",
+ adp->device->channel->status,
+ adp->device->channel->error);
+ ad_invalidatequeue(adp, NULL);
+ return ATA_OP_FINISHED;
+ }
+ adp->device->channel->dma->start(adp->device->channel, request->data, request->bytecount,
+ request->flags & ADR_F_READ);
+ return ATA_OP_CONTINUES;
+ }
+ return ATA_OP_FINISHED;
+}
+
+static void
+ad_free(struct ad_request *request)
+{
+ request->softc->tags[request->tag] = NULL;
+ free(request, M_AD);
+}
+
+static void
+ad_invalidatequeue(struct ad_softc *adp, struct ad_request *request)
+{
+ /* if tags in use invalidate all other outstanding transfers */
+ if (adp->flags & AD_F_TAG_ENABLED) {
+ struct ad_request *tmpreq;
+ int tag;
+
+ ata_prtdev(adp->device, "invalidating queued requests\n");
+ for (tag = 0; tag <= adp->num_tags; tag++) {
+ tmpreq = adp->tags[tag];
+ adp->tags[tag] = NULL;
+ if (tmpreq == request || tmpreq == NULL)
+ continue;
+ untimeout((timeout_t *)ad_timeout, tmpreq, tmpreq->timeout_handle);
+ TAILQ_INSERT_HEAD(&adp->device->channel->ata_queue, tmpreq, chain);
+ }
+ adp->outstanding = 0;
+ if (ata_command(adp->device, ATA_C_NOP,
+ 0, 0, ATA_C_F_FLUSHQUEUE, ATA_WAIT_READY))
+ ata_prtdev(adp->device, "flush queue failed\n");
+ }
+}
+
+static int
+ad_tagsupported(struct ad_softc *adp)
+{
+ /* check for controllers that we know doesn't support tags */
+ switch (adp->device->channel->chiptype) {
+ case ATA_PDC20265: case ATA_PDC20263: case ATA_PDC20267:
+ case ATA_PDC20246: case ATA_PDC20262:
+ return 0;
+ }
+
+ /* check that drive does DMA, has tags enabled, and is one we know works */
+ if (adp->device->mode >= ATA_DMA && adp->device->param->support.queued &&
+ adp->device->param->enabled.queued) {
+
+ /* IBM DTTA series needs transfers <= 64K for tags to work properly */
+ if (!strncmp(adp->device->param->model, "IBM-DTTA", 8)) {
+ adp->max_iosize = 128 * DEV_BSIZE;
+ return 1;
+ }
+
+ /* IBM DJNA series has broken tags, corrupts data */
+ if (!strncmp(adp->device->param->model, "IBM-DJNA", 8))
+ return 0;
+
+ /* IBM DPTA & IBM DTLA series supports tags */
+ if (!strncmp(adp->device->param->model, "IBM-DPTA", 8) ||
+ !strncmp(adp->device->param->model, "IBM-DTLA", 8))
+ return 1;
+
+ /* IBM IC series ATA drives supports tags */
+ if (!strncmp(adp->device->param->model, "IC", 2) &&
+ (!strncmp(adp->device->param->model + 8, "AT", 2) ||
+ !strncmp(adp->device->param->model + 8, "AV", 2)))
+ return 1;
+ }
+ return 0;
+}
+
+static void
+ad_timeout(struct ad_request *request)
+{
+ struct ad_softc *adp = request->softc;
+
+ adp->device->channel->running = NULL;
+ request->timeout_handle.callout = NULL;
+ ata_prtdev(adp->device, "%s command timeout tag=%d serv=%d - resetting\n",
+ (request->flags & ADR_F_READ) ? "READ" : "WRITE",
+ request->tag, request->serv);
+
+ if (request->flags & ADR_F_DMA_USED) {
+ adp->device->channel->dma->stop(adp->device->channel);
+ ad_invalidatequeue(adp, request);
+ if (request->retries == AD_MAX_RETRIES) {
+ adp->device->setmode(adp->device, ATA_PIO_MAX);
+ ata_prtdev(adp->device, "trying fallback to PIO mode\n");
+ request->retries = 0;
+ }
+ }
+
+ /* if retries still permit, reinject this request */
+ if (request->retries++ < AD_MAX_RETRIES) {
+ TAILQ_INSERT_HEAD(&adp->device->channel->ata_queue, request, chain);
+ }
+ else {
+ /* retries all used up, return error */
+ request->bp->bio_error = EIO;
+ request->bp->bio_flags |= BIO_ERROR;
+ biodone(request->bp);
+ ad_free(request);
+ }
+ ata_reinit(adp->device->channel);
+}
+
+void
+ad_reinit(struct ata_device *atadev)
+{
+ struct ad_softc *adp = atadev->driver;
+
+ /* reinit disk parameters */
+ ad_invalidatequeue(atadev->driver, NULL);
+ ata_command(atadev, ATA_C_SET_MULTI, 0,
+ adp->transfersize / DEV_BSIZE, 0, ATA_WAIT_READY);
+ atadev->setmode(atadev, adp->device->mode);
+}
+
+void
+ad_print(struct ad_softc *adp)
+{
+ if (bootverbose) {
+ ata_prtdev(adp->device, "<%.40s/%.8s> ATA-%d disk at ata%d-%s\n",
+ adp->device->param->model, adp->device->param->revision,
+ ad_version(adp->device->param->version_major),
+ device_get_unit(adp->device->channel->dev),
+ (adp->device->unit == ATA_MASTER) ? "master" : "slave");
+
+ ata_prtdev(adp->device,
+ "%lluMB (%llu sectors), %llu C, %u H, %u S, %u B\n",
+ (unsigned long long)(adp->total_secs /
+ ((1024L*1024L)/DEV_BSIZE)),
+ (unsigned long long)adp->total_secs,
+ (unsigned long long)(adp->total_secs /
+ (adp->heads * adp->sectors)),
+ adp->heads, adp->sectors, DEV_BSIZE);
+
+ ata_prtdev(adp->device, "%d secs/int, %d depth queue, %s%s\n",
+ adp->transfersize / DEV_BSIZE, adp->num_tags + 1,
+ (adp->flags & AD_F_TAG_ENABLED) ? "tagged " : "",
+ ata_mode2str(adp->device->mode));
+
+ ata_prtdev(adp->device, "piomode=%d dmamode=%d udmamode=%d cblid=%d\n",
+ ata_pmode(adp->device->param), ata_wmode(adp->device->param),
+ ata_umode(adp->device->param),
+ adp->device->param->hwres_cblid);
+
+ }
+ else
+ ata_prtdev(adp->device,"%lluMB <%.40s> [%lld/%d/%d] at ata%d-%s %s%s\n",
+ (unsigned long long)(adp->total_secs /
+ ((1024L * 1024L) / DEV_BSIZE)),
+ adp->device->param->model,
+ (unsigned long long)(adp->total_secs /
+ (adp->heads * adp->sectors)),
+ adp->heads, adp->sectors,
+ device_get_unit(adp->device->channel->dev),
+ (adp->device->unit == ATA_MASTER) ? "master" : "slave",
+ (adp->flags & AD_F_TAG_ENABLED) ? "tagged " : "",
+ ata_mode2str(adp->device->mode));
+}
+
+static int
+ad_version(u_int16_t version)
+{
+ int bit;
+
+ if (version == 0xffff)
+ return 0;
+ for (bit = 15; bit >= 0; bit--)
+ if (version & (1<<bit))
+ return bit;
+ return 0;
+}
OpenPOWER on IntegriCloud