/*- * Copyright (c) 1998,1999,2000,2001 Søren Schmidt * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer, * without modification, immediately at the beginning of the file. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * 3. The name of the author may not be used to endorse or promote products * derived from this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. * * $FreeBSD$ */ #include "opt_global.h" #include "opt_ata.h" #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include /* device structures */ static d_open_t adopen; static d_close_t adclose; static d_strategy_t adstrategy; static d_dump_t addump; static struct cdevsw ad_cdevsw = { /* open */ adopen, /* close */ adclose, /* read */ physread, /* write */ physwrite, /* ioctl */ noioctl, /* poll */ nopoll, /* mmap */ nommap, /* strategy */ adstrategy, /* name */ "ad", /* maj */ 116, /* dump */ addump, /* psize */ nopsize, /* flags */ D_DISK, }; static struct cdevsw addisk_cdevsw; /* prototypes */ static void ad_invalidatequeue(struct ad_softc *, struct ad_request *); static int ad_tagsupported(struct ad_softc *); static void ad_timeout(struct ad_request *); static void ad_free(struct ad_request *); static int ad_version(u_int16_t); /* internal vars */ static u_int32_t adp_lun_map = 0; static MALLOC_DEFINE(M_AD, "AD driver", "ATA disk driver"); static int ata_dma = 1; static int ata_wc = 0; static int ata_tags = 0; TUNABLE_INT("hw.ata.ata_dma", &ata_dma); TUNABLE_INT("hw.ata.wc", &ata_wc); TUNABLE_INT("hw.ata.tags", &ata_tags); /* sysctl vars */ SYSCTL_DECL(_hw_ata); SYSCTL_INT(_hw_ata, OID_AUTO, ata_dma, CTLFLAG_RD, &ata_dma, 0, "ATA disk DMA mode control"); SYSCTL_INT(_hw_ata, OID_AUTO, wc, CTLFLAG_RD, &ata_wc, 0, "ATA disk write caching"); SYSCTL_INT(_hw_ata, OID_AUTO, tags, CTLFLAG_RD, &ata_tags, 0, "ATA disk tagged queuing support"); /* defines */ #define AD_MAX_RETRIES 3 #define AD_PARAM ATA_PARAM(adp->controller, adp->unit) /* experimental cache flush on BIO_ORDERED */ #undef ATA_FLUSHCACHE_ON void ad_attach(struct ata_softc *scp, int device) { struct ad_softc *adp; dev_t dev; if (!(adp = malloc(sizeof(struct ad_softc), M_AD, M_NOWAIT | M_ZERO))) { ata_printf(scp, device, "failed to allocate driver storage\n"); return; } adp->controller = scp; adp->unit = device; #ifdef ATA_STATIC_ID adp->lun = (device_get_unit(scp->dev) << 1) + ATA_DEV(device); #else adp->lun = ata_get_lun(&adp_lun_map); #endif ata_set_name(scp, device, "ad", adp->lun); adp->heads = AD_PARAM->heads; adp->sectors = AD_PARAM->sectors; adp->total_secs = AD_PARAM->cylinders * adp->heads * adp->sectors; /* does this device need oldstyle CHS addressing */ if (!ad_version(AD_PARAM->version_major) || !(AD_PARAM->atavalid & ATA_FLAG_54_58) || !AD_PARAM->lba_size) adp->flags |= AD_F_CHS_USED; /* use the 28bit LBA size if valid */ if (AD_PARAM->cylinders == 16383 && adp->total_secs < AD_PARAM->lba_size) adp->total_secs = AD_PARAM->lba_size; /* use the 48bit LBA size if valid */ if (AD_PARAM->support.address48) adp->total_secs = AD_PARAM->lba_size48; /* use multiple sectors/interrupt if device supports it */ adp->transfersize = DEV_BSIZE; if (ad_version(AD_PARAM->version_major)) { int secsperint = max(1, min(AD_PARAM->sectors_intr, 16)); if (!ata_command(adp->controller, adp->unit, ATA_C_SET_MULTI, 0, secsperint, 0, ATA_WAIT_INTR) && !ata_wait(adp->controller, adp->unit, 0)) adp->transfersize *= secsperint; } /* enable read cacheing if not default on device */ if (ata_command(adp->controller, adp->unit, ATA_C_SETFEATURES, 0, 0, ATA_C_F_ENAB_RCACHE, ATA_WAIT_INTR)) ata_printf(scp, device, "enabling readahead cache failed\n"); /* enable write cacheing if allowed and not default on device */ if (ata_wc || (ata_tags && ad_tagsupported(adp))) { if (ata_command(adp->controller, adp->unit, ATA_C_SETFEATURES, 0, 0, ATA_C_F_ENAB_WCACHE, ATA_WAIT_INTR)) ata_printf(scp, device, "enabling write cache failed\n"); } else { if (ata_command(adp->controller, adp->unit, ATA_C_SETFEATURES, 0, 0, ATA_C_F_DIS_WCACHE, ATA_WAIT_INTR)) ata_printf(scp, device, "disabling write cache failed\n"); } /* use DMA if allowed and if drive/controller supports it */ if (ata_dma) ata_dmainit(adp->controller, adp->unit, ata_pmode(AD_PARAM), ata_wmode(AD_PARAM), ata_umode(AD_PARAM)); else ata_dmainit(adp->controller, adp->unit, ata_pmode(AD_PARAM), -1, -1); /* use tagged queueing if allowed and supported */ if (ata_tags && ad_tagsupported(adp)) { adp->num_tags = AD_PARAM->queuelen; adp->flags |= AD_F_TAG_ENABLED; adp->controller->flags |= ATA_QUEUED; if (ata_command(adp->controller, adp->unit, ATA_C_SETFEATURES, 0, 0, ATA_C_F_DIS_RELIRQ, ATA_WAIT_INTR)) ata_printf(scp, device, "disabling release interrupt failed\n"); if (ata_command(adp->controller, adp->unit, ATA_C_SETFEATURES, 0, 0, ATA_C_F_DIS_SRVIRQ, ATA_WAIT_INTR)) ata_printf(scp, device, "disabling service interrupt failed\n"); } devstat_add_entry(&adp->stats, "ad", adp->lun, DEV_BSIZE, DEVSTAT_NO_ORDERED_TAGS, DEVSTAT_TYPE_DIRECT | DEVSTAT_TYPE_IF_IDE, DEVSTAT_PRIORITY_DISK); dev = disk_create(adp->lun, &adp->disk, 0, &ad_cdevsw, &addisk_cdevsw); dev->si_drv1 = adp; dev->si_iosize_max = 256 * DEV_BSIZE; adp->dev = dev; bioq_init(&adp->queue); /* if this disk belongs to an ATA RAID dont print the probe */ if (!ar_probe(adp)) adp->flags |= AD_F_RAID_SUBDISK; else ad_print(adp, ""); /* construct the disklabel */ bzero(&adp->disk.d_label, sizeof(struct disklabel)); adp->disk.d_label.d_secsize = DEV_BSIZE; adp->disk.d_label.d_nsectors = adp->sectors; adp->disk.d_label.d_ntracks = adp->heads; adp->disk.d_label.d_ncylinders = adp->total_secs/(adp->heads*adp->sectors); adp->disk.d_label.d_secpercyl = adp->sectors * adp->heads; adp->disk.d_label.d_secperunit = adp->total_secs; /* store our softc signalling we are ready to go */ scp->dev_softc[ATA_DEV(device)] = adp; } void ad_detach(struct ad_softc *adp, int flush) { struct ad_request *request; struct bio *bp; adp->flags |= AD_F_DETACHING; if (adp->flags & AD_F_RAID_SUBDISK) printf("WARNING! detaching RAID subdisk, danger ahead\n"); ata_printf(adp->controller, adp->unit, "removed from configuration\n"); ad_invalidatequeue(adp, NULL); TAILQ_FOREACH(request, &adp->controller->ata_queue, chain) { if (request->device != adp) continue; TAILQ_REMOVE(&adp->controller->ata_queue, request, chain); biofinish(request->bp, NULL, ENXIO); ad_free(request); } while ((bp = bioq_first(&adp->queue))) { biofinish(bp, NULL, ENXIO); } disk_invalidate(&adp->disk); disk_destroy(adp->dev); devstat_remove_entry(&adp->stats); if (flush) { if (ata_command(adp->controller, adp->unit, ATA_C_FLUSHCACHE, 0, 0, 0, ATA_WAIT_READY)) ata_printf(adp->controller, adp->unit, "flushing cache on detach failed\n"); } ata_free_lun(&adp_lun_map, adp->lun); adp->controller->dev_softc[ATA_DEV(adp->unit)] = NULL; free(adp, M_AD); } static int adopen(dev_t dev, int flags, int fmt, struct thread *td) { struct ad_softc *adp = dev->si_drv1; if (adp->flags & AD_F_RAID_SUBDISK) return EBUSY; return 0; } static int adclose(dev_t dev, int flags, int fmt, struct thread *td) { struct ad_softc *adp = dev->si_drv1; if (ata_command(adp->controller, adp->unit, ATA_C_FLUSHCACHE, 0, 0, 0, ATA_WAIT_READY)) ata_printf(adp->controller, adp->unit, "flushing cache on close failed\n"); return 0; } static void adstrategy(struct bio *bp) { struct ad_softc *adp = bp->bio_dev->si_drv1; int s; if (adp->flags & AD_F_DETACHING) { biofinish(bp, NULL, ENXIO); return; } s = splbio(); bioqdisksort(&adp->queue, bp); ata_start(adp->controller); splx(s); } int addump(dev_t dev) { struct ad_softc *adp = dev->si_drv1; struct ad_request request; u_int count, blkno, secsize; vm_offset_t addr = 0; long blkcnt; int dumppages = MAXDUMPPGS; int error; int i; if ((error = disk_dumpcheck(dev, &count, &blkno, &secsize))) return error; if (!adp) return ENXIO; /* force PIO mode for dumps */ adp->controller->mode[ATA_DEV(adp->unit)] = ATA_PIO; ata_reinit(adp->controller); blkcnt = howmany(PAGE_SIZE, secsize); while (count > 0) { caddr_t va = NULL; DELAY(1000); if ((count / blkcnt) < dumppages) dumppages = count / blkcnt; for (i = 0; i < dumppages; ++i) { vm_offset_t a = addr + (i * PAGE_SIZE); if (is_physical_memory(a)) va = pmap_kenter_temporary(trunc_page(a), i); else va = pmap_kenter_temporary(trunc_page(0), i); } bzero(&request, sizeof(struct ad_request)); request.device = adp; request.blockaddr = blkno; request.bytecount = PAGE_SIZE * dumppages; request.data = va; while (request.bytecount > 0) { ad_transfer(&request); if (request.flags & ADR_F_ERROR) return EIO; request.donecount += request.currentsize; request.bytecount -= request.currentsize; DELAY(20); } if (dumpstatus(addr, (off_t)count * DEV_BSIZE) < 0) return EINTR; blkno += blkcnt * dumppages; count -= blkcnt * dumppages; addr += PAGE_SIZE * dumppages; } if (ata_wait(adp->controller, adp->unit, ATA_S_READY | ATA_S_DSC) < 0) ata_printf(adp->controller, adp->unit, "timeout waiting for final ready\n"); return 0; } void ad_start(struct ad_softc *adp) { struct bio *bp = bioq_first(&adp->queue); struct ad_request *request; int tag = 0; if (!bp) return; #ifdef ATA_FLUSHCACHE_ON /* * if BIO_ORDERED is set cache should be flushed, if there are * any outstanding requests, hold off and wait for them to finish */ if (adp->flags & AD_F_TAG_ENABLED && bp->bio_flags & BIO_ORDERED && adp->outstanding > 0) return; #endif /* if tagged queueing enabled get next free tag */ if (adp->flags & AD_F_TAG_ENABLED) { while (tag <= adp->num_tags && adp->tags[tag]) tag++; if (tag > adp->num_tags ) return; } if (!(request = malloc(sizeof(struct ad_request), M_AD, M_NOWAIT|M_ZERO))) { ata_printf(adp->controller, adp->unit, "out of memory in start\n"); return; } /* setup request */ request->device = adp; request->bp = bp; request->blockaddr = bp->bio_pblkno; request->bytecount = bp->bio_bcount; request->data = bp->bio_data; request->tag = tag; if (bp->bio_cmd == BIO_READ) request->flags |= ADR_F_READ; if (adp->controller->mode[ATA_DEV(adp->unit)] >= ATA_DMA) { if (!(request->dmatab = ata_dmaalloc(adp->controller, adp->unit))) adp->controller->mode[ATA_DEV(adp->unit)] = ATA_PIO; } /* insert in tag array */ adp->tags[tag] = request; /* remove from drive queue */ bioq_remove(&adp->queue, bp); /* link onto controller queue */ TAILQ_INSERT_TAIL(&adp->controller->ata_queue, request, chain); } int ad_transfer(struct ad_request *request) { struct ad_softc *adp; u_int64_t lba; u_int32_t count, max_count; u_int8_t cmd; int flags = ATA_IMMEDIATE; /* get request params */ adp = request->device; /* calculate transfer details */ lba = request->blockaddr + (request->donecount / DEV_BSIZE); if (request->donecount == 0) { /* start timeout for this transfer */ if (dumping) request->timeout_handle.callout = NULL; else request->timeout_handle = timeout((timeout_t*)ad_timeout, request, 10 * hz); /* setup transfer parameters !! 65536 for 48bit SOS XXX */ count = howmany(request->bytecount, DEV_BSIZE); max_count = AD_PARAM->support.address48 ? 65536 : 256; if (count > max_count) { ata_printf(adp->controller, adp->unit, "count %d size transfers not supported\n", count); count = max_count; } if (adp->flags & AD_F_CHS_USED) { int sector = (lba % adp->sectors) + 1; int cylinder = lba / (adp->sectors * adp->heads); int head = (lba % (adp->sectors * adp->heads)) / adp->sectors; lba = (sector&0xff) | ((cylinder&0xffff)<<8) | ((head&0xf)<<24); flags |= ATA_USE_CHS; } /* setup first transfer length */ request->currentsize = min(request->bytecount, adp->transfersize); devstat_start_transaction(&adp->stats); /* does this drive & transfer work with DMA ? */ request->flags &= ~ADR_F_DMA_USED; if (adp->controller->mode[ATA_DEV(adp->unit)] >= ATA_DMA && !ata_dmasetup(adp->controller, adp->unit, request->dmatab, request->data, request->bytecount)) { request->flags |= ADR_F_DMA_USED; request->currentsize = request->bytecount; /* do we have tags enabled ? */ if (adp->flags & AD_F_TAG_ENABLED) { cmd = (request->flags & ADR_F_READ) ? ATA_C_READ_DMA_QUEUED : ATA_C_WRITE_DMA_QUEUED; if (ata_command(adp->controller, adp->unit, cmd, lba, request->tag << 3, count, flags)) { ata_printf(adp->controller, adp->unit, "error executing command"); goto transfer_failed; } if (ata_wait(adp->controller, adp->unit, ATA_S_READY)) { ata_printf(adp->controller, adp->unit, "timeout waiting for READY\n"); goto transfer_failed; } adp->outstanding++; /* if ATA bus RELEASE check for SERVICE */ if (adp->flags & AD_F_TAG_ENABLED && ATA_INB(adp->controller->r_io, ATA_IREASON) & ATA_I_RELEASE) return ad_service(adp, 1); } else { cmd = (request->flags & ADR_F_READ) ? ATA_C_READ_DMA : ATA_C_WRITE_DMA; if (ata_command(adp->controller, adp->unit, cmd, lba, count, 0, flags)) { ata_printf(adp->controller, adp->unit, "error executing command"); goto transfer_failed; } #if 0 /* * wait for data transfer phase * * well this should be here acording to specs, but * promise controllers doesn't like it, they lockup! * thats probably why tags doesn't work on the promise * as this is needed there... */ if (ata_wait(adp->controller, adp->unit, ATA_S_READY | ATA_S_DRQ)) { ata_printf(adp->controller, adp->unit, "timeout waiting for data phase\n"); goto transfer_failed; } #endif } /* start transfer, return and wait for interrupt */ ata_dmastart(adp->controller, adp->unit, request->dmatab, request->flags & ADR_F_READ); return ATA_OP_CONTINUES; } /* does this drive support multi sector transfers ? */ if (request->currentsize > DEV_BSIZE) cmd = request->flags&ADR_F_READ ? ATA_C_READ_MUL : ATA_C_WRITE_MUL; /* just plain old single sector transfer */ else cmd = request->flags&ADR_F_READ ? ATA_C_READ : ATA_C_WRITE; if (ata_command(adp->controller, adp->unit, cmd, lba, count, 0, flags)){ ata_printf(adp->controller, adp->unit, "error executing command"); goto transfer_failed; } } /* calculate this transfer length */ request->currentsize = min(request->bytecount, adp->transfersize); /* if this is a PIO read operation, return and wait for interrupt */ if (request->flags & ADR_F_READ) return ATA_OP_CONTINUES; /* ready to write PIO data ? */ if (ata_wait(adp->controller, adp->unit, (ATA_S_READY | ATA_S_DSC | ATA_S_DRQ)) < 0) { ata_printf(adp->controller, adp->unit, "timeout waiting for DRQ"); goto transfer_failed; } /* output the data */ if (adp->controller->flags & ATA_USE_16BIT) ATA_OUTSW(adp->controller->r_io, ATA_DATA, (void *)((uintptr_t)request->data + request->donecount), request->currentsize / sizeof(int16_t)); else ATA_OUTSL(adp->controller->r_io, ATA_DATA, (void *)((uintptr_t)request->data + request->donecount), request->currentsize / sizeof(int32_t)); return ATA_OP_CONTINUES; transfer_failed: untimeout((timeout_t *)ad_timeout, request, request->timeout_handle); ad_invalidatequeue(adp, request); printf(" - resetting\n"); /* if retries still permit, reinject this request */ if (request->retries++ < AD_MAX_RETRIES) TAILQ_INSERT_HEAD(&adp->controller->ata_queue, request, chain); else { /* retries all used up, return error */ request->bp->bio_error = EIO; request->bp->bio_flags |= BIO_ERROR; request->bp->bio_resid = request->bytecount; biofinish(request->bp, &adp->stats, 0); ad_free(request); } ata_reinit(adp->controller); return ATA_OP_CONTINUES; } int ad_interrupt(struct ad_request *request) { struct ad_softc *adp = request->device; int dma_stat = 0; #ifdef ATA_FLUSHCACHE_ON if (request->flags & ADR_F_FLUSHCACHE) goto finish; #endif /* finish DMA transfer */ if (request->flags & ADR_F_DMA_USED) dma_stat = ata_dmadone(adp->controller); /* do we have a corrected soft error ? */ if (adp->controller->status & ATA_S_CORR) diskerr(request->bp, "soft error (ECC corrected)", request->blockaddr + (request->donecount / DEV_BSIZE), &adp->disk.d_label); /* did any real errors happen ? */ if ((adp->controller->status & ATA_S_ERROR) || (request->flags & ADR_F_DMA_USED && dma_stat & ATA_BMSTAT_ERROR)) { adp->controller->error = ATA_INB(adp->controller->r_io, ATA_ERROR); diskerr(request->bp, (adp->controller->error & ATA_E_ICRC) ? "UDMA ICRC error" : "hard error", request->blockaddr + (request->donecount / DEV_BSIZE), &adp->disk.d_label); /* if this is a UDMA CRC error, reinject request */ if (request->flags & ADR_F_DMA_USED && adp->controller->error & ATA_E_ICRC) { untimeout((timeout_t *)ad_timeout, request,request->timeout_handle); ad_invalidatequeue(adp, request); if (request->retries++ < AD_MAX_RETRIES) printf(" retrying\n"); else { ata_dmainit(adp->controller, adp->unit, ata_pmode(AD_PARAM), -1, -1); printf(" falling back to PIO mode\n"); } TAILQ_INSERT_HEAD(&adp->controller->ata_queue, request, chain); return ATA_OP_FINISHED; } /* if using DMA, try once again in PIO mode */ if (request->flags & ADR_F_DMA_USED) { untimeout((timeout_t *)ad_timeout, request,request->timeout_handle); ad_invalidatequeue(adp, request); ata_dmainit(adp->controller, adp->unit, ata_pmode(AD_PARAM), -1, -1); request->flags |= ADR_F_FORCE_PIO; TAILQ_INSERT_HEAD(&adp->controller->ata_queue, request, chain); return ATA_OP_FINISHED; } request->flags |= ADR_F_ERROR; printf(" status=%02x error=%02x\n", adp->controller->status, adp->controller->error); } /* if we arrived here with forced PIO mode, DMA doesn't work right */ if (request->flags & ADR_F_FORCE_PIO) ata_printf(adp->controller, adp->unit, "DMA problem fallback to PIO mode\n"); /* if this was a PIO read operation, get the data */ if (!(request->flags & ADR_F_DMA_USED) && (request->flags & (ADR_F_READ | ADR_F_ERROR)) == ADR_F_READ) { /* ready to receive data? */ if ((adp->controller->status & (ATA_S_READY | ATA_S_DSC | ATA_S_DRQ)) != (ATA_S_READY | ATA_S_DSC | ATA_S_DRQ)) ata_printf(adp->controller, adp->unit, "read interrupt arrived early"); if (ata_wait(adp->controller, adp->unit, (ATA_S_READY | ATA_S_DSC | ATA_S_DRQ)) != 0) { ata_printf(adp->controller, adp->unit, "read error detected (too) late"); request->flags |= ADR_F_ERROR; } else { /* data ready, read in */ if (adp->controller->flags & ATA_USE_16BIT) ATA_INSW(adp->controller->r_io, ATA_DATA, (void *)((uintptr_t)request->data + request->donecount), request->currentsize / sizeof(int16_t)); else ATA_INSL(adp->controller->r_io, ATA_DATA, (void *)((uintptr_t)request->data + request->donecount), request->currentsize / sizeof(int32_t)); } } /* finish up transfer */ if (request->flags & ADR_F_ERROR) { request->bp->bio_error = EIO; request->bp->bio_flags |= BIO_ERROR; } else { request->bytecount -= request->currentsize; request->donecount += request->currentsize; if (request->bytecount > 0) { ad_transfer(request); return ATA_OP_CONTINUES; } } /* disarm timeout for this transfer */ untimeout((timeout_t *)ad_timeout, request, request->timeout_handle); request->bp->bio_resid = request->bytecount; #ifdef ATA_FLUSHCACHE_ON if (request->bp->bio_flags & BIO_ORDERED) { request->flags |= ADR_F_FLUSHCACHE; if (ata_command(adp->controller, adp->unit, ATA_C_FLUSHCACHE, 0, 0, 0, ATA_IMMEDIATE)) ata_printf(adp->controller, adp->unit, "flushing cache failed\n"); else return ATA_OP_CONTINUES; } finish: #endif biofinish(request->bp, &adp->stats, 0); ad_free(request); adp->outstanding--; /* check for SERVICE (tagged operations only) */ return ad_service(adp, 1); } int ad_service(struct ad_softc *adp, int change) { /* do we have to check the other device on this channel ? */ if (adp->controller->flags & ATA_QUEUED && change) { int device = adp->unit; if (adp->unit == ATA_MASTER) { if (adp->controller->devices & ATA_ATA_SLAVE && ((struct ad_softc *) (adp->controller->dev_softc[ATA_DEV(ATA_SLAVE)]))->flags & AD_F_TAG_ENABLED) device = ATA_SLAVE; } else { if (adp->controller->devices & ATA_ATA_MASTER && ((struct ad_softc *) (adp->controller->dev_softc[ATA_DEV(ATA_MASTER)]))->flags & AD_F_TAG_ENABLED) device = ATA_MASTER; } if (device != adp->unit && ((struct ad_softc *) (adp->controller->dev_softc[ATA_DEV(device)]))->outstanding > 0) { ATA_OUTB(adp->controller->r_io, ATA_DRIVE, ATA_D_IBM | device); adp = adp->controller->dev_softc[ATA_DEV(device)]; DELAY(1); } } adp->controller->status = ATA_INB(adp->controller->r_altio, ATA_ALTSTAT); /* do we have a SERVICE request from the drive ? */ if (adp->flags & AD_F_TAG_ENABLED && adp->outstanding > 0 && adp->controller->status & ATA_S_SERVICE) { struct ad_request *request; int tag; /* check for error */ if (adp->controller->status & ATA_S_ERROR) { ata_printf(adp->controller, adp->unit, "Oops! controller says s=0x%02x e=0x%02x\n", adp->controller->status, adp->controller->error); ad_invalidatequeue(adp, NULL); return ATA_OP_FINISHED; } /* issue SERVICE cmd */ if (ata_command(adp->controller, adp->unit, ATA_C_SERVICE, 0, 0, 0, ATA_IMMEDIATE)) { ata_printf(adp->controller, adp->unit, "problem executing SERVICE cmd\n"); ad_invalidatequeue(adp, NULL); return ATA_OP_FINISHED; } /* setup the transfer environment when ready */ if (ata_wait(adp->controller, adp->unit, ATA_S_READY)) { ata_printf(adp->controller, adp->unit, "problem issueing SERVICE tag=%d s=0x%02x e=0x%02x\n", ATA_INB(adp->controller->r_io, ATA_COUNT) >> 3, adp->controller->status, adp->controller->error); ad_invalidatequeue(adp, NULL); return ATA_OP_FINISHED; } tag = ATA_INB(adp->controller->r_io, ATA_COUNT) >> 3; if (!(request = adp->tags[tag])) { ata_printf(adp->controller, adp->unit, "no request for tag=%d\n", tag); ad_invalidatequeue(adp, NULL); return ATA_OP_FINISHED; } adp->controller->active = ATA_ACTIVE_ATA; adp->controller->running = request; request->serv++; /* start DMA transfer when ready */ if (ata_wait(adp->controller, adp->unit, ATA_S_READY | ATA_S_DRQ)) { ata_printf(adp->controller, adp->unit, "timeout waiting for data phase s=%02x e=%02x\n", adp->controller->status, adp->controller->error); ad_invalidatequeue(adp, NULL); return ATA_OP_FINISHED; } ata_dmastart(adp->controller, adp->unit, request->dmatab, request->flags & ADR_F_READ); return ATA_OP_CONTINUES; } return ATA_OP_FINISHED; } static void ad_free(struct ad_request *request) { int s = splbio(); if (request->dmatab) free(request->dmatab, M_DEVBUF); request->device->tags[request->tag] = NULL; free(request, M_AD); splx(s); } static void ad_invalidatequeue(struct ad_softc *adp, struct ad_request *request) { /* if tags used invalidate all other tagged transfers */ if (adp->flags & AD_F_TAG_ENABLED) { struct ad_request *tmpreq; int tag; ata_printf(adp->controller, adp->unit,"invalidating queued requests\n"); for (tag = 0; tag <= adp->num_tags; tag++) { tmpreq = adp->tags[tag]; adp->tags[tag] = NULL; if (tmpreq == request || tmpreq == NULL) continue; untimeout((timeout_t *)ad_timeout, tmpreq, tmpreq->timeout_handle); TAILQ_INSERT_HEAD(&adp->controller->ata_queue, tmpreq, chain); } if (ata_command(adp->controller, adp->unit, ATA_C_NOP, 0, 0, ATA_C_F_FLUSHQUEUE, ATA_WAIT_READY)) ata_printf(adp->controller, adp->unit, "flush queue failed\n"); adp->outstanding = 0; } } static int ad_tagsupported(struct ad_softc *adp) { const char *drives[] = {"IBM-DPTA", "IBM-DTLA", NULL}; int i = 0; switch (adp->controller->chiptype) { case 0x4d33105a: /* Promises before TX2 doesn't work with tagged queuing */ case 0x4d38105a: case 0x0d30105a: case 0x4d30105a: return 0; } /* check that drive does DMA, has tags enabled, and is one we know works */ if (adp->controller->mode[ATA_DEV(adp->unit)] >= ATA_DMA && AD_PARAM->support.queued && AD_PARAM->enabled.queued) { while (drives[i] != NULL) { if (!strncmp(AD_PARAM->model, drives[i], strlen(drives[i]))) return 1; i++; } /* * check IBM's new obscure way of naming drives * we want "IC" (IBM CORP) and "AT" or "AV" (ATA interface) * but doesn't care about the other info (size, capacity etc) */ if (!strncmp(AD_PARAM->model, "IC", 2) && (!strncmp(AD_PARAM->model + 8, "AT", 2) || !strncmp(AD_PARAM->model + 8, "AV", 2))) return 1; } return 0; } static void ad_timeout(struct ad_request *request) { struct ad_softc *adp = request->device; int s = splbio(); adp->controller->running = NULL; ata_printf(adp->controller, adp->unit, "%s command timeout tag=%d serv=%d - resetting\n", (request->flags & ADR_F_READ) ? "READ" : "WRITE", request->tag, request->serv); if (request->flags & ADR_F_DMA_USED) { ata_dmadone(adp->controller); ad_invalidatequeue(adp, request); if (request->retries == AD_MAX_RETRIES) { ata_dmainit(adp->controller, adp->unit, ata_pmode(AD_PARAM), -1, -1); ata_printf(adp->controller, adp->unit, "trying fallback to PIO mode\n"); request->retries = 0; } } /* if retries still permit, reinject this request */ if (request->retries++ < AD_MAX_RETRIES) { TAILQ_INSERT_HEAD(&adp->controller->ata_queue, request, chain); } else { /* retries all used up, return error */ request->bp->bio_error = EIO; request->bp->bio_flags |= BIO_ERROR; biofinish(request->bp, &adp->stats, 0); ad_free(request); } ata_reinit(adp->controller); splx(s); } void ad_reinit(struct ad_softc *adp) { /* reinit disk parameters */ ad_invalidatequeue(adp, NULL); ata_command(adp->controller, adp->unit, ATA_C_SET_MULTI, 0, adp->transfersize / DEV_BSIZE, 0, ATA_WAIT_INTR); if (adp->controller->mode[ATA_DEV(adp->unit)] >= ATA_DMA) ata_dmainit(adp->controller, adp->unit, ata_pmode(AD_PARAM), ata_wmode(AD_PARAM), ata_umode(AD_PARAM)); else ata_dmainit(adp->controller, adp->unit, ata_pmode(AD_PARAM), -1, -1); } void ad_print(struct ad_softc *adp, char *prepend) { if (prepend) printf("%s", prepend); if (bootverbose) { ata_printf(adp->controller, adp->unit, "<%.40s/%.8s> ATA-%d disk at ata%d-%s\n", AD_PARAM->model, AD_PARAM->revision, ad_version(AD_PARAM->version_major), device_get_unit(adp->controller->dev), (adp->unit == ATA_MASTER) ? "master" : "slave"); if (prepend) printf("%s", prepend); ata_printf(adp->controller, adp->unit, "%lluMB (%llu sectors), %llu C, %u H, %u S, %u B\n", (unsigned long long)(adp->total_secs / ((1024L*1024L)/DEV_BSIZE)), (unsigned long long) adp->total_secs, (unsigned long long) (adp->total_secs / (adp->heads * adp->sectors)), adp->heads, adp->sectors, DEV_BSIZE); if (prepend) printf("%s", prepend); ata_printf(adp->controller, adp->unit, "%d secs/int, %d depth queue, %s%s\n", adp->transfersize / DEV_BSIZE, adp->num_tags + 1, (adp->flags & AD_F_TAG_ENABLED) ? "tagged " : "", ata_mode2str(adp->controller->mode[ATA_DEV(adp->unit)])); if (prepend) printf("%s", prepend); ata_printf(adp->controller, adp->unit, "piomode=%d dmamode=%d udmamode=%d cblid=%d\n", ata_pmode(AD_PARAM), ata_wmode(AD_PARAM), ata_umode(AD_PARAM), AD_PARAM->hwres_cblid); } else ata_printf(adp->controller, adp->unit, "%lluMB <%.40s> [%lld/%d/%d] at ata%d-%s %s%s\n", (unsigned long long)(adp->total_secs / ((1024L * 1024L) / DEV_BSIZE)), AD_PARAM->model, (unsigned long long)(adp->total_secs / (adp->heads*adp->sectors)), adp->heads, adp->sectors, device_get_unit(adp->controller->dev), (adp->unit == ATA_MASTER) ? "master" : "slave", (adp->flags & AD_F_TAG_ENABLED) ? "tagged " : "", ata_mode2str(adp->controller->mode[ATA_DEV(adp->unit)])); } static int ad_version(u_int16_t version) { int bit; if (version == 0xffff) return 0; for (bit = 15; bit >= 0; bit--) if (version & (1<