diff options
author | glebius <glebius@FreeBSD.org> | 2005-12-08 16:11:45 +0000 |
---|---|---|
committer | glebius <glebius@FreeBSD.org> | 2005-12-08 16:11:45 +0000 |
commit | 32bd4e565a3a4b0c4830285475d1acb39946ff70 (patch) | |
tree | fa5e6f16fea77fd4115ced5d9203bfa853cab714 /sys/dev/bge | |
parent | d6ef4fe5b252612d44020eecf6e3feb9b8233dd5 (diff) | |
download | FreeBSD-src-32bd4e565a3a4b0c4830285475d1acb39946ff70.zip FreeBSD-src-32bd4e565a3a4b0c4830285475d1acb39946ff70.tar.gz |
A big rewrite of receive Jumbo frame handling. Remove the local Jumbo
cluster allocator, that wasn't MPSAFE. Instead, utilize our new generic
UMA jumbo cluster allocator. Since UMA gives us a 9k piece that is contigous
in virtual memory, but isn't contigous in physical memory we need to handle
a few segments. To deal with this we utilize Tigon chip feature - extended
RX descriptors, that can handle up to four DMA segments for one frame.
Details:
o Remove bge_alloc_jumbo_mem(), bge_free_jumbo_mem(),
bge_jalloc(), bge_jfree() functions.
o Remove SLIST heads, bge_jumbo_tag, bge_jumbo_map from softc.
o Use extended RX BDs for Jumbo receive producer ring, and
initialize it appropriately.
o New bge_newbuf_jumbo():
- Allocate an mbuf with Jumbo cluster with help of m_cljget().
- Load the cluster for DMA with help of bus_dmamap_load_mbuf_sg().
- Assert that we got 3 segments in the DMA mapping.
- Fill in these 3 segments into the extended RX descriptor.
Diffstat (limited to 'sys/dev/bge')
-rw-r--r-- | sys/dev/bge/if_bge.c | 282 | ||||
-rw-r--r-- | sys/dev/bge/if_bgereg.h | 41 |
2 files changed, 75 insertions, 248 deletions
diff --git a/sys/dev/bge/if_bge.c b/sys/dev/bge/if_bge.c index b1d2076..8073305 100644 --- a/sys/dev/bge/if_bge.c +++ b/sys/dev/bge/if_bge.c @@ -79,7 +79,6 @@ __FBSDID("$FreeBSD$"); #include <sys/kernel.h> #include <sys/module.h> #include <sys/socket.h> -#include <sys/queue.h> #include <net/if.h> #include <net/if_arp.h> @@ -237,10 +236,6 @@ static int bge_read_eeprom (struct bge_softc *, caddr_t, int, int); static void bge_setmulti (struct bge_softc *); static void bge_handle_events (struct bge_softc *); -static int bge_alloc_jumbo_mem (struct bge_softc *); -static void bge_free_jumbo_mem (struct bge_softc *); -static void *bge_jalloc (struct bge_softc *); -static void bge_jfree (void *, void *); static int bge_newbuf_std (struct bge_softc *, int, struct mbuf *); static int bge_newbuf_jumbo (struct bge_softc *, int, struct mbuf *); static int bge_init_rx_ring_std (struct bge_softc *); @@ -448,7 +443,6 @@ bge_dma_map_tx_desc(arg, segs, nseg, mapsize, error) return; } - #ifdef notdef static u_int8_t bge_vpd_readbyte(sc, addr) @@ -747,156 +741,6 @@ bge_handle_events(sc) } /* - * Memory management for jumbo frames. - */ - -static int -bge_alloc_jumbo_mem(sc) - struct bge_softc *sc; -{ - caddr_t ptr; - register int i, error; - struct bge_jpool_entry *entry; - - /* Create tag for jumbo buffer block */ - - error = bus_dma_tag_create(sc->bge_cdata.bge_parent_tag, - PAGE_SIZE, 0, BUS_SPACE_MAXADDR, BUS_SPACE_MAXADDR, NULL, - NULL, BGE_JMEM, 1, BGE_JMEM, 0, NULL, NULL, - &sc->bge_cdata.bge_jumbo_tag); - - if (error) { - printf("bge%d: could not allocate jumbo dma tag\n", - sc->bge_unit); - return (ENOMEM); - } - - /* Allocate DMA'able memory for jumbo buffer block */ - - error = bus_dmamem_alloc(sc->bge_cdata.bge_jumbo_tag, - (void **)&sc->bge_ldata.bge_jumbo_buf, BUS_DMA_NOWAIT, - &sc->bge_cdata.bge_jumbo_map); - - if (error) - return (ENOMEM); - - SLIST_INIT(&sc->bge_jfree_listhead); - SLIST_INIT(&sc->bge_jinuse_listhead); - - /* - * Now divide it up into 9K pieces and save the addresses - * in an array. - */ - ptr = sc->bge_ldata.bge_jumbo_buf; - for (i = 0; i < BGE_JSLOTS; i++) { - sc->bge_cdata.bge_jslots[i] = ptr; - ptr += BGE_JLEN; - entry = malloc(sizeof(struct bge_jpool_entry), - M_DEVBUF, M_NOWAIT); - if (entry == NULL) { - bge_free_jumbo_mem(sc); - sc->bge_ldata.bge_jumbo_buf = NULL; - printf("bge%d: no memory for jumbo " - "buffer queue!\n", sc->bge_unit); - return(ENOBUFS); - } - entry->slot = i; - SLIST_INSERT_HEAD(&sc->bge_jfree_listhead, - entry, jpool_entries); - } - - return(0); -} - -static void -bge_free_jumbo_mem(sc) - struct bge_softc *sc; -{ - int i; - struct bge_jpool_entry *entry; - - for (i = 0; i < BGE_JSLOTS; i++) { - entry = SLIST_FIRST(&sc->bge_jfree_listhead); - SLIST_REMOVE_HEAD(&sc->bge_jfree_listhead, jpool_entries); - free(entry, M_DEVBUF); - } - - /* Destroy jumbo buffer block */ - - if (sc->bge_ldata.bge_rx_jumbo_ring) - bus_dmamem_free(sc->bge_cdata.bge_jumbo_tag, - sc->bge_ldata.bge_jumbo_buf, - sc->bge_cdata.bge_jumbo_map); - - if (sc->bge_cdata.bge_rx_jumbo_ring_map) - bus_dmamap_destroy(sc->bge_cdata.bge_jumbo_tag, - sc->bge_cdata.bge_jumbo_map); - - if (sc->bge_cdata.bge_jumbo_tag) - bus_dma_tag_destroy(sc->bge_cdata.bge_jumbo_tag); - - return; -} - -/* - * Allocate a jumbo buffer. - */ -static void * -bge_jalloc(sc) - struct bge_softc *sc; -{ - struct bge_jpool_entry *entry; - - entry = SLIST_FIRST(&sc->bge_jfree_listhead); - - if (entry == NULL) { - printf("bge%d: no free jumbo buffers\n", sc->bge_unit); - return(NULL); - } - - SLIST_REMOVE_HEAD(&sc->bge_jfree_listhead, jpool_entries); - SLIST_INSERT_HEAD(&sc->bge_jinuse_listhead, entry, jpool_entries); - return(sc->bge_cdata.bge_jslots[entry->slot]); -} - -/* - * Release a jumbo buffer. - */ -static void -bge_jfree(buf, args) - void *buf; - void *args; -{ - struct bge_jpool_entry *entry; - struct bge_softc *sc; - int i; - - /* Extract the softc struct pointer. */ - sc = (struct bge_softc *)args; - - if (sc == NULL) - panic("bge_jfree: can't find softc pointer!"); - - /* calculate the slot this buffer belongs to */ - - i = ((vm_offset_t)buf - - (vm_offset_t)sc->bge_ldata.bge_jumbo_buf) / BGE_JLEN; - - if ((i < 0) || (i >= BGE_JSLOTS)) - panic("bge_jfree: asked to free buffer that we don't manage!"); - - entry = SLIST_FIRST(&sc->bge_jinuse_listhead); - if (entry == NULL) - panic("bge_jfree: buffer not in use!"); - entry->slot = i; - SLIST_REMOVE_HEAD(&sc->bge_jinuse_listhead, jpool_entries); - SLIST_INSERT_HEAD(&sc->bge_jfree_listhead, entry, jpool_entries); - - return; -} - - -/* * Intialize a standard receive ring descriptor. */ static int @@ -967,68 +811,66 @@ bge_newbuf_jumbo(sc, i, m) int i; struct mbuf *m; { + bus_dma_segment_t segs[BGE_NSEG_JUMBO]; + struct bge_extrx_bd *r; struct mbuf *m_new = NULL; - struct bge_rx_bd *r; - struct bge_dmamap_arg ctx; + int nsegs; int error; if (m == NULL) { - caddr_t *buf = NULL; - - /* Allocate the mbuf. */ MGETHDR(m_new, M_DONTWAIT, MT_DATA); - if (m_new == NULL) { + if (m_new == NULL) return(ENOBUFS); - } - /* Allocate the jumbo buffer */ - buf = bge_jalloc(sc); - if (buf == NULL) { + m_cljget(m_new, M_DONTWAIT, MJUM9BYTES); + if (!(m_new->m_flags & M_EXT)) { m_freem(m_new); - printf("bge%d: jumbo allocation failed " - "-- packet dropped!\n", sc->bge_unit); return(ENOBUFS); } - - /* Attach the buffer to the mbuf. */ - m_new->m_data = (void *) buf; - m_new->m_len = m_new->m_pkthdr.len = BGE_JUMBO_FRAMELEN; - MEXTADD(m_new, buf, BGE_JUMBO_FRAMELEN, bge_jfree, - (struct bge_softc *)sc, 0, EXT_NET_DRV); + m_new->m_len = m_new->m_pkthdr.len = MJUM9BYTES; } else { m_new = m; + m_new->m_len = m_new->m_pkthdr.len = MJUM9BYTES; m_new->m_data = m_new->m_ext.ext_buf; - m_new->m_ext.ext_size = BGE_JUMBO_FRAMELEN; } if (!sc->bge_rx_alignment_bug) m_adj(m_new, ETHER_ALIGN); - /* Set up the descriptor. */ - sc->bge_cdata.bge_rx_jumbo_chain[i] = m_new; - r = &sc->bge_ldata.bge_rx_jumbo_ring[i]; - ctx.bge_maxsegs = 1; - ctx.sc = sc; - error = bus_dmamap_load(sc->bge_cdata.bge_mtag_jumbo, - sc->bge_cdata.bge_rx_jumbo_dmamap[i], mtod(m_new, void *), - m_new->m_len, bge_dma_map_addr, &ctx, BUS_DMA_NOWAIT); - if (error || ctx.bge_maxsegs == 0) { - if (m == NULL) { - sc->bge_cdata.bge_rx_jumbo_chain[i] = NULL; + + error = bus_dmamap_load_mbuf_sg(sc->bge_cdata.bge_mtag_jumbo, + sc->bge_cdata.bge_rx_jumbo_dmamap[i], + m_new, segs, &nsegs, BUS_DMA_NOWAIT); + if (error) { + if (m == NULL) m_freem(m_new); - } - return(ENOMEM); + return(error); } - r->bge_addr.bge_addr_lo = htole32(BGE_ADDR_LO(ctx.bge_busaddr)); - r->bge_addr.bge_addr_hi = htole32(BGE_ADDR_HI(ctx.bge_busaddr)); - r->bge_flags = htole16(BGE_RXBDFLAG_END|BGE_RXBDFLAG_JUMBO_RING); - r->bge_len = htole16(m_new->m_len); + KASSERT(nsegs == BGE_NSEG_JUMBO, ("%s: %d segments", __func__, nsegs)); + + sc->bge_cdata.bge_rx_jumbo_chain[i] = m_new; + + /* + * Fill in the extended RX buffer descriptor. + */ + r = &sc->bge_ldata.bge_rx_jumbo_ring[i]; + r->bge_addr0.bge_addr_lo = htole32(BGE_ADDR_LO(segs[0].ds_addr)); + r->bge_addr0.bge_addr_hi = htole32(BGE_ADDR_HI(segs[0].ds_addr)); + r->bge_len0 = htole16(segs[0].ds_len); + r->bge_addr1.bge_addr_lo = htole32(BGE_ADDR_LO(segs[1].ds_addr)); + r->bge_addr1.bge_addr_hi = htole32(BGE_ADDR_HI(segs[1].ds_addr)); + r->bge_len1 = htole16(segs[1].ds_len); + r->bge_addr2.bge_addr_lo = htole32(BGE_ADDR_LO(segs[2].ds_addr)); + r->bge_addr2.bge_addr_hi = htole32(BGE_ADDR_HI(segs[2].ds_addr)); + r->bge_len2 = htole16(segs[2].ds_len); + r->bge_len3 = htole16(0); + r->bge_flags = htole16(BGE_RXBDFLAG_JUMBO_RING|BGE_RXBDFLAG_END); r->bge_idx = htole16(i); bus_dmamap_sync(sc->bge_cdata.bge_mtag, sc->bge_cdata.bge_rx_jumbo_dmamap[i], BUS_DMASYNC_PREREAD); - return(0); + return (0); } /* @@ -1082,8 +924,8 @@ static int bge_init_rx_ring_jumbo(sc) struct bge_softc *sc; { - int i; struct bge_rcb *rcb; + int i; for (i = 0; i < BGE_JUMBO_RX_RING_CNT; i++) { if (bge_newbuf_jumbo(sc, i, NULL) == ENOBUFS) @@ -1097,7 +939,8 @@ bge_init_rx_ring_jumbo(sc) sc->bge_jumbo = i - 1; rcb = &sc->bge_ldata.bge_info.bge_jumbo_rx_rcb; - rcb->bge_maxlen_flags = BGE_RCB_MAXLEN_FLAGS(0, 0); + rcb->bge_maxlen_flags = BGE_RCB_MAXLEN_FLAGS(0, + BGE_RCB_FLAG_USE_EXT_RX_BD); CSR_WRITE_4(sc, BGE_RX_JUMBO_RCB_MAXLEN_FLAGS, rcb->bge_maxlen_flags); CSR_WRITE_4(sc, BGE_MBX_RX_JUMBO_PROD_LO, sc->bge_jumbo); @@ -1119,7 +962,7 @@ bge_free_rx_ring_jumbo(sc) sc->bge_cdata.bge_rx_jumbo_dmamap[i]); } bzero((char *)&sc->bge_ldata.bge_rx_jumbo_ring[i], - sizeof(struct bge_rx_bd)); + sizeof(struct bge_extrx_bd)); } return; @@ -1470,9 +1313,8 @@ bge_blockinit(sc) bus_dmamap_sync(sc->bge_cdata.bge_rx_jumbo_ring_tag, sc->bge_cdata.bge_rx_jumbo_ring_map, BUS_DMASYNC_PREREAD); - rcb->bge_maxlen_flags = - BGE_RCB_MAXLEN_FLAGS(BGE_MAX_FRAMELEN, - BGE_RCB_FLAG_RING_DISABLED); + rcb->bge_maxlen_flags = BGE_RCB_MAXLEN_FLAGS(0, + BGE_RCB_FLAG_USE_EXT_RX_BD|BGE_RCB_FLAG_RING_DISABLED); if (sc->bge_extram) rcb->bge_nicaddr = BGE_EXT_JUMBO_RX_RINGS; else @@ -1943,7 +1785,7 @@ bge_dma_alloc(dev) device_t dev; { struct bge_softc *sc; - int nseg, i, error; + int i, error; struct bge_dmamap_arg ctx; sc = device_get_softc(dev); @@ -1951,7 +1793,6 @@ bge_dma_alloc(dev) /* * Allocate the parent bus DMA tag appropriate for PCI. */ -#define BGE_NSEG_NEW 32 error = bus_dma_tag_create(NULL, /* parent */ PAGE_SIZE, 0, /* alignment, boundary */ BUS_SPACE_MAXADDR, /* lowaddr */ @@ -1966,11 +1807,10 @@ bge_dma_alloc(dev) /* * Create tag for RX mbufs. */ - nseg = 32; error = bus_dma_tag_create(sc->bge_cdata.bge_parent_tag, 1, 0, BUS_SPACE_MAXADDR, BUS_SPACE_MAXADDR, NULL, - NULL, MCLBYTES * nseg, nseg, MCLBYTES, BUS_DMA_ALLOCNOW, NULL, NULL, - &sc->bge_cdata.bge_mtag); + NULL, MCLBYTES * BGE_NSEG_NEW, BGE_NSEG_NEW, MCLBYTES, + BUS_DMA_ALLOCNOW, NULL, NULL, &sc->bge_cdata.bge_mtag); if (error) { device_printf(dev, "could not allocate dma tag\n"); @@ -2057,8 +1897,8 @@ bge_dma_alloc(dev) error = bus_dma_tag_create(sc->bge_cdata.bge_parent_tag, 1, 0, BUS_SPACE_MAXADDR, BUS_SPACE_MAXADDR, NULL, - NULL, MCLBYTES * nseg, nseg, BGE_JLEN, 0, NULL, NULL, - &sc->bge_cdata.bge_mtag_jumbo); + NULL, MJUM9BYTES, BGE_NSEG_JUMBO, PAGE_SIZE, + 0, NULL, NULL, &sc->bge_cdata.bge_mtag_jumbo); if (error) { device_printf(dev, "could not allocate dma tag\n"); @@ -2066,7 +1906,6 @@ bge_dma_alloc(dev) } /* Create tag for jumbo RX ring */ - error = bus_dma_tag_create(sc->bge_cdata.bge_parent_tag, PAGE_SIZE, 0, BUS_SPACE_MAXADDR, BUS_SPACE_MAXADDR, NULL, NULL, BGE_JUMBO_RX_RING_SZ, 1, BGE_JUMBO_RX_RING_SZ, 0, @@ -2078,18 +1917,14 @@ bge_dma_alloc(dev) } /* Allocate DMA'able memory for jumbo RX ring */ - error = bus_dmamem_alloc(sc->bge_cdata.bge_rx_jumbo_ring_tag, - (void **)&sc->bge_ldata.bge_rx_jumbo_ring, BUS_DMA_NOWAIT, + (void **)&sc->bge_ldata.bge_rx_jumbo_ring, + BUS_DMA_NOWAIT | BUS_DMA_ZERO, &sc->bge_cdata.bge_rx_jumbo_ring_map); if (error) return (ENOMEM); - bzero((char *)sc->bge_ldata.bge_rx_jumbo_ring, - BGE_JUMBO_RX_RING_SZ); - /* Load the address of the jumbo RX ring */ - ctx.bge_maxsegs = 1; ctx.sc = sc; @@ -2393,21 +2228,6 @@ bge_attach(dev) goto fail; } - /* - * Try to allocate memory for jumbo buffers. - * The 5705 does not appear to support jumbo frames. - */ - if (sc->bge_asicrev != BGE_ASICREV_BCM5705 && - sc->bge_asicrev != BGE_ASICREV_BCM5750) { - if (bge_alloc_jumbo_mem(sc)) { - printf("bge%d: jumbo buffer allocation " - "failed\n", sc->bge_unit); - bge_release_resources(sc); - error = ENXIO; - goto fail; - } - } - /* Set default tuneable values. */ sc->bge_stat_ticks = BGE_TICKS_PER_SEC; sc->bge_rx_coal_ticks = 150; @@ -2484,7 +2304,6 @@ bge_attach(dev) bge_ifmedia_upd, bge_ifmedia_sts)) { printf("bge%d: MII without any PHY!\n", sc->bge_unit); bge_release_resources(sc); - bge_free_jumbo_mem(sc); error = ENXIO; goto fail; } @@ -2562,9 +2381,6 @@ bge_detach(dev) } bge_release_resources(sc); - if (sc->bge_asicrev != BGE_ASICREV_BCM5705 && - sc->bge_asicrev != BGE_ASICREV_BCM5750) - bge_free_jumbo_mem(sc); return(0); } @@ -2739,7 +2555,7 @@ bge_reset(sc) * on the receive return list. * * Note: we have to be able to handle two possibilities here: - * 1) the frame is from the jumbo recieve ring + * 1) the frame is from the jumbo receive ring * 2) the frame is from the standard receive ring */ diff --git a/sys/dev/bge/if_bgereg.h b/sys/dev/bge/if_bgereg.h index 5ad6a5f..139d6ec 100644 --- a/sys/dev/bge/if_bgereg.h +++ b/sys/dev/bge/if_bgereg.h @@ -1811,6 +1811,27 @@ struct bge_rx_bd { u_int32_t bge_opaque; }; +struct bge_extrx_bd { + bge_hostaddr bge_addr1; + bge_hostaddr bge_addr2; + bge_hostaddr bge_addr3; + u_int16_t bge_len2; + u_int16_t bge_len1; + u_int16_t bge_rsvd1; + u_int16_t bge_len3; + bge_hostaddr bge_addr0; + u_int16_t bge_len0; + u_int16_t bge_idx; + u_int16_t bge_flags; + u_int16_t bge_type; + u_int16_t bge_tcp_udp_csum; + u_int16_t bge_ip_csum; + u_int16_t bge_vlan_tag; + u_int16_t bge_error_flag; + u_int32_t bge_rsvd0; + u_int32_t bge_opaque; +}; + #define BGE_RXBDFLAG_END 0x0004 #define BGE_RXBDFLAG_JUMBO_RING 0x0020 #define BGE_RXBDFLAG_VLAN_TAG 0x0040 @@ -2145,7 +2166,6 @@ struct bge_gib { #define BGE_MAX_FRAMELEN 1536 #define BGE_JUMBO_FRAMELEN 9018 #define BGE_JUMBO_MTU (BGE_JUMBO_FRAMELEN-ETHER_HDR_LEN-ETHER_CRC_LEN) -#define BGE_PAGE_SIZE PAGE_SIZE #define BGE_MIN_FRAMELEN 60 /* @@ -2214,6 +2234,9 @@ struct vpd_key { #define BGE_RESID (BGE_JPAGESZ - (BGE_JLEN * BGE_JSLOTS) % BGE_JPAGESZ) #define BGE_JMEM ((BGE_JLEN * BGE_JSLOTS) + BGE_RESID) +#define BGE_NSEG_JUMBO (MJUM9BYTES/PAGE_SIZE + 1) +#define BGE_NSEG_NEW 32 + /* * Ring structures. Most of these reside in host memory and we tell * the NIC where they are via the ring control blocks. The exceptions @@ -2224,7 +2247,7 @@ struct vpd_key { struct bge_ring_data { struct bge_rx_bd *bge_rx_std_ring; bus_addr_t bge_rx_std_ring_paddr; - struct bge_rx_bd *bge_rx_jumbo_ring; + struct bge_extrx_bd *bge_rx_jumbo_ring; bus_addr_t bge_rx_jumbo_ring_paddr; struct bge_rx_bd *bge_rx_return_ring; bus_addr_t bge_rx_return_ring_paddr; @@ -2234,14 +2257,13 @@ struct bge_ring_data { bus_addr_t bge_status_block_paddr; struct bge_stats *bge_stats; bus_addr_t bge_stats_paddr; - void *bge_jumbo_buf; struct bge_gib bge_info; }; #define BGE_STD_RX_RING_SZ \ (sizeof(struct bge_rx_bd) * BGE_STD_RX_RING_CNT) #define BGE_JUMBO_RX_RING_SZ \ - (sizeof(struct bge_rx_bd) * BGE_JUMBO_RX_RING_CNT) + (sizeof(struct bge_extrx_bd) * BGE_JUMBO_RX_RING_CNT) #define BGE_TX_RING_SZ \ (sizeof(struct bge_tx_bd) * BGE_TX_RING_CNT) #define BGE_RX_RTN_RING_SZ(x) \ @@ -2264,7 +2286,6 @@ struct bge_chain_data { bus_dma_tag_t bge_tx_ring_tag; bus_dma_tag_t bge_status_tag; bus_dma_tag_t bge_stats_tag; - bus_dma_tag_t bge_jumbo_tag; bus_dma_tag_t bge_mtag; /* mbuf mapping tag */ bus_dma_tag_t bge_mtag_jumbo; /* mbuf mapping tag */ bus_dmamap_t bge_tx_dmamap[BGE_TX_RING_CNT]; @@ -2276,12 +2297,9 @@ struct bge_chain_data { bus_dmamap_t bge_rx_return_ring_map; bus_dmamap_t bge_status_map; bus_dmamap_t bge_stats_map; - bus_dmamap_t bge_jumbo_map; struct mbuf *bge_tx_chain[BGE_TX_RING_CNT]; struct mbuf *bge_rx_std_chain[BGE_STD_RX_RING_CNT]; struct mbuf *bge_rx_jumbo_chain[BGE_JUMBO_RX_RING_CNT]; - /* Stick the jumbo mem management stuff here too. */ - caddr_t bge_jslots[BGE_JSLOTS]; }; struct bge_dmamap_arg { @@ -2304,11 +2322,6 @@ struct bge_type { #define BGE_TIMEOUT 100000 #define BGE_TXCONS_UNSET 0xFFFF /* impossible value */ -struct bge_jpool_entry { - int slot; - SLIST_ENTRY(bge_jpool_entry) jpool_entries; -}; - struct bge_bcom_hack { int reg; int val; @@ -2343,8 +2356,6 @@ struct bge_softc { u_int16_t bge_return_ring_cnt; u_int16_t bge_std; /* current std ring head */ u_int16_t bge_jumbo; /* current jumo ring head */ - SLIST_HEAD(__bge_jfreehead, bge_jpool_entry) bge_jfree_listhead; - SLIST_HEAD(__bge_jinusehead, bge_jpool_entry) bge_jinuse_listhead; u_int32_t bge_stat_ticks; u_int32_t bge_rx_coal_ticks; u_int32_t bge_tx_coal_ticks; |