1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
|
/*-
* Copyright (c) 1999 The NetBSD Foundation, Inc.
* All rights reserved.
*
* This code is derived from software contributed to The NetBSD Foundation
* by Paul Kranenburg.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
* ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
* TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
* BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
* POSSIBILITY OF SUCH DAMAGE.
*
* from: NetBSD: hmevar.h,v 1.5 2000/06/25 01:10:04 eeh Exp
*
* $FreeBSD$
*/
#include <sys/callout.h>
/*
* Number of receive and transmit descriptors. For each receive descriptor,
* an mbuf cluster is allocated and set up to receive a packet, and a dma map
* is created. Therefore, this number should not be too high to not waste
* memory.
* TX descriptors have no static cost, except for the memory directly allocated
* for them. TX queue elements (the number of which is fixed by HME_NTXQ) hold
* the software state for a transmit job; each has a dmamap allocated for it.
* There may be multiple descriptors allocated to a single queue element.
* HME_NTXQ and HME_NTXSEGS are completely arbitrary.
*/
#define HME_NRXDESC 128
#define HME_NTXDESC 256
#define HME_NTXQ 64
#define HME_NTXSEGS 16
/* Maximum size of a mapped RX buffer. */
#define HME_BUFSZ 1600
/*
* RX DMA descriptor. The descriptors are preallocated; the dma map is
* reused.
*/
struct hme_rxdesc {
struct mbuf *hrx_m;
bus_dmamap_t hrx_dmamap;
};
/* Lazily leave at least one burst size grace space. */
#define HME_DESC_RXLEN(sc, d) \
ulmin(HME_BUFSZ, (d)->hrx_m->m_len - (sc)->sc_burst)
struct hme_txdesc {
struct mbuf *htx_m;
bus_dmamap_t htx_dmamap;
int htx_lastdesc;
STAILQ_ENTRY(hme_txdesc) htx_q;
};
STAILQ_HEAD(hme_txdq, hme_txdesc);
struct hme_ring {
/* Ring Descriptors */
caddr_t rb_membase; /* Packet buffer: CPU address */
bus_addr_t rb_dmabase; /* Packet buffer: DMA address */
caddr_t rb_txd; /* Transmit descriptors */
bus_addr_t rb_txddma; /* DMA address of same */
caddr_t rb_rxd; /* Receive descriptors */
bus_addr_t rb_rxddma; /* DMA address of same */
/* Ring Descriptor state */
int rb_tdhead, rb_tdtail;
int rb_rdtail;
int rb_td_nbusy;
/* Descriptors */
struct hme_rxdesc rb_rxdesc[HME_NRXDESC];
struct hme_txdesc rb_txdesc[HME_NTXQ];
struct hme_txdq rb_txfreeq;
struct hme_txdq rb_txbusyq;
bus_dmamap_t rb_spare_dmamap;
};
struct hme_softc {
struct ifnet *sc_ifp;
struct ifmedia sc_ifmedia;
device_t sc_dev;
device_t sc_miibus;
struct mii_data *sc_mii; /* MII media control */
u_char sc_enaddr[ETHER_ADDR_LEN];
struct callout sc_tick_ch; /* tick callout */
int sc_wdog_timer; /* watchdog timer */
/* The following bus handles are to be provided by the bus front-end */
bus_dma_tag_t sc_pdmatag; /* bus dma parent tag */
bus_dma_tag_t sc_cdmatag; /* control bus dma tag */
bus_dmamap_t sc_cdmamap; /* control bus dma handle */
bus_dma_tag_t sc_rdmatag; /* RX bus dma tag */
bus_dma_tag_t sc_tdmatag; /* RX bus dma tag */
bus_space_handle_t sc_sebh; /* HME Global registers */
bus_space_handle_t sc_erxh; /* HME ERX registers */
bus_space_handle_t sc_etxh; /* HME ETX registers */
bus_space_handle_t sc_mach; /* HME MAC registers */
bus_space_handle_t sc_mifh; /* HME MIF registers */
bus_space_tag_t sc_sebt; /* HME Global registers */
bus_space_tag_t sc_erxt; /* HME ERX registers */
bus_space_tag_t sc_etxt; /* HME ETX registers */
bus_space_tag_t sc_mact; /* HME MAC registers */
bus_space_tag_t sc_mift; /* HME MIF registers */
int sc_burst; /* DVMA burst size in effect */
int sc_phys[2]; /* MII instance -> PHY map */
u_int sc_flags;
#define HME_LINK (1 << 0) /* link is up */
#define HME_PCI (1 << 1) /* PCI busses are little-endian */
int sc_ifflags;
int sc_csum_features;
/* Ring descriptor */
struct hme_ring sc_rb;
struct mtx sc_lock;
};
#define HME_LOCK(_sc) mtx_lock(&(_sc)->sc_lock)
#define HME_UNLOCK(_sc) mtx_unlock(&(_sc)->sc_lock)
#define HME_LOCK_ASSERT(_sc, _what) mtx_assert(&(_sc)->sc_lock, (_what))
extern devclass_t hme_devclass;
int hme_config(struct hme_softc *);
void hme_detach(struct hme_softc *);
void hme_suspend(struct hme_softc *);
void hme_resume(struct hme_softc *);
void hme_intr(void *);
/* MII methods & callbacks */
int hme_mii_readreg(device_t, int, int);
int hme_mii_writereg(device_t, int, int, int);
void hme_mii_statchg(device_t);
|