summaryrefslogtreecommitdiffstats
path: root/bsps/powerpc/beatnik/net/if_gfe/if_gfevar.h
diff options
context:
space:
mode:
Diffstat (limited to 'bsps/powerpc/beatnik/net/if_gfe/if_gfevar.h')
-rw-r--r--bsps/powerpc/beatnik/net/if_gfe/if_gfevar.h225
1 files changed, 225 insertions, 0 deletions
diff --git a/bsps/powerpc/beatnik/net/if_gfe/if_gfevar.h b/bsps/powerpc/beatnik/net/if_gfe/if_gfevar.h
new file mode 100644
index 0000000000..cbb9609cf8
--- /dev/null
+++ b/bsps/powerpc/beatnik/net/if_gfe/if_gfevar.h
@@ -0,0 +1,225 @@
+#ifndef IF_GFEVAR_H
+#define IF_GFEVAR_H
+/* $NetBSD: if_gfevar.h,v 1.4.10.1 2005/04/29 11:28:56 kent Exp $ */
+
+/*
+ * Copyright (c) 2002 Allegro Networks, Inc., Wasabi Systems, Inc.
+ * All rights reserved.
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions
+ * are met:
+ * 1. Redistributions of source code must retain the above copyright
+ * notice, this list of conditions and the following disclaimer.
+ * 2. Redistributions in binary form must reproduce the above copyright
+ * notice, this list of conditions and the following disclaimer in the
+ * documentation and/or other materials provided with the distribution.
+ * 3. All advertising materials mentioning features or use of this software
+ * must display the following acknowledgement:
+ * This product includes software developed for the NetBSD Project by
+ * Allegro Networks, Inc., and Wasabi Systems, Inc.
+ * 4. The name of Allegro Networks, Inc. may not be used to endorse
+ * or promote products derived from this software without specific prior
+ * written permission.
+ * 5. The name of Wasabi Systems, Inc. may not be used to endorse
+ * or promote products derived from this software without specific prior
+ * written permission.
+ *
+ * THIS SOFTWARE IS PROVIDED BY ALLEGRO NETWORKS, INC. AND
+ * WASABI SYSTEMS, INC. ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES,
+ * INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY
+ * AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
+ * IN NO EVENT SHALL EITHER ALLEGRO NETWORKS, INC. OR WASABI SYSTEMS, INC.
+ * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
+ * POSSIBILITY OF SUCH DAMAGE.
+ */
+
+/* NOTE: GE_RXDESC_MAX * 16 <= GE_RXDESC_MEMSIZE */
+/* NOTE: the driver needs 4*GE_RXDESC_MAX mbuf clusters (4 queues) */
+#ifndef __rtems__
+#define GE_RXDESC_MEMSIZE (1 * PAGE_SIZE)
+#define GE_RXDESC_MAX 64
+#define GE_RXBUF_SIZE 2048
+#define GE_RXBUF_MEMSIZE (GE_RXDESC_MAX*GE_RXBUF_SIZE)
+#else
+#define GE_RXDESC_MEMSIZE (GE_RXDESC_MAX * sizeof(struct gt_eth_desc))
+#define GE_RXDESC_MAX (sc->num_rxdesc)
+#define GE_RXBUF_MEMSIZE 0
+#endif
+
+#define GE_RXBUF_NSEGS ((GE_RXBUF_MEMSIZE/PAGE_SIZE)+1)
+#define GE_DMSEG_MAX (GE_RXBUF_NSEGS)
+
+struct gfe_dmamem {
+ bus_dmamap_t gdm_map; /* dmamem'ed memory */
+#ifdef __rtems__
+ void *gdm_unaligned_buf;
+#endif
+ caddr_t gdm_kva; /* kva of tx memory */
+ int gdm_nsegs; /* # of segment in gdm_segs */
+ int gdm_maxsegs; /* maximum # of segments allowed */
+ size_t gdm_size; /* size of memory region */
+ bus_dma_segment_t gdm_segs[GE_DMSEG_MAX]; /* dma segment of tx memory */
+};
+
+/* With a 4096 page size, we get 256 descriptors per page.
+ */
+#ifndef __rtems__
+#define GE_TXDESC_MEMSIZE (1 * PAGE_SIZE)
+#define GE_TXDESC_MAX (GE_TXDESC_MEMSIZE / 16)
+#define GE_TXBUF_SIZE (4 * PAGE_SIZE)
+#else
+#define GE_TXDESC_MEMSIZE (sc->num_txdesc * sizeof(struct gt_eth_desc))
+#define GE_TXDESC_MAX (sc->num_txdesc)
+#endif
+
+struct gfe_txqueue {
+ struct ifqueue txq_pendq; /* these are ready to go to the GT */
+ struct ifqueue txq_sentq;
+ struct gfe_dmamem txq_desc_mem; /* transmit descriptor memory */
+#ifndef __rtems__
+ struct gfe_dmamem txq_buf_mem; /* transmit buffer memory */
+#endif
+ unsigned int txq_lo; /* next to be given to GT */
+ unsigned int txq_fi; /* next to be returned to CPU */
+#ifndef __rtems__
+ unsigned int txq_ei_gapcount; /* counter until next EI */
+#endif
+ unsigned int txq_nactive; /* number of active descriptors */
+#ifndef __rtems__
+ unsigned int txq_outptr; /* where to put next transmit packet */
+ unsigned int txq_inptr; /* start of 1st queued tx packet */
+#endif
+ uint32_t txq_intrbits; /* bits to write to EIMR */
+ uint32_t txq_esdcmrbits; /* bits to write to ESDCMR */
+ uint32_t txq_epsrbits; /* bits to test with EPSR */
+ volatile struct gt_eth_desc *txq_descs; /* ptr to tx descriptors */
+ bus_addr_t txq_ectdp; /* offset to cur. tx desc ptr reg */
+ bus_addr_t txq_desc_busaddr; /* bus addr of tx descriptors */
+#ifndef __rtems__
+ bus_addr_t txq_buf_busaddr; /* bus addr of tx buffers */
+#endif
+};
+
+/* With a 4096 page size, we get 256 descriptors per page. We want 1024
+ * which will give us about 8ms of 64 byte packets (2ms for each priority
+ * queue).
+ */
+
+#ifndef __rtems__
+struct gfe_rxbuf {
+ uint8_t rb_data[GE_RXBUF_SIZE];
+};
+#endif
+
+struct gfe_rxqueue {
+ struct gfe_dmamem rxq_desc_mem; /* receive descriptor memory */
+#ifndef __rtems__
+ struct gfe_dmamem rxq_buf_mem; /* receive buffer memory */
+ struct mbuf *rxq_curpkt; /* mbuf for current packet */
+#endif
+ volatile struct gt_eth_desc *rxq_descs;
+#ifndef __rtems__
+ struct gfe_rxbuf *rxq_bufs;
+#else
+ struct mbuf **rxq_bufs;
+#endif
+ unsigned int rxq_fi; /* next to be returned to CPU */
+ unsigned int rxq_active; /* # of descriptors given to GT */
+ uint32_t rxq_intrbits; /* bits to write to EIMR */
+ bus_addr_t rxq_desc_busaddr; /* bus addr of rx descriptors */
+ uint32_t rxq_cmdsts; /* save cmdsts from first descriptor */
+ bus_size_t rxq_efrdp;
+ bus_size_t rxq_ecrdp;
+};
+
+enum gfe_txprio {
+ GE_TXPRIO_HI=1,
+ GE_TXPRIO_LO=0,
+ GE_TXPRIO_NONE=2
+};
+enum gfe_rxprio {
+ GE_RXPRIO_HI=3,
+ GE_RXPRIO_MEDHI=2,
+ GE_RXPRIO_MEDLO=1,
+ GE_RXPRIO_LO=0
+};
+
+#ifdef __rtems__
+#define sc_ec arpcom
+#define ec_if ac_if
+#define sc_dev arpcom
+#define dv_xname ac_if.if_name
+#endif
+
+struct gfe_softc {
+#ifndef __rtems__
+ struct device sc_dev; /* must be first */
+ struct ethercom sc_ec; /* common ethernet glue */
+ struct callout sc_co; /* resource recovery */
+ mii_data_t sc_mii; /* mii interface */
+
+ /*
+ *
+ */
+ bus_space_tag_t sc_gt_memt;
+ bus_space_handle_t sc_gt_memh;
+ bus_space_handle_t sc_memh; /* subregion for ethernet */
+ bus_dma_tag_t sc_dmat;
+#else
+ struct arpcom sc_ec;
+ unsigned sc_gt_memh;
+ unsigned sc_memh;
+ unsigned char irq_no;
+ rtems_id tid;
+ int sc_phyaddr;
+ int num_rxdesc, num_txdesc;
+#endif
+ int sc_macno; /* which mac? 0, 1, or 2 */
+
+ unsigned int sc_tickflags;
+#define GE_TICK_TX_IFSTART 0x0001
+#define GE_TICK_RX_RESTART 0x0002
+ unsigned int sc_flags;
+#define GE_ALLMULTI 0x0001
+#define GE_PHYSTSCHG 0x0002
+#define GE_RXACTIVE 0x0004
+#define GE_NOFREE 0x0008 /* Don't free on disable */
+ uint32_t sc_pcr; /* current EPCR value */
+ uint32_t sc_pcxr; /* current EPCXR value */
+ uint32_t sc_intrmask; /* current EIMR value */
+ uint32_t sc_idlemask; /* suspended EIMR bits */
+ size_t sc_max_frame_length; /* maximum frame length */
+
+ /*
+ * Hash table related members
+ */
+ struct gfe_dmamem sc_hash_mem; /* dma'ble hash table */
+ uint64_t *sc_hashtable;
+ unsigned int sc_hashmask; /* 0x1ff or 0x1fff */
+
+ /*
+ * Transmit related members
+ */
+ struct gfe_txqueue sc_txq[2]; /* High & Low transmit queues */
+
+ /*
+ * Receive related members
+ */
+ struct gfe_rxqueue sc_rxq[4]; /* Hi/MedHi/MedLo/Lo receive queues */
+};
+
+#ifdef __rtems__
+int
+gfe_mii_read(int phy, void *arg, unsigned reg, uint32_t *pval);
+
+int
+gfe_mii_write(int phy, void *arg, unsigned reg, uint32_t value);
+#endif
+
+#endif