diff options
Diffstat (limited to 'bsps/sparc/shared/tmtc/grtm.c')
-rw-r--r-- | bsps/sparc/shared/tmtc/grtm.c | 1623 |
1 files changed, 1623 insertions, 0 deletions
diff --git a/bsps/sparc/shared/tmtc/grtm.c b/bsps/sparc/shared/tmtc/grtm.c new file mode 100644 index 0000000000..bd7d98f972 --- /dev/null +++ b/bsps/sparc/shared/tmtc/grtm.c @@ -0,0 +1,1623 @@ +/* GRTM CCSDS Telemetry Encoder driver + * + * COPYRIGHT (c) 2007. + * Cobham Gaisler AB. + * + * The license and distribution terms for this file may be + * found in the file LICENSE in this distribution or at + * http://www.rtems.org/license/LICENSE. + */ + +#include <bsp.h> +#include <rtems/libio.h> +#include <stdlib.h> +#include <stdio.h> +#include <string.h> +#include <assert.h> +#include <ctype.h> +#include <malloc.h> +#include <rtems/bspIo.h> + +#include <drvmgr/drvmgr.h> +#include <ambapp.h> +#include <drvmgr/ambapp_bus.h> +#include <bsp/grtm.h> + +/* map via rtems_interrupt_lock_* API: */ +#define SPIN_DECLARE(lock) RTEMS_INTERRUPT_LOCK_MEMBER(lock) +#define SPIN_INIT(lock, name) rtems_interrupt_lock_initialize(lock, name) +#define SPIN_LOCK(lock, level) rtems_interrupt_lock_acquire_isr(lock, &level) +#define SPIN_LOCK_IRQ(lock, level) rtems_interrupt_lock_acquire(lock, &level) +#define SPIN_UNLOCK(lock, level) rtems_interrupt_lock_release_isr(lock, &level) +#define SPIN_UNLOCK_IRQ(lock, level) rtems_interrupt_lock_release(lock, &level) +#define SPIN_IRQFLAGS(k) rtems_interrupt_lock_context k +#define SPIN_ISR_IRQFLAGS(k) SPIN_IRQFLAGS(k) + +/* +#define DEBUG +#define DEBUGFUNCS +*/ + +#include <bsp/debug_defs.h> + +/* GRTM register map */ +struct grtm_regs { + volatile unsigned int dma_ctrl; /* DMA Control Register (0x00) */ + volatile unsigned int dma_status; /* DMA Status Register (0x04) */ + volatile unsigned int dma_len; /* DMA Length Register (0x08) */ + volatile unsigned int dma_bd; /* DMA Descriptor Pointer Register (0x0c) */ + + volatile unsigned int dma_cfg; /* DMA Configuration Register (0x10) */ + volatile unsigned int revision; /* GRTM Revision Register (0x14) */ + + int unused0[(0x80-0x18)/4]; + + volatile unsigned int ctrl; /* TM Control Register (0x80) */ + volatile unsigned int status; /* TM Status Register (0x84) */ + volatile unsigned int cfg; /* TM Configuration Register (0x88) */ + volatile unsigned int size; /* TM Size Register (0x8c) */ + + volatile unsigned int phy; /* TM Physical Layer Register (0x90) */ + volatile unsigned int code; /* TM Coding Sub-Layer Register (0x94) */ + volatile unsigned int asmr; /* TM Attached Synchronization Marker Register (0x98) */ + + int unused1; + + volatile unsigned int all_frm; /* TM All Frames Generation Register (0xa0) */ + volatile unsigned int mst_frm; /* TM Master Channel Frame Generation Register (0xa4) */ + volatile unsigned int idle_frm; /* TM Idle Frame Generation Register (0xa8) */ + + int unused2[(0xc0-0xac)/4]; + + volatile unsigned int fsh[4]; /* TM FSH/Insert Zone Registers (0xc0..0xcc) */ + + volatile unsigned int ocf; /* TM Operational Control Field Register (0xd0) */ +}; + +/* DMA Control Register (0x00) */ +#define GRTM_DMA_CTRL_EN_BIT 0 +#define GRTM_DMA_CTRL_IE_BIT 1 +#define GRTM_DMA_CTRL_TXRST_BIT 2 +#define GRTM_DMA_CTRL_RST_BIT 3 +#define GRTM_DMA_CTRL_TFIE_BIT 4 + +#define GRTM_DMA_CTRL_EN (1<<GRTM_DMA_CTRL_EN_BIT) +#define GRTM_DMA_CTRL_IE (1<<GRTM_DMA_CTRL_IE_BIT) +#define GRTM_DMA_CTRL_TXRST (1<<GRTM_DMA_CTRL_TXRST_BIT) +#define GRTM_DMA_CTRL_RST (1<<GRTM_DMA_CTRL_RST_BIT) +#define GRTM_DMA_CTRL_TFIE (1<<GRTM_DMA_CTRL_TFIE_BIT) + +/* DMA Status Register (0x04) */ +#define GRTM_DMA_STS_TE_BIT 0 +#define GRTM_DMA_STS_TI_BIT 1 +#define GRTM_DMA_STS_TA_BIT 2 +#define GRTM_DMA_STS_TFF_BIT 3 +#define GRTM_DMA_STS_TFS_BIT 4 + +#define GRTM_DMA_STS_TE (1<<GRTM_DMA_STS_TE_BIT) +#define GRTM_DMA_STS_TI (1<<GRTM_DMA_STS_TI_BIT) +#define GRTM_DMA_STS_TA (1<<GRTM_DMA_STS_TA_BIT) +#define GRTM_DMA_STS_TFF (1<<GRTM_DMA_STS_TFF_BIT) +#define GRTM_DMA_STS_TFS (1<<GRTM_DMA_STS_TFS_BIT) +#define GRTM_DMA_STS_ALL 0x1f + +/* DMA Length Register (0x08) */ +#define GRTM_DMA_LEN_LEN_BIT 0 +#define GRTM_DMA_LEN_LIM_BIT 16 + +#define GRTM_DMA_LEN_LEN (0x7ff<<GRTM_DMA_LEN_LEN_BIT) +#define GRTM_DMA_LEN_LIM (0x3ff<<GRTM_DMA_LEN_LIM_BIT) + +/* DMA Descriptor Pointer Register (0x0c) */ +#define GRTM_DMA_BD_INDEX_BIT 0 +#define GRTM_DMA_BD_BASE_BIT 10 + +#define GRTM_DMA_BD_INDEX (0x3ff<<GRTM_DMA_BD_INDEX_BIT) +#define GRTM_DMA_BD_BASE (0xfffffc<<GRTM_DMA_BD_BASE_BIT) + +/* DMA Configuration Register (0x10) */ +#define GRTM_DMA_CFG_BLKSZ_BIT 0 +#define GRTM_DMA_CFG_FIFOSZ_BIT 16 + +#define GRTM_DMA_CFG_BLKSZ (0xffff<<GRTM_DMA_CFG_BLKSZ_BIT) +#define GRTM_DMA_CFG_FIFOSZ (0xffff<<GRTM_DMA_CFG_FIFOSZ_BIT) + +/* TM Control Register (0x80) */ +#define GRTM_CTRL_EN_BIT 0 + +#define GRTM_CTRL_EN (1<<GRTM_CTRL_EN_BIT) + +/* TM Status Register (0x84) - Unused */ + +/* TM Configuration Register (0x88) */ +#define GRTM_CFG_SC_BIT 0 +#define GRTM_CFG_SP_BIT 1 +#define GRTM_CFG_CE_BIT 2 +#define GRTM_CFG_NRZ_BIT 3 +#define GRTM_CFG_PSR_BIT 4 +#define GRTM_CFG_TE_BIT 5 +#define GRTM_CFG_RSDEP_BIT 6 +#define GRTM_CFG_RS_BIT 9 +#define GRTM_CFG_AASM_BIT 11 +#define GRTM_CFG_FECF_BIT 12 +#define GRTM_CFG_OCF_BIT 13 +#define GRTM_CFG_EVC_BIT 14 +#define GRTM_CFG_IDLE_BIT 15 +#define GRTM_CFG_FSH_BIT 16 +#define GRTM_CFG_MCG_BIT 17 +#define GRTM_CFG_IZ_BIT 18 +#define GRTM_CFG_FHEC_BIT 19 +#define GRTM_CFG_AOS_BIT 20 +#define GRTM_CFG_CIF_BIT 21 +#define GRTM_CFG_OCFB_BIT 22 + +#define GRTM_CFG_SC (1<<GRTM_CFG_SC_BIT) +#define GRTM_CFG_SP (1<<GRTM_CFG_SP_BIT) +#define GRTM_CFG_CE (1<<GRTM_CFG_CE_BIT) +#define GRTM_CFG_NRZ (1<<GRTM_CFG_NRZ_BIT) +#define GRTM_CFG_PSR (1<<GRTM_CFG_PSR_BIT) +#define GRTM_CFG_TE (1<<GRTM_CFG_TE_BIT) +#define GRTM_CFG_RSDEP (0x7<<GRTM_CFG_RSDEP_BIT) +#define GRTM_CFG_RS (0x3<<GRTM_CFG_RS_BIT) +#define GRTM_CFG_AASM (1<<GRTM_CFG_AASM_BIT) +#define GRTM_CFG_FECF (1<<GRTM_CFG_FECF_BIT) +#define GRTM_CFG_OCF (1<<GRTM_CFG_OCF_BIT) +#define GRTM_CFG_EVC (1<<GRTM_CFG_EVC_BIT) +#define GRTM_CFG_IDLE (1<<GRTM_CFG_IDLE_BIT) +#define GRTM_CFG_FSH (1<<GRTM_CFG_FSH_BIT) +#define GRTM_CFG_MCG (1<<GRTM_CFG_MCG_BIT) +#define GRTM_CFG_IZ (1<<GRTM_CFG_IZ_BIT) +#define GRTM_CFG_FHEC (1<<GRTM_CFG_FHEC_BIT) +#define GRTM_CFG_AOS (1<<GRTM_CFG_AOS_BIT) +#define GRTM_CFG_CIF (1<<GRTM_CFG_CIF_BIT) +#define GRTM_CFG_OCFB (1<<GRTM_CFG_OCFB_BIT) + +/* TM Size Register (0x8c) */ +#define GRTM_SIZE_BLKSZ_BIT 0 +#define GRTM_SIZE_FIFOSZ_BIT 8 +#define GRTM_SIZE_LEN_BIT 20 + +#define GRTM_SIZE_BLKSZ (0xff<<GRTM_SIZE_BLKSZ_BIT) +#define GRTM_SIZE_FIFOSZ (0xfff<<GRTM_SIZE_FIFOSZ_BIT) +#define GRTM_SIZE_LEN (0xfff<<GRTM_SIZE_LEN_BIT) + +/* TM Physical Layer Register (0x90) */ +#define GRTM_PHY_SUB_BIT 0 +#define GRTM_PHY_SCF_BIT 15 +#define GRTM_PHY_SYM_BIT 16 +#define GRTM_PHY_SF_BIT 31 + +#define GRTM_PHY_SUB (0x7fff<<GRTM_PHY_SUB_BIT) +#define GRTM_PHY_SCF (1<<GRTM_PHY_SCF_BIT) +#define GRTM_PHY_SYM (0x7fff<<GRTM_PHY_SYM_BIT) +#define GRTM_PHY_SF (1<<GRTM_PHY_SF_BIT) + +/* TM Coding Sub-Layer Register (0x94) */ +#define GRTM_CODE_SC_BIT 0 +#define GRTM_CODE_SP_BIT 1 +#define GRTM_CODE_CERATE_BIT 2 +#define GRTM_CODE_CE_BIT 5 +#define GRTM_CODE_NRZ_BIT 6 +#define GRTM_CODE_PSR_BIT 7 +#define GRTM_CODE_RS8_BIT 11 +#define GRTM_CODE_RSDEP_BIT 12 +#define GRTM_CODE_RS_BIT 15 +#define GRTM_CODE_AASM_BIT 16 +#define GRTM_CODE_CSEL_BIT 17 + +#define GRTM_CODE_SC (1<<GRTM_CODE_SC_BIT) +#define GRTM_CODE_SP (1<<GRTM_CODE_SP_BIT) +#define GRTM_CODE_CERATE (0x7<<GRTM_CODE_CERATE_BIT) +#define GRTM_CODE_CE (1<<GRTM_CODE_CE_BIT) +#define GRTM_CODE_NRZ (1<<GRTM_CODE_NRZ_BIT) +#define GRTM_CODE_PSR (1<<GRTM_CODE_PSR_BIT) +#define GRTM_CODE_RS8 (1<<GRTM_CODE_RS8_BIT) +#define GRTM_CODE_RSDEP (0x7<<GRTM_CODE_RSDEP_BIT) +#define GRTM_CODE_RS (1<<GRTM_CODE_RS_BIT) +#define GRTM_CODE_AASM (1<<GRTM_CODE_AASM_BIT) +#define GRTM_CODE_CSEL (0x3<<GRTM_CODE_CSEL_BIT) + +/* TM Attached Synchronization Marker Register (0x98) */ +#define GRTM_ASM_BIT 0 + +#define GRTM_ASM 0xffffffff + +/* TM All Frames Generation Register (0xa0) */ +#define GRTM_ALL_LEN_BIT 0 +#define GRTM_ALL_VER_BIT 12 +#define GRTM_ALL_FHEC_BIT 14 +#define GRTM_ALL_FECF_BIT 15 +#define GRTM_ALL_IZ_BIT 16 +#define GRTM_ALL_IZLEN_BIT 17 + +#define GRTM_ALL_LEN (0x7ff<<GRTM_ALL_LEN_BIT) +#define GRTM_ALL_VER (0x3<<GRTM_ALL_VER_BIT) +#define GRTM_ALL_FHEC (1<<GRTM_ALL_FHEC_BIT) +#define GRTM_ALL_FECF (1<<GRTM_ALL_FECF_BIT) +#define GRTM_ALL_IZ (1<<GRTM_ALL_IZ_BIT) +#define GRTM_ALL_IZLEN (0x1f<<GRTM_ALL_IZLEN_BIT) + +/* TM Master Channel Frame Generation Register (0xa4) */ +#define GRTM_MST_OW_BIT 0 +#define GRTM_MST_OCF_BIT 1 +#define GRTM_MST_FSH_BIT 2 +#define GRTM_MST_MC_BIT 3 +#define GRTM_MST_MCCNTR_BIT 24 + +#define GRTM_MST_OW (1<<GRTM_MST_OW_BIT) +#define GRTM_MST_OCF (1<<GRTM_MST_OCF_BIT) +#define GRTM_MST_FSH (1<<GRTM_MST_FSH_BIT) +#define GRTM_MST_MC (0xff<<GRTM_MST_MC_BIT) + +/* TM Idle Frame Generation Register (0xa8) */ +#define GRTM_IDLE_SCID_BIT 0 +#define GRTM_IDLE_VCID_BIT 10 +#define GRTM_IDLE_MC_BIT 16 +#define GRTM_IDLE_VCC_BIT 17 +#define GRTM_IDLE_FSH_BIT 18 +#define GRTM_IDLE_EVC_BIT 19 +#define GRTM_IDLE_OCF_BIT 20 +#define GRTM_IDLE_IDLE_BIT 21 +#define GRTM_IDLE_MCCNTR_BIT 24 + +#define GRTM_IDLE_SCID (0x3ff<<GRTM_IDLE_SCID_BIT) +#define GRTM_IDLE_VCID (0x3f<<GRTM_IDLE_VCID_BIT) +#define GRTM_IDLE_MC (1<<GRTM_IDLE_MC_BIT) +#define GRTM_IDLE_VCC (1<<GRTM_IDLE_VCC_BIT) +#define GRTM_IDLE_FSH (1<<GRTM_IDLE_FSH_BIT) +#define GRTM_IDLE_EVC (1<<GRTM_IDLE_EVC_BIT) +#define GRTM_IDLE_OCF (1<<GRTM_IDLE_OCF_BIT) +#define GRTM_IDLE_IDLE (1<<GRTM_IDLE_IDLE_BIT) +#define GRTM_IDLE_MCCNTR (0xff<<GRTM_IDLE_MCCNTR_BIT) + +/* TM FSH/Insert Zone Registers (0xc0..0xcc) */ +#define GRTM_FSH_DATA_BIT 0 + +#define GRTM_FSH_DATA 0xffffffff + + +/* TM Operational Control Field Register (0xd0) */ +#define GRTM_OCF_CLCW_BIT 0 + +#define GRTM_OCF_CLCW 0xffffffff + + +/* GRTM Revision 0 */ +#define GRTM_REV0_DMA_CTRL_TXRDY_BIT 5 +#define GRTM_REV0_DMA_CTRL_TXRDY (1<<GRTM_REV0_DMA_CTRL_TXRDY_BIT) + +/* GRTM Revision 1 */ +#define GRTM_REV1_DMA_STS_TXRDY_BIT 6 +#define GRTM_REV1_DMA_STS_TXSTAT_BIT 7 +#define GRTM_REV1_DMA_STS_TXRDY (1<<GRTM_REV1_DMA_STS_TXRDY_BIT) +#define GRTM_REV1_DMA_STS_TXSTAT (1<<GRTM_REV1_DMA_STS_TXSTAT_BIT) + +#define GRTM_REV1_REV_SREV_BIT 0 +#define GRTM_REV1_REV_MREV_BIT 8 +#define GRTM_REV1_REV_TIRQ_BIT 16 +#define GRTM_REV1_REV_SREV (0xff<<GRTM_REV1_REV_SREV_BIT) +#define GRTM_REV1_REV_MREV (0xff<<GRTM_REV1_REV_MREV_BIT) +#define GRTM_REV1_REV_TIRQ (1<<GRTM_REV1_REV_TIRQ_BIT) + + +/* GRTM transmit descriptor (0x400 Alignment need) */ +struct grtm_bd { + volatile unsigned int ctrl; + unsigned int address; +}; + +#define GRTM_BD_EN_BIT 0 +#define GRTM_BD_WR_BIT 1 +#define GRTM_BD_IE_BIT 2 +#define GRTM_BD_FECFB_BIT 3 +#define GRTM_BD_IZB_BIT 4 +#define GRTM_BD_FHECB_BIT 5 +#define GRTM_BD_OCFB_BIT 6 +#define GRTM_BD_FSHB_BIT 7 +#define GRTM_BD_MCB_BIT 8 +#define GRTM_BD_VCE_BIT 9 +#define GRTM_BD_TS_BIT 14 +#define GRTM_BD_UE_BIT 15 + +#define GRTM_BD_EN (1<<GRTM_BD_EN_BIT) +#define GRTM_BD_WR (1<<GRTM_BD_WR_BIT) +#define GRTM_BD_IE (1<<GRTM_BD_IE_BIT) +#define GRTM_BD_FECFB (1<<GRTM_BD_FECFB_BIT) +#define GRTM_BD_IZB (1<<GRTM_BD_IZB_BIT) +#define GRTM_BD_FHECB (1<<GRTM_BD_FHECB_BIT) +#define GRTM_BD_OCFB (1<<GRTM_BD_OCFB_BIT) +#define GRTM_BD_FSHB (1<<GRTM_BD_FSHB_BIT) +#define GRTM_BD_MCB (1<<GRTM_BD_MCB_BIT) +#define GRTM_BD_VCE (1<<GRTM_BD_VCE_BIT) +#define GRTM_BD_TS (1<<GRTM_BD_TS_BIT) +#define GRTM_BD_UE (1<<GRTM_BD_UE_BIT) + +/* Load register */ + +#define READ_REG(address) (*(volatile unsigned int *)address) + +/* Driver functions */ +static rtems_device_driver grtm_initialize(rtems_device_major_number major, rtems_device_minor_number minor, void *arg); +static rtems_device_driver grtm_open(rtems_device_major_number major, rtems_device_minor_number minor, void *arg); +static rtems_device_driver grtm_close(rtems_device_major_number major, rtems_device_minor_number minor, void *arg); +static rtems_device_driver grtm_read(rtems_device_major_number major, rtems_device_minor_number minor, void *arg); +static rtems_device_driver grtm_write(rtems_device_major_number major, rtems_device_minor_number minor, void *arg); +static rtems_device_driver grtm_ioctl(rtems_device_major_number major, rtems_device_minor_number minor, void *arg); + +#define GRTM_DRIVER_TABLE_ENTRY { grtm_initialize, grtm_open, grtm_close, grtm_read, grtm_write, grtm_ioctl } + +static rtems_driver_address_table grtm_driver = GRTM_DRIVER_TABLE_ENTRY; + +/* Structure that connects BD with SoftWare Frame */ +struct grtm_ring { + struct grtm_ring *next; + struct grtm_bd *bd; + struct grtm_frame *frm; +}; + +struct grtm_priv { + struct drvmgr_dev *dev; /* Driver manager device */ + char devName[32]; /* Device Name */ + struct grtm_regs *regs; + int irq; + int minor; + int subrev; /* GRTM Revision */ + SPIN_DECLARE(devlock); /* spin-lock ISR protection */ + + int open; + int running; + + struct grtm_bd *bds; + void *_bds; + + /* Interrupt generation */ + int enable_cnt_curr;/* Down counter, when 0 the interrupt bit is set for next descriptor */ + volatile int handling_transmission; /* Tells ISR if user are active changing descriptors/queues */ + + struct grtm_ring *_ring; /* Root of ring */ + struct grtm_ring *ring; /* Next ring to use for new frames to be transmitted */ + struct grtm_ring *ring_end; /* Oldest activated ring used */ + + /* Collections of frames Ready to sent/ Scheduled for transmission/Sent + * frames waiting for the user to reclaim + */ + struct grtm_list ready; /* Frames Waiting for free BDs */ + struct grtm_list scheduled; /* Frames in BDs beeing transmitted */ + struct grtm_list sent; /* Sent Frames waiting for user to reclaim and reuse */ + + /* Number of frames in the lists */ + int ready_cnt; /* Number of ready frames */ + int scheduled_cnt; /* Number of scheduled frames */ + int sent_cnt; /* Number of sent frames */ + + struct grtm_ioc_hw hw_avail; /* Hardware support available */ + struct grtm_ioc_config config; + struct grtm_ioc_stats stats; + + rtems_id sem_tx; +}; + +/* Prototypes */ +static void *grtm_memalign(unsigned int boundary, unsigned int length, void *realbuf); +static void grtm_hw_reset(struct grtm_priv *pDev); +static void grtm_interrupt(void *arg); + +/* Common Global Variables */ +static rtems_id grtm_dev_sem; +static int grtm_driver_io_registered = 0; +static rtems_device_major_number grtm_driver_io_major = 0; + +/******************* Driver manager interface ***********************/ + +/* Driver prototypes */ +static int grtm_register_io(rtems_device_major_number *m); +static int grtm_device_init(struct grtm_priv *pDev); + +static int grtm_init2(struct drvmgr_dev *dev); +static int grtm_init3(struct drvmgr_dev *dev); + +static struct drvmgr_drv_ops grtm_ops = +{ + {NULL, grtm_init2, grtm_init3, NULL}, + NULL, + NULL +}; + +static struct amba_dev_id grtm_ids[] = +{ + {VENDOR_GAISLER, GAISLER_GRTM}, + {0, 0} /* Mark end of table */ +}; + +static struct amba_drv_info grtm_drv_info = +{ + { + DRVMGR_OBJ_DRV, /* Driver */ + NULL, /* Next driver */ + NULL, /* Device list */ + DRIVER_AMBAPP_GAISLER_GRTM_ID, /* Driver ID */ + "GRTM_DRV", /* Driver Name */ + DRVMGR_BUS_TYPE_AMBAPP, /* Bus Type */ + &grtm_ops, + NULL, /* Funcs */ + 0, /* No devices yet */ + 0, + }, + &grtm_ids[0] +}; + +void grtm_register_drv (void) +{ + DBG("Registering GRTM driver\n"); + drvmgr_drv_register(&grtm_drv_info.general); +} + +static int grtm_init2(struct drvmgr_dev *dev) +{ + struct grtm_priv *priv; + + DBG("GRTM[%d] on bus %s\n", dev->minor_drv, dev->parent->dev->name); + priv = dev->priv = malloc(sizeof(struct grtm_priv)); + if ( !priv ) + return DRVMGR_NOMEM; + memset(priv, 0, sizeof(*priv)); + priv->dev = dev; + + /* This core will not find other cores, so we wait for init2() */ + + return DRVMGR_OK; +} + +static int grtm_init3(struct drvmgr_dev *dev) +{ + struct grtm_priv *priv; + char prefix[32]; + rtems_status_code status; + + priv = dev->priv; + + /* Do initialization */ + + if ( grtm_driver_io_registered == 0) { + /* Register the I/O driver only once for all cores */ + if ( grtm_register_io(&grtm_driver_io_major) ) { + /* Failed to register I/O driver */ + dev->priv = NULL; + return DRVMGR_FAIL; + } + + grtm_driver_io_registered = 1; + } + + /* I/O system registered and initialized + * Now we take care of device initialization. + */ + if ( grtm_device_init(priv) ) { + return DRVMGR_FAIL; + } + + /* Get Filesystem name prefix */ + prefix[0] = '\0'; + if ( drvmgr_get_dev_prefix(dev, prefix) ) { + /* Failed to get prefix, make sure of a unique FS name + * by using the driver minor. + */ + sprintf(priv->devName, "/dev/grtm%d", dev->minor_drv); + } else { + /* Got special prefix, this means we have a bus prefix + * And we should use our "bus minor" + */ + sprintf(priv->devName, "/dev/%sgrtm%d", prefix, dev->minor_bus); + } + + SPIN_INIT(&priv->devlock, priv->devName); + + /* Register Device */ + status = rtems_io_register_name(priv->devName, grtm_driver_io_major, dev->minor_drv); + if (status != RTEMS_SUCCESSFUL) { + return DRVMGR_FAIL; + } + + return DRVMGR_OK; +} + +/******************* Driver Implementation ***********************/ + +static int grtm_register_io(rtems_device_major_number *m) +{ + rtems_status_code r; + + if ((r = rtems_io_register_driver(0, &grtm_driver, m)) == RTEMS_SUCCESSFUL) { + DBG("GRTM driver successfully registered, major: %d\n", *m); + } else { + switch(r) { + case RTEMS_TOO_MANY: + printk("GRTM rtems_io_register_driver failed: RTEMS_TOO_MANY\n"); + return -1; + case RTEMS_INVALID_NUMBER: + printk("GRTM rtems_io_register_driver failed: RTEMS_INVALID_NUMBER\n"); + return -1; + case RTEMS_RESOURCE_IN_USE: + printk("GRTM rtems_io_register_driver failed: RTEMS_RESOURCE_IN_USE\n"); + return -1; + default: + printk("GRTM rtems_io_register_driver failed\n"); + return -1; + } + } + return 0; +} + +static int grtm_device_init(struct grtm_priv *pDev) +{ + struct amba_dev_info *ambadev; + struct ambapp_core *pnpinfo; + union drvmgr_key_value *value; + + /* Get device information from AMBA PnP information */ + ambadev = (struct amba_dev_info *)pDev->dev->businfo; + if ( ambadev == NULL ) { + return -1; + } + pnpinfo = &ambadev->info; + pDev->irq = pnpinfo->irq; + pDev->regs = (struct grtm_regs *)pnpinfo->apb_slv->start; + pDev->minor = pDev->dev->minor_drv; + pDev->open = 0; + pDev->running = 0; + + /* Create Binary RX Semaphore with count = 0 */ + if ( rtems_semaphore_create(rtems_build_name('G', 'R', 'M', '0' + pDev->minor), + 0, + RTEMS_FIFO|RTEMS_SIMPLE_BINARY_SEMAPHORE|RTEMS_NO_INHERIT_PRIORITY|\ + RTEMS_LOCAL|RTEMS_NO_PRIORITY_CEILING, + 0, + &pDev->sem_tx) != RTEMS_SUCCESSFUL ) { + return -1; + } + + /* Allocate Memory for Buffer Descriptor Table, or let user provide a custom + * address. + */ + value = drvmgr_dev_key_get(pDev->dev, "bdTabAdr", DRVMGR_KT_POINTER); + if ( value ) { + pDev->bds = (struct grtm_bd *)value->ptr; + pDev->_bds = (void *)value->ptr; + } else { + pDev->bds = (struct grtm_bd *)grtm_memalign(0x400, 0x400, &pDev->_bds); + } + if ( !pDev->bds ) { + DBG("GRTM: Failed to allocate descriptor table\n"); + return -1; + } + memset(pDev->bds, 0, 0x400); + + pDev->_ring = malloc(sizeof(struct grtm_ring) * 128); + if ( !pDev->_ring ) { + return -1; + } + + /* Reset Hardware before attaching IRQ handler */ + grtm_hw_reset(pDev); + + /* Read SUB revision number, ignore */ + pDev->subrev = (READ_REG(&pDev->regs->revision) & GRTM_REV1_REV_SREV) + >> GRTM_REV1_REV_SREV_BIT; + + return 0; +} + + +static inline void grtm_list_clr(struct grtm_list *list) +{ + list->head = NULL; + list->tail = NULL; +} + +static void grtm_hw_reset(struct grtm_priv *pDev) +{ + /* Reset Core */ + pDev->regs->dma_ctrl = GRTM_DMA_CTRL_RST; +} + +static void grtm_hw_get_implementation(struct grtm_priv *pDev, struct grtm_ioc_hw *hwcfg) +{ + unsigned int cfg = READ_REG(&pDev->regs->cfg); + + hwcfg->cs = (cfg & GRTM_CFG_SC) ? 1:0; + hwcfg->sp = (cfg & GRTM_CFG_SP) ? 1:0; + hwcfg->ce = (cfg & GRTM_CFG_CE) ? 1:0; + hwcfg->nrz = (cfg & GRTM_CFG_NRZ) ? 1:0; + hwcfg->psr = (cfg & GRTM_CFG_PSR) ? 1:0; + hwcfg->te = (cfg & GRTM_CFG_TE) ? 1:0; + hwcfg->rsdep = (cfg & GRTM_CFG_RSDEP)>>GRTM_CFG_RSDEP_BIT; + hwcfg->rs = (cfg & GRTM_CFG_RS)>>GRTM_CFG_RS_BIT; + hwcfg->aasm = (cfg & GRTM_CFG_AASM) ? 1:0; + hwcfg->fecf = (cfg & GRTM_CFG_FECF) ? 1:0; + hwcfg->ocf = (cfg & GRTM_CFG_OCF) ? 1:0; + hwcfg->evc = (cfg & GRTM_CFG_EVC) ? 1:0; + hwcfg->idle = (cfg & GRTM_CFG_IDLE) ? 1:0; + hwcfg->fsh = (cfg & GRTM_CFG_FSH) ? 1:0; + hwcfg->mcg = (cfg & GRTM_CFG_MCG) ? 1:0; + hwcfg->iz = (cfg & GRTM_CFG_IZ) ? 1:0; + hwcfg->fhec = (cfg & GRTM_CFG_FHEC) ? 1:0; + hwcfg->aos = (cfg & GRTM_CFG_AOS) ? 1:0; + hwcfg->cif = (cfg & GRTM_CFG_CIF) ? 1:0; + hwcfg->ocfb = (cfg & GRTM_CFG_OCFB) ? 1:0; + + cfg = READ_REG(&pDev->regs->dma_cfg); + hwcfg->blk_size = (cfg & GRTM_DMA_CFG_BLKSZ) >> GRTM_DMA_CFG_BLKSZ_BIT; + hwcfg->fifo_size= (cfg & GRTM_DMA_CFG_FIFOSZ) >> GRTM_DMA_CFG_FIFOSZ_BIT; +} + + +/* TODO: Implement proper default calculation from hardware configuration */ +static void grtm_hw_get_default_modes(struct grtm_ioc_config *cfg, struct grtm_ioc_hw *hwcfg) +{ + cfg->mode = GRTM_MODE_TM; + cfg->frame_length = 223; + cfg->limit = 0; /* Make driver auto configure it on START, user may override with non-zero value */ + cfg->as_marker = 0x1ACFFC1D; + + /* Physical */ + cfg->phy_subrate = 1; + cfg->phy_symbolrate = 1; + cfg->phy_opts = 0; + + /* Coding Layer */ + cfg->code_rsdep = 1; + cfg->code_ce_rate = 0; + cfg->code_csel = 0; + cfg->code_opts = 0; + + /* All Frame Generation */ + cfg->all_izlen = 0; + cfg->all_opts = GRTM_IOC_ALL_FECF; + + /* Master Channel Frame Generation */ + if ( hwcfg->mcg ) { + cfg->mf_opts = GRTM_IOC_MF_MC; + } else { + cfg->mf_opts = 0; + } + + /* Idle Frame Generation */ + cfg->idle_scid = 0; + cfg->idle_vcid = 0; + if ( hwcfg->idle ) { + cfg->idle_opts = GRTM_IOC_IDLE_EN; + } else { + cfg->idle_opts = 0; + } + + /* Interrupt options */ + cfg->blocking = 0; /* non-blocking mode is default */ + cfg->enable_cnt = 16; /* generate interrupt every 16 descriptor */ + cfg->isr_desc_proc = 1; /* Let interrupt handler do descriptor processing */ + cfg->timeout = RTEMS_NO_TIMEOUT; + +} + +static void *grtm_memalign(unsigned int boundary, unsigned int length, void *realbuf) +{ + *(int *)realbuf = (int)malloc(length+boundary); + DBG("GRTM: Alloced %d (0x%x) bytes, requested: %d\n",length+boundary,length+boundary,length); + return (void *)(((*(unsigned int *)realbuf)+boundary) & ~(boundary-1)); +} + +static int grtm_hw_set_config(struct grtm_priv *pDev, struct grtm_ioc_config *cfg, struct grtm_ioc_hw *hwcfg) +{ + struct grtm_regs *regs = pDev->regs; + unsigned int tmp; + unsigned int limit; + + if ( cfg->limit == 0 ) { + /* Calculate Limit */ + if ( cfg->frame_length > hwcfg->blk_size ) { + limit = hwcfg->blk_size*2; + } else { + limit = cfg->frame_length; + } + } else { + /* Use user configured limit */ + limit = cfg->limit; + } + + /* Frame Length and Limit */ + regs->dma_len = (((limit-1) << GRTM_DMA_LEN_LIM_BIT) & GRTM_DMA_LEN_LIM)| + (((cfg->frame_length-1) << GRTM_DMA_LEN_LEN_BIT) & GRTM_DMA_LEN_LEN); + + /* Physical layer options */ + tmp = (cfg->phy_opts & (GRTM_IOC_PHY_SCF|GRTM_IOC_PHY_SF)) | + (((cfg->phy_symbolrate-1)<<GRTM_PHY_SYM_BIT) & GRTM_PHY_SYM) | (((cfg->phy_subrate-1)<<GRTM_PHY_SUB_BIT) & GRTM_PHY_SUB); + regs->phy = tmp; + + /* Coding Sub-layer Options */ + tmp = (cfg->code_opts & GRTM_IOC_CODE_ALL) | ((cfg->code_csel<<GRTM_CODE_CSEL_BIT) & GRTM_CODE_CSEL) | + (((cfg->code_rsdep-1)<<GRTM_CODE_RSDEP_BIT) & GRTM_CODE_RSDEP) | ((cfg->code_ce_rate<<GRTM_CODE_CERATE_BIT) & GRTM_CODE_CERATE); + regs->code = tmp; + + /* Attached synchronization marker register */ + regs->asmr = cfg->as_marker; + + /* All Frames Generation */ + tmp = ((cfg->all_opts & GRTM_IOC_ALL_ALL)<<14) | + ((cfg->all_izlen<<GRTM_ALL_IZLEN_BIT) & GRTM_ALL_IZLEN) | + ((cfg->mode<<GRTM_ALL_VER_BIT) & GRTM_ALL_VER); + regs->all_frm = tmp; + + /* Master Frame Generation */ + regs->mst_frm = cfg->mf_opts & GRTM_IOC_MF_ALL; + + /* Idle frame Generation */ + tmp = ((cfg->idle_opts & GRTM_IOC_IDLE_ALL) << 16) | + ((cfg->idle_vcid << GRTM_IDLE_VCID_BIT) & GRTM_IDLE_VCID) | + ((cfg->idle_scid << GRTM_IDLE_SCID_BIT) & GRTM_IDLE_SCID); + regs->idle_frm = tmp; + + return 0; +} + +static int grtm_start(struct grtm_priv *pDev) +{ + struct grtm_regs *regs = pDev->regs; + int i; + struct grtm_ioc_config *cfg = &pDev->config; + unsigned int txrdy; + + /* Clear Descriptors */ + memset(pDev->bds,0,0x400); + + /* Clear stats */ + memset(&pDev->stats,0,sizeof(struct grtm_ioc_stats)); + + /* Init Descriptor Ring */ + memset(pDev->_ring,0,sizeof(struct grtm_ring)*128); + for(i=0;i<127;i++){ + pDev->_ring[i].next = &pDev->_ring[i+1]; + pDev->_ring[i].bd = &pDev->bds[i]; + pDev->_ring[i].frm = NULL; + } + pDev->_ring[127].next = &pDev->_ring[0]; + pDev->_ring[127].bd = &pDev->bds[127]; + pDev->_ring[127].frm = NULL; + + pDev->ring = &pDev->_ring[0]; + pDev->ring_end = &pDev->_ring[0]; + + /* Clear Scheduled, Ready and Sent list */ + grtm_list_clr(&pDev->ready); + grtm_list_clr(&pDev->scheduled); + grtm_list_clr(&pDev->sent); + + /* Software init */ + pDev->handling_transmission = 0; + + /* Reset the transmitter */ + regs->dma_ctrl = GRTM_DMA_CTRL_TXRST; + regs->dma_ctrl = 0; /* Leave Reset */ + + /* Clear old interrupts */ + regs->dma_status = GRTM_DMA_STS_ALL; + + /* Set Descriptor Pointer Base register to point to first descriptor */ + drvmgr_translate_check(pDev->dev, CPUMEM_TO_DMA, (void *)pDev->bds, + (void **)®s->dma_bd, 0x400); + + /* Set hardware options as defined by config */ + if ( grtm_hw_set_config(pDev, cfg, &pDev->hw_avail) ) { + return RTEMS_IO_ERROR; + } + + /* Enable TM Transmitter */ + regs->ctrl = GRTM_CTRL_EN; + + /* Wait for TXRDY to be cleared */ + i=1000; + while( i > 0 ) { + asm volatile ("nop"::); + i--; + } + + /* Check transmitter startup OK */ + i = 1000000; + do { + /* Location of TXRDY Bit is different for different revisions */ + if ( pDev->subrev == 0 ) { + txrdy = READ_REG(®s->dma_ctrl) & + GRTM_REV0_DMA_CTRL_TXRDY; + } else { + txrdy = READ_REG(®s->dma_status) & + GRTM_REV1_DMA_STS_TXRDY; + } + if (txrdy != 0) + break; + + asm volatile ("nop"::); + } while ( --i > 0 ); + if ( i == 0 ) { + /* Reset Failed */ + DBG("GRTM: start: Reseting transmitter failed (%d)\n",i); + return RTEMS_IO_ERROR; + } + DBG("GRTM: reset time %d\n",i); + + /* Everything is configured, the TM transmitter is started + * and idle frames has been sent. + */ + + /* Mark running before enabling the DMA transmitter */ + pDev->running = 1; + + /* Enable interrupts (Error and DMA TX) */ + regs->dma_ctrl = GRTM_DMA_CTRL_IE; + + DBG("GRTM: STARTED\n"); + + return RTEMS_SUCCESSFUL; +} + +static void grtm_stop(struct grtm_priv *pDev) +{ + struct grtm_regs *regs = pDev->regs; + + /* Disable the transmitter & Interrupts */ + regs->dma_ctrl = 0; + + /* Clear any pending interrupt */ + regs->dma_status = GRTM_DMA_STS_ALL; + + DBG("GRTM: STOPPED\n"); + + /* Flush semaphore in case a thread is stuck waiting for TX Interrupts */ + rtems_semaphore_flush(pDev->sem_tx); +} + +static rtems_device_driver grtm_open( + rtems_device_major_number major, + rtems_device_minor_number minor, + void *arg) +{ + struct grtm_priv *pDev; + struct drvmgr_dev *dev; + + FUNCDBG(); + + if ( drvmgr_get_dev(&grtm_drv_info.general, minor, &dev) ) { + DBG("Wrong minor %d\n", minor); + return RTEMS_INVALID_NUMBER; + } + pDev = (struct grtm_priv *)dev->priv; + + /* Wait until we get semaphore */ + if ( rtems_semaphore_obtain(grtm_dev_sem, RTEMS_WAIT, RTEMS_NO_TIMEOUT) != RTEMS_SUCCESSFUL ){ + return RTEMS_INTERNAL_ERROR; + } + + /* Is device in use? */ + if ( pDev->open ){ + rtems_semaphore_release(grtm_dev_sem); + return RTEMS_RESOURCE_IN_USE; + } + + /* Mark device taken */ + pDev->open = 1; + + rtems_semaphore_release(grtm_dev_sem); + + DBG("grtm_open: OPENED minor %d (pDev: 0x%x)\n",pDev->minor,(unsigned int)pDev); + + /* Set defaults */ + pDev->config.timeout = RTEMS_NO_TIMEOUT; /* no timeout (wait forever) */ + pDev->config.blocking = 0; /* polling mode */ + + pDev->running = 0; /* not in running mode yet */ + + memset(&pDev->config,0,sizeof(pDev->config)); + + /* The core has been reset when we execute here, so it is possible + * to read out what HW is implemented from core. + */ + grtm_hw_get_implementation(pDev, &pDev->hw_avail); + + /* Get default modes */ + grtm_hw_get_default_modes(&pDev->config,&pDev->hw_avail); + + return RTEMS_SUCCESSFUL; +} + +static rtems_device_driver grtm_close(rtems_device_major_number major, rtems_device_minor_number minor, void *arg) +{ + struct grtm_priv *pDev; + struct drvmgr_dev *dev; + + FUNCDBG(); + + if ( drvmgr_get_dev(&grtm_drv_info.general, minor, &dev) ) { + return RTEMS_INVALID_NUMBER; + } + pDev = (struct grtm_priv *)dev->priv; + + if ( pDev->running ){ + drvmgr_interrupt_unregister(dev, 0, grtm_interrupt, pDev); + grtm_stop(pDev); + pDev->running = 0; + } + + /* Reset core */ + grtm_hw_reset(pDev); + + /* Clear descriptor area just for sure */ + memset(pDev->bds, 0, 0x400); + + /* Mark not open */ + pDev->open = 0; + + return RTEMS_SUCCESSFUL; +} + +static rtems_device_driver grtm_read(rtems_device_major_number major, rtems_device_minor_number minor, void *arg) +{ + FUNCDBG(); + return RTEMS_NOT_IMPLEMENTED; +} + +static rtems_device_driver grtm_write(rtems_device_major_number major, rtems_device_minor_number minor, void *arg) +{ + FUNCDBG(); + return RTEMS_NOT_IMPLEMENTED; +} + +/* Scans the desciptor table for scheduled frames that has been sent, + * and moves these frames from the head of the scheduled queue to the + * tail of the sent queue. + * + * Also, for all frames the status is updated. + * + * Return Value + * Number of frames freed. + */ +static int grtm_free_sent(struct grtm_priv *pDev) +{ + struct grtm_ring *curr; + struct grtm_frame *last_frm, *first_frm; + int freed_frame_cnt=0; + unsigned int ctrl; + + curr = pDev->ring_end; + + /* Step into TX ring to find sent frames */ + if ( !curr->frm ){ + /* No scheduled frames, abort */ + return 0; + } + + /* There has been messages scheduled ==> scheduled messages may have been + * transmitted and needs to be collected. + */ + + first_frm = curr->frm; + + /* Loop until first enabled unsent frame is found. + * A unused descriptor is indicated by an unassigned frm field + */ + while ( curr->frm && !((ctrl=READ_REG(&curr->bd->ctrl)) & GRTM_BD_EN) ){ + /* Handle one sent Frame */ + + /* Remember last handled frame so that insertion/removal from + * frames lists go fast. + */ + last_frm = curr->frm; + + /* 1. Set flags to indicate error(s) and other information */ + last_frm->flags |= GRTM_FLAGS_SENT; /* Mark sent */ + + /* Update Stats */ + pDev->stats.frames_sent++; + + /* Did packet encounter link error? */ + if ( ctrl & GRTM_BD_UE ) { + pDev->stats.err_underrun++; + last_frm->flags |= GRRM_FLAGS_ERR; + } + + curr->frm = NULL; /* Mark unused */ + + /* Increment */ + curr = curr->next; + freed_frame_cnt++; + } + + /* 1. Remove all handled frames from scheduled queue + * 2. Put all handled frames into sent queue + */ + if ( freed_frame_cnt > 0 ){ + + /* Save TX ring posistion */ + pDev->ring_end = curr; + + /* Remove all sent frames from scheduled list */ + if ( pDev->scheduled.tail == last_frm ){ + /* All scheduled frames sent... */ + pDev->scheduled.head = NULL; + pDev->scheduled.tail = NULL; + }else{ + pDev->scheduled.head = last_frm->next; + } + last_frm->next = NULL; + + /* Put all sent frames into "Sent queue" for user to + * collect, later on. + */ + if ( !pDev->sent.head ){ + /* Sent queue empty */ + pDev->sent.head = first_frm; + pDev->sent.tail = last_frm; + }else{ + pDev->sent.tail->next = first_frm; + pDev->sent.tail = last_frm; + } + } + return freed_frame_cnt; +} + + +/* Moves as many frames in the ready queue (as there are free descriptors for) + * to the scheduled queue. The free descriptors are then assigned one frame + * each and enabled for transmission. + * + * Return Value + * Returns number of frames moved from ready to scheduled queue + */ +static int grtm_schedule_ready(struct grtm_priv *pDev) +{ + int cnt; + unsigned int ctrl, dmactrl; + struct grtm_ring *curr_bd; + struct grtm_frame *curr_frm, *last_frm; + + if ( !pDev->ready.head ){ + return 0; + } + + cnt=0; + curr_frm = pDev->ready.head; + curr_bd = pDev->ring; + while( !curr_bd->frm ){ + /* Assign frame to descriptor */ + curr_bd->frm = curr_frm; + + /* Prepare descriptor address. Three cases: + * - GRTM core on same bus as CPU ==> no translation (Address used by CPU = address used by GRTM) + * - GRTM core on remote bus, and payload address given as used by CPU ==> Translation needed + * - GRTM core on remote bus, and payload address given as used by GRTM ==> no translation [ USER does custom translation] + */ + if ( curr_frm->flags & (GRTM_FLAGS_TRANSLATE|GRTM_FLAGS_TRANSLATE_AND_REMEMBER) ) { + /* Do translation */ + drvmgr_translate(pDev->dev, CPUMEM_TO_DMA, (void *)curr_frm->payload, (void **)&curr_bd->bd->address); + if ( curr_frm->flags & GRTM_FLAGS_TRANSLATE_AND_REMEMBER ) { + if ( curr_frm->payload != (unsigned int *)curr_bd->bd->address ) { + /* Translation needed */ + curr_frm->flags &= ~GRTM_FLAGS_TRANSLATE_AND_REMEMBER; + curr_frm->flags |= GRTM_FLAGS_TRANSLATE; + } else { + /* No Trnaslation needed */ + curr_frm->flags &= ~(GRTM_FLAGS_TRANSLATE|GRTM_FLAGS_TRANSLATE_AND_REMEMBER); + } + } + } else { + /* Custom translation or no translation needed */ + curr_bd->bd->address = (unsigned int)curr_frm->payload; + } + + ctrl = GRTM_BD_EN; + if ( curr_bd->next == pDev->_ring ){ + ctrl |= GRTM_BD_WR; /* Wrap around */ + } + /* Apply user options/flags */ + ctrl |= (curr_frm->flags & GRTM_FLAGS_MASK); + + /* Is this Frame going to be an interrupt Frame? */ + if ( (--pDev->enable_cnt_curr) <= 0 ){ + if ( pDev->config.enable_cnt == 0 ){ + pDev->enable_cnt_curr = 0x3fffffff; + }else{ + pDev->enable_cnt_curr = pDev->config.enable_cnt; + ctrl |= GRTM_BD_IE; + } + } + + /* Enable descriptor */ + curr_bd->bd->ctrl = ctrl; + + last_frm = curr_frm; + curr_bd = curr_bd->next; + cnt++; + + /* Get Next Frame from Ready Queue */ + if ( curr_frm == pDev->ready.tail ){ + /* Handled all in ready queue. */ + curr_frm = NULL; + break; + } + curr_frm = curr_frm->next; + } + + /* Has frames have been scheduled? */ + if ( cnt > 0 ){ + /* Make last frame mark end of chain, probably pointless... */ + last_frm->next = NULL; + + /* Insert scheduled packets into scheduled queue */ + if ( !pDev->scheduled.head ){ + /* empty scheduled queue */ + pDev->scheduled.head = pDev->ready.head; + pDev->scheduled.tail = last_frm; + }else{ + pDev->scheduled.tail->next = pDev->ready.head; + pDev->scheduled.tail = last_frm; + } + + /* Remove scheduled packets from ready queue */ + pDev->ready.head = curr_frm; + if ( !curr_frm ){ + pDev->ready.tail = NULL; + } + + /* Update TX ring posistion */ + pDev->ring = curr_bd; + + /* Make hardware aware of the newly enabled descriptors */ + dmactrl = READ_REG(&pDev->regs->dma_ctrl); + dmactrl &= ~(GRTM_DMA_CTRL_TXRST | GRTM_DMA_CTRL_RST); + dmactrl |= GRTM_DMA_CTRL_EN; + pDev->regs->dma_ctrl = dmactrl; + } + + return cnt; +} + +static void grtm_tx_process(struct grtm_priv *pDev) +{ + int num; + + /* Free used descriptors and put the sent frame into the "Sent queue" + * (SCHEDULED->SENT) + */ + num = grtm_free_sent(pDev); + pDev->scheduled_cnt -= num; + pDev->sent_cnt += num; + + /* Use all available free descriptors there are frames for + * in the ready queue. + * (READY->SCHEDULED) + */ + if (pDev->running) { + num = grtm_schedule_ready(pDev); + pDev->ready_cnt -= num; + pDev->scheduled_cnt += num; + } +} + +/* + * The TX lock protects user tasks from the ISR. If TX DMA interrupt occurs + * while the user task is processing the TX DMA descriptors the ISR will + * ignore interrupt the request by not processing the DMA table since that + * is done by the user task anyway. In SMP, when a user task enters the TX DMA + * processing while the ISR (on another CPU) is also processing the user task + * will loop waiting for the ISR to complete. + */ +static int grtm_request_txlock(struct grtm_priv *pDev, int block) +{ + SPIN_IRQFLAGS(irqflags); + int got_lock = 0; + + do { + SPIN_LOCK_IRQ(&pDev->devlock, irqflags); + if (pDev->handling_transmission == 0) { + pDev->handling_transmission = 1; + got_lock = 1; + } + SPIN_UNLOCK_IRQ(&pDev->devlock, irqflags); + } while (!got_lock && block); + + return got_lock; +} + +static inline int grtm_request_txlock_isr(struct grtm_priv *pDev) +{ + SPIN_ISR_IRQFLAGS(irqflags); + int got_lock = 0; + + SPIN_LOCK(&pDev->devlock, irqflags); + if (pDev->handling_transmission == 0) { + pDev->handling_transmission = 1; + got_lock = 1; + } + SPIN_UNLOCK(&pDev->devlock, irqflags); + + return got_lock; +} + +static inline void grtm_release_txlock(struct grtm_priv *pDev) +{ + pDev->handling_transmission = 0; +} + +static rtems_device_driver grtm_ioctl(rtems_device_major_number major, rtems_device_minor_number minor, void *arg) +{ + struct grtm_priv *pDev; + struct drvmgr_dev *dev; + rtems_libio_ioctl_args_t *ioarg = (rtems_libio_ioctl_args_t *)arg; + unsigned int *data; + int status; + struct grtm_ioc_config *cfg; + struct grtm_ioc_hw_status *hwregs; + struct grtm_list *chain; + struct grtm_frame *curr; + struct grtm_ioc_hw *hwimpl; + struct grtm_ioc_stats *stats; + int num,ret; + + FUNCDBG(); + + if ( drvmgr_get_dev(&grtm_drv_info.general, minor, &dev) ) { + return RTEMS_INVALID_NUMBER; + } + pDev = (struct grtm_priv *)dev->priv; + + if (!ioarg) + return RTEMS_INVALID_NAME; + + data = ioarg->buffer; + ioarg->ioctl_return = 0; + switch(ioarg->command) { + case GRTM_IOC_START: + if ( pDev->running ) { + return RTEMS_RESOURCE_IN_USE; /* EBUSY */ + } + if ( (status=grtm_start(pDev)) != RTEMS_SUCCESSFUL ){ + return status; + } + /* Register ISR & Enable interrupt */ + drvmgr_interrupt_register(dev, 0, "grtm", grtm_interrupt, pDev); + + /* Read and write are now open... */ + break; + + case GRTM_IOC_STOP: + if ( !pDev->running ) { + return RTEMS_RESOURCE_IN_USE; + } + + /* Disable interrupts */ + drvmgr_interrupt_unregister(dev, 0, grtm_interrupt, pDev); + grtm_stop(pDev); + pDev->running = 0; + break; + + case GRTM_IOC_ISSTARTED: + if ( !pDev->running ) { + return RTEMS_RESOURCE_IN_USE; + } + break; + + case GRTM_IOC_SET_BLOCKING_MODE: + if ( (unsigned int)data > GRTM_BLKMODE_BLK ) { + return RTEMS_INVALID_NAME; + } + DBG("GRTM: Set blocking mode: %d\n",(unsigned int)data); + pDev->config.blocking = (unsigned int)data; + break; + + case GRTM_IOC_SET_TIMEOUT: + DBG("GRTM: Timeout: %d\n",(unsigned int)data); + pDev->config.timeout = (rtems_interval)data; + break; + + case GRTM_IOC_SET_CONFIG: + cfg = (struct grtm_ioc_config *)data; + if ( !cfg ) { + return RTEMS_INVALID_NAME; + } + + if ( pDev->running ) { + return RTEMS_RESOURCE_IN_USE; + } + + pDev->config = *cfg; + break; + + case GRTM_IOC_GET_STATS: + stats = (struct grtm_ioc_stats *)data; + if ( !stats ) { + return RTEMS_INVALID_NAME; + } + memcpy(stats,&pDev->stats,sizeof(struct grtm_ioc_stats)); + break; + + case GRTM_IOC_CLR_STATS: + memset(&pDev->stats,0,sizeof(struct grtm_ioc_stats)); + break; + + case GRTM_IOC_GET_CONFIG: + cfg = (struct grtm_ioc_config *)data; + if ( !cfg ) { + return RTEMS_INVALID_NAME; + } + + *cfg = pDev->config; + break; + + case GRTM_IOC_GET_OCFREG: + if ( !pDev->hw_avail.ocf ) { + /* Hardware does not implement the OCF register */ + return RTEMS_NOT_DEFINED; + } + if ( !data ) { + return RTEMS_INVALID_NAME; + } + *(unsigned int **)data = (unsigned int *)&pDev->regs->ocf; + break; + + case GRTM_IOC_GET_HW_IMPL: + hwimpl = (struct grtm_ioc_hw *)data; + if ( !hwimpl ) { + return RTEMS_INVALID_NAME; + } + *hwimpl = pDev->hw_avail; + break; + + case GRTM_IOC_GET_HW_STATUS: + hwregs = (struct grtm_ioc_hw_status *)data; + if ( !hwregs ) { + return RTEMS_INVALID_NAME; + } + /* We disable interrupt in order to get a snapshot of the registers */ +/* TODO: implement hwregs */ + break; + + /* Put a chain of frames at the back of the "Ready frames" queue. This + * triggers the driver to put frames from the Ready queue into unused + * available descriptors. (Ready -> Scheduled) + */ + + case GRTM_IOC_SEND: + if ( !pDev->running ){ + return RTEMS_RESOURCE_IN_USE; + } + + /* Get pointer to frame chain wished be sent */ + chain = (struct grtm_list *)ioarg->buffer; + if ( !chain ){ + /* No new frames to send ==> just trigger hardware + * to send previously made ready frames to be sent. + * If someone else is processing the DMA we igore the + * request. + */ + if (grtm_request_txlock(pDev, 0)) { + grtm_tx_process(pDev); + grtm_release_txlock(pDev); + } + break; + } + if ( !chain->tail || !chain->head ){ + return RTEMS_INVALID_NAME; + } + + DBG("GRTM_SEND: head: 0x%x, tail: 0x%x\n",chain->head,chain->tail); + + /* Mark ready frames unsent by clearing GRTM_FLAGS_SENT of all frames */ + + num = 0; + curr = chain->head; + while(curr != chain->tail){ + curr->flags = curr->flags & ~(GRTM_FLAGS_SENT|GRRM_FLAGS_ERR); + curr = curr->next; + num++; + } + curr->flags = curr->flags & ~(GRTM_FLAGS_SENT|GRRM_FLAGS_ERR); + num++; + + /* wait until we get the device lock */ + grtm_request_txlock(pDev, 1); + + /* 1. Put frames into ready queue + * (New Frames->READY) + */ + if ( pDev->ready.head ){ + /* Frames already on ready queue (no free descriptors previously) ==> + * Put frames at end of ready queue + */ + pDev->ready.tail->next = chain->head; + pDev->ready.tail = chain->tail; + chain->tail->next = NULL; + }else{ + /* All frames is put into the ready queue for later processing */ + pDev->ready.head = chain->head; + pDev->ready.tail = chain->tail; + chain->tail->next = NULL; + } + pDev->ready_cnt += num; /* Added 'num' frames to ready queue */ + + /* 2. SCHEDULED->SENT + * 3. READY->SCHEDULED + */ + grtm_tx_process(pDev); + grtm_release_txlock(pDev); + break; + + /* Take all available sent frames from the "Sent frames" queue. + * If no frames has been sent, the thread may get blocked if in blocking + * mode. The blocking mode is not available if driver is not in running mode. + * + * Note this ioctl may return success even if the driver is not in STARTED mode. + * This is because in case of a error (link error of similar) and the driver switch + * from START to STOP mode we must still be able to get our frames back. + * + * Note in case the driver fails to send a frame for some reason (link error), + * the sent flag is set to 0 indicating a failure. + * + */ + case GRTM_IOC_RECLAIM: + /* Get pointer to were to place reaped chain */ + chain = (struct grtm_list *)ioarg->buffer; + if ( !chain ){ + return RTEMS_INVALID_NAME; + } + + /* Lock out interrupt handler */ + grtm_request_txlock(pDev, 1); + + do { + /* Process descriptor table and populate with new + * buffers: + * * SCHEDULED->SENT + * * READY->SCHEDULED + */ + grtm_tx_process(pDev); + + /* Are there any frames on the sent queue waiting to be + * reclaimed? + */ + + if ( !pDev->sent.head ){ + /* No frames to reclaim - no frame in sent queue. + * Instead we block thread until frames have been sent + * if in blocking mode. + */ + if ( pDev->running && pDev->config.blocking ){ + ret = rtems_semaphore_obtain(pDev->sem_tx,RTEMS_WAIT,pDev->config.timeout); + if ( ret == RTEMS_TIMEOUT ) { + grtm_release_txlock(pDev); + return RTEMS_TIMEOUT; + } else if ( ret == RTEMS_SUCCESSFUL ) { + /* There might be frames available, go check */ + continue; + } else { + /* any error (driver closed, internal error etc.) */ + grtm_release_txlock(pDev); + return RTEMS_UNSATISFIED; + } + + }else{ + /* non-blocking mode, we quit */ + chain->head = NULL; + chain->tail = NULL; + /* do not lock out interrupt handler any more */ + grtm_release_txlock(pDev); + return RTEMS_TIMEOUT; + } + }else{ + /* Take all sent framess from sent queue to userspace queue */ + chain->head = pDev->sent.head; + chain->tail = pDev->sent.tail; + chain->tail->next = NULL; /* Just for sure */ + + /* Mark no Sent */ + grtm_list_clr(&pDev->sent); + pDev->sent_cnt = 0; + + DBG("TX_RECLAIM: head: 0x%x, tail: 0x%x\n",chain->head,chain->tail); + break; + } + + }while(1); + + /* do not lock out interrupt handler any more */ + grtm_release_txlock(pDev); + break; + + default: + return RTEMS_NOT_DEFINED; + } + return RTEMS_SUCCESSFUL; +} + +static void grtm_interrupt(void *arg) +{ + struct grtm_priv *pDev = arg; + struct grtm_regs *regs = pDev->regs; + unsigned int status; + + /* Clear interrupt by reading it */ + status = READ_REG(®s->dma_status); + + /* Spurious Interrupt? */ + if ( !pDev->running || !status) + return; + + regs->dma_status = status; + + if ( status & GRTM_DMA_STS_TFF ){ + pDev->stats.err_transfer_frame++; + } + + if ( status & GRTM_DMA_STS_TA ){ + pDev->stats.err_ahb++; + } + + if ( status & GRTM_DMA_STS_TE ){ + pDev->stats.err_tx++; + } + + if ( status & GRTM_DMA_STS_TI ){ + + if ( pDev->config.isr_desc_proc) { + if (grtm_request_txlock_isr(pDev)) { + grtm_tx_process(pDev); + grtm_release_txlock(pDev); + } + +#if 0 + if ( (pDev->config.blocking==GRTM_BLKMODE_COMPLETE) && pDev->timeout ){ + /* Signal to thread only if enough data is available */ + if ( pDev->wait_for_frames > grtm_data_avail(pDev) ){ + /* Not enough data available */ + goto procceed_processing_interrupts; + } + + /* Enough number of frames has been transmitted which means that + * the waiting thread should be woken up. + */ + rtems_semaphore_release(pDev->sem_tx); + } +#endif + } + + if ( pDev->config.blocking == GRTM_BLKMODE_BLK ) { + /* Blocking mode */ + +#if 0 + /* Disable further Interrupts until handled by waiting task. */ + regs->dma_ctrl = READ_REG(®s->dma_ctrl) & ~GRTM_DMA_CTRL_IE; +#endif + + /* Signal Semaphore to wake waiting thread in ioctl(SEND|RECLAIM) */ + rtems_semaphore_release(pDev->sem_tx); + } + + } +#if 0 +procceed_processing_interrupts: + ; +#endif +} + +static rtems_device_driver grtm_initialize( + rtems_device_major_number major, + rtems_device_minor_number unused, + void *arg + ) +{ + /* Device Semaphore created with count = 1 */ + if ( rtems_semaphore_create(rtems_build_name('G', 'R', 'T', 'M'), + 1, + RTEMS_FIFO|RTEMS_NO_INHERIT_PRIORITY|RTEMS_LOCAL|RTEMS_NO_PRIORITY_CEILING, + 0, + &grtm_dev_sem) != RTEMS_SUCCESSFUL ) { + return RTEMS_INTERNAL_ERROR; + } + + return RTEMS_SUCCESSFUL; +} |