2005-04-16 16:20:36 -06:00
|
|
|
/*
|
|
|
|
* Toshiba TC86C001 ("Goku-S") USB Device Controller driver
|
|
|
|
*
|
|
|
|
* Copyright (C) 2000-2002 Lineo
|
|
|
|
* by Stuart Lynne, Tom Rushworth, and Bruce Balden
|
|
|
|
* Copyright (C) 2002 Toshiba Corporation
|
|
|
|
* Copyright (C) 2003 MontaVista Software (source@mvista.com)
|
|
|
|
*
|
|
|
|
* This file is licensed under the terms of the GNU General Public
|
|
|
|
* License version 2. This program is licensed "as is" without any
|
|
|
|
* warranty of any kind, whether express or implied.
|
|
|
|
*/
|
|
|
|
|
|
|
|
/*
|
|
|
|
* This device has ep0 and three semi-configurable bulk/interrupt endpoints.
|
|
|
|
*
|
|
|
|
* - Endpoint numbering is fixed: ep{1,2,3}-bulk
|
|
|
|
* - Gadget drivers can choose ep maxpacket (8/16/32/64)
|
|
|
|
* - Gadget drivers can choose direction (IN, OUT)
|
|
|
|
* - DMA works with ep1 (OUT transfers) and ep2 (IN transfers).
|
|
|
|
*/
|
|
|
|
|
|
|
|
#undef DEBUG
|
|
|
|
// #define VERBOSE /* extra debug messages (success too) */
|
|
|
|
// #define USB_TRACE /* packet-level success messages */
|
|
|
|
|
|
|
|
#include <linux/kernel.h>
|
|
|
|
#include <linux/module.h>
|
|
|
|
#include <linux/pci.h>
|
|
|
|
#include <linux/delay.h>
|
|
|
|
#include <linux/ioport.h>
|
|
|
|
#include <linux/sched.h>
|
|
|
|
#include <linux/slab.h>
|
|
|
|
#include <linux/smp_lock.h>
|
|
|
|
#include <linux/errno.h>
|
|
|
|
#include <linux/init.h>
|
|
|
|
#include <linux/timer.h>
|
|
|
|
#include <linux/list.h>
|
|
|
|
#include <linux/interrupt.h>
|
|
|
|
#include <linux/proc_fs.h>
|
|
|
|
#include <linux/device.h>
|
|
|
|
#include <linux/usb_ch9.h>
|
|
|
|
#include <linux/usb_gadget.h>
|
|
|
|
|
|
|
|
#include <asm/byteorder.h>
|
|
|
|
#include <asm/io.h>
|
|
|
|
#include <asm/irq.h>
|
|
|
|
#include <asm/system.h>
|
|
|
|
#include <asm/unaligned.h>
|
|
|
|
|
|
|
|
|
|
|
|
#include "goku_udc.h"
|
|
|
|
|
|
|
|
#define DRIVER_DESC "TC86C001 USB Device Controller"
|
|
|
|
#define DRIVER_VERSION "30-Oct 2003"
|
|
|
|
|
|
|
|
#define DMA_ADDR_INVALID (~(dma_addr_t)0)
|
|
|
|
|
|
|
|
static const char driver_name [] = "goku_udc";
|
|
|
|
static const char driver_desc [] = DRIVER_DESC;
|
|
|
|
|
|
|
|
MODULE_AUTHOR("source@mvista.com");
|
|
|
|
MODULE_DESCRIPTION(DRIVER_DESC);
|
|
|
|
MODULE_LICENSE("GPL");
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
* IN dma behaves ok under testing, though the IN-dma abort paths don't
|
|
|
|
* seem to behave quite as expected. Used by default.
|
|
|
|
*
|
|
|
|
* OUT dma documents design problems handling the common "short packet"
|
2005-05-07 14:05:52 -06:00
|
|
|
* transfer termination policy; it couldn't be enabled by default, even
|
2005-04-16 16:20:36 -06:00
|
|
|
* if the OUT-dma abort problems had a resolution.
|
|
|
|
*/
|
|
|
|
static unsigned use_dma = 1;
|
|
|
|
|
|
|
|
#if 0
|
|
|
|
//#include <linux/moduleparam.h>
|
|
|
|
/* "modprobe goku_udc use_dma=1" etc
|
|
|
|
* 0 to disable dma
|
|
|
|
* 1 to use IN dma only (normal operation)
|
|
|
|
* 2 to use IN and OUT dma
|
|
|
|
*/
|
|
|
|
module_param(use_dma, uint, S_IRUGO);
|
|
|
|
#endif
|
|
|
|
|
|
|
|
/*-------------------------------------------------------------------------*/
|
|
|
|
|
|
|
|
static void nuke(struct goku_ep *, int status);
|
|
|
|
|
|
|
|
static inline void
|
|
|
|
command(struct goku_udc_regs __iomem *regs, int command, unsigned epnum)
|
|
|
|
{
|
|
|
|
writel(COMMAND_EP(epnum) | command, ®s->Command);
|
|
|
|
udelay(300);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
goku_ep_enable(struct usb_ep *_ep, const struct usb_endpoint_descriptor *desc)
|
|
|
|
{
|
|
|
|
struct goku_udc *dev;
|
|
|
|
struct goku_ep *ep;
|
|
|
|
u32 mode;
|
|
|
|
u16 max;
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
ep = container_of(_ep, struct goku_ep, ep);
|
|
|
|
if (!_ep || !desc || ep->desc
|
|
|
|
|| desc->bDescriptorType != USB_DT_ENDPOINT)
|
|
|
|
return -EINVAL;
|
|
|
|
dev = ep->dev;
|
|
|
|
if (ep == &dev->ep[0])
|
|
|
|
return -EINVAL;
|
|
|
|
if (!dev->driver || dev->gadget.speed == USB_SPEED_UNKNOWN)
|
|
|
|
return -ESHUTDOWN;
|
|
|
|
if (ep->num != (desc->bEndpointAddress & 0x0f))
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
switch (desc->bmAttributes & USB_ENDPOINT_XFERTYPE_MASK) {
|
|
|
|
case USB_ENDPOINT_XFER_BULK:
|
|
|
|
case USB_ENDPOINT_XFER_INT:
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
if ((readl(ep->reg_status) & EPxSTATUS_EP_MASK)
|
|
|
|
!= EPxSTATUS_EP_INVALID)
|
|
|
|
return -EBUSY;
|
|
|
|
|
|
|
|
/* enabling the no-toggle interrupt mode would need an api hook */
|
|
|
|
mode = 0;
|
|
|
|
max = le16_to_cpu(get_unaligned(&desc->wMaxPacketSize));
|
|
|
|
switch (max) {
|
|
|
|
case 64: mode++;
|
|
|
|
case 32: mode++;
|
|
|
|
case 16: mode++;
|
|
|
|
case 8: mode <<= 3;
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
mode |= 2 << 1; /* bulk, or intr-with-toggle */
|
|
|
|
|
|
|
|
/* ep1/ep2 dma direction is chosen early; it works in the other
|
|
|
|
* direction, with pio. be cautious with out-dma.
|
|
|
|
*/
|
|
|
|
ep->is_in = (USB_DIR_IN & desc->bEndpointAddress) != 0;
|
|
|
|
if (ep->is_in) {
|
|
|
|
mode |= 1;
|
|
|
|
ep->dma = (use_dma != 0) && (ep->num == UDC_MSTRD_ENDPOINT);
|
|
|
|
} else {
|
|
|
|
ep->dma = (use_dma == 2) && (ep->num == UDC_MSTWR_ENDPOINT);
|
|
|
|
if (ep->dma)
|
|
|
|
DBG(dev, "%s out-dma hides short packets\n",
|
|
|
|
ep->ep.name);
|
|
|
|
}
|
|
|
|
|
|
|
|
spin_lock_irqsave(&ep->dev->lock, flags);
|
|
|
|
|
|
|
|
/* ep1 and ep2 can do double buffering and/or dma */
|
|
|
|
if (ep->num < 3) {
|
|
|
|
struct goku_udc_regs __iomem *regs = ep->dev->regs;
|
|
|
|
u32 tmp;
|
|
|
|
|
|
|
|
/* double buffer except (for now) with pio in */
|
|
|
|
tmp = ((ep->dma || !ep->is_in)
|
|
|
|
? 0x10 /* double buffered */
|
|
|
|
: 0x11 /* single buffer */
|
|
|
|
) << ep->num;
|
|
|
|
tmp |= readl(®s->EPxSingle);
|
|
|
|
writel(tmp, ®s->EPxSingle);
|
|
|
|
|
|
|
|
tmp = (ep->dma ? 0x10/*dma*/ : 0x11/*pio*/) << ep->num;
|
|
|
|
tmp |= readl(®s->EPxBCS);
|
|
|
|
writel(tmp, ®s->EPxBCS);
|
|
|
|
}
|
|
|
|
writel(mode, ep->reg_mode);
|
|
|
|
command(ep->dev->regs, COMMAND_RESET, ep->num);
|
|
|
|
ep->ep.maxpacket = max;
|
|
|
|
ep->stopped = 0;
|
|
|
|
ep->desc = desc;
|
|
|
|
spin_unlock_irqrestore(&ep->dev->lock, flags);
|
|
|
|
|
|
|
|
DBG(dev, "enable %s %s %s maxpacket %u\n", ep->ep.name,
|
|
|
|
ep->is_in ? "IN" : "OUT",
|
|
|
|
ep->dma ? "dma" : "pio",
|
|
|
|
max);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ep_reset(struct goku_udc_regs __iomem *regs, struct goku_ep *ep)
|
|
|
|
{
|
|
|
|
struct goku_udc *dev = ep->dev;
|
|
|
|
|
|
|
|
if (regs) {
|
|
|
|
command(regs, COMMAND_INVALID, ep->num);
|
|
|
|
if (ep->num) {
|
|
|
|
if (ep->num == UDC_MSTWR_ENDPOINT)
|
|
|
|
dev->int_enable &= ~(INT_MSTWREND
|
|
|
|
|INT_MSTWRTMOUT);
|
|
|
|
else if (ep->num == UDC_MSTRD_ENDPOINT)
|
|
|
|
dev->int_enable &= ~INT_MSTRDEND;
|
|
|
|
dev->int_enable &= ~INT_EPxDATASET (ep->num);
|
|
|
|
} else
|
|
|
|
dev->int_enable &= ~INT_EP0;
|
|
|
|
writel(dev->int_enable, ®s->int_enable);
|
|
|
|
readl(®s->int_enable);
|
|
|
|
if (ep->num < 3) {
|
|
|
|
struct goku_udc_regs __iomem *r = ep->dev->regs;
|
|
|
|
u32 tmp;
|
|
|
|
|
|
|
|
tmp = readl(&r->EPxSingle);
|
|
|
|
tmp &= ~(0x11 << ep->num);
|
|
|
|
writel(tmp, &r->EPxSingle);
|
|
|
|
|
|
|
|
tmp = readl(&r->EPxBCS);
|
|
|
|
tmp &= ~(0x11 << ep->num);
|
|
|
|
writel(tmp, &r->EPxBCS);
|
|
|
|
}
|
|
|
|
/* reset dma in case we're still using it */
|
|
|
|
if (ep->dma) {
|
|
|
|
u32 master;
|
|
|
|
|
|
|
|
master = readl(®s->dma_master) & MST_RW_BITS;
|
|
|
|
if (ep->num == UDC_MSTWR_ENDPOINT) {
|
|
|
|
master &= ~MST_W_BITS;
|
|
|
|
master |= MST_WR_RESET;
|
|
|
|
} else {
|
|
|
|
master &= ~MST_R_BITS;
|
|
|
|
master |= MST_RD_RESET;
|
|
|
|
}
|
|
|
|
writel(master, ®s->dma_master);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
ep->ep.maxpacket = MAX_FIFO_SIZE;
|
|
|
|
ep->desc = NULL;
|
|
|
|
ep->stopped = 1;
|
|
|
|
ep->irqs = 0;
|
|
|
|
ep->dma = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int goku_ep_disable(struct usb_ep *_ep)
|
|
|
|
{
|
|
|
|
struct goku_ep *ep;
|
|
|
|
struct goku_udc *dev;
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
ep = container_of(_ep, struct goku_ep, ep);
|
|
|
|
if (!_ep || !ep->desc)
|
|
|
|
return -ENODEV;
|
|
|
|
dev = ep->dev;
|
|
|
|
if (dev->ep0state == EP0_SUSPEND)
|
|
|
|
return -EBUSY;
|
|
|
|
|
|
|
|
VDBG(dev, "disable %s\n", _ep->name);
|
|
|
|
|
|
|
|
spin_lock_irqsave(&dev->lock, flags);
|
|
|
|
nuke(ep, -ESHUTDOWN);
|
|
|
|
ep_reset(dev->regs, ep);
|
|
|
|
spin_unlock_irqrestore(&dev->lock, flags);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*-------------------------------------------------------------------------*/
|
|
|
|
|
|
|
|
static struct usb_request *
|
2005-10-21 01:21:58 -06:00
|
|
|
goku_alloc_request(struct usb_ep *_ep, gfp_t gfp_flags)
|
2005-04-16 16:20:36 -06:00
|
|
|
{
|
|
|
|
struct goku_request *req;
|
|
|
|
|
|
|
|
if (!_ep)
|
|
|
|
return NULL;
|
2006-02-27 14:34:10 -07:00
|
|
|
req = kzalloc(sizeof *req, gfp_flags);
|
2005-04-16 16:20:36 -06:00
|
|
|
if (!req)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
req->req.dma = DMA_ADDR_INVALID;
|
|
|
|
INIT_LIST_HEAD(&req->queue);
|
|
|
|
return &req->req;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
goku_free_request(struct usb_ep *_ep, struct usb_request *_req)
|
|
|
|
{
|
|
|
|
struct goku_request *req;
|
|
|
|
|
|
|
|
if (!_ep || !_req)
|
|
|
|
return;
|
|
|
|
|
|
|
|
req = container_of(_req, struct goku_request, req);
|
|
|
|
WARN_ON(!list_empty(&req->queue));
|
|
|
|
kfree(req);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*-------------------------------------------------------------------------*/
|
|
|
|
|
|
|
|
#undef USE_KMALLOC
|
|
|
|
|
|
|
|
/* many common platforms have dma-coherent caches, which means that it's
|
|
|
|
* safe to use kmalloc() memory for all i/o buffers without using any
|
|
|
|
* cache flushing calls. (unless you're trying to share cache lines
|
|
|
|
* between dma and non-dma activities, which is a slow idea in any case.)
|
|
|
|
*
|
|
|
|
* other platforms need more care, with 2.6 having a moderately general
|
|
|
|
* solution except for the common "buffer is smaller than a page" case.
|
|
|
|
*/
|
|
|
|
#if defined(CONFIG_X86)
|
|
|
|
#define USE_KMALLOC
|
|
|
|
|
2005-05-07 14:05:52 -06:00
|
|
|
#elif defined(CONFIG_MIPS) && !defined(CONFIG_DMA_NONCOHERENT)
|
2005-04-16 16:20:36 -06:00
|
|
|
#define USE_KMALLOC
|
|
|
|
|
|
|
|
#elif defined(CONFIG_PPC) && !defined(CONFIG_NOT_COHERENT_CACHE)
|
|
|
|
#define USE_KMALLOC
|
|
|
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
/* allocating buffers this way eliminates dma mapping overhead, which
|
|
|
|
* on some platforms will mean eliminating a per-io buffer copy. with
|
|
|
|
* some kinds of system caches, further tweaks may still be needed.
|
|
|
|
*/
|
|
|
|
static void *
|
|
|
|
goku_alloc_buffer(struct usb_ep *_ep, unsigned bytes,
|
2005-10-21 01:21:58 -06:00
|
|
|
dma_addr_t *dma, gfp_t gfp_flags)
|
2005-04-16 16:20:36 -06:00
|
|
|
{
|
|
|
|
void *retval;
|
|
|
|
struct goku_ep *ep;
|
|
|
|
|
|
|
|
ep = container_of(_ep, struct goku_ep, ep);
|
|
|
|
if (!_ep)
|
|
|
|
return NULL;
|
|
|
|
*dma = DMA_ADDR_INVALID;
|
|
|
|
|
|
|
|
#if defined(USE_KMALLOC)
|
|
|
|
retval = kmalloc(bytes, gfp_flags);
|
|
|
|
if (retval)
|
|
|
|
*dma = virt_to_phys(retval);
|
|
|
|
#else
|
|
|
|
if (ep->dma) {
|
|
|
|
/* the main problem with this call is that it wastes memory
|
|
|
|
* on typical 1/N page allocations: it allocates 1-N pages.
|
|
|
|
*/
|
|
|
|
#warning Using dma_alloc_coherent even with buffers smaller than a page.
|
|
|
|
retval = dma_alloc_coherent(&ep->dev->pdev->dev,
|
|
|
|
bytes, dma, gfp_flags);
|
|
|
|
} else
|
|
|
|
retval = kmalloc(bytes, gfp_flags);
|
|
|
|
#endif
|
|
|
|
return retval;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
goku_free_buffer(struct usb_ep *_ep, void *buf, dma_addr_t dma, unsigned bytes)
|
|
|
|
{
|
|
|
|
/* free memory into the right allocator */
|
|
|
|
#ifndef USE_KMALLOC
|
|
|
|
if (dma != DMA_ADDR_INVALID) {
|
|
|
|
struct goku_ep *ep;
|
|
|
|
|
|
|
|
ep = container_of(_ep, struct goku_ep, ep);
|
|
|
|
if (!_ep)
|
|
|
|
return;
|
|
|
|
dma_free_coherent(&ep->dev->pdev->dev, bytes, buf, dma);
|
|
|
|
} else
|
|
|
|
#endif
|
|
|
|
kfree (buf);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*-------------------------------------------------------------------------*/
|
|
|
|
|
|
|
|
static void
|
|
|
|
done(struct goku_ep *ep, struct goku_request *req, int status)
|
|
|
|
{
|
|
|
|
struct goku_udc *dev;
|
|
|
|
unsigned stopped = ep->stopped;
|
|
|
|
|
|
|
|
list_del_init(&req->queue);
|
|
|
|
|
|
|
|
if (likely(req->req.status == -EINPROGRESS))
|
|
|
|
req->req.status = status;
|
|
|
|
else
|
|
|
|
status = req->req.status;
|
|
|
|
|
|
|
|
dev = ep->dev;
|
|
|
|
if (req->mapped) {
|
|
|
|
pci_unmap_single(dev->pdev, req->req.dma, req->req.length,
|
|
|
|
ep->is_in ? PCI_DMA_TODEVICE : PCI_DMA_FROMDEVICE);
|
|
|
|
req->req.dma = DMA_ADDR_INVALID;
|
|
|
|
req->mapped = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
#ifndef USB_TRACE
|
|
|
|
if (status && status != -ESHUTDOWN)
|
|
|
|
#endif
|
|
|
|
VDBG(dev, "complete %s req %p stat %d len %u/%u\n",
|
|
|
|
ep->ep.name, &req->req, status,
|
|
|
|
req->req.actual, req->req.length);
|
|
|
|
|
|
|
|
/* don't modify queue heads during completion callback */
|
|
|
|
ep->stopped = 1;
|
|
|
|
spin_unlock(&dev->lock);
|
|
|
|
req->req.complete(&ep->ep, &req->req);
|
|
|
|
spin_lock(&dev->lock);
|
|
|
|
ep->stopped = stopped;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*-------------------------------------------------------------------------*/
|
|
|
|
|
|
|
|
static inline int
|
|
|
|
write_packet(u32 __iomem *fifo, u8 *buf, struct goku_request *req, unsigned max)
|
|
|
|
{
|
|
|
|
unsigned length, count;
|
|
|
|
|
|
|
|
length = min(req->req.length - req->req.actual, max);
|
|
|
|
req->req.actual += length;
|
|
|
|
|
|
|
|
count = length;
|
|
|
|
while (likely(count--))
|
|
|
|
writel(*buf++, fifo);
|
|
|
|
return length;
|
|
|
|
}
|
|
|
|
|
|
|
|
// return: 0 = still running, 1 = completed, negative = errno
|
|
|
|
static int write_fifo(struct goku_ep *ep, struct goku_request *req)
|
|
|
|
{
|
|
|
|
struct goku_udc *dev = ep->dev;
|
|
|
|
u32 tmp;
|
|
|
|
u8 *buf;
|
|
|
|
unsigned count;
|
|
|
|
int is_last;
|
|
|
|
|
|
|
|
tmp = readl(&dev->regs->DataSet);
|
|
|
|
buf = req->req.buf + req->req.actual;
|
|
|
|
prefetch(buf);
|
|
|
|
|
|
|
|
dev = ep->dev;
|
|
|
|
if (unlikely(ep->num == 0 && dev->ep0state != EP0_IN))
|
|
|
|
return -EL2HLT;
|
|
|
|
|
|
|
|
/* NOTE: just single-buffered PIO-IN for now. */
|
|
|
|
if (unlikely((tmp & DATASET_A(ep->num)) != 0))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
/* clear our "packet available" irq */
|
|
|
|
if (ep->num != 0)
|
|
|
|
writel(~INT_EPxDATASET(ep->num), &dev->regs->int_status);
|
|
|
|
|
|
|
|
count = write_packet(ep->reg_fifo, buf, req, ep->ep.maxpacket);
|
|
|
|
|
|
|
|
/* last packet often short (sometimes a zlp, especially on ep0) */
|
|
|
|
if (unlikely(count != ep->ep.maxpacket)) {
|
|
|
|
writel(~(1<<ep->num), &dev->regs->EOP);
|
|
|
|
if (ep->num == 0) {
|
|
|
|
dev->ep[0].stopped = 1;
|
|
|
|
dev->ep0state = EP0_STATUS;
|
|
|
|
}
|
|
|
|
is_last = 1;
|
|
|
|
} else {
|
|
|
|
if (likely(req->req.length != req->req.actual)
|
|
|
|
|| req->req.zero)
|
|
|
|
is_last = 0;
|
|
|
|
else
|
|
|
|
is_last = 1;
|
|
|
|
}
|
|
|
|
#if 0 /* printk seemed to trash is_last...*/
|
|
|
|
//#ifdef USB_TRACE
|
|
|
|
VDBG(dev, "wrote %s %u bytes%s IN %u left %p\n",
|
|
|
|
ep->ep.name, count, is_last ? "/last" : "",
|
|
|
|
req->req.length - req->req.actual, req);
|
|
|
|
#endif
|
|
|
|
|
|
|
|
/* requests complete when all IN data is in the FIFO,
|
|
|
|
* or sometimes later, if a zlp was needed.
|
|
|
|
*/
|
|
|
|
if (is_last) {
|
|
|
|
done(ep, req, 0);
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int read_fifo(struct goku_ep *ep, struct goku_request *req)
|
|
|
|
{
|
|
|
|
struct goku_udc_regs __iomem *regs;
|
|
|
|
u32 size, set;
|
|
|
|
u8 *buf;
|
|
|
|
unsigned bufferspace, is_short, dbuff;
|
|
|
|
|
|
|
|
regs = ep->dev->regs;
|
|
|
|
top:
|
|
|
|
buf = req->req.buf + req->req.actual;
|
|
|
|
prefetchw(buf);
|
|
|
|
|
|
|
|
if (unlikely(ep->num == 0 && ep->dev->ep0state != EP0_OUT))
|
|
|
|
return -EL2HLT;
|
|
|
|
|
|
|
|
dbuff = (ep->num == 1 || ep->num == 2);
|
|
|
|
do {
|
|
|
|
/* ack dataset irq matching the status we'll handle */
|
|
|
|
if (ep->num != 0)
|
|
|
|
writel(~INT_EPxDATASET(ep->num), ®s->int_status);
|
|
|
|
|
|
|
|
set = readl(®s->DataSet) & DATASET_AB(ep->num);
|
|
|
|
size = readl(®s->EPxSizeLA[ep->num]);
|
|
|
|
bufferspace = req->req.length - req->req.actual;
|
|
|
|
|
|
|
|
/* usually do nothing without an OUT packet */
|
|
|
|
if (likely(ep->num != 0 || bufferspace != 0)) {
|
|
|
|
if (unlikely(set == 0))
|
|
|
|
break;
|
|
|
|
/* use ep1/ep2 double-buffering for OUT */
|
|
|
|
if (!(size & PACKET_ACTIVE))
|
|
|
|
size = readl(®s->EPxSizeLB[ep->num]);
|
|
|
|
if (!(size & PACKET_ACTIVE)) // "can't happen"
|
|
|
|
break;
|
|
|
|
size &= DATASIZE; /* EPxSizeH == 0 */
|
|
|
|
|
|
|
|
/* ep0out no-out-data case for set_config, etc */
|
|
|
|
} else
|
|
|
|
size = 0;
|
|
|
|
|
|
|
|
/* read all bytes from this packet */
|
|
|
|
req->req.actual += size;
|
|
|
|
is_short = (size < ep->ep.maxpacket);
|
|
|
|
#ifdef USB_TRACE
|
|
|
|
VDBG(ep->dev, "read %s %u bytes%s OUT req %p %u/%u\n",
|
|
|
|
ep->ep.name, size, is_short ? "/S" : "",
|
|
|
|
req, req->req.actual, req->req.length);
|
|
|
|
#endif
|
|
|
|
while (likely(size-- != 0)) {
|
|
|
|
u8 byte = (u8) readl(ep->reg_fifo);
|
|
|
|
|
|
|
|
if (unlikely(bufferspace == 0)) {
|
|
|
|
/* this happens when the driver's buffer
|
|
|
|
* is smaller than what the host sent.
|
|
|
|
* discard the extra data in this packet.
|
|
|
|
*/
|
|
|
|
if (req->req.status != -EOVERFLOW)
|
|
|
|
DBG(ep->dev, "%s overflow %u\n",
|
|
|
|
ep->ep.name, size);
|
|
|
|
req->req.status = -EOVERFLOW;
|
|
|
|
} else {
|
|
|
|
*buf++ = byte;
|
|
|
|
bufferspace--;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* completion */
|
|
|
|
if (unlikely(is_short || req->req.actual == req->req.length)) {
|
|
|
|
if (unlikely(ep->num == 0)) {
|
|
|
|
/* non-control endpoints now usable? */
|
|
|
|
if (ep->dev->req_config)
|
|
|
|
writel(ep->dev->configured
|
|
|
|
? USBSTATE_CONFIGURED
|
|
|
|
: 0,
|
|
|
|
®s->UsbState);
|
|
|
|
/* ep0out status stage */
|
|
|
|
writel(~(1<<0), ®s->EOP);
|
|
|
|
ep->stopped = 1;
|
|
|
|
ep->dev->ep0state = EP0_STATUS;
|
|
|
|
}
|
|
|
|
done(ep, req, 0);
|
|
|
|
|
|
|
|
/* empty the second buffer asap */
|
|
|
|
if (dbuff && !list_empty(&ep->queue)) {
|
|
|
|
req = list_entry(ep->queue.next,
|
|
|
|
struct goku_request, queue);
|
|
|
|
goto top;
|
|
|
|
}
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
} while (dbuff);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void
|
|
|
|
pio_irq_enable(struct goku_udc *dev,
|
|
|
|
struct goku_udc_regs __iomem *regs, int epnum)
|
|
|
|
{
|
|
|
|
dev->int_enable |= INT_EPxDATASET (epnum);
|
|
|
|
writel(dev->int_enable, ®s->int_enable);
|
|
|
|
/* write may still be posted */
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void
|
|
|
|
pio_irq_disable(struct goku_udc *dev,
|
|
|
|
struct goku_udc_regs __iomem *regs, int epnum)
|
|
|
|
{
|
|
|
|
dev->int_enable &= ~INT_EPxDATASET (epnum);
|
|
|
|
writel(dev->int_enable, ®s->int_enable);
|
|
|
|
/* write may still be posted */
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void
|
|
|
|
pio_advance(struct goku_ep *ep)
|
|
|
|
{
|
|
|
|
struct goku_request *req;
|
|
|
|
|
|
|
|
if (unlikely(list_empty (&ep->queue)))
|
|
|
|
return;
|
|
|
|
req = list_entry(ep->queue.next, struct goku_request, queue);
|
|
|
|
(ep->is_in ? write_fifo : read_fifo)(ep, req);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*-------------------------------------------------------------------------*/
|
|
|
|
|
|
|
|
// return: 0 = q running, 1 = q stopped, negative = errno
|
|
|
|
static int start_dma(struct goku_ep *ep, struct goku_request *req)
|
|
|
|
{
|
|
|
|
struct goku_udc_regs __iomem *regs = ep->dev->regs;
|
|
|
|
u32 master;
|
|
|
|
u32 start = req->req.dma;
|
|
|
|
u32 end = start + req->req.length - 1;
|
|
|
|
|
|
|
|
master = readl(®s->dma_master) & MST_RW_BITS;
|
|
|
|
|
|
|
|
/* re-init the bits affecting IN dma; careful with zlps */
|
|
|
|
if (likely(ep->is_in)) {
|
|
|
|
if (unlikely(master & MST_RD_ENA)) {
|
|
|
|
DBG (ep->dev, "start, IN active dma %03x!!\n",
|
|
|
|
master);
|
|
|
|
// return -EL2HLT;
|
|
|
|
}
|
|
|
|
writel(end, ®s->in_dma_end);
|
|
|
|
writel(start, ®s->in_dma_start);
|
|
|
|
|
|
|
|
master &= ~MST_R_BITS;
|
|
|
|
if (unlikely(req->req.length == 0))
|
|
|
|
master = MST_RD_ENA | MST_RD_EOPB;
|
|
|
|
else if ((req->req.length % ep->ep.maxpacket) != 0
|
|
|
|
|| req->req.zero)
|
|
|
|
master = MST_RD_ENA | MST_EOPB_ENA;
|
|
|
|
else
|
|
|
|
master = MST_RD_ENA | MST_EOPB_DIS;
|
|
|
|
|
|
|
|
ep->dev->int_enable |= INT_MSTRDEND;
|
|
|
|
|
|
|
|
/* Goku DMA-OUT merges short packets, which plays poorly with
|
|
|
|
* protocols where short packets mark the transfer boundaries.
|
|
|
|
* The chip supports a nonstandard policy with INT_MSTWRTMOUT,
|
|
|
|
* ending transfers after 3 SOFs; we don't turn it on.
|
|
|
|
*/
|
|
|
|
} else {
|
|
|
|
if (unlikely(master & MST_WR_ENA)) {
|
|
|
|
DBG (ep->dev, "start, OUT active dma %03x!!\n",
|
|
|
|
master);
|
|
|
|
// return -EL2HLT;
|
|
|
|
}
|
|
|
|
writel(end, ®s->out_dma_end);
|
|
|
|
writel(start, ®s->out_dma_start);
|
|
|
|
|
|
|
|
master &= ~MST_W_BITS;
|
|
|
|
master |= MST_WR_ENA | MST_TIMEOUT_DIS;
|
|
|
|
|
|
|
|
ep->dev->int_enable |= INT_MSTWREND|INT_MSTWRTMOUT;
|
|
|
|
}
|
|
|
|
|
|
|
|
writel(master, ®s->dma_master);
|
|
|
|
writel(ep->dev->int_enable, ®s->int_enable);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void dma_advance(struct goku_udc *dev, struct goku_ep *ep)
|
|
|
|
{
|
|
|
|
struct goku_request *req;
|
|
|
|
struct goku_udc_regs __iomem *regs = ep->dev->regs;
|
|
|
|
u32 master;
|
|
|
|
|
|
|
|
master = readl(®s->dma_master);
|
|
|
|
|
|
|
|
if (unlikely(list_empty(&ep->queue))) {
|
|
|
|
stop:
|
|
|
|
if (ep->is_in)
|
|
|
|
dev->int_enable &= ~INT_MSTRDEND;
|
|
|
|
else
|
|
|
|
dev->int_enable &= ~(INT_MSTWREND|INT_MSTWRTMOUT);
|
|
|
|
writel(dev->int_enable, ®s->int_enable);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
req = list_entry(ep->queue.next, struct goku_request, queue);
|
|
|
|
|
|
|
|
/* normal hw dma completion (not abort) */
|
|
|
|
if (likely(ep->is_in)) {
|
|
|
|
if (unlikely(master & MST_RD_ENA))
|
|
|
|
return;
|
|
|
|
req->req.actual = readl(®s->in_dma_current);
|
|
|
|
} else {
|
|
|
|
if (unlikely(master & MST_WR_ENA))
|
|
|
|
return;
|
|
|
|
|
|
|
|
/* hardware merges short packets, and also hides packet
|
|
|
|
* overruns. a partial packet MAY be in the fifo here.
|
|
|
|
*/
|
|
|
|
req->req.actual = readl(®s->out_dma_current);
|
|
|
|
}
|
|
|
|
req->req.actual -= req->req.dma;
|
|
|
|
req->req.actual++;
|
|
|
|
|
|
|
|
#ifdef USB_TRACE
|
|
|
|
VDBG(dev, "done %s %s dma, %u/%u bytes, req %p\n",
|
|
|
|
ep->ep.name, ep->is_in ? "IN" : "OUT",
|
|
|
|
req->req.actual, req->req.length, req);
|
|
|
|
#endif
|
|
|
|
done(ep, req, 0);
|
|
|
|
if (list_empty(&ep->queue))
|
|
|
|
goto stop;
|
|
|
|
req = list_entry(ep->queue.next, struct goku_request, queue);
|
|
|
|
(void) start_dma(ep, req);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void abort_dma(struct goku_ep *ep, int status)
|
|
|
|
{
|
|
|
|
struct goku_udc_regs __iomem *regs = ep->dev->regs;
|
|
|
|
struct goku_request *req;
|
|
|
|
u32 curr, master;
|
|
|
|
|
|
|
|
/* NAK future host requests, hoping the implicit delay lets the
|
|
|
|
* dma engine finish reading (or writing) its latest packet and
|
|
|
|
* empty the dma buffer (up to 16 bytes).
|
|
|
|
*
|
|
|
|
* This avoids needing to clean up a partial packet in the fifo;
|
|
|
|
* we can't do that for IN without side effects to HALT and TOGGLE.
|
|
|
|
*/
|
|
|
|
command(regs, COMMAND_FIFO_DISABLE, ep->num);
|
|
|
|
req = list_entry(ep->queue.next, struct goku_request, queue);
|
|
|
|
master = readl(®s->dma_master) & MST_RW_BITS;
|
|
|
|
|
|
|
|
/* FIXME using these resets isn't usably documented. this may
|
|
|
|
* not work unless it's followed by disabling the endpoint.
|
|
|
|
*
|
|
|
|
* FIXME the OUT reset path doesn't even behave consistently.
|
|
|
|
*/
|
|
|
|
if (ep->is_in) {
|
|
|
|
if (unlikely((readl(®s->dma_master) & MST_RD_ENA) == 0))
|
|
|
|
goto finished;
|
|
|
|
curr = readl(®s->in_dma_current);
|
|
|
|
|
|
|
|
writel(curr, ®s->in_dma_end);
|
|
|
|
writel(curr, ®s->in_dma_start);
|
|
|
|
|
|
|
|
master &= ~MST_R_BITS;
|
|
|
|
master |= MST_RD_RESET;
|
|
|
|
writel(master, ®s->dma_master);
|
|
|
|
|
|
|
|
if (readl(®s->dma_master) & MST_RD_ENA)
|
|
|
|
DBG(ep->dev, "IN dma active after reset!\n");
|
|
|
|
|
|
|
|
} else {
|
|
|
|
if (unlikely((readl(®s->dma_master) & MST_WR_ENA) == 0))
|
|
|
|
goto finished;
|
|
|
|
curr = readl(®s->out_dma_current);
|
|
|
|
|
|
|
|
writel(curr, ®s->out_dma_end);
|
|
|
|
writel(curr, ®s->out_dma_start);
|
|
|
|
|
|
|
|
master &= ~MST_W_BITS;
|
|
|
|
master |= MST_WR_RESET;
|
|
|
|
writel(master, ®s->dma_master);
|
|
|
|
|
|
|
|
if (readl(®s->dma_master) & MST_WR_ENA)
|
|
|
|
DBG(ep->dev, "OUT dma active after reset!\n");
|
|
|
|
}
|
|
|
|
req->req.actual = (curr - req->req.dma) + 1;
|
|
|
|
req->req.status = status;
|
|
|
|
|
|
|
|
VDBG(ep->dev, "%s %s %s %d/%d\n", __FUNCTION__, ep->ep.name,
|
|
|
|
ep->is_in ? "IN" : "OUT",
|
|
|
|
req->req.actual, req->req.length);
|
|
|
|
|
|
|
|
command(regs, COMMAND_FIFO_ENABLE, ep->num);
|
|
|
|
|
|
|
|
return;
|
|
|
|
|
|
|
|
finished:
|
|
|
|
/* dma already completed; no abort needed */
|
|
|
|
command(regs, COMMAND_FIFO_ENABLE, ep->num);
|
|
|
|
req->req.actual = req->req.length;
|
|
|
|
req->req.status = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*-------------------------------------------------------------------------*/
|
|
|
|
|
|
|
|
static int
|
2005-10-21 01:21:58 -06:00
|
|
|
goku_queue(struct usb_ep *_ep, struct usb_request *_req, gfp_t gfp_flags)
|
2005-04-16 16:20:36 -06:00
|
|
|
{
|
|
|
|
struct goku_request *req;
|
|
|
|
struct goku_ep *ep;
|
|
|
|
struct goku_udc *dev;
|
|
|
|
unsigned long flags;
|
|
|
|
int status;
|
|
|
|
|
|
|
|
/* always require a cpu-view buffer so pio works */
|
|
|
|
req = container_of(_req, struct goku_request, req);
|
|
|
|
if (unlikely(!_req || !_req->complete
|
|
|
|
|| !_req->buf || !list_empty(&req->queue)))
|
|
|
|
return -EINVAL;
|
|
|
|
ep = container_of(_ep, struct goku_ep, ep);
|
|
|
|
if (unlikely(!_ep || (!ep->desc && ep->num != 0)))
|
|
|
|
return -EINVAL;
|
|
|
|
dev = ep->dev;
|
|
|
|
if (unlikely(!dev->driver || dev->gadget.speed == USB_SPEED_UNKNOWN))
|
|
|
|
return -ESHUTDOWN;
|
|
|
|
|
|
|
|
/* can't touch registers when suspended */
|
|
|
|
if (dev->ep0state == EP0_SUSPEND)
|
|
|
|
return -EBUSY;
|
|
|
|
|
|
|
|
/* set up dma mapping in case the caller didn't */
|
|
|
|
if (ep->dma && _req->dma == DMA_ADDR_INVALID) {
|
|
|
|
_req->dma = pci_map_single(dev->pdev, _req->buf, _req->length,
|
|
|
|
ep->is_in ? PCI_DMA_TODEVICE : PCI_DMA_FROMDEVICE);
|
|
|
|
req->mapped = 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
#ifdef USB_TRACE
|
|
|
|
VDBG(dev, "%s queue req %p, len %u buf %p\n",
|
|
|
|
_ep->name, _req, _req->length, _req->buf);
|
|
|
|
#endif
|
|
|
|
|
|
|
|
spin_lock_irqsave(&dev->lock, flags);
|
|
|
|
|
|
|
|
_req->status = -EINPROGRESS;
|
|
|
|
_req->actual = 0;
|
|
|
|
|
|
|
|
/* for ep0 IN without premature status, zlp is required and
|
|
|
|
* writing EOP starts the status stage (OUT).
|
|
|
|
*/
|
|
|
|
if (unlikely(ep->num == 0 && ep->is_in))
|
|
|
|
_req->zero = 1;
|
|
|
|
|
|
|
|
/* kickstart this i/o queue? */
|
|
|
|
status = 0;
|
|
|
|
if (list_empty(&ep->queue) && likely(!ep->stopped)) {
|
|
|
|
/* dma: done after dma completion IRQ (or error)
|
|
|
|
* pio: done after last fifo operation
|
|
|
|
*/
|
|
|
|
if (ep->dma)
|
|
|
|
status = start_dma(ep, req);
|
|
|
|
else
|
|
|
|
status = (ep->is_in ? write_fifo : read_fifo)(ep, req);
|
|
|
|
|
|
|
|
if (unlikely(status != 0)) {
|
|
|
|
if (status > 0)
|
|
|
|
status = 0;
|
|
|
|
req = NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
} /* else pio or dma irq handler advances the queue. */
|
|
|
|
|
|
|
|
if (likely(req != 0))
|
|
|
|
list_add_tail(&req->queue, &ep->queue);
|
|
|
|
|
|
|
|
if (likely(!list_empty(&ep->queue))
|
|
|
|
&& likely(ep->num != 0)
|
|
|
|
&& !ep->dma
|
|
|
|
&& !(dev->int_enable & INT_EPxDATASET (ep->num)))
|
|
|
|
pio_irq_enable(dev, dev->regs, ep->num);
|
|
|
|
|
|
|
|
spin_unlock_irqrestore(&dev->lock, flags);
|
|
|
|
|
|
|
|
/* pci writes may still be posted */
|
|
|
|
return status;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* dequeue ALL requests */
|
|
|
|
static void nuke(struct goku_ep *ep, int status)
|
|
|
|
{
|
|
|
|
struct goku_request *req;
|
|
|
|
|
|
|
|
ep->stopped = 1;
|
|
|
|
if (list_empty(&ep->queue))
|
|
|
|
return;
|
|
|
|
if (ep->dma)
|
|
|
|
abort_dma(ep, status);
|
|
|
|
while (!list_empty(&ep->queue)) {
|
|
|
|
req = list_entry(ep->queue.next, struct goku_request, queue);
|
|
|
|
done(ep, req, status);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* dequeue JUST ONE request */
|
|
|
|
static int goku_dequeue(struct usb_ep *_ep, struct usb_request *_req)
|
|
|
|
{
|
|
|
|
struct goku_request *req;
|
|
|
|
struct goku_ep *ep;
|
|
|
|
struct goku_udc *dev;
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
ep = container_of(_ep, struct goku_ep, ep);
|
|
|
|
if (!_ep || !_req || (!ep->desc && ep->num != 0))
|
|
|
|
return -EINVAL;
|
|
|
|
dev = ep->dev;
|
|
|
|
if (!dev->driver)
|
|
|
|
return -ESHUTDOWN;
|
|
|
|
|
|
|
|
/* we can't touch (dma) registers when suspended */
|
|
|
|
if (dev->ep0state == EP0_SUSPEND)
|
|
|
|
return -EBUSY;
|
|
|
|
|
|
|
|
VDBG(dev, "%s %s %s %s %p\n", __FUNCTION__, _ep->name,
|
|
|
|
ep->is_in ? "IN" : "OUT",
|
|
|
|
ep->dma ? "dma" : "pio",
|
|
|
|
_req);
|
|
|
|
|
|
|
|
spin_lock_irqsave(&dev->lock, flags);
|
|
|
|
|
|
|
|
/* make sure it's actually queued on this endpoint */
|
|
|
|
list_for_each_entry (req, &ep->queue, queue) {
|
|
|
|
if (&req->req == _req)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (&req->req != _req) {
|
|
|
|
spin_unlock_irqrestore (&dev->lock, flags);
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (ep->dma && ep->queue.next == &req->queue && !ep->stopped) {
|
|
|
|
abort_dma(ep, -ECONNRESET);
|
|
|
|
done(ep, req, -ECONNRESET);
|
|
|
|
dma_advance(dev, ep);
|
|
|
|
} else if (!list_empty(&req->queue))
|
|
|
|
done(ep, req, -ECONNRESET);
|
|
|
|
else
|
|
|
|
req = NULL;
|
|
|
|
spin_unlock_irqrestore(&dev->lock, flags);
|
|
|
|
|
|
|
|
return req ? 0 : -EOPNOTSUPP;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*-------------------------------------------------------------------------*/
|
|
|
|
|
|
|
|
static void goku_clear_halt(struct goku_ep *ep)
|
|
|
|
{
|
|
|
|
// assert (ep->num !=0)
|
|
|
|
VDBG(ep->dev, "%s clear halt\n", ep->ep.name);
|
|
|
|
command(ep->dev->regs, COMMAND_SETDATA0, ep->num);
|
|
|
|
command(ep->dev->regs, COMMAND_STALL_CLEAR, ep->num);
|
|
|
|
if (ep->stopped) {
|
|
|
|
ep->stopped = 0;
|
|
|
|
if (ep->dma) {
|
|
|
|
struct goku_request *req;
|
|
|
|
|
|
|
|
if (list_empty(&ep->queue))
|
|
|
|
return;
|
|
|
|
req = list_entry(ep->queue.next, struct goku_request,
|
|
|
|
queue);
|
|
|
|
(void) start_dma(ep, req);
|
|
|
|
} else
|
|
|
|
pio_advance(ep);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static int goku_set_halt(struct usb_ep *_ep, int value)
|
|
|
|
{
|
|
|
|
struct goku_ep *ep;
|
|
|
|
unsigned long flags;
|
|
|
|
int retval = 0;
|
|
|
|
|
|
|
|
if (!_ep)
|
|
|
|
return -ENODEV;
|
|
|
|
ep = container_of (_ep, struct goku_ep, ep);
|
|
|
|
|
|
|
|
if (ep->num == 0) {
|
|
|
|
if (value) {
|
|
|
|
ep->dev->ep0state = EP0_STALL;
|
|
|
|
ep->dev->ep[0].stopped = 1;
|
|
|
|
} else
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
/* don't change EPxSTATUS_EP_INVALID to READY */
|
|
|
|
} else if (!ep->desc) {
|
|
|
|
DBG(ep->dev, "%s %s inactive?\n", __FUNCTION__, ep->ep.name);
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
spin_lock_irqsave(&ep->dev->lock, flags);
|
|
|
|
if (!list_empty(&ep->queue))
|
|
|
|
retval = -EAGAIN;
|
|
|
|
else if (ep->is_in && value
|
|
|
|
/* data in (either) packet buffer? */
|
|
|
|
&& (readl(&ep->dev->regs->DataSet)
|
|
|
|
& DATASET_AB(ep->num)))
|
|
|
|
retval = -EAGAIN;
|
|
|
|
else if (!value)
|
|
|
|
goku_clear_halt(ep);
|
|
|
|
else {
|
|
|
|
ep->stopped = 1;
|
|
|
|
VDBG(ep->dev, "%s set halt\n", ep->ep.name);
|
|
|
|
command(ep->dev->regs, COMMAND_STALL, ep->num);
|
|
|
|
readl(ep->reg_status);
|
|
|
|
}
|
|
|
|
spin_unlock_irqrestore(&ep->dev->lock, flags);
|
|
|
|
return retval;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int goku_fifo_status(struct usb_ep *_ep)
|
|
|
|
{
|
|
|
|
struct goku_ep *ep;
|
|
|
|
struct goku_udc_regs __iomem *regs;
|
|
|
|
u32 size;
|
|
|
|
|
|
|
|
if (!_ep)
|
|
|
|
return -ENODEV;
|
|
|
|
ep = container_of(_ep, struct goku_ep, ep);
|
|
|
|
|
|
|
|
/* size is only reported sanely for OUT */
|
|
|
|
if (ep->is_in)
|
|
|
|
return -EOPNOTSUPP;
|
|
|
|
|
|
|
|
/* ignores 16-byte dma buffer; SizeH == 0 */
|
|
|
|
regs = ep->dev->regs;
|
|
|
|
size = readl(®s->EPxSizeLA[ep->num]) & DATASIZE;
|
|
|
|
size += readl(®s->EPxSizeLB[ep->num]) & DATASIZE;
|
|
|
|
VDBG(ep->dev, "%s %s %u\n", __FUNCTION__, ep->ep.name, size);
|
|
|
|
return size;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void goku_fifo_flush(struct usb_ep *_ep)
|
|
|
|
{
|
|
|
|
struct goku_ep *ep;
|
|
|
|
struct goku_udc_regs __iomem *regs;
|
|
|
|
u32 size;
|
|
|
|
|
|
|
|
if (!_ep)
|
|
|
|
return;
|
|
|
|
ep = container_of(_ep, struct goku_ep, ep);
|
|
|
|
VDBG(ep->dev, "%s %s\n", __FUNCTION__, ep->ep.name);
|
|
|
|
|
|
|
|
/* don't change EPxSTATUS_EP_INVALID to READY */
|
|
|
|
if (!ep->desc && ep->num != 0) {
|
|
|
|
DBG(ep->dev, "%s %s inactive?\n", __FUNCTION__, ep->ep.name);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
regs = ep->dev->regs;
|
|
|
|
size = readl(®s->EPxSizeLA[ep->num]);
|
|
|
|
size &= DATASIZE;
|
|
|
|
|
|
|
|
/* Non-desirable behavior: FIFO_CLEAR also clears the
|
|
|
|
* endpoint halt feature. For OUT, we _could_ just read
|
|
|
|
* the bytes out (PIO, if !ep->dma); for in, no choice.
|
|
|
|
*/
|
|
|
|
if (size)
|
|
|
|
command(regs, COMMAND_FIFO_CLEAR, ep->num);
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct usb_ep_ops goku_ep_ops = {
|
|
|
|
.enable = goku_ep_enable,
|
|
|
|
.disable = goku_ep_disable,
|
|
|
|
|
|
|
|
.alloc_request = goku_alloc_request,
|
|
|
|
.free_request = goku_free_request,
|
|
|
|
|
|
|
|
.alloc_buffer = goku_alloc_buffer,
|
|
|
|
.free_buffer = goku_free_buffer,
|
|
|
|
|
|
|
|
.queue = goku_queue,
|
|
|
|
.dequeue = goku_dequeue,
|
|
|
|
|
|
|
|
.set_halt = goku_set_halt,
|
|
|
|
.fifo_status = goku_fifo_status,
|
|
|
|
.fifo_flush = goku_fifo_flush,
|
|
|
|
};
|
|
|
|
|
|
|
|
/*-------------------------------------------------------------------------*/
|
|
|
|
|
|
|
|
static int goku_get_frame(struct usb_gadget *_gadget)
|
|
|
|
{
|
|
|
|
return -EOPNOTSUPP;
|
|
|
|
}
|
|
|
|
|
|
|
|
static const struct usb_gadget_ops goku_ops = {
|
|
|
|
.get_frame = goku_get_frame,
|
|
|
|
// no remote wakeup
|
|
|
|
// not selfpowered
|
|
|
|
};
|
|
|
|
|
|
|
|
/*-------------------------------------------------------------------------*/
|
|
|
|
|
|
|
|
static inline char *dmastr(void)
|
|
|
|
{
|
|
|
|
if (use_dma == 0)
|
|
|
|
return "(dma disabled)";
|
|
|
|
else if (use_dma == 2)
|
|
|
|
return "(dma IN and OUT)";
|
|
|
|
else
|
|
|
|
return "(dma IN)";
|
|
|
|
}
|
|
|
|
|
|
|
|
#ifdef CONFIG_USB_GADGET_DEBUG_FILES
|
|
|
|
|
|
|
|
static const char proc_node_name [] = "driver/udc";
|
|
|
|
|
|
|
|
#define FOURBITS "%s%s%s%s"
|
|
|
|
#define EIGHTBITS FOURBITS FOURBITS
|
|
|
|
|
|
|
|
static void
|
|
|
|
dump_intmask(const char *label, u32 mask, char **next, unsigned *size)
|
|
|
|
{
|
|
|
|
int t;
|
|
|
|
|
|
|
|
/* int_status is the same format ... */
|
|
|
|
t = scnprintf(*next, *size,
|
|
|
|
"%s %05X =" FOURBITS EIGHTBITS EIGHTBITS "\n",
|
|
|
|
label, mask,
|
|
|
|
(mask & INT_PWRDETECT) ? " power" : "",
|
|
|
|
(mask & INT_SYSERROR) ? " sys" : "",
|
|
|
|
(mask & INT_MSTRDEND) ? " in-dma" : "",
|
|
|
|
(mask & INT_MSTWRTMOUT) ? " wrtmo" : "",
|
|
|
|
|
|
|
|
(mask & INT_MSTWREND) ? " out-dma" : "",
|
|
|
|
(mask & INT_MSTWRSET) ? " wrset" : "",
|
|
|
|
(mask & INT_ERR) ? " err" : "",
|
|
|
|
(mask & INT_SOF) ? " sof" : "",
|
|
|
|
|
|
|
|
(mask & INT_EP3NAK) ? " ep3nak" : "",
|
|
|
|
(mask & INT_EP2NAK) ? " ep2nak" : "",
|
|
|
|
(mask & INT_EP1NAK) ? " ep1nak" : "",
|
|
|
|
(mask & INT_EP3DATASET) ? " ep3" : "",
|
|
|
|
|
|
|
|
(mask & INT_EP2DATASET) ? " ep2" : "",
|
|
|
|
(mask & INT_EP1DATASET) ? " ep1" : "",
|
|
|
|
(mask & INT_STATUSNAK) ? " ep0snak" : "",
|
|
|
|
(mask & INT_STATUS) ? " ep0status" : "",
|
|
|
|
|
|
|
|
(mask & INT_SETUP) ? " setup" : "",
|
|
|
|
(mask & INT_ENDPOINT0) ? " ep0" : "",
|
|
|
|
(mask & INT_USBRESET) ? " reset" : "",
|
|
|
|
(mask & INT_SUSPEND) ? " suspend" : "");
|
|
|
|
*size -= t;
|
|
|
|
*next += t;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
static int
|
|
|
|
udc_proc_read(char *buffer, char **start, off_t off, int count,
|
|
|
|
int *eof, void *_dev)
|
|
|
|
{
|
|
|
|
char *buf = buffer;
|
|
|
|
struct goku_udc *dev = _dev;
|
|
|
|
struct goku_udc_regs __iomem *regs = dev->regs;
|
|
|
|
char *next = buf;
|
|
|
|
unsigned size = count;
|
|
|
|
unsigned long flags;
|
|
|
|
int i, t, is_usb_connected;
|
|
|
|
u32 tmp;
|
|
|
|
|
|
|
|
if (off != 0)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
local_irq_save(flags);
|
|
|
|
|
|
|
|
/* basic device status */
|
|
|
|
tmp = readl(®s->power_detect);
|
|
|
|
is_usb_connected = tmp & PW_DETECT;
|
|
|
|
t = scnprintf(next, size,
|
|
|
|
"%s - %s\n"
|
|
|
|
"%s version: %s %s\n"
|
|
|
|
"Gadget driver: %s\n"
|
|
|
|
"Host %s, %s\n"
|
|
|
|
"\n",
|
|
|
|
pci_name(dev->pdev), driver_desc,
|
|
|
|
driver_name, DRIVER_VERSION, dmastr(),
|
|
|
|
dev->driver ? dev->driver->driver.name : "(none)",
|
|
|
|
is_usb_connected
|
|
|
|
? ((tmp & PW_PULLUP) ? "full speed" : "powered")
|
|
|
|
: "disconnected",
|
|
|
|
({char *tmp;
|
|
|
|
switch(dev->ep0state){
|
|
|
|
case EP0_DISCONNECT: tmp = "ep0_disconnect"; break;
|
|
|
|
case EP0_IDLE: tmp = "ep0_idle"; break;
|
|
|
|
case EP0_IN: tmp = "ep0_in"; break;
|
|
|
|
case EP0_OUT: tmp = "ep0_out"; break;
|
|
|
|
case EP0_STATUS: tmp = "ep0_status"; break;
|
|
|
|
case EP0_STALL: tmp = "ep0_stall"; break;
|
|
|
|
case EP0_SUSPEND: tmp = "ep0_suspend"; break;
|
|
|
|
default: tmp = "ep0_?"; break;
|
|
|
|
} tmp; })
|
|
|
|
);
|
|
|
|
size -= t;
|
|
|
|
next += t;
|
|
|
|
|
|
|
|
dump_intmask("int_status", readl(®s->int_status), &next, &size);
|
|
|
|
dump_intmask("int_enable", readl(®s->int_enable), &next, &size);
|
|
|
|
|
|
|
|
if (!is_usb_connected || !dev->driver || (tmp & PW_PULLUP) == 0)
|
|
|
|
goto done;
|
|
|
|
|
|
|
|
/* registers for (active) device and ep0 */
|
|
|
|
t = scnprintf(next, size, "\nirqs %lu\ndataset %02x "
|
|
|
|
"single.bcs %02x.%02x state %x addr %u\n",
|
|
|
|
dev->irqs, readl(®s->DataSet),
|
|
|
|
readl(®s->EPxSingle), readl(®s->EPxBCS),
|
|
|
|
readl(®s->UsbState),
|
|
|
|
readl(®s->address));
|
|
|
|
size -= t;
|
|
|
|
next += t;
|
|
|
|
|
|
|
|
tmp = readl(®s->dma_master);
|
|
|
|
t = scnprintf(next, size,
|
|
|
|
"dma %03X =" EIGHTBITS "%s %s\n", tmp,
|
|
|
|
(tmp & MST_EOPB_DIS) ? " eopb-" : "",
|
|
|
|
(tmp & MST_EOPB_ENA) ? " eopb+" : "",
|
|
|
|
(tmp & MST_TIMEOUT_DIS) ? " tmo-" : "",
|
|
|
|
(tmp & MST_TIMEOUT_ENA) ? " tmo+" : "",
|
|
|
|
|
|
|
|
(tmp & MST_RD_EOPB) ? " eopb" : "",
|
|
|
|
(tmp & MST_RD_RESET) ? " in_reset" : "",
|
|
|
|
(tmp & MST_WR_RESET) ? " out_reset" : "",
|
|
|
|
(tmp & MST_RD_ENA) ? " IN" : "",
|
|
|
|
|
|
|
|
(tmp & MST_WR_ENA) ? " OUT" : "",
|
|
|
|
(tmp & MST_CONNECTION)
|
|
|
|
? "ep1in/ep2out"
|
|
|
|
: "ep1out/ep2in");
|
|
|
|
size -= t;
|
|
|
|
next += t;
|
|
|
|
|
|
|
|
/* dump endpoint queues */
|
|
|
|
for (i = 0; i < 4; i++) {
|
|
|
|
struct goku_ep *ep = &dev->ep [i];
|
|
|
|
struct goku_request *req;
|
|
|
|
int t;
|
|
|
|
|
|
|
|
if (i && !ep->desc)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
tmp = readl(ep->reg_status);
|
|
|
|
t = scnprintf(next, size,
|
|
|
|
"%s %s max %u %s, irqs %lu, "
|
|
|
|
"status %02x (%s) " FOURBITS "\n",
|
|
|
|
ep->ep.name,
|
|
|
|
ep->is_in ? "in" : "out",
|
|
|
|
ep->ep.maxpacket,
|
|
|
|
ep->dma ? "dma" : "pio",
|
|
|
|
ep->irqs,
|
|
|
|
tmp, ({ char *s;
|
|
|
|
switch (tmp & EPxSTATUS_EP_MASK) {
|
|
|
|
case EPxSTATUS_EP_READY:
|
|
|
|
s = "ready"; break;
|
|
|
|
case EPxSTATUS_EP_DATAIN:
|
|
|
|
s = "packet"; break;
|
|
|
|
case EPxSTATUS_EP_FULL:
|
|
|
|
s = "full"; break;
|
|
|
|
case EPxSTATUS_EP_TX_ERR: // host will retry
|
|
|
|
s = "tx_err"; break;
|
|
|
|
case EPxSTATUS_EP_RX_ERR:
|
|
|
|
s = "rx_err"; break;
|
|
|
|
case EPxSTATUS_EP_BUSY: /* ep0 only */
|
|
|
|
s = "busy"; break;
|
|
|
|
case EPxSTATUS_EP_STALL:
|
|
|
|
s = "stall"; break;
|
|
|
|
case EPxSTATUS_EP_INVALID: // these "can't happen"
|
|
|
|
s = "invalid"; break;
|
|
|
|
default:
|
|
|
|
s = "?"; break;
|
|
|
|
}; s; }),
|
|
|
|
(tmp & EPxSTATUS_TOGGLE) ? "data1" : "data0",
|
|
|
|
(tmp & EPxSTATUS_SUSPEND) ? " suspend" : "",
|
|
|
|
(tmp & EPxSTATUS_FIFO_DISABLE) ? " disable" : "",
|
|
|
|
(tmp & EPxSTATUS_STAGE_ERROR) ? " ep0stat" : ""
|
|
|
|
);
|
|
|
|
if (t <= 0 || t > size)
|
|
|
|
goto done;
|
|
|
|
size -= t;
|
|
|
|
next += t;
|
|
|
|
|
|
|
|
if (list_empty(&ep->queue)) {
|
|
|
|
t = scnprintf(next, size, "\t(nothing queued)\n");
|
|
|
|
if (t <= 0 || t > size)
|
|
|
|
goto done;
|
|
|
|
size -= t;
|
|
|
|
next += t;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
list_for_each_entry(req, &ep->queue, queue) {
|
|
|
|
if (ep->dma && req->queue.prev == &ep->queue) {
|
|
|
|
if (i == UDC_MSTRD_ENDPOINT)
|
|
|
|
tmp = readl(®s->in_dma_current);
|
|
|
|
else
|
|
|
|
tmp = readl(®s->out_dma_current);
|
|
|
|
tmp -= req->req.dma;
|
|
|
|
tmp++;
|
|
|
|
} else
|
|
|
|
tmp = req->req.actual;
|
|
|
|
|
|
|
|
t = scnprintf(next, size,
|
|
|
|
"\treq %p len %u/%u buf %p\n",
|
|
|
|
&req->req, tmp, req->req.length,
|
|
|
|
req->req.buf);
|
|
|
|
if (t <= 0 || t > size)
|
|
|
|
goto done;
|
|
|
|
size -= t;
|
|
|
|
next += t;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
done:
|
|
|
|
local_irq_restore(flags);
|
|
|
|
*eof = 1;
|
|
|
|
return count - size;
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif /* CONFIG_USB_GADGET_DEBUG_FILES */
|
|
|
|
|
|
|
|
/*-------------------------------------------------------------------------*/
|
|
|
|
|
|
|
|
static void udc_reinit (struct goku_udc *dev)
|
|
|
|
{
|
|
|
|
static char *names [] = { "ep0", "ep1-bulk", "ep2-bulk", "ep3-bulk" };
|
|
|
|
|
|
|
|
unsigned i;
|
|
|
|
|
|
|
|
INIT_LIST_HEAD (&dev->gadget.ep_list);
|
|
|
|
dev->gadget.ep0 = &dev->ep [0].ep;
|
|
|
|
dev->gadget.speed = USB_SPEED_UNKNOWN;
|
|
|
|
dev->ep0state = EP0_DISCONNECT;
|
|
|
|
dev->irqs = 0;
|
|
|
|
|
|
|
|
for (i = 0; i < 4; i++) {
|
|
|
|
struct goku_ep *ep = &dev->ep[i];
|
|
|
|
|
|
|
|
ep->num = i;
|
|
|
|
ep->ep.name = names[i];
|
|
|
|
ep->reg_fifo = &dev->regs->ep_fifo [i];
|
|
|
|
ep->reg_status = &dev->regs->ep_status [i];
|
|
|
|
ep->reg_mode = &dev->regs->ep_mode[i];
|
|
|
|
|
|
|
|
ep->ep.ops = &goku_ep_ops;
|
|
|
|
list_add_tail (&ep->ep.ep_list, &dev->gadget.ep_list);
|
|
|
|
ep->dev = dev;
|
|
|
|
INIT_LIST_HEAD (&ep->queue);
|
|
|
|
|
|
|
|
ep_reset(NULL, ep);
|
|
|
|
}
|
|
|
|
|
|
|
|
dev->ep[0].reg_mode = NULL;
|
|
|
|
dev->ep[0].ep.maxpacket = MAX_EP0_SIZE;
|
|
|
|
list_del_init (&dev->ep[0].ep.ep_list);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void udc_reset(struct goku_udc *dev)
|
|
|
|
{
|
|
|
|
struct goku_udc_regs __iomem *regs = dev->regs;
|
|
|
|
|
|
|
|
writel(0, ®s->power_detect);
|
|
|
|
writel(0, ®s->int_enable);
|
|
|
|
readl(®s->int_enable);
|
|
|
|
dev->int_enable = 0;
|
|
|
|
|
|
|
|
/* deassert reset, leave USB D+ at hi-Z (no pullup)
|
|
|
|
* don't let INT_PWRDETECT sequence begin
|
|
|
|
*/
|
|
|
|
udelay(250);
|
|
|
|
writel(PW_RESETB, ®s->power_detect);
|
|
|
|
readl(®s->int_enable);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void ep0_start(struct goku_udc *dev)
|
|
|
|
{
|
|
|
|
struct goku_udc_regs __iomem *regs = dev->regs;
|
|
|
|
unsigned i;
|
|
|
|
|
|
|
|
VDBG(dev, "%s\n", __FUNCTION__);
|
|
|
|
|
|
|
|
udc_reset(dev);
|
|
|
|
udc_reinit (dev);
|
|
|
|
//writel(MST_EOPB_ENA | MST_TIMEOUT_ENA, ®s->dma_master);
|
|
|
|
|
|
|
|
/* hw handles set_address, set_feature, get_status; maybe more */
|
|
|
|
writel( G_REQMODE_SET_INTF | G_REQMODE_GET_INTF
|
|
|
|
| G_REQMODE_SET_CONF | G_REQMODE_GET_CONF
|
|
|
|
| G_REQMODE_GET_DESC
|
|
|
|
| G_REQMODE_CLEAR_FEAT
|
|
|
|
, ®s->reqmode);
|
|
|
|
|
|
|
|
for (i = 0; i < 4; i++)
|
|
|
|
dev->ep[i].irqs = 0;
|
|
|
|
|
|
|
|
/* can't modify descriptors after writing UsbReady */
|
|
|
|
for (i = 0; i < DESC_LEN; i++)
|
|
|
|
writel(0, ®s->descriptors[i]);
|
|
|
|
writel(0, ®s->UsbReady);
|
|
|
|
|
|
|
|
/* expect ep0 requests when the host drops reset */
|
|
|
|
writel(PW_RESETB | PW_PULLUP, ®s->power_detect);
|
|
|
|
dev->int_enable = INT_DEVWIDE | INT_EP0;
|
|
|
|
writel(dev->int_enable, &dev->regs->int_enable);
|
|
|
|
readl(®s->int_enable);
|
|
|
|
dev->gadget.speed = USB_SPEED_FULL;
|
|
|
|
dev->ep0state = EP0_IDLE;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void udc_enable(struct goku_udc *dev)
|
|
|
|
{
|
|
|
|
/* start enumeration now, or after power detect irq */
|
|
|
|
if (readl(&dev->regs->power_detect) & PW_DETECT)
|
|
|
|
ep0_start(dev);
|
|
|
|
else {
|
|
|
|
DBG(dev, "%s\n", __FUNCTION__);
|
|
|
|
dev->int_enable = INT_PWRDETECT;
|
|
|
|
writel(dev->int_enable, &dev->regs->int_enable);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/*-------------------------------------------------------------------------*/
|
|
|
|
|
|
|
|
/* keeping it simple:
|
|
|
|
* - one bus driver, initted first;
|
|
|
|
* - one function driver, initted second
|
|
|
|
*/
|
|
|
|
|
|
|
|
static struct goku_udc *the_controller;
|
|
|
|
|
|
|
|
/* when a driver is successfully registered, it will receive
|
|
|
|
* control requests including set_configuration(), which enables
|
|
|
|
* non-control requests. then usb traffic follows until a
|
|
|
|
* disconnect is reported. then a host may connect again, or
|
|
|
|
* the driver might get unbound.
|
|
|
|
*/
|
|
|
|
int usb_gadget_register_driver(struct usb_gadget_driver *driver)
|
|
|
|
{
|
|
|
|
struct goku_udc *dev = the_controller;
|
|
|
|
int retval;
|
|
|
|
|
|
|
|
if (!driver
|
|
|
|
|| driver->speed != USB_SPEED_FULL
|
|
|
|
|| !driver->bind
|
|
|
|
|| !driver->unbind
|
|
|
|
|| !driver->disconnect
|
|
|
|
|| !driver->setup)
|
|
|
|
return -EINVAL;
|
|
|
|
if (!dev)
|
|
|
|
return -ENODEV;
|
|
|
|
if (dev->driver)
|
|
|
|
return -EBUSY;
|
|
|
|
|
|
|
|
/* hook up the driver */
|
|
|
|
driver->driver.bus = NULL;
|
|
|
|
dev->driver = driver;
|
|
|
|
dev->gadget.dev.driver = &driver->driver;
|
|
|
|
retval = driver->bind(&dev->gadget);
|
|
|
|
if (retval) {
|
|
|
|
DBG(dev, "bind to driver %s --> error %d\n",
|
|
|
|
driver->driver.name, retval);
|
|
|
|
dev->driver = NULL;
|
|
|
|
dev->gadget.dev.driver = NULL;
|
|
|
|
return retval;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* then enable host detection and ep0; and we're ready
|
|
|
|
* for set_configuration as well as eventual disconnect.
|
|
|
|
*/
|
|
|
|
udc_enable(dev);
|
|
|
|
|
|
|
|
DBG(dev, "registered gadget driver '%s'\n", driver->driver.name);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(usb_gadget_register_driver);
|
|
|
|
|
|
|
|
static void
|
|
|
|
stop_activity(struct goku_udc *dev, struct usb_gadget_driver *driver)
|
|
|
|
{
|
|
|
|
unsigned i;
|
|
|
|
|
|
|
|
DBG (dev, "%s\n", __FUNCTION__);
|
|
|
|
|
|
|
|
if (dev->gadget.speed == USB_SPEED_UNKNOWN)
|
|
|
|
driver = NULL;
|
|
|
|
|
|
|
|
/* disconnect gadget driver after quiesceing hw and the driver */
|
|
|
|
udc_reset (dev);
|
|
|
|
for (i = 0; i < 4; i++)
|
|
|
|
nuke(&dev->ep [i], -ESHUTDOWN);
|
|
|
|
if (driver) {
|
|
|
|
spin_unlock(&dev->lock);
|
|
|
|
driver->disconnect(&dev->gadget);
|
|
|
|
spin_lock(&dev->lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (dev->driver)
|
|
|
|
udc_enable(dev);
|
|
|
|
}
|
|
|
|
|
|
|
|
int usb_gadget_unregister_driver(struct usb_gadget_driver *driver)
|
|
|
|
{
|
|
|
|
struct goku_udc *dev = the_controller;
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
if (!dev)
|
|
|
|
return -ENODEV;
|
|
|
|
if (!driver || driver != dev->driver)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&dev->lock, flags);
|
|
|
|
dev->driver = NULL;
|
|
|
|
stop_activity(dev, driver);
|
|
|
|
spin_unlock_irqrestore(&dev->lock, flags);
|
|
|
|
|
|
|
|
driver->unbind(&dev->gadget);
|
|
|
|
|
|
|
|
DBG(dev, "unregistered driver '%s'\n", driver->driver.name);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(usb_gadget_unregister_driver);
|
|
|
|
|
|
|
|
|
|
|
|
/*-------------------------------------------------------------------------*/
|
|
|
|
|
|
|
|
static void ep0_setup(struct goku_udc *dev)
|
|
|
|
{
|
|
|
|
struct goku_udc_regs __iomem *regs = dev->regs;
|
|
|
|
struct usb_ctrlrequest ctrl;
|
|
|
|
int tmp;
|
|
|
|
|
|
|
|
/* read SETUP packet and enter DATA stage */
|
|
|
|
ctrl.bRequestType = readl(®s->bRequestType);
|
|
|
|
ctrl.bRequest = readl(®s->bRequest);
|
2005-05-07 14:05:52 -06:00
|
|
|
ctrl.wValue = cpu_to_le16((readl(®s->wValueH) << 8)
|
|
|
|
| readl(®s->wValueL));
|
|
|
|
ctrl.wIndex = cpu_to_le16((readl(®s->wIndexH) << 8)
|
|
|
|
| readl(®s->wIndexL));
|
|
|
|
ctrl.wLength = cpu_to_le16((readl(®s->wLengthH) << 8)
|
|
|
|
| readl(®s->wLengthL));
|
2005-04-16 16:20:36 -06:00
|
|
|
writel(0, ®s->SetupRecv);
|
|
|
|
|
|
|
|
nuke(&dev->ep[0], 0);
|
|
|
|
dev->ep[0].stopped = 0;
|
|
|
|
if (likely(ctrl.bRequestType & USB_DIR_IN)) {
|
|
|
|
dev->ep[0].is_in = 1;
|
|
|
|
dev->ep0state = EP0_IN;
|
|
|
|
/* detect early status stages */
|
|
|
|
writel(ICONTROL_STATUSNAK, &dev->regs->IntControl);
|
|
|
|
} else {
|
|
|
|
dev->ep[0].is_in = 0;
|
|
|
|
dev->ep0state = EP0_OUT;
|
|
|
|
|
|
|
|
/* NOTE: CLEAR_FEATURE is done in software so that we can
|
|
|
|
* synchronize transfer restarts after bulk IN stalls. data
|
|
|
|
* won't even enter the fifo until the halt is cleared.
|
|
|
|
*/
|
|
|
|
switch (ctrl.bRequest) {
|
|
|
|
case USB_REQ_CLEAR_FEATURE:
|
|
|
|
switch (ctrl.bRequestType) {
|
|
|
|
case USB_RECIP_ENDPOINT:
|
2005-05-07 14:05:52 -06:00
|
|
|
tmp = le16_to_cpu(ctrl.wIndex) & 0x0f;
|
2005-04-16 16:20:36 -06:00
|
|
|
/* active endpoint */
|
|
|
|
if (tmp > 3 || (!dev->ep[tmp].desc && tmp != 0))
|
|
|
|
goto stall;
|
2005-05-07 14:05:52 -06:00
|
|
|
if (ctrl.wIndex & __constant_cpu_to_le16(
|
|
|
|
USB_DIR_IN)) {
|
2005-04-16 16:20:36 -06:00
|
|
|
if (!dev->ep[tmp].is_in)
|
|
|
|
goto stall;
|
|
|
|
} else {
|
|
|
|
if (dev->ep[tmp].is_in)
|
|
|
|
goto stall;
|
|
|
|
}
|
2005-05-07 14:05:52 -06:00
|
|
|
if (ctrl.wValue != __constant_cpu_to_le16(
|
|
|
|
USB_ENDPOINT_HALT))
|
2005-04-16 16:20:36 -06:00
|
|
|
goto stall;
|
|
|
|
if (tmp)
|
|
|
|
goku_clear_halt(&dev->ep[tmp]);
|
|
|
|
succeed:
|
|
|
|
/* start ep0out status stage */
|
|
|
|
writel(~(1<<0), ®s->EOP);
|
|
|
|
dev->ep[0].stopped = 1;
|
|
|
|
dev->ep0state = EP0_STATUS;
|
|
|
|
return;
|
|
|
|
case USB_RECIP_DEVICE:
|
|
|
|
/* device remote wakeup: always clear */
|
2005-05-07 14:05:52 -06:00
|
|
|
if (ctrl.wValue != __constant_cpu_to_le16(1))
|
2005-04-16 16:20:36 -06:00
|
|
|
goto stall;
|
|
|
|
VDBG(dev, "clear dev remote wakeup\n");
|
|
|
|
goto succeed;
|
|
|
|
case USB_RECIP_INTERFACE:
|
|
|
|
goto stall;
|
|
|
|
default: /* pass to gadget driver */
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
#ifdef USB_TRACE
|
|
|
|
VDBG(dev, "SETUP %02x.%02x v%04x i%04x l%04x\n",
|
|
|
|
ctrl.bRequestType, ctrl.bRequest,
|
2005-05-07 14:05:52 -06:00
|
|
|
le16_to_cpu(ctrl.wValue), le16_to_cpu(ctrl.wIndex),
|
|
|
|
le16_to_cpu(ctrl.wLength));
|
2005-04-16 16:20:36 -06:00
|
|
|
#endif
|
|
|
|
|
|
|
|
/* hw wants to know when we're configured (or not) */
|
|
|
|
dev->req_config = (ctrl.bRequest == USB_REQ_SET_CONFIGURATION
|
|
|
|
&& ctrl.bRequestType == USB_RECIP_DEVICE);
|
|
|
|
if (unlikely(dev->req_config))
|
2005-05-07 14:05:52 -06:00
|
|
|
dev->configured = (ctrl.wValue != __constant_cpu_to_le16(0));
|
2005-04-16 16:20:36 -06:00
|
|
|
|
|
|
|
/* delegate everything to the gadget driver.
|
|
|
|
* it may respond after this irq handler returns.
|
|
|
|
*/
|
|
|
|
spin_unlock (&dev->lock);
|
|
|
|
tmp = dev->driver->setup(&dev->gadget, &ctrl);
|
|
|
|
spin_lock (&dev->lock);
|
|
|
|
if (unlikely(tmp < 0)) {
|
|
|
|
stall:
|
|
|
|
#ifdef USB_TRACE
|
|
|
|
VDBG(dev, "req %02x.%02x protocol STALL; err %d\n",
|
|
|
|
ctrl.bRequestType, ctrl.bRequest, tmp);
|
|
|
|
#endif
|
|
|
|
command(regs, COMMAND_STALL, 0);
|
|
|
|
dev->ep[0].stopped = 1;
|
|
|
|
dev->ep0state = EP0_STALL;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* expect at least one data or status stage irq */
|
|
|
|
}
|
|
|
|
|
|
|
|
#define ACK(irqbit) { \
|
|
|
|
stat &= ~irqbit; \
|
|
|
|
writel(~irqbit, ®s->int_status); \
|
|
|
|
handled = 1; \
|
|
|
|
}
|
|
|
|
|
IRQ: Maintain regs pointer globally rather than passing to IRQ handlers
Maintain a per-CPU global "struct pt_regs *" variable which can be used instead
of passing regs around manually through all ~1800 interrupt handlers in the
Linux kernel.
The regs pointer is used in few places, but it potentially costs both stack
space and code to pass it around. On the FRV arch, removing the regs parameter
from all the genirq function results in a 20% speed up of the IRQ exit path
(ie: from leaving timer_interrupt() to leaving do_IRQ()).
Where appropriate, an arch may override the generic storage facility and do
something different with the variable. On FRV, for instance, the address is
maintained in GR28 at all times inside the kernel as part of general exception
handling.
Having looked over the code, it appears that the parameter may be handed down
through up to twenty or so layers of functions. Consider a USB character
device attached to a USB hub, attached to a USB controller that posts its
interrupts through a cascaded auxiliary interrupt controller. A character
device driver may want to pass regs to the sysrq handler through the input
layer which adds another few layers of parameter passing.
I've build this code with allyesconfig for x86_64 and i386. I've runtested the
main part of the code on FRV and i386, though I can't test most of the drivers.
I've also done partial conversion for powerpc and MIPS - these at least compile
with minimal configurations.
This will affect all archs. Mostly the changes should be relatively easy.
Take do_IRQ(), store the regs pointer at the beginning, saving the old one:
struct pt_regs *old_regs = set_irq_regs(regs);
And put the old one back at the end:
set_irq_regs(old_regs);
Don't pass regs through to generic_handle_irq() or __do_IRQ().
In timer_interrupt(), this sort of change will be necessary:
- update_process_times(user_mode(regs));
- profile_tick(CPU_PROFILING, regs);
+ update_process_times(user_mode(get_irq_regs()));
+ profile_tick(CPU_PROFILING);
I'd like to move update_process_times()'s use of get_irq_regs() into itself,
except that i386, alone of the archs, uses something other than user_mode().
Some notes on the interrupt handling in the drivers:
(*) input_dev() is now gone entirely. The regs pointer is no longer stored in
the input_dev struct.
(*) finish_unlinks() in drivers/usb/host/ohci-q.c needs checking. It does
something different depending on whether it's been supplied with a regs
pointer or not.
(*) Various IRQ handler function pointers have been moved to type
irq_handler_t.
Signed-Off-By: David Howells <dhowells@redhat.com>
(cherry picked from 1b16e7ac850969f38b375e511e3fa2f474a33867 commit)
2006-10-05 07:55:46 -06:00
|
|
|
static irqreturn_t goku_irq(int irq, void *_dev)
|
2005-04-16 16:20:36 -06:00
|
|
|
{
|
|
|
|
struct goku_udc *dev = _dev;
|
|
|
|
struct goku_udc_regs __iomem *regs = dev->regs;
|
|
|
|
struct goku_ep *ep;
|
|
|
|
u32 stat, handled = 0;
|
|
|
|
unsigned i, rescans = 5;
|
|
|
|
|
|
|
|
spin_lock(&dev->lock);
|
|
|
|
|
|
|
|
rescan:
|
|
|
|
stat = readl(®s->int_status) & dev->int_enable;
|
|
|
|
if (!stat)
|
|
|
|
goto done;
|
|
|
|
dev->irqs++;
|
|
|
|
|
|
|
|
/* device-wide irqs */
|
|
|
|
if (unlikely(stat & INT_DEVWIDE)) {
|
|
|
|
if (stat & INT_SYSERROR) {
|
|
|
|
ERROR(dev, "system error\n");
|
|
|
|
stop_activity(dev, dev->driver);
|
|
|
|
stat = 0;
|
|
|
|
handled = 1;
|
|
|
|
// FIXME have a neater way to prevent re-enumeration
|
|
|
|
dev->driver = NULL;
|
|
|
|
goto done;
|
|
|
|
}
|
|
|
|
if (stat & INT_PWRDETECT) {
|
|
|
|
writel(~stat, ®s->int_status);
|
|
|
|
if (readl(&dev->regs->power_detect) & PW_DETECT) {
|
|
|
|
VDBG(dev, "connect\n");
|
|
|
|
ep0_start(dev);
|
|
|
|
} else {
|
|
|
|
DBG(dev, "disconnect\n");
|
|
|
|
if (dev->gadget.speed == USB_SPEED_FULL)
|
|
|
|
stop_activity(dev, dev->driver);
|
|
|
|
dev->ep0state = EP0_DISCONNECT;
|
|
|
|
dev->int_enable = INT_DEVWIDE;
|
|
|
|
writel(dev->int_enable, &dev->regs->int_enable);
|
|
|
|
}
|
|
|
|
stat = 0;
|
|
|
|
handled = 1;
|
|
|
|
goto done;
|
|
|
|
}
|
|
|
|
if (stat & INT_SUSPEND) {
|
|
|
|
ACK(INT_SUSPEND);
|
|
|
|
if (readl(®s->ep_status[0]) & EPxSTATUS_SUSPEND) {
|
|
|
|
switch (dev->ep0state) {
|
|
|
|
case EP0_DISCONNECT:
|
|
|
|
case EP0_SUSPEND:
|
|
|
|
goto pm_next;
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
DBG(dev, "USB suspend\n");
|
|
|
|
dev->ep0state = EP0_SUSPEND;
|
|
|
|
if (dev->gadget.speed != USB_SPEED_UNKNOWN
|
|
|
|
&& dev->driver
|
|
|
|
&& dev->driver->suspend) {
|
|
|
|
spin_unlock(&dev->lock);
|
|
|
|
dev->driver->suspend(&dev->gadget);
|
|
|
|
spin_lock(&dev->lock);
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
if (dev->ep0state != EP0_SUSPEND) {
|
|
|
|
DBG(dev, "bogus USB resume %d\n",
|
|
|
|
dev->ep0state);
|
|
|
|
goto pm_next;
|
|
|
|
}
|
|
|
|
DBG(dev, "USB resume\n");
|
|
|
|
dev->ep0state = EP0_IDLE;
|
|
|
|
if (dev->gadget.speed != USB_SPEED_UNKNOWN
|
|
|
|
&& dev->driver
|
|
|
|
&& dev->driver->resume) {
|
|
|
|
spin_unlock(&dev->lock);
|
|
|
|
dev->driver->resume(&dev->gadget);
|
|
|
|
spin_lock(&dev->lock);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
pm_next:
|
|
|
|
if (stat & INT_USBRESET) { /* hub reset done */
|
|
|
|
ACK(INT_USBRESET);
|
|
|
|
INFO(dev, "USB reset done, gadget %s\n",
|
|
|
|
dev->driver->driver.name);
|
|
|
|
}
|
|
|
|
// and INT_ERR on some endpoint's crc/bitstuff/... problem
|
|
|
|
}
|
|
|
|
|
|
|
|
/* progress ep0 setup, data, or status stages.
|
|
|
|
* no transition {EP0_STATUS, EP0_STALL} --> EP0_IDLE; saves irqs
|
|
|
|
*/
|
|
|
|
if (stat & INT_SETUP) {
|
|
|
|
ACK(INT_SETUP);
|
|
|
|
dev->ep[0].irqs++;
|
|
|
|
ep0_setup(dev);
|
|
|
|
}
|
|
|
|
if (stat & INT_STATUSNAK) {
|
|
|
|
ACK(INT_STATUSNAK|INT_ENDPOINT0);
|
|
|
|
if (dev->ep0state == EP0_IN) {
|
|
|
|
ep = &dev->ep[0];
|
|
|
|
ep->irqs++;
|
|
|
|
nuke(ep, 0);
|
|
|
|
writel(~(1<<0), ®s->EOP);
|
|
|
|
dev->ep0state = EP0_STATUS;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (stat & INT_ENDPOINT0) {
|
|
|
|
ACK(INT_ENDPOINT0);
|
|
|
|
ep = &dev->ep[0];
|
|
|
|
ep->irqs++;
|
|
|
|
pio_advance(ep);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* dma completion */
|
|
|
|
if (stat & INT_MSTRDEND) { /* IN */
|
|
|
|
ACK(INT_MSTRDEND);
|
|
|
|
ep = &dev->ep[UDC_MSTRD_ENDPOINT];
|
|
|
|
ep->irqs++;
|
|
|
|
dma_advance(dev, ep);
|
|
|
|
}
|
|
|
|
if (stat & INT_MSTWREND) { /* OUT */
|
|
|
|
ACK(INT_MSTWREND);
|
|
|
|
ep = &dev->ep[UDC_MSTWR_ENDPOINT];
|
|
|
|
ep->irqs++;
|
|
|
|
dma_advance(dev, ep);
|
|
|
|
}
|
|
|
|
if (stat & INT_MSTWRTMOUT) { /* OUT */
|
|
|
|
ACK(INT_MSTWRTMOUT);
|
|
|
|
ep = &dev->ep[UDC_MSTWR_ENDPOINT];
|
|
|
|
ep->irqs++;
|
|
|
|
ERROR(dev, "%s write timeout ?\n", ep->ep.name);
|
|
|
|
// reset dma? then dma_advance()
|
|
|
|
}
|
|
|
|
|
|
|
|
/* pio */
|
|
|
|
for (i = 1; i < 4; i++) {
|
|
|
|
u32 tmp = INT_EPxDATASET(i);
|
|
|
|
|
|
|
|
if (!(stat & tmp))
|
|
|
|
continue;
|
|
|
|
ep = &dev->ep[i];
|
|
|
|
pio_advance(ep);
|
|
|
|
if (list_empty (&ep->queue))
|
|
|
|
pio_irq_disable(dev, regs, i);
|
|
|
|
stat &= ~tmp;
|
|
|
|
handled = 1;
|
|
|
|
ep->irqs++;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (rescans--)
|
|
|
|
goto rescan;
|
|
|
|
|
|
|
|
done:
|
|
|
|
(void)readl(®s->int_enable);
|
|
|
|
spin_unlock(&dev->lock);
|
|
|
|
if (stat)
|
|
|
|
DBG(dev, "unhandled irq status: %05x (%05x, %05x)\n", stat,
|
|
|
|
readl(®s->int_status), dev->int_enable);
|
|
|
|
return IRQ_RETVAL(handled);
|
|
|
|
}
|
|
|
|
|
|
|
|
#undef ACK
|
|
|
|
|
|
|
|
/*-------------------------------------------------------------------------*/
|
|
|
|
|
|
|
|
static void gadget_release(struct device *_dev)
|
|
|
|
{
|
|
|
|
struct goku_udc *dev = dev_get_drvdata(_dev);
|
|
|
|
|
|
|
|
kfree(dev);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* tear down the binding between this driver and the pci device */
|
|
|
|
|
|
|
|
static void goku_remove(struct pci_dev *pdev)
|
|
|
|
{
|
|
|
|
struct goku_udc *dev = pci_get_drvdata(pdev);
|
|
|
|
|
|
|
|
DBG(dev, "%s\n", __FUNCTION__);
|
|
|
|
/* start with the driver above us */
|
|
|
|
if (dev->driver) {
|
|
|
|
/* should have been done already by driver model core */
|
|
|
|
WARN(dev, "pci remove, driver '%s' is still registered\n",
|
|
|
|
dev->driver->driver.name);
|
|
|
|
usb_gadget_unregister_driver(dev->driver);
|
|
|
|
}
|
|
|
|
|
|
|
|
#ifdef CONFIG_USB_GADGET_DEBUG_FILES
|
|
|
|
remove_proc_entry(proc_node_name, NULL);
|
|
|
|
#endif
|
|
|
|
if (dev->regs)
|
|
|
|
udc_reset(dev);
|
|
|
|
if (dev->got_irq)
|
|
|
|
free_irq(pdev->irq, dev);
|
|
|
|
if (dev->regs)
|
|
|
|
iounmap(dev->regs);
|
|
|
|
if (dev->got_region)
|
|
|
|
release_mem_region(pci_resource_start (pdev, 0),
|
|
|
|
pci_resource_len (pdev, 0));
|
|
|
|
if (dev->enabled)
|
|
|
|
pci_disable_device(pdev);
|
|
|
|
device_unregister(&dev->gadget.dev);
|
|
|
|
|
|
|
|
pci_set_drvdata(pdev, NULL);
|
|
|
|
dev->regs = NULL;
|
|
|
|
the_controller = NULL;
|
|
|
|
|
|
|
|
INFO(dev, "unbind\n");
|
|
|
|
}
|
|
|
|
|
|
|
|
/* wrap this driver around the specified pci device, but
|
|
|
|
* don't respond over USB until a gadget driver binds to us.
|
|
|
|
*/
|
|
|
|
|
|
|
|
static int goku_probe(struct pci_dev *pdev, const struct pci_device_id *id)
|
|
|
|
{
|
|
|
|
struct goku_udc *dev = NULL;
|
|
|
|
unsigned long resource, len;
|
|
|
|
void __iomem *base = NULL;
|
|
|
|
int retval;
|
|
|
|
|
|
|
|
/* if you want to support more than one controller in a system,
|
|
|
|
* usb_gadget_driver_{register,unregister}() must change.
|
|
|
|
*/
|
|
|
|
if (the_controller) {
|
|
|
|
WARN(dev, "ignoring %s\n", pci_name(pdev));
|
|
|
|
return -EBUSY;
|
|
|
|
}
|
|
|
|
if (!pdev->irq) {
|
|
|
|
printk(KERN_ERR "Check PCI %s IRQ setup!\n", pci_name(pdev));
|
|
|
|
retval = -ENODEV;
|
|
|
|
goto done;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* alloc, and start init */
|
2006-12-06 21:33:17 -07:00
|
|
|
dev = kmalloc (sizeof *dev, GFP_KERNEL);
|
2005-04-16 16:20:36 -06:00
|
|
|
if (dev == NULL){
|
|
|
|
pr_debug("enomem %s\n", pci_name(pdev));
|
|
|
|
retval = -ENOMEM;
|
|
|
|
goto done;
|
|
|
|
}
|
|
|
|
|
|
|
|
memset(dev, 0, sizeof *dev);
|
|
|
|
spin_lock_init(&dev->lock);
|
|
|
|
dev->pdev = pdev;
|
|
|
|
dev->gadget.ops = &goku_ops;
|
|
|
|
|
|
|
|
/* the "gadget" abstracts/virtualizes the controller */
|
|
|
|
strcpy(dev->gadget.dev.bus_id, "gadget");
|
|
|
|
dev->gadget.dev.parent = &pdev->dev;
|
|
|
|
dev->gadget.dev.dma_mask = pdev->dev.dma_mask;
|
|
|
|
dev->gadget.dev.release = gadget_release;
|
|
|
|
dev->gadget.name = driver_name;
|
|
|
|
|
|
|
|
/* now all the pci goodies ... */
|
|
|
|
retval = pci_enable_device(pdev);
|
|
|
|
if (retval < 0) {
|
|
|
|
DBG(dev, "can't enable, %d\n", retval);
|
|
|
|
goto done;
|
|
|
|
}
|
|
|
|
dev->enabled = 1;
|
|
|
|
|
|
|
|
resource = pci_resource_start(pdev, 0);
|
|
|
|
len = pci_resource_len(pdev, 0);
|
|
|
|
if (!request_mem_region(resource, len, driver_name)) {
|
|
|
|
DBG(dev, "controller already in use\n");
|
|
|
|
retval = -EBUSY;
|
|
|
|
goto done;
|
|
|
|
}
|
|
|
|
dev->got_region = 1;
|
|
|
|
|
|
|
|
base = ioremap_nocache(resource, len);
|
|
|
|
if (base == NULL) {
|
|
|
|
DBG(dev, "can't map memory\n");
|
|
|
|
retval = -EFAULT;
|
|
|
|
goto done;
|
|
|
|
}
|
|
|
|
dev->regs = (struct goku_udc_regs __iomem *) base;
|
|
|
|
|
|
|
|
pci_set_drvdata(pdev, dev);
|
|
|
|
INFO(dev, "%s\n", driver_desc);
|
|
|
|
INFO(dev, "version: " DRIVER_VERSION " %s\n", dmastr());
|
2006-06-20 02:21:29 -06:00
|
|
|
INFO(dev, "irq %d, pci mem %p\n", pdev->irq, base);
|
2005-04-16 16:20:36 -06:00
|
|
|
|
|
|
|
/* init to known state, then setup irqs */
|
|
|
|
udc_reset(dev);
|
|
|
|
udc_reinit (dev);
|
2006-07-01 20:29:44 -06:00
|
|
|
if (request_irq(pdev->irq, goku_irq, IRQF_SHARED/*|IRQF_SAMPLE_RANDOM*/,
|
2005-04-16 16:20:36 -06:00
|
|
|
driver_name, dev) != 0) {
|
2006-06-20 02:21:29 -06:00
|
|
|
DBG(dev, "request interrupt %d failed\n", pdev->irq);
|
2005-04-16 16:20:36 -06:00
|
|
|
retval = -EBUSY;
|
|
|
|
goto done;
|
|
|
|
}
|
|
|
|
dev->got_irq = 1;
|
|
|
|
if (use_dma)
|
|
|
|
pci_set_master(pdev);
|
|
|
|
|
|
|
|
|
|
|
|
#ifdef CONFIG_USB_GADGET_DEBUG_FILES
|
|
|
|
create_proc_read_entry(proc_node_name, 0, NULL, udc_proc_read, dev);
|
|
|
|
#endif
|
|
|
|
|
|
|
|
/* done */
|
|
|
|
the_controller = dev;
|
|
|
|
device_register(&dev->gadget.dev);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
done:
|
|
|
|
if (dev)
|
|
|
|
goku_remove (pdev);
|
|
|
|
return retval;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*-------------------------------------------------------------------------*/
|
|
|
|
|
|
|
|
static struct pci_device_id pci_ids [] = { {
|
|
|
|
.class = ((PCI_CLASS_SERIAL_USB << 8) | 0xfe),
|
|
|
|
.class_mask = ~0,
|
|
|
|
.vendor = 0x102f, /* Toshiba */
|
|
|
|
.device = 0x0107, /* this UDC */
|
|
|
|
.subvendor = PCI_ANY_ID,
|
|
|
|
.subdevice = PCI_ANY_ID,
|
|
|
|
|
|
|
|
}, { /* end: all zeroes */ }
|
|
|
|
};
|
|
|
|
MODULE_DEVICE_TABLE (pci, pci_ids);
|
|
|
|
|
|
|
|
static struct pci_driver goku_pci_driver = {
|
|
|
|
.name = (char *) driver_name,
|
|
|
|
.id_table = pci_ids,
|
|
|
|
|
|
|
|
.probe = goku_probe,
|
|
|
|
.remove = goku_remove,
|
|
|
|
|
|
|
|
/* FIXME add power management support */
|
|
|
|
};
|
|
|
|
|
|
|
|
static int __init init (void)
|
|
|
|
{
|
|
|
|
return pci_register_driver (&goku_pci_driver);
|
|
|
|
}
|
|
|
|
module_init (init);
|
|
|
|
|
|
|
|
static void __exit cleanup (void)
|
|
|
|
{
|
|
|
|
pci_unregister_driver (&goku_pci_driver);
|
|
|
|
}
|
|
|
|
module_exit (cleanup);
|