linux/arch/mips/dec/ioasic-irq.c
<<
>>
Prefs
   1// SPDX-License-Identifier: GPL-2.0-or-later
   2/*
   3 *      DEC I/O ASIC interrupts.
   4 *
   5 *      Copyright (c) 2002, 2003, 2013  Maciej W. Rozycki
   6 */
   7
   8#include <linux/init.h>
   9#include <linux/irq.h>
  10#include <linux/types.h>
  11
  12#include <asm/dec/ioasic.h>
  13#include <asm/dec/ioasic_addrs.h>
  14#include <asm/dec/ioasic_ints.h>
  15
  16static int ioasic_irq_base;
  17
  18static void unmask_ioasic_irq(struct irq_data *d)
  19{
  20        u32 simr;
  21
  22        simr = ioasic_read(IO_REG_SIMR);
  23        simr |= (1 << (d->irq - ioasic_irq_base));
  24        ioasic_write(IO_REG_SIMR, simr);
  25}
  26
  27static void mask_ioasic_irq(struct irq_data *d)
  28{
  29        u32 simr;
  30
  31        simr = ioasic_read(IO_REG_SIMR);
  32        simr &= ~(1 << (d->irq - ioasic_irq_base));
  33        ioasic_write(IO_REG_SIMR, simr);
  34}
  35
  36static void ack_ioasic_irq(struct irq_data *d)
  37{
  38        mask_ioasic_irq(d);
  39        fast_iob();
  40}
  41
  42static struct irq_chip ioasic_irq_type = {
  43        .name = "IO-ASIC",
  44        .irq_ack = ack_ioasic_irq,
  45        .irq_mask = mask_ioasic_irq,
  46        .irq_mask_ack = ack_ioasic_irq,
  47        .irq_unmask = unmask_ioasic_irq,
  48};
  49
  50static void clear_ioasic_dma_irq(struct irq_data *d)
  51{
  52        u32 sir;
  53
  54        sir = ~(1 << (d->irq - ioasic_irq_base));
  55        ioasic_write(IO_REG_SIR, sir);
  56        fast_iob();
  57}
  58
  59static struct irq_chip ioasic_dma_irq_type = {
  60        .name = "IO-ASIC-DMA",
  61        .irq_ack = clear_ioasic_dma_irq,
  62        .irq_mask = mask_ioasic_irq,
  63        .irq_unmask = unmask_ioasic_irq,
  64        .irq_eoi = clear_ioasic_dma_irq,
  65};
  66
  67/*
  68 * I/O ASIC implements two kinds of DMA interrupts, informational and
  69 * error interrupts.
  70 *
  71 * The formers do not stop DMA and should be cleared as soon as possible
  72 * so that if they retrigger before the handler has completed, usually as
  73 * a side effect of actions taken by the handler, then they are reissued.
  74 * These use the `handle_edge_irq' handler that clears the request right
  75 * away.
  76 *
  77 * The latters stop DMA and do not resume it until the interrupt has been
  78 * cleared.  This cannot be done until after a corrective action has been
  79 * taken and this also means they will not retrigger.  Therefore they use
  80 * the `handle_fasteoi_irq' handler that only clears the request on the
  81 * way out.  Because MIPS processor interrupt inputs, one of which the I/O
  82 * ASIC is cascaded to, are level-triggered it is recommended that error
  83 * DMA interrupt action handlers are registered with the IRQF_ONESHOT flag
  84 * set so that they are run with the interrupt line masked.
  85 *
  86 * This mask has `1' bits in the positions of informational interrupts.
  87 */
  88#define IO_IRQ_DMA_INFO                                                 \
  89        (IO_IRQ_MASK(IO_INR_SCC0A_RXDMA) |                              \
  90         IO_IRQ_MASK(IO_INR_SCC1A_RXDMA) |                              \
  91         IO_IRQ_MASK(IO_INR_ISDN_TXDMA) |                               \
  92         IO_IRQ_MASK(IO_INR_ISDN_RXDMA) |                               \
  93         IO_IRQ_MASK(IO_INR_ASC_DMA))
  94
  95void __init init_ioasic_irqs(int base)
  96{
  97        int i;
  98
  99        /* Mask interrupts. */
 100        ioasic_write(IO_REG_SIMR, 0);
 101        fast_iob();
 102
 103        for (i = base; i < base + IO_INR_DMA; i++)
 104                irq_set_chip_and_handler(i, &ioasic_irq_type,
 105                                         handle_level_irq);
 106        for (; i < base + IO_IRQ_LINES; i++)
 107                irq_set_chip_and_handler(i, &ioasic_dma_irq_type,
 108                                         1 << (i - base) & IO_IRQ_DMA_INFO ?
 109                                         handle_edge_irq : handle_fasteoi_irq);
 110
 111        ioasic_irq_base = base;
 112}
 113