|
|
|
|
/*
|
|
|
|
|
* linux/arch/alpha/kernel/sys_titan.c
|
|
|
|
|
*
|
|
|
|
|
* Copyright (C) 1995 David A Rusling
|
|
|
|
|
* Copyright (C) 1996, 1999 Jay A Estabrook
|
|
|
|
|
* Copyright (C) 1998, 1999 Richard Henderson
|
|
|
|
|
* Copyright (C) 1999, 2000 Jeff Wiedemeier
|
|
|
|
|
*
|
|
|
|
|
* Code supporting TITAN systems (EV6+TITAN), currently:
|
|
|
|
|
* Privateer
|
|
|
|
|
* Falcon
|
|
|
|
|
* Granite
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
|
#include <linux/kernel.h>
|
|
|
|
|
#include <linux/types.h>
|
|
|
|
|
#include <linux/mm.h>
|
|
|
|
|
#include <linux/sched.h>
|
|
|
|
|
#include <linux/pci.h>
|
|
|
|
|
#include <linux/init.h>
|
|
|
|
|
#include <linux/bitops.h>
|
|
|
|
|
|
|
|
|
|
#include <asm/ptrace.h>
|
|
|
|
|
#include <asm/system.h>
|
|
|
|
|
#include <asm/dma.h>
|
|
|
|
|
#include <asm/irq.h>
|
|
|
|
|
#include <asm/mmu_context.h>
|
|
|
|
|
#include <asm/io.h>
|
|
|
|
|
#include <asm/pgtable.h>
|
|
|
|
|
#include <asm/core_titan.h>
|
|
|
|
|
#include <asm/hwrpb.h>
|
|
|
|
|
#include <asm/tlbflush.h>
|
|
|
|
|
|
|
|
|
|
#include "proto.h"
|
|
|
|
|
#include "irq_impl.h"
|
|
|
|
|
#include "pci_impl.h"
|
|
|
|
|
#include "machvec_impl.h"
|
|
|
|
|
#include "err_impl.h"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Titan generic
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Titan supports up to 4 CPUs
|
|
|
|
|
*/
|
|
|
|
|
static unsigned long titan_cpu_irq_affinity[4] = { ~0UL, ~0UL, ~0UL, ~0UL };
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Mask is set (1) if enabled
|
|
|
|
|
*/
|
|
|
|
|
static unsigned long titan_cached_irq_mask;
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Need SMP-safe access to interrupt CSRs
|
|
|
|
|
*/
|
|
|
|
|
DEFINE_SPINLOCK(titan_irq_lock);
|
|
|
|
|
|
|
|
|
|
static void
|
|
|
|
|
titan_update_irq_hw(unsigned long mask)
|
|
|
|
|
{
|
|
|
|
|
register titan_cchip *cchip = TITAN_cchip;
|
|
|
|
|
unsigned long isa_enable = 1UL << 55;
|
|
|
|
|
register int bcpu = boot_cpuid;
|
|
|
|
|
|
|
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
|
cpumask_t cpm = cpu_present_map;
|
|
|
|
|
volatile unsigned long *dim0, *dim1, *dim2, *dim3;
|
|
|
|
|
unsigned long mask0, mask1, mask2, mask3, dummy;
|
|
|
|
|
|
|
|
|
|
mask &= ~isa_enable;
|
|
|
|
|
mask0 = mask & titan_cpu_irq_affinity[0];
|
|
|
|
|
mask1 = mask & titan_cpu_irq_affinity[1];
|
|
|
|
|
mask2 = mask & titan_cpu_irq_affinity[2];
|
|
|
|
|
mask3 = mask & titan_cpu_irq_affinity[3];
|
|
|
|
|
|
|
|
|
|
if (bcpu == 0) mask0 |= isa_enable;
|
|
|
|
|
else if (bcpu == 1) mask1 |= isa_enable;
|
|
|
|
|
else if (bcpu == 2) mask2 |= isa_enable;
|
|
|
|
|
else mask3 |= isa_enable;
|
|
|
|
|
|
|
|
|
|
dim0 = &cchip->dim0.csr;
|
|
|
|
|
dim1 = &cchip->dim1.csr;
|
|
|
|
|
dim2 = &cchip->dim2.csr;
|
|
|
|
|
dim3 = &cchip->dim3.csr;
|
|
|
|
|
if (!cpu_isset(0, cpm)) dim0 = &dummy;
|
|
|
|
|
if (!cpu_isset(1, cpm)) dim1 = &dummy;
|
|
|
|
|
if (!cpu_isset(2, cpm)) dim2 = &dummy;
|
|
|
|
|
if (!cpu_isset(3, cpm)) dim3 = &dummy;
|
|
|
|
|
|
|
|
|
|
*dim0 = mask0;
|
|
|
|
|
*dim1 = mask1;
|
|
|
|
|
*dim2 = mask2;
|
|
|
|
|
*dim3 = mask3;
|
|
|
|
|
mb();
|
|
|
|
|
*dim0;
|
|
|
|
|
*dim1;
|
|
|
|
|
*dim2;
|
|
|
|
|
*dim3;
|
|
|
|
|
#else
|
|
|
|
|
volatile unsigned long *dimB;
|
|
|
|
|
dimB = &cchip->dim0.csr;
|
|
|
|
|
if (bcpu == 1) dimB = &cchip->dim1.csr;
|
|
|
|
|
else if (bcpu == 2) dimB = &cchip->dim2.csr;
|
|
|
|
|
else if (bcpu == 3) dimB = &cchip->dim3.csr;
|
|
|
|
|
|
|
|
|
|
*dimB = mask | isa_enable;
|
|
|
|
|
mb();
|
|
|
|
|
*dimB;
|
|
|
|
|
#endif
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static inline void
|
|
|
|
|
titan_enable_irq(unsigned int irq)
|
|
|
|
|
{
|
|
|
|
|
spin_lock(&titan_irq_lock);
|
|
|
|
|
titan_cached_irq_mask |= 1UL << (irq - 16);
|
|
|
|
|
titan_update_irq_hw(titan_cached_irq_mask);
|
|
|
|
|
spin_unlock(&titan_irq_lock);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static inline void
|
|
|
|
|
titan_disable_irq(unsigned int irq)
|
|
|
|
|
{
|
|
|
|
|
spin_lock(&titan_irq_lock);
|
|
|
|
|
titan_cached_irq_mask &= ~(1UL << (irq - 16));
|
|
|
|
|
titan_update_irq_hw(titan_cached_irq_mask);
|
|
|
|
|
spin_unlock(&titan_irq_lock);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static unsigned int
|
|
|
|
|
titan_startup_irq(unsigned int irq)
|
|
|
|
|
{
|
|
|
|
|
titan_enable_irq(irq);
|
|
|
|
|
return 0; /* never anything pending */
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void
|
|
|
|
|
titan_end_irq(unsigned int irq)
|
|
|
|
|
{
|
|
|
|
|
if (!(irq_desc[irq].status & (IRQ_DISABLED|IRQ_INPROGRESS)))
|
|
|
|
|
titan_enable_irq(irq);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void
|
|
|
|
|
titan_cpu_set_irq_affinity(unsigned int irq, cpumask_t affinity)
|
|
|
|
|
{
|
|
|
|
|
int cpu;
|
|
|
|
|
|
|
|
|
|
for (cpu = 0; cpu < 4; cpu++) {
|
|
|
|
|
if (cpu_isset(cpu, affinity))
|
|
|
|
|
titan_cpu_irq_affinity[cpu] |= 1UL << irq;
|
|
|
|
|
else
|
|
|
|
|
titan_cpu_irq_affinity[cpu] &= ~(1UL << irq);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void
|
|
|
|
|
titan_set_irq_affinity(unsigned int irq, cpumask_t affinity)
|
|
|
|
|
{
|
|
|
|
|
spin_lock(&titan_irq_lock);
|
|
|
|
|
titan_cpu_set_irq_affinity(irq - 16, affinity);
|
|
|
|
|
titan_update_irq_hw(titan_cached_irq_mask);
|
|
|
|
|
spin_unlock(&titan_irq_lock);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void
|
|
|
|
|
titan_device_interrupt(unsigned long vector)
|
|
|
|
|
{
|
|
|
|
|
printk("titan_device_interrupt: NOT IMPLEMENTED YET!! \n");
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void
|
|
|
|
|
titan_srm_device_interrupt(unsigned long vector)
|
|
|
|
|
{
|
|
|
|
|
int irq;
|
|
|
|
|
|
|
|
|
|
irq = (vector - 0x800) >> 4;
|
|
|
|
|
handle_irq(irq);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
static void __init
|
|
|
|
|
init_titan_irqs(struct hw_interrupt_type * ops, int imin, int imax)
|
|
|
|
|
{
|
|
|
|
|
long i;
|
|
|
|
|
for (i = imin; i <= imax; ++i) {
|
|
|
|
|
irq_desc[i].status = IRQ_DISABLED | IRQ_LEVEL;
|
[PATCH] genirq: rename desc->handler to desc->chip
This patch-queue improves the generic IRQ layer to be truly generic, by adding
various abstractions and features to it, without impacting existing
functionality.
While the queue can be best described as "fix and improve everything in the
generic IRQ layer that we could think of", and thus it consists of many
smaller features and lots of cleanups, the one feature that stands out most is
the new 'irq chip' abstraction.
The irq-chip abstraction is about describing and coding and IRQ controller
driver by mapping its raw hardware capabilities [and quirks, if needed] in a
straightforward way, without having to think about "IRQ flow"
(level/edge/etc.) type of details.
This stands in contrast with the current 'irq-type' model of genirq
architectures, which 'mixes' raw hardware capabilities with 'flow' details.
The patchset supports both types of irq controller designs at once, and
converts i386 and x86_64 to the new irq-chip design.
As a bonus side-effect of the irq-chip approach, chained interrupt controllers
(master/slave PIC constructs, etc.) are now supported by design as well.
The end result of this patchset intends to be simpler architecture-level code
and more consolidation between architectures.
We reused many bits of code and many concepts from Russell King's ARM IRQ
layer, the merging of which was one of the motivations for this patchset.
This patch:
rename desc->handler to desc->chip.
Originally i did not want to do this, because it's a big patch. But having
both "desc->handler", "desc->handle_irq" and "action->handler" caused a
large degree of confusion and made the code appear alot less clean than it
truly is.
I have also attempted a dual approach as well by introducing a
desc->chip alias - but that just wasnt robust enough and broke
frequently.
So lets get over with this quickly. The conversion was done automatically
via scripts and converts all the code in the kernel.
This renaming patch is the first one amongst the patches, so that the
remaining patches can stay flexible and can be merged and split up
without having some big monolithic patch act as a merge barrier.
[akpm@osdl.org: build fix]
[akpm@osdl.org: another build fix]
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
19 years ago
|
|
|
|
irq_desc[i].chip = ops;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static struct hw_interrupt_type titan_irq_type = {
|
|
|
|
|
.typename = "TITAN",
|
|
|
|
|
.startup = titan_startup_irq,
|
|
|
|
|
.shutdown = titan_disable_irq,
|
|
|
|
|
.enable = titan_enable_irq,
|
|
|
|
|
.disable = titan_disable_irq,
|
|
|
|
|
.ack = titan_disable_irq,
|
|
|
|
|
.end = titan_end_irq,
|
|
|
|
|
.set_affinity = titan_set_irq_affinity,
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
static irqreturn_t
|
|
|
|
|
titan_intr_nop(int irq, void *dev_id)
|
|
|
|
|
{
|
|
|
|
|
/*
|
|
|
|
|
* This is a NOP interrupt handler for the purposes of
|
|
|
|
|
* event counting -- just return.
|
|
|
|
|
*/
|
|
|
|
|
return IRQ_HANDLED;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void __init
|
|
|
|
|
titan_init_irq(void)
|
|
|
|
|
{
|
|
|
|
|
if (alpha_using_srm && !alpha_mv.device_interrupt)
|
|
|
|
|
alpha_mv.device_interrupt = titan_srm_device_interrupt;
|
|
|
|
|
if (!alpha_mv.device_interrupt)
|
|
|
|
|
alpha_mv.device_interrupt = titan_device_interrupt;
|
|
|
|
|
|
|
|
|
|
titan_update_irq_hw(0);
|
|
|
|
|
|
|
|
|
|
init_titan_irqs(&titan_irq_type, 16, 63 + 16);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void __init
|
|
|
|
|
titan_legacy_init_irq(void)
|
|
|
|
|
{
|
|
|
|
|
/* init the legacy dma controller */
|
|
|
|
|
outb(0, DMA1_RESET_REG);
|
|
|
|
|
outb(0, DMA2_RESET_REG);
|
|
|
|
|
outb(DMA_MODE_CASCADE, DMA2_MODE_REG);
|
|
|
|
|
outb(0, DMA2_MASK_REG);
|
|
|
|
|
|
|
|
|
|
/* init the legacy irq controller */
|
|
|
|
|
init_i8259a_irqs();
|
|
|
|
|
|
|
|
|
|
/* init the titan irqs */
|
|
|
|
|
titan_init_irq();
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
void
|
|
|
|
|
titan_dispatch_irqs(u64 mask)
|
|
|
|
|
{
|
|
|
|
|
unsigned long vector;
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Mask down to those interrupts which are enable on this processor
|
|
|
|
|
*/
|
|
|
|
|
mask &= titan_cpu_irq_affinity[smp_processor_id()];
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Dispatch all requested interrupts
|
|
|
|
|
*/
|
|
|
|
|
while (mask) {
|
|
|
|
|
/* convert to SRM vector... priority is <63> -> <0> */
|
|
|
|
|
vector = 63 - __kernel_ctlz(mask);
|
|
|
|
|
mask &= ~(1UL << vector); /* clear it out */
|
|
|
|
|
vector = 0x900 + (vector << 4); /* convert to SRM vector */
|
|
|
|
|
|
|
|
|
|
/* dispatch it */
|
|
|
|
|
alpha_mv.device_interrupt(vector);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Titan Family
|
|
|
|
|
*/
|
|
|
|
|
static void __init
|
|
|
|
|
titan_request_irq(unsigned int irq, irq_handler_t handler,
|
|
|
|
|
unsigned long irqflags, const char *devname,
|
|
|
|
|
void *dev_id)
|
|
|
|
|
{
|
|
|
|
|
int err;
|
|
|
|
|
err = request_irq(irq, handler, irqflags, devname, dev_id);
|
|
|
|
|
if (err) {
|
|
|
|
|
printk("titan_request_irq for IRQ %d returned %d; ignoring\n",
|
|
|
|
|
irq, err);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void __init
|
|
|
|
|
titan_late_init(void)
|
|
|
|
|
{
|
|
|
|
|
/*
|
|
|
|
|
* Enable the system error interrupts. These interrupts are
|
|
|
|
|
* all reported to the kernel as machine checks, so the handler
|
|
|
|
|
* is a nop so it can be called to count the individual events.
|
|
|
|
|
*/
|
|
|
|
|
titan_request_irq(63+16, titan_intr_nop, IRQF_DISABLED,
|
|
|
|
|
"CChip Error", NULL);
|
|
|
|
|
titan_request_irq(62+16, titan_intr_nop, IRQF_DISABLED,
|
|
|
|
|
"PChip 0 H_Error", NULL);
|
|
|
|
|
titan_request_irq(61+16, titan_intr_nop, IRQF_DISABLED,
|
|
|
|
|
"PChip 1 H_Error", NULL);
|
|
|
|
|
titan_request_irq(60+16, titan_intr_nop, IRQF_DISABLED,
|
|
|
|
|
"PChip 0 C_Error", NULL);
|
|
|
|
|
titan_request_irq(59+16, titan_intr_nop, IRQF_DISABLED,
|
|
|
|
|
"PChip 1 C_Error", NULL);
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Register our error handlers.
|
|
|
|
|
*/
|
|
|
|
|
titan_register_error_handlers();
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Check if the console left us any error logs.
|
|
|
|
|
*/
|
|
|
|
|
cdl_check_console_data_log();
|
|
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static int __devinit
|
|
|
|
|
titan_map_irq(struct pci_dev *dev, u8 slot, u8 pin)
|
|
|
|
|
{
|
|
|
|
|
u8 intline;
|
|
|
|
|
int irq;
|
|
|
|
|
|
|
|
|
|
/* Get the current intline. */
|
|
|
|
|
pci_read_config_byte(dev, PCI_INTERRUPT_LINE, &intline);
|
|
|
|
|
irq = intline;
|
|
|
|
|
|
|
|
|
|
/* Is it explicitly routed through ISA? */
|
|
|
|
|
if ((irq & 0xF0) == 0xE0)
|
|
|
|
|
return irq;
|
|
|
|
|
|
|
|
|
|
/* Offset by 16 to make room for ISA interrupts 0 - 15. */
|
|
|
|
|
return irq + 16;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void __init
|
|
|
|
|
titan_init_pci(void)
|
|
|
|
|
{
|
|
|
|
|
/*
|
|
|
|
|
* This isn't really the right place, but there's some init
|
|
|
|
|
* that needs to be done after everything is basically up.
|
|
|
|
|
*/
|
|
|
|
|
titan_late_init();
|
|
|
|
|
|
|
|
|
|
pci_probe_only = 1;
|
|
|
|
|
common_init_pci();
|
|
|
|
|
SMC669_Init(0);
|
|
|
|
|
locate_and_init_vga(NULL);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Privateer
|
|
|
|
|
*/
|
|
|
|
|
static void __init
|
|
|
|
|
privateer_init_pci(void)
|
|
|
|
|
{
|
|
|
|
|
/*
|
|
|
|
|
* Hook a couple of extra err interrupts that the
|
|
|
|
|
* common titan code won't.
|
|
|
|
|
*/
|
|
|
|
|
titan_request_irq(53+16, titan_intr_nop, IRQF_DISABLED,
|
|
|
|
|
"NMI", NULL);
|
|
|
|
|
titan_request_irq(50+16, titan_intr_nop, IRQF_DISABLED,
|
|
|
|
|
"Temperature Warning", NULL);
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Finish with the common version.
|
|
|
|
|
*/
|
|
|
|
|
return titan_init_pci();
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* The System Vectors.
|
|
|
|
|
*/
|
|
|
|
|
struct alpha_machine_vector titan_mv __initmv = {
|
|
|
|
|
.vector_name = "TITAN",
|
|
|
|
|
DO_EV6_MMU,
|
|
|
|
|
DO_DEFAULT_RTC,
|
|
|
|
|
DO_TITAN_IO,
|
|
|
|
|
.machine_check = titan_machine_check,
|
|
|
|
|
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
|
|
|
|
|
.min_io_address = DEFAULT_IO_BASE,
|
|
|
|
|
.min_mem_address = DEFAULT_MEM_BASE,
|
|
|
|
|
.pci_dac_offset = TITAN_DAC_OFFSET,
|
|
|
|
|
|
|
|
|
|
.nr_irqs = 80, /* 64 + 16 */
|
|
|
|
|
/* device_interrupt will be filled in by titan_init_irq */
|
|
|
|
|
|
|
|
|
|
.agp_info = titan_agp_info,
|
|
|
|
|
|
|
|
|
|
.init_arch = titan_init_arch,
|
|
|
|
|
.init_irq = titan_legacy_init_irq,
|
|
|
|
|
.init_rtc = common_init_rtc,
|
|
|
|
|
.init_pci = titan_init_pci,
|
|
|
|
|
|
|
|
|
|
.kill_arch = titan_kill_arch,
|
|
|
|
|
.pci_map_irq = titan_map_irq,
|
|
|
|
|
.pci_swizzle = common_swizzle,
|
|
|
|
|
};
|
|
|
|
|
ALIAS_MV(titan)
|
|
|
|
|
|
|
|
|
|
struct alpha_machine_vector privateer_mv __initmv = {
|
|
|
|
|
.vector_name = "PRIVATEER",
|
|
|
|
|
DO_EV6_MMU,
|
|
|
|
|
DO_DEFAULT_RTC,
|
|
|
|
|
DO_TITAN_IO,
|
|
|
|
|
.machine_check = privateer_machine_check,
|
|
|
|
|
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
|
|
|
|
|
.min_io_address = DEFAULT_IO_BASE,
|
|
|
|
|
.min_mem_address = DEFAULT_MEM_BASE,
|
|
|
|
|
.pci_dac_offset = TITAN_DAC_OFFSET,
|
|
|
|
|
|
|
|
|
|
.nr_irqs = 80, /* 64 + 16 */
|
|
|
|
|
/* device_interrupt will be filled in by titan_init_irq */
|
|
|
|
|
|
|
|
|
|
.agp_info = titan_agp_info,
|
|
|
|
|
|
|
|
|
|
.init_arch = titan_init_arch,
|
|
|
|
|
.init_irq = titan_legacy_init_irq,
|
|
|
|
|
.init_rtc = common_init_rtc,
|
|
|
|
|
.init_pci = privateer_init_pci,
|
|
|
|
|
|
|
|
|
|
.kill_arch = titan_kill_arch,
|
|
|
|
|
.pci_map_irq = titan_map_irq,
|
|
|
|
|
.pci_swizzle = common_swizzle,
|
|
|
|
|
};
|
|
|
|
|
/* No alpha_mv alias for privateer since we compile it
|
|
|
|
|
in unconditionally with titan; setup_arch knows how to cope. */
|