1/* 2 * Linux/Meta general interrupt handling code 3 * 4 */ 5 6#include <linux/kernel.h> 7#include <linux/interrupt.h> 8#include <linux/init.h> 9#include <linux/irqchip/metag-ext.h> 10#include <linux/irqchip/metag.h> 11#include <linux/irqdomain.h> 12#include <linux/ratelimit.h> 13 14#include <asm/core_reg.h> 15#include <asm/mach/arch.h> 16#include <asm/uaccess.h> 17 18#ifdef CONFIG_4KSTACKS 19union irq_ctx { 20 struct thread_info tinfo; 21 u32 stack[THREAD_SIZE/sizeof(u32)]; 22}; 23 24static union irq_ctx *hardirq_ctx[NR_CPUS] __read_mostly; 25static union irq_ctx *softirq_ctx[NR_CPUS] __read_mostly; 26#endif 27 28static struct irq_domain *root_domain; 29 30static unsigned int startup_meta_irq(struct irq_data *data) 31{ 32 tbi_startup_interrupt(data->hwirq); 33 return 0; 34} 35 36static void shutdown_meta_irq(struct irq_data *data) 37{ 38 tbi_shutdown_interrupt(data->hwirq); 39} 40 41void do_IRQ(int irq, struct pt_regs *regs) 42{ 43 struct pt_regs *old_regs = set_irq_regs(regs); 44#ifdef CONFIG_4KSTACKS 45 struct irq_desc *desc; 46 union irq_ctx *curctx, *irqctx; 47 u32 *isp; 48#endif 49 50 irq_enter(); 51 52 irq = irq_linear_revmap(root_domain, irq); 53 54#ifdef CONFIG_DEBUG_STACKOVERFLOW 55 /* Debugging check for stack overflow: is there less than 1KB free? */ 56 { 57 unsigned long sp; 58 59 sp = __core_reg_get(A0StP); 60 sp &= THREAD_SIZE - 1; 61 62 if (unlikely(sp > (THREAD_SIZE - 1024))) 63 pr_err("Stack overflow in do_IRQ: %ld\n", sp); 64 } 65#endif 66 67 68#ifdef CONFIG_4KSTACKS 69 curctx = (union irq_ctx *) current_thread_info(); 70 irqctx = hardirq_ctx[smp_processor_id()]; 71 72 /* 73 * this is where we switch to the IRQ stack. However, if we are 74 * already using the IRQ stack (because we interrupted a hardirq 75 * handler) we can't do that and just have to keep using the 76 * current stack (which is the irq stack already after all) 77 */ 78 if (curctx != irqctx) { 79 /* build the stack frame on the IRQ stack */ 80 isp = (u32 *) ((char *)irqctx + sizeof(struct thread_info)); 81 irqctx->tinfo.task = curctx->tinfo.task; 82 83 /* 84 * Copy the softirq bits in preempt_count so that the 85 * softirq checks work in the hardirq context. 86 */ 87 irqctx->tinfo.preempt_count = 88 (irqctx->tinfo.preempt_count & ~SOFTIRQ_MASK) | 89 (curctx->tinfo.preempt_count & SOFTIRQ_MASK); 90 91 desc = irq_to_desc(irq); 92 93 asm volatile ( 94 "MOV D0.5,%0\n" 95 "MOV D1Ar1,%1\n" 96 "MOV D1RtP,%2\n" 97 "MOV D0Ar2,%3\n" 98 "SWAP A0StP,D0.5\n" 99 "SWAP PC,D1RtP\n" 100 "MOV A0StP,D0.5\n" 101 : 102 : "r" (isp), "r" (irq), "r" (desc->handle_irq), 103 "r" (desc) 104 : "memory", "cc", "D1Ar1", "D0Ar2", "D1Ar3", "D0Ar4", 105 "D1Ar5", "D0Ar6", "D0Re0", "D1Re0", "D0.4", "D1RtP", 106 "D0.5" 107 ); 108 } else 109#endif 110 generic_handle_irq(irq); 111 112 irq_exit(); 113 114 set_irq_regs(old_regs); 115} 116 117#ifdef CONFIG_4KSTACKS 118 119static char softirq_stack[NR_CPUS * THREAD_SIZE] __page_aligned_bss; 120 121static char hardirq_stack[NR_CPUS * THREAD_SIZE] __page_aligned_bss; 122 123/* 124 * allocate per-cpu stacks for hardirq and for softirq processing 125 */ 126void irq_ctx_init(int cpu) 127{ 128 union irq_ctx *irqctx; 129 130 if (hardirq_ctx[cpu]) 131 return; 132 133 irqctx = (union irq_ctx *) &hardirq_stack[cpu * THREAD_SIZE]; 134 irqctx->tinfo.task = NULL; 135 irqctx->tinfo.cpu = cpu; 136 irqctx->tinfo.preempt_count = HARDIRQ_OFFSET; 137 irqctx->tinfo.addr_limit = MAKE_MM_SEG(0); 138 139 hardirq_ctx[cpu] = irqctx; 140 141 irqctx = (union irq_ctx *) &softirq_stack[cpu * THREAD_SIZE]; 142 irqctx->tinfo.task = NULL; 143 irqctx->tinfo.cpu = cpu; 144 irqctx->tinfo.preempt_count = 0; 145 irqctx->tinfo.addr_limit = MAKE_MM_SEG(0); 146 147 softirq_ctx[cpu] = irqctx; 148 149 pr_info("CPU %u irqstacks, hard=%p soft=%p\n", 150 cpu, hardirq_ctx[cpu], softirq_ctx[cpu]); 151} 152 153void irq_ctx_exit(int cpu) 154{ 155 hardirq_ctx[smp_processor_id()] = NULL; 156} 157 158extern asmlinkage void __do_softirq(void); 159 160void do_softirq_own_stack(void) 161{ 162 struct thread_info *curctx; 163 union irq_ctx *irqctx; 164 u32 *isp; 165 166 curctx = current_thread_info(); 167 irqctx = softirq_ctx[smp_processor_id()]; 168 irqctx->tinfo.task = curctx->task; 169 170 /* build the stack frame on the softirq stack */ 171 isp = (u32 *) ((char *)irqctx + sizeof(struct thread_info)); 172 173 asm volatile ( 174 "MOV D0.5,%0\n" 175 "SWAP A0StP,D0.5\n" 176 "CALLR D1RtP,___do_softirq\n" 177 "MOV A0StP,D0.5\n" 178 : 179 : "r" (isp) 180 : "memory", "cc", "D1Ar1", "D0Ar2", "D1Ar3", "D0Ar4", 181 "D1Ar5", "D0Ar6", "D0Re0", "D1Re0", "D0.4", "D1RtP", 182 "D0.5" 183 ); 184} 185#endif 186 187static struct irq_chip meta_irq_type = { 188 .name = "META-IRQ", 189 .irq_startup = startup_meta_irq, 190 .irq_shutdown = shutdown_meta_irq, 191}; 192 193/** 194 * tbisig_map() - Map a TBI signal number to a virtual IRQ number. 195 * @hw: Number of the TBI signal. Must be in range. 196 * 197 * Returns: The virtual IRQ number of the TBI signal number IRQ specified by 198 * @hw. 199 */ 200int tbisig_map(unsigned int hw) 201{ 202 return irq_create_mapping(root_domain, hw); 203} 204 205/** 206 * metag_tbisig_map() - map a tbi signal to a Linux virtual IRQ number 207 * @d: root irq domain 208 * @irq: virtual irq number 209 * @hw: hardware irq number (TBI signal number) 210 * 211 * This sets up a virtual irq for a specified TBI signal number. 212 */ 213static int metag_tbisig_map(struct irq_domain *d, unsigned int irq, 214 irq_hw_number_t hw) 215{ 216#ifdef CONFIG_SMP 217 irq_set_chip_and_handler(irq, &meta_irq_type, handle_percpu_irq); 218#else 219 irq_set_chip_and_handler(irq, &meta_irq_type, handle_simple_irq); 220#endif 221 return 0; 222} 223 224static const struct irq_domain_ops metag_tbisig_domain_ops = { 225 .map = metag_tbisig_map, 226}; 227 228/* 229 * void init_IRQ(void) 230 * 231 * Parameters: None 232 * 233 * Returns: Nothing 234 * 235 * This function should be called during kernel startup to initialize 236 * the IRQ handling routines. 237 */ 238void __init init_IRQ(void) 239{ 240 root_domain = irq_domain_add_linear(NULL, 32, 241 &metag_tbisig_domain_ops, NULL); 242 if (unlikely(!root_domain)) 243 panic("init_IRQ: cannot add root IRQ domain"); 244 245 irq_ctx_init(smp_processor_id()); 246 247 init_internal_IRQ(); 248 init_external_IRQ(); 249 250 if (machine_desc->init_irq) 251 machine_desc->init_irq(); 252} 253 254int __init arch_probe_nr_irqs(void) 255{ 256 if (machine_desc->nr_irqs) 257 nr_irqs = machine_desc->nr_irqs; 258 return 0; 259} 260 261#ifdef CONFIG_HOTPLUG_CPU 262/* 263 * The CPU has been marked offline. Migrate IRQs off this CPU. If 264 * the affinity settings do not allow other CPUs, force them onto any 265 * available CPU. 266 */ 267void migrate_irqs(void) 268{ 269 unsigned int i, cpu = smp_processor_id(); 270 271 for_each_active_irq(i) { 272 struct irq_data *data = irq_get_irq_data(i); 273 unsigned int newcpu; 274 275 if (irqd_is_per_cpu(data)) 276 continue; 277 278 if (!cpumask_test_cpu(cpu, data->affinity)) 279 continue; 280 281 newcpu = cpumask_any_and(data->affinity, cpu_online_mask); 282 283 if (newcpu >= nr_cpu_ids) { 284 pr_info_ratelimited("IRQ%u no longer affine to CPU%u\n", 285 i, cpu); 286 287 cpumask_setall(data->affinity); 288 } 289 irq_set_affinity(i, data->affinity); 290 } 291} 292#endif /* CONFIG_HOTPLUG_CPU */ 293