/* * Atheros AR71xx/AR724x/AR913x specific interrupt handling * * Copyright (C) 2010-2011 Jaiganesh Narayanan * Copyright (C) 2008-2011 Gabor Juhos * Copyright (C) 2008 Imre Kaloz * * Parts of this file are based on Atheros' 2.6.15/2.6.31 BSP * * This program is free software; you can redistribute it and/or modify it * under the terms of the GNU General Public License version 2 as published * by the Free Software Foundation. */ #include #include #include #include #include #include #include #include #include #include "common.h" #include "machtypes.h" #if defined(CONFIG_AVM_POWER) #include #endif #if defined(CONFIG_AVM_SIMPLE_PROFILING) #include #endif /*--- #if defined(CONFIG_AVM_SIMPLE_PROFILING) ---*/ static struct irq_chip ip2_chip; static struct irq_chip ip3_chip; static void ath79_misc_irq_handler(struct irq_desc *desc) { void __iomem *base = ath79_reset_base; u32 pending; pending = __raw_readl(base + AR71XX_RESET_REG_MISC_INT_STATUS) & __raw_readl(base + AR71XX_RESET_REG_MISC_INT_ENABLE); if (!pending) { spurious_interrupt(); return; } while (pending) { int bit = __ffs(pending); #if defined(CONFIG_AVM_SIMPLE_PROFILING) avm_simple_profiling_log(avm_profile_data_type_hw_irq_begin, (unsigned int)(irq_desc + ATH79_MISC_IRQ(bit)), ATH79_MISC_IRQ(bit)); #endif /*--- #if defined(CONFIG_AVM_SIMPLE_PROFILING) ---*/ generic_handle_irq(ATH79_MISC_IRQ(bit)); #if defined(CONFIG_AVM_SIMPLE_PROFILING) avm_simple_profiling_log(avm_profile_data_type_hw_irq_end, (unsigned int)(irq_desc + ATH79_MISC_IRQ(bit)), ATH79_MISC_IRQ(bit)); #endif /*--- #if defined(CONFIG_AVM_SIMPLE_PROFILING) ---*/ pending &= ~BIT(bit); } } static void ar71xx_misc_irq_unmask(struct irq_data *d) { unsigned int irq = d->irq - ATH79_MISC_IRQ_BASE; void __iomem *base = ath79_reset_base; u32 t; t = __raw_readl(base + AR71XX_RESET_REG_MISC_INT_ENABLE); __raw_writel(t | (1 << irq), base + AR71XX_RESET_REG_MISC_INT_ENABLE); /* flush write */ __raw_readl(base + AR71XX_RESET_REG_MISC_INT_ENABLE); } static void ar71xx_misc_irq_mask(struct irq_data *d) { unsigned int irq = d->irq - ATH79_MISC_IRQ_BASE; void __iomem *base = ath79_reset_base; u32 t; t = __raw_readl(base + AR71XX_RESET_REG_MISC_INT_ENABLE); __raw_writel(t & ~(1 << irq), base + AR71XX_RESET_REG_MISC_INT_ENABLE); /* flush write */ __raw_readl(base + AR71XX_RESET_REG_MISC_INT_ENABLE); } static void ar724x_misc_irq_ack(struct irq_data *d) { unsigned int irq = d->irq - ATH79_MISC_IRQ_BASE; void __iomem *base = ath79_reset_base; u32 t; t = __raw_readl(base + AR71XX_RESET_REG_MISC_INT_STATUS); __raw_writel(t & ~(1 << irq), base + AR71XX_RESET_REG_MISC_INT_STATUS); /* flush write */ __raw_readl(base + AR71XX_RESET_REG_MISC_INT_STATUS); } static struct irq_chip ath79_misc_irq_chip = { .name = "MISC", .irq_unmask = ar71xx_misc_irq_unmask, .irq_mask = ar71xx_misc_irq_mask, }; static void __init ath79_misc_irq_init(void) { void __iomem *base = ath79_reset_base; int i; __raw_writel(0, base + AR71XX_RESET_REG_MISC_INT_ENABLE); __raw_writel(0, base + AR71XX_RESET_REG_MISC_INT_STATUS); if (soc_is_ar71xx() || soc_is_ar913x()) ath79_misc_irq_chip.irq_mask_ack = ar71xx_misc_irq_mask; else if (soc_is_ar724x() || soc_is_ar933x() || soc_is_ar934x() || soc_is_qca953x() || soc_is_qca955x() || soc_is_qca956x() || soc_is_qcn550x() || soc_is_tp9343()) ath79_misc_irq_chip.irq_ack = ar724x_misc_irq_ack; else BUG(); for (i = ATH79_MISC_IRQ_BASE; i < ATH79_MISC_IRQ_BASE + ATH79_MISC_IRQ_COUNT; i++) { irq_set_chip_and_handler(i, &ath79_misc_irq_chip, handle_level_irq); } irq_set_chained_handler(ATH79_CPU_IRQ(6), ath79_misc_irq_handler); } static void ar934x_ip2_irq_dispatch(struct irq_desc *desc) { u32 status; status = ath79_reset_rr(AR934X_RESET_REG_PCIE_WMAC_INT_STATUS); if (status & AR934X_PCIE_WMAC_INT_PCIE_ALL) { ath79_ddr_wb_flush(3); generic_handle_irq(ATH79_IP2_IRQ(0)); } else if (status & AR934X_PCIE_WMAC_INT_WMAC_ALL) { ath79_ddr_wb_flush(4); generic_handle_irq(ATH79_IP2_IRQ(1)); } else { spurious_interrupt(); } } static void ar934x_ip2_irq_init(void) { int i; for (i = ATH79_IP2_IRQ_BASE; i < ATH79_IP2_IRQ_BASE + ATH79_IP2_IRQ_COUNT; i++) irq_set_chip_and_handler(i, &ip2_chip, handle_level_irq); irq_set_chained_handler(ATH79_CPU_IRQ(2), ar934x_ip2_irq_dispatch); } static void qca953x_ip2_irq_dispatch(struct irq_desc *desc) { u32 status; status = ath79_reset_rr(QCA953X_RESET_REG_PCIE_WMAC_INT_STATUS); if (status & QCA953X_PCIE_WMAC_INT_PCIE_ALL) { ath79_ddr_wb_flush(3); generic_handle_irq(ATH79_IP2_IRQ(0)); } else if (status & QCA953X_PCIE_WMAC_INT_WMAC_ALL) { ath79_ddr_wb_flush(4); generic_handle_irq(ATH79_IP2_IRQ(1)); } else { spurious_interrupt(); } } static void qca953x_irq_init(void) { int i; for (i = ATH79_IP2_IRQ_BASE; i < ATH79_IP2_IRQ_BASE + ATH79_IP2_IRQ_COUNT; i++) irq_set_chip_and_handler(i, &ip2_chip, handle_level_irq); irq_set_chained_handler(ATH79_CPU_IRQ(2), qca953x_ip2_irq_dispatch); } static void qca955x_ip2_irq_dispatch(struct irq_desc *desc) { u32 status; status = ath79_reset_rr(QCA955X_RESET_REG_EXT_INT_STATUS); status &= QCA955X_EXT_INT_PCIE_RC1_ALL | QCA955X_EXT_INT_WMAC_ALL; if (status == 0) { spurious_interrupt(); return; } if (status & QCA955X_EXT_INT_PCIE_RC1_ALL) { /* TODO: flush DDR? */ generic_handle_irq(ATH79_IP2_IRQ(0)); } if (status & QCA955X_EXT_INT_WMAC_ALL) { /* TODO: flush DDR? */ generic_handle_irq(ATH79_IP2_IRQ(1)); } } static void qca955x_ip3_irq_dispatch(struct irq_desc *desc) { u32 status; status = ath79_reset_rr(QCA955X_RESET_REG_EXT_INT_STATUS); status &= QCA955X_EXT_INT_PCIE_RC2_ALL | QCA955X_EXT_INT_USB1 | QCA955X_EXT_INT_USB2; if (status == 0) { spurious_interrupt(); return; } if (status & QCA955X_EXT_INT_USB1) { /* TODO: flush DDR? */ generic_handle_irq(ATH79_IP3_IRQ(0)); } if (status & QCA955X_EXT_INT_USB2) { /* TODO: flush DDR? */ generic_handle_irq(ATH79_IP3_IRQ(1)); } if (status & QCA955X_EXT_INT_PCIE_RC2_ALL) { /* TODO: flush DDR? */ generic_handle_irq(ATH79_IP3_IRQ(2)); } } static void qca955x_irq_init(void) { int i; for (i = ATH79_IP2_IRQ_BASE; i < ATH79_IP2_IRQ_BASE + ATH79_IP2_IRQ_COUNT; i++) irq_set_chip_and_handler(i, &ip2_chip, handle_level_irq); irq_set_chained_handler(ATH79_CPU_IRQ(2), qca955x_ip2_irq_dispatch); for (i = ATH79_IP3_IRQ_BASE; i < ATH79_IP3_IRQ_BASE + ATH79_IP3_IRQ_COUNT; i++) irq_set_chip_and_handler(i, &ip3_chip, handle_level_irq); irq_set_chained_handler(ATH79_CPU_IRQ(3), qca955x_ip3_irq_dispatch); } /* * The IP2/IP3 lines are tied to a PCI/WMAC/USB device. Drivers for * these devices typically allocate coherent DMA memory, however the * DMA controller may still have some unsynchronized data in the FIFO. * Issue a flush in the handlers to ensure that the driver sees * the update. * * This array map the interrupt lines to the DDR write buffer channels. */ static unsigned irq_wb_chan[8] = { -1, -1, -1, -1, -1, -1, -1, -1, }; static void qca956x_ip2_irq_dispatch(struct irq_desc *desc) { u32 status; status = ath79_reset_rr(QCA956X_RESET_REG_EXT_INT_STATUS); status &= QCA956X_EXT_INT_PCIE_RC1_ALL | QCA956X_EXT_INT_WMAC_ALL; if (status == 0) { spurious_interrupt(); return; } if (status & QCA956X_EXT_INT_PCIE_RC1_ALL) { /* TODO: flush DDR? */ generic_handle_irq(ATH79_IP2_IRQ(0)); } if (status & QCA956X_EXT_INT_WMAC_ALL) { /* TODO: flsuh DDR? */ generic_handle_irq(ATH79_IP2_IRQ(1)); } } static void qca956x_ip3_irq_dispatch(struct irq_desc *desc) { u32 status; status = ath79_reset_rr(QCA956X_RESET_REG_EXT_INT_STATUS); status &= QCA956X_EXT_INT_PCIE_RC2_ALL | QCA956X_EXT_INT_USB1 | QCA956X_EXT_INT_USB2; if (status == 0) { spurious_interrupt(); return; } if (status & QCA956X_EXT_INT_USB1) { /* TODO: flush DDR? */ generic_handle_irq(ATH79_IP3_IRQ(0)); } if (status & QCA956X_EXT_INT_USB2) { /* TODO: flush DDR? */ generic_handle_irq(ATH79_IP3_IRQ(1)); } if (status & QCA956X_EXT_INT_PCIE_RC2_ALL) { /* TODO: flush DDR? */ generic_handle_irq(ATH79_IP3_IRQ(2)); } } static void qca956x_qcn550x_enable_timer_cb(void) { u32 misc; misc = ath79_reset_rr(AR71XX_RESET_REG_MISC_INT_ENABLE); misc |= MISC_INT_MIPS_SI_TIMERINT_MASK; ath79_reset_wr(AR71XX_RESET_REG_MISC_INT_ENABLE, misc); } static void qca956x_irq_init(void) { int i; for (i = ATH79_IP2_IRQ_BASE; i < ATH79_IP2_IRQ_BASE + ATH79_IP2_IRQ_COUNT; i++) irq_set_chip_and_handler(i, &ip2_chip, handle_level_irq); irq_set_chained_handler(ATH79_CPU_IRQ(2), qca956x_ip2_irq_dispatch); for (i = ATH79_IP3_IRQ_BASE; i < ATH79_IP3_IRQ_BASE + ATH79_IP3_IRQ_COUNT; i++) irq_set_chip_and_handler(i, &ip3_chip, handle_level_irq); irq_set_chained_handler(ATH79_CPU_IRQ(3), qca956x_ip3_irq_dispatch); /* QCA956x timer init workaround has to be applied right before setting * up the clock. Else, there will be no jiffies */ late_time_init = &qca956x_qcn550x_enable_timer_cb; } static void qcn550x_ip2_irq_dispatch(struct irq_desc *desc) { u32 status; status = ath79_reset_rr(QCN550X_RESET_REG_EXT_INT_STATUS); status &= QCN550X_EXT_INT_PCIE_RC1_ALL | QCN550X_EXT_INT_WMAC_ALL; if (status == 0) { spurious_interrupt(); return; } if (status & QCN550X_EXT_INT_PCIE_RC1_ALL) { /* TODO: flush DDR? */ generic_handle_irq(ATH79_IP2_IRQ(0)); } if (status & QCN550X_EXT_INT_WMAC_ALL) { /* TODO: flsuh DDR? */ generic_handle_irq(ATH79_IP2_IRQ(1)); } } static void qcn550x_ip3_irq_dispatch(struct irq_desc *desc) { u32 status; status = ath79_reset_rr(QCN550X_RESET_REG_EXT_INT_STATUS); status &= QCN550X_EXT_INT_PCIE_RC2_ALL | QCN550X_EXT_INT_USB1; if (status == 0) { spurious_interrupt(); return; } if (status & QCN550X_EXT_INT_USB1) { /* TODO: flush DDR? */ generic_handle_irq(ATH79_IP3_IRQ(0)); } if (status & QCN550X_EXT_INT_PCIE_RC2_ALL) { /* TODO: flush DDR? */ generic_handle_irq(ATH79_IP3_IRQ(2)); } } static void qcn550x_irq_init(void) { int i; for (i = ATH79_IP2_IRQ_BASE; i < ATH79_IP2_IRQ_BASE + ATH79_IP2_IRQ_COUNT; i++) irq_set_chip_and_handler(i, &ip2_chip, handle_level_irq); irq_set_chained_handler(ATH79_CPU_IRQ(2), qcn550x_ip2_irq_dispatch); for (i = ATH79_IP3_IRQ_BASE; i < ATH79_IP3_IRQ_BASE + ATH79_IP3_IRQ_COUNT; i++) irq_set_chip_and_handler(i, &ip3_chip, handle_level_irq); irq_set_chained_handler(ATH79_CPU_IRQ(3), qcn550x_ip3_irq_dispatch); /* QCN550x timer init workaround has to be applied right before setting * up the clock. Else, there will be no jiffies */ late_time_init = &qca956x_qcn550x_enable_timer_cb; } #if defined(CONFIG_AVM_SIMPLE_PROFILING) asmlinkage void plat_irq_dispatch(struct pt_regs *regs __maybe_unused) #else asmlinkage void plat_irq_dispatch(void) #endif /*--- #if defined(CONFIG_AVM_SIMPLE_PROFILING) ---*/ { unsigned long pending; int irq; #if defined(CONFIG_AVM_POWER) avm_cpu_wait_end(); /*--- auch wenn es r4k_wait_irqoff gibt: trotzdem aufrufen, um system-load-Ausgabe zu triggern ---*/ #endif/*--- #if defined(CONFIG_AVM_POWER) ---*/ #if defined(CONFIG_AVM_SIMPLE_PROFILING) avm_simple_profiling_enter_irq_context(regs->cp0_epc, regs->regs[31]); #endif /*--- #if defined(CONFIG_AVM_SIMPLE_PROFILING) ---*/ do { pending = read_c0_status() & read_c0_cause() & ST0_IM; if (!pending) { spurious_interrupt(); return; } pending >>= CAUSEB_IP; irq = fls(pending) - 1; if (irq < ARRAY_SIZE(irq_wb_chan) && irq_wb_chan[irq] != -1) ath79_ddr_wb_flush(irq_wb_chan[irq]); #if defined(CONFIG_AVM_SIMPLE_PROFILING) avm_simple_profiling_log(avm_profile_data_type_hw_irq_begin, MIPS_CPU_IRQ_BASE + irq, irq); #endif /*--- #if defined(CONFIG_AVM_SIMPLE_PROFILING) ---*/ do_IRQ(MIPS_CPU_IRQ_BASE + irq); #if defined(CONFIG_AVM_SIMPLE_PROFILING) avm_simple_profiling_log(avm_profile_data_type_hw_irq_end, MIPS_CPU_IRQ_BASE + irq, irq); #endif /*--- #if defined(CONFIG_AVM_SIMPLE_PROFILING) ---*/ pending &= ~BIT(irq); } while(pending); } #ifdef CONFIG_IRQCHIP static int misc_map(struct irq_domain *d, unsigned int irq, irq_hw_number_t hw) { irq_set_chip_and_handler(irq, &ath79_misc_irq_chip, handle_level_irq); return 0; } static const struct irq_domain_ops misc_irq_domain_ops = { .xlate = irq_domain_xlate_onecell, .map = misc_map, }; static int ip2_map(struct irq_domain *d, unsigned int irq, irq_hw_number_t hw) { irq_set_chip_and_handler(irq, &ip2_chip, handle_level_irq); return 0; } static const struct irq_domain_ops ip2_irq_domain_ops = { .xlate = irq_domain_xlate_onecell, .map = ip2_map, }; static int ip3_map(struct irq_domain *d, unsigned int irq, irq_hw_number_t hw) { irq_set_chip_and_handler(irq, &ip3_chip, handle_level_irq); return 0; } static const struct irq_domain_ops ip3_irq_domain_ops = { .xlate = irq_domain_xlate_onecell, .map = ip3_map, }; static int __init ath79_misc_intc_of_init( struct device_node *node, struct device_node *parent) { void __iomem *base = ath79_reset_base; struct irq_domain *domain; int irq; irq = irq_of_parse_and_map(node, 0); if (!irq) panic("Failed to get MISC IRQ"); domain = irq_domain_add_legacy(node, ATH79_MISC_IRQ_COUNT, ATH79_MISC_IRQ_BASE, 0, &misc_irq_domain_ops, NULL); if (!domain) panic("Failed to add MISC irqdomain"); /* Disable and clear all interrupts */ __raw_writel(0, base + AR71XX_RESET_REG_MISC_INT_ENABLE); __raw_writel(0, base + AR71XX_RESET_REG_MISC_INT_STATUS); irq_set_chained_handler(irq, ath79_misc_irq_handler); return 0; } static int __init ar7100_misc_intc_of_init( struct device_node *node, struct device_node *parent) { ath79_misc_irq_chip.irq_mask_ack = ar71xx_misc_irq_mask; return ath79_misc_intc_of_init(node, parent); } IRQCHIP_DECLARE(ar7100_misc_intc, "qca,ar7100-misc-intc", ar7100_misc_intc_of_init); static int __init ar7240_misc_intc_of_init( struct device_node *node, struct device_node *parent) { ath79_misc_irq_chip.irq_ack = ar724x_misc_irq_ack; return ath79_misc_intc_of_init(node, parent); } IRQCHIP_DECLARE(ar7240_misc_intc, "qca,ar7240-misc-intc", ar7240_misc_intc_of_init); static int __init ath79_ip2_irq_of_init( struct device_node *node, struct device_node *parent) { struct irq_domain *domain; int irq; irq = irq_of_parse_and_map(node, 0); if (!irq) panic("Failed to get IP2 IRQ"); domain = irq_domain_add_legacy(node, ATH79_IP2_IRQ_COUNT, ATH79_IP2_IRQ_BASE, 0, &ip2_irq_domain_ops, NULL); if (!domain) panic("Failed to add IP2 irqdomain"); if (soc_is_ar934x()) irq_set_chained_handler(ATH79_CPU_IRQ(2), ar934x_ip2_irq_dispatch); else if (soc_is_qca953x()) irq_set_chained_handler(ATH79_CPU_IRQ(2), qca953x_ip2_irq_dispatch); else if (soc_is_qca955x()) irq_set_chained_handler(ATH79_CPU_IRQ(2), qca955x_ip2_irq_dispatch); else if (soc_is_qca956x() || soc_is_tp9343()) irq_set_chained_handler(ATH79_CPU_IRQ(2), qca956x_ip2_irq_dispatch); else if (soc_is_qcn550x()) irq_set_chained_handler(ATH79_CPU_IRQ(2), qcn550x_ip2_irq_dispatch); return 0; } IRQCHIP_DECLARE(ath79_ip2_intc, "qca,ath79-ip2-intc", ath79_ip2_irq_of_init); static int __init ath79_ip3_irq_of_init( struct device_node *node, struct device_node *parent) { struct irq_domain *domain; int irq; irq = irq_of_parse_and_map(node, 0); if (!irq) panic("Failed to get IP3 IRQ"); domain = irq_domain_add_legacy(node, ATH79_IP3_IRQ_COUNT, ATH79_IP3_IRQ_BASE, 0, &ip3_irq_domain_ops, NULL); if (!domain) panic("Failed to add IP3 irqdomain"); if (soc_is_qca955x()) irq_set_chained_handler(ATH79_CPU_IRQ(3), qca955x_ip3_irq_dispatch); else if (soc_is_qca956x() || soc_is_tp9343()) { irq_set_chained_handler(ATH79_CPU_IRQ(3), qca956x_ip3_irq_dispatch); } else if (soc_is_qcn550x()) { irq_set_chained_handler(ATH79_CPU_IRQ(3), qcn550x_ip3_irq_dispatch); } return 0; } IRQCHIP_DECLARE(ath79_ip3_intc, "qca,ath79-ip3-intc", ath79_ip3_irq_of_init); static int __init ar79_cpu_intc_of_init( struct device_node *node, struct device_node *parent) { int err, i, count; /* Fill the irq_wb_chan table */ count = of_count_phandle_with_args( node, "qca,ddr-wb-channels", "#qca,ddr-wb-channel-cells"); for (i = 0; i < count; i++) { struct of_phandle_args args; u32 irq = i; of_property_read_u32_index( node, "qca,ddr-wb-channel-interrupts", i, &irq); if (irq >= ARRAY_SIZE(irq_wb_chan)) continue; err = of_parse_phandle_with_args( node, "qca,ddr-wb-channels", "#qca,ddr-wb-channel-cells", i, &args); if (err) return err; irq_wb_chan[irq] = args.args[0]; pr_info("IRQ: Set flush channel of IRQ%d to %d\n", irq, args.args[0]); } return mips_cpu_irq_of_init(node, parent); } IRQCHIP_DECLARE(ar79_cpu_intc, "qca,ar7100-cpu-intc", ar79_cpu_intc_of_init); #endif static void ath79_ip2_disable(struct irq_data *data) { disable_irq(ATH79_CPU_IRQ(2)); } static void ath79_ip2_enable(struct irq_data *data) { enable_irq(ATH79_CPU_IRQ(2)); } static void ath79_ip3_disable(struct irq_data *data) { disable_irq(ATH79_CPU_IRQ(3)); } static void ath79_ip3_enable(struct irq_data *data) { enable_irq(ATH79_CPU_IRQ(3)); } void __init arch_init_irq(void) { ip2_chip = dummy_irq_chip; ip2_chip.irq_disable = ath79_ip2_disable; ip2_chip.irq_enable = ath79_ip2_enable; ip3_chip = dummy_irq_chip; ip3_chip.irq_disable = ath79_ip3_disable; ip3_chip.irq_enable = ath79_ip3_enable; if (mips_machtype == ATH79_MACH_GENERIC_OF || ath79_mach_is_avm(mips_machtype)) { irqchip_init(); if (soc_is_qca956x() || soc_is_tp9343() || soc_is_qcn550x()) { /* * QCA956x timer init workaround has to be applied * right before setting up the clock. Else, there will * be no jiffies */ late_time_init = &qca956x_qcn550x_enable_timer_cb; } return; } if (soc_is_ar71xx() || soc_is_ar724x() || soc_is_ar913x() || soc_is_ar933x()) { irq_wb_chan[2] = 3; irq_wb_chan[3] = 2; } else if (soc_is_ar934x() || soc_is_qca953x()) { irq_wb_chan[3] = 2; } mips_cpu_irq_init(); ath79_misc_irq_init(); if (soc_is_ar934x()) ar934x_ip2_irq_init(); else if (soc_is_qca953x()) qca953x_irq_init(); else if (soc_is_qca955x()) qca955x_irq_init(); else if (soc_is_qca956x() || soc_is_tp9343()) qca956x_irq_init(); else if (soc_is_qcn550x()) qcn550x_irq_init(); }