| // SPDX-License-Identifier: GPL-2.0 |
| // Copyright (C) 2018 Hangzhou C-SKY Microsystems co.,ltd. |
| |
| #include <linux/kernel.h> |
| #include <linux/init.h> |
| #include <linux/of.h> |
| #include <linux/of_address.h> |
| #include <linux/module.h> |
| #include <linux/irqdomain.h> |
| #include <linux/irqchip.h> |
| #include <linux/irq.h> |
| #include <linux/interrupt.h> |
| #include <linux/smp.h> |
| #include <linux/io.h> |
| #include <asm/irq.h> |
| #include <asm/traps.h> |
| #include <asm/reg_ops.h> |
| |
| static struct irq_domain *root_domain; |
| static void __iomem *INTCG_base; |
| static void __iomem *INTCL_base; |
| |
| #define IPI_IRQ 15 |
| #define INTC_IRQS 256 |
| #define COMM_IRQ_BASE 32 |
| |
| #define INTCG_SIZE 0x8000 |
| #define INTCL_SIZE 0x1000 |
| |
| #define INTCG_ICTLR 0x0 |
| #define INTCG_CICFGR 0x100 |
| #define INTCG_CIDSTR 0x1000 |
| |
| #define INTCL_PICTLR 0x0 |
| #define INTCL_CFGR 0x14 |
| #define INTCL_SIGR 0x60 |
| #define INTCL_RDYIR 0x6c |
| #define INTCL_SENR 0xa0 |
| #define INTCL_CENR 0xa4 |
| #define INTCL_CACR 0xb4 |
| |
| static DEFINE_PER_CPU(void __iomem *, intcl_reg); |
| |
| static unsigned long *__trigger; |
| |
| #define IRQ_OFFSET(irq) ((irq < COMM_IRQ_BASE) ? irq : (irq - COMM_IRQ_BASE)) |
| |
| #define TRIG_BYTE_OFFSET(i) ((((i) * 2) / 32) * 4) |
| #define TRIG_BIT_OFFSET(i) (((i) * 2) % 32) |
| |
| #define TRIG_VAL(trigger, irq) (trigger << TRIG_BIT_OFFSET(IRQ_OFFSET(irq))) |
| #define TRIG_VAL_MSK(irq) (~(3 << TRIG_BIT_OFFSET(IRQ_OFFSET(irq)))) |
| |
| #define TRIG_BASE(irq) \ |
| (TRIG_BYTE_OFFSET(IRQ_OFFSET(irq)) + ((irq < COMM_IRQ_BASE) ? \ |
| (this_cpu_read(intcl_reg) + INTCL_CFGR) : (INTCG_base + INTCG_CICFGR))) |
| |
| static DEFINE_SPINLOCK(setup_lock); |
| static void setup_trigger(unsigned long irq, unsigned long trigger) |
| { |
| unsigned int tmp; |
| |
| spin_lock(&setup_lock); |
| |
| /* setup trigger */ |
| tmp = readl_relaxed(TRIG_BASE(irq)) & TRIG_VAL_MSK(irq); |
| |
| writel_relaxed(tmp | TRIG_VAL(trigger, irq), TRIG_BASE(irq)); |
| |
| spin_unlock(&setup_lock); |
| } |
| |
| static void csky_mpintc_handler(struct pt_regs *regs) |
| { |
| void __iomem *reg_base = this_cpu_read(intcl_reg); |
| |
| handle_domain_irq(root_domain, |
| readl_relaxed(reg_base + INTCL_RDYIR), regs); |
| } |
| |
| static void csky_mpintc_enable(struct irq_data *d) |
| { |
| void __iomem *reg_base = this_cpu_read(intcl_reg); |
| |
| setup_trigger(d->hwirq, __trigger[d->hwirq]); |
| |
| writel_relaxed(d->hwirq, reg_base + INTCL_SENR); |
| } |
| |
| static void csky_mpintc_disable(struct irq_data *d) |
| { |
| void __iomem *reg_base = this_cpu_read(intcl_reg); |
| |
| writel_relaxed(d->hwirq, reg_base + INTCL_CENR); |
| } |
| |
| static void csky_mpintc_eoi(struct irq_data *d) |
| { |
| void __iomem *reg_base = this_cpu_read(intcl_reg); |
| |
| writel_relaxed(d->hwirq, reg_base + INTCL_CACR); |
| } |
| |
| static int csky_mpintc_set_type(struct irq_data *d, unsigned int type) |
| { |
| switch (type & IRQ_TYPE_SENSE_MASK) { |
| case IRQ_TYPE_LEVEL_HIGH: |
| __trigger[d->hwirq] = 0; |
| break; |
| case IRQ_TYPE_LEVEL_LOW: |
| __trigger[d->hwirq] = 1; |
| break; |
| case IRQ_TYPE_EDGE_RISING: |
| __trigger[d->hwirq] = 2; |
| break; |
| case IRQ_TYPE_EDGE_FALLING: |
| __trigger[d->hwirq] = 3; |
| break; |
| default: |
| return -EINVAL; |
| } |
| |
| return 0; |
| } |
| |
| #ifdef CONFIG_SMP |
| static int csky_irq_set_affinity(struct irq_data *d, |
| const struct cpumask *mask_val, |
| bool force) |
| { |
| unsigned int cpu; |
| unsigned int offset = 4 * (d->hwirq - COMM_IRQ_BASE); |
| |
| if (!force) |
| cpu = cpumask_any_and(mask_val, cpu_online_mask); |
| else |
| cpu = cpumask_first(mask_val); |
| |
| if (cpu >= nr_cpu_ids) |
| return -EINVAL; |
| |
| /* |
| * The csky,mpintc could support auto irq deliver, but it only |
| * could deliver external irq to one cpu or all cpus. So it |
| * doesn't support deliver external irq to a group of cpus |
| * with cpu_mask. |
| * SO we only use auto deliver mode when affinity mask_val is |
| * equal to cpu_present_mask. |
| * |
| */ |
| if (cpumask_equal(mask_val, cpu_present_mask)) |
| cpu = 0; |
| else |
| cpu |= BIT(31); |
| |
| writel_relaxed(cpu, INTCG_base + INTCG_CIDSTR + offset); |
| |
| irq_data_update_effective_affinity(d, cpumask_of(cpu)); |
| |
| return IRQ_SET_MASK_OK_DONE; |
| } |
| #endif |
| |
| static struct irq_chip csky_irq_chip = { |
| .name = "C-SKY SMP Intc", |
| .irq_eoi = csky_mpintc_eoi, |
| .irq_enable = csky_mpintc_enable, |
| .irq_disable = csky_mpintc_disable, |
| .irq_set_type = csky_mpintc_set_type, |
| #ifdef CONFIG_SMP |
| .irq_set_affinity = csky_irq_set_affinity, |
| #endif |
| }; |
| |
| static int csky_irqdomain_map(struct irq_domain *d, unsigned int irq, |
| irq_hw_number_t hwirq) |
| { |
| if (hwirq < COMM_IRQ_BASE) { |
| irq_set_percpu_devid(irq); |
| irq_set_chip_and_handler(irq, &csky_irq_chip, |
| handle_percpu_irq); |
| } else { |
| irq_set_chip_and_handler(irq, &csky_irq_chip, |
| handle_fasteoi_irq); |
| } |
| |
| return 0; |
| } |
| |
| static int csky_irq_domain_xlate_cells(struct irq_domain *d, |
| struct device_node *ctrlr, const u32 *intspec, |
| unsigned int intsize, unsigned long *out_hwirq, |
| unsigned int *out_type) |
| { |
| if (WARN_ON(intsize < 1)) |
| return -EINVAL; |
| |
| *out_hwirq = intspec[0]; |
| if (intsize > 1) |
| *out_type = intspec[1] & IRQ_TYPE_SENSE_MASK; |
| else |
| *out_type = IRQ_TYPE_LEVEL_HIGH; |
| |
| return 0; |
| } |
| |
| static const struct irq_domain_ops csky_irqdomain_ops = { |
| .map = csky_irqdomain_map, |
| .xlate = csky_irq_domain_xlate_cells, |
| }; |
| |
| #ifdef CONFIG_SMP |
| static void csky_mpintc_send_ipi(const struct cpumask *mask) |
| { |
| void __iomem *reg_base = this_cpu_read(intcl_reg); |
| |
| /* |
| * INTCL_SIGR[3:0] INTID |
| * INTCL_SIGR[8:15] CPUMASK |
| */ |
| writel_relaxed((*cpumask_bits(mask)) << 8 | IPI_IRQ, |
| reg_base + INTCL_SIGR); |
| } |
| #endif |
| |
| /* C-SKY multi processor interrupt controller */ |
| static int __init |
| csky_mpintc_init(struct device_node *node, struct device_node *parent) |
| { |
| int ret; |
| unsigned int cpu, nr_irq; |
| #ifdef CONFIG_SMP |
| unsigned int ipi_irq; |
| #endif |
| |
| if (parent) |
| return 0; |
| |
| ret = of_property_read_u32(node, "csky,num-irqs", &nr_irq); |
| if (ret < 0) |
| nr_irq = INTC_IRQS; |
| |
| __trigger = kcalloc(nr_irq, sizeof(unsigned long), GFP_KERNEL); |
| if (__trigger == NULL) |
| return -ENXIO; |
| |
| if (INTCG_base == NULL) { |
| INTCG_base = ioremap(mfcr("cr<31, 14>"), |
| INTCL_SIZE*nr_cpu_ids + INTCG_SIZE); |
| if (INTCG_base == NULL) |
| return -EIO; |
| |
| INTCL_base = INTCG_base + INTCG_SIZE; |
| |
| writel_relaxed(BIT(0), INTCG_base + INTCG_ICTLR); |
| } |
| |
| root_domain = irq_domain_add_linear(node, nr_irq, &csky_irqdomain_ops, |
| NULL); |
| if (!root_domain) |
| return -ENXIO; |
| |
| /* for every cpu */ |
| for_each_present_cpu(cpu) { |
| per_cpu(intcl_reg, cpu) = INTCL_base + (INTCL_SIZE * cpu); |
| writel_relaxed(BIT(0), per_cpu(intcl_reg, cpu) + INTCL_PICTLR); |
| } |
| |
| set_handle_irq(&csky_mpintc_handler); |
| |
| #ifdef CONFIG_SMP |
| ipi_irq = irq_create_mapping(root_domain, IPI_IRQ); |
| if (!ipi_irq) |
| return -EIO; |
| |
| set_send_ipi(&csky_mpintc_send_ipi, ipi_irq); |
| #endif |
| |
| return 0; |
| } |
| IRQCHIP_DECLARE(csky_mpintc, "csky,mpintc", csky_mpintc_init); |