1  // SPDX-License-Identifier: GPL-2.0-or-later
2  /*
3   * Copyright 2016 IBM Corporation.
4   */
5  #include <linux/types.h>
6  #include <linux/kernel.h>
7  #include <linux/irq.h>
8  #include <linux/smp.h>
9  #include <linux/interrupt.h>
10  #include <linux/irqdomain.h>
11  #include <linux/cpu.h>
12  #include <linux/of.h>
13  
14  #include <asm/smp.h>
15  #include <asm/irq.h>
16  #include <asm/errno.h>
17  #include <asm/xics.h>
18  #include <asm/io.h>
19  #include <asm/opal.h>
20  #include <asm/kvm_ppc.h>
21  
icp_opal_teardown_cpu(void)22  static void icp_opal_teardown_cpu(void)
23  {
24  	int hw_cpu = hard_smp_processor_id();
25  
26  	/* Clear any pending IPI */
27  	opal_int_set_mfrr(hw_cpu, 0xff);
28  }
29  
icp_opal_flush_ipi(void)30  static void icp_opal_flush_ipi(void)
31  {
32  	/*
33  	 * We take the ipi irq but and never return so we need to EOI the IPI,
34  	 * but want to leave our priority 0.
35  	 *
36  	 * Should we check all the other interrupts too?
37  	 * Should we be flagging idle loop instead?
38  	 * Or creating some task to be scheduled?
39  	 */
40  	if (opal_int_eoi((0x00 << 24) | XICS_IPI) > 0)
41  		force_external_irq_replay();
42  }
43  
icp_opal_get_xirr(void)44  static unsigned int icp_opal_get_xirr(void)
45  {
46  	unsigned int kvm_xirr;
47  	__be32 hw_xirr;
48  	int64_t rc;
49  
50  	/* Handle an interrupt latched by KVM first */
51  	kvm_xirr = kvmppc_get_xics_latch();
52  	if (kvm_xirr)
53  		return kvm_xirr;
54  
55  	/* Then ask OPAL */
56  	rc = opal_int_get_xirr(&hw_xirr, false);
57  	if (rc < 0)
58  		return 0;
59  	return be32_to_cpu(hw_xirr);
60  }
61  
icp_opal_get_irq(void)62  static unsigned int icp_opal_get_irq(void)
63  {
64  	unsigned int xirr;
65  	unsigned int vec;
66  	unsigned int irq;
67  
68  	xirr = icp_opal_get_xirr();
69  	vec = xirr & 0x00ffffff;
70  	if (vec == XICS_IRQ_SPURIOUS)
71  		return 0;
72  
73  	irq = irq_find_mapping(xics_host, vec);
74  	if (likely(irq)) {
75  		xics_push_cppr(vec);
76  		return irq;
77  	}
78  
79  	/* We don't have a linux mapping, so have rtas mask it. */
80  	xics_mask_unknown_vec(vec);
81  
82  	/* We might learn about it later, so EOI it */
83  	if (opal_int_eoi(xirr) > 0)
84  		force_external_irq_replay();
85  
86  	return 0;
87  }
88  
icp_opal_set_cpu_priority(unsigned char cppr)89  static void icp_opal_set_cpu_priority(unsigned char cppr)
90  {
91  	/*
92  	 * Here be dragons. The caller has asked to allow only IPI's and not
93  	 * external interrupts. But OPAL XIVE doesn't support that. So instead
94  	 * of allowing no interrupts allow all. That's still not right, but
95  	 * currently the only caller who does this is xics_migrate_irqs_away()
96  	 * and it works in that case.
97  	 */
98  	if (cppr >= DEFAULT_PRIORITY)
99  		cppr = LOWEST_PRIORITY;
100  
101  	xics_set_base_cppr(cppr);
102  	opal_int_set_cppr(cppr);
103  	iosync();
104  }
105  
icp_opal_eoi(struct irq_data * d)106  static void icp_opal_eoi(struct irq_data *d)
107  {
108  	unsigned int hw_irq = (unsigned int)irqd_to_hwirq(d);
109  	int64_t rc;
110  
111  	iosync();
112  	rc = opal_int_eoi((xics_pop_cppr() << 24) | hw_irq);
113  
114  	/*
115  	 * EOI tells us whether there are more interrupts to fetch.
116  	 *
117  	 * Some HW implementations might not be able to send us another
118  	 * external interrupt in that case, so we force a replay.
119  	 */
120  	if (rc > 0)
121  		force_external_irq_replay();
122  }
123  
124  #ifdef CONFIG_SMP
125  
icp_opal_cause_ipi(int cpu)126  static void icp_opal_cause_ipi(int cpu)
127  {
128  	int hw_cpu = get_hard_smp_processor_id(cpu);
129  
130  	kvmppc_set_host_ipi(cpu);
131  	opal_int_set_mfrr(hw_cpu, IPI_PRIORITY);
132  }
133  
icp_opal_ipi_action(int irq,void * dev_id)134  static irqreturn_t icp_opal_ipi_action(int irq, void *dev_id)
135  {
136  	int cpu = smp_processor_id();
137  
138  	kvmppc_clear_host_ipi(cpu);
139  	opal_int_set_mfrr(get_hard_smp_processor_id(cpu), 0xff);
140  
141  	return smp_ipi_demux();
142  }
143  
144  /*
145   * Called when an interrupt is received on an off-line CPU to
146   * clear the interrupt, so that the CPU can go back to nap mode.
147   */
icp_opal_flush_interrupt(void)148  void icp_opal_flush_interrupt(void)
149  {
150  	unsigned int xirr;
151  	unsigned int vec;
152  
153  	do {
154  		xirr = icp_opal_get_xirr();
155  		vec = xirr & 0x00ffffff;
156  		if (vec == XICS_IRQ_SPURIOUS)
157  			break;
158  		if (vec == XICS_IPI) {
159  			/* Clear pending IPI */
160  			int cpu = smp_processor_id();
161  			kvmppc_clear_host_ipi(cpu);
162  			opal_int_set_mfrr(get_hard_smp_processor_id(cpu), 0xff);
163  		} else {
164  			pr_err("XICS: hw interrupt 0x%x to offline cpu, "
165  			       "disabling\n", vec);
166  			xics_mask_unknown_vec(vec);
167  		}
168  
169  		/* EOI the interrupt */
170  	} while (opal_int_eoi(xirr) > 0);
171  }
172  
173  #endif /* CONFIG_SMP */
174  
175  static const struct icp_ops icp_opal_ops = {
176  	.get_irq	= icp_opal_get_irq,
177  	.eoi		= icp_opal_eoi,
178  	.set_priority	= icp_opal_set_cpu_priority,
179  	.teardown_cpu	= icp_opal_teardown_cpu,
180  	.flush_ipi	= icp_opal_flush_ipi,
181  #ifdef CONFIG_SMP
182  	.ipi_action	= icp_opal_ipi_action,
183  	.cause_ipi	= icp_opal_cause_ipi,
184  #endif
185  };
186  
icp_opal_init(void)187  int __init icp_opal_init(void)
188  {
189  	struct device_node *np;
190  
191  	np = of_find_compatible_node(NULL, NULL, "ibm,opal-intc");
192  	if (!np)
193  		return -ENODEV;
194  
195  	icp_ops = &icp_opal_ops;
196  
197  	printk("XICS: Using OPAL ICP fallbacks\n");
198  
199  	of_node_put(np);
200  	return 0;
201  }
202  
203