1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34#define pr_fmt(fmt) "xen:" KBUILD_MODNAME ": " fmt
35
36#include <linux/linkage.h>
37#include <linux/interrupt.h>
38#include <linux/irq.h>
39#include <linux/smp.h>
40#include <linux/percpu.h>
41#include <linux/cpu.h>
42
43#include <asm/barrier.h>
44#include <asm/sync_bitops.h>
45#include <asm/xen/hypercall.h>
46#include <asm/xen/hypervisor.h>
47
48#include <xen/xen.h>
49#include <xen/xen-ops.h>
50#include <xen/events.h>
51#include <xen/interface/xen.h>
52#include <xen/interface/event_channel.h>
53#include <xen/page.h>
54
55#include "events_internal.h"
56
57#define EVENT_WORDS_PER_PAGE (XEN_PAGE_SIZE / sizeof(event_word_t))
58#define MAX_EVENT_ARRAY_PAGES (EVTCHN_FIFO_NR_CHANNELS / EVENT_WORDS_PER_PAGE)
59
60struct evtchn_fifo_queue {
61 uint32_t head[EVTCHN_FIFO_MAX_QUEUES];
62};
63
64static DEFINE_PER_CPU(struct evtchn_fifo_control_block *, cpu_control_block);
65static DEFINE_PER_CPU(struct evtchn_fifo_queue, cpu_queue);
66static event_word_t *event_array[MAX_EVENT_ARRAY_PAGES] __read_mostly;
67static unsigned event_array_pages __read_mostly;
68
69
70
71
72#if BITS_PER_LONG > 32
73
74#define BM(w) (unsigned long *)((unsigned long)w & ~0x7UL)
75#define EVTCHN_FIFO_BIT(b, w) \
76 (((unsigned long)w & 0x4UL) ? (EVTCHN_FIFO_ ##b + 32) : EVTCHN_FIFO_ ##b)
77
78#else
79
80#define BM(w) ((unsigned long *)(w))
81#define EVTCHN_FIFO_BIT(b, w) EVTCHN_FIFO_ ##b
82
83#endif
84
85static inline event_word_t *event_word_from_port(unsigned port)
86{
87 unsigned i = port / EVENT_WORDS_PER_PAGE;
88
89 return event_array[i] + port % EVENT_WORDS_PER_PAGE;
90}
91
92static unsigned evtchn_fifo_max_channels(void)
93{
94 return EVTCHN_FIFO_NR_CHANNELS;
95}
96
97static unsigned evtchn_fifo_nr_channels(void)
98{
99 return event_array_pages * EVENT_WORDS_PER_PAGE;
100}
101
102static int init_control_block(int cpu,
103 struct evtchn_fifo_control_block *control_block)
104{
105 struct evtchn_fifo_queue *q = &per_cpu(cpu_queue, cpu);
106 struct evtchn_init_control init_control;
107 unsigned int i;
108
109
110 clear_page(control_block);
111 for (i = 0; i < EVTCHN_FIFO_MAX_QUEUES; i++)
112 q->head[i] = 0;
113
114 init_control.control_gfn = virt_to_gfn(control_block);
115 init_control.offset = 0;
116 init_control.vcpu = xen_vcpu_nr(cpu);
117
118 return HYPERVISOR_event_channel_op(EVTCHNOP_init_control, &init_control);
119}
120
121static void free_unused_array_pages(void)
122{
123 unsigned i;
124
125 for (i = event_array_pages; i < MAX_EVENT_ARRAY_PAGES; i++) {
126 if (!event_array[i])
127 break;
128 free_page((unsigned long)event_array[i]);
129 event_array[i] = NULL;
130 }
131}
132
133static void init_array_page(event_word_t *array_page)
134{
135 unsigned i;
136
137 for (i = 0; i < EVENT_WORDS_PER_PAGE; i++)
138 array_page[i] = 1 << EVTCHN_FIFO_MASKED;
139}
140
141static int evtchn_fifo_setup(struct irq_info *info)
142{
143 unsigned port = info->evtchn;
144 unsigned new_array_pages;
145 int ret;
146
147 new_array_pages = port / EVENT_WORDS_PER_PAGE + 1;
148
149 if (new_array_pages > MAX_EVENT_ARRAY_PAGES)
150 return -EINVAL;
151
152 while (event_array_pages < new_array_pages) {
153 void *array_page;
154 struct evtchn_expand_array expand_array;
155
156
157 array_page = event_array[event_array_pages];
158 if (!array_page) {
159 array_page = (void *)__get_free_page(GFP_KERNEL);
160 if (array_page == NULL) {
161 ret = -ENOMEM;
162 goto error;
163 }
164 event_array[event_array_pages] = array_page;
165 }
166
167
168 init_array_page(array_page);
169
170 expand_array.array_gfn = virt_to_gfn(array_page);
171
172 ret = HYPERVISOR_event_channel_op(EVTCHNOP_expand_array, &expand_array);
173 if (ret < 0)
174 goto error;
175
176 event_array_pages++;
177 }
178 return 0;
179
180 error:
181 if (event_array_pages == 0)
182 panic("xen: unable to expand event array with initial page (%d)\n", ret);
183 else
184 pr_err("unable to expand event array (%d)\n", ret);
185 free_unused_array_pages();
186 return ret;
187}
188
189static void evtchn_fifo_bind_to_cpu(struct irq_info *info, unsigned cpu)
190{
191
192}
193
194static void evtchn_fifo_clear_pending(unsigned port)
195{
196 event_word_t *word = event_word_from_port(port);
197 sync_clear_bit(EVTCHN_FIFO_BIT(PENDING, word), BM(word));
198}
199
200static void evtchn_fifo_set_pending(unsigned port)
201{
202 event_word_t *word = event_word_from_port(port);
203 sync_set_bit(EVTCHN_FIFO_BIT(PENDING, word), BM(word));
204}
205
206static bool evtchn_fifo_is_pending(unsigned port)
207{
208 event_word_t *word = event_word_from_port(port);
209 return sync_test_bit(EVTCHN_FIFO_BIT(PENDING, word), BM(word));
210}
211
212static bool evtchn_fifo_test_and_set_mask(unsigned port)
213{
214 event_word_t *word = event_word_from_port(port);
215 return sync_test_and_set_bit(EVTCHN_FIFO_BIT(MASKED, word), BM(word));
216}
217
218static void evtchn_fifo_mask(unsigned port)
219{
220 event_word_t *word = event_word_from_port(port);
221 sync_set_bit(EVTCHN_FIFO_BIT(MASKED, word), BM(word));
222}
223
224static bool evtchn_fifo_is_masked(unsigned port)
225{
226 event_word_t *word = event_word_from_port(port);
227 return sync_test_bit(EVTCHN_FIFO_BIT(MASKED, word), BM(word));
228}
229
230
231
232static void clear_masked(volatile event_word_t *word)
233{
234 event_word_t new, old, w;
235
236 w = *word;
237
238 do {
239 old = w & ~(1 << EVTCHN_FIFO_BUSY);
240 new = old & ~(1 << EVTCHN_FIFO_MASKED);
241 w = sync_cmpxchg(word, old, new);
242 } while (w != old);
243}
244
245static void evtchn_fifo_unmask(unsigned port)
246{
247 event_word_t *word = event_word_from_port(port);
248
249 BUG_ON(!irqs_disabled());
250
251 clear_masked(word);
252 if (evtchn_fifo_is_pending(port)) {
253 struct evtchn_unmask unmask = { .port = port };
254 (void)HYPERVISOR_event_channel_op(EVTCHNOP_unmask, &unmask);
255 }
256}
257
258static uint32_t clear_linked(volatile event_word_t *word)
259{
260 event_word_t new, old, w;
261
262 w = *word;
263
264 do {
265 old = w;
266 new = (w & ~((1 << EVTCHN_FIFO_LINKED)
267 | EVTCHN_FIFO_LINK_MASK));
268 } while ((w = sync_cmpxchg(word, old, new)) != old);
269
270 return w & EVTCHN_FIFO_LINK_MASK;
271}
272
273static void handle_irq_for_port(unsigned port)
274{
275 int irq;
276
277 irq = get_evtchn_to_irq(port);
278 if (irq != -1)
279 generic_handle_irq(irq);
280}
281
282static void consume_one_event(unsigned cpu,
283 struct evtchn_fifo_control_block *control_block,
284 unsigned priority, unsigned long *ready,
285 bool drop)
286{
287 struct evtchn_fifo_queue *q = &per_cpu(cpu_queue, cpu);
288 uint32_t head;
289 unsigned port;
290 event_word_t *word;
291
292 head = q->head[priority];
293
294
295
296
297
298 if (head == 0) {
299 virt_rmb();
300 head = control_block->head[priority];
301 }
302
303 port = head;
304 word = event_word_from_port(port);
305 head = clear_linked(word);
306
307
308
309
310
311
312
313
314 if (head == 0)
315 clear_bit(priority, ready);
316
317 if (evtchn_fifo_is_pending(port) && !evtchn_fifo_is_masked(port)) {
318 if (unlikely(drop))
319 pr_warn("Dropping pending event for port %u\n", port);
320 else
321 handle_irq_for_port(port);
322 }
323
324 q->head[priority] = head;
325}
326
327static void __evtchn_fifo_handle_events(unsigned cpu, bool drop)
328{
329 struct evtchn_fifo_control_block *control_block;
330 unsigned long ready;
331 unsigned q;
332
333 control_block = per_cpu(cpu_control_block, cpu);
334
335 ready = xchg(&control_block->ready, 0);
336
337 while (ready) {
338 q = find_first_bit(&ready, EVTCHN_FIFO_MAX_QUEUES);
339 consume_one_event(cpu, control_block, q, &ready, drop);
340 ready |= xchg(&control_block->ready, 0);
341 }
342}
343
344static void evtchn_fifo_handle_events(unsigned cpu)
345{
346 __evtchn_fifo_handle_events(cpu, false);
347}
348
349static void evtchn_fifo_resume(void)
350{
351 unsigned cpu;
352
353 for_each_possible_cpu(cpu) {
354 void *control_block = per_cpu(cpu_control_block, cpu);
355 int ret;
356
357 if (!control_block)
358 continue;
359
360
361
362
363
364
365 if (!cpu_online(cpu)) {
366 free_page((unsigned long)control_block);
367 per_cpu(cpu_control_block, cpu) = NULL;
368 continue;
369 }
370
371 ret = init_control_block(cpu, control_block);
372 BUG_ON(ret < 0);
373 }
374
375
376
377
378
379
380 event_array_pages = 0;
381}
382
383static const struct evtchn_ops evtchn_ops_fifo = {
384 .max_channels = evtchn_fifo_max_channels,
385 .nr_channels = evtchn_fifo_nr_channels,
386 .setup = evtchn_fifo_setup,
387 .bind_to_cpu = evtchn_fifo_bind_to_cpu,
388 .clear_pending = evtchn_fifo_clear_pending,
389 .set_pending = evtchn_fifo_set_pending,
390 .is_pending = evtchn_fifo_is_pending,
391 .test_and_set_mask = evtchn_fifo_test_and_set_mask,
392 .mask = evtchn_fifo_mask,
393 .unmask = evtchn_fifo_unmask,
394 .handle_events = evtchn_fifo_handle_events,
395 .resume = evtchn_fifo_resume,
396};
397
398static int evtchn_fifo_alloc_control_block(unsigned cpu)
399{
400 void *control_block = NULL;
401 int ret = -ENOMEM;
402
403 control_block = (void *)__get_free_page(GFP_KERNEL);
404 if (control_block == NULL)
405 goto error;
406
407 ret = init_control_block(cpu, control_block);
408 if (ret < 0)
409 goto error;
410
411 per_cpu(cpu_control_block, cpu) = control_block;
412
413 return 0;
414
415 error:
416 free_page((unsigned long)control_block);
417 return ret;
418}
419
420static int xen_evtchn_cpu_prepare(unsigned int cpu)
421{
422 if (!per_cpu(cpu_control_block, cpu))
423 return evtchn_fifo_alloc_control_block(cpu);
424 return 0;
425}
426
427static int xen_evtchn_cpu_dead(unsigned int cpu)
428{
429 __evtchn_fifo_handle_events(cpu, true);
430 return 0;
431}
432
433int __init xen_evtchn_fifo_init(void)
434{
435 int cpu = get_cpu();
436 int ret;
437
438 ret = evtchn_fifo_alloc_control_block(cpu);
439 if (ret < 0)
440 goto out;
441
442 pr_info("Using FIFO-based ABI\n");
443
444 evtchn_ops = &evtchn_ops_fifo;
445
446 cpuhp_setup_state_nocalls(CPUHP_XEN_EVTCHN_PREPARE,
447 "xen/evtchn:prepare",
448 xen_evtchn_cpu_prepare, xen_evtchn_cpu_dead);
449out:
450 put_cpu();
451 return ret;
452}
453