1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17#include <linux/cpufreq.h>
18#include <linux/delay.h>
19#include <linux/init.h>
20#include <linux/jiffies.h>
21#include <linux/kthread.h>
22#include <linux/oprofile.h>
23#include <linux/percpu.h>
24#include <linux/smp.h>
25#include <linux/spinlock.h>
26#include <linux/timer.h>
27#include <asm/cell-pmu.h>
28#include <asm/cputable.h>
29#include <asm/firmware.h>
30#include <asm/io.h>
31#include <asm/oprofile_impl.h>
32#include <asm/processor.h>
33#include <asm/prom.h>
34#include <asm/ptrace.h>
35#include <asm/reg.h>
36#include <asm/rtas.h>
37#include <asm/cell-regs.h>
38
39#include "../platforms/cell/interrupt.h"
40#include "cell/pr_util.h"
41
42#define PPU_PROFILING 0
43#define SPU_PROFILING_CYCLES 1
44#define SPU_PROFILING_EVENTS 2
45
46#define SPU_EVENT_NUM_START 4100
47#define SPU_EVENT_NUM_STOP 4399
48#define SPU_PROFILE_EVENT_ADDR 4363
49#define SPU_PROFILE_EVENT_ADDR_MASK_A 0x146
50#define SPU_PROFILE_EVENT_ADDR_MASK_B 0x186
51
52#define NUM_SPUS_PER_NODE 8
53#define SPU_CYCLES_EVENT_NUM 2
54
55#define PPU_CYCLES_EVENT_NUM 1
56#define PPU_CYCLES_GRP_NUM 1
57
58
59#define CBE_COUNT_ALL_CYCLES 0x42800000
60
61#define NUM_THREADS 2
62
63
64#define NUM_DEBUG_BUS_WORDS 4
65#define NUM_INPUT_BUS_WORDS 2
66
67#define MAX_SPU_COUNT 0xFFFFFF
68
69
70
71
72
73#define NUM_INTERVAL_CYC 0xFFFFFFFF - 10
74
75
76
77
78
79
80static unsigned int spu_cycle_reset;
81static unsigned int profiling_mode;
82static int spu_evnt_phys_spu_indx;
83
84struct pmc_cntrl_data {
85 unsigned long vcntr;
86 unsigned long evnts;
87 unsigned long masks;
88 unsigned long enabled;
89};
90
91
92
93
94struct pm_signal {
95 u16 cpu;
96 u16 sub_unit;
97 short int signal_group;
98 u8 bus_word;
99
100
101 u8 bit;
102};
103
104
105
106
107enum {
108 SUBFUNC_RESET = 1,
109 SUBFUNC_ACTIVATE = 2,
110 SUBFUNC_DEACTIVATE = 3,
111
112 PASSTHRU_IGNORE = 0,
113 PASSTHRU_ENABLE = 1,
114 PASSTHRU_DISABLE = 2,
115};
116
117struct pm_cntrl {
118 u16 enable;
119 u16 stop_at_max;
120 u16 trace_mode;
121 u16 freeze;
122 u16 count_mode;
123 u16 spu_addr_trace;
124 u8 trace_buf_ovflw;
125};
126
127static struct {
128 u32 group_control;
129 u32 debug_bus_control;
130 struct pm_cntrl pm_cntrl;
131 u32 pm07_cntrl[NR_PHYS_CTRS];
132} pm_regs;
133
134#define GET_SUB_UNIT(x) ((x & 0x0000f000) >> 12)
135#define GET_BUS_WORD(x) ((x & 0x000000f0) >> 4)
136#define GET_BUS_TYPE(x) ((x & 0x00000300) >> 8)
137#define GET_POLARITY(x) ((x & 0x00000002) >> 1)
138#define GET_COUNT_CYCLES(x) (x & 0x00000001)
139#define GET_INPUT_CONTROL(x) ((x & 0x00000004) >> 2)
140
141static DEFINE_PER_CPU(unsigned long[NR_PHYS_CTRS], pmc_values);
142static unsigned long spu_pm_cnt[MAX_NUMNODES * NUM_SPUS_PER_NODE];
143static struct pmc_cntrl_data pmc_cntrl[NUM_THREADS][NR_PHYS_CTRS];
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168static u32 hdw_thread;
169
170static u32 virt_cntr_inter_mask;
171static struct timer_list timer_virt_cntr;
172static struct timer_list timer_spu_event_swap;
173
174
175
176
177
178
179static struct pm_signal pm_signal[NR_PHYS_CTRS];
180static int pm_rtas_token;
181static int spu_rtas_token;
182
183static u32 reset_value[NR_PHYS_CTRS];
184static int num_counters;
185static int oprofile_running;
186static DEFINE_SPINLOCK(cntr_lock);
187
188static u32 ctr_enabled;
189
190static unsigned char input_bus[NUM_INPUT_BUS_WORDS];
191
192
193
194
195static int
196rtas_ibm_cbe_perftools(int subfunc, int passthru,
197 void *address, unsigned long length)
198{
199 u64 paddr = __pa(address);
200
201 return rtas_call(pm_rtas_token, 5, 1, NULL, subfunc,
202 passthru, paddr >> 32, paddr & 0xffffffff, length);
203}
204
205static void pm_rtas_reset_signals(u32 node)
206{
207 int ret;
208 struct pm_signal pm_signal_local;
209
210
211
212
213
214
215
216
217
218
219
220 pm_signal_local.cpu = node;
221 pm_signal_local.signal_group = 21;
222 pm_signal_local.bus_word = 1;
223 pm_signal_local.sub_unit = 0;
224 pm_signal_local.bit = 0;
225
226 ret = rtas_ibm_cbe_perftools(SUBFUNC_RESET, PASSTHRU_DISABLE,
227 &pm_signal_local,
228 sizeof(struct pm_signal));
229
230 if (unlikely(ret))
231
232
233
234
235
236 printk(KERN_WARNING "%s: rtas returned: %d\n",
237 __func__, ret);
238}
239
240static int pm_rtas_activate_signals(u32 node, u32 count)
241{
242 int ret;
243 int i, j;
244 struct pm_signal pm_signal_local[NR_PHYS_CTRS];
245
246
247
248
249
250
251
252
253
254 i = 0;
255 for (j = 0; j < count; j++) {
256 if (pm_signal[j].signal_group != PPU_CYCLES_GRP_NUM) {
257
258
259 pm_signal_local[i].cpu = node;
260 pm_signal_local[i].signal_group
261 = pm_signal[j].signal_group;
262 pm_signal_local[i].bus_word = pm_signal[j].bus_word;
263 pm_signal_local[i].sub_unit = pm_signal[j].sub_unit;
264 pm_signal_local[i].bit = pm_signal[j].bit;
265 i++;
266 }
267 }
268
269 if (i != 0) {
270 ret = rtas_ibm_cbe_perftools(SUBFUNC_ACTIVATE, PASSTHRU_ENABLE,
271 pm_signal_local,
272 i * sizeof(struct pm_signal));
273
274 if (unlikely(ret)) {
275 printk(KERN_WARNING "%s: rtas returned: %d\n",
276 __func__, ret);
277 return -EIO;
278 }
279 }
280
281 return 0;
282}
283
284
285
286
287static void set_pm_event(u32 ctr, int event, u32 unit_mask)
288{
289 struct pm_signal *p;
290 u32 signal_bit;
291 u32 bus_word, bus_type, count_cycles, polarity, input_control;
292 int j, i;
293
294 if (event == PPU_CYCLES_EVENT_NUM) {
295
296 pm_regs.pm07_cntrl[ctr] = CBE_COUNT_ALL_CYCLES;
297 p = &(pm_signal[ctr]);
298 p->signal_group = PPU_CYCLES_GRP_NUM;
299 p->bus_word = 1;
300 p->sub_unit = 0;
301 p->bit = 0;
302 goto out;
303 } else {
304 pm_regs.pm07_cntrl[ctr] = 0;
305 }
306
307 bus_word = GET_BUS_WORD(unit_mask);
308 bus_type = GET_BUS_TYPE(unit_mask);
309 count_cycles = GET_COUNT_CYCLES(unit_mask);
310 polarity = GET_POLARITY(unit_mask);
311 input_control = GET_INPUT_CONTROL(unit_mask);
312 signal_bit = (event % 100);
313
314 p = &(pm_signal[ctr]);
315
316 p->signal_group = event / 100;
317 p->bus_word = bus_word;
318 p->sub_unit = GET_SUB_UNIT(unit_mask);
319
320 pm_regs.pm07_cntrl[ctr] = 0;
321 pm_regs.pm07_cntrl[ctr] |= PM07_CTR_COUNT_CYCLES(count_cycles);
322 pm_regs.pm07_cntrl[ctr] |= PM07_CTR_POLARITY(polarity);
323 pm_regs.pm07_cntrl[ctr] |= PM07_CTR_INPUT_CONTROL(input_control);
324
325
326
327
328
329
330
331
332
333
334 if (input_control == 0) {
335 if (signal_bit > 31) {
336 signal_bit -= 32;
337 if (bus_word == 0x3)
338 bus_word = 0x2;
339 else if (bus_word == 0xc)
340 bus_word = 0x8;
341 }
342
343 if ((bus_type == 0) && p->signal_group >= 60)
344 bus_type = 2;
345 if ((bus_type == 1) && p->signal_group >= 50)
346 bus_type = 0;
347
348 pm_regs.pm07_cntrl[ctr] |= PM07_CTR_INPUT_MUX(signal_bit);
349 } else {
350 pm_regs.pm07_cntrl[ctr] = 0;
351 p->bit = signal_bit;
352 }
353
354 for (i = 0; i < NUM_DEBUG_BUS_WORDS; i++) {
355 if (bus_word & (1 << i)) {
356 pm_regs.debug_bus_control |=
357 (bus_type << (30 - (2 * i)));
358
359 for (j = 0; j < NUM_INPUT_BUS_WORDS; j++) {
360 if (input_bus[j] == 0xff) {
361 input_bus[j] = i;
362 pm_regs.group_control |=
363 (i << (30 - (2 * j)));
364
365 break;
366 }
367 }
368 }
369 }
370out:
371 ;
372}
373
374static void write_pm_cntrl(int cpu)
375{
376
377
378
379
380
381 u32 val = 0;
382 if (pm_regs.pm_cntrl.enable == 1)
383 val |= CBE_PM_ENABLE_PERF_MON;
384
385 if (pm_regs.pm_cntrl.stop_at_max == 1)
386 val |= CBE_PM_STOP_AT_MAX;
387
388 if (pm_regs.pm_cntrl.trace_mode != 0)
389 val |= CBE_PM_TRACE_MODE_SET(pm_regs.pm_cntrl.trace_mode);
390
391 if (pm_regs.pm_cntrl.trace_buf_ovflw == 1)
392 val |= CBE_PM_TRACE_BUF_OVFLW(pm_regs.pm_cntrl.trace_buf_ovflw);
393 if (pm_regs.pm_cntrl.freeze == 1)
394 val |= CBE_PM_FREEZE_ALL_CTRS;
395
396 val |= CBE_PM_SPU_ADDR_TRACE_SET(pm_regs.pm_cntrl.spu_addr_trace);
397
398
399
400
401
402 val |= CBE_PM_COUNT_MODE_SET(pm_regs.pm_cntrl.count_mode);
403 cbe_write_pm(cpu, pm_control, val);
404}
405
406static inline void
407set_count_mode(u32 kernel, u32 user)
408{
409
410
411
412
413
414 if (kernel) {
415 if (user)
416 pm_regs.pm_cntrl.count_mode = CBE_COUNT_ALL_MODES;
417 else
418 pm_regs.pm_cntrl.count_mode =
419 CBE_COUNT_SUPERVISOR_MODE;
420 } else {
421 if (user)
422 pm_regs.pm_cntrl.count_mode = CBE_COUNT_PROBLEM_MODE;
423 else
424 pm_regs.pm_cntrl.count_mode =
425 CBE_COUNT_HYPERVISOR_MODE;
426 }
427}
428
429static inline void enable_ctr(u32 cpu, u32 ctr, u32 *pm07_cntrl)
430{
431
432 pm07_cntrl[ctr] |= CBE_PM_CTR_ENABLE;
433 cbe_write_pm07_control(cpu, ctr, pm07_cntrl[ctr]);
434}
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455static void cell_virtual_cntr(unsigned long data)
456{
457 int i, prev_hdw_thread, next_hdw_thread;
458 u32 cpu;
459 unsigned long flags;
460
461
462
463
464
465
466 spin_lock_irqsave(&cntr_lock, flags);
467
468 prev_hdw_thread = hdw_thread;
469
470
471 hdw_thread = 1 ^ hdw_thread;
472 next_hdw_thread = hdw_thread;
473
474 pm_regs.group_control = 0;
475 pm_regs.debug_bus_control = 0;
476
477 for (i = 0; i < NUM_INPUT_BUS_WORDS; i++)
478 input_bus[i] = 0xff;
479
480
481
482
483
484 for (i = 0; i < num_counters; i++)
485 set_pm_event(i,
486 pmc_cntrl[next_hdw_thread][i].evnts,
487 pmc_cntrl[next_hdw_thread][i].masks);
488
489
490
491
492
493 for_each_online_cpu(cpu) {
494 if (cbe_get_hw_thread_id(cpu))
495 continue;
496
497
498
499
500
501 cbe_disable_pm(cpu);
502 cbe_disable_pm_interrupts(cpu);
503 for (i = 0; i < num_counters; i++) {
504 per_cpu(pmc_values, cpu + prev_hdw_thread)[i]
505 = cbe_read_ctr(cpu, i);
506
507 if (per_cpu(pmc_values, cpu + next_hdw_thread)[i]
508 == 0xFFFFFFFF)
509
510
511
512
513
514
515
516
517
518
519 cbe_write_ctr(cpu, i, 0xFFFFFFF0);
520 else
521 cbe_write_ctr(cpu, i,
522 per_cpu(pmc_values,
523 cpu +
524 next_hdw_thread)[i]);
525 }
526
527
528
529
530
531
532 for (i = 0; i < num_counters; i++) {
533 if (pmc_cntrl[next_hdw_thread][i].enabled) {
534
535
536
537
538
539 enable_ctr(cpu, i,
540 pm_regs.pm07_cntrl);
541 } else {
542 cbe_write_pm07_control(cpu, i, 0);
543 }
544 }
545
546
547 cbe_enable_pm_interrupts(cpu, next_hdw_thread,
548 virt_cntr_inter_mask);
549 cbe_enable_pm(cpu);
550 }
551
552 spin_unlock_irqrestore(&cntr_lock, flags);
553
554 mod_timer(&timer_virt_cntr, jiffies + HZ / 10);
555}
556
557static void start_virt_cntrs(void)
558{
559 init_timer(&timer_virt_cntr);
560 timer_virt_cntr.function = cell_virtual_cntr;
561 timer_virt_cntr.data = 0UL;
562 timer_virt_cntr.expires = jiffies + HZ / 10;
563 add_timer(&timer_virt_cntr);
564}
565
566static int cell_reg_setup_spu_cycles(struct op_counter_config *ctr,
567 struct op_system_config *sys, int num_ctrs)
568{
569 spu_cycle_reset = ctr[0].count;
570
571
572
573
574
575 spu_rtas_token = rtas_token("ibm,cbe-spu-perftools");
576
577 if (unlikely(spu_rtas_token == RTAS_UNKNOWN_SERVICE)) {
578 printk(KERN_ERR
579 "%s: rtas token ibm,cbe-spu-perftools unknown\n",
580 __func__);
581 return -EIO;
582 }
583 return 0;
584}
585
586
587
588
589
590
591
592
593static void spu_evnt_swap(unsigned long data)
594{
595 int node;
596 int cur_phys_spu, nxt_phys_spu, cur_spu_evnt_phys_spu_indx;
597 unsigned long flags;
598 int cpu;
599 int ret;
600 u32 interrupt_mask;
601
602
603
604 interrupt_mask = CBE_PM_CTR_OVERFLOW_INTR(0);
605
606 hdw_thread = 0;
607
608
609
610
611 spin_lock_irqsave(&cntr_lock, flags);
612
613 cur_spu_evnt_phys_spu_indx = spu_evnt_phys_spu_indx;
614
615 if (++(spu_evnt_phys_spu_indx) == NUM_SPUS_PER_NODE)
616 spu_evnt_phys_spu_indx = 0;
617
618 pm_signal[0].sub_unit = spu_evnt_phys_spu_indx;
619 pm_signal[1].sub_unit = spu_evnt_phys_spu_indx;
620 pm_signal[2].sub_unit = spu_evnt_phys_spu_indx;
621
622
623 for_each_online_cpu(cpu) {
624 if (cbe_get_hw_thread_id(cpu))
625 continue;
626
627 node = cbe_cpu_to_node(cpu);
628 cur_phys_spu = (node * NUM_SPUS_PER_NODE)
629 + cur_spu_evnt_phys_spu_indx;
630 nxt_phys_spu = (node * NUM_SPUS_PER_NODE)
631 + spu_evnt_phys_spu_indx;
632
633
634
635
636
637 cbe_disable_pm(cpu);
638 cbe_disable_pm_interrupts(cpu);
639
640 spu_pm_cnt[cur_phys_spu]
641 = cbe_read_ctr(cpu, 0);
642
643
644
645
646
647 if (spu_pm_cnt[nxt_phys_spu] >= 0xFFFFFFFF)
648 cbe_write_ctr(cpu, 0, 0xFFFFFFF0);
649 else
650 cbe_write_ctr(cpu, 0, spu_pm_cnt[nxt_phys_spu]);
651
652 pm_rtas_reset_signals(cbe_cpu_to_node(cpu));
653
654
655
656
657
658 ret = pm_rtas_activate_signals(cbe_cpu_to_node(cpu), 3);
659 if (ret)
660 printk(KERN_ERR "%s: pm_rtas_activate_signals failed, "
661 "SPU event swap\n", __func__);
662
663
664
665 cbe_write_pm(cpu, trace_address, 0);
666
667 enable_ctr(cpu, 0, pm_regs.pm07_cntrl);
668
669
670 cbe_enable_pm_interrupts(cpu, hdw_thread,
671 interrupt_mask);
672 cbe_enable_pm(cpu);
673 }
674
675 spin_unlock_irqrestore(&cntr_lock, flags);
676
677
678 mod_timer(&timer_spu_event_swap, jiffies + HZ / 25);
679}
680
681static void start_spu_event_swap(void)
682{
683 init_timer(&timer_spu_event_swap);
684 timer_spu_event_swap.function = spu_evnt_swap;
685 timer_spu_event_swap.data = 0UL;
686 timer_spu_event_swap.expires = jiffies + HZ / 25;
687 add_timer(&timer_spu_event_swap);
688}
689
690static int cell_reg_setup_spu_events(struct op_counter_config *ctr,
691 struct op_system_config *sys, int num_ctrs)
692{
693 int i;
694
695
696
697 spu_evnt_phys_spu_indx = 0;
698
699
700
701
702
703
704 pm_rtas_token = rtas_token("ibm,cbe-perftools");
705
706 if (unlikely(pm_rtas_token == RTAS_UNKNOWN_SERVICE)) {
707 printk(KERN_ERR
708 "%s: rtas token ibm,cbe-perftools unknown\n",
709 __func__);
710 return -EIO;
711 }
712
713
714
715
716
717 pm_regs.pm_cntrl.trace_buf_ovflw = 1;
718
719
720
721
722
723 pm_regs.pm_cntrl.trace_mode = 2;
724
725 pm_regs.pm_cntrl.spu_addr_trace = 0x1;
726
727
728
729
730
731 pm_signal[1].signal_group = SPU_PROFILE_EVENT_ADDR / 100;
732 pm_signal[1].bus_word = GET_BUS_WORD(SPU_PROFILE_EVENT_ADDR_MASK_A);
733 pm_signal[1].bit = SPU_PROFILE_EVENT_ADDR % 100;
734 pm_signal[1].sub_unit = spu_evnt_phys_spu_indx;
735
736 pm_signal[2].signal_group = SPU_PROFILE_EVENT_ADDR / 100;
737 pm_signal[2].bus_word = GET_BUS_WORD(SPU_PROFILE_EVENT_ADDR_MASK_B);
738 pm_signal[2].bit = SPU_PROFILE_EVENT_ADDR % 100;
739 pm_signal[2].sub_unit = spu_evnt_phys_spu_indx;
740
741
742
743
744 num_counters = 1;
745 set_pm_event(0, ctr[0].event, ctr[0].unit_mask);
746
747 reset_value[0] = 0xFFFFFFFF - ctr[0].count;
748
749
750 ctr_enabled |= 1;
751
752
753 for (i=0; i < MAX_NUMNODES * NUM_SPUS_PER_NODE; i++)
754 spu_pm_cnt[i] = reset_value[0];
755
756 return 0;
757}
758
759static int cell_reg_setup_ppu(struct op_counter_config *ctr,
760 struct op_system_config *sys, int num_ctrs)
761{
762
763 int i, j, cpu;
764
765 num_counters = num_ctrs;
766
767 if (unlikely(num_ctrs > NR_PHYS_CTRS)) {
768 printk(KERN_ERR
769 "%s: Oprofile, number of specified events " \
770 "exceeds number of physical counters\n",
771 __func__);
772 return -EIO;
773 }
774
775 set_count_mode(sys->enable_kernel, sys->enable_user);
776
777
778 for (i = 0; i < num_ctrs; ++i) {
779
780 pmc_cntrl[0][i].evnts = ctr[i].event;
781 pmc_cntrl[0][i].masks = ctr[i].unit_mask;
782 pmc_cntrl[0][i].enabled = ctr[i].enabled;
783 pmc_cntrl[0][i].vcntr = i;
784
785 for_each_possible_cpu(j)
786 per_cpu(pmc_values, j)[i] = 0;
787 }
788
789
790
791
792
793 for (i = 0; i < num_ctrs; ++i) {
794 if ((ctr[i].event >= 2100) && (ctr[i].event <= 2111))
795 pmc_cntrl[1][i].evnts = ctr[i].event + 19;
796 else if (ctr[i].event == 2203)
797 pmc_cntrl[1][i].evnts = ctr[i].event;
798 else if ((ctr[i].event >= 2200) && (ctr[i].event <= 2215))
799 pmc_cntrl[1][i].evnts = ctr[i].event + 16;
800 else
801 pmc_cntrl[1][i].evnts = ctr[i].event;
802
803 pmc_cntrl[1][i].masks = ctr[i].unit_mask;
804 pmc_cntrl[1][i].enabled = ctr[i].enabled;
805 pmc_cntrl[1][i].vcntr = i;
806 }
807
808 for (i = 0; i < NUM_INPUT_BUS_WORDS; i++)
809 input_bus[i] = 0xff;
810
811
812
813
814
815
816
817
818 for (i = 0; i < num_counters; ++i) {
819
820 if (pmc_cntrl[0][i].enabled) {
821
822 reset_value[i] = 0xFFFFFFFF - ctr[i].count;
823 set_pm_event(i,
824 pmc_cntrl[0][i].evnts,
825 pmc_cntrl[0][i].masks);
826
827
828 ctr_enabled |= (1 << i);
829 }
830 }
831
832
833 for_each_online_cpu(cpu)
834 for (i = 0; i < num_counters; ++i) {
835 per_cpu(pmc_values, cpu)[i] = reset_value[i];
836 }
837
838 return 0;
839}
840
841
842
843static int cell_reg_setup(struct op_counter_config *ctr,
844 struct op_system_config *sys, int num_ctrs)
845{
846 int ret=0;
847 spu_cycle_reset = 0;
848
849
850
851
852 pm_regs.group_control = 0;
853 pm_regs.debug_bus_control = 0;
854 pm_regs.pm_cntrl.stop_at_max = 1;
855 pm_regs.pm_cntrl.trace_mode = 0;
856 pm_regs.pm_cntrl.freeze = 1;
857 pm_regs.pm_cntrl.trace_buf_ovflw = 0;
858 pm_regs.pm_cntrl.spu_addr_trace = 0;
859
860
861
862
863
864
865
866 pm_rtas_token = rtas_token("ibm,cbe-perftools");
867
868 if (unlikely(pm_rtas_token == RTAS_UNKNOWN_SERVICE)) {
869 printk(KERN_ERR
870 "%s: rtas token ibm,cbe-perftools unknown\n",
871 __func__);
872 return -EIO;
873 }
874
875 if (ctr[0].event == SPU_CYCLES_EVENT_NUM) {
876 profiling_mode = SPU_PROFILING_CYCLES;
877 ret = cell_reg_setup_spu_cycles(ctr, sys, num_ctrs);
878 } else if ((ctr[0].event >= SPU_EVENT_NUM_START) &&
879 (ctr[0].event <= SPU_EVENT_NUM_STOP)) {
880 profiling_mode = SPU_PROFILING_EVENTS;
881 spu_cycle_reset = ctr[0].count;
882
883
884
885
886
887
888
889 cell_reg_setup_spu_events(ctr, sys, num_ctrs);
890 } else {
891 profiling_mode = PPU_PROFILING;
892 ret = cell_reg_setup_ppu(ctr, sys, num_ctrs);
893 }
894
895 return ret;
896}
897
898
899
900
901static int cell_cpu_setup(struct op_counter_config *cntr)
902{
903 u32 cpu = smp_processor_id();
904 u32 num_enabled = 0;
905 int i;
906 int ret;
907
908
909
910
911
912 if (profiling_mode == SPU_PROFILING_CYCLES)
913 return 0;
914
915
916
917
918 if (cbe_get_hw_thread_id(cpu))
919 return 0;
920
921
922 cbe_disable_pm(cpu);
923 cbe_disable_pm_interrupts(cpu);
924
925 cbe_write_pm(cpu, pm_start_stop, 0);
926 cbe_write_pm(cpu, group_control, pm_regs.group_control);
927 cbe_write_pm(cpu, debug_bus_control, pm_regs.debug_bus_control);
928 write_pm_cntrl(cpu);
929
930 for (i = 0; i < num_counters; ++i) {
931 if (ctr_enabled & (1 << i)) {
932 pm_signal[num_enabled].cpu = cbe_cpu_to_node(cpu);
933 num_enabled++;
934 }
935 }
936
937
938
939
940
941 if (profiling_mode == SPU_PROFILING_EVENTS) {
942
943
944
945 ret = pm_rtas_activate_signals(cbe_cpu_to_node(cpu),
946 num_enabled+2);
947
948
949
950 cbe_write_pm(cpu, pm_interval, NUM_INTERVAL_CYC);
951 return ret;
952 } else
953 return pm_rtas_activate_signals(cbe_cpu_to_node(cpu),
954 num_enabled);
955}
956
957#define ENTRIES 303
958#define MAXLFSR 0xFFFFFF
959
960
961static int initial_lfsr[] = {
962 8221349, 12579195, 5379618, 10097839, 7512963, 7519310, 3955098, 10753424,
963 15507573, 7458917, 285419, 2641121, 9780088, 3915503, 6668768, 1548716,
964 4885000, 8774424, 9650099, 2044357, 2304411, 9326253, 10332526, 4421547,
965 3440748, 10179459, 13332843, 10375561, 1313462, 8375100, 5198480, 6071392,
966 9341783, 1526887, 3985002, 1439429, 13923762, 7010104, 11969769, 4547026,
967 2040072, 4025602, 3437678, 7939992, 11444177, 4496094, 9803157, 10745556,
968 3671780, 4257846, 5662259, 13196905, 3237343, 12077182, 16222879, 7587769,
969 14706824, 2184640, 12591135, 10420257, 7406075, 3648978, 11042541, 15906893,
970 11914928, 4732944, 10695697, 12928164, 11980531, 4430912, 11939291, 2917017,
971 6119256, 4172004, 9373765, 8410071, 14788383, 5047459, 5474428, 1737756,
972 15967514, 13351758, 6691285, 8034329, 2856544, 14394753, 11310160, 12149558,
973 7487528, 7542781, 15668898, 12525138, 12790975, 3707933, 9106617, 1965401,
974 16219109, 12801644, 2443203, 4909502, 8762329, 3120803, 6360315, 9309720,
975 15164599, 10844842, 4456529, 6667610, 14924259, 884312, 6234963, 3326042,
976 15973422, 13919464, 5272099, 6414643, 3909029, 2764324, 5237926, 4774955,
977 10445906, 4955302, 5203726, 10798229, 11443419, 2303395, 333836, 9646934,
978 3464726, 4159182, 568492, 995747, 10318756, 13299332, 4836017, 8237783,
979 3878992, 2581665, 11394667, 5672745, 14412947, 3159169, 9094251, 16467278,
980 8671392, 15230076, 4843545, 7009238, 15504095, 1494895, 9627886, 14485051,
981 8304291, 252817, 12421642, 16085736, 4774072, 2456177, 4160695, 15409741,
982 4902868, 5793091, 13162925, 16039714, 782255, 11347835, 14884586, 366972,
983 16308990, 11913488, 13390465, 2958444, 10340278, 1177858, 1319431, 10426302,
984 2868597, 126119, 5784857, 5245324, 10903900, 16436004, 3389013, 1742384,
985 14674502, 10279218, 8536112, 10364279, 6877778, 14051163, 1025130, 6072469,
986 1988305, 8354440, 8216060, 16342977, 13112639, 3976679, 5913576, 8816697,
987 6879995, 14043764, 3339515, 9364420, 15808858, 12261651, 2141560, 5636398,
988 10345425, 10414756, 781725, 6155650, 4746914, 5078683, 7469001, 6799140,
989 10156444, 9667150, 10116470, 4133858, 2121972, 1124204, 1003577, 1611214,
990 14304602, 16221850, 13878465, 13577744, 3629235, 8772583, 10881308, 2410386,
991 7300044, 5378855, 9301235, 12755149, 4977682, 8083074, 10327581, 6395087,
992 9155434, 15501696, 7514362, 14520507, 15808945, 3244584, 4741962, 9658130,
993 14336147, 8654727, 7969093, 15759799, 14029445, 5038459, 9894848, 8659300,
994 13699287, 8834306, 10712885, 14753895, 10410465, 3373251, 309501, 9561475,
995 5526688, 14647426, 14209836, 5339224, 207299, 14069911, 8722990, 2290950,
996 3258216, 12505185, 6007317, 9218111, 14661019, 10537428, 11731949, 9027003,
997 6641507, 9490160, 200241, 9720425, 16277895, 10816638, 1554761, 10431375,
998 7467528, 6790302, 3429078, 14633753, 14428997, 11463204, 3576212, 2003426,
999 6123687, 820520, 9992513, 15784513, 5778891, 6428165, 8388607
1000};
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055#define V2_16 (0x1 << 16)
1056#define V2_19 (0x1 << 19)
1057#define V2_22 (0x1 << 22)
1058
1059static int calculate_lfsr(int n)
1060{
1061
1062
1063
1064
1065 int index;
1066
1067 if ((n >> 16) == 0)
1068 index = 0;
1069 else if (((n - V2_16) >> 19) == 0)
1070 index = ((n - V2_16) >> 12) + 1;
1071 else if (((n - V2_16 - V2_19) >> 22) == 0)
1072 index = ((n - V2_16 - V2_19) >> 15 ) + 1 + 128;
1073 else if (((n - V2_16 - V2_19 - V2_22) >> 24) == 0)
1074 index = ((n - V2_16 - V2_19 - V2_22) >> 18 ) + 1 + 256;
1075 else
1076 index = ENTRIES-1;
1077
1078
1079 if ((index >= ENTRIES) || (index < 0))
1080 index = ENTRIES-1;
1081
1082 return initial_lfsr[index];
1083}
1084
1085static int pm_rtas_activate_spu_profiling(u32 node)
1086{
1087 int ret, i;
1088 struct pm_signal pm_signal_local[NUM_SPUS_PER_NODE];
1089
1090
1091
1092
1093
1094 for (i = 0; i < ARRAY_SIZE(pm_signal_local); i++) {
1095 pm_signal_local[i].cpu = node;
1096 pm_signal_local[i].signal_group = 41;
1097
1098 pm_signal_local[i].bus_word = 1 << i / 2;
1099
1100 pm_signal_local[i].sub_unit = i;
1101 pm_signal_local[i].bit = 63;
1102 }
1103
1104 ret = rtas_ibm_cbe_perftools(SUBFUNC_ACTIVATE,
1105 PASSTHRU_ENABLE, pm_signal_local,
1106 (ARRAY_SIZE(pm_signal_local)
1107 * sizeof(struct pm_signal)));
1108
1109 if (unlikely(ret)) {
1110 printk(KERN_WARNING "%s: rtas returned: %d\n",
1111 __func__, ret);
1112 return -EIO;
1113 }
1114
1115 return 0;
1116}
1117
1118#ifdef CONFIG_CPU_FREQ
1119static int
1120oprof_cpufreq_notify(struct notifier_block *nb, unsigned long val, void *data)
1121{
1122 int ret = 0;
1123 struct cpufreq_freqs *frq = data;
1124 if ((val == CPUFREQ_PRECHANGE && frq->old < frq->new) ||
1125 (val == CPUFREQ_POSTCHANGE && frq->old > frq->new) ||
1126 (val == CPUFREQ_RESUMECHANGE || val == CPUFREQ_SUSPENDCHANGE))
1127 set_spu_profiling_frequency(frq->new, spu_cycle_reset);
1128 return ret;
1129}
1130
1131static struct notifier_block cpu_freq_notifier_block = {
1132 .notifier_call = oprof_cpufreq_notify
1133};
1134#endif
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144static void cell_global_stop_spu_cycles(void)
1145{
1146 int subfunc, rtn_value;
1147 unsigned int lfsr_value;
1148 int cpu;
1149
1150 oprofile_running = 0;
1151 smp_wmb();
1152
1153#ifdef CONFIG_CPU_FREQ
1154 cpufreq_unregister_notifier(&cpu_freq_notifier_block,
1155 CPUFREQ_TRANSITION_NOTIFIER);
1156#endif
1157
1158 for_each_online_cpu(cpu) {
1159 if (cbe_get_hw_thread_id(cpu))
1160 continue;
1161
1162 subfunc = 3;
1163
1164
1165
1166 lfsr_value = 0x8f100000;
1167
1168 rtn_value = rtas_call(spu_rtas_token, 3, 1, NULL,
1169 subfunc, cbe_cpu_to_node(cpu),
1170 lfsr_value);
1171
1172 if (unlikely(rtn_value != 0)) {
1173 printk(KERN_ERR
1174 "%s: rtas call ibm,cbe-spu-perftools " \
1175 "failed, return = %d\n",
1176 __func__, rtn_value);
1177 }
1178
1179
1180 pm_rtas_reset_signals(cbe_cpu_to_node(cpu));
1181 }
1182
1183 stop_spu_profiling_cycles();
1184}
1185
1186static void cell_global_stop_spu_events(void)
1187{
1188 int cpu;
1189 oprofile_running = 0;
1190
1191 stop_spu_profiling_events();
1192 smp_wmb();
1193
1194 for_each_online_cpu(cpu) {
1195 if (cbe_get_hw_thread_id(cpu))
1196 continue;
1197
1198 cbe_sync_irq(cbe_cpu_to_node(cpu));
1199
1200 cbe_disable_pm(cpu);
1201 cbe_write_pm07_control(cpu, 0, 0);
1202
1203
1204 pm_rtas_reset_signals(cbe_cpu_to_node(cpu));
1205
1206
1207 cbe_disable_pm_interrupts(cpu);
1208 }
1209 del_timer_sync(&timer_spu_event_swap);
1210}
1211
1212static void cell_global_stop_ppu(void)
1213{
1214 int cpu;
1215
1216
1217
1218
1219
1220
1221 del_timer_sync(&timer_virt_cntr);
1222 oprofile_running = 0;
1223 smp_wmb();
1224
1225 for_each_online_cpu(cpu) {
1226 if (cbe_get_hw_thread_id(cpu))
1227 continue;
1228
1229 cbe_sync_irq(cbe_cpu_to_node(cpu));
1230
1231 cbe_disable_pm(cpu);
1232
1233
1234 pm_rtas_reset_signals(cbe_cpu_to_node(cpu));
1235
1236
1237 cbe_disable_pm_interrupts(cpu);
1238 }
1239}
1240
1241static void cell_global_stop(void)
1242{
1243 if (profiling_mode == PPU_PROFILING)
1244 cell_global_stop_ppu();
1245 else if (profiling_mode == SPU_PROFILING_EVENTS)
1246 cell_global_stop_spu_events();
1247 else
1248 cell_global_stop_spu_cycles();
1249}
1250
1251static int cell_global_start_spu_cycles(struct op_counter_config *ctr)
1252{
1253 int subfunc;
1254 unsigned int lfsr_value;
1255 int cpu;
1256 int ret;
1257 int rtas_error;
1258 unsigned int cpu_khzfreq = 0;
1259
1260
1261
1262
1263
1264
1265#ifdef CONFIG_CPU_FREQ
1266 ret = cpufreq_register_notifier(&cpu_freq_notifier_block,
1267 CPUFREQ_TRANSITION_NOTIFIER);
1268 if (ret < 0)
1269
1270 printk(KERN_ERR "CPU freq change registration failed: %d\n",
1271 ret);
1272
1273 else
1274 cpu_khzfreq = cpufreq_quick_get(smp_processor_id());
1275#endif
1276
1277 set_spu_profiling_frequency(cpu_khzfreq, spu_cycle_reset);
1278
1279 for_each_online_cpu(cpu) {
1280 if (cbe_get_hw_thread_id(cpu))
1281 continue;
1282
1283
1284
1285
1286
1287
1288 cbe_write_pm(cpu, pm_control, 0);
1289
1290 if (spu_cycle_reset > MAX_SPU_COUNT)
1291
1292 lfsr_value = calculate_lfsr(MAX_SPU_COUNT-1);
1293 else
1294 lfsr_value = calculate_lfsr(spu_cycle_reset);
1295
1296
1297 if (lfsr_value == 0)
1298 lfsr_value = calculate_lfsr(1);
1299
1300 lfsr_value = lfsr_value << 8;
1301
1302
1303
1304
1305 ret = pm_rtas_activate_spu_profiling(cbe_cpu_to_node(cpu));
1306
1307 if (unlikely(ret)) {
1308 rtas_error = ret;
1309 goto out;
1310 }
1311
1312
1313 subfunc = 2;
1314
1315
1316 ret = rtas_call(spu_rtas_token, 3, 1, NULL, subfunc,
1317 cbe_cpu_to_node(cpu), lfsr_value);
1318
1319 if (unlikely(ret != 0)) {
1320 printk(KERN_ERR
1321 "%s: rtas call ibm,cbe-spu-perftools failed, " \
1322 "return = %d\n", __func__, ret);
1323 rtas_error = -EIO;
1324 goto out;
1325 }
1326 }
1327
1328 rtas_error = start_spu_profiling_cycles(spu_cycle_reset);
1329 if (rtas_error)
1330 goto out_stop;
1331
1332 oprofile_running = 1;
1333 return 0;
1334
1335out_stop:
1336 cell_global_stop_spu_cycles();
1337out:
1338 return rtas_error;
1339}
1340
1341static int cell_global_start_spu_events(struct op_counter_config *ctr)
1342{
1343 int cpu;
1344 u32 interrupt_mask = 0;
1345 int rtn = 0;
1346
1347 hdw_thread = 0;
1348
1349
1350
1351
1352
1353
1354
1355
1356
1357
1358
1359
1360
1361
1362
1363 for_each_online_cpu(cpu) {
1364 if (cbe_get_hw_thread_id(cpu))
1365 continue;
1366
1367
1368
1369
1370
1371
1372
1373
1374 if (ctr_enabled & 1) {
1375 cbe_write_ctr(cpu, 0, reset_value[0]);
1376 enable_ctr(cpu, 0, pm_regs.pm07_cntrl);
1377 interrupt_mask |=
1378 CBE_PM_CTR_OVERFLOW_INTR(0);
1379 } else {
1380
1381 cbe_write_pm07_control(cpu, 0, 0);
1382 }
1383
1384 cbe_get_and_clear_pm_interrupts(cpu);
1385 cbe_enable_pm_interrupts(cpu, hdw_thread, interrupt_mask);
1386 cbe_enable_pm(cpu);
1387
1388
1389 cbe_write_pm(cpu, trace_address, 0);
1390 }
1391
1392
1393
1394
1395
1396 start_spu_event_swap();
1397 start_spu_profiling_events();
1398 oprofile_running = 1;
1399 smp_wmb();
1400
1401 return rtn;
1402}
1403
1404static int cell_global_start_ppu(struct op_counter_config *ctr)
1405{
1406 u32 cpu, i;
1407 u32 interrupt_mask = 0;
1408
1409
1410
1411
1412
1413 for_each_online_cpu(cpu) {
1414 if (cbe_get_hw_thread_id(cpu))
1415 continue;
1416
1417 interrupt_mask = 0;
1418
1419 for (i = 0; i < num_counters; ++i) {
1420 if (ctr_enabled & (1 << i)) {
1421 cbe_write_ctr(cpu, i, reset_value[i]);
1422 enable_ctr(cpu, i, pm_regs.pm07_cntrl);
1423 interrupt_mask |= CBE_PM_CTR_OVERFLOW_INTR(i);
1424 } else {
1425
1426 cbe_write_pm07_control(cpu, i, 0);
1427 }
1428 }
1429
1430 cbe_get_and_clear_pm_interrupts(cpu);
1431 cbe_enable_pm_interrupts(cpu, hdw_thread, interrupt_mask);
1432 cbe_enable_pm(cpu);
1433 }
1434
1435 virt_cntr_inter_mask = interrupt_mask;
1436 oprofile_running = 1;
1437 smp_wmb();
1438
1439
1440
1441
1442
1443
1444
1445 start_virt_cntrs();
1446
1447 return 0;
1448}
1449
1450static int cell_global_start(struct op_counter_config *ctr)
1451{
1452 if (profiling_mode == SPU_PROFILING_CYCLES)
1453 return cell_global_start_spu_cycles(ctr);
1454 else if (profiling_mode == SPU_PROFILING_EVENTS)
1455 return cell_global_start_spu_events(ctr);
1456 else
1457 return cell_global_start_ppu(ctr);
1458}
1459
1460
1461
1462
1463
1464
1465
1466
1467
1468
1469
1470
1471
1472
1473
1474
1475
1476
1477
1478
1479
1480
1481
1482
1483
1484
1485
1486
1487
1488static void cell_handle_interrupt_spu(struct pt_regs *regs,
1489 struct op_counter_config *ctr)
1490{
1491 u32 cpu, cpu_tmp;
1492 u64 trace_entry;
1493 u32 interrupt_mask;
1494 u64 trace_buffer[2];
1495 u64 last_trace_buffer;
1496 u32 sample;
1497 u32 trace_addr;
1498 unsigned long sample_array_lock_flags;
1499 int spu_num;
1500 unsigned long flags;
1501
1502
1503
1504
1505 cpu = smp_processor_id();
1506 spin_lock_irqsave(&cntr_lock, flags);
1507
1508 cpu_tmp = cpu;
1509 cbe_disable_pm(cpu);
1510
1511 interrupt_mask = cbe_get_and_clear_pm_interrupts(cpu);
1512
1513 sample = 0xABCDEF;
1514 trace_entry = 0xfedcba;
1515 last_trace_buffer = 0xdeadbeaf;
1516
1517 if ((oprofile_running == 1) && (interrupt_mask != 0)) {
1518
1519 cbe_write_pm(cpu, pm_interval, 0);
1520
1521
1522 if ((interrupt_mask & CBE_PM_CTR_OVERFLOW_INTR(0))
1523 && ctr[0].enabled)
1524
1525
1526
1527
1528 cbe_write_ctr(cpu, 0, reset_value[0]);
1529
1530 trace_addr = cbe_read_pm(cpu, trace_address);
1531
1532 while (!(trace_addr & CBE_PM_TRACE_BUF_EMPTY)) {
1533
1534
1535
1536
1537
1538 cbe_read_trace_buffer(cpu, trace_buffer);
1539 trace_addr = cbe_read_pm(cpu, trace_address);
1540 }
1541
1542
1543
1544
1545
1546
1547
1548
1549
1550
1551
1552
1553
1554
1555
1556
1557 trace_entry = trace_buffer[0]
1558 & 0x00000000FFFF0000;
1559
1560
1561
1562
1563 sample = trace_entry >> 14;
1564 last_trace_buffer = trace_buffer[0];
1565
1566 spu_num = spu_evnt_phys_spu_indx
1567 + (cbe_cpu_to_node(cpu) * NUM_SPUS_PER_NODE);
1568
1569
1570
1571
1572 spin_lock_irqsave(&oprof_spu_smpl_arry_lck,
1573 sample_array_lock_flags);
1574 spu_sync_buffer(spu_num, &sample, 1);
1575 spin_unlock_irqrestore(&oprof_spu_smpl_arry_lck,
1576 sample_array_lock_flags);
1577
1578 smp_wmb();
1579
1580
1581
1582
1583
1584 cbe_write_pm(cpu, pm_interval, NUM_INTERVAL_CYC);
1585 cbe_enable_pm_interrupts(cpu, hdw_thread,
1586 virt_cntr_inter_mask);
1587
1588
1589 cbe_write_pm(cpu, trace_address, 0);
1590 cbe_write_pm(cpu, pm_interval, NUM_INTERVAL_CYC);
1591
1592
1593
1594
1595
1596
1597
1598
1599
1600 write_pm_cntrl(cpu);
1601 cbe_enable_pm(cpu);
1602 }
1603 spin_unlock_irqrestore(&cntr_lock, flags);
1604}
1605
1606static void cell_handle_interrupt_ppu(struct pt_regs *regs,
1607 struct op_counter_config *ctr)
1608{
1609 u32 cpu;
1610 u64 pc;
1611 int is_kernel;
1612 unsigned long flags = 0;
1613 u32 interrupt_mask;
1614 int i;
1615
1616 cpu = smp_processor_id();
1617
1618
1619
1620
1621
1622
1623 spin_lock_irqsave(&cntr_lock, flags);
1624
1625
1626
1627
1628
1629
1630
1631 cbe_disable_pm(cpu);
1632
1633 interrupt_mask = cbe_get_and_clear_pm_interrupts(cpu);
1634
1635
1636
1637
1638
1639
1640
1641
1642 if ((oprofile_running == 1) && (interrupt_mask != 0)) {
1643 pc = regs->nip;
1644 is_kernel = is_kernel_addr(pc);
1645
1646 for (i = 0; i < num_counters; ++i) {
1647 if ((interrupt_mask & CBE_PM_CTR_OVERFLOW_INTR(i))
1648 && ctr[i].enabled) {
1649 oprofile_add_ext_sample(pc, regs, i, is_kernel);
1650 cbe_write_ctr(cpu, i, reset_value[i]);
1651 }
1652 }
1653
1654
1655
1656
1657
1658
1659
1660
1661
1662 cbe_enable_pm_interrupts(cpu, hdw_thread,
1663 virt_cntr_inter_mask);
1664
1665
1666
1667
1668
1669
1670
1671
1672
1673
1674 cbe_enable_pm(cpu);
1675 }
1676 spin_unlock_irqrestore(&cntr_lock, flags);
1677}
1678
1679static void cell_handle_interrupt(struct pt_regs *regs,
1680 struct op_counter_config *ctr)
1681{
1682 if (profiling_mode == PPU_PROFILING)
1683 cell_handle_interrupt_ppu(regs, ctr);
1684 else
1685 cell_handle_interrupt_spu(regs, ctr);
1686}
1687
1688
1689
1690
1691
1692
1693static int cell_sync_start(void)
1694{
1695 if ((profiling_mode == SPU_PROFILING_CYCLES) ||
1696 (profiling_mode == SPU_PROFILING_EVENTS))
1697 return spu_sync_start();
1698 else
1699 return DO_GENERIC_SYNC;
1700}
1701
1702static int cell_sync_stop(void)
1703{
1704 if ((profiling_mode == SPU_PROFILING_CYCLES) ||
1705 (profiling_mode == SPU_PROFILING_EVENTS))
1706 return spu_sync_stop();
1707 else
1708 return 1;
1709}
1710
1711struct op_powerpc_model op_model_cell = {
1712 .reg_setup = cell_reg_setup,
1713 .cpu_setup = cell_cpu_setup,
1714 .global_start = cell_global_start,
1715 .global_stop = cell_global_stop,
1716 .sync_start = cell_sync_start,
1717 .sync_stop = cell_sync_stop,
1718 .handle_interrupt = cell_handle_interrupt,
1719};
1720