1
2
3
4
5
6
7
8
9
10
11
12
13#define DEBUG
14
15#include <linux/export.h>
16#include <linux/string.h>
17#include <linux/sched.h>
18#include <linux/init.h>
19#include <linux/kernel.h>
20#include <linux/reboot.h>
21#include <linux/delay.h>
22#include <linux/initrd.h>
23#include <linux/seq_file.h>
24#include <linux/ioport.h>
25#include <linux/console.h>
26#include <linux/utsname.h>
27#include <linux/tty.h>
28#include <linux/root_dev.h>
29#include <linux/notifier.h>
30#include <linux/cpu.h>
31#include <linux/unistd.h>
32#include <linux/serial.h>
33#include <linux/serial_8250.h>
34#include <linux/bootmem.h>
35#include <linux/pci.h>
36#include <linux/lockdep.h>
37#include <linux/memblock.h>
38#include <linux/hugetlb.h>
39#include <linux/memory.h>
40#include <linux/nmi.h>
41
42#include <asm/io.h>
43#include <asm/kdump.h>
44#include <asm/prom.h>
45#include <asm/processor.h>
46#include <asm/pgtable.h>
47#include <asm/smp.h>
48#include <asm/elf.h>
49#include <asm/machdep.h>
50#include <asm/paca.h>
51#include <asm/time.h>
52#include <asm/cputable.h>
53#include <asm/sections.h>
54#include <asm/btext.h>
55#include <asm/nvram.h>
56#include <asm/setup.h>
57#include <asm/rtas.h>
58#include <asm/iommu.h>
59#include <asm/serial.h>
60#include <asm/cache.h>
61#include <asm/page.h>
62#include <asm/mmu.h>
63#include <asm/firmware.h>
64#include <asm/xmon.h>
65#include <asm/udbg.h>
66#include <asm/kexec.h>
67#include <asm/mmu_context.h>
68#include <asm/code-patching.h>
69#include <asm/kvm_ppc.h>
70#include <asm/hugetlb.h>
71#include <asm/epapr_hcalls.h>
72
73#ifdef DEBUG
74#define DBG(fmt...) udbg_printf(fmt)
75#else
76#define DBG(fmt...)
77#endif
78
79int spinning_secondaries;
80u64 ppc64_pft_size;
81
82
83
84
85struct ppc64_caches ppc64_caches = {
86 .dline_size = 0x40,
87 .log_dline_size = 6,
88 .iline_size = 0x40,
89 .log_iline_size = 6
90};
91EXPORT_SYMBOL_GPL(ppc64_caches);
92
93
94
95
96
97int dcache_bsize;
98int icache_bsize;
99int ucache_bsize;
100
101#if defined(CONFIG_PPC_BOOK3E) && defined(CONFIG_SMP)
102static void setup_tlb_core_data(void)
103{
104 int cpu;
105
106 BUILD_BUG_ON(offsetof(struct tlb_core_data, lock) != 0);
107
108 for_each_possible_cpu(cpu) {
109 int first = cpu_first_thread_sibling(cpu);
110
111 paca[cpu].tcd_ptr = &paca[first].tcd;
112
113
114
115
116
117
118 if (smt_enabled_at_boot >= 2 &&
119 !mmu_has_feature(MMU_FTR_USE_TLBRSRV) &&
120 book3e_htw_mode != PPC_HTW_E6500) {
121
122 WARN_ONCE("%s: unsupported MMU configuration -- expect problems\n",
123 __func__);
124 }
125 }
126}
127#else
128static void setup_tlb_core_data(void)
129{
130}
131#endif
132
133#ifdef CONFIG_SMP
134
135static char *smt_enabled_cmdline;
136
137
138static void check_smt_enabled(void)
139{
140 struct device_node *dn;
141 const char *smt_option;
142
143
144 smt_enabled_at_boot = threads_per_core;
145
146
147 if (smt_enabled_cmdline) {
148 if (!strcmp(smt_enabled_cmdline, "on"))
149 smt_enabled_at_boot = threads_per_core;
150 else if (!strcmp(smt_enabled_cmdline, "off"))
151 smt_enabled_at_boot = 0;
152 else {
153 int smt;
154 int rc;
155
156 rc = kstrtoint(smt_enabled_cmdline, 10, &smt);
157 if (!rc)
158 smt_enabled_at_boot =
159 min(threads_per_core, smt);
160 }
161 } else {
162 dn = of_find_node_by_path("/options");
163 if (dn) {
164 smt_option = of_get_property(dn, "ibm,smt-enabled",
165 NULL);
166
167 if (smt_option) {
168 if (!strcmp(smt_option, "on"))
169 smt_enabled_at_boot = threads_per_core;
170 else if (!strcmp(smt_option, "off"))
171 smt_enabled_at_boot = 0;
172 }
173
174 of_node_put(dn);
175 }
176 }
177}
178
179
180static int __init early_smt_enabled(char *p)
181{
182 smt_enabled_cmdline = p;
183 return 0;
184}
185early_param("smt-enabled", early_smt_enabled);
186
187#else
188#define check_smt_enabled()
189#endif
190
191
192static void fixup_boot_paca(void)
193{
194
195 get_paca()->cpu_start = 1;
196
197 get_paca()->data_offset = 0;
198}
199
200static void cpu_ready_for_interrupts(void)
201{
202
203 get_paca()->kernel_msr = MSR_KERNEL;
204
205
206
207
208
209
210 if (cpu_has_feature(CPU_FTR_HVMODE) &&
211 cpu_has_feature(CPU_FTR_ARCH_207S)) {
212 unsigned long lpcr = mfspr(SPRN_LPCR);
213 mtspr(SPRN_LPCR, lpcr | LPCR_AIL_3);
214 }
215}
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236void __init early_setup(unsigned long dt_ptr)
237{
238 static __initdata struct paca_struct boot_paca;
239
240
241
242
243 identify_cpu(0, mfspr(SPRN_PVR));
244
245
246 initialise_paca(&boot_paca, 0);
247 setup_paca(&boot_paca);
248 fixup_boot_paca();
249
250
251 lockdep_init();
252
253
254
255
256 udbg_early_init();
257
258 DBG(" -> early_setup(), dt_ptr: 0x%lx\n", dt_ptr);
259
260
261
262
263
264
265 early_init_devtree(__va(dt_ptr));
266
267 epapr_paravirt_early_init();
268
269
270 setup_paca(&paca[boot_cpuid]);
271 fixup_boot_paca();
272
273
274 probe_machine();
275
276 setup_kdump_trampoline();
277
278 DBG("Found, Initializing memory management...\n");
279
280
281 early_init_mmu();
282
283
284
285
286
287
288 cpu_ready_for_interrupts();
289
290
291 kvm_cma_reserve();
292
293
294
295
296
297
298 reserve_hugetlb_gpages();
299
300 DBG(" <- early_setup()\n");
301
302#ifdef CONFIG_PPC_EARLY_DEBUG_BOOTX
303
304
305
306
307
308
309
310
311 btext_map();
312#endif
313}
314
315#ifdef CONFIG_SMP
316void early_setup_secondary(void)
317{
318
319 get_paca()->soft_enabled = 0;
320
321
322 early_init_mmu_secondary();
323
324
325
326
327
328
329 cpu_ready_for_interrupts();
330}
331
332#endif
333
334#if defined(CONFIG_SMP) || defined(CONFIG_KEXEC)
335void smp_release_cpus(void)
336{
337 unsigned long *ptr;
338 int i;
339
340 DBG(" -> smp_release_cpus()\n");
341
342
343
344
345
346
347
348 ptr = (unsigned long *)((unsigned long)&__secondary_hold_spinloop
349 - PHYSICAL_START);
350 *ptr = ppc_function_entry(generic_secondary_smp_init);
351
352
353 for (i = 0; i < 100000; i++) {
354 mb();
355 HMT_low();
356 if (spinning_secondaries == 0)
357 break;
358 udelay(1);
359 }
360 DBG("spinning_secondaries = %d\n", spinning_secondaries);
361
362 DBG(" <- smp_release_cpus()\n");
363}
364#endif
365
366
367
368
369
370
371
372
373static void __init initialize_cache_info(void)
374{
375 struct device_node *np;
376 unsigned long num_cpus = 0;
377
378 DBG(" -> initialize_cache_info()\n");
379
380 for_each_node_by_type(np, "cpu") {
381 num_cpus += 1;
382
383
384
385
386
387 if (num_cpus == 1) {
388 const __be32 *sizep, *lsizep;
389 u32 size, lsize;
390
391 size = 0;
392 lsize = cur_cpu_spec->dcache_bsize;
393 sizep = of_get_property(np, "d-cache-size", NULL);
394 if (sizep != NULL)
395 size = be32_to_cpu(*sizep);
396 lsizep = of_get_property(np, "d-cache-block-size",
397 NULL);
398
399 if (lsizep == NULL)
400 lsizep = of_get_property(np,
401 "d-cache-line-size",
402 NULL);
403 if (lsizep != NULL)
404 lsize = be32_to_cpu(*lsizep);
405 if (sizep == NULL || lsizep == NULL)
406 DBG("Argh, can't find dcache properties ! "
407 "sizep: %p, lsizep: %p\n", sizep, lsizep);
408
409 ppc64_caches.dsize = size;
410 ppc64_caches.dline_size = lsize;
411 ppc64_caches.log_dline_size = __ilog2(lsize);
412 ppc64_caches.dlines_per_page = PAGE_SIZE / lsize;
413
414 size = 0;
415 lsize = cur_cpu_spec->icache_bsize;
416 sizep = of_get_property(np, "i-cache-size", NULL);
417 if (sizep != NULL)
418 size = be32_to_cpu(*sizep);
419 lsizep = of_get_property(np, "i-cache-block-size",
420 NULL);
421 if (lsizep == NULL)
422 lsizep = of_get_property(np,
423 "i-cache-line-size",
424 NULL);
425 if (lsizep != NULL)
426 lsize = be32_to_cpu(*lsizep);
427 if (sizep == NULL || lsizep == NULL)
428 DBG("Argh, can't find icache properties ! "
429 "sizep: %p, lsizep: %p\n", sizep, lsizep);
430
431 ppc64_caches.isize = size;
432 ppc64_caches.iline_size = lsize;
433 ppc64_caches.log_iline_size = __ilog2(lsize);
434 ppc64_caches.ilines_per_page = PAGE_SIZE / lsize;
435 }
436 }
437
438 DBG(" <- initialize_cache_info()\n");
439}
440
441
442
443
444
445
446void __init setup_system(void)
447{
448 DBG(" -> setup_system()\n");
449
450
451
452
453 do_feature_fixups(cur_cpu_spec->cpu_features,
454 &__start___ftr_fixup, &__stop___ftr_fixup);
455 do_feature_fixups(cur_cpu_spec->mmu_features,
456 &__start___mmu_ftr_fixup, &__stop___mmu_ftr_fixup);
457 do_feature_fixups(powerpc_firmware_features,
458 &__start___fw_ftr_fixup, &__stop___fw_ftr_fixup);
459 do_lwsync_fixups(cur_cpu_spec->cpu_features,
460 &__start___lwsync_fixup, &__stop___lwsync_fixup);
461 do_final_fixups();
462
463
464
465
466 unflatten_device_tree();
467
468
469
470
471
472 initialize_cache_info();
473
474#ifdef CONFIG_PPC_RTAS
475
476
477
478 rtas_initialize();
479#endif
480
481
482
483
484 check_for_initrd();
485
486
487
488
489
490
491 if (ppc_md.init_early)
492 ppc_md.init_early();
493
494
495
496
497
498
499 find_legacy_serial_ports();
500
501
502
503
504 register_early_udbg_console();
505
506
507
508
509 xmon_setup();
510
511 smp_setup_cpu_maps();
512 check_smt_enabled();
513 setup_tlb_core_data();
514
515
516
517
518
519#if defined(CONFIG_SMP) && !defined(CONFIG_PPC_FSL_BOOK3E)
520
521
522
523 smp_release_cpus();
524#endif
525
526 pr_info("Starting Linux PPC64 %s\n", init_utsname()->version);
527
528 pr_info("-----------------------------------------------------\n");
529 pr_info("ppc64_pft_size = 0x%llx\n", ppc64_pft_size);
530 pr_info("phys_mem_size = 0x%llx\n", memblock_phys_mem_size());
531
532 if (ppc64_caches.dline_size != 0x80)
533 pr_info("dcache_line_size = 0x%x\n", ppc64_caches.dline_size);
534 if (ppc64_caches.iline_size != 0x80)
535 pr_info("icache_line_size = 0x%x\n", ppc64_caches.iline_size);
536
537 pr_info("cpu_features = 0x%016lx\n", cur_cpu_spec->cpu_features);
538 pr_info(" possible = 0x%016lx\n", CPU_FTRS_POSSIBLE);
539 pr_info(" always = 0x%016lx\n", CPU_FTRS_ALWAYS);
540 pr_info("cpu_user_features = 0x%08x 0x%08x\n", cur_cpu_spec->cpu_user_features,
541 cur_cpu_spec->cpu_user_features2);
542 pr_info("mmu_features = 0x%08x\n", cur_cpu_spec->mmu_features);
543 pr_info("firmware_features = 0x%016lx\n", powerpc_firmware_features);
544
545#ifdef CONFIG_PPC_STD_MMU_64
546 if (htab_address)
547 pr_info("htab_address = 0x%p\n", htab_address);
548
549 pr_info("htab_hash_mask = 0x%lx\n", htab_hash_mask);
550#endif
551
552 if (PHYSICAL_START > 0)
553 pr_info("physical_start = 0x%llx\n",
554 (unsigned long long)PHYSICAL_START);
555 pr_info("-----------------------------------------------------\n");
556
557 DBG(" <- setup_system()\n");
558}
559
560
561
562
563
564
565static u64 safe_stack_limit(void)
566{
567#ifdef CONFIG_PPC_BOOK3E
568
569 if (mmu_has_feature(MMU_FTR_TYPE_FSL_E))
570 return linear_map_top;
571
572 return 1ul << 30;
573#else
574
575 if (mmu_has_feature(MMU_FTR_1T_SEGMENT))
576 return 1UL << SID_SHIFT_1T;
577 return 1UL << SID_SHIFT;
578#endif
579}
580
581static void __init irqstack_early_init(void)
582{
583 u64 limit = safe_stack_limit();
584 unsigned int i;
585
586
587
588
589
590 for_each_possible_cpu(i) {
591 softirq_ctx[i] = (struct thread_info *)
592 __va(memblock_alloc_base(THREAD_SIZE,
593 THREAD_SIZE, limit));
594 hardirq_ctx[i] = (struct thread_info *)
595 __va(memblock_alloc_base(THREAD_SIZE,
596 THREAD_SIZE, limit));
597 }
598}
599
600#ifdef CONFIG_PPC_BOOK3E
601static void __init exc_lvl_early_init(void)
602{
603 unsigned int i;
604 unsigned long sp;
605
606 for_each_possible_cpu(i) {
607 sp = memblock_alloc(THREAD_SIZE, THREAD_SIZE);
608 critirq_ctx[i] = (struct thread_info *)__va(sp);
609 paca[i].crit_kstack = __va(sp + THREAD_SIZE);
610
611 sp = memblock_alloc(THREAD_SIZE, THREAD_SIZE);
612 dbgirq_ctx[i] = (struct thread_info *)__va(sp);
613 paca[i].dbg_kstack = __va(sp + THREAD_SIZE);
614
615 sp = memblock_alloc(THREAD_SIZE, THREAD_SIZE);
616 mcheckirq_ctx[i] = (struct thread_info *)__va(sp);
617 paca[i].mc_kstack = __va(sp + THREAD_SIZE);
618 }
619
620 if (cpu_has_feature(CPU_FTR_DEBUG_LVL_EXC))
621 patch_exception(0x040, exc_debug_debug_book3e);
622}
623#else
624#define exc_lvl_early_init()
625#endif
626
627
628
629
630
631
632static void __init emergency_stack_init(void)
633{
634 u64 limit;
635 unsigned int i;
636
637
638
639
640
641
642
643
644
645
646 limit = min(safe_stack_limit(), ppc64_rma_size);
647
648 for_each_possible_cpu(i) {
649 unsigned long sp;
650 sp = memblock_alloc_base(THREAD_SIZE, THREAD_SIZE, limit);
651 sp += THREAD_SIZE;
652 paca[i].emergency_sp = __va(sp);
653
654#ifdef CONFIG_PPC_BOOK3S_64
655
656 sp = memblock_alloc_base(THREAD_SIZE, THREAD_SIZE, limit);
657 sp += THREAD_SIZE;
658 paca[i].mc_emergency_sp = __va(sp);
659#endif
660 }
661}
662
663
664
665
666
667void __init setup_arch(char **cmdline_p)
668{
669 *cmdline_p = boot_command_line;
670
671
672
673
674
675
676 dcache_bsize = ppc64_caches.dline_size;
677 icache_bsize = ppc64_caches.iline_size;
678
679 if (ppc_md.panic)
680 setup_panic();
681
682 init_mm.start_code = (unsigned long)_stext;
683 init_mm.end_code = (unsigned long) _etext;
684 init_mm.end_data = (unsigned long) _edata;
685 init_mm.brk = klimit;
686#ifdef CONFIG_PPC_64K_PAGES
687 init_mm.context.pte_frag = NULL;
688#endif
689 irqstack_early_init();
690 exc_lvl_early_init();
691 emergency_stack_init();
692
693 initmem_init();
694
695#ifdef CONFIG_DUMMY_CONSOLE
696 conswitchp = &dummy_con;
697#endif
698
699 if (ppc_md.setup_arch)
700 ppc_md.setup_arch();
701
702 paging_init();
703
704
705 mmu_context_init();
706
707
708 if ((unsigned long)_stext & 0xffff)
709 panic("Kernelbase not 64K-aligned (0x%lx)!\n",
710 (unsigned long)_stext);
711}
712
713#ifdef CONFIG_SMP
714#define PCPU_DYN_SIZE ()
715
716static void * __init pcpu_fc_alloc(unsigned int cpu, size_t size, size_t align)
717{
718 return __alloc_bootmem_node(NODE_DATA(cpu_to_node(cpu)), size, align,
719 __pa(MAX_DMA_ADDRESS));
720}
721
722static void __init pcpu_fc_free(void *ptr, size_t size)
723{
724 free_bootmem(__pa(ptr), size);
725}
726
727static int pcpu_cpu_distance(unsigned int from, unsigned int to)
728{
729 if (cpu_to_node(from) == cpu_to_node(to))
730 return LOCAL_DISTANCE;
731 else
732 return REMOTE_DISTANCE;
733}
734
735unsigned long __per_cpu_offset[NR_CPUS] __read_mostly;
736EXPORT_SYMBOL(__per_cpu_offset);
737
738void __init setup_per_cpu_areas(void)
739{
740 const size_t dyn_size = PERCPU_MODULE_RESERVE + PERCPU_DYNAMIC_RESERVE;
741 size_t atom_size;
742 unsigned long delta;
743 unsigned int cpu;
744 int rc;
745
746
747
748
749
750
751 if (mmu_linear_psize == MMU_PAGE_4K)
752 atom_size = PAGE_SIZE;
753 else
754 atom_size = 1 << 20;
755
756 rc = pcpu_embed_first_chunk(0, dyn_size, atom_size, pcpu_cpu_distance,
757 pcpu_fc_alloc, pcpu_fc_free);
758 if (rc < 0)
759 panic("cannot initialize percpu area (err=%d)", rc);
760
761 delta = (unsigned long)pcpu_base_addr - (unsigned long)__per_cpu_start;
762 for_each_possible_cpu(cpu) {
763 __per_cpu_offset[cpu] = delta + pcpu_unit_offsets[cpu];
764 paca[cpu].data_offset = __per_cpu_offset[cpu];
765 }
766}
767#endif
768
769#ifdef CONFIG_MEMORY_HOTPLUG_SPARSE
770unsigned long memory_block_size_bytes(void)
771{
772 if (ppc_md.memory_block_size)
773 return ppc_md.memory_block_size();
774
775 return MIN_MEMORY_BLOCK_SIZE;
776}
777#endif
778
779#if defined(CONFIG_PPC_INDIRECT_PIO) || defined(CONFIG_PPC_INDIRECT_MMIO)
780struct ppc_pci_io ppc_pci_io;
781EXPORT_SYMBOL(ppc_pci_io);
782#endif
783
784#ifdef CONFIG_HARDLOCKUP_DETECTOR
785u64 hw_nmi_get_sample_period(int watchdog_thresh)
786{
787 return ppc_proc_freq * watchdog_thresh;
788}
789
790
791
792
793
794static int __init disable_hardlockup_detector(void)
795{
796 hardlockup_detector_disable();
797
798 return 0;
799}
800early_initcall(disable_hardlockup_detector);
801#endif
802