linux/arch/x86/kernel/setup.c
<<
>>
Prefs
   1/*
   2 *  Copyright (C) 1995  Linus Torvalds
   3 *
   4 *  Support of BIGMEM added by Gerhard Wichert, Siemens AG, July 1999
   5 *
   6 *  Memory region support
   7 *      David Parsons <orc@pell.chi.il.us>, July-August 1999
   8 *
   9 *  Added E820 sanitization routine (removes overlapping memory regions);
  10 *  Brian Moyle <bmoyle@mvista.com>, February 2001
  11 *
  12 * Moved CPU detection code to cpu/${cpu}.c
  13 *    Patrick Mochel <mochel@osdl.org>, March 2002
  14 *
  15 *  Provisions for empty E820 memory regions (reported by certain BIOSes).
  16 *  Alex Achenbach <xela@slit.de>, December 2002.
  17 *
  18 */
  19
  20/*
  21 * This file handles the architecture-dependent parts of initialization
  22 */
  23
  24#include <linux/sched.h>
  25#include <linux/mm.h>
  26#include <linux/mmzone.h>
  27#include <linux/screen_info.h>
  28#include <linux/ioport.h>
  29#include <linux/acpi.h>
  30#include <linux/sfi.h>
  31#include <linux/apm_bios.h>
  32#include <linux/initrd.h>
  33#include <linux/bootmem.h>
  34#include <linux/seq_file.h>
  35#include <linux/console.h>
  36#include <linux/mca.h>
  37#include <linux/root_dev.h>
  38#include <linux/highmem.h>
  39#include <linux/module.h>
  40#include <linux/efi.h>
  41#include <linux/init.h>
  42#include <linux/edd.h>
  43#include <linux/iscsi_ibft.h>
  44#include <linux/nodemask.h>
  45#include <linux/kexec.h>
  46#include <linux/dmi.h>
  47#include <linux/pfn.h>
  48#include <linux/pci.h>
  49#include <asm/pci-direct.h>
  50#include <linux/init_ohci1394_dma.h>
  51#include <linux/kvm_para.h>
  52
  53#include <linux/errno.h>
  54#include <linux/kernel.h>
  55#include <linux/stddef.h>
  56#include <linux/unistd.h>
  57#include <linux/ptrace.h>
  58#include <linux/slab.h>
  59#include <linux/user.h>
  60#include <linux/delay.h>
  61
  62#include <linux/kallsyms.h>
  63#include <linux/cpufreq.h>
  64#include <linux/dma-mapping.h>
  65#include <linux/ctype.h>
  66#include <linux/uaccess.h>
  67
  68#include <linux/percpu.h>
  69#include <linux/crash_dump.h>
  70#include <linux/tboot.h>
  71
  72#include <video/edid.h>
  73
  74#include <asm/mtrr.h>
  75#include <asm/apic.h>
  76#include <asm/e820.h>
  77#include <asm/mpspec.h>
  78#include <asm/setup.h>
  79#include <asm/efi.h>
  80#include <asm/timer.h>
  81#include <asm/i8259.h>
  82#include <asm/sections.h>
  83#include <asm/dmi.h>
  84#include <asm/io_apic.h>
  85#include <asm/ist.h>
  86#include <asm/vmi.h>
  87#include <asm/setup_arch.h>
  88#include <asm/bios_ebda.h>
  89#include <asm/cacheflush.h>
  90#include <asm/processor.h>
  91#include <asm/bugs.h>
  92
  93#include <asm/system.h>
  94#include <asm/vsyscall.h>
  95#include <asm/cpu.h>
  96#include <asm/desc.h>
  97#include <asm/dma.h>
  98#include <asm/iommu.h>
  99#include <asm/gart.h>
 100#include <asm/mmu_context.h>
 101#include <asm/proto.h>
 102
 103#include <asm/paravirt.h>
 104#include <asm/hypervisor.h>
 105
 106#include <asm/percpu.h>
 107#include <asm/topology.h>
 108#include <asm/apicdef.h>
 109#ifdef CONFIG_X86_64
 110#include <asm/numa_64.h>
 111#endif
 112
 113/*
 114 * end_pfn only includes RAM, while max_pfn_mapped includes all e820 entries.
 115 * The direct mapping extends to max_pfn_mapped, so that we can directly access
 116 * apertures, ACPI and other tables without having to play with fixmaps.
 117 */
 118unsigned long max_low_pfn_mapped;
 119unsigned long max_pfn_mapped;
 120
 121RESERVE_BRK(dmi_alloc, 65536);
 122
 123unsigned int boot_cpu_id __read_mostly;
 124
 125static __initdata unsigned long _brk_start = (unsigned long)__brk_base;
 126unsigned long _brk_end = (unsigned long)__brk_base;
 127
 128#ifdef CONFIG_X86_64
 129int default_cpu_present_to_apicid(int mps_cpu)
 130{
 131        return __default_cpu_present_to_apicid(mps_cpu);
 132}
 133
 134int default_check_phys_apicid_present(int phys_apicid)
 135{
 136        return __default_check_phys_apicid_present(phys_apicid);
 137}
 138#endif
 139
 140#ifndef CONFIG_DEBUG_BOOT_PARAMS
 141struct boot_params __initdata boot_params;
 142#else
 143struct boot_params boot_params;
 144#endif
 145
 146/*
 147 * Machine setup..
 148 */
 149static struct resource data_resource = {
 150        .name   = "Kernel data",
 151        .start  = 0,
 152        .end    = 0,
 153        .flags  = IORESOURCE_BUSY | IORESOURCE_MEM
 154};
 155
 156static struct resource code_resource = {
 157        .name   = "Kernel code",
 158        .start  = 0,
 159        .end    = 0,
 160        .flags  = IORESOURCE_BUSY | IORESOURCE_MEM
 161};
 162
 163static struct resource bss_resource = {
 164        .name   = "Kernel bss",
 165        .start  = 0,
 166        .end    = 0,
 167        .flags  = IORESOURCE_BUSY | IORESOURCE_MEM
 168};
 169
 170
 171#ifdef CONFIG_X86_32
 172/* cpu data as detected by the assembly code in head.S */
 173struct cpuinfo_x86 new_cpu_data __cpuinitdata = {0, 0, 0, 0, -1, 1, 0, 0, -1};
 174/* common cpu data for all cpus */
 175struct cpuinfo_x86 boot_cpu_data __read_mostly = {0, 0, 0, 0, -1, 1, 0, 0, -1};
 176EXPORT_SYMBOL(boot_cpu_data);
 177static void set_mca_bus(int x)
 178{
 179#ifdef CONFIG_MCA
 180        MCA_bus = x;
 181#endif
 182}
 183
 184unsigned int def_to_bigsmp;
 185
 186/* for MCA, but anyone else can use it if they want */
 187unsigned int machine_id;
 188unsigned int machine_submodel_id;
 189unsigned int BIOS_revision;
 190
 191struct apm_info apm_info;
 192EXPORT_SYMBOL(apm_info);
 193
 194#if defined(CONFIG_X86_SPEEDSTEP_SMI) || \
 195        defined(CONFIG_X86_SPEEDSTEP_SMI_MODULE)
 196struct ist_info ist_info;
 197EXPORT_SYMBOL(ist_info);
 198#else
 199struct ist_info ist_info;
 200#endif
 201
 202#else
 203struct cpuinfo_x86 boot_cpu_data __read_mostly = {
 204        .x86_phys_bits = MAX_PHYSMEM_BITS,
 205};
 206EXPORT_SYMBOL(boot_cpu_data);
 207#endif
 208
 209
 210#if !defined(CONFIG_X86_PAE) || defined(CONFIG_X86_64)
 211unsigned long mmu_cr4_features;
 212#else
 213unsigned long mmu_cr4_features = X86_CR4_PAE;
 214#endif
 215
 216/* Boot loader ID and version as integers, for the benefit of proc_dointvec */
 217int bootloader_type, bootloader_version;
 218
 219/*
 220 * Setup options
 221 */
 222struct screen_info screen_info;
 223EXPORT_SYMBOL(screen_info);
 224struct edid_info edid_info;
 225EXPORT_SYMBOL_GPL(edid_info);
 226
 227extern int root_mountflags;
 228
 229unsigned long saved_video_mode;
 230
 231#define RAMDISK_IMAGE_START_MASK        0x07FF
 232#define RAMDISK_PROMPT_FLAG             0x8000
 233#define RAMDISK_LOAD_FLAG               0x4000
 234
 235static char __initdata command_line[COMMAND_LINE_SIZE];
 236#ifdef CONFIG_CMDLINE_BOOL
 237static char __initdata builtin_cmdline[COMMAND_LINE_SIZE] = CONFIG_CMDLINE;
 238#endif
 239
 240#if defined(CONFIG_EDD) || defined(CONFIG_EDD_MODULE)
 241struct edd edd;
 242#ifdef CONFIG_EDD_MODULE
 243EXPORT_SYMBOL(edd);
 244#endif
 245/**
 246 * copy_edd() - Copy the BIOS EDD information
 247 *              from boot_params into a safe place.
 248 *
 249 */
 250static inline void copy_edd(void)
 251{
 252     memcpy(edd.mbr_signature, boot_params.edd_mbr_sig_buffer,
 253            sizeof(edd.mbr_signature));
 254     memcpy(edd.edd_info, boot_params.eddbuf, sizeof(edd.edd_info));
 255     edd.mbr_signature_nr = boot_params.edd_mbr_sig_buf_entries;
 256     edd.edd_info_nr = boot_params.eddbuf_entries;
 257}
 258#else
 259static inline void copy_edd(void)
 260{
 261}
 262#endif
 263
 264void * __init extend_brk(size_t size, size_t align)
 265{
 266        size_t mask = align - 1;
 267        void *ret;
 268
 269        BUG_ON(_brk_start == 0);
 270        BUG_ON(align & mask);
 271
 272        _brk_end = (_brk_end + mask) & ~mask;
 273        BUG_ON((char *)(_brk_end + size) > __brk_limit);
 274
 275        ret = (void *)_brk_end;
 276        _brk_end += size;
 277
 278        memset(ret, 0, size);
 279
 280        return ret;
 281}
 282
 283#ifdef CONFIG_X86_64
 284static void __init init_gbpages(void)
 285{
 286        if (direct_gbpages && cpu_has_gbpages)
 287                printk(KERN_INFO "Using GB pages for direct mapping\n");
 288        else
 289                direct_gbpages = 0;
 290}
 291#else
 292static inline void init_gbpages(void)
 293{
 294}
 295#endif
 296
 297static void __init reserve_brk(void)
 298{
 299        if (_brk_end > _brk_start)
 300                reserve_early(__pa(_brk_start), __pa(_brk_end), "BRK");
 301
 302        /* Mark brk area as locked down and no longer taking any
 303           new allocations */
 304        _brk_start = 0;
 305}
 306
 307#ifdef CONFIG_BLK_DEV_INITRD
 308
 309#define MAX_MAP_CHUNK   (NR_FIX_BTMAPS << PAGE_SHIFT)
 310static void __init relocate_initrd(void)
 311{
 312
 313        u64 ramdisk_image = boot_params.hdr.ramdisk_image;
 314        u64 ramdisk_size  = boot_params.hdr.ramdisk_size;
 315        u64 end_of_lowmem = max_low_pfn_mapped << PAGE_SHIFT;
 316        u64 ramdisk_here;
 317        unsigned long slop, clen, mapaddr;
 318        char *p, *q;
 319
 320        /* We need to move the initrd down into lowmem */
 321        ramdisk_here = find_e820_area(0, end_of_lowmem, ramdisk_size,
 322                                         PAGE_SIZE);
 323
 324        if (ramdisk_here == -1ULL)
 325                panic("Cannot find place for new RAMDISK of size %lld\n",
 326                         ramdisk_size);
 327
 328        /* Note: this includes all the lowmem currently occupied by
 329           the initrd, we rely on that fact to keep the data intact. */
 330        reserve_early(ramdisk_here, ramdisk_here + ramdisk_size,
 331                         "NEW RAMDISK");
 332        initrd_start = ramdisk_here + PAGE_OFFSET;
 333        initrd_end   = initrd_start + ramdisk_size;
 334        printk(KERN_INFO "Allocated new RAMDISK: %08llx - %08llx\n",
 335                         ramdisk_here, ramdisk_here + ramdisk_size);
 336
 337        q = (char *)initrd_start;
 338
 339        /* Copy any lowmem portion of the initrd */
 340        if (ramdisk_image < end_of_lowmem) {
 341                clen = end_of_lowmem - ramdisk_image;
 342                p = (char *)__va(ramdisk_image);
 343                memcpy(q, p, clen);
 344                q += clen;
 345                ramdisk_image += clen;
 346                ramdisk_size  -= clen;
 347        }
 348
 349        /* Copy the highmem portion of the initrd */
 350        while (ramdisk_size) {
 351                slop = ramdisk_image & ~PAGE_MASK;
 352                clen = ramdisk_size;
 353                if (clen > MAX_MAP_CHUNK-slop)
 354                        clen = MAX_MAP_CHUNK-slop;
 355                mapaddr = ramdisk_image & PAGE_MASK;
 356                p = early_memremap(mapaddr, clen+slop);
 357                memcpy(q, p+slop, clen);
 358                early_iounmap(p, clen+slop);
 359                q += clen;
 360                ramdisk_image += clen;
 361                ramdisk_size  -= clen;
 362        }
 363        /* high pages is not converted by early_res_to_bootmem */
 364        ramdisk_image = boot_params.hdr.ramdisk_image;
 365        ramdisk_size  = boot_params.hdr.ramdisk_size;
 366        printk(KERN_INFO "Move RAMDISK from %016llx - %016llx to"
 367                " %08llx - %08llx\n",
 368                ramdisk_image, ramdisk_image + ramdisk_size - 1,
 369                ramdisk_here, ramdisk_here + ramdisk_size - 1);
 370}
 371
 372static void __init reserve_initrd(void)
 373{
 374        u64 ramdisk_image = boot_params.hdr.ramdisk_image;
 375        u64 ramdisk_size  = boot_params.hdr.ramdisk_size;
 376        u64 ramdisk_end   = ramdisk_image + ramdisk_size;
 377        u64 end_of_lowmem = max_low_pfn_mapped << PAGE_SHIFT;
 378
 379        if (!boot_params.hdr.type_of_loader ||
 380            !ramdisk_image || !ramdisk_size)
 381                return;         /* No initrd provided by bootloader */
 382
 383        initrd_start = 0;
 384
 385        if (ramdisk_size >= (end_of_lowmem>>1)) {
 386                free_early(ramdisk_image, ramdisk_end);
 387                printk(KERN_ERR "initrd too large to handle, "
 388                       "disabling initrd\n");
 389                return;
 390        }
 391
 392        printk(KERN_INFO "RAMDISK: %08llx - %08llx\n", ramdisk_image,
 393                        ramdisk_end);
 394
 395
 396        if (ramdisk_end <= end_of_lowmem) {
 397                /* All in lowmem, easy case */
 398                /*
 399                 * don't need to reserve again, already reserved early
 400                 * in i386_start_kernel
 401                 */
 402                initrd_start = ramdisk_image + PAGE_OFFSET;
 403                initrd_end = initrd_start + ramdisk_size;
 404                return;
 405        }
 406
 407        relocate_initrd();
 408
 409        free_early(ramdisk_image, ramdisk_end);
 410}
 411#else
 412static void __init reserve_initrd(void)
 413{
 414}
 415#endif /* CONFIG_BLK_DEV_INITRD */
 416
 417static void __init parse_setup_data(void)
 418{
 419        struct setup_data *data;
 420        u64 pa_data;
 421
 422        if (boot_params.hdr.version < 0x0209)
 423                return;
 424        pa_data = boot_params.hdr.setup_data;
 425        while (pa_data) {
 426                data = early_memremap(pa_data, PAGE_SIZE);
 427                switch (data->type) {
 428                case SETUP_E820_EXT:
 429                        parse_e820_ext(data, pa_data);
 430                        break;
 431                default:
 432                        break;
 433                }
 434                pa_data = data->next;
 435                early_iounmap(data, PAGE_SIZE);
 436        }
 437}
 438
 439static void __init e820_reserve_setup_data(void)
 440{
 441        struct setup_data *data;
 442        u64 pa_data;
 443        int found = 0;
 444
 445        if (boot_params.hdr.version < 0x0209)
 446                return;
 447        pa_data = boot_params.hdr.setup_data;
 448        while (pa_data) {
 449                data = early_memremap(pa_data, sizeof(*data));
 450                e820_update_range(pa_data, sizeof(*data)+data->len,
 451                         E820_RAM, E820_RESERVED_KERN);
 452                found = 1;
 453                pa_data = data->next;
 454                early_iounmap(data, sizeof(*data));
 455        }
 456        if (!found)
 457                return;
 458
 459        sanitize_e820_map(e820.map, ARRAY_SIZE(e820.map), &e820.nr_map);
 460        memcpy(&e820_saved, &e820, sizeof(struct e820map));
 461        printk(KERN_INFO "extended physical RAM map:\n");
 462        e820_print_map("reserve setup_data");
 463}
 464
 465static void __init reserve_early_setup_data(void)
 466{
 467        struct setup_data *data;
 468        u64 pa_data;
 469        char buf[32];
 470
 471        if (boot_params.hdr.version < 0x0209)
 472                return;
 473        pa_data = boot_params.hdr.setup_data;
 474        while (pa_data) {
 475                data = early_memremap(pa_data, sizeof(*data));
 476                sprintf(buf, "setup data %x", data->type);
 477                reserve_early(pa_data, pa_data+sizeof(*data)+data->len, buf);
 478                pa_data = data->next;
 479                early_iounmap(data, sizeof(*data));
 480        }
 481}
 482
 483/*
 484 * --------- Crashkernel reservation ------------------------------
 485 */
 486
 487#ifdef CONFIG_KEXEC
 488
 489/**
 490 * Reserve @size bytes of crashkernel memory at any suitable offset.
 491 *
 492 * @size: Size of the crashkernel memory to reserve.
 493 * Returns the base address on success, and -1ULL on failure.
 494 */
 495static
 496unsigned long long __init find_and_reserve_crashkernel(unsigned long long size)
 497{
 498        const unsigned long long alignment = 16<<20;    /* 16M */
 499        unsigned long long start = 0LL;
 500
 501        while (1) {
 502                int ret;
 503
 504                start = find_e820_area(start, ULONG_MAX, size, alignment);
 505                if (start == -1ULL)
 506                        return start;
 507
 508                /* try to reserve it */
 509                ret = reserve_bootmem_generic(start, size, BOOTMEM_EXCLUSIVE);
 510                if (ret >= 0)
 511                        return start;
 512
 513                start += alignment;
 514        }
 515}
 516
 517static inline unsigned long long get_total_mem(void)
 518{
 519        unsigned long long total;
 520
 521        total = max_low_pfn - min_low_pfn;
 522#ifdef CONFIG_HIGHMEM
 523        total += highend_pfn - highstart_pfn;
 524#endif
 525
 526        return total << PAGE_SHIFT;
 527}
 528
 529static void __init reserve_crashkernel(void)
 530{
 531        unsigned long long total_mem;
 532        unsigned long long crash_size, crash_base;
 533        int ret;
 534
 535        total_mem = get_total_mem();
 536
 537        ret = parse_crashkernel(boot_command_line, total_mem,
 538                        &crash_size, &crash_base);
 539        if (ret != 0 || crash_size <= 0)
 540                return;
 541
 542        /* 0 means: find the address automatically */
 543        if (crash_base <= 0) {
 544                crash_base = find_and_reserve_crashkernel(crash_size);
 545                if (crash_base == -1ULL) {
 546                        pr_info("crashkernel reservation failed. "
 547                                "No suitable area found.\n");
 548                        return;
 549                }
 550        } else {
 551                ret = reserve_bootmem_generic(crash_base, crash_size,
 552                                        BOOTMEM_EXCLUSIVE);
 553                if (ret < 0) {
 554                        pr_info("crashkernel reservation failed - "
 555                                "memory is in use\n");
 556                        return;
 557                }
 558        }
 559
 560        printk(KERN_INFO "Reserving %ldMB of memory at %ldMB "
 561                        "for crashkernel (System RAM: %ldMB)\n",
 562                        (unsigned long)(crash_size >> 20),
 563                        (unsigned long)(crash_base >> 20),
 564                        (unsigned long)(total_mem >> 20));
 565
 566        crashk_res.start = crash_base;
 567        crashk_res.end   = crash_base + crash_size - 1;
 568        insert_resource(&iomem_resource, &crashk_res);
 569}
 570#else
 571static void __init reserve_crashkernel(void)
 572{
 573}
 574#endif
 575
 576static struct resource standard_io_resources[] = {
 577        { .name = "dma1", .start = 0x00, .end = 0x1f,
 578                .flags = IORESOURCE_BUSY | IORESOURCE_IO },
 579        { .name = "pic1", .start = 0x20, .end = 0x21,
 580                .flags = IORESOURCE_BUSY | IORESOURCE_IO },
 581        { .name = "timer0", .start = 0x40, .end = 0x43,
 582                .flags = IORESOURCE_BUSY | IORESOURCE_IO },
 583        { .name = "timer1", .start = 0x50, .end = 0x53,
 584                .flags = IORESOURCE_BUSY | IORESOURCE_IO },
 585        { .name = "keyboard", .start = 0x60, .end = 0x60,
 586                .flags = IORESOURCE_BUSY | IORESOURCE_IO },
 587        { .name = "keyboard", .start = 0x64, .end = 0x64,
 588                .flags = IORESOURCE_BUSY | IORESOURCE_IO },
 589        { .name = "dma page reg", .start = 0x80, .end = 0x8f,
 590                .flags = IORESOURCE_BUSY | IORESOURCE_IO },
 591        { .name = "pic2", .start = 0xa0, .end = 0xa1,
 592                .flags = IORESOURCE_BUSY | IORESOURCE_IO },
 593        { .name = "dma2", .start = 0xc0, .end = 0xdf,
 594                .flags = IORESOURCE_BUSY | IORESOURCE_IO },
 595        { .name = "fpu", .start = 0xf0, .end = 0xff,
 596                .flags = IORESOURCE_BUSY | IORESOURCE_IO }
 597};
 598
 599void __init reserve_standard_io_resources(void)
 600{
 601        int i;
 602
 603        /* request I/O space for devices used on all i[345]86 PCs */
 604        for (i = 0; i < ARRAY_SIZE(standard_io_resources); i++)
 605                request_resource(&ioport_resource, &standard_io_resources[i]);
 606
 607}
 608
 609/*
 610 * Note: elfcorehdr_addr is not just limited to vmcore. It is also used by
 611 * is_kdump_kernel() to determine if we are booting after a panic. Hence
 612 * ifdef it under CONFIG_CRASH_DUMP and not CONFIG_PROC_VMCORE.
 613 */
 614
 615#ifdef CONFIG_CRASH_DUMP
 616/* elfcorehdr= specifies the location of elf core header
 617 * stored by the crashed kernel. This option will be passed
 618 * by kexec loader to the capture kernel.
 619 */
 620static int __init setup_elfcorehdr(char *arg)
 621{
 622        char *end;
 623        if (!arg)
 624                return -EINVAL;
 625        elfcorehdr_addr = memparse(arg, &end);
 626        return end > arg ? 0 : -EINVAL;
 627}
 628early_param("elfcorehdr", setup_elfcorehdr);
 629#endif
 630
 631#ifdef CONFIG_X86_RESERVE_LOW_64K
 632static int __init dmi_low_memory_corruption(const struct dmi_system_id *d)
 633{
 634        printk(KERN_NOTICE
 635                "%s detected: BIOS may corrupt low RAM, working around it.\n",
 636                d->ident);
 637
 638        e820_update_range(0, 0x10000, E820_RAM, E820_RESERVED);
 639        sanitize_e820_map(e820.map, ARRAY_SIZE(e820.map), &e820.nr_map);
 640
 641        return 0;
 642}
 643#endif
 644
 645/* List of systems that have known low memory corruption BIOS problems */
 646static struct dmi_system_id __initdata bad_bios_dmi_table[] = {
 647#ifdef CONFIG_X86_RESERVE_LOW_64K
 648        {
 649                .callback = dmi_low_memory_corruption,
 650                .ident = "AMI BIOS",
 651                .matches = {
 652                        DMI_MATCH(DMI_BIOS_VENDOR, "American Megatrends Inc."),
 653                },
 654        },
 655        {
 656                .callback = dmi_low_memory_corruption,
 657                .ident = "Phoenix BIOS",
 658                .matches = {
 659                        DMI_MATCH(DMI_BIOS_VENDOR, "Phoenix Technologies"),
 660                },
 661        },
 662        {
 663                .callback = dmi_low_memory_corruption,
 664                .ident = "Phoenix/MSC BIOS",
 665                .matches = {
 666                        DMI_MATCH(DMI_BIOS_VENDOR, "Phoenix/MSC"),
 667                },
 668        },
 669        {
 670        /*
 671         * AMI BIOS with low memory corruption was found on Intel DG45ID board.
 672         * It hase different DMI_BIOS_VENDOR = "Intel Corp.", for now we will
 673         * match only DMI_BOARD_NAME and see if there is more bad products
 674         * with this vendor.
 675         */
 676                .callback = dmi_low_memory_corruption,
 677                .ident = "AMI BIOS",
 678                .matches = {
 679                        DMI_MATCH(DMI_BOARD_NAME, "DG45ID"),
 680                },
 681        },
 682#endif
 683        {}
 684};
 685
 686/*
 687 * Determine if we were loaded by an EFI loader.  If so, then we have also been
 688 * passed the efi memmap, systab, etc., so we should use these data structures
 689 * for initialization.  Note, the efi init code path is determined by the
 690 * global efi_enabled. This allows the same kernel image to be used on existing
 691 * systems (with a traditional BIOS) as well as on EFI systems.
 692 */
 693/*
 694 * setup_arch - architecture-specific boot-time initializations
 695 *
 696 * Note: On x86_64, fixmaps are ready for use even before this is called.
 697 */
 698
 699void __init setup_arch(char **cmdline_p)
 700{
 701#ifdef CONFIG_X86_32
 702        memcpy(&boot_cpu_data, &new_cpu_data, sizeof(new_cpu_data));
 703        visws_early_detect();
 704#else
 705        printk(KERN_INFO "Command line: %s\n", boot_command_line);
 706#endif
 707
 708        /* VMI may relocate the fixmap; do this before touching ioremap area */
 709        vmi_init();
 710
 711        early_cpu_init();
 712        early_ioremap_init();
 713
 714        ROOT_DEV = old_decode_dev(boot_params.hdr.root_dev);
 715        screen_info = boot_params.screen_info;
 716        edid_info = boot_params.edid_info;
 717#ifdef CONFIG_X86_32
 718        apm_info.bios = boot_params.apm_bios_info;
 719        ist_info = boot_params.ist_info;
 720        if (boot_params.sys_desc_table.length != 0) {
 721                set_mca_bus(boot_params.sys_desc_table.table[3] & 0x2);
 722                machine_id = boot_params.sys_desc_table.table[0];
 723                machine_submodel_id = boot_params.sys_desc_table.table[1];
 724                BIOS_revision = boot_params.sys_desc_table.table[2];
 725        }
 726#endif
 727        saved_video_mode = boot_params.hdr.vid_mode;
 728        bootloader_type = boot_params.hdr.type_of_loader;
 729        if ((bootloader_type >> 4) == 0xe) {
 730                bootloader_type &= 0xf;
 731                bootloader_type |= (boot_params.hdr.ext_loader_type+0x10) << 4;
 732        }
 733        bootloader_version  = bootloader_type & 0xf;
 734        bootloader_version |= boot_params.hdr.ext_loader_ver << 4;
 735
 736#ifdef CONFIG_BLK_DEV_RAM
 737        rd_image_start = boot_params.hdr.ram_size & RAMDISK_IMAGE_START_MASK;
 738        rd_prompt = ((boot_params.hdr.ram_size & RAMDISK_PROMPT_FLAG) != 0);
 739        rd_doload = ((boot_params.hdr.ram_size & RAMDISK_LOAD_FLAG) != 0);
 740#endif
 741#ifdef CONFIG_EFI
 742        if (!strncmp((char *)&boot_params.efi_info.efi_loader_signature,
 743#ifdef CONFIG_X86_32
 744                     "EL32",
 745#else
 746                     "EL64",
 747#endif
 748         4)) {
 749                efi_enabled = 1;
 750                efi_reserve_early();
 751        }
 752#endif
 753
 754        x86_init.oem.arch_setup();
 755
 756        setup_memory_map();
 757        parse_setup_data();
 758        /* update the e820_saved too */
 759        e820_reserve_setup_data();
 760
 761        copy_edd();
 762
 763        if (!boot_params.hdr.root_flags)
 764                root_mountflags &= ~MS_RDONLY;
 765        init_mm.start_code = (unsigned long) _text;
 766        init_mm.end_code = (unsigned long) _etext;
 767        init_mm.end_data = (unsigned long) _edata;
 768        init_mm.brk = _brk_end;
 769
 770        code_resource.start = virt_to_phys(_text);
 771        code_resource.end = virt_to_phys(_etext)-1;
 772        data_resource.start = virt_to_phys(_etext);
 773        data_resource.end = virt_to_phys(_edata)-1;
 774        bss_resource.start = virt_to_phys(&__bss_start);
 775        bss_resource.end = virt_to_phys(&__bss_stop)-1;
 776
 777#ifdef CONFIG_CMDLINE_BOOL
 778#ifdef CONFIG_CMDLINE_OVERRIDE
 779        strlcpy(boot_command_line, builtin_cmdline, COMMAND_LINE_SIZE);
 780#else
 781        if (builtin_cmdline[0]) {
 782                /* append boot loader cmdline to builtin */
 783                strlcat(builtin_cmdline, " ", COMMAND_LINE_SIZE);
 784                strlcat(builtin_cmdline, boot_command_line, COMMAND_LINE_SIZE);
 785                strlcpy(boot_command_line, builtin_cmdline, COMMAND_LINE_SIZE);
 786        }
 787#endif
 788#endif
 789
 790        strlcpy(command_line, boot_command_line, COMMAND_LINE_SIZE);
 791        *cmdline_p = command_line;
 792
 793#ifdef CONFIG_X86_64
 794        /*
 795         * Must call this twice: Once just to detect whether hardware doesn't
 796         * support NX (so that the early EHCI debug console setup can safely
 797         * call set_fixmap(), and then again after parsing early parameters to
 798         * honor the respective command line option.
 799         */
 800        check_efer();
 801#endif
 802
 803        parse_early_param();
 804
 805#ifdef CONFIG_X86_64
 806        check_efer();
 807#endif
 808
 809        /* Must be before kernel pagetables are setup */
 810        vmi_activate();
 811
 812        /* after early param, so could get panic from serial */
 813        reserve_early_setup_data();
 814
 815        if (acpi_mps_check()) {
 816#ifdef CONFIG_X86_LOCAL_APIC
 817                disable_apic = 1;
 818#endif
 819                setup_clear_cpu_cap(X86_FEATURE_APIC);
 820        }
 821
 822#ifdef CONFIG_PCI
 823        if (pci_early_dump_regs)
 824                early_dump_pci_devices();
 825#endif
 826
 827        finish_e820_parsing();
 828
 829        if (efi_enabled)
 830                efi_init();
 831
 832        dmi_scan_machine();
 833
 834        dmi_check_system(bad_bios_dmi_table);
 835
 836        /*
 837         * VMware detection requires dmi to be available, so this
 838         * needs to be done after dmi_scan_machine, for the BP.
 839         */
 840        init_hypervisor_platform();
 841
 842        x86_init.resources.probe_roms();
 843
 844        /* after parse_early_param, so could debug it */
 845        insert_resource(&iomem_resource, &code_resource);
 846        insert_resource(&iomem_resource, &data_resource);
 847        insert_resource(&iomem_resource, &bss_resource);
 848
 849
 850#ifdef CONFIG_X86_32
 851        if (ppro_with_ram_bug()) {
 852                e820_update_range(0x70000000ULL, 0x40000ULL, E820_RAM,
 853                                  E820_RESERVED);
 854                sanitize_e820_map(e820.map, ARRAY_SIZE(e820.map), &e820.nr_map);
 855                printk(KERN_INFO "fixed physical RAM map:\n");
 856                e820_print_map("bad_ppro");
 857        }
 858#else
 859        early_gart_iommu_check();
 860#endif
 861
 862        /*
 863         * partially used pages are not usable - thus
 864         * we are rounding upwards:
 865         */
 866        max_pfn = e820_end_of_ram_pfn();
 867
 868        /* preallocate 4k for mptable mpc */
 869        early_reserve_e820_mpc_new();
 870        /* update e820 for memory not covered by WB MTRRs */
 871        mtrr_bp_init();
 872        if (mtrr_trim_uncached_memory(max_pfn))
 873                max_pfn = e820_end_of_ram_pfn();
 874
 875#ifdef CONFIG_X86_32
 876        /* max_low_pfn get updated here */
 877        find_low_pfn_range();
 878#else
 879        num_physpages = max_pfn;
 880
 881        check_x2apic();
 882
 883        /* How many end-of-memory variables you have, grandma! */
 884        /* need this before calling reserve_initrd */
 885        if (max_pfn > (1UL<<(32 - PAGE_SHIFT)))
 886                max_low_pfn = e820_end_of_low_ram_pfn();
 887        else
 888                max_low_pfn = max_pfn;
 889
 890        high_memory = (void *)__va(max_pfn * PAGE_SIZE - 1) + 1;
 891        max_pfn_mapped = KERNEL_IMAGE_SIZE >> PAGE_SHIFT;
 892#endif
 893
 894#ifdef CONFIG_X86_CHECK_BIOS_CORRUPTION
 895        setup_bios_corruption_check();
 896#endif
 897
 898        printk(KERN_DEBUG "initial memory mapped : 0 - %08lx\n",
 899                        max_pfn_mapped<<PAGE_SHIFT);
 900
 901        reserve_brk();
 902
 903        init_gbpages();
 904
 905        /* max_pfn_mapped is updated here */
 906        max_low_pfn_mapped = init_memory_mapping(0, max_low_pfn<<PAGE_SHIFT);
 907        max_pfn_mapped = max_low_pfn_mapped;
 908
 909#ifdef CONFIG_X86_64
 910        if (max_pfn > max_low_pfn) {
 911                max_pfn_mapped = init_memory_mapping(1UL<<32,
 912                                                     max_pfn<<PAGE_SHIFT);
 913                /* can we preseve max_low_pfn ?*/
 914                max_low_pfn = max_pfn;
 915        }
 916#endif
 917
 918        /*
 919         * NOTE: On x86-32, only from this point on, fixmaps are ready for use.
 920         */
 921
 922#ifdef CONFIG_PROVIDE_OHCI1394_DMA_INIT
 923        if (init_ohci1394_dma_early)
 924                init_ohci1394_dma_on_all_controllers();
 925#endif
 926
 927        reserve_initrd();
 928
 929        vsmp_init();
 930
 931        io_delay_init();
 932
 933        /*
 934         * Parse the ACPI tables for possible boot-time SMP configuration.
 935         */
 936        acpi_boot_table_init();
 937
 938        early_acpi_boot_init();
 939
 940#ifdef CONFIG_ACPI_NUMA
 941        /*
 942         * Parse SRAT to discover nodes.
 943         */
 944        acpi_numa_init();
 945#endif
 946
 947        initmem_init(0, max_pfn);
 948
 949#ifdef CONFIG_ACPI_SLEEP
 950        /*
 951         * Reserve low memory region for sleep support.
 952         */
 953        acpi_reserve_bootmem();
 954#endif
 955        /*
 956         * Find and reserve possible boot-time SMP configuration:
 957         */
 958        find_smp_config();
 959
 960        reserve_crashkernel();
 961
 962#ifdef CONFIG_X86_64
 963        /*
 964         * dma32_reserve_bootmem() allocates bootmem which may conflict
 965         * with the crashkernel command line, so do that after
 966         * reserve_crashkernel()
 967         */
 968        dma32_reserve_bootmem();
 969#endif
 970
 971        reserve_ibft_region();
 972
 973#ifdef CONFIG_KVM_CLOCK
 974        kvmclock_init();
 975#endif
 976
 977        x86_init.paging.pagetable_setup_start(swapper_pg_dir);
 978        paging_init();
 979        x86_init.paging.pagetable_setup_done(swapper_pg_dir);
 980
 981        tboot_probe();
 982
 983#ifdef CONFIG_X86_64
 984        map_vsyscall();
 985#endif
 986
 987        generic_apic_probe();
 988
 989        early_quirks();
 990
 991        /*
 992         * Read APIC and some other early information from ACPI tables.
 993         */
 994        acpi_boot_init();
 995
 996        sfi_init();
 997
 998        /*
 999         * get boot-time SMP configuration:
1000         */
1001        if (smp_found_config)
1002                get_smp_config();
1003
1004        prefill_possible_map();
1005
1006#ifdef CONFIG_X86_64
1007        init_cpu_to_node();
1008#endif
1009
1010        init_apic_mappings();
1011        ioapic_init_mappings();
1012
1013        /* need to wait for io_apic is mapped */
1014        probe_nr_irqs_gsi();
1015
1016        kvm_guest_init();
1017
1018        e820_reserve_resources();
1019        e820_mark_nosave_regions(max_low_pfn);
1020
1021        x86_init.resources.reserve_resources();
1022
1023        e820_setup_gap();
1024
1025#ifdef CONFIG_VT
1026#if defined(CONFIG_VGA_CONSOLE)
1027        if (!efi_enabled || (efi_mem_type(0xa0000) != EFI_CONVENTIONAL_MEMORY))
1028                conswitchp = &vga_con;
1029#elif defined(CONFIG_DUMMY_CONSOLE)
1030        conswitchp = &dummy_con;
1031#endif
1032#endif
1033        x86_init.oem.banner();
1034}
1035
1036#ifdef CONFIG_X86_32
1037
1038static struct resource video_ram_resource = {
1039        .name   = "Video RAM area",
1040        .start  = 0xa0000,
1041        .end    = 0xbffff,
1042        .flags  = IORESOURCE_BUSY | IORESOURCE_MEM
1043};
1044
1045void __init i386_reserve_resources(void)
1046{
1047        request_resource(&iomem_resource, &video_ram_resource);
1048        reserve_standard_io_resources();
1049}
1050
1051#endif /* CONFIG_X86_32 */
1052