]>
Commit | Line | Data |
---|---|---|
1 | /* | |
2 | * Copyright (C) 1995 Linus Torvalds | |
3 | * | |
4 | * Support of BIGMEM added by Gerhard Wichert, Siemens AG, July 1999 | |
5 | * | |
6 | * Memory region support | |
7 | * David Parsons <orc@pell.chi.il.us>, July-August 1999 | |
8 | * | |
9 | * Added E820 sanitization routine (removes overlapping memory regions); | |
10 | * Brian Moyle <bmoyle@mvista.com>, February 2001 | |
11 | * | |
12 | * Moved CPU detection code to cpu/${cpu}.c | |
13 | * Patrick Mochel <mochel@osdl.org>, March 2002 | |
14 | * | |
15 | * Provisions for empty E820 memory regions (reported by certain BIOSes). | |
16 | * Alex Achenbach <xela@slit.de>, December 2002. | |
17 | * | |
18 | */ | |
19 | ||
20 | /* | |
21 | * This file handles the architecture-dependent parts of initialization | |
22 | */ | |
23 | ||
24 | #include <linux/sched.h> | |
25 | #include <linux/mm.h> | |
26 | #include <linux/mmzone.h> | |
27 | #include <linux/screen_info.h> | |
28 | #include <linux/ioport.h> | |
29 | #include <linux/acpi.h> | |
30 | #include <linux/sfi.h> | |
31 | #include <linux/apm_bios.h> | |
32 | #include <linux/initrd.h> | |
33 | #include <linux/bootmem.h> | |
34 | #include <linux/seq_file.h> | |
35 | #include <linux/console.h> | |
36 | #include <linux/mca.h> | |
37 | #include <linux/root_dev.h> | |
38 | #include <linux/highmem.h> | |
39 | #include <linux/module.h> | |
40 | #include <linux/efi.h> | |
41 | #include <linux/init.h> | |
42 | #include <linux/edd.h> | |
43 | #include <linux/iscsi_ibft.h> | |
44 | #include <linux/nodemask.h> | |
45 | #include <linux/kexec.h> | |
46 | #include <linux/dmi.h> | |
47 | #include <linux/pfn.h> | |
48 | #include <linux/pci.h> | |
49 | #include <asm/pci-direct.h> | |
50 | #include <linux/init_ohci1394_dma.h> | |
51 | #include <linux/kvm_para.h> | |
52 | ||
53 | #include <linux/errno.h> | |
54 | #include <linux/kernel.h> | |
55 | #include <linux/stddef.h> | |
56 | #include <linux/unistd.h> | |
57 | #include <linux/ptrace.h> | |
58 | #include <linux/slab.h> | |
59 | #include <linux/user.h> | |
60 | #include <linux/delay.h> | |
61 | ||
62 | #include <linux/kallsyms.h> | |
63 | #include <linux/cpufreq.h> | |
64 | #include <linux/dma-mapping.h> | |
65 | #include <linux/ctype.h> | |
66 | #include <linux/uaccess.h> | |
67 | ||
68 | #include <linux/percpu.h> | |
69 | #include <linux/crash_dump.h> | |
70 | #include <linux/tboot.h> | |
71 | ||
72 | #include <video/edid.h> | |
73 | ||
74 | #include <asm/mtrr.h> | |
75 | #include <asm/apic.h> | |
76 | #include <asm/e820.h> | |
77 | #include <asm/mpspec.h> | |
78 | #include <asm/setup.h> | |
79 | #include <asm/efi.h> | |
80 | #include <asm/timer.h> | |
81 | #include <asm/i8259.h> | |
82 | #include <asm/sections.h> | |
83 | #include <asm/dmi.h> | |
84 | #include <asm/io_apic.h> | |
85 | #include <asm/ist.h> | |
86 | #include <asm/vmi.h> | |
87 | #include <asm/setup_arch.h> | |
88 | #include <asm/bios_ebda.h> | |
89 | #include <asm/cacheflush.h> | |
90 | #include <asm/processor.h> | |
91 | #include <asm/bugs.h> | |
92 | ||
93 | #include <asm/system.h> | |
94 | #include <asm/vsyscall.h> | |
95 | #include <asm/cpu.h> | |
96 | #include <asm/desc.h> | |
97 | #include <asm/dma.h> | |
98 | #include <asm/iommu.h> | |
99 | #include <asm/gart.h> | |
100 | #include <asm/mmu_context.h> | |
101 | #include <asm/proto.h> | |
102 | ||
103 | #include <asm/paravirt.h> | |
104 | #include <asm/hypervisor.h> | |
105 | ||
106 | #include <asm/percpu.h> | |
107 | #include <asm/topology.h> | |
108 | #include <asm/apicdef.h> | |
109 | #include <asm/k8.h> | |
110 | #ifdef CONFIG_X86_64 | |
111 | #include <asm/numa_64.h> | |
112 | #endif | |
113 | ||
114 | /* | |
115 | * end_pfn only includes RAM, while max_pfn_mapped includes all e820 entries. | |
116 | * The direct mapping extends to max_pfn_mapped, so that we can directly access | |
117 | * apertures, ACPI and other tables without having to play with fixmaps. | |
118 | */ | |
119 | unsigned long max_low_pfn_mapped; | |
120 | unsigned long max_pfn_mapped; | |
121 | ||
122 | RESERVE_BRK(dmi_alloc, 65536); | |
123 | ||
124 | unsigned int boot_cpu_id __read_mostly; | |
125 | ||
126 | static __initdata unsigned long _brk_start = (unsigned long)__brk_base; | |
127 | unsigned long _brk_end = (unsigned long)__brk_base; | |
128 | ||
129 | #ifdef CONFIG_X86_64 | |
130 | int default_cpu_present_to_apicid(int mps_cpu) | |
131 | { | |
132 | return __default_cpu_present_to_apicid(mps_cpu); | |
133 | } | |
134 | ||
135 | int default_check_phys_apicid_present(int phys_apicid) | |
136 | { | |
137 | return __default_check_phys_apicid_present(phys_apicid); | |
138 | } | |
139 | #endif | |
140 | ||
141 | #ifndef CONFIG_DEBUG_BOOT_PARAMS | |
142 | struct boot_params __initdata boot_params; | |
143 | #else | |
144 | struct boot_params boot_params; | |
145 | #endif | |
146 | ||
147 | /* | |
148 | * Machine setup.. | |
149 | */ | |
150 | static struct resource data_resource = { | |
151 | .name = "Kernel data", | |
152 | .start = 0, | |
153 | .end = 0, | |
154 | .flags = IORESOURCE_BUSY | IORESOURCE_MEM | |
155 | }; | |
156 | ||
157 | static struct resource code_resource = { | |
158 | .name = "Kernel code", | |
159 | .start = 0, | |
160 | .end = 0, | |
161 | .flags = IORESOURCE_BUSY | IORESOURCE_MEM | |
162 | }; | |
163 | ||
164 | static struct resource bss_resource = { | |
165 | .name = "Kernel bss", | |
166 | .start = 0, | |
167 | .end = 0, | |
168 | .flags = IORESOURCE_BUSY | IORESOURCE_MEM | |
169 | }; | |
170 | ||
171 | ||
172 | #ifdef CONFIG_X86_32 | |
173 | /* cpu data as detected by the assembly code in head.S */ | |
174 | struct cpuinfo_x86 new_cpu_data __cpuinitdata = {0, 0, 0, 0, -1, 1, 0, 0, -1}; | |
175 | /* common cpu data for all cpus */ | |
176 | struct cpuinfo_x86 boot_cpu_data __read_mostly = {0, 0, 0, 0, -1, 1, 0, 0, -1}; | |
177 | EXPORT_SYMBOL(boot_cpu_data); | |
178 | static void set_mca_bus(int x) | |
179 | { | |
180 | #ifdef CONFIG_MCA | |
181 | MCA_bus = x; | |
182 | #endif | |
183 | } | |
184 | ||
185 | unsigned int def_to_bigsmp; | |
186 | ||
187 | /* for MCA, but anyone else can use it if they want */ | |
188 | unsigned int machine_id; | |
189 | unsigned int machine_submodel_id; | |
190 | unsigned int BIOS_revision; | |
191 | ||
192 | struct apm_info apm_info; | |
193 | EXPORT_SYMBOL(apm_info); | |
194 | ||
195 | #if defined(CONFIG_X86_SPEEDSTEP_SMI) || \ | |
196 | defined(CONFIG_X86_SPEEDSTEP_SMI_MODULE) | |
197 | struct ist_info ist_info; | |
198 | EXPORT_SYMBOL(ist_info); | |
199 | #else | |
200 | struct ist_info ist_info; | |
201 | #endif | |
202 | ||
203 | #else | |
204 | struct cpuinfo_x86 boot_cpu_data __read_mostly = { | |
205 | .x86_phys_bits = MAX_PHYSMEM_BITS, | |
206 | }; | |
207 | EXPORT_SYMBOL(boot_cpu_data); | |
208 | #endif | |
209 | ||
210 | ||
211 | #if !defined(CONFIG_X86_PAE) || defined(CONFIG_X86_64) | |
212 | unsigned long mmu_cr4_features; | |
213 | #else | |
214 | unsigned long mmu_cr4_features = X86_CR4_PAE; | |
215 | #endif | |
216 | ||
217 | /* Boot loader ID and version as integers, for the benefit of proc_dointvec */ | |
218 | int bootloader_type, bootloader_version; | |
219 | ||
220 | /* | |
221 | * Setup options | |
222 | */ | |
223 | struct screen_info screen_info; | |
224 | EXPORT_SYMBOL(screen_info); | |
225 | struct edid_info edid_info; | |
226 | EXPORT_SYMBOL_GPL(edid_info); | |
227 | ||
228 | extern int root_mountflags; | |
229 | ||
230 | unsigned long saved_video_mode; | |
231 | ||
232 | #define RAMDISK_IMAGE_START_MASK 0x07FF | |
233 | #define RAMDISK_PROMPT_FLAG 0x8000 | |
234 | #define RAMDISK_LOAD_FLAG 0x4000 | |
235 | ||
236 | static char __initdata command_line[COMMAND_LINE_SIZE]; | |
237 | #ifdef CONFIG_CMDLINE_BOOL | |
238 | static char __initdata builtin_cmdline[COMMAND_LINE_SIZE] = CONFIG_CMDLINE; | |
239 | #endif | |
240 | ||
241 | #if defined(CONFIG_EDD) || defined(CONFIG_EDD_MODULE) | |
242 | struct edd edd; | |
243 | #ifdef CONFIG_EDD_MODULE | |
244 | EXPORT_SYMBOL(edd); | |
245 | #endif | |
246 | /** | |
247 | * copy_edd() - Copy the BIOS EDD information | |
248 | * from boot_params into a safe place. | |
249 | * | |
250 | */ | |
251 | static inline void copy_edd(void) | |
252 | { | |
253 | memcpy(edd.mbr_signature, boot_params.edd_mbr_sig_buffer, | |
254 | sizeof(edd.mbr_signature)); | |
255 | memcpy(edd.edd_info, boot_params.eddbuf, sizeof(edd.edd_info)); | |
256 | edd.mbr_signature_nr = boot_params.edd_mbr_sig_buf_entries; | |
257 | edd.edd_info_nr = boot_params.eddbuf_entries; | |
258 | } | |
259 | #else | |
260 | static inline void copy_edd(void) | |
261 | { | |
262 | } | |
263 | #endif | |
264 | ||
265 | void * __init extend_brk(size_t size, size_t align) | |
266 | { | |
267 | size_t mask = align - 1; | |
268 | void *ret; | |
269 | ||
270 | BUG_ON(_brk_start == 0); | |
271 | BUG_ON(align & mask); | |
272 | ||
273 | _brk_end = (_brk_end + mask) & ~mask; | |
274 | BUG_ON((char *)(_brk_end + size) > __brk_limit); | |
275 | ||
276 | ret = (void *)_brk_end; | |
277 | _brk_end += size; | |
278 | ||
279 | memset(ret, 0, size); | |
280 | ||
281 | return ret; | |
282 | } | |
283 | ||
284 | #ifdef CONFIG_X86_64 | |
285 | static void __init init_gbpages(void) | |
286 | { | |
287 | if (direct_gbpages && cpu_has_gbpages) | |
288 | printk(KERN_INFO "Using GB pages for direct mapping\n"); | |
289 | else | |
290 | direct_gbpages = 0; | |
291 | } | |
292 | #else | |
293 | static inline void init_gbpages(void) | |
294 | { | |
295 | } | |
296 | #endif | |
297 | ||
298 | static void __init reserve_brk(void) | |
299 | { | |
300 | if (_brk_end > _brk_start) | |
301 | reserve_early(__pa(_brk_start), __pa(_brk_end), "BRK"); | |
302 | ||
303 | /* Mark brk area as locked down and no longer taking any | |
304 | new allocations */ | |
305 | _brk_start = 0; | |
306 | } | |
307 | ||
308 | #ifdef CONFIG_BLK_DEV_INITRD | |
309 | ||
310 | #define MAX_MAP_CHUNK (NR_FIX_BTMAPS << PAGE_SHIFT) | |
311 | static void __init relocate_initrd(void) | |
312 | { | |
313 | ||
314 | u64 ramdisk_image = boot_params.hdr.ramdisk_image; | |
315 | u64 ramdisk_size = boot_params.hdr.ramdisk_size; | |
316 | u64 end_of_lowmem = max_low_pfn_mapped << PAGE_SHIFT; | |
317 | u64 ramdisk_here; | |
318 | unsigned long slop, clen, mapaddr; | |
319 | char *p, *q; | |
320 | ||
321 | /* We need to move the initrd down into lowmem */ | |
322 | ramdisk_here = find_e820_area(0, end_of_lowmem, ramdisk_size, | |
323 | PAGE_SIZE); | |
324 | ||
325 | if (ramdisk_here == -1ULL) | |
326 | panic("Cannot find place for new RAMDISK of size %lld\n", | |
327 | ramdisk_size); | |
328 | ||
329 | /* Note: this includes all the lowmem currently occupied by | |
330 | the initrd, we rely on that fact to keep the data intact. */ | |
331 | reserve_early(ramdisk_here, ramdisk_here + ramdisk_size, | |
332 | "NEW RAMDISK"); | |
333 | initrd_start = ramdisk_here + PAGE_OFFSET; | |
334 | initrd_end = initrd_start + ramdisk_size; | |
335 | printk(KERN_INFO "Allocated new RAMDISK: %08llx - %08llx\n", | |
336 | ramdisk_here, ramdisk_here + ramdisk_size); | |
337 | ||
338 | q = (char *)initrd_start; | |
339 | ||
340 | /* Copy any lowmem portion of the initrd */ | |
341 | if (ramdisk_image < end_of_lowmem) { | |
342 | clen = end_of_lowmem - ramdisk_image; | |
343 | p = (char *)__va(ramdisk_image); | |
344 | memcpy(q, p, clen); | |
345 | q += clen; | |
346 | ramdisk_image += clen; | |
347 | ramdisk_size -= clen; | |
348 | } | |
349 | ||
350 | /* Copy the highmem portion of the initrd */ | |
351 | while (ramdisk_size) { | |
352 | slop = ramdisk_image & ~PAGE_MASK; | |
353 | clen = ramdisk_size; | |
354 | if (clen > MAX_MAP_CHUNK-slop) | |
355 | clen = MAX_MAP_CHUNK-slop; | |
356 | mapaddr = ramdisk_image & PAGE_MASK; | |
357 | p = early_memremap(mapaddr, clen+slop); | |
358 | memcpy(q, p+slop, clen); | |
359 | early_iounmap(p, clen+slop); | |
360 | q += clen; | |
361 | ramdisk_image += clen; | |
362 | ramdisk_size -= clen; | |
363 | } | |
364 | /* high pages is not converted by early_res_to_bootmem */ | |
365 | ramdisk_image = boot_params.hdr.ramdisk_image; | |
366 | ramdisk_size = boot_params.hdr.ramdisk_size; | |
367 | printk(KERN_INFO "Move RAMDISK from %016llx - %016llx to" | |
368 | " %08llx - %08llx\n", | |
369 | ramdisk_image, ramdisk_image + ramdisk_size - 1, | |
370 | ramdisk_here, ramdisk_here + ramdisk_size - 1); | |
371 | } | |
372 | ||
373 | static void __init reserve_initrd(void) | |
374 | { | |
375 | u64 ramdisk_image = boot_params.hdr.ramdisk_image; | |
376 | u64 ramdisk_size = boot_params.hdr.ramdisk_size; | |
377 | u64 ramdisk_end = ramdisk_image + ramdisk_size; | |
378 | u64 end_of_lowmem = max_low_pfn_mapped << PAGE_SHIFT; | |
379 | ||
380 | if (!boot_params.hdr.type_of_loader || | |
381 | !ramdisk_image || !ramdisk_size) | |
382 | return; /* No initrd provided by bootloader */ | |
383 | ||
384 | initrd_start = 0; | |
385 | ||
386 | if (ramdisk_size >= (end_of_lowmem>>1)) { | |
387 | free_early(ramdisk_image, ramdisk_end); | |
388 | printk(KERN_ERR "initrd too large to handle, " | |
389 | "disabling initrd\n"); | |
390 | return; | |
391 | } | |
392 | ||
393 | printk(KERN_INFO "RAMDISK: %08llx - %08llx\n", ramdisk_image, | |
394 | ramdisk_end); | |
395 | ||
396 | ||
397 | if (ramdisk_end <= end_of_lowmem) { | |
398 | /* All in lowmem, easy case */ | |
399 | /* | |
400 | * don't need to reserve again, already reserved early | |
401 | * in i386_start_kernel | |
402 | */ | |
403 | initrd_start = ramdisk_image + PAGE_OFFSET; | |
404 | initrd_end = initrd_start + ramdisk_size; | |
405 | return; | |
406 | } | |
407 | ||
408 | relocate_initrd(); | |
409 | ||
410 | free_early(ramdisk_image, ramdisk_end); | |
411 | } | |
412 | #else | |
413 | static void __init reserve_initrd(void) | |
414 | { | |
415 | } | |
416 | #endif /* CONFIG_BLK_DEV_INITRD */ | |
417 | ||
418 | static void __init parse_setup_data(void) | |
419 | { | |
420 | struct setup_data *data; | |
421 | u64 pa_data; | |
422 | ||
423 | if (boot_params.hdr.version < 0x0209) | |
424 | return; | |
425 | pa_data = boot_params.hdr.setup_data; | |
426 | while (pa_data) { | |
427 | data = early_memremap(pa_data, PAGE_SIZE); | |
428 | switch (data->type) { | |
429 | case SETUP_E820_EXT: | |
430 | parse_e820_ext(data, pa_data); | |
431 | break; | |
432 | default: | |
433 | break; | |
434 | } | |
435 | pa_data = data->next; | |
436 | early_iounmap(data, PAGE_SIZE); | |
437 | } | |
438 | } | |
439 | ||
440 | static void __init e820_reserve_setup_data(void) | |
441 | { | |
442 | struct setup_data *data; | |
443 | u64 pa_data; | |
444 | int found = 0; | |
445 | ||
446 | if (boot_params.hdr.version < 0x0209) | |
447 | return; | |
448 | pa_data = boot_params.hdr.setup_data; | |
449 | while (pa_data) { | |
450 | data = early_memremap(pa_data, sizeof(*data)); | |
451 | e820_update_range(pa_data, sizeof(*data)+data->len, | |
452 | E820_RAM, E820_RESERVED_KERN); | |
453 | found = 1; | |
454 | pa_data = data->next; | |
455 | early_iounmap(data, sizeof(*data)); | |
456 | } | |
457 | if (!found) | |
458 | return; | |
459 | ||
460 | sanitize_e820_map(e820.map, ARRAY_SIZE(e820.map), &e820.nr_map); | |
461 | memcpy(&e820_saved, &e820, sizeof(struct e820map)); | |
462 | printk(KERN_INFO "extended physical RAM map:\n"); | |
463 | e820_print_map("reserve setup_data"); | |
464 | } | |
465 | ||
466 | static void __init reserve_early_setup_data(void) | |
467 | { | |
468 | struct setup_data *data; | |
469 | u64 pa_data; | |
470 | char buf[32]; | |
471 | ||
472 | if (boot_params.hdr.version < 0x0209) | |
473 | return; | |
474 | pa_data = boot_params.hdr.setup_data; | |
475 | while (pa_data) { | |
476 | data = early_memremap(pa_data, sizeof(*data)); | |
477 | sprintf(buf, "setup data %x", data->type); | |
478 | reserve_early(pa_data, pa_data+sizeof(*data)+data->len, buf); | |
479 | pa_data = data->next; | |
480 | early_iounmap(data, sizeof(*data)); | |
481 | } | |
482 | } | |
483 | ||
484 | /* | |
485 | * --------- Crashkernel reservation ------------------------------ | |
486 | */ | |
487 | ||
488 | #ifdef CONFIG_KEXEC | |
489 | ||
490 | static inline unsigned long long get_total_mem(void) | |
491 | { | |
492 | unsigned long long total; | |
493 | ||
494 | total = max_pfn - min_low_pfn; | |
495 | ||
496 | return total << PAGE_SHIFT; | |
497 | } | |
498 | ||
499 | static void __init reserve_crashkernel(void) | |
500 | { | |
501 | unsigned long long total_mem; | |
502 | unsigned long long crash_size, crash_base; | |
503 | int ret; | |
504 | ||
505 | total_mem = get_total_mem(); | |
506 | ||
507 | ret = parse_crashkernel(boot_command_line, total_mem, | |
508 | &crash_size, &crash_base); | |
509 | if (ret != 0 || crash_size <= 0) | |
510 | return; | |
511 | ||
512 | /* 0 means: find the address automatically */ | |
513 | if (crash_base <= 0) { | |
514 | const unsigned long long alignment = 16<<20; /* 16M */ | |
515 | ||
516 | crash_base = find_e820_area(alignment, ULONG_MAX, crash_size, | |
517 | alignment); | |
518 | if (crash_base == -1ULL) { | |
519 | pr_info("crashkernel reservation failed - No suitable area found.\n"); | |
520 | return; | |
521 | } | |
522 | } else { | |
523 | unsigned long long start; | |
524 | ||
525 | start = find_e820_area(crash_base, ULONG_MAX, crash_size, | |
526 | 1<<20); | |
527 | if (start != crash_base) { | |
528 | pr_info("crashkernel reservation failed - memory is in use.\n"); | |
529 | return; | |
530 | } | |
531 | } | |
532 | reserve_early(crash_base, crash_base + crash_size, "CRASH KERNEL"); | |
533 | ||
534 | printk(KERN_INFO "Reserving %ldMB of memory at %ldMB " | |
535 | "for crashkernel (System RAM: %ldMB)\n", | |
536 | (unsigned long)(crash_size >> 20), | |
537 | (unsigned long)(crash_base >> 20), | |
538 | (unsigned long)(total_mem >> 20)); | |
539 | ||
540 | crashk_res.start = crash_base; | |
541 | crashk_res.end = crash_base + crash_size - 1; | |
542 | insert_resource(&iomem_resource, &crashk_res); | |
543 | } | |
544 | #else | |
545 | static void __init reserve_crashkernel(void) | |
546 | { | |
547 | } | |
548 | #endif | |
549 | ||
550 | static struct resource standard_io_resources[] = { | |
551 | { .name = "dma1", .start = 0x00, .end = 0x1f, | |
552 | .flags = IORESOURCE_BUSY | IORESOURCE_IO }, | |
553 | { .name = "pic1", .start = 0x20, .end = 0x21, | |
554 | .flags = IORESOURCE_BUSY | IORESOURCE_IO }, | |
555 | { .name = "timer0", .start = 0x40, .end = 0x43, | |
556 | .flags = IORESOURCE_BUSY | IORESOURCE_IO }, | |
557 | { .name = "timer1", .start = 0x50, .end = 0x53, | |
558 | .flags = IORESOURCE_BUSY | IORESOURCE_IO }, | |
559 | { .name = "keyboard", .start = 0x60, .end = 0x60, | |
560 | .flags = IORESOURCE_BUSY | IORESOURCE_IO }, | |
561 | { .name = "keyboard", .start = 0x64, .end = 0x64, | |
562 | .flags = IORESOURCE_BUSY | IORESOURCE_IO }, | |
563 | { .name = "dma page reg", .start = 0x80, .end = 0x8f, | |
564 | .flags = IORESOURCE_BUSY | IORESOURCE_IO }, | |
565 | { .name = "pic2", .start = 0xa0, .end = 0xa1, | |
566 | .flags = IORESOURCE_BUSY | IORESOURCE_IO }, | |
567 | { .name = "dma2", .start = 0xc0, .end = 0xdf, | |
568 | .flags = IORESOURCE_BUSY | IORESOURCE_IO }, | |
569 | { .name = "fpu", .start = 0xf0, .end = 0xff, | |
570 | .flags = IORESOURCE_BUSY | IORESOURCE_IO } | |
571 | }; | |
572 | ||
573 | void __init reserve_standard_io_resources(void) | |
574 | { | |
575 | int i; | |
576 | ||
577 | /* request I/O space for devices used on all i[345]86 PCs */ | |
578 | for (i = 0; i < ARRAY_SIZE(standard_io_resources); i++) | |
579 | request_resource(&ioport_resource, &standard_io_resources[i]); | |
580 | ||
581 | } | |
582 | ||
583 | /* | |
584 | * Note: elfcorehdr_addr is not just limited to vmcore. It is also used by | |
585 | * is_kdump_kernel() to determine if we are booting after a panic. Hence | |
586 | * ifdef it under CONFIG_CRASH_DUMP and not CONFIG_PROC_VMCORE. | |
587 | */ | |
588 | ||
589 | #ifdef CONFIG_CRASH_DUMP | |
590 | /* elfcorehdr= specifies the location of elf core header | |
591 | * stored by the crashed kernel. This option will be passed | |
592 | * by kexec loader to the capture kernel. | |
593 | */ | |
594 | static int __init setup_elfcorehdr(char *arg) | |
595 | { | |
596 | char *end; | |
597 | if (!arg) | |
598 | return -EINVAL; | |
599 | elfcorehdr_addr = memparse(arg, &end); | |
600 | return end > arg ? 0 : -EINVAL; | |
601 | } | |
602 | early_param("elfcorehdr", setup_elfcorehdr); | |
603 | #endif | |
604 | ||
605 | #ifdef CONFIG_X86_RESERVE_LOW_64K | |
606 | static int __init dmi_low_memory_corruption(const struct dmi_system_id *d) | |
607 | { | |
608 | printk(KERN_NOTICE | |
609 | "%s detected: BIOS may corrupt low RAM, working around it.\n", | |
610 | d->ident); | |
611 | ||
612 | e820_update_range(0, 0x10000, E820_RAM, E820_RESERVED); | |
613 | sanitize_e820_map(e820.map, ARRAY_SIZE(e820.map), &e820.nr_map); | |
614 | ||
615 | return 0; | |
616 | } | |
617 | #endif | |
618 | ||
619 | /* List of systems that have known low memory corruption BIOS problems */ | |
620 | static struct dmi_system_id __initdata bad_bios_dmi_table[] = { | |
621 | #ifdef CONFIG_X86_RESERVE_LOW_64K | |
622 | { | |
623 | .callback = dmi_low_memory_corruption, | |
624 | .ident = "AMI BIOS", | |
625 | .matches = { | |
626 | DMI_MATCH(DMI_BIOS_VENDOR, "American Megatrends Inc."), | |
627 | }, | |
628 | }, | |
629 | { | |
630 | .callback = dmi_low_memory_corruption, | |
631 | .ident = "Phoenix BIOS", | |
632 | .matches = { | |
633 | DMI_MATCH(DMI_BIOS_VENDOR, "Phoenix Technologies"), | |
634 | }, | |
635 | }, | |
636 | { | |
637 | /* | |
638 | * AMI BIOS with low memory corruption was found on Intel DG45ID board. | |
639 | * It hase different DMI_BIOS_VENDOR = "Intel Corp.", for now we will | |
640 | * match only DMI_BOARD_NAME and see if there is more bad products | |
641 | * with this vendor. | |
642 | */ | |
643 | .callback = dmi_low_memory_corruption, | |
644 | .ident = "AMI BIOS", | |
645 | .matches = { | |
646 | DMI_MATCH(DMI_BOARD_NAME, "DG45ID"), | |
647 | }, | |
648 | }, | |
649 | #endif | |
650 | {} | |
651 | }; | |
652 | ||
653 | /* | |
654 | * Determine if we were loaded by an EFI loader. If so, then we have also been | |
655 | * passed the efi memmap, systab, etc., so we should use these data structures | |
656 | * for initialization. Note, the efi init code path is determined by the | |
657 | * global efi_enabled. This allows the same kernel image to be used on existing | |
658 | * systems (with a traditional BIOS) as well as on EFI systems. | |
659 | */ | |
660 | /* | |
661 | * setup_arch - architecture-specific boot-time initializations | |
662 | * | |
663 | * Note: On x86_64, fixmaps are ready for use even before this is called. | |
664 | */ | |
665 | ||
666 | void __init setup_arch(char **cmdline_p) | |
667 | { | |
668 | int acpi = 0; | |
669 | int k8 = 0; | |
670 | ||
671 | #ifdef CONFIG_X86_32 | |
672 | memcpy(&boot_cpu_data, &new_cpu_data, sizeof(new_cpu_data)); | |
673 | visws_early_detect(); | |
674 | #else | |
675 | printk(KERN_INFO "Command line: %s\n", boot_command_line); | |
676 | #endif | |
677 | ||
678 | /* VMI may relocate the fixmap; do this before touching ioremap area */ | |
679 | vmi_init(); | |
680 | ||
681 | early_cpu_init(); | |
682 | early_ioremap_init(); | |
683 | ||
684 | ROOT_DEV = old_decode_dev(boot_params.hdr.root_dev); | |
685 | screen_info = boot_params.screen_info; | |
686 | edid_info = boot_params.edid_info; | |
687 | #ifdef CONFIG_X86_32 | |
688 | apm_info.bios = boot_params.apm_bios_info; | |
689 | ist_info = boot_params.ist_info; | |
690 | if (boot_params.sys_desc_table.length != 0) { | |
691 | set_mca_bus(boot_params.sys_desc_table.table[3] & 0x2); | |
692 | machine_id = boot_params.sys_desc_table.table[0]; | |
693 | machine_submodel_id = boot_params.sys_desc_table.table[1]; | |
694 | BIOS_revision = boot_params.sys_desc_table.table[2]; | |
695 | } | |
696 | #endif | |
697 | saved_video_mode = boot_params.hdr.vid_mode; | |
698 | bootloader_type = boot_params.hdr.type_of_loader; | |
699 | if ((bootloader_type >> 4) == 0xe) { | |
700 | bootloader_type &= 0xf; | |
701 | bootloader_type |= (boot_params.hdr.ext_loader_type+0x10) << 4; | |
702 | } | |
703 | bootloader_version = bootloader_type & 0xf; | |
704 | bootloader_version |= boot_params.hdr.ext_loader_ver << 4; | |
705 | ||
706 | #ifdef CONFIG_BLK_DEV_RAM | |
707 | rd_image_start = boot_params.hdr.ram_size & RAMDISK_IMAGE_START_MASK; | |
708 | rd_prompt = ((boot_params.hdr.ram_size & RAMDISK_PROMPT_FLAG) != 0); | |
709 | rd_doload = ((boot_params.hdr.ram_size & RAMDISK_LOAD_FLAG) != 0); | |
710 | #endif | |
711 | #ifdef CONFIG_EFI | |
712 | if (!strncmp((char *)&boot_params.efi_info.efi_loader_signature, | |
713 | #ifdef CONFIG_X86_32 | |
714 | "EL32", | |
715 | #else | |
716 | "EL64", | |
717 | #endif | |
718 | 4)) { | |
719 | efi_enabled = 1; | |
720 | efi_reserve_early(); | |
721 | } | |
722 | #endif | |
723 | ||
724 | x86_init.oem.arch_setup(); | |
725 | ||
726 | setup_memory_map(); | |
727 | parse_setup_data(); | |
728 | /* update the e820_saved too */ | |
729 | e820_reserve_setup_data(); | |
730 | ||
731 | copy_edd(); | |
732 | ||
733 | if (!boot_params.hdr.root_flags) | |
734 | root_mountflags &= ~MS_RDONLY; | |
735 | init_mm.start_code = (unsigned long) _text; | |
736 | init_mm.end_code = (unsigned long) _etext; | |
737 | init_mm.end_data = (unsigned long) _edata; | |
738 | init_mm.brk = _brk_end; | |
739 | ||
740 | code_resource.start = virt_to_phys(_text); | |
741 | code_resource.end = virt_to_phys(_etext)-1; | |
742 | data_resource.start = virt_to_phys(_etext); | |
743 | data_resource.end = virt_to_phys(_edata)-1; | |
744 | bss_resource.start = virt_to_phys(&__bss_start); | |
745 | bss_resource.end = virt_to_phys(&__bss_stop)-1; | |
746 | ||
747 | #ifdef CONFIG_CMDLINE_BOOL | |
748 | #ifdef CONFIG_CMDLINE_OVERRIDE | |
749 | strlcpy(boot_command_line, builtin_cmdline, COMMAND_LINE_SIZE); | |
750 | #else | |
751 | if (builtin_cmdline[0]) { | |
752 | /* append boot loader cmdline to builtin */ | |
753 | strlcat(builtin_cmdline, " ", COMMAND_LINE_SIZE); | |
754 | strlcat(builtin_cmdline, boot_command_line, COMMAND_LINE_SIZE); | |
755 | strlcpy(boot_command_line, builtin_cmdline, COMMAND_LINE_SIZE); | |
756 | } | |
757 | #endif | |
758 | #endif | |
759 | ||
760 | strlcpy(command_line, boot_command_line, COMMAND_LINE_SIZE); | |
761 | *cmdline_p = command_line; | |
762 | ||
763 | /* | |
764 | * x86_configure_nx() is called before parse_early_param() to detect | |
765 | * whether hardware doesn't support NX (so that the early EHCI debug | |
766 | * console setup can safely call set_fixmap()). It may then be called | |
767 | * again from within noexec_setup() during parsing early parameters | |
768 | * to honor the respective command line option. | |
769 | */ | |
770 | x86_configure_nx(); | |
771 | ||
772 | parse_early_param(); | |
773 | ||
774 | x86_report_nx(); | |
775 | ||
776 | /* Must be before kernel pagetables are setup */ | |
777 | vmi_activate(); | |
778 | ||
779 | /* after early param, so could get panic from serial */ | |
780 | reserve_early_setup_data(); | |
781 | ||
782 | if (acpi_mps_check()) { | |
783 | #ifdef CONFIG_X86_LOCAL_APIC | |
784 | disable_apic = 1; | |
785 | #endif | |
786 | setup_clear_cpu_cap(X86_FEATURE_APIC); | |
787 | } | |
788 | ||
789 | #ifdef CONFIG_PCI | |
790 | if (pci_early_dump_regs) | |
791 | early_dump_pci_devices(); | |
792 | #endif | |
793 | ||
794 | finish_e820_parsing(); | |
795 | ||
796 | if (efi_enabled) | |
797 | efi_init(); | |
798 | ||
799 | dmi_scan_machine(); | |
800 | ||
801 | dmi_check_system(bad_bios_dmi_table); | |
802 | ||
803 | /* | |
804 | * VMware detection requires dmi to be available, so this | |
805 | * needs to be done after dmi_scan_machine, for the BP. | |
806 | */ | |
807 | init_hypervisor_platform(); | |
808 | ||
809 | x86_init.resources.probe_roms(); | |
810 | ||
811 | /* after parse_early_param, so could debug it */ | |
812 | insert_resource(&iomem_resource, &code_resource); | |
813 | insert_resource(&iomem_resource, &data_resource); | |
814 | insert_resource(&iomem_resource, &bss_resource); | |
815 | ||
816 | ||
817 | #ifdef CONFIG_X86_32 | |
818 | if (ppro_with_ram_bug()) { | |
819 | e820_update_range(0x70000000ULL, 0x40000ULL, E820_RAM, | |
820 | E820_RESERVED); | |
821 | sanitize_e820_map(e820.map, ARRAY_SIZE(e820.map), &e820.nr_map); | |
822 | printk(KERN_INFO "fixed physical RAM map:\n"); | |
823 | e820_print_map("bad_ppro"); | |
824 | } | |
825 | #else | |
826 | early_gart_iommu_check(); | |
827 | #endif | |
828 | ||
829 | /* | |
830 | * partially used pages are not usable - thus | |
831 | * we are rounding upwards: | |
832 | */ | |
833 | max_pfn = e820_end_of_ram_pfn(); | |
834 | ||
835 | /* preallocate 4k for mptable mpc */ | |
836 | early_reserve_e820_mpc_new(); | |
837 | /* update e820 for memory not covered by WB MTRRs */ | |
838 | mtrr_bp_init(); | |
839 | if (mtrr_trim_uncached_memory(max_pfn)) | |
840 | max_pfn = e820_end_of_ram_pfn(); | |
841 | ||
842 | #ifdef CONFIG_X86_32 | |
843 | /* max_low_pfn get updated here */ | |
844 | find_low_pfn_range(); | |
845 | #else | |
846 | num_physpages = max_pfn; | |
847 | ||
848 | check_x2apic(); | |
849 | ||
850 | /* How many end-of-memory variables you have, grandma! */ | |
851 | /* need this before calling reserve_initrd */ | |
852 | if (max_pfn > (1UL<<(32 - PAGE_SHIFT))) | |
853 | max_low_pfn = e820_end_of_low_ram_pfn(); | |
854 | else | |
855 | max_low_pfn = max_pfn; | |
856 | ||
857 | high_memory = (void *)__va(max_pfn * PAGE_SIZE - 1) + 1; | |
858 | max_pfn_mapped = KERNEL_IMAGE_SIZE >> PAGE_SHIFT; | |
859 | #endif | |
860 | ||
861 | #ifdef CONFIG_X86_CHECK_BIOS_CORRUPTION | |
862 | setup_bios_corruption_check(); | |
863 | #endif | |
864 | ||
865 | printk(KERN_DEBUG "initial memory mapped : 0 - %08lx\n", | |
866 | max_pfn_mapped<<PAGE_SHIFT); | |
867 | ||
868 | reserve_brk(); | |
869 | ||
870 | #ifdef CONFIG_ACPI_SLEEP | |
871 | /* | |
872 | * Reserve low memory region for sleep support. | |
873 | * even before init_memory_mapping | |
874 | */ | |
875 | acpi_reserve_wakeup_memory(); | |
876 | #endif | |
877 | init_gbpages(); | |
878 | ||
879 | /* max_pfn_mapped is updated here */ | |
880 | max_low_pfn_mapped = init_memory_mapping(0, max_low_pfn<<PAGE_SHIFT); | |
881 | max_pfn_mapped = max_low_pfn_mapped; | |
882 | ||
883 | #ifdef CONFIG_X86_64 | |
884 | if (max_pfn > max_low_pfn) { | |
885 | max_pfn_mapped = init_memory_mapping(1UL<<32, | |
886 | max_pfn<<PAGE_SHIFT); | |
887 | /* can we preseve max_low_pfn ?*/ | |
888 | max_low_pfn = max_pfn; | |
889 | } | |
890 | #endif | |
891 | ||
892 | /* | |
893 | * NOTE: On x86-32, only from this point on, fixmaps are ready for use. | |
894 | */ | |
895 | ||
896 | #ifdef CONFIG_PROVIDE_OHCI1394_DMA_INIT | |
897 | if (init_ohci1394_dma_early) | |
898 | init_ohci1394_dma_on_all_controllers(); | |
899 | #endif | |
900 | ||
901 | reserve_initrd(); | |
902 | ||
903 | reserve_crashkernel(); | |
904 | ||
905 | vsmp_init(); | |
906 | ||
907 | io_delay_init(); | |
908 | ||
909 | /* | |
910 | * Parse the ACPI tables for possible boot-time SMP configuration. | |
911 | */ | |
912 | acpi_boot_table_init(); | |
913 | ||
914 | early_acpi_boot_init(); | |
915 | ||
916 | /* | |
917 | * Find and reserve possible boot-time SMP configuration: | |
918 | */ | |
919 | find_smp_config(); | |
920 | ||
921 | #ifdef CONFIG_ACPI_NUMA | |
922 | /* | |
923 | * Parse SRAT to discover nodes. | |
924 | */ | |
925 | acpi = acpi_numa_init(); | |
926 | #endif | |
927 | ||
928 | #ifdef CONFIG_K8_NUMA | |
929 | if (!acpi) | |
930 | k8 = !k8_numa_init(0, max_pfn); | |
931 | #endif | |
932 | ||
933 | initmem_init(0, max_pfn, acpi, k8); | |
934 | ||
935 | #ifdef CONFIG_X86_64 | |
936 | /* | |
937 | * dma32_reserve_bootmem() allocates bootmem which may conflict | |
938 | * with the crashkernel command line, so do that after | |
939 | * reserve_crashkernel() | |
940 | */ | |
941 | dma32_reserve_bootmem(); | |
942 | #endif | |
943 | ||
944 | reserve_ibft_region(); | |
945 | ||
946 | #ifdef CONFIG_KVM_CLOCK | |
947 | kvmclock_init(); | |
948 | #endif | |
949 | ||
950 | x86_init.paging.pagetable_setup_start(swapper_pg_dir); | |
951 | paging_init(); | |
952 | x86_init.paging.pagetable_setup_done(swapper_pg_dir); | |
953 | ||
954 | tboot_probe(); | |
955 | ||
956 | #ifdef CONFIG_X86_64 | |
957 | map_vsyscall(); | |
958 | #endif | |
959 | ||
960 | generic_apic_probe(); | |
961 | ||
962 | early_quirks(); | |
963 | ||
964 | /* | |
965 | * Read APIC and some other early information from ACPI tables. | |
966 | */ | |
967 | acpi_boot_init(); | |
968 | ||
969 | sfi_init(); | |
970 | ||
971 | /* | |
972 | * get boot-time SMP configuration: | |
973 | */ | |
974 | if (smp_found_config) | |
975 | get_smp_config(); | |
976 | ||
977 | prefill_possible_map(); | |
978 | ||
979 | #ifdef CONFIG_X86_64 | |
980 | init_cpu_to_node(); | |
981 | #endif | |
982 | ||
983 | init_apic_mappings(); | |
984 | ioapic_init_mappings(); | |
985 | ||
986 | /* need to wait for io_apic is mapped */ | |
987 | probe_nr_irqs_gsi(); | |
988 | ||
989 | kvm_guest_init(); | |
990 | ||
991 | e820_reserve_resources(); | |
992 | e820_mark_nosave_regions(max_low_pfn); | |
993 | ||
994 | x86_init.resources.reserve_resources(); | |
995 | ||
996 | e820_setup_gap(); | |
997 | ||
998 | #ifdef CONFIG_VT | |
999 | #if defined(CONFIG_VGA_CONSOLE) | |
1000 | if (!efi_enabled || (efi_mem_type(0xa0000) != EFI_CONVENTIONAL_MEMORY)) | |
1001 | conswitchp = &vga_con; | |
1002 | #elif defined(CONFIG_DUMMY_CONSOLE) | |
1003 | conswitchp = &dummy_con; | |
1004 | #endif | |
1005 | #endif | |
1006 | x86_init.oem.banner(); | |
1007 | } | |
1008 | ||
1009 | #ifdef CONFIG_X86_32 | |
1010 | ||
1011 | static struct resource video_ram_resource = { | |
1012 | .name = "Video RAM area", | |
1013 | .start = 0xa0000, | |
1014 | .end = 0xbffff, | |
1015 | .flags = IORESOURCE_BUSY | IORESOURCE_MEM | |
1016 | }; | |
1017 | ||
1018 | void __init i386_reserve_resources(void) | |
1019 | { | |
1020 | request_resource(&iomem_resource, &video_ram_resource); | |
1021 | reserve_standard_io_resources(); | |
1022 | } | |
1023 | ||
1024 | #endif /* CONFIG_X86_32 */ |