[PATCH] x86-64: Fix off by one error in IOMMU boundary checking
[GitHub/mt8127/android_kernel_alcatel_ttab.git] / arch / x86_64 / kernel / setup.c
CommitLineData
1da177e4
LT
1/*
2 * linux/arch/x86-64/kernel/setup.c
3 *
4 * Copyright (C) 1995 Linus Torvalds
5 *
6 * Nov 2001 Dave Jones <davej@suse.de>
7 * Forked from i386 setup code.
1da177e4
LT
8 */
9
10/*
11 * This file handles the architecture-dependent parts of initialization
12 */
13
14#include <linux/errno.h>
15#include <linux/sched.h>
16#include <linux/kernel.h>
17#include <linux/mm.h>
18#include <linux/stddef.h>
19#include <linux/unistd.h>
20#include <linux/ptrace.h>
21#include <linux/slab.h>
22#include <linux/user.h>
23#include <linux/a.out.h>
894673ee 24#include <linux/screen_info.h>
1da177e4
LT
25#include <linux/ioport.h>
26#include <linux/delay.h>
1da177e4
LT
27#include <linux/init.h>
28#include <linux/initrd.h>
29#include <linux/highmem.h>
30#include <linux/bootmem.h>
31#include <linux/module.h>
32#include <asm/processor.h>
33#include <linux/console.h>
34#include <linux/seq_file.h>
aac04b32 35#include <linux/crash_dump.h>
1da177e4
LT
36#include <linux/root_dev.h>
37#include <linux/pci.h>
38#include <linux/acpi.h>
39#include <linux/kallsyms.h>
40#include <linux/edd.h>
bbfceef4 41#include <linux/mmzone.h>
5f5609df 42#include <linux/kexec.h>
95235ca2 43#include <linux/cpufreq.h>
e9928674 44#include <linux/dmi.h>
17a941d8 45#include <linux/dma-mapping.h>
681558fd 46#include <linux/ctype.h>
bbfceef4 47
1da177e4
LT
48#include <asm/mtrr.h>
49#include <asm/uaccess.h>
50#include <asm/system.h>
51#include <asm/io.h>
52#include <asm/smp.h>
53#include <asm/msr.h>
54#include <asm/desc.h>
55#include <video/edid.h>
56#include <asm/e820.h>
57#include <asm/dma.h>
58#include <asm/mpspec.h>
59#include <asm/mmu_context.h>
60#include <asm/bootsetup.h>
61#include <asm/proto.h>
62#include <asm/setup.h>
63#include <asm/mach_apic.h>
64#include <asm/numa.h>
2bc0414e 65#include <asm/sections.h>
f2d3efed 66#include <asm/dmi.h>
1da177e4
LT
67
68/*
69 * Machine setup..
70 */
71
6c231b7b 72struct cpuinfo_x86 boot_cpu_data __read_mostly;
2ee60e17 73EXPORT_SYMBOL(boot_cpu_data);
1da177e4
LT
74
75unsigned long mmu_cr4_features;
76
1da177e4
LT
77/* Boot loader ID as an integer, for the benefit of proc_dointvec */
78int bootloader_type;
79
80unsigned long saved_video_mode;
81
f2d3efed
AK
82/*
83 * Early DMI memory
84 */
85int dmi_alloc_index;
86char dmi_alloc_data[DMI_MAX_DATA];
87
1da177e4
LT
88/*
89 * Setup options
90 */
1da177e4 91struct screen_info screen_info;
2ee60e17 92EXPORT_SYMBOL(screen_info);
1da177e4
LT
93struct sys_desc_table_struct {
94 unsigned short length;
95 unsigned char table[0];
96};
97
98struct edid_info edid_info;
ba70710e 99EXPORT_SYMBOL_GPL(edid_info);
1da177e4
LT
100
101extern int root_mountflags;
1da177e4 102
adf48856 103char __initdata command_line[COMMAND_LINE_SIZE];
1da177e4
LT
104
105struct resource standard_io_resources[] = {
106 { .name = "dma1", .start = 0x00, .end = 0x1f,
107 .flags = IORESOURCE_BUSY | IORESOURCE_IO },
108 { .name = "pic1", .start = 0x20, .end = 0x21,
109 .flags = IORESOURCE_BUSY | IORESOURCE_IO },
110 { .name = "timer0", .start = 0x40, .end = 0x43,
111 .flags = IORESOURCE_BUSY | IORESOURCE_IO },
112 { .name = "timer1", .start = 0x50, .end = 0x53,
113 .flags = IORESOURCE_BUSY | IORESOURCE_IO },
114 { .name = "keyboard", .start = 0x60, .end = 0x6f,
115 .flags = IORESOURCE_BUSY | IORESOURCE_IO },
116 { .name = "dma page reg", .start = 0x80, .end = 0x8f,
117 .flags = IORESOURCE_BUSY | IORESOURCE_IO },
118 { .name = "pic2", .start = 0xa0, .end = 0xa1,
119 .flags = IORESOURCE_BUSY | IORESOURCE_IO },
120 { .name = "dma2", .start = 0xc0, .end = 0xdf,
121 .flags = IORESOURCE_BUSY | IORESOURCE_IO },
122 { .name = "fpu", .start = 0xf0, .end = 0xff,
123 .flags = IORESOURCE_BUSY | IORESOURCE_IO }
124};
125
1da177e4
LT
126#define IORESOURCE_RAM (IORESOURCE_BUSY | IORESOURCE_MEM)
127
128struct resource data_resource = {
129 .name = "Kernel data",
130 .start = 0,
131 .end = 0,
132 .flags = IORESOURCE_RAM,
133};
134struct resource code_resource = {
135 .name = "Kernel code",
136 .start = 0,
137 .end = 0,
138 .flags = IORESOURCE_RAM,
139};
140
141#define IORESOURCE_ROM (IORESOURCE_BUSY | IORESOURCE_READONLY | IORESOURCE_MEM)
142
143static struct resource system_rom_resource = {
144 .name = "System ROM",
145 .start = 0xf0000,
146 .end = 0xfffff,
147 .flags = IORESOURCE_ROM,
148};
149
150static struct resource extension_rom_resource = {
151 .name = "Extension ROM",
152 .start = 0xe0000,
153 .end = 0xeffff,
154 .flags = IORESOURCE_ROM,
155};
156
157static struct resource adapter_rom_resources[] = {
158 { .name = "Adapter ROM", .start = 0xc8000, .end = 0,
159 .flags = IORESOURCE_ROM },
160 { .name = "Adapter ROM", .start = 0, .end = 0,
161 .flags = IORESOURCE_ROM },
162 { .name = "Adapter ROM", .start = 0, .end = 0,
163 .flags = IORESOURCE_ROM },
164 { .name = "Adapter ROM", .start = 0, .end = 0,
165 .flags = IORESOURCE_ROM },
166 { .name = "Adapter ROM", .start = 0, .end = 0,
167 .flags = IORESOURCE_ROM },
168 { .name = "Adapter ROM", .start = 0, .end = 0,
169 .flags = IORESOURCE_ROM }
170};
171
1da177e4
LT
172static struct resource video_rom_resource = {
173 .name = "Video ROM",
174 .start = 0xc0000,
175 .end = 0xc7fff,
176 .flags = IORESOURCE_ROM,
177};
178
179static struct resource video_ram_resource = {
180 .name = "Video RAM area",
181 .start = 0xa0000,
182 .end = 0xbffff,
183 .flags = IORESOURCE_RAM,
184};
185
186#define romsignature(x) (*(unsigned short *)(x) == 0xaa55)
187
188static int __init romchecksum(unsigned char *rom, unsigned long length)
189{
190 unsigned char *p, sum = 0;
191
192 for (p = rom; p < rom + length; p++)
193 sum += *p;
194 return sum == 0;
195}
196
197static void __init probe_roms(void)
198{
199 unsigned long start, length, upper;
200 unsigned char *rom;
201 int i;
202
203 /* video rom */
204 upper = adapter_rom_resources[0].start;
205 for (start = video_rom_resource.start; start < upper; start += 2048) {
206 rom = isa_bus_to_virt(start);
207 if (!romsignature(rom))
208 continue;
209
210 video_rom_resource.start = start;
211
212 /* 0 < length <= 0x7f * 512, historically */
213 length = rom[2] * 512;
214
215 /* if checksum okay, trust length byte */
216 if (length && romchecksum(rom, length))
217 video_rom_resource.end = start + length - 1;
218
219 request_resource(&iomem_resource, &video_rom_resource);
220 break;
221 }
222
223 start = (video_rom_resource.end + 1 + 2047) & ~2047UL;
224 if (start < upper)
225 start = upper;
226
227 /* system rom */
228 request_resource(&iomem_resource, &system_rom_resource);
229 upper = system_rom_resource.start;
230
231 /* check for extension rom (ignore length byte!) */
232 rom = isa_bus_to_virt(extension_rom_resource.start);
233 if (romsignature(rom)) {
234 length = extension_rom_resource.end - extension_rom_resource.start + 1;
235 if (romchecksum(rom, length)) {
236 request_resource(&iomem_resource, &extension_rom_resource);
237 upper = extension_rom_resource.start;
238 }
239 }
240
241 /* check for adapter roms on 2k boundaries */
9d0ef4fd
AK
242 for (i = 0; i < ARRAY_SIZE(adapter_rom_resources) && start < upper;
243 start += 2048) {
1da177e4
LT
244 rom = isa_bus_to_virt(start);
245 if (!romsignature(rom))
246 continue;
247
248 /* 0 < length <= 0x7f * 512, historically */
249 length = rom[2] * 512;
250
251 /* but accept any length that fits if checksum okay */
252 if (!length || start + length > upper || !romchecksum(rom, length))
253 continue;
254
255 adapter_rom_resources[i].start = start;
256 adapter_rom_resources[i].end = start + length - 1;
257 request_resource(&iomem_resource, &adapter_rom_resources[i]);
258
259 start = adapter_rom_resources[i++].end & ~2047UL;
260 }
261}
262
2c8c0e6b
AK
263#ifdef CONFIG_PROC_VMCORE
264/* elfcorehdr= specifies the location of elf core header
265 * stored by the crashed kernel. This option will be passed
266 * by kexec loader to the capture kernel.
267 */
268static int __init setup_elfcorehdr(char *arg)
681558fd 269{
2c8c0e6b
AK
270 char *end;
271 if (!arg)
272 return -EINVAL;
273 elfcorehdr_addr = memparse(arg, &end);
274 return end > arg ? 0 : -EINVAL;
681558fd 275}
2c8c0e6b 276early_param("elfcorehdr", setup_elfcorehdr);
e2c03888
AK
277#endif
278
2b97690f 279#ifndef CONFIG_NUMA
bbfceef4
MT
280static void __init
281contig_initmem_init(unsigned long start_pfn, unsigned long end_pfn)
1da177e4 282{
bbfceef4
MT
283 unsigned long bootmap_size, bootmap;
284
bbfceef4
MT
285 bootmap_size = bootmem_bootmap_pages(end_pfn)<<PAGE_SHIFT;
286 bootmap = find_e820_area(0, end_pfn<<PAGE_SHIFT, bootmap_size);
287 if (bootmap == -1L)
288 panic("Cannot find bootmem map of size %ld\n",bootmap_size);
289 bootmap_size = init_bootmem(bootmap >> PAGE_SHIFT, end_pfn);
5cb248ab
MG
290 e820_register_active_regions(0, start_pfn, end_pfn);
291 free_bootmem_with_active_regions(0, end_pfn);
bbfceef4 292 reserve_bootmem(bootmap, bootmap_size);
1da177e4
LT
293}
294#endif
295
1da177e4
LT
296#if defined(CONFIG_EDD) || defined(CONFIG_EDD_MODULE)
297struct edd edd;
298#ifdef CONFIG_EDD_MODULE
299EXPORT_SYMBOL(edd);
300#endif
301/**
302 * copy_edd() - Copy the BIOS EDD information
303 * from boot_params into a safe place.
304 *
305 */
306static inline void copy_edd(void)
307{
308 memcpy(edd.mbr_signature, EDD_MBR_SIGNATURE, sizeof(edd.mbr_signature));
309 memcpy(edd.edd_info, EDD_BUF, sizeof(edd.edd_info));
310 edd.mbr_signature_nr = EDD_MBR_SIG_NR;
311 edd.edd_info_nr = EDD_NR;
312}
313#else
314static inline void copy_edd(void)
315{
316}
317#endif
318
319#define EBDA_ADDR_POINTER 0x40E
ac71d12c
AK
320
321unsigned __initdata ebda_addr;
322unsigned __initdata ebda_size;
323
324static void discover_ebda(void)
1da177e4 325{
ac71d12c 326 /*
1da177e4
LT
327 * there is a real-mode segmented pointer pointing to the
328 * 4K EBDA area at 0x40E
329 */
ac71d12c
AK
330 ebda_addr = *(unsigned short *)EBDA_ADDR_POINTER;
331 ebda_addr <<= 4;
332
333 ebda_size = *(unsigned short *)(unsigned long)ebda_addr;
334
335 /* Round EBDA up to pages */
336 if (ebda_size == 0)
337 ebda_size = 1;
338 ebda_size <<= 10;
339 ebda_size = round_up(ebda_size + (ebda_addr & ~PAGE_MASK), PAGE_SIZE);
340 if (ebda_size > 64*1024)
341 ebda_size = 64*1024;
1da177e4
LT
342}
343
344void __init setup_arch(char **cmdline_p)
345{
adf48856 346 printk(KERN_INFO "Command line: %s\n", boot_command_line);
43c85c9c 347
1da177e4 348 ROOT_DEV = old_decode_dev(ORIG_ROOT_DEV);
1da177e4
LT
349 screen_info = SCREEN_INFO;
350 edid_info = EDID_INFO;
351 saved_video_mode = SAVED_VIDEO_MODE;
352 bootloader_type = LOADER_TYPE;
353
354#ifdef CONFIG_BLK_DEV_RAM
355 rd_image_start = RAMDISK_FLAGS & RAMDISK_IMAGE_START_MASK;
356 rd_prompt = ((RAMDISK_FLAGS & RAMDISK_PROMPT_FLAG) != 0);
357 rd_doload = ((RAMDISK_FLAGS & RAMDISK_LOAD_FLAG) != 0);
358#endif
359 setup_memory_region();
360 copy_edd();
361
362 if (!MOUNT_ROOT_RDONLY)
363 root_mountflags &= ~MS_RDONLY;
364 init_mm.start_code = (unsigned long) &_text;
365 init_mm.end_code = (unsigned long) &_etext;
366 init_mm.end_data = (unsigned long) &_edata;
367 init_mm.brk = (unsigned long) &_end;
368
369 code_resource.start = virt_to_phys(&_text);
370 code_resource.end = virt_to_phys(&_etext)-1;
371 data_resource.start = virt_to_phys(&_etext);
372 data_resource.end = virt_to_phys(&_edata)-1;
373
1da177e4
LT
374 early_identify_cpu(&boot_cpu_data);
375
adf48856 376 strlcpy(command_line, boot_command_line, COMMAND_LINE_SIZE);
2c8c0e6b
AK
377 *cmdline_p = command_line;
378
379 parse_early_param();
380
381 finish_e820_parsing();
9ca33eb6 382
5cb248ab 383 e820_register_active_regions(0, 0, -1UL);
1da177e4
LT
384 /*
385 * partially used pages are not usable - thus
386 * we are rounding upwards:
387 */
388 end_pfn = e820_end_of_ram();
caff0710 389 num_physpages = end_pfn;
1da177e4
LT
390
391 check_efer();
392
ac71d12c
AK
393 discover_ebda();
394
1da177e4
LT
395 init_memory_mapping(0, (end_pfn_map << PAGE_SHIFT));
396
f2d3efed
AK
397 dmi_scan_machine();
398
f6c2e333
SS
399 zap_low_mappings(0);
400
888ba6c6 401#ifdef CONFIG_ACPI
1da177e4
LT
402 /*
403 * Initialize the ACPI boot-time table parser (gets the RSDP and SDT).
404 * Call this early for SRAT node setup.
405 */
406 acpi_boot_table_init();
407#endif
408
caff0710
JB
409 /* How many end-of-memory variables you have, grandma! */
410 max_low_pfn = end_pfn;
411 max_pfn = end_pfn;
412 high_memory = (void *)__va(end_pfn * PAGE_SIZE - 1) + 1;
413
5cb248ab
MG
414 /* Remove active ranges so rediscovery with NUMA-awareness happens */
415 remove_all_active_ranges();
416
1da177e4
LT
417#ifdef CONFIG_ACPI_NUMA
418 /*
419 * Parse SRAT to discover nodes.
420 */
421 acpi_numa_init();
422#endif
423
2b97690f 424#ifdef CONFIG_NUMA
1da177e4
LT
425 numa_initmem_init(0, end_pfn);
426#else
bbfceef4 427 contig_initmem_init(0, end_pfn);
1da177e4
LT
428#endif
429
430 /* Reserve direct mapping */
431 reserve_bootmem_generic(table_start << PAGE_SHIFT,
432 (table_end - table_start) << PAGE_SHIFT);
433
434 /* reserve kernel */
ceee8822
AK
435 reserve_bootmem_generic(__pa_symbol(&_text),
436 __pa_symbol(&_end) - __pa_symbol(&_text));
1da177e4
LT
437
438 /*
439 * reserve physical page 0 - it's a special BIOS page on many boxes,
440 * enabling clean reboots, SMP operation, laptop functions.
441 */
442 reserve_bootmem_generic(0, PAGE_SIZE);
443
444 /* reserve ebda region */
ac71d12c
AK
445 if (ebda_addr)
446 reserve_bootmem_generic(ebda_addr, ebda_size);
076422d2
AS
447#ifdef CONFIG_NUMA
448 /* reserve nodemap region */
449 if (nodemap_addr)
450 reserve_bootmem_generic(nodemap_addr, nodemap_size);
451#endif
1da177e4
LT
452
453#ifdef CONFIG_SMP
454 /*
455 * But first pinch a few for the stack/trampoline stuff
456 * FIXME: Don't need the extra page at 4K, but need to fix
457 * trampoline before removing it. (see the GDT stuff)
458 */
459 reserve_bootmem_generic(PAGE_SIZE, PAGE_SIZE);
460
461 /* Reserve SMP trampoline */
462 reserve_bootmem_generic(SMP_TRAMPOLINE_BASE, PAGE_SIZE);
463#endif
464
465#ifdef CONFIG_ACPI_SLEEP
466 /*
467 * Reserve low memory region for sleep support.
468 */
469 acpi_reserve_bootmem();
470#endif
1da177e4
LT
471 /*
472 * Find and reserve possible boot-time SMP configuration:
473 */
474 find_smp_config();
1da177e4
LT
475#ifdef CONFIG_BLK_DEV_INITRD
476 if (LOADER_TYPE && INITRD_START) {
477 if (INITRD_START + INITRD_SIZE <= (end_pfn << PAGE_SHIFT)) {
478 reserve_bootmem_generic(INITRD_START, INITRD_SIZE);
19e5d9c0 479 initrd_start = INITRD_START + PAGE_OFFSET;
1da177e4
LT
480 initrd_end = initrd_start+INITRD_SIZE;
481 }
482 else {
483 printk(KERN_ERR "initrd extends beyond end of memory "
484 "(0x%08lx > 0x%08lx)\ndisabling initrd\n",
485 (unsigned long)(INITRD_START + INITRD_SIZE),
486 (unsigned long)(end_pfn << PAGE_SHIFT));
487 initrd_start = 0;
488 }
489 }
490#endif
5f5609df
EB
491#ifdef CONFIG_KEXEC
492 if (crashk_res.start != crashk_res.end) {
00212fef 493 reserve_bootmem_generic(crashk_res.start,
5f5609df
EB
494 crashk_res.end - crashk_res.start + 1);
495 }
496#endif
0d317fb7 497
1da177e4
LT
498 paging_init();
499
f157cbb1 500#ifdef CONFIG_PCI
dfa4698c 501 early_quirks();
f157cbb1 502#endif
1da177e4 503
51f62e18
AR
504 /*
505 * set this early, so we dont allocate cpu0
506 * if MADT list doesnt list BSP first
507 * mpparse.c/MP_processor_info() allocates logical cpu numbers.
508 */
509 cpu_set(0, cpu_present_map);
888ba6c6 510#ifdef CONFIG_ACPI
1da177e4
LT
511 /*
512 * Read APIC and some other early information from ACPI tables.
513 */
514 acpi_boot_init();
515#endif
516
05b3cbd8
RT
517 init_cpu_to_node();
518
1da177e4
LT
519 /*
520 * get boot-time SMP configuration:
521 */
522 if (smp_found_config)
523 get_smp_config();
524 init_apic_mappings();
1da177e4
LT
525
526 /*
527 * Request address space for all standard RAM and ROM resources
528 * and also for regions reported as reserved by the e820.
529 */
530 probe_roms();
531 e820_reserve_resources();
e8eff5ac 532 e820_mark_nosave_regions();
1da177e4
LT
533
534 request_resource(&iomem_resource, &video_ram_resource);
535
536 {
537 unsigned i;
538 /* request I/O space for devices used on all i[345]86 PCs */
9d0ef4fd 539 for (i = 0; i < ARRAY_SIZE(standard_io_resources); i++)
1da177e4
LT
540 request_resource(&ioport_resource, &standard_io_resources[i]);
541 }
542
a1e97782 543 e820_setup_gap();
1da177e4 544
1da177e4
LT
545#ifdef CONFIG_VT
546#if defined(CONFIG_VGA_CONSOLE)
547 conswitchp = &vga_con;
548#elif defined(CONFIG_DUMMY_CONSOLE)
549 conswitchp = &dummy_con;
550#endif
551#endif
552}
553
e6982c67 554static int __cpuinit get_model_name(struct cpuinfo_x86 *c)
1da177e4
LT
555{
556 unsigned int *v;
557
ebfcaa96 558 if (c->extended_cpuid_level < 0x80000004)
1da177e4
LT
559 return 0;
560
561 v = (unsigned int *) c->x86_model_id;
562 cpuid(0x80000002, &v[0], &v[1], &v[2], &v[3]);
563 cpuid(0x80000003, &v[4], &v[5], &v[6], &v[7]);
564 cpuid(0x80000004, &v[8], &v[9], &v[10], &v[11]);
565 c->x86_model_id[48] = 0;
566 return 1;
567}
568
569
e6982c67 570static void __cpuinit display_cacheinfo(struct cpuinfo_x86 *c)
1da177e4
LT
571{
572 unsigned int n, dummy, eax, ebx, ecx, edx;
573
ebfcaa96 574 n = c->extended_cpuid_level;
1da177e4
LT
575
576 if (n >= 0x80000005) {
577 cpuid(0x80000005, &dummy, &ebx, &ecx, &edx);
578 printk(KERN_INFO "CPU: L1 I Cache: %dK (%d bytes/line), D cache %dK (%d bytes/line)\n",
579 edx>>24, edx&0xFF, ecx>>24, ecx&0xFF);
580 c->x86_cache_size=(ecx>>24)+(edx>>24);
581 /* On K8 L1 TLB is inclusive, so don't count it */
582 c->x86_tlbsize = 0;
583 }
584
585 if (n >= 0x80000006) {
586 cpuid(0x80000006, &dummy, &ebx, &ecx, &edx);
587 ecx = cpuid_ecx(0x80000006);
588 c->x86_cache_size = ecx >> 16;
589 c->x86_tlbsize += ((ebx >> 16) & 0xfff) + (ebx & 0xfff);
590
591 printk(KERN_INFO "CPU: L2 Cache: %dK (%d bytes/line)\n",
592 c->x86_cache_size, ecx & 0xFF);
593 }
594
595 if (n >= 0x80000007)
596 cpuid(0x80000007, &dummy, &dummy, &dummy, &c->x86_power);
597 if (n >= 0x80000008) {
598 cpuid(0x80000008, &eax, &dummy, &dummy, &dummy);
599 c->x86_virt_bits = (eax >> 8) & 0xff;
600 c->x86_phys_bits = eax & 0xff;
601 }
602}
603
3f098c26
AK
604#ifdef CONFIG_NUMA
605static int nearby_node(int apicid)
606{
607 int i;
608 for (i = apicid - 1; i >= 0; i--) {
609 int node = apicid_to_node[i];
610 if (node != NUMA_NO_NODE && node_online(node))
611 return node;
612 }
613 for (i = apicid + 1; i < MAX_LOCAL_APIC; i++) {
614 int node = apicid_to_node[i];
615 if (node != NUMA_NO_NODE && node_online(node))
616 return node;
617 }
618 return first_node(node_online_map); /* Shouldn't happen */
619}
620#endif
621
63518644
AK
622/*
623 * On a AMD dual core setup the lower bits of the APIC id distingush the cores.
624 * Assumes number of cores is a power of two.
625 */
626static void __init amd_detect_cmp(struct cpuinfo_x86 *c)
627{
628#ifdef CONFIG_SMP
b41e2939 629 unsigned bits;
3f098c26 630#ifdef CONFIG_NUMA
f3fa8ebc 631 int cpu = smp_processor_id();
3f098c26 632 int node = 0;
60c1bc82 633 unsigned apicid = hard_smp_processor_id();
3f098c26 634#endif
faee9a5d 635 unsigned ecx = cpuid_ecx(0x80000008);
b41e2939 636
faee9a5d 637 c->x86_max_cores = (ecx & 0xff) + 1;
b41e2939 638
faee9a5d
AK
639 /* CPU telling us the core id bits shift? */
640 bits = (ecx >> 12) & 0xF;
641
642 /* Otherwise recompute */
643 if (bits == 0) {
644 while ((1 << bits) < c->x86_max_cores)
645 bits++;
646 }
b41e2939
AK
647
648 /* Low order bits define the core id (index of core in socket) */
f3fa8ebc 649 c->cpu_core_id = c->phys_proc_id & ((1 << bits)-1);
b41e2939 650 /* Convert the APIC ID into the socket ID */
f3fa8ebc 651 c->phys_proc_id = phys_pkg_id(bits);
63518644
AK
652
653#ifdef CONFIG_NUMA
f3fa8ebc 654 node = c->phys_proc_id;
3f098c26
AK
655 if (apicid_to_node[apicid] != NUMA_NO_NODE)
656 node = apicid_to_node[apicid];
657 if (!node_online(node)) {
658 /* Two possibilities here:
659 - The CPU is missing memory and no node was created.
660 In that case try picking one from a nearby CPU
661 - The APIC IDs differ from the HyperTransport node IDs
662 which the K8 northbridge parsing fills in.
663 Assume they are all increased by a constant offset,
664 but in the same order as the HT nodeids.
665 If that doesn't result in a usable node fall back to the
666 path for the previous case. */
f3fa8ebc 667 int ht_nodeid = apicid - (cpu_data[0].phys_proc_id << bits);
3f098c26
AK
668 if (ht_nodeid >= 0 &&
669 apicid_to_node[ht_nodeid] != NUMA_NO_NODE)
670 node = apicid_to_node[ht_nodeid];
671 /* Pick a nearby node */
672 if (!node_online(node))
673 node = nearby_node(apicid);
674 }
69d81fcd 675 numa_set_node(cpu, node);
3f098c26 676
e42f9437 677 printk(KERN_INFO "CPU %d/%x -> Node %d\n", cpu, apicid, node);
63518644 678#endif
63518644
AK
679#endif
680}
1da177e4 681
ed77504b 682static void __cpuinit init_amd(struct cpuinfo_x86 *c)
1da177e4 683{
7bcd3f34 684 unsigned level;
1da177e4 685
bc5e8fdf
LT
686#ifdef CONFIG_SMP
687 unsigned long value;
688
7d318d77
AK
689 /*
690 * Disable TLB flush filter by setting HWCR.FFDIS on K8
691 * bit 6 of msr C001_0015
692 *
693 * Errata 63 for SH-B3 steppings
694 * Errata 122 for all steppings (F+ have it disabled by default)
695 */
696 if (c->x86 == 15) {
697 rdmsrl(MSR_K8_HWCR, value);
698 value |= 1 << 6;
699 wrmsrl(MSR_K8_HWCR, value);
700 }
bc5e8fdf
LT
701#endif
702
1da177e4
LT
703 /* Bit 31 in normal CPUID used for nonstandard 3DNow ID;
704 3DNow is IDd by bit 31 in extended CPUID (1*32+31) anyway */
705 clear_bit(0*32+31, &c->x86_capability);
706
7bcd3f34
AK
707 /* On C+ stepping K8 rep microcode works well for copy/memset */
708 level = cpuid_eax(1);
709 if (c->x86 == 15 && ((level >= 0x0f48 && level < 0x0f50) || level >= 0x0f58))
710 set_bit(X86_FEATURE_REP_GOOD, &c->x86_capability);
711
18bd057b
AK
712 /* Enable workaround for FXSAVE leak */
713 if (c->x86 >= 6)
714 set_bit(X86_FEATURE_FXSAVE_LEAK, &c->x86_capability);
715
e42f9437
RS
716 level = get_model_name(c);
717 if (!level) {
1da177e4
LT
718 switch (c->x86) {
719 case 15:
720 /* Should distinguish Models here, but this is only
721 a fallback anyways. */
722 strcpy(c->x86_model_id, "Hammer");
723 break;
724 }
725 }
726 display_cacheinfo(c);
727
130951cc
AK
728 /* c->x86_power is 8000_0007 edx. Bit 8 is constant TSC */
729 if (c->x86_power & (1<<8))
730 set_bit(X86_FEATURE_CONSTANT_TSC, &c->x86_capability);
731
faee9a5d
AK
732 /* Multi core CPU? */
733 if (c->extended_cpuid_level >= 0x80000008)
63518644 734 amd_detect_cmp(c);
1da177e4 735
240cd6a8
AK
736 /* Fix cpuid4 emulation for more */
737 num_cache_leaves = 3;
2049336f 738
61677965
AK
739 /* RDTSC can be speculated around */
740 clear_bit(X86_FEATURE_SYNC_RDTSC, &c->x86_capability);
1da177e4
LT
741}
742
e6982c67 743static void __cpuinit detect_ht(struct cpuinfo_x86 *c)
1da177e4
LT
744{
745#ifdef CONFIG_SMP
746 u32 eax, ebx, ecx, edx;
94605eff 747 int index_msb, core_bits;
94605eff
SS
748
749 cpuid(1, &eax, &ebx, &ecx, &edx);
750
94605eff 751
e42f9437 752 if (!cpu_has(c, X86_FEATURE_HT))
1da177e4 753 return;
e42f9437
RS
754 if (cpu_has(c, X86_FEATURE_CMP_LEGACY))
755 goto out;
1da177e4 756
1da177e4 757 smp_num_siblings = (ebx & 0xff0000) >> 16;
94605eff 758
1da177e4
LT
759 if (smp_num_siblings == 1) {
760 printk(KERN_INFO "CPU: Hyper-Threading is disabled\n");
94605eff
SS
761 } else if (smp_num_siblings > 1 ) {
762
1da177e4
LT
763 if (smp_num_siblings > NR_CPUS) {
764 printk(KERN_WARNING "CPU: Unsupported number of the siblings %d", smp_num_siblings);
765 smp_num_siblings = 1;
766 return;
767 }
94605eff
SS
768
769 index_msb = get_count_order(smp_num_siblings);
f3fa8ebc 770 c->phys_proc_id = phys_pkg_id(index_msb);
3dd9d514 771
94605eff 772 smp_num_siblings = smp_num_siblings / c->x86_max_cores;
3dd9d514 773
94605eff
SS
774 index_msb = get_count_order(smp_num_siblings) ;
775
776 core_bits = get_count_order(c->x86_max_cores);
3dd9d514 777
f3fa8ebc 778 c->cpu_core_id = phys_pkg_id(index_msb) &
94605eff 779 ((1 << core_bits) - 1);
1da177e4 780 }
e42f9437
RS
781out:
782 if ((c->x86_max_cores * smp_num_siblings) > 1) {
783 printk(KERN_INFO "CPU: Physical Processor ID: %d\n", c->phys_proc_id);
784 printk(KERN_INFO "CPU: Processor Core ID: %d\n", c->cpu_core_id);
785 }
786
1da177e4
LT
787#endif
788}
789
3dd9d514
AK
790/*
791 * find out the number of processor cores on the die
792 */
e6982c67 793static int __cpuinit intel_num_cpu_cores(struct cpuinfo_x86 *c)
3dd9d514 794{
2bbc419f 795 unsigned int eax, t;
3dd9d514
AK
796
797 if (c->cpuid_level < 4)
798 return 1;
799
2bbc419f 800 cpuid_count(4, 0, &eax, &t, &t, &t);
3dd9d514
AK
801
802 if (eax & 0x1f)
803 return ((eax >> 26) + 1);
804 else
805 return 1;
806}
807
df0cc26b
AK
808static void srat_detect_node(void)
809{
810#ifdef CONFIG_NUMA
ddea7be0 811 unsigned node;
df0cc26b 812 int cpu = smp_processor_id();
e42f9437 813 int apicid = hard_smp_processor_id();
df0cc26b
AK
814
815 /* Don't do the funky fallback heuristics the AMD version employs
816 for now. */
e42f9437 817 node = apicid_to_node[apicid];
df0cc26b 818 if (node == NUMA_NO_NODE)
0d015324 819 node = first_node(node_online_map);
69d81fcd 820 numa_set_node(cpu, node);
df0cc26b 821
c31fbb1a 822 printk(KERN_INFO "CPU %d/%x -> Node %d\n", cpu, apicid, node);
df0cc26b
AK
823#endif
824}
825
e6982c67 826static void __cpuinit init_intel(struct cpuinfo_x86 *c)
1da177e4
LT
827{
828 /* Cache sizes */
829 unsigned n;
830
831 init_intel_cacheinfo(c);
0080e667
VP
832 if (c->cpuid_level > 9 ) {
833 unsigned eax = cpuid_eax(10);
834 /* Check for version and the number of counters */
835 if ((eax & 0xff) && (((eax>>8) & 0xff) > 1))
836 set_bit(X86_FEATURE_ARCH_PERFMON, &c->x86_capability);
837 }
838
36b2a8d5
SE
839 if (cpu_has_ds) {
840 unsigned int l1, l2;
841 rdmsr(MSR_IA32_MISC_ENABLE, l1, l2);
ee58fad5
SE
842 if (!(l1 & (1<<11)))
843 set_bit(X86_FEATURE_BTS, c->x86_capability);
36b2a8d5
SE
844 if (!(l1 & (1<<12)))
845 set_bit(X86_FEATURE_PEBS, c->x86_capability);
846 }
847
ebfcaa96 848 n = c->extended_cpuid_level;
1da177e4
LT
849 if (n >= 0x80000008) {
850 unsigned eax = cpuid_eax(0x80000008);
851 c->x86_virt_bits = (eax >> 8) & 0xff;
852 c->x86_phys_bits = eax & 0xff;
af9c142d
SL
853 /* CPUID workaround for Intel 0F34 CPU */
854 if (c->x86_vendor == X86_VENDOR_INTEL &&
855 c->x86 == 0xF && c->x86_model == 0x3 &&
856 c->x86_mask == 0x4)
857 c->x86_phys_bits = 36;
1da177e4
LT
858 }
859
860 if (c->x86 == 15)
861 c->x86_cache_alignment = c->x86_clflush_size * 2;
39b3a791
AK
862 if ((c->x86 == 0xf && c->x86_model >= 0x03) ||
863 (c->x86 == 0x6 && c->x86_model >= 0x0e))
c29601e9 864 set_bit(X86_FEATURE_CONSTANT_TSC, &c->x86_capability);
27fbe5b2
AK
865 if (c->x86 == 6)
866 set_bit(X86_FEATURE_REP_GOOD, &c->x86_capability);
f3d73707
AV
867 if (c->x86 == 15)
868 set_bit(X86_FEATURE_SYNC_RDTSC, &c->x86_capability);
869 else
870 clear_bit(X86_FEATURE_SYNC_RDTSC, &c->x86_capability);
94605eff 871 c->x86_max_cores = intel_num_cpu_cores(c);
df0cc26b
AK
872
873 srat_detect_node();
1da177e4
LT
874}
875
672289e9 876static void __cpuinit get_cpu_vendor(struct cpuinfo_x86 *c)
1da177e4
LT
877{
878 char *v = c->x86_vendor_id;
879
880 if (!strcmp(v, "AuthenticAMD"))
881 c->x86_vendor = X86_VENDOR_AMD;
882 else if (!strcmp(v, "GenuineIntel"))
883 c->x86_vendor = X86_VENDOR_INTEL;
884 else
885 c->x86_vendor = X86_VENDOR_UNKNOWN;
886}
887
888struct cpu_model_info {
889 int vendor;
890 int family;
891 char *model_names[16];
892};
893
894/* Do some early cpuid on the boot CPU to get some parameter that are
895 needed before check_bugs. Everything advanced is in identify_cpu
896 below. */
e6982c67 897void __cpuinit early_identify_cpu(struct cpuinfo_x86 *c)
1da177e4
LT
898{
899 u32 tfms;
900
901 c->loops_per_jiffy = loops_per_jiffy;
902 c->x86_cache_size = -1;
903 c->x86_vendor = X86_VENDOR_UNKNOWN;
904 c->x86_model = c->x86_mask = 0; /* So far unknown... */
905 c->x86_vendor_id[0] = '\0'; /* Unset */
906 c->x86_model_id[0] = '\0'; /* Unset */
907 c->x86_clflush_size = 64;
908 c->x86_cache_alignment = c->x86_clflush_size;
94605eff 909 c->x86_max_cores = 1;
ebfcaa96 910 c->extended_cpuid_level = 0;
1da177e4
LT
911 memset(&c->x86_capability, 0, sizeof c->x86_capability);
912
913 /* Get vendor name */
914 cpuid(0x00000000, (unsigned int *)&c->cpuid_level,
915 (unsigned int *)&c->x86_vendor_id[0],
916 (unsigned int *)&c->x86_vendor_id[8],
917 (unsigned int *)&c->x86_vendor_id[4]);
918
919 get_cpu_vendor(c);
920
921 /* Initialize the standard set of capabilities */
922 /* Note that the vendor-specific code below might override */
923
924 /* Intel-defined flags: level 0x00000001 */
925 if (c->cpuid_level >= 0x00000001) {
926 __u32 misc;
927 cpuid(0x00000001, &tfms, &misc, &c->x86_capability[4],
928 &c->x86_capability[0]);
929 c->x86 = (tfms >> 8) & 0xf;
930 c->x86_model = (tfms >> 4) & 0xf;
931 c->x86_mask = tfms & 0xf;
f5f786d0 932 if (c->x86 == 0xf)
1da177e4 933 c->x86 += (tfms >> 20) & 0xff;
f5f786d0 934 if (c->x86 >= 0x6)
1da177e4 935 c->x86_model += ((tfms >> 16) & 0xF) << 4;
1da177e4
LT
936 if (c->x86_capability[0] & (1<<19))
937 c->x86_clflush_size = ((misc >> 8) & 0xff) * 8;
1da177e4
LT
938 } else {
939 /* Have CPUID level 0 only - unheard of */
940 c->x86 = 4;
941 }
a158608b
AK
942
943#ifdef CONFIG_SMP
f3fa8ebc 944 c->phys_proc_id = (cpuid_ebx(1) >> 24) & 0xff;
a158608b 945#endif
1da177e4
LT
946}
947
948/*
949 * This does the hard work of actually picking apart the CPU stuff...
950 */
e6982c67 951void __cpuinit identify_cpu(struct cpuinfo_x86 *c)
1da177e4
LT
952{
953 int i;
954 u32 xlvl;
955
956 early_identify_cpu(c);
957
958 /* AMD-defined flags: level 0x80000001 */
959 xlvl = cpuid_eax(0x80000000);
ebfcaa96 960 c->extended_cpuid_level = xlvl;
1da177e4
LT
961 if ((xlvl & 0xffff0000) == 0x80000000) {
962 if (xlvl >= 0x80000001) {
963 c->x86_capability[1] = cpuid_edx(0x80000001);
5b7abc6f 964 c->x86_capability[6] = cpuid_ecx(0x80000001);
1da177e4
LT
965 }
966 if (xlvl >= 0x80000004)
967 get_model_name(c); /* Default name */
968 }
969
970 /* Transmeta-defined flags: level 0x80860001 */
971 xlvl = cpuid_eax(0x80860000);
972 if ((xlvl & 0xffff0000) == 0x80860000) {
973 /* Don't set x86_cpuid_level here for now to not confuse. */
974 if (xlvl >= 0x80860001)
975 c->x86_capability[2] = cpuid_edx(0x80860001);
976 }
977
1e9f28fa
SS
978 c->apicid = phys_pkg_id(0);
979
1da177e4
LT
980 /*
981 * Vendor-specific initialization. In this section we
982 * canonicalize the feature flags, meaning if there are
983 * features a certain CPU supports which CPUID doesn't
984 * tell us, CPUID claiming incorrect flags, or other bugs,
985 * we handle them here.
986 *
987 * At the end of this section, c->x86_capability better
988 * indicate the features this CPU genuinely supports!
989 */
990 switch (c->x86_vendor) {
991 case X86_VENDOR_AMD:
992 init_amd(c);
993 break;
994
995 case X86_VENDOR_INTEL:
996 init_intel(c);
997 break;
998
999 case X86_VENDOR_UNKNOWN:
1000 default:
1001 display_cacheinfo(c);
1002 break;
1003 }
1004
1005 select_idle_routine(c);
1006 detect_ht(c);
1da177e4
LT
1007
1008 /*
1009 * On SMP, boot_cpu_data holds the common feature set between
1010 * all CPUs; so make sure that we indicate which features are
1011 * common between the CPUs. The first time this routine gets
1012 * executed, c == &boot_cpu_data.
1013 */
1014 if (c != &boot_cpu_data) {
1015 /* AND the already accumulated flags with these */
1016 for (i = 0 ; i < NCAPINTS ; i++)
1017 boot_cpu_data.x86_capability[i] &= c->x86_capability[i];
1018 }
1019
1020#ifdef CONFIG_X86_MCE
1021 mcheck_init(c);
1022#endif
3b520b23
SL
1023 if (c == &boot_cpu_data)
1024 mtrr_bp_init();
1025 else
1026 mtrr_ap_init();
1da177e4 1027#ifdef CONFIG_NUMA
3019e8eb 1028 numa_add_cpu(smp_processor_id());
1da177e4
LT
1029#endif
1030}
1031
1032
e6982c67 1033void __cpuinit print_cpu_info(struct cpuinfo_x86 *c)
1da177e4
LT
1034{
1035 if (c->x86_model_id[0])
1036 printk("%s", c->x86_model_id);
1037
1038 if (c->x86_mask || c->cpuid_level >= 0)
1039 printk(" stepping %02x\n", c->x86_mask);
1040 else
1041 printk("\n");
1042}
1043
1044/*
1045 * Get CPU information for use by the procfs.
1046 */
1047
1048static int show_cpuinfo(struct seq_file *m, void *v)
1049{
1050 struct cpuinfo_x86 *c = v;
1051
1052 /*
1053 * These flag bits must match the definitions in <asm/cpufeature.h>.
1054 * NULL means this bit is undefined or reserved; either way it doesn't
1055 * have meaning as far as Linux is concerned. Note that it's important
1056 * to realize there is a difference between this table and CPUID -- if
1057 * applications want to get the raw CPUID data, they should access
1058 * /dev/cpu/<cpu_nr>/cpuid instead.
1059 */
1060 static char *x86_cap_flags[] = {
1061 /* Intel-defined */
1062 "fpu", "vme", "de", "pse", "tsc", "msr", "pae", "mce",
1063 "cx8", "apic", NULL, "sep", "mtrr", "pge", "mca", "cmov",
1064 "pat", "pse36", "pn", "clflush", NULL, "dts", "acpi", "mmx",
1065 "fxsr", "sse", "sse2", "ss", "ht", "tm", "ia64", NULL,
1066
1067 /* AMD-defined */
3c3b73b6 1068 NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL,
1da177e4
LT
1069 NULL, NULL, NULL, "syscall", NULL, NULL, NULL, NULL,
1070 NULL, NULL, NULL, NULL, "nx", NULL, "mmxext", NULL,
7b0e8501 1071 NULL, "fxsr_opt", NULL, "rdtscp", NULL, "lm", "3dnowext", "3dnow",
1da177e4
LT
1072
1073 /* Transmeta-defined */
1074 "recovery", "longrun", NULL, "lrti", NULL, NULL, NULL, NULL,
1075 NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL,
1076 NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL,
1077 NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL,
1078
1079 /* Other (Linux-defined) */
622dcaf9 1080 "cxmmx", NULL, "cyrix_arr", "centaur_mcr", NULL,
c29601e9 1081 "constant_tsc", NULL, NULL,
d167a518 1082 "up", NULL, NULL, NULL, NULL, NULL, NULL, NULL,
1da177e4
LT
1083 NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL,
1084 NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL,
1085
1086 /* Intel-defined (#2) */
9d95dd84 1087 "pni", NULL, NULL, "monitor", "ds_cpl", "vmx", "smx", "est",
dcf10307
DJ
1088 "tm2", "ssse3", "cid", NULL, NULL, "cx16", "xtpr", NULL,
1089 NULL, NULL, "dca", NULL, NULL, NULL, NULL, NULL,
1da177e4
LT
1090 NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL,
1091
5b7abc6f
PA
1092 /* VIA/Cyrix/Centaur-defined */
1093 NULL, NULL, "rng", "rng_en", NULL, NULL, "ace", "ace_en",
1094 NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL,
1095 NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL,
1096 NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL,
1097
1da177e4 1098 /* AMD-defined (#2) */
3f98bc49 1099 "lahf_lm", "cmp_legacy", "svm", NULL, "cr8_legacy", NULL, NULL, NULL,
1da177e4
LT
1100 NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL,
1101 NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL,
5b7abc6f 1102 NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL,
1da177e4
LT
1103 };
1104 static char *x86_power_flags[] = {
1105 "ts", /* temperature sensor */
1106 "fid", /* frequency id control */
1107 "vid", /* voltage id control */
1108 "ttp", /* thermal trip */
1109 "tm",
3f98bc49
AK
1110 "stc",
1111 NULL,
39b3a791 1112 /* nothing */ /* constant_tsc - moved to flags */
1da177e4
LT
1113 };
1114
1115
1116#ifdef CONFIG_SMP
1117 if (!cpu_online(c-cpu_data))
1118 return 0;
1119#endif
1120
1121 seq_printf(m,"processor\t: %u\n"
1122 "vendor_id\t: %s\n"
1123 "cpu family\t: %d\n"
1124 "model\t\t: %d\n"
1125 "model name\t: %s\n",
1126 (unsigned)(c-cpu_data),
1127 c->x86_vendor_id[0] ? c->x86_vendor_id : "unknown",
1128 c->x86,
1129 (int)c->x86_model,
1130 c->x86_model_id[0] ? c->x86_model_id : "unknown");
1131
1132 if (c->x86_mask || c->cpuid_level >= 0)
1133 seq_printf(m, "stepping\t: %d\n", c->x86_mask);
1134 else
1135 seq_printf(m, "stepping\t: unknown\n");
1136
1137 if (cpu_has(c,X86_FEATURE_TSC)) {
95235ca2
VP
1138 unsigned int freq = cpufreq_quick_get((unsigned)(c-cpu_data));
1139 if (!freq)
1140 freq = cpu_khz;
1da177e4 1141 seq_printf(m, "cpu MHz\t\t: %u.%03u\n",
95235ca2 1142 freq / 1000, (freq % 1000));
1da177e4
LT
1143 }
1144
1145 /* Cache size */
1146 if (c->x86_cache_size >= 0)
1147 seq_printf(m, "cache size\t: %d KB\n", c->x86_cache_size);
1148
1149#ifdef CONFIG_SMP
94605eff 1150 if (smp_num_siblings * c->x86_max_cores > 1) {
db468681 1151 int cpu = c - cpu_data;
f3fa8ebc 1152 seq_printf(m, "physical id\t: %d\n", c->phys_proc_id);
94605eff 1153 seq_printf(m, "siblings\t: %d\n", cpus_weight(cpu_core_map[cpu]));
f3fa8ebc 1154 seq_printf(m, "core id\t\t: %d\n", c->cpu_core_id);
94605eff 1155 seq_printf(m, "cpu cores\t: %d\n", c->booted_cores);
db468681 1156 }
1da177e4
LT
1157#endif
1158
1159 seq_printf(m,
1160 "fpu\t\t: yes\n"
1161 "fpu_exception\t: yes\n"
1162 "cpuid level\t: %d\n"
1163 "wp\t\t: yes\n"
1164 "flags\t\t:",
1165 c->cpuid_level);
1166
1167 {
1168 int i;
1169 for ( i = 0 ; i < 32*NCAPINTS ; i++ )
3d1712c9 1170 if (cpu_has(c, i) && x86_cap_flags[i] != NULL)
1da177e4
LT
1171 seq_printf(m, " %s", x86_cap_flags[i]);
1172 }
1173
1174 seq_printf(m, "\nbogomips\t: %lu.%02lu\n",
1175 c->loops_per_jiffy/(500000/HZ),
1176 (c->loops_per_jiffy/(5000/HZ)) % 100);
1177
1178 if (c->x86_tlbsize > 0)
1179 seq_printf(m, "TLB size\t: %d 4K pages\n", c->x86_tlbsize);
1180 seq_printf(m, "clflush size\t: %d\n", c->x86_clflush_size);
1181 seq_printf(m, "cache_alignment\t: %d\n", c->x86_cache_alignment);
1182
1183 seq_printf(m, "address sizes\t: %u bits physical, %u bits virtual\n",
1184 c->x86_phys_bits, c->x86_virt_bits);
1185
1186 seq_printf(m, "power management:");
1187 {
1188 unsigned i;
1189 for (i = 0; i < 32; i++)
1190 if (c->x86_power & (1 << i)) {
3f98bc49
AK
1191 if (i < ARRAY_SIZE(x86_power_flags) &&
1192 x86_power_flags[i])
1193 seq_printf(m, "%s%s",
1194 x86_power_flags[i][0]?" ":"",
1195 x86_power_flags[i]);
1da177e4
LT
1196 else
1197 seq_printf(m, " [%d]", i);
1198 }
1199 }
1da177e4 1200
d31ddaa1 1201 seq_printf(m, "\n\n");
1da177e4
LT
1202
1203 return 0;
1204}
1205
1206static void *c_start(struct seq_file *m, loff_t *pos)
1207{
1208 return *pos < NR_CPUS ? cpu_data + *pos : NULL;
1209}
1210
1211static void *c_next(struct seq_file *m, void *v, loff_t *pos)
1212{
1213 ++*pos;
1214 return c_start(m, pos);
1215}
1216
1217static void c_stop(struct seq_file *m, void *v)
1218{
1219}
1220
1221struct seq_operations cpuinfo_op = {
1222 .start =c_start,
1223 .next = c_next,
1224 .stop = c_stop,
1225 .show = show_cpuinfo,
1226};
e9928674 1227
9c63f873 1228#if defined(CONFIG_INPUT_PCSPKR) || defined(CONFIG_INPUT_PCSPKR_MODULE)
160bd18e
MP
1229#include <linux/platform_device.h>
1230static __init int add_pcspkr(void)
1231{
1232 struct platform_device *pd;
1233 int ret;
1234
1235 pd = platform_device_alloc("pcspkr", -1);
1236 if (!pd)
1237 return -ENOMEM;
1238
1239 ret = platform_device_add(pd);
1240 if (ret)
1241 platform_device_put(pd);
1242
1243 return ret;
1244}
1245device_initcall(add_pcspkr);
1246#endif