]> git.ipfire.org Git - thirdparty/kernel/stable.git/blob - drivers/acpi/acpi_processor.c
KVM: x86/mmu: Remove unnecessary ‘NULL’ values from sptep
[thirdparty/kernel/stable.git] / drivers / acpi / acpi_processor.c
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3 * acpi_processor.c - ACPI processor enumeration support
4 *
5 * Copyright (C) 2001, 2002 Andy Grover <andrew.grover@intel.com>
6 * Copyright (C) 2001, 2002 Paul Diefenbaugh <paul.s.diefenbaugh@intel.com>
7 * Copyright (C) 2004 Dominik Brodowski <linux@brodo.de>
8 * Copyright (C) 2004 Anil S Keshavamurthy <anil.s.keshavamurthy@intel.com>
9 * Copyright (C) 2013, Intel Corporation
10 * Rafael J. Wysocki <rafael.j.wysocki@intel.com>
11 */
12 #define pr_fmt(fmt) "ACPI: " fmt
13
14 #include <linux/acpi.h>
15 #include <linux/device.h>
16 #include <linux/dmi.h>
17 #include <linux/kernel.h>
18 #include <linux/module.h>
19 #include <linux/pci.h>
20 #include <linux/platform_device.h>
21
22 #include <acpi/processor.h>
23
24 #include <asm/cpu.h>
25
26 #include <xen/xen.h>
27
28 #include "internal.h"
29
30 DEFINE_PER_CPU(struct acpi_processor *, processors);
31 EXPORT_PER_CPU_SYMBOL(processors);
32
33 /* Errata Handling */
34 struct acpi_processor_errata errata __read_mostly;
35 EXPORT_SYMBOL_GPL(errata);
36
37 static int acpi_processor_errata_piix4(struct pci_dev *dev)
38 {
39 u8 value1 = 0;
40 u8 value2 = 0;
41
42
43 if (!dev)
44 return -EINVAL;
45
46 /*
47 * Note that 'dev' references the PIIX4 ACPI Controller.
48 */
49
50 switch (dev->revision) {
51 case 0:
52 dev_dbg(&dev->dev, "Found PIIX4 A-step\n");
53 break;
54 case 1:
55 dev_dbg(&dev->dev, "Found PIIX4 B-step\n");
56 break;
57 case 2:
58 dev_dbg(&dev->dev, "Found PIIX4E\n");
59 break;
60 case 3:
61 dev_dbg(&dev->dev, "Found PIIX4M\n");
62 break;
63 default:
64 dev_dbg(&dev->dev, "Found unknown PIIX4\n");
65 break;
66 }
67
68 switch (dev->revision) {
69
70 case 0: /* PIIX4 A-step */
71 case 1: /* PIIX4 B-step */
72 /*
73 * See specification changes #13 ("Manual Throttle Duty Cycle")
74 * and #14 ("Enabling and Disabling Manual Throttle"), plus
75 * erratum #5 ("STPCLK# Deassertion Time") from the January
76 * 2002 PIIX4 specification update. Applies to only older
77 * PIIX4 models.
78 */
79 errata.piix4.throttle = 1;
80 fallthrough;
81
82 case 2: /* PIIX4E */
83 case 3: /* PIIX4M */
84 /*
85 * See erratum #18 ("C3 Power State/BMIDE and Type-F DMA
86 * Livelock") from the January 2002 PIIX4 specification update.
87 * Applies to all PIIX4 models.
88 */
89
90 /*
91 * BM-IDE
92 * ------
93 * Find the PIIX4 IDE Controller and get the Bus Master IDE
94 * Status register address. We'll use this later to read
95 * each IDE controller's DMA status to make sure we catch all
96 * DMA activity.
97 */
98 dev = pci_get_subsys(PCI_VENDOR_ID_INTEL,
99 PCI_DEVICE_ID_INTEL_82371AB,
100 PCI_ANY_ID, PCI_ANY_ID, NULL);
101 if (dev) {
102 errata.piix4.bmisx = pci_resource_start(dev, 4);
103 pci_dev_put(dev);
104 }
105
106 /*
107 * Type-F DMA
108 * ----------
109 * Find the PIIX4 ISA Controller and read the Motherboard
110 * DMA controller's status to see if Type-F (Fast) DMA mode
111 * is enabled (bit 7) on either channel. Note that we'll
112 * disable C3 support if this is enabled, as some legacy
113 * devices won't operate well if fast DMA is disabled.
114 */
115 dev = pci_get_subsys(PCI_VENDOR_ID_INTEL,
116 PCI_DEVICE_ID_INTEL_82371AB_0,
117 PCI_ANY_ID, PCI_ANY_ID, NULL);
118 if (dev) {
119 pci_read_config_byte(dev, 0x76, &value1);
120 pci_read_config_byte(dev, 0x77, &value2);
121 if ((value1 & 0x80) || (value2 & 0x80))
122 errata.piix4.fdma = 1;
123 pci_dev_put(dev);
124 }
125
126 break;
127 }
128
129 if (errata.piix4.bmisx)
130 dev_dbg(&dev->dev, "Bus master activity detection (BM-IDE) erratum enabled\n");
131 if (errata.piix4.fdma)
132 dev_dbg(&dev->dev, "Type-F DMA livelock erratum (C3 disabled)\n");
133
134 return 0;
135 }
136
137 static int acpi_processor_errata(void)
138 {
139 int result = 0;
140 struct pci_dev *dev = NULL;
141
142 /*
143 * PIIX4
144 */
145 dev = pci_get_subsys(PCI_VENDOR_ID_INTEL,
146 PCI_DEVICE_ID_INTEL_82371AB_3, PCI_ANY_ID,
147 PCI_ANY_ID, NULL);
148 if (dev) {
149 result = acpi_processor_errata_piix4(dev);
150 pci_dev_put(dev);
151 }
152
153 return result;
154 }
155
156 /* Create a platform device to represent a CPU frequency control mechanism. */
157 static void cpufreq_add_device(const char *name)
158 {
159 struct platform_device *pdev;
160
161 pdev = platform_device_register_simple(name, PLATFORM_DEVID_NONE, NULL, 0);
162 if (IS_ERR(pdev))
163 pr_info("%s device creation failed: %ld\n", name, PTR_ERR(pdev));
164 }
165
166 #ifdef CONFIG_X86
167 /* Check presence of Processor Clocking Control by searching for \_SB.PCCH. */
168 static void __init acpi_pcc_cpufreq_init(void)
169 {
170 acpi_status status;
171 acpi_handle handle;
172
173 status = acpi_get_handle(NULL, "\\_SB", &handle);
174 if (ACPI_FAILURE(status))
175 return;
176
177 if (acpi_has_method(handle, "PCCH"))
178 cpufreq_add_device("pcc-cpufreq");
179 }
180 #else
181 static void __init acpi_pcc_cpufreq_init(void) {}
182 #endif /* CONFIG_X86 */
183
184 /* Initialization */
185 #ifdef CONFIG_ACPI_HOTPLUG_CPU
186 int __weak acpi_map_cpu(acpi_handle handle,
187 phys_cpuid_t physid, u32 acpi_id, int *pcpu)
188 {
189 return -ENODEV;
190 }
191
192 int __weak acpi_unmap_cpu(int cpu)
193 {
194 return -ENODEV;
195 }
196
197 int __weak arch_register_cpu(int cpu)
198 {
199 return -ENODEV;
200 }
201
202 void __weak arch_unregister_cpu(int cpu) {}
203
204 static int acpi_processor_hotadd_init(struct acpi_processor *pr)
205 {
206 unsigned long long sta;
207 acpi_status status;
208 int ret;
209
210 if (invalid_phys_cpuid(pr->phys_id))
211 return -ENODEV;
212
213 status = acpi_evaluate_integer(pr->handle, "_STA", NULL, &sta);
214 if (ACPI_FAILURE(status) || !(sta & ACPI_STA_DEVICE_PRESENT))
215 return -ENODEV;
216
217 cpu_maps_update_begin();
218 cpus_write_lock();
219
220 ret = acpi_map_cpu(pr->handle, pr->phys_id, pr->acpi_id, &pr->id);
221 if (ret)
222 goto out;
223
224 ret = arch_register_cpu(pr->id);
225 if (ret) {
226 acpi_unmap_cpu(pr->id);
227 goto out;
228 }
229
230 /*
231 * CPU got hot-added, but cpu_data is not initialized yet. Set a flag
232 * to delay cpu_idle/throttling initialization and do it when the CPU
233 * gets online for the first time.
234 */
235 pr_info("CPU%d has been hot-added\n", pr->id);
236 pr->flags.need_hotplug_init = 1;
237
238 out:
239 cpus_write_unlock();
240 cpu_maps_update_done();
241 return ret;
242 }
243 #else
244 static inline int acpi_processor_hotadd_init(struct acpi_processor *pr)
245 {
246 return -ENODEV;
247 }
248 #endif /* CONFIG_ACPI_HOTPLUG_CPU */
249
250 static int acpi_processor_get_info(struct acpi_device *device)
251 {
252 union acpi_object object = { 0 };
253 struct acpi_buffer buffer = { sizeof(union acpi_object), &object };
254 struct acpi_processor *pr = acpi_driver_data(device);
255 int device_declaration = 0;
256 acpi_status status = AE_OK;
257 static int cpu0_initialized;
258 unsigned long long value;
259
260 acpi_processor_errata();
261
262 /*
263 * Check to see if we have bus mastering arbitration control. This
264 * is required for proper C3 usage (to maintain cache coherency).
265 */
266 if (acpi_gbl_FADT.pm2_control_block && acpi_gbl_FADT.pm2_control_length) {
267 pr->flags.bm_control = 1;
268 dev_dbg(&device->dev, "Bus mastering arbitration control present\n");
269 } else
270 dev_dbg(&device->dev, "No bus mastering arbitration control\n");
271
272 if (!strcmp(acpi_device_hid(device), ACPI_PROCESSOR_OBJECT_HID)) {
273 /* Declared with "Processor" statement; match ProcessorID */
274 status = acpi_evaluate_object(pr->handle, NULL, NULL, &buffer);
275 if (ACPI_FAILURE(status)) {
276 dev_err(&device->dev,
277 "Failed to evaluate processor object (0x%x)\n",
278 status);
279 return -ENODEV;
280 }
281
282 pr->acpi_id = object.processor.proc_id;
283 } else {
284 /*
285 * Declared with "Device" statement; match _UID.
286 */
287 status = acpi_evaluate_integer(pr->handle, METHOD_NAME__UID,
288 NULL, &value);
289 if (ACPI_FAILURE(status)) {
290 dev_err(&device->dev,
291 "Failed to evaluate processor _UID (0x%x)\n",
292 status);
293 return -ENODEV;
294 }
295 device_declaration = 1;
296 pr->acpi_id = value;
297 }
298
299 if (acpi_duplicate_processor_id(pr->acpi_id)) {
300 if (pr->acpi_id == 0xff)
301 dev_info_once(&device->dev,
302 "Entry not well-defined, consider updating BIOS\n");
303 else
304 dev_err(&device->dev,
305 "Failed to get unique processor _UID (0x%x)\n",
306 pr->acpi_id);
307 return -ENODEV;
308 }
309
310 pr->phys_id = acpi_get_phys_id(pr->handle, device_declaration,
311 pr->acpi_id);
312 if (invalid_phys_cpuid(pr->phys_id))
313 dev_dbg(&device->dev, "Failed to get CPU physical ID.\n");
314
315 pr->id = acpi_map_cpuid(pr->phys_id, pr->acpi_id);
316 if (!cpu0_initialized) {
317 cpu0_initialized = 1;
318 /*
319 * Handle UP system running SMP kernel, with no CPU
320 * entry in MADT
321 */
322 if (!acpi_has_cpu_in_madt() && invalid_logical_cpuid(pr->id) &&
323 (num_online_cpus() == 1))
324 pr->id = 0;
325 /*
326 * Check availability of Processor Performance Control by
327 * looking at the presence of the _PCT object under the first
328 * processor definition.
329 */
330 if (acpi_has_method(pr->handle, "_PCT"))
331 cpufreq_add_device("acpi-cpufreq");
332 }
333
334 /*
335 * Extra Processor objects may be enumerated on MP systems with
336 * less than the max # of CPUs. They should be ignored _iff
337 * they are physically not present.
338 *
339 * NOTE: Even if the processor has a cpuid, it may not be present
340 * because cpuid <-> apicid mapping is persistent now.
341 */
342 if (invalid_logical_cpuid(pr->id) || !cpu_present(pr->id)) {
343 int ret = acpi_processor_hotadd_init(pr);
344
345 if (ret)
346 return ret;
347 }
348
349 /*
350 * On some boxes several processors use the same processor bus id.
351 * But they are located in different scope. For example:
352 * \_SB.SCK0.CPU0
353 * \_SB.SCK1.CPU0
354 * Rename the processor device bus id. And the new bus id will be
355 * generated as the following format:
356 * CPU+CPU ID.
357 */
358 sprintf(acpi_device_bid(device), "CPU%X", pr->id);
359 dev_dbg(&device->dev, "Processor [%d:%d]\n", pr->id, pr->acpi_id);
360
361 if (!object.processor.pblk_address)
362 dev_dbg(&device->dev, "No PBLK (NULL address)\n");
363 else if (object.processor.pblk_length != 6)
364 dev_err(&device->dev, "Invalid PBLK length [%d]\n",
365 object.processor.pblk_length);
366 else {
367 pr->throttling.address = object.processor.pblk_address;
368 pr->throttling.duty_offset = acpi_gbl_FADT.duty_offset;
369 pr->throttling.duty_width = acpi_gbl_FADT.duty_width;
370
371 pr->pblk = object.processor.pblk_address;
372 }
373
374 /*
375 * If ACPI describes a slot number for this CPU, we can use it to
376 * ensure we get the right value in the "physical id" field
377 * of /proc/cpuinfo
378 */
379 status = acpi_evaluate_integer(pr->handle, "_SUN", NULL, &value);
380 if (ACPI_SUCCESS(status))
381 arch_fix_phys_package_id(pr->id, value);
382
383 return 0;
384 }
385
386 /*
387 * Do not put anything in here which needs the core to be online.
388 * For example MSR access or setting up things which check for cpuinfo_x86
389 * (cpu_data(cpu)) values, like CPU feature flags, family, model, etc.
390 * Such things have to be put in and set up by the processor driver's .probe().
391 */
392 static DEFINE_PER_CPU(void *, processor_device_array);
393
394 static int acpi_processor_add(struct acpi_device *device,
395 const struct acpi_device_id *id)
396 {
397 struct acpi_processor *pr;
398 struct device *dev;
399 int result = 0;
400
401 pr = kzalloc(sizeof(struct acpi_processor), GFP_KERNEL);
402 if (!pr)
403 return -ENOMEM;
404
405 if (!zalloc_cpumask_var(&pr->throttling.shared_cpu_map, GFP_KERNEL)) {
406 result = -ENOMEM;
407 goto err_free_pr;
408 }
409
410 pr->handle = device->handle;
411 strcpy(acpi_device_name(device), ACPI_PROCESSOR_DEVICE_NAME);
412 strcpy(acpi_device_class(device), ACPI_PROCESSOR_CLASS);
413 device->driver_data = pr;
414
415 result = acpi_processor_get_info(device);
416 if (result) /* Processor is not physically present or unavailable */
417 return 0;
418
419 BUG_ON(pr->id >= nr_cpu_ids);
420
421 /*
422 * Buggy BIOS check.
423 * ACPI id of processors can be reported wrongly by the BIOS.
424 * Don't trust it blindly
425 */
426 if (per_cpu(processor_device_array, pr->id) != NULL &&
427 per_cpu(processor_device_array, pr->id) != device) {
428 dev_warn(&device->dev,
429 "BIOS reported wrong ACPI id %d for the processor\n",
430 pr->id);
431 /* Give up, but do not abort the namespace scan. */
432 goto err;
433 }
434 /*
435 * processor_device_array is not cleared on errors to allow buggy BIOS
436 * checks.
437 */
438 per_cpu(processor_device_array, pr->id) = device;
439 per_cpu(processors, pr->id) = pr;
440
441 dev = get_cpu_device(pr->id);
442 if (!dev) {
443 result = -ENODEV;
444 goto err;
445 }
446
447 result = acpi_bind_one(dev, device);
448 if (result)
449 goto err;
450
451 pr->dev = dev;
452
453 /* Trigger the processor driver's .probe() if present. */
454 if (device_attach(dev) >= 0)
455 return 1;
456
457 dev_err(dev, "Processor driver could not be attached\n");
458 acpi_unbind_one(dev);
459
460 err:
461 free_cpumask_var(pr->throttling.shared_cpu_map);
462 device->driver_data = NULL;
463 per_cpu(processors, pr->id) = NULL;
464 err_free_pr:
465 kfree(pr);
466 return result;
467 }
468
469 #ifdef CONFIG_ACPI_HOTPLUG_CPU
470 /* Removal */
471 static void acpi_processor_remove(struct acpi_device *device)
472 {
473 struct acpi_processor *pr;
474
475 if (!device || !acpi_driver_data(device))
476 return;
477
478 pr = acpi_driver_data(device);
479 if (pr->id >= nr_cpu_ids)
480 goto out;
481
482 /*
483 * The only reason why we ever get here is CPU hot-removal. The CPU is
484 * already offline and the ACPI device removal locking prevents it from
485 * being put back online at this point.
486 *
487 * Unbind the driver from the processor device and detach it from the
488 * ACPI companion object.
489 */
490 device_release_driver(pr->dev);
491 acpi_unbind_one(pr->dev);
492
493 /* Clean up. */
494 per_cpu(processor_device_array, pr->id) = NULL;
495 per_cpu(processors, pr->id) = NULL;
496
497 cpu_maps_update_begin();
498 cpus_write_lock();
499
500 /* Remove the CPU. */
501 arch_unregister_cpu(pr->id);
502 acpi_unmap_cpu(pr->id);
503
504 cpus_write_unlock();
505 cpu_maps_update_done();
506
507 try_offline_node(cpu_to_node(pr->id));
508
509 out:
510 free_cpumask_var(pr->throttling.shared_cpu_map);
511 kfree(pr);
512 }
513 #endif /* CONFIG_ACPI_HOTPLUG_CPU */
514
515 #ifdef CONFIG_ARCH_MIGHT_HAVE_ACPI_PDC
516 bool __init processor_physically_present(acpi_handle handle)
517 {
518 int cpuid, type;
519 u32 acpi_id;
520 acpi_status status;
521 acpi_object_type acpi_type;
522 unsigned long long tmp;
523 union acpi_object object = {};
524 struct acpi_buffer buffer = { sizeof(union acpi_object), &object };
525
526 status = acpi_get_type(handle, &acpi_type);
527 if (ACPI_FAILURE(status))
528 return false;
529
530 switch (acpi_type) {
531 case ACPI_TYPE_PROCESSOR:
532 status = acpi_evaluate_object(handle, NULL, NULL, &buffer);
533 if (ACPI_FAILURE(status))
534 return false;
535 acpi_id = object.processor.proc_id;
536 break;
537 case ACPI_TYPE_DEVICE:
538 status = acpi_evaluate_integer(handle, METHOD_NAME__UID,
539 NULL, &tmp);
540 if (ACPI_FAILURE(status))
541 return false;
542 acpi_id = tmp;
543 break;
544 default:
545 return false;
546 }
547
548 if (xen_initial_domain())
549 /*
550 * When running as a Xen dom0 the number of processors Linux
551 * sees can be different from the real number of processors on
552 * the system, and we still need to execute _PDC or _OSC for
553 * all of them.
554 */
555 return xen_processor_present(acpi_id);
556
557 type = (acpi_type == ACPI_TYPE_DEVICE) ? 1 : 0;
558 cpuid = acpi_get_cpuid(handle, type, acpi_id);
559
560 return !invalid_logical_cpuid(cpuid);
561 }
562
563 /* vendor specific UUID indicating an Intel platform */
564 static u8 sb_uuid_str[] = "4077A616-290C-47BE-9EBD-D87058713953";
565
566 static acpi_status __init acpi_processor_osc(acpi_handle handle, u32 lvl,
567 void *context, void **rv)
568 {
569 u32 capbuf[2] = {};
570 struct acpi_osc_context osc_context = {
571 .uuid_str = sb_uuid_str,
572 .rev = 1,
573 .cap.length = 8,
574 .cap.pointer = capbuf,
575 };
576 acpi_status status;
577
578 if (!processor_physically_present(handle))
579 return AE_OK;
580
581 arch_acpi_set_proc_cap_bits(&capbuf[OSC_SUPPORT_DWORD]);
582
583 status = acpi_run_osc(handle, &osc_context);
584 if (ACPI_FAILURE(status))
585 return status;
586
587 kfree(osc_context.ret.pointer);
588
589 return AE_OK;
590 }
591
592 static bool __init acpi_early_processor_osc(void)
593 {
594 acpi_status status;
595
596 acpi_proc_quirk_mwait_check();
597
598 status = acpi_walk_namespace(ACPI_TYPE_PROCESSOR, ACPI_ROOT_OBJECT,
599 ACPI_UINT32_MAX, acpi_processor_osc, NULL,
600 NULL, NULL);
601 if (ACPI_FAILURE(status))
602 return false;
603
604 status = acpi_get_devices(ACPI_PROCESSOR_DEVICE_HID, acpi_processor_osc,
605 NULL, NULL);
606 if (ACPI_FAILURE(status))
607 return false;
608
609 return true;
610 }
611
612 void __init acpi_early_processor_control_setup(void)
613 {
614 if (acpi_early_processor_osc()) {
615 pr_info("_OSC evaluated successfully for all CPUs\n");
616 } else {
617 pr_info("_OSC evaluation for CPUs failed, trying _PDC\n");
618 acpi_early_processor_set_pdc();
619 }
620 }
621 #endif
622
623 /*
624 * The following ACPI IDs are known to be suitable for representing as
625 * processor devices.
626 */
627 static const struct acpi_device_id processor_device_ids[] = {
628
629 { ACPI_PROCESSOR_OBJECT_HID, },
630 { ACPI_PROCESSOR_DEVICE_HID, },
631
632 { }
633 };
634
635 static struct acpi_scan_handler processor_handler = {
636 .ids = processor_device_ids,
637 .attach = acpi_processor_add,
638 #ifdef CONFIG_ACPI_HOTPLUG_CPU
639 .detach = acpi_processor_remove,
640 #endif
641 .hotplug = {
642 .enabled = true,
643 },
644 };
645
646 static int acpi_processor_container_attach(struct acpi_device *dev,
647 const struct acpi_device_id *id)
648 {
649 return 1;
650 }
651
652 static const struct acpi_device_id processor_container_ids[] = {
653 { ACPI_PROCESSOR_CONTAINER_HID, },
654 { }
655 };
656
657 static struct acpi_scan_handler processor_container_handler = {
658 .ids = processor_container_ids,
659 .attach = acpi_processor_container_attach,
660 };
661
662 /* The number of the unique processor IDs */
663 static int nr_unique_ids __initdata;
664
665 /* The number of the duplicate processor IDs */
666 static int nr_duplicate_ids;
667
668 /* Used to store the unique processor IDs */
669 static int unique_processor_ids[] __initdata = {
670 [0 ... NR_CPUS - 1] = -1,
671 };
672
673 /* Used to store the duplicate processor IDs */
674 static int duplicate_processor_ids[] = {
675 [0 ... NR_CPUS - 1] = -1,
676 };
677
678 static void __init processor_validated_ids_update(int proc_id)
679 {
680 int i;
681
682 if (nr_unique_ids == NR_CPUS||nr_duplicate_ids == NR_CPUS)
683 return;
684
685 /*
686 * Firstly, compare the proc_id with duplicate IDs, if the proc_id is
687 * already in the IDs, do nothing.
688 */
689 for (i = 0; i < nr_duplicate_ids; i++) {
690 if (duplicate_processor_ids[i] == proc_id)
691 return;
692 }
693
694 /*
695 * Secondly, compare the proc_id with unique IDs, if the proc_id is in
696 * the IDs, put it in the duplicate IDs.
697 */
698 for (i = 0; i < nr_unique_ids; i++) {
699 if (unique_processor_ids[i] == proc_id) {
700 duplicate_processor_ids[nr_duplicate_ids] = proc_id;
701 nr_duplicate_ids++;
702 return;
703 }
704 }
705
706 /*
707 * Lastly, the proc_id is a unique ID, put it in the unique IDs.
708 */
709 unique_processor_ids[nr_unique_ids] = proc_id;
710 nr_unique_ids++;
711 }
712
713 static acpi_status __init acpi_processor_ids_walk(acpi_handle handle,
714 u32 lvl,
715 void *context,
716 void **rv)
717 {
718 acpi_status status;
719 acpi_object_type acpi_type;
720 unsigned long long uid;
721 union acpi_object object = { 0 };
722 struct acpi_buffer buffer = { sizeof(union acpi_object), &object };
723
724 status = acpi_get_type(handle, &acpi_type);
725 if (ACPI_FAILURE(status))
726 return status;
727
728 switch (acpi_type) {
729 case ACPI_TYPE_PROCESSOR:
730 status = acpi_evaluate_object(handle, NULL, NULL, &buffer);
731 if (ACPI_FAILURE(status))
732 goto err;
733 uid = object.processor.proc_id;
734 break;
735
736 case ACPI_TYPE_DEVICE:
737 status = acpi_evaluate_integer(handle, "_UID", NULL, &uid);
738 if (ACPI_FAILURE(status))
739 goto err;
740 break;
741 default:
742 goto err;
743 }
744
745 processor_validated_ids_update(uid);
746 return AE_OK;
747
748 err:
749 /* Exit on error, but don't abort the namespace walk */
750 acpi_handle_info(handle, "Invalid processor object\n");
751 return AE_OK;
752
753 }
754
755 static void __init acpi_processor_check_duplicates(void)
756 {
757 /* check the correctness for all processors in ACPI namespace */
758 acpi_walk_namespace(ACPI_TYPE_PROCESSOR, ACPI_ROOT_OBJECT,
759 ACPI_UINT32_MAX,
760 acpi_processor_ids_walk,
761 NULL, NULL, NULL);
762 acpi_get_devices(ACPI_PROCESSOR_DEVICE_HID, acpi_processor_ids_walk,
763 NULL, NULL);
764 }
765
766 bool acpi_duplicate_processor_id(int proc_id)
767 {
768 int i;
769
770 /*
771 * compare the proc_id with duplicate IDs, if the proc_id is already
772 * in the duplicate IDs, return true, otherwise, return false.
773 */
774 for (i = 0; i < nr_duplicate_ids; i++) {
775 if (duplicate_processor_ids[i] == proc_id)
776 return true;
777 }
778 return false;
779 }
780
781 void __init acpi_processor_init(void)
782 {
783 acpi_processor_check_duplicates();
784 acpi_scan_add_handler_with_hotplug(&processor_handler, "processor");
785 acpi_scan_add_handler(&processor_container_handler);
786 acpi_pcc_cpufreq_init();
787 }
788
789 #ifdef CONFIG_ACPI_PROCESSOR_CSTATE
790 /**
791 * acpi_processor_claim_cst_control - Request _CST control from the platform.
792 */
793 bool acpi_processor_claim_cst_control(void)
794 {
795 static bool cst_control_claimed;
796 acpi_status status;
797
798 if (!acpi_gbl_FADT.cst_control || cst_control_claimed)
799 return true;
800
801 status = acpi_os_write_port(acpi_gbl_FADT.smi_command,
802 acpi_gbl_FADT.cst_control, 8);
803 if (ACPI_FAILURE(status)) {
804 pr_warn("ACPI: Failed to claim processor _CST control\n");
805 return false;
806 }
807
808 cst_control_claimed = true;
809 return true;
810 }
811 EXPORT_SYMBOL_GPL(acpi_processor_claim_cst_control);
812
813 /**
814 * acpi_processor_evaluate_cst - Evaluate the processor _CST control method.
815 * @handle: ACPI handle of the processor object containing the _CST.
816 * @cpu: The numeric ID of the target CPU.
817 * @info: Object write the C-states information into.
818 *
819 * Extract the C-state information for the given CPU from the output of the _CST
820 * control method under the corresponding ACPI processor object (or processor
821 * device object) and populate @info with it.
822 *
823 * If any ACPI_ADR_SPACE_FIXED_HARDWARE C-states are found, invoke
824 * acpi_processor_ffh_cstate_probe() to verify them and update the
825 * cpu_cstate_entry data for @cpu.
826 */
827 int acpi_processor_evaluate_cst(acpi_handle handle, u32 cpu,
828 struct acpi_processor_power *info)
829 {
830 struct acpi_buffer buffer = { ACPI_ALLOCATE_BUFFER, NULL };
831 union acpi_object *cst;
832 acpi_status status;
833 u64 count;
834 int last_index = 0;
835 int i, ret = 0;
836
837 status = acpi_evaluate_object(handle, "_CST", NULL, &buffer);
838 if (ACPI_FAILURE(status)) {
839 acpi_handle_debug(handle, "No _CST\n");
840 return -ENODEV;
841 }
842
843 cst = buffer.pointer;
844
845 /* There must be at least 2 elements. */
846 if (!cst || cst->type != ACPI_TYPE_PACKAGE || cst->package.count < 2) {
847 acpi_handle_warn(handle, "Invalid _CST output\n");
848 ret = -EFAULT;
849 goto end;
850 }
851
852 count = cst->package.elements[0].integer.value;
853
854 /* Validate the number of C-states. */
855 if (count < 1 || count != cst->package.count - 1) {
856 acpi_handle_warn(handle, "Inconsistent _CST data\n");
857 ret = -EFAULT;
858 goto end;
859 }
860
861 for (i = 1; i <= count; i++) {
862 union acpi_object *element;
863 union acpi_object *obj;
864 struct acpi_power_register *reg;
865 struct acpi_processor_cx cx;
866
867 /*
868 * If there is not enough space for all C-states, skip the
869 * excess ones and log a warning.
870 */
871 if (last_index >= ACPI_PROCESSOR_MAX_POWER - 1) {
872 acpi_handle_warn(handle,
873 "No room for more idle states (limit: %d)\n",
874 ACPI_PROCESSOR_MAX_POWER - 1);
875 break;
876 }
877
878 memset(&cx, 0, sizeof(cx));
879
880 element = &cst->package.elements[i];
881 if (element->type != ACPI_TYPE_PACKAGE) {
882 acpi_handle_info(handle, "_CST C%d type(%x) is not package, skip...\n",
883 i, element->type);
884 continue;
885 }
886
887 if (element->package.count != 4) {
888 acpi_handle_info(handle, "_CST C%d package count(%d) is not 4, skip...\n",
889 i, element->package.count);
890 continue;
891 }
892
893 obj = &element->package.elements[0];
894
895 if (obj->type != ACPI_TYPE_BUFFER) {
896 acpi_handle_info(handle, "_CST C%d package element[0] type(%x) is not buffer, skip...\n",
897 i, obj->type);
898 continue;
899 }
900
901 reg = (struct acpi_power_register *)obj->buffer.pointer;
902
903 obj = &element->package.elements[1];
904 if (obj->type != ACPI_TYPE_INTEGER) {
905 acpi_handle_info(handle, "_CST C[%d] package element[1] type(%x) is not integer, skip...\n",
906 i, obj->type);
907 continue;
908 }
909
910 cx.type = obj->integer.value;
911 /*
912 * There are known cases in which the _CST output does not
913 * contain C1, so if the type of the first state found is not
914 * C1, leave an empty slot for C1 to be filled in later.
915 */
916 if (i == 1 && cx.type != ACPI_STATE_C1)
917 last_index = 1;
918
919 cx.address = reg->address;
920 cx.index = last_index + 1;
921
922 if (reg->space_id == ACPI_ADR_SPACE_FIXED_HARDWARE) {
923 if (!acpi_processor_ffh_cstate_probe(cpu, &cx, reg)) {
924 /*
925 * In the majority of cases _CST describes C1 as
926 * a FIXED_HARDWARE C-state, but if the command
927 * line forbids using MWAIT, use CSTATE_HALT for
928 * C1 regardless.
929 */
930 if (cx.type == ACPI_STATE_C1 &&
931 boot_option_idle_override == IDLE_NOMWAIT) {
932 cx.entry_method = ACPI_CSTATE_HALT;
933 snprintf(cx.desc, ACPI_CX_DESC_LEN, "ACPI HLT");
934 } else {
935 cx.entry_method = ACPI_CSTATE_FFH;
936 }
937 } else if (cx.type == ACPI_STATE_C1) {
938 /*
939 * In the special case of C1, FIXED_HARDWARE can
940 * be handled by executing the HLT instruction.
941 */
942 cx.entry_method = ACPI_CSTATE_HALT;
943 snprintf(cx.desc, ACPI_CX_DESC_LEN, "ACPI HLT");
944 } else {
945 acpi_handle_info(handle, "_CST C%d declares FIXED_HARDWARE C-state but not supported in hardware, skip...\n",
946 i);
947 continue;
948 }
949 } else if (reg->space_id == ACPI_ADR_SPACE_SYSTEM_IO) {
950 cx.entry_method = ACPI_CSTATE_SYSTEMIO;
951 snprintf(cx.desc, ACPI_CX_DESC_LEN, "ACPI IOPORT 0x%x",
952 cx.address);
953 } else {
954 acpi_handle_info(handle, "_CST C%d space_id(%x) neither FIXED_HARDWARE nor SYSTEM_IO, skip...\n",
955 i, reg->space_id);
956 continue;
957 }
958
959 if (cx.type == ACPI_STATE_C1)
960 cx.valid = 1;
961
962 obj = &element->package.elements[2];
963 if (obj->type != ACPI_TYPE_INTEGER) {
964 acpi_handle_info(handle, "_CST C%d package element[2] type(%x) not integer, skip...\n",
965 i, obj->type);
966 continue;
967 }
968
969 cx.latency = obj->integer.value;
970
971 obj = &element->package.elements[3];
972 if (obj->type != ACPI_TYPE_INTEGER) {
973 acpi_handle_info(handle, "_CST C%d package element[3] type(%x) not integer, skip...\n",
974 i, obj->type);
975 continue;
976 }
977
978 memcpy(&info->states[++last_index], &cx, sizeof(cx));
979 }
980
981 acpi_handle_info(handle, "Found %d idle states\n", last_index);
982
983 info->count = last_index;
984
985 end:
986 kfree(buffer.pointer);
987
988 return ret;
989 }
990 EXPORT_SYMBOL_GPL(acpi_processor_evaluate_cst);
991 #endif /* CONFIG_ACPI_PROCESSOR_CSTATE */