1/* SPDX-License-Identifier: GPL-2.0-or-later */
2#ifndef _ASM_X86_ACPI_H
3#define _ASM_X86_ACPI_H
4
5/*
6 * Copyright (C) 2001 Paul Diefenbaugh <paul.s.diefenbaugh@intel.com>
7 * Copyright (C) 2001 Patrick Mochel <mochel@osdl.org>
8 */
9#include <acpi/proc_cap_intel.h>
10
11#include <asm/numa.h>
12#include <asm/fixmap.h>
13#include <asm/processor.h>
14#include <asm/mmu.h>
15#include <asm/mpspec.h>
16#include <asm/x86_init.h>
17#include <asm/cpufeature.h>
18#include <asm/irq_vectors.h>
19
20#ifdef CONFIG_ACPI_APEI
21# include <asm/pgtable_types.h>
22#endif
23
24#ifdef CONFIG_ACPI
25extern int acpi_lapic;
26extern int acpi_ioapic;
27extern int acpi_noirq;
28extern int acpi_strict;
29extern int acpi_disabled;
30extern int acpi_pci_disabled;
31extern int acpi_skip_timer_override;
32extern int acpi_use_timer_override;
33extern int acpi_fix_pin2_polarity;
34extern int acpi_disable_cmcff;
35extern bool acpi_int_src_ovr[NR_IRQS_LEGACY];
36
37extern u8 acpi_sci_flags;
38extern u32 acpi_sci_override_gsi;
39void acpi_pic_sci_set_trigger(unsigned int, u16);
40
41struct device;
42
43extern int (*__acpi_register_gsi)(struct device *dev, u32 gsi,
44 int trigger, int polarity);
45extern void (*__acpi_unregister_gsi)(u32 gsi);
46
47static inline void disable_acpi(void)
48{
49 acpi_disabled = 1;
50 acpi_pci_disabled = 1;
51 acpi_noirq = 1;
52}
53
54extern int acpi_gsi_to_irq(u32 gsi, unsigned int *irq);
55
56static inline void acpi_noirq_set(void) { acpi_noirq = 1; }
57static inline void acpi_disable_pci(void)
58{
59 acpi_pci_disabled = 1;
60 acpi_noirq_set();
61}
62
63/* Low-level suspend routine. */
64extern int (*acpi_suspend_lowlevel)(void);
65
66/* Physical address to resume after wakeup */
67unsigned long acpi_get_wakeup_address(void);
68
69static inline bool acpi_skip_set_wakeup_address(void)
70{
71 return cpu_feature_enabled(X86_FEATURE_XENPV);
72}
73
74#define acpi_skip_set_wakeup_address acpi_skip_set_wakeup_address
75
76/*
77 * Check if the CPU can handle C2 and deeper
78 */
79static inline unsigned int acpi_processor_cstate_check(unsigned int max_cstate)
80{
81 /*
82 * Early models (<=5) of AMD Opterons are not supposed to go into
83 * C2 state.
84 *
85 * Steppings 0x0A and later are good
86 */
87 if (boot_cpu_data.x86 == 0x0F &&
88 boot_cpu_data.x86_vendor == X86_VENDOR_AMD &&
89 boot_cpu_data.x86_model <= 0x05 &&
90 boot_cpu_data.x86_stepping < 0x0A)
91 return 1;
92 else if (boot_cpu_has(X86_BUG_AMD_APIC_C1E))
93 return 1;
94 else
95 return max_cstate;
96}
97
98static inline bool arch_has_acpi_pdc(void)
99{
100 struct cpuinfo_x86 *c = &cpu_data(0);
101 return (c->x86_vendor == X86_VENDOR_INTEL ||
102 c->x86_vendor == X86_VENDOR_CENTAUR);
103}
104
105static inline void arch_acpi_set_proc_cap_bits(u32 *cap)
106{
107 struct cpuinfo_x86 *c = &cpu_data(0);
108
109 *cap |= ACPI_PROC_CAP_C_CAPABILITY_SMP;
110
111 /* Enable coordination with firmware's _TSD info */
112 *cap |= ACPI_PROC_CAP_SMP_T_SWCOORD;
113
114 if (cpu_has(c, X86_FEATURE_EST))
115 *cap |= ACPI_PROC_CAP_EST_CAPABILITY_SWSMP;
116
117 if (cpu_has(c, X86_FEATURE_ACPI))
118 *cap |= ACPI_PROC_CAP_T_FFH;
119
120 if (cpu_has(c, X86_FEATURE_HWP))
121 *cap |= ACPI_PROC_CAP_COLLAB_PROC_PERF;
122
123 /*
124 * If mwait/monitor is unsupported, C_C1_FFH and
125 * C2/C3_FFH will be disabled.
126 */
127 if (!cpu_has(c, X86_FEATURE_MWAIT) ||
128 boot_option_idle_override == IDLE_NOMWAIT)
129 *cap &= ~(ACPI_PROC_CAP_C_C1_FFH | ACPI_PROC_CAP_C_C2C3_FFH);
130}
131
132static inline bool acpi_has_cpu_in_madt(void)
133{
134 return !!acpi_lapic;
135}
136
137#define ACPI_HAVE_ARCH_SET_ROOT_POINTER
138static inline void acpi_arch_set_root_pointer(u64 addr)
139{
140 x86_init.acpi.set_root_pointer(addr);
141}
142
143#define ACPI_HAVE_ARCH_GET_ROOT_POINTER
144static inline u64 acpi_arch_get_root_pointer(void)
145{
146 return x86_init.acpi.get_root_pointer();
147}
148
149void acpi_generic_reduced_hw_init(void);
150
151void x86_default_set_root_pointer(u64 addr);
152u64 x86_default_get_root_pointer(void);
153
154#else /* !CONFIG_ACPI */
155
156#define acpi_lapic 0
157#define acpi_ioapic 0
158#define acpi_disable_cmcff 0
159static inline void acpi_noirq_set(void) { }
160static inline void acpi_disable_pci(void) { }
161static inline void disable_acpi(void) { }
162
163static inline void acpi_generic_reduced_hw_init(void) { }
164
165static inline void x86_default_set_root_pointer(u64 addr) { }
166
167static inline u64 x86_default_get_root_pointer(void)
168{
169 return 0;
170}
171
172#endif /* !CONFIG_ACPI */
173
174#define ARCH_HAS_POWER_INIT 1
175
176#ifdef CONFIG_ACPI_NUMA
177extern int x86_acpi_numa_init(void);
178#endif /* CONFIG_ACPI_NUMA */
179
180struct cper_ia_proc_ctx;
181
182#ifdef CONFIG_ACPI_APEI
183static inline pgprot_t arch_apei_get_mem_attribute(phys_addr_t addr)
184{
185 /*
186 * We currently have no way to look up the EFI memory map
187 * attributes for a region in a consistent way, because the
188 * memmap is discarded after efi_free_boot_services(). So if
189 * you call efi_mem_attributes() during boot and at runtime,
190 * you could theoretically see different attributes.
191 *
192 * We are yet to see any x86 platforms that require anything
193 * other than PAGE_KERNEL (some ARM64 platforms require the
194 * equivalent of PAGE_KERNEL_NOCACHE). Additionally, if SME
195 * is active, the ACPI information will not be encrypted,
196 * so return PAGE_KERNEL_NOENC until we know differently.
197 */
198 return PAGE_KERNEL_NOENC;
199}
200
201int arch_apei_report_x86_error(struct cper_ia_proc_ctx *ctx_info,
202 u64 lapic_id);
203#else
204static inline int arch_apei_report_x86_error(struct cper_ia_proc_ctx *ctx_info,
205 u64 lapic_id)
206{
207 return -EINVAL;
208}
209#endif
210
211#define ACPI_TABLE_UPGRADE_MAX_PHYS (max_low_pfn_mapped << PAGE_SHIFT)
212
213#endif /* _ASM_X86_ACPI_H */
214

source code of linux/arch/x86/include/asm/acpi.h