^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2) * Written by: Matthew Dobson, IBM Corporation
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4) * Copyright (C) 2002, IBM Corp.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 5) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 6) * All rights reserved.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 7) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 8) * This program is free software; you can redistribute it and/or modify
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 9) * it under the terms of the GNU General Public License as published by
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 10) * the Free Software Foundation; either version 2 of the License, or
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 11) * (at your option) any later version.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 12) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 13) * This program is distributed in the hope that it will be useful, but
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 14) * WITHOUT ANY WARRANTY; without even the implied warranty of
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 15) * MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE, GOOD TITLE or
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 16) * NON INFRINGEMENT. See the GNU General Public License for more
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 17) * details.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 18) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 19) * You should have received a copy of the GNU General Public License
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 20) * along with this program; if not, write to the Free Software
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 21) * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 22) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 23) * Send feedback to <colpatch@us.ibm.com>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 24) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 25) #ifndef _ASM_X86_TOPOLOGY_H
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 26) #define _ASM_X86_TOPOLOGY_H
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 27)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 28) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 29) * to preserve the visibility of NUMA_NO_NODE definition,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 30) * moved to there from here. May be used independent of
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 31) * CONFIG_NUMA.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 32) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 33) #include <linux/numa.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 34)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 35) #ifdef CONFIG_NUMA
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 36) #include <linux/cpumask.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 37)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 38) #include <asm/mpspec.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 39) #include <asm/percpu.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 40)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 41) /* Mappings between logical cpu number and node number */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 42) DECLARE_EARLY_PER_CPU(int, x86_cpu_to_node_map);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 43)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 44) #ifdef CONFIG_DEBUG_PER_CPU_MAPS
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 45) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 46) * override generic percpu implementation of cpu_to_node
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 47) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 48) extern int __cpu_to_node(int cpu);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 49) #define cpu_to_node __cpu_to_node
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 50)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 51) extern int early_cpu_to_node(int cpu);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 52)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 53) #else /* !CONFIG_DEBUG_PER_CPU_MAPS */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 54)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 55) /* Same function but used if called before per_cpu areas are setup */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 56) static inline int early_cpu_to_node(int cpu)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 57) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 58) return early_per_cpu(x86_cpu_to_node_map, cpu);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 59) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 60)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 61) #endif /* !CONFIG_DEBUG_PER_CPU_MAPS */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 62)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 63) /* Mappings between node number and cpus on that node. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 64) extern cpumask_var_t node_to_cpumask_map[MAX_NUMNODES];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 65)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 66) #ifdef CONFIG_DEBUG_PER_CPU_MAPS
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 67) extern const struct cpumask *cpumask_of_node(int node);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 68) #else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 69) /* Returns a pointer to the cpumask of CPUs on Node 'node'. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 70) static inline const struct cpumask *cpumask_of_node(int node)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 71) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 72) return node_to_cpumask_map[node];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 73) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 74) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 75)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 76) extern void setup_node_to_cpumask_map(void);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 77)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 78) #define pcibus_to_node(bus) __pcibus_to_node(bus)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 79)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 80) extern int __node_distance(int, int);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 81) #define node_distance(a, b) __node_distance(a, b)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 82)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 83) #else /* !CONFIG_NUMA */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 84)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 85) static inline int numa_node_id(void)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 86) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 87) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 88) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 89) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 90) * indicate override:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 91) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 92) #define numa_node_id numa_node_id
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 93)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 94) static inline int early_cpu_to_node(int cpu)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 95) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 96) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 97) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 98)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 99) static inline void setup_node_to_cpumask_map(void) { }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 100)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 101) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 102)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 103) #include <asm-generic/topology.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 104)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 105) extern const struct cpumask *cpu_coregroup_mask(int cpu);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 106)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 107) #define topology_logical_package_id(cpu) (cpu_data(cpu).logical_proc_id)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 108) #define topology_physical_package_id(cpu) (cpu_data(cpu).phys_proc_id)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 109) #define topology_logical_die_id(cpu) (cpu_data(cpu).logical_die_id)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 110) #define topology_die_id(cpu) (cpu_data(cpu).cpu_die_id)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 111) #define topology_core_id(cpu) (cpu_data(cpu).cpu_core_id)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 112)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 113) extern unsigned int __max_die_per_package;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 114)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 115) #ifdef CONFIG_SMP
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 116) #define topology_die_cpumask(cpu) (per_cpu(cpu_die_map, cpu))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 117) #define topology_core_cpumask(cpu) (per_cpu(cpu_core_map, cpu))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 118) #define topology_sibling_cpumask(cpu) (per_cpu(cpu_sibling_map, cpu))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 119)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 120) extern unsigned int __max_logical_packages;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 121) #define topology_max_packages() (__max_logical_packages)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 122)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 123) static inline int topology_max_die_per_package(void)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 124) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 125) return __max_die_per_package;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 126) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 127)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 128) extern int __max_smt_threads;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 129)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 130) static inline int topology_max_smt_threads(void)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 131) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 132) return __max_smt_threads;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 133) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 134)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 135) int topology_update_package_map(unsigned int apicid, unsigned int cpu);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 136) int topology_update_die_map(unsigned int dieid, unsigned int cpu);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 137) int topology_phys_to_logical_pkg(unsigned int pkg);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 138) int topology_phys_to_logical_die(unsigned int die, unsigned int cpu);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 139) bool topology_is_primary_thread(unsigned int cpu);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 140) bool topology_smt_supported(void);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 141) #else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 142) #define topology_max_packages() (1)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 143) static inline int
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 144) topology_update_package_map(unsigned int apicid, unsigned int cpu) { return 0; }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 145) static inline int
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 146) topology_update_die_map(unsigned int dieid, unsigned int cpu) { return 0; }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 147) static inline int topology_phys_to_logical_pkg(unsigned int pkg) { return 0; }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 148) static inline int topology_phys_to_logical_die(unsigned int die,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 149) unsigned int cpu) { return 0; }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 150) static inline int topology_max_die_per_package(void) { return 1; }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 151) static inline int topology_max_smt_threads(void) { return 1; }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 152) static inline bool topology_is_primary_thread(unsigned int cpu) { return true; }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 153) static inline bool topology_smt_supported(void) { return false; }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 154) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 155)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 156) static inline void arch_fix_phys_package_id(int num, u32 slot)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 157) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 158) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 159)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 160) struct pci_bus;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 161) int x86_pci_root_bus_node(int bus);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 162) void x86_pci_root_bus_resources(int bus, struct list_head *resources);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 163)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 164) extern bool x86_topology_update;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 165)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 166) #ifdef CONFIG_SCHED_MC_PRIO
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 167) #include <asm/percpu.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 168)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 169) DECLARE_PER_CPU_READ_MOSTLY(int, sched_core_priority);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 170) extern unsigned int __read_mostly sysctl_sched_itmt_enabled;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 171)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 172) /* Interface to set priority of a cpu */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 173) void sched_set_itmt_core_prio(int prio, int core_cpu);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 174)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 175) /* Interface to notify scheduler that system supports ITMT */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 176) int sched_set_itmt_support(void);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 177)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 178) /* Interface to notify scheduler that system revokes ITMT support */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 179) void sched_clear_itmt_support(void);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 180)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 181) #else /* CONFIG_SCHED_MC_PRIO */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 182)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 183) #define sysctl_sched_itmt_enabled 0
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 184) static inline void sched_set_itmt_core_prio(int prio, int core_cpu)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 185) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 186) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 187) static inline int sched_set_itmt_support(void)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 188) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 189) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 190) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 191) static inline void sched_clear_itmt_support(void)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 192) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 193) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 194) #endif /* CONFIG_SCHED_MC_PRIO */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 195)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 196) #if defined(CONFIG_SMP) && defined(CONFIG_X86_64)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 197) #include <asm/cpufeature.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 198)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 199) DECLARE_STATIC_KEY_FALSE(arch_scale_freq_key);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 200)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 201) #define arch_scale_freq_invariant() static_branch_likely(&arch_scale_freq_key)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 202)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 203) DECLARE_PER_CPU(unsigned long, arch_freq_scale);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 204)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 205) static inline long arch_scale_freq_capacity(int cpu)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 206) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 207) return per_cpu(arch_freq_scale, cpu);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 208) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 209) #define arch_scale_freq_capacity arch_scale_freq_capacity
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 210)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 211) extern void arch_scale_freq_tick(void);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 212) #define arch_scale_freq_tick arch_scale_freq_tick
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 213)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 214) extern void arch_set_max_freq_ratio(bool turbo_disabled);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 215) #else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 216) static inline void arch_set_max_freq_ratio(bool turbo_disabled)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 217) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 218) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 219) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 220)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 221) #endif /* _ASM_X86_TOPOLOGY_H */