Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 | // SPDX-License-Identifier: GPL-2.0 /* * ACPI 6.6 based NUMA setup for RISCV * Lots of code was borrowed from arch/arm64/kernel/acpi_numa.c * * Copyright 2004 Andi Kleen, SuSE Labs. * Copyright (C) 2013-2016, Linaro Ltd. * Author: Hanjun Guo <hanjun.guo@linaro.org> * Copyright (C) 2024 Intel Corporation. * * Reads the ACPI SRAT table to figure out what memory belongs to which CPUs. * * Called from acpi_numa_init while reading the SRAT and SLIT tables. * Assumes all memory regions belonging to a single proximity domain * are in one chunk. Holes between them will be included in the node. */ #define pr_fmt(fmt) "ACPI: NUMA: " fmt #include <linux/acpi.h> #include <linux/bitmap.h> #include <linux/kernel.h> #include <linux/mm.h> #include <linux/memblock.h> #include <linux/mmzone.h> #include <linux/module.h> #include <linux/topology.h> #include <asm/numa.h> static int acpi_early_node_map[NR_CPUS] __initdata = { [0 ... NR_CPUS - 1] = NUMA_NO_NODE }; int __init acpi_numa_get_nid(unsigned int cpu) { return acpi_early_node_map[cpu]; } static inline int get_cpu_for_acpi_id(u32 uid) { int cpu; for (cpu = 0; cpu < nr_cpu_ids; cpu++) if (uid == get_acpi_id_for_cpu(cpu)) return cpu; return -EINVAL; } static int __init acpi_parse_rintc_pxm(union acpi_subtable_headers *header, const unsigned long end) { struct acpi_srat_rintc_affinity *pa; int cpu, pxm, node; if (srat_disabled()) return -EINVAL; pa = (struct acpi_srat_rintc_affinity *)header; if (!pa) return -EINVAL; if (!(pa->flags & ACPI_SRAT_RINTC_ENABLED)) return 0; pxm = pa->proximity_domain; node = pxm_to_node(pxm); /* * If we can't map the UID to a logical cpu this * means that the UID is not part of possible cpus * so we do not need a NUMA mapping for it, skip * the SRAT entry and keep parsing. */ cpu = get_cpu_for_acpi_id(pa->acpi_processor_uid); if (cpu < 0) return 0; acpi_early_node_map[cpu] = node; pr_info("SRAT: PXM %d -> HARTID 0x%lx -> Node %d\n", pxm, cpuid_to_hartid_map(cpu), node); return 0; } void __init acpi_map_cpus_to_nodes(void) { int i; /* * In ACPI, SMP and CPU NUMA information is provided in separate * static tables, namely the MADT and the SRAT. * * Thus, it is simpler to first create the cpu logical map through * an MADT walk and then map the logical cpus to their node ids * as separate steps. */ acpi_table_parse_entries(ACPI_SIG_SRAT, sizeof(struct acpi_table_srat), ACPI_SRAT_TYPE_RINTC_AFFINITY, acpi_parse_rintc_pxm, 0); for (i = 0; i < nr_cpu_ids; i++) early_map_cpu_to_node(i, acpi_numa_get_nid(i)); } /* Callback for Proximity Domain -> logical node ID mapping */ void __init acpi_numa_rintc_affinity_init(struct acpi_srat_rintc_affinity *pa) { int pxm, node; if (srat_disabled()) return; if (pa->header.length < sizeof(struct acpi_srat_rintc_affinity)) { pr_err("SRAT: Invalid SRAT header length: %d\n", pa->header.length); bad_srat(); return; } if (!(pa->flags & ACPI_SRAT_RINTC_ENABLED)) return; pxm = pa->proximity_domain; node = acpi_map_pxm_to_node(pxm); if (node == NUMA_NO_NODE) { pr_err("SRAT: Too many proximity domains %d\n", pxm); bad_srat(); return; } node_set(node, numa_nodes_parsed); } |