// SPDX-License-Identifier: GPL-2.0 /* * acpi.c - Architecture-Specific Low-Level ACPI Boot Support * * Author: Jianmin Lv * Huacai Chen * Copyright (C) 2020-2022 Loongson Technology Corporation Limited */ #include #include #include #include #include #include #include #include #include #include #include int acpi_disabled; EXPORT_SYMBOL(acpi_disabled); int acpi_noirq; int acpi_pci_disabled; EXPORT_SYMBOL(acpi_pci_disabled); int acpi_strict = 1; /* We have no workarounds on LoongArch */ int num_processors; int disabled_cpus; u64 acpi_saved_sp; #define PREFIX "ACPI: " struct acpi_madt_core_pic acpi_core_pic[MAX_CORE_PIC]; void __init __iomem * __acpi_map_table(unsigned long phys, unsigned long size) { if (!phys || !size) return NULL; return early_memremap(phys, size); } void __init __acpi_unmap_table(void __iomem *map, unsigned long size) { if (!map || !size) return; early_memunmap(map, size); } void __iomem *acpi_os_ioremap(acpi_physical_address phys, acpi_size size) { if (!memblock_is_memory(phys)) return ioremap(phys, size); else return ioremap_cache(phys, size); } #ifdef CONFIG_SMP static int set_processor_mask(u32 id, u32 pass) { int cpu = -1, cpuid = id; if (num_processors >= NR_CPUS) { pr_warn(PREFIX "nr_cpus limit of %i reached." " processor 0x%x ignored.\n", NR_CPUS, cpuid); return -ENODEV; } if (cpuid == loongson_sysconf.boot_cpu_id) cpu = 0; switch (pass) { case 1: /* Pass 1 handle enabled processors */ if (cpu < 0) cpu = find_first_zero_bit(cpumask_bits(cpu_present_mask), NR_CPUS); num_processors++; set_cpu_present(cpu, true); break; case 2: /* Pass 2 handle disabled processors */ if (cpu < 0) cpu = find_first_zero_bit(cpumask_bits(cpu_possible_mask), NR_CPUS); disabled_cpus++; break; default: return cpu; } set_cpu_possible(cpu, true); __cpu_number_map[cpuid] = cpu; __cpu_logical_map[cpu] = cpuid; return cpu; } #endif static int __init acpi_parse_p1_processor(union acpi_subtable_headers *header, const unsigned long end) { struct acpi_madt_core_pic *processor = NULL; processor = (struct acpi_madt_core_pic *)header; if (BAD_MADT_ENTRY(processor, end)) return -EINVAL; acpi_table_print_madt_entry(&header->common); #ifdef CONFIG_SMP acpi_core_pic[processor->core_id] = *processor; if (processor->flags & ACPI_MADT_ENABLED) set_processor_mask(processor->core_id, 1); #endif return 0; } static int __init acpi_parse_p2_processor(union acpi_subtable_headers *header, const unsigned long end) { struct acpi_madt_core_pic *processor = NULL; processor = (struct acpi_madt_core_pic *)header; if (BAD_MADT_ENTRY(processor, end)) return -EINVAL; #ifdef CONFIG_SMP if (!(processor->flags & ACPI_MADT_ENABLED)) set_processor_mask(processor->core_id, 2); #endif return 0; } static int __init acpi_parse_eio_master(union acpi_subtable_headers *header, const unsigned long end) { static int core = 0; struct acpi_madt_eio_pic *eiointc = NULL; eiointc = (struct acpi_madt_eio_pic *)header; if (BAD_MADT_ENTRY(eiointc, end)) return -EINVAL; core = eiointc->node * CORES_PER_EIO_NODE; set_bit(core, loongson_sysconf.cores_io_master); return 0; } static void __init acpi_process_madt(void) { #ifdef CONFIG_SMP int i; for (i = 0; i < NR_CPUS; i++) { __cpu_number_map[i] = -1; __cpu_logical_map[i] = -1; } #endif acpi_table_parse_madt(ACPI_MADT_TYPE_CORE_PIC, acpi_parse_p1_processor, MAX_CORE_PIC); acpi_table_parse_madt(ACPI_MADT_TYPE_CORE_PIC, acpi_parse_p2_processor, MAX_CORE_PIC); acpi_table_parse_madt(ACPI_MADT_TYPE_EIO_PIC, acpi_parse_eio_master, MAX_IO_PICS); loongson_sysconf.nr_cpus = num_processors; } int pptt_enabled; int __init parse_acpi_topology(void) { int cpu, topology_id; for_each_possible_cpu(cpu) { topology_id = find_acpi_cpu_topology(cpu, 0); if (topology_id < 0) { pr_warn("Invalid BIOS PPTT\n"); return -ENOENT; } if (acpi_pptt_cpu_is_thread(cpu) <= 0) cpu_data[cpu].core = topology_id; else { topology_id = find_acpi_cpu_topology(cpu, 1); if (topology_id < 0) return -ENOENT; cpu_data[cpu].core = topology_id; } } pptt_enabled = 1; return 0; } #ifndef CONFIG_SUSPEND int (*acpi_suspend_lowlevel)(void); #else int (*acpi_suspend_lowlevel)(void) = loongarch_acpi_suspend; #endif void __init acpi_boot_table_init(void) { /* * If acpi_disabled, bail out */ if (acpi_disabled) goto fdt_earlycon; /* * Initialize the ACPI boot-time table parser. */ if (acpi_table_init()) { disable_acpi(); goto fdt_earlycon; } loongson_sysconf.boot_cpu_id = read_csr_cpuid(); /* * Process the Multiple APIC Description Table (MADT), if present */ acpi_process_madt(); /* Do not enable ACPI SPCR console by default */ acpi_parse_spcr(earlycon_acpi_spcr_enable, false); if (IS_ENABLED(CONFIG_ACPI_BGRT)) acpi_table_parse(ACPI_SIG_BGRT, acpi_parse_bgrt); return; fdt_earlycon: if (earlycon_acpi_spcr_enable) early_init_dt_scan_chosen_stdout(); } #ifdef CONFIG_ACPI_NUMA static __init int setup_node(int pxm) { return acpi_map_pxm_to_node(pxm); } /* * Callback for SLIT parsing. pxm_to_node() returns NUMA_NO_NODE for * I/O localities since SRAT does not list them. I/O localities are * not supported at this point. */ unsigned int numa_distance_cnt; static inline unsigned int get_numa_distances_cnt(struct acpi_table_slit *slit) { return slit->locality_count; } void __init numa_set_distance(int from, int to, int distance) { if ((u8)distance != distance || (from == to && distance != LOCAL_DISTANCE)) { pr_warn_once("Warning: invalid distance parameter, from=%d to=%d distance=%d\n", from, to, distance); return; } node_distances[from][to] = distance; } /* Callback for Proximity Domain -> CPUID mapping */ void __init acpi_numa_processor_affinity_init(struct acpi_srat_cpu_affinity *pa) { int pxm, node; if (srat_disabled()) return; if (pa->header.length != sizeof(struct acpi_srat_cpu_affinity)) { bad_srat(); return; } if ((pa->flags & ACPI_SRAT_CPU_ENABLED) == 0) return; pxm = pa->proximity_domain_lo; if (acpi_srat_revision >= 2) { pxm |= (pa->proximity_domain_hi[0] << 8); pxm |= (pa->proximity_domain_hi[1] << 16); pxm |= (pa->proximity_domain_hi[2] << 24); } node = setup_node(pxm); if (node < 0) { pr_err("SRAT: Too many proximity domains %x\n", pxm); bad_srat(); return; } if (pa->apic_id >= CONFIG_NR_CPUS) { pr_info("SRAT: PXM %u -> CPU 0x%02x -> Node %u skipped apicid that is too big\n", pxm, pa->apic_id, node); return; } early_numa_add_cpu(pa->apic_id, node); set_cpuid_to_node(pa->apic_id, node); node_set(node, numa_nodes_parsed); pr_info("SRAT: PXM %u -> CPU 0x%02x -> Node %u\n", pxm, pa->apic_id, node); } #endif void __init arch_reserve_mem_area(acpi_physical_address addr, size_t size) { memblock_reserve(addr, size); } #ifdef CONFIG_ACPI_HOTPLUG_CPU #include static int __ref acpi_map_cpu2node(acpi_handle handle, int cpu, int physid) { #ifdef CONFIG_ACPI_NUMA int nid; nid = acpi_get_node(handle); if (nid != NUMA_NO_NODE) nid = early_cpu_to_node(cpu); if (nid != NUMA_NO_NODE) { set_cpuid_to_node(physid, nid); node_set(nid, numa_nodes_parsed); set_cpu_numa_node(cpu, nid); cpumask_set_cpu(cpu, cpumask_of_node(nid)); } #endif return 0; } int acpi_map_cpu(acpi_handle handle, phys_cpuid_t physid, u32 acpi_id, int *pcpu) { int cpu; cpu = cpu_number_map(physid); if (cpu < 0 || cpu >= nr_cpu_ids) { pr_info(PREFIX "Unable to map lapic to logical cpu number\n"); return -ERANGE; } num_processors++; set_cpu_present(cpu, true); acpi_map_cpu2node(handle, cpu, physid); *pcpu = cpu; return 0; } EXPORT_SYMBOL(acpi_map_cpu); int acpi_unmap_cpu(int cpu) { #ifdef CONFIG_ACPI_NUMA set_cpuid_to_node(cpu_logical_map(cpu), NUMA_NO_NODE); #endif set_cpu_present(cpu, false); num_processors--; pr_info("cpu%d hot remove!\n", cpu); return 0; } EXPORT_SYMBOL(acpi_unmap_cpu); #endif /* CONFIG_ACPI_HOTPLUG_CPU */