// SPDX-License-Identifier: GPL-2.0-only /* * Copyright (c) 2024, Google LLC. * Pasha Tatashin */ #include "iommu-pages.h" #include #include #include #define IOPTDESC_MATCH(pg_elm, elm) \ static_assert(offsetof(struct page, pg_elm) == \ offsetof(struct ioptdesc, elm)) IOPTDESC_MATCH(flags, __page_flags); IOPTDESC_MATCH(lru, iopt_freelist_elm); /* Ensure bit 0 is clear */ IOPTDESC_MATCH(mapping, __page_mapping); IOPTDESC_MATCH(private, _private); IOPTDESC_MATCH(page_type, __page_type); IOPTDESC_MATCH(_refcount, __page_refcount); #ifdef CONFIG_MEMCG IOPTDESC_MATCH(memcg_data, memcg_data); #endif #undef IOPTDESC_MATCH static_assert(sizeof(struct ioptdesc) <= sizeof(struct page)); static inline size_t ioptdesc_mem_size(struct ioptdesc *desc) { return 1UL << (folio_order(ioptdesc_folio(desc)) + PAGE_SHIFT); } /** * iommu_alloc_pages_node_sz - Allocate a zeroed page of a given size from * specific NUMA node * @nid: memory NUMA node id * @gfp: buddy allocator flags * @size: Memory size to allocate, rounded up to a power of 2 * * Returns the virtual address of the allocated page. The page must be freed * either by calling iommu_free_pages() or via iommu_put_pages_list(). The * returned allocation is round_up_pow_two(size) big, and is physically aligned * to its size. */ void *iommu_alloc_pages_node_sz(int nid, gfp_t gfp, size_t size) { struct ioptdesc *iopt; unsigned long pgcnt; struct folio *folio; unsigned int order; /* This uses page_address() on the memory. */ if (WARN_ON(gfp & __GFP_HIGHMEM)) return NULL; /* * Currently sub page allocations result in a full page being returned. */ order = get_order(size); /* * __folio_alloc_node() does not handle NUMA_NO_NODE like * alloc_pages_node() did. */ if (nid == NUMA_NO_NODE) nid = numa_mem_id(); folio = __folio_alloc_node(gfp | __GFP_ZERO, order, nid); if (unlikely(!folio)) return NULL; iopt = folio_ioptdesc(folio); iopt->incoherent = false; /* * All page allocations that should be reported to as "iommu-pagetables" * to userspace must use one of the functions below. This includes * allocations of page-tables and other per-iommu_domain configuration * structures. * * This is necessary for the proper accounting as IOMMU state can be * rather large, i.e. multiple gigabytes in size. */ pgcnt = 1UL << order; mod_node_page_state(folio_pgdat(folio), NR_IOMMU_PAGES, pgcnt); lruvec_stat_mod_folio(folio, NR_SECONDARY_PAGETABLE, pgcnt); return folio_address(folio); } EXPORT_SYMBOL_GPL(iommu_alloc_pages_node_sz); static void __iommu_free_desc(struct ioptdesc *iopt) { struct folio *folio = ioptdesc_folio(iopt); const unsigned long pgcnt = folio_nr_pages(folio); if (IOMMU_PAGES_USE_DMA_API) WARN_ON_ONCE(iopt->incoherent); mod_node_page_state(folio_pgdat(folio), NR_IOMMU_PAGES, -pgcnt); lruvec_stat_mod_folio(folio, NR_SECONDARY_PAGETABLE, -pgcnt); folio_put(folio); } /** * iommu_free_pages - free pages * @virt: virtual address of the page to be freed. * * The page must have have been allocated by iommu_alloc_pages_node_sz() */ void iommu_free_pages(void *virt) { if (!virt) return; __iommu_free_desc(virt_to_ioptdesc(virt)); } EXPORT_SYMBOL_GPL(iommu_free_pages); /** * iommu_put_pages_list - free a list of pages. * @list: The list of pages to be freed * * Frees a list of pages allocated by iommu_alloc_pages_node_sz(). On return the * passed list is invalid, the caller must use IOMMU_PAGES_LIST_INIT to reinit * the list if it expects to use it again. */ void iommu_put_pages_list(struct iommu_pages_list *list) { struct ioptdesc *iopt, *tmp; list_for_each_entry_safe(iopt, tmp, &list->pages, iopt_freelist_elm) __iommu_free_desc(iopt); } EXPORT_SYMBOL_GPL(iommu_put_pages_list); /** * iommu_pages_start_incoherent - Setup the page for cache incoherent operation * @virt: The page to setup * @dma_dev: The iommu device * * For incoherent memory this will use the DMA API to manage the cache flushing * on some arches. This is a lot of complexity compared to just calling * arch_sync_dma_for_device(), but it is what the existing ARM iommu drivers * have been doing. The DMA API requires keeping track of the DMA map and * freeing it when required. This keeps track of the dma map inside the ioptdesc * so that error paths are simple for the caller. */ int iommu_pages_start_incoherent(void *virt, struct device *dma_dev) { struct ioptdesc *iopt = virt_to_ioptdesc(virt); dma_addr_t dma; if (WARN_ON(iopt->incoherent)) return -EINVAL; if (!IOMMU_PAGES_USE_DMA_API) { iommu_pages_flush_incoherent(dma_dev, virt, 0, ioptdesc_mem_size(iopt)); } else { dma = dma_map_single(dma_dev, virt, ioptdesc_mem_size(iopt), DMA_TO_DEVICE); if (dma_mapping_error(dma_dev, dma)) return -EINVAL; /* * The DMA API is not allowed to do anything other than DMA * direct. It would be nice to also check * dev_is_dma_coherent(dma_dev)); */ if (WARN_ON(dma != virt_to_phys(virt))) { dma_unmap_single(dma_dev, dma, ioptdesc_mem_size(iopt), DMA_TO_DEVICE); return -EOPNOTSUPP; } } iopt->incoherent = 1; return 0; } EXPORT_SYMBOL_GPL(iommu_pages_start_incoherent); /** * iommu_pages_start_incoherent_list - Make a list of pages incoherent * @list: The list of pages to setup * @dma_dev: The iommu device * * Perform iommu_pages_start_incoherent() across all of list. * * If this fails the caller must call iommu_pages_stop_incoherent_list(). */ int iommu_pages_start_incoherent_list(struct iommu_pages_list *list, struct device *dma_dev) { struct ioptdesc *cur; int ret; list_for_each_entry(cur, &list->pages, iopt_freelist_elm) { if (WARN_ON(cur->incoherent)) continue; ret = iommu_pages_start_incoherent( folio_address(ioptdesc_folio(cur)), dma_dev); if (ret) return ret; } return 0; } EXPORT_SYMBOL_GPL(iommu_pages_start_incoherent_list); /** * iommu_pages_stop_incoherent_list - Undo incoherence across a list * @list: The list of pages to release * @dma_dev: The iommu device * * Revert iommu_pages_start_incoherent() across all of the list. Pages that did * not call or succeed iommu_pages_start_incoherent() will be ignored. */ #if IOMMU_PAGES_USE_DMA_API void iommu_pages_stop_incoherent_list(struct iommu_pages_list *list, struct device *dma_dev) { struct ioptdesc *cur; list_for_each_entry(cur, &list->pages, iopt_freelist_elm) { struct folio *folio = ioptdesc_folio(cur); if (!cur->incoherent) continue; dma_unmap_single(dma_dev, virt_to_phys(folio_address(folio)), ioptdesc_mem_size(cur), DMA_TO_DEVICE); cur->incoherent = 0; } } EXPORT_SYMBOL_GPL(iommu_pages_stop_incoherent_list); /** * iommu_pages_free_incoherent - Free an incoherent page * @virt: virtual address of the page to be freed. * @dma_dev: The iommu device * * If the page is incoherent it made coherent again then freed. */ void iommu_pages_free_incoherent(void *virt, struct device *dma_dev) { struct ioptdesc *iopt = virt_to_ioptdesc(virt); if (iopt->incoherent) { dma_unmap_single(dma_dev, virt_to_phys(virt), ioptdesc_mem_size(iopt), DMA_TO_DEVICE); iopt->incoherent = 0; } __iommu_free_desc(iopt); } EXPORT_SYMBOL_GPL(iommu_pages_free_incoherent); #endif