treewide: Rename "unencrypted" to "decrypted" Back then when the whole SME machinery started getting mainlined, it was agreed that for simplicity, clarity and sanity's sake, the terms denoting encrypted and not-encrypted memory should be "encrypted" and "decrypted". And the majority of the code sticks to that convention except those two. So rename them. The intent of "encrypted" and "decrypted" is to represent the binary concept of memory which is encrypted and of memory which is not. The further differentiation between decrypted and unencrypted memory is not needed in the code (for now) and therefore keep things simple by using solely the two terms. No functional changes. [ Convert forgotten s390 and ppc function variants. ] Reported-by: kbuild test robot <lkp@intel.com> Signed-off-by: Borislav Petkov <bp@suse.de>
diff --git a/arch/powerpc/include/asm/mem_encrypt.h b/arch/powerpc/include/asm/mem_encrypt.h index ba9dab0..f0705cd 100644 --- a/arch/powerpc/include/asm/mem_encrypt.h +++ b/arch/powerpc/include/asm/mem_encrypt.h
@@ -15,7 +15,7 @@ static inline bool mem_encrypt_active(void) return is_secure_guest(); } -static inline bool force_dma_unencrypted(struct device *dev) +static inline bool force_dma_decrypted(struct device *dev) { return is_secure_guest(); }
diff --git a/arch/powerpc/platforms/pseries/Kconfig b/arch/powerpc/platforms/pseries/Kconfig index 24c1836..a78e2c3 100644 --- a/arch/powerpc/platforms/pseries/Kconfig +++ b/arch/powerpc/platforms/pseries/Kconfig
@@ -151,7 +151,7 @@ depends on PPC_PSERIES select SWIOTLB select ARCH_HAS_MEM_ENCRYPT - select ARCH_HAS_FORCE_DMA_UNENCRYPTED + select ARCH_HAS_FORCE_DMA_DECRYPTED help There are certain POWER platforms which support secure guests using the Protected Execution Facility, with the help of an Ultravisor
diff --git a/arch/s390/Kconfig b/arch/s390/Kconfig index 8abe775..ab1dbb7 100644 --- a/arch/s390/Kconfig +++ b/arch/s390/Kconfig
@@ -192,7 +192,7 @@ select VIRT_CPU_ACCOUNTING select ARCH_HAS_SCALED_CPUTIME select HAVE_NMI - select ARCH_HAS_FORCE_DMA_UNENCRYPTED + select ARCH_HAS_FORCE_DMA_DECRYPTED select SWIOTLB select GENERIC_ALLOCATOR
diff --git a/arch/s390/mm/init.c b/arch/s390/mm/init.c index ac44bd7..5fed85f 100644 --- a/arch/s390/mm/init.c +++ b/arch/s390/mm/init.c
@@ -157,7 +157,7 @@ int set_memory_decrypted(unsigned long addr, int numpages) } /* are we a protected virtualization guest? */ -bool force_dma_unencrypted(struct device *dev) +bool force_dma_decrypted(struct device *dev) { return is_prot_virt_guest(); }
diff --git a/arch/x86/Kconfig b/arch/x86/Kconfig index beea770..2ae904f 100644 --- a/arch/x86/Kconfig +++ b/arch/x86/Kconfig
@@ -1525,7 +1525,7 @@ depends on X86_64 && CPU_SUP_AMD select DYNAMIC_PHYSICAL_MASK select ARCH_USE_MEMREMAP_PROT - select ARCH_HAS_FORCE_DMA_UNENCRYPTED + select ARCH_HAS_FORCE_DMA_DECRYPTED ---help--- Say yes to enable support for the encryption of system memory. This requires an AMD processor that supports Secure Memory
diff --git a/arch/x86/mm/mem_encrypt.c b/arch/x86/mm/mem_encrypt.c index a03614b..66d09f2 100644 --- a/arch/x86/mm/mem_encrypt.c +++ b/arch/x86/mm/mem_encrypt.c
@@ -350,8 +350,8 @@ bool sev_active(void) return sme_me_mask && sev_enabled; } -/* Override for DMA direct allocation check - ARCH_HAS_FORCE_DMA_UNENCRYPTED */ -bool force_dma_unencrypted(struct device *dev) +/* Override for DMA direct allocation check - ARCH_HAS_FORCE_DMA_DECRYPTED */ +bool force_dma_decrypted(struct device *dev) { /* * For SEV, all DMA must be to unencrypted addresses.
diff --git a/include/linux/dma-direct.h b/include/linux/dma-direct.h index 24b8684..9f95584 100644 --- a/include/linux/dma-direct.h +++ b/include/linux/dma-direct.h
@@ -26,14 +26,14 @@ static inline phys_addr_t __dma_to_phys(struct device *dev, dma_addr_t dev_addr) } #endif /* !CONFIG_ARCH_HAS_PHYS_TO_DMA */ -#ifdef CONFIG_ARCH_HAS_FORCE_DMA_UNENCRYPTED -bool force_dma_unencrypted(struct device *dev); +#ifdef CONFIG_ARCH_HAS_FORCE_DMA_DECRYPTED +bool force_dma_decrypted(struct device *dev); #else -static inline bool force_dma_unencrypted(struct device *dev) +static inline bool force_dma_decrypted(struct device *dev) { return false; } -#endif /* CONFIG_ARCH_HAS_FORCE_DMA_UNENCRYPTED */ +#endif /* CONFIG_ARCH_HAS_FORCE_DMA_DECRYPTED */ /* * If memory encryption is supported, phys_to_dma will set the memory encryption
diff --git a/kernel/dma/Kconfig b/kernel/dma/Kconfig index 4c103a2..55c4147 100644 --- a/kernel/dma/Kconfig +++ b/kernel/dma/Kconfig
@@ -51,7 +51,7 @@ config ARCH_HAS_DMA_PREP_COHERENT bool -config ARCH_HAS_FORCE_DMA_UNENCRYPTED +config ARCH_HAS_FORCE_DMA_DECRYPTED bool config DMA_NONCOHERENT_CACHE_SYNC
diff --git a/kernel/dma/direct.c b/kernel/dma/direct.c index ac7956c..a0576c0 100644 --- a/kernel/dma/direct.c +++ b/kernel/dma/direct.c
@@ -26,7 +26,7 @@ unsigned int zone_dma_bits __ro_after_init = 24; static inline dma_addr_t phys_to_dma_direct(struct device *dev, phys_addr_t phys) { - if (force_dma_unencrypted(dev)) + if (force_dma_decrypted(dev)) return __phys_to_dma(dev, phys); return phys_to_dma(dev, phys); } @@ -49,7 +49,7 @@ static gfp_t __dma_direct_optimal_gfp_mask(struct device *dev, u64 dma_mask, { u64 dma_limit = min_not_zero(dma_mask, dev->bus_dma_limit); - if (force_dma_unencrypted(dev)) + if (force_dma_decrypted(dev)) *phys_limit = __dma_to_phys(dev, dma_limit); else *phys_limit = dma_to_phys(dev, dma_limit); @@ -138,7 +138,7 @@ void *dma_direct_alloc_pages(struct device *dev, size_t size, return NULL; if ((attrs & DMA_ATTR_NO_KERNEL_MAPPING) && - !force_dma_unencrypted(dev)) { + !force_dma_decrypted(dev)) { /* remove any dirty cache lines on the kernel alias */ if (!PageHighMem(page)) arch_dma_prep_coherent(page, size); @@ -179,7 +179,7 @@ void *dma_direct_alloc_pages(struct device *dev, size_t size, } ret = page_address(page); - if (force_dma_unencrypted(dev)) + if (force_dma_decrypted(dev)) set_memory_decrypted((unsigned long)ret, 1 << get_order(size)); memset(ret, 0, size); @@ -190,7 +190,7 @@ void *dma_direct_alloc_pages(struct device *dev, size_t size, ret = uncached_kernel_address(ret); } done: - if (force_dma_unencrypted(dev)) + if (force_dma_decrypted(dev)) *dma_handle = __phys_to_dma(dev, page_to_phys(page)); else *dma_handle = phys_to_dma(dev, page_to_phys(page)); @@ -203,7 +203,7 @@ void dma_direct_free_pages(struct device *dev, size_t size, void *cpu_addr, unsigned int page_order = get_order(size); if ((attrs & DMA_ATTR_NO_KERNEL_MAPPING) && - !force_dma_unencrypted(dev)) { + !force_dma_decrypted(dev)) { /* cpu_addr is a struct page cookie, not a kernel address */ dma_free_contiguous(dev, cpu_addr, size); return; @@ -213,7 +213,7 @@ void dma_direct_free_pages(struct device *dev, size_t size, void *cpu_addr, dma_free_from_pool(cpu_addr, PAGE_ALIGN(size))) return; - if (force_dma_unencrypted(dev)) + if (force_dma_decrypted(dev)) set_memory_encrypted((unsigned long)cpu_addr, 1 << page_order); if (IS_ENABLED(CONFIG_DMA_REMAP) && is_vmalloc_addr(cpu_addr))
diff --git a/kernel/dma/mapping.c b/kernel/dma/mapping.c index 98e3d87..dbd0605 100644 --- a/kernel/dma/mapping.c +++ b/kernel/dma/mapping.c
@@ -154,7 +154,7 @@ EXPORT_SYMBOL(dma_get_sgtable_attrs); */ pgprot_t dma_pgprot(struct device *dev, pgprot_t prot, unsigned long attrs) { - if (force_dma_unencrypted(dev)) + if (force_dma_decrypted(dev)) prot = pgprot_decrypted(prot); if (dev_is_dma_coherent(dev) || (IS_ENABLED(CONFIG_DMA_NONCOHERENT_CACHE_SYNC) &&