|  | // SPDX-License-Identifier: GPL-2.0 | 
|  | /* | 
|  | * Copyright (C) 2002 Richard Henderson | 
|  | * Copyright (C) 2001 Rusty Russell, 2002, 2010 Rusty Russell IBM. | 
|  | * Copyright (C) 2023 Luis Chamberlain <mcgrof@kernel.org> | 
|  | * Copyright (C) 2024 Mike Rapoport IBM. | 
|  | */ | 
|  |  | 
|  | #define pr_fmt(fmt) "execmem: " fmt | 
|  |  | 
|  | #include <linux/mm.h> | 
|  | #include <linux/mutex.h> | 
|  | #include <linux/vmalloc.h> | 
|  | #include <linux/execmem.h> | 
|  | #include <linux/maple_tree.h> | 
|  | #include <linux/set_memory.h> | 
|  | #include <linux/moduleloader.h> | 
|  | #include <linux/text-patching.h> | 
|  |  | 
|  | #include <asm/tlbflush.h> | 
|  |  | 
|  | #include "internal.h" | 
|  |  | 
|  | static struct execmem_info *execmem_info __ro_after_init; | 
|  | static struct execmem_info default_execmem_info __ro_after_init; | 
|  |  | 
|  | #ifdef CONFIG_MMU | 
|  | static void *execmem_vmalloc(struct execmem_range *range, size_t size, | 
|  | pgprot_t pgprot, unsigned long vm_flags) | 
|  | { | 
|  | bool kasan = range->flags & EXECMEM_KASAN_SHADOW; | 
|  | gfp_t gfp_flags = GFP_KERNEL | __GFP_NOWARN; | 
|  | unsigned int align = range->alignment; | 
|  | unsigned long start = range->start; | 
|  | unsigned long end = range->end; | 
|  | void *p; | 
|  |  | 
|  | if (kasan) | 
|  | vm_flags |= VM_DEFER_KMEMLEAK; | 
|  |  | 
|  | if (vm_flags & VM_ALLOW_HUGE_VMAP) | 
|  | align = PMD_SIZE; | 
|  |  | 
|  | p = __vmalloc_node_range(size, align, start, end, gfp_flags, | 
|  | pgprot, vm_flags, NUMA_NO_NODE, | 
|  | __builtin_return_address(0)); | 
|  | if (!p && range->fallback_start) { | 
|  | start = range->fallback_start; | 
|  | end = range->fallback_end; | 
|  | p = __vmalloc_node_range(size, align, start, end, gfp_flags, | 
|  | pgprot, vm_flags, NUMA_NO_NODE, | 
|  | __builtin_return_address(0)); | 
|  | } | 
|  |  | 
|  | if (!p) { | 
|  | pr_warn_ratelimited("unable to allocate memory\n"); | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | if (kasan && (kasan_alloc_module_shadow(p, size, GFP_KERNEL) < 0)) { | 
|  | vfree(p); | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | return p; | 
|  | } | 
|  |  | 
|  | struct vm_struct *execmem_vmap(size_t size) | 
|  | { | 
|  | struct execmem_range *range = &execmem_info->ranges[EXECMEM_MODULE_DATA]; | 
|  | struct vm_struct *area; | 
|  |  | 
|  | area = __get_vm_area_node(size, range->alignment, PAGE_SHIFT, VM_ALLOC, | 
|  | range->start, range->end, NUMA_NO_NODE, | 
|  | GFP_KERNEL, __builtin_return_address(0)); | 
|  | if (!area && range->fallback_start) | 
|  | area = __get_vm_area_node(size, range->alignment, PAGE_SHIFT, VM_ALLOC, | 
|  | range->fallback_start, range->fallback_end, | 
|  | NUMA_NO_NODE, GFP_KERNEL, __builtin_return_address(0)); | 
|  |  | 
|  | return area; | 
|  | } | 
|  | #else | 
|  | static void *execmem_vmalloc(struct execmem_range *range, size_t size, | 
|  | pgprot_t pgprot, unsigned long vm_flags) | 
|  | { | 
|  | return vmalloc(size); | 
|  | } | 
|  | #endif /* CONFIG_MMU */ | 
|  |  | 
|  | #ifdef CONFIG_ARCH_HAS_EXECMEM_ROX | 
|  | struct execmem_cache { | 
|  | struct mutex mutex; | 
|  | struct maple_tree busy_areas; | 
|  | struct maple_tree free_areas; | 
|  | unsigned int pending_free_cnt;	/* protected by mutex */ | 
|  | }; | 
|  |  | 
|  | /* delay to schedule asynchronous free if fast path free fails */ | 
|  | #define FREE_DELAY	(msecs_to_jiffies(10)) | 
|  |  | 
|  | /* mark entries in busy_areas that should be freed asynchronously */ | 
|  | #define PENDING_FREE_MASK	(1 << (PAGE_SHIFT - 1)) | 
|  |  | 
|  | static struct execmem_cache execmem_cache = { | 
|  | .mutex = __MUTEX_INITIALIZER(execmem_cache.mutex), | 
|  | .busy_areas = MTREE_INIT_EXT(busy_areas, MT_FLAGS_LOCK_EXTERN, | 
|  | execmem_cache.mutex), | 
|  | .free_areas = MTREE_INIT_EXT(free_areas, MT_FLAGS_LOCK_EXTERN, | 
|  | execmem_cache.mutex), | 
|  | }; | 
|  |  | 
|  | static inline unsigned long mas_range_len(struct ma_state *mas) | 
|  | { | 
|  | return mas->last - mas->index + 1; | 
|  | } | 
|  |  | 
|  | static int execmem_set_direct_map_valid(struct vm_struct *vm, bool valid) | 
|  | { | 
|  | unsigned int nr = (1 << get_vm_area_page_order(vm)); | 
|  | unsigned int updated = 0; | 
|  | int err = 0; | 
|  |  | 
|  | for (int i = 0; i < vm->nr_pages; i += nr) { | 
|  | err = set_direct_map_valid_noflush(vm->pages[i], nr, valid); | 
|  | if (err) | 
|  | goto err_restore; | 
|  | updated += nr; | 
|  | } | 
|  |  | 
|  | return 0; | 
|  |  | 
|  | err_restore: | 
|  | for (int i = 0; i < updated; i += nr) | 
|  | set_direct_map_valid_noflush(vm->pages[i], nr, !valid); | 
|  |  | 
|  | return err; | 
|  | } | 
|  |  | 
|  | static int execmem_force_rw(void *ptr, size_t size) | 
|  | { | 
|  | unsigned int nr = PAGE_ALIGN(size) >> PAGE_SHIFT; | 
|  | unsigned long addr = (unsigned long)ptr; | 
|  | int ret; | 
|  |  | 
|  | ret = set_memory_nx(addr, nr); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | return set_memory_rw(addr, nr); | 
|  | } | 
|  |  | 
|  | int execmem_restore_rox(void *ptr, size_t size) | 
|  | { | 
|  | unsigned int nr = PAGE_ALIGN(size) >> PAGE_SHIFT; | 
|  | unsigned long addr = (unsigned long)ptr; | 
|  |  | 
|  | return set_memory_rox(addr, nr); | 
|  | } | 
|  |  | 
|  | static void execmem_cache_clean(struct work_struct *work) | 
|  | { | 
|  | struct maple_tree *free_areas = &execmem_cache.free_areas; | 
|  | struct mutex *mutex = &execmem_cache.mutex; | 
|  | MA_STATE(mas, free_areas, 0, ULONG_MAX); | 
|  | void *area; | 
|  |  | 
|  | mutex_lock(mutex); | 
|  | mas_for_each(&mas, area, ULONG_MAX) { | 
|  | size_t size = mas_range_len(&mas); | 
|  |  | 
|  | if (IS_ALIGNED(size, PMD_SIZE) && | 
|  | IS_ALIGNED(mas.index, PMD_SIZE)) { | 
|  | struct vm_struct *vm = find_vm_area(area); | 
|  |  | 
|  | execmem_set_direct_map_valid(vm, true); | 
|  | mas_store_gfp(&mas, NULL, GFP_KERNEL); | 
|  | vfree(area); | 
|  | } | 
|  | } | 
|  | mutex_unlock(mutex); | 
|  | } | 
|  |  | 
|  | static DECLARE_WORK(execmem_cache_clean_work, execmem_cache_clean); | 
|  |  | 
|  | static int execmem_cache_add_locked(void *ptr, size_t size, gfp_t gfp_mask) | 
|  | { | 
|  | struct maple_tree *free_areas = &execmem_cache.free_areas; | 
|  | unsigned long addr = (unsigned long)ptr; | 
|  | MA_STATE(mas, free_areas, addr - 1, addr + 1); | 
|  | unsigned long lower, upper; | 
|  | void *area = NULL; | 
|  |  | 
|  | lower = addr; | 
|  | upper = addr + size - 1; | 
|  |  | 
|  | area = mas_walk(&mas); | 
|  | if (area && mas.last == addr - 1) | 
|  | lower = mas.index; | 
|  |  | 
|  | area = mas_next(&mas, ULONG_MAX); | 
|  | if (area && mas.index == addr + size) | 
|  | upper = mas.last; | 
|  |  | 
|  | mas_set_range(&mas, lower, upper); | 
|  | return mas_store_gfp(&mas, (void *)lower, gfp_mask); | 
|  | } | 
|  |  | 
|  | static int execmem_cache_add(void *ptr, size_t size, gfp_t gfp_mask) | 
|  | { | 
|  | guard(mutex)(&execmem_cache.mutex); | 
|  |  | 
|  | return execmem_cache_add_locked(ptr, size, gfp_mask); | 
|  | } | 
|  |  | 
|  | static bool within_range(struct execmem_range *range, struct ma_state *mas, | 
|  | size_t size) | 
|  | { | 
|  | unsigned long addr = mas->index; | 
|  |  | 
|  | if (addr >= range->start && addr + size < range->end) | 
|  | return true; | 
|  |  | 
|  | if (range->fallback_start && | 
|  | addr >= range->fallback_start && addr + size < range->fallback_end) | 
|  | return true; | 
|  |  | 
|  | return false; | 
|  | } | 
|  |  | 
|  | static void *__execmem_cache_alloc(struct execmem_range *range, size_t size) | 
|  | { | 
|  | struct maple_tree *free_areas = &execmem_cache.free_areas; | 
|  | struct maple_tree *busy_areas = &execmem_cache.busy_areas; | 
|  | MA_STATE(mas_free, free_areas, 0, ULONG_MAX); | 
|  | MA_STATE(mas_busy, busy_areas, 0, ULONG_MAX); | 
|  | struct mutex *mutex = &execmem_cache.mutex; | 
|  | unsigned long addr, last, area_size = 0; | 
|  | void *area, *ptr = NULL; | 
|  | int err; | 
|  |  | 
|  | mutex_lock(mutex); | 
|  | mas_for_each(&mas_free, area, ULONG_MAX) { | 
|  | area_size = mas_range_len(&mas_free); | 
|  |  | 
|  | if (area_size >= size && within_range(range, &mas_free, size)) | 
|  | break; | 
|  | } | 
|  |  | 
|  | if (area_size < size) | 
|  | goto out_unlock; | 
|  |  | 
|  | addr = mas_free.index; | 
|  | last = mas_free.last; | 
|  |  | 
|  | /* insert allocated size to busy_areas at range [addr, addr + size) */ | 
|  | mas_set_range(&mas_busy, addr, addr + size - 1); | 
|  | err = mas_store_gfp(&mas_busy, (void *)addr, GFP_KERNEL); | 
|  | if (err) | 
|  | goto out_unlock; | 
|  |  | 
|  | mas_store_gfp(&mas_free, NULL, GFP_KERNEL); | 
|  | if (area_size > size) { | 
|  | void *ptr = (void *)(addr + size); | 
|  |  | 
|  | /* | 
|  | * re-insert remaining free size to free_areas at range | 
|  | * [addr + size, last] | 
|  | */ | 
|  | mas_set_range(&mas_free, addr + size, last); | 
|  | err = mas_store_gfp(&mas_free, ptr, GFP_KERNEL); | 
|  | if (err) { | 
|  | mas_store_gfp(&mas_busy, NULL, GFP_KERNEL); | 
|  | goto out_unlock; | 
|  | } | 
|  | } | 
|  | ptr = (void *)addr; | 
|  |  | 
|  | out_unlock: | 
|  | mutex_unlock(mutex); | 
|  | return ptr; | 
|  | } | 
|  |  | 
|  | static int execmem_cache_populate(struct execmem_range *range, size_t size) | 
|  | { | 
|  | unsigned long vm_flags = VM_ALLOW_HUGE_VMAP; | 
|  | struct vm_struct *vm; | 
|  | size_t alloc_size; | 
|  | int err = -ENOMEM; | 
|  | void *p; | 
|  |  | 
|  | alloc_size = round_up(size, PMD_SIZE); | 
|  | p = execmem_vmalloc(range, alloc_size, PAGE_KERNEL, vm_flags); | 
|  | if (!p) { | 
|  | alloc_size = size; | 
|  | p = execmem_vmalloc(range, alloc_size, PAGE_KERNEL, vm_flags); | 
|  | } | 
|  |  | 
|  | if (!p) | 
|  | return err; | 
|  |  | 
|  | vm = find_vm_area(p); | 
|  | if (!vm) | 
|  | goto err_free_mem; | 
|  |  | 
|  | /* fill memory with instructions that will trap */ | 
|  | execmem_fill_trapping_insns(p, alloc_size); | 
|  |  | 
|  | err = set_memory_rox((unsigned long)p, vm->nr_pages); | 
|  | if (err) | 
|  | goto err_free_mem; | 
|  |  | 
|  | err = execmem_cache_add(p, alloc_size, GFP_KERNEL); | 
|  | if (err) | 
|  | goto err_reset_direct_map; | 
|  |  | 
|  | return 0; | 
|  |  | 
|  | err_reset_direct_map: | 
|  | execmem_set_direct_map_valid(vm, true); | 
|  | err_free_mem: | 
|  | vfree(p); | 
|  | return err; | 
|  | } | 
|  |  | 
|  | static void *execmem_cache_alloc(struct execmem_range *range, size_t size) | 
|  | { | 
|  | void *p; | 
|  | int err; | 
|  |  | 
|  | p = __execmem_cache_alloc(range, size); | 
|  | if (p) | 
|  | return p; | 
|  |  | 
|  | err = execmem_cache_populate(range, size); | 
|  | if (err) | 
|  | return NULL; | 
|  |  | 
|  | return __execmem_cache_alloc(range, size); | 
|  | } | 
|  |  | 
|  | static inline bool is_pending_free(void *ptr) | 
|  | { | 
|  | return ((unsigned long)ptr & PENDING_FREE_MASK); | 
|  | } | 
|  |  | 
|  | static inline void *pending_free_set(void *ptr) | 
|  | { | 
|  | return (void *)((unsigned long)ptr | PENDING_FREE_MASK); | 
|  | } | 
|  |  | 
|  | static inline void *pending_free_clear(void *ptr) | 
|  | { | 
|  | return (void *)((unsigned long)ptr & ~PENDING_FREE_MASK); | 
|  | } | 
|  |  | 
|  | static int __execmem_cache_free(struct ma_state *mas, void *ptr, gfp_t gfp_mask) | 
|  | { | 
|  | size_t size = mas_range_len(mas); | 
|  | int err; | 
|  |  | 
|  | err = execmem_force_rw(ptr, size); | 
|  | if (err) | 
|  | return err; | 
|  |  | 
|  | execmem_fill_trapping_insns(ptr, size); | 
|  | execmem_restore_rox(ptr, size); | 
|  |  | 
|  | err = execmem_cache_add_locked(ptr, size, gfp_mask); | 
|  | if (err) | 
|  | return err; | 
|  |  | 
|  | mas_store_gfp(mas, NULL, gfp_mask); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static void execmem_cache_free_slow(struct work_struct *work); | 
|  | static DECLARE_DELAYED_WORK(execmem_cache_free_work, execmem_cache_free_slow); | 
|  |  | 
|  | static void execmem_cache_free_slow(struct work_struct *work) | 
|  | { | 
|  | struct maple_tree *busy_areas = &execmem_cache.busy_areas; | 
|  | MA_STATE(mas, busy_areas, 0, ULONG_MAX); | 
|  | void *area; | 
|  |  | 
|  | guard(mutex)(&execmem_cache.mutex); | 
|  |  | 
|  | if (!execmem_cache.pending_free_cnt) | 
|  | return; | 
|  |  | 
|  | mas_for_each(&mas, area, ULONG_MAX) { | 
|  | if (!is_pending_free(area)) | 
|  | continue; | 
|  |  | 
|  | area = pending_free_clear(area); | 
|  | if (__execmem_cache_free(&mas, area, GFP_KERNEL)) | 
|  | continue; | 
|  |  | 
|  | execmem_cache.pending_free_cnt--; | 
|  | } | 
|  |  | 
|  | if (execmem_cache.pending_free_cnt) | 
|  | schedule_delayed_work(&execmem_cache_free_work, FREE_DELAY); | 
|  | else | 
|  | schedule_work(&execmem_cache_clean_work); | 
|  | } | 
|  |  | 
|  | static bool execmem_cache_free(void *ptr) | 
|  | { | 
|  | struct maple_tree *busy_areas = &execmem_cache.busy_areas; | 
|  | unsigned long addr = (unsigned long)ptr; | 
|  | MA_STATE(mas, busy_areas, addr, addr); | 
|  | void *area; | 
|  | int err; | 
|  |  | 
|  | guard(mutex)(&execmem_cache.mutex); | 
|  |  | 
|  | area = mas_walk(&mas); | 
|  | if (!area) | 
|  | return false; | 
|  |  | 
|  | err = __execmem_cache_free(&mas, area, GFP_KERNEL | __GFP_NORETRY); | 
|  | if (err) { | 
|  | /* | 
|  | * mas points to exact slot we've got the area from, nothing | 
|  | * else can modify the tree because of the mutex, so there | 
|  | * won't be any allocations in mas_store_gfp() and it will just | 
|  | * change the pointer. | 
|  | */ | 
|  | area = pending_free_set(area); | 
|  | mas_store_gfp(&mas, area, GFP_KERNEL); | 
|  | execmem_cache.pending_free_cnt++; | 
|  | schedule_delayed_work(&execmem_cache_free_work, FREE_DELAY); | 
|  | return true; | 
|  | } | 
|  |  | 
|  | schedule_work(&execmem_cache_clean_work); | 
|  |  | 
|  | return true; | 
|  | } | 
|  |  | 
|  | #else /* CONFIG_ARCH_HAS_EXECMEM_ROX */ | 
|  | /* | 
|  | * when ROX cache is not used the permissions defined by architectures for | 
|  | * execmem ranges that are updated before use (e.g. EXECMEM_MODULE_TEXT) must | 
|  | * be writable anyway | 
|  | */ | 
|  | static inline int execmem_force_rw(void *ptr, size_t size) | 
|  | { | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static void *execmem_cache_alloc(struct execmem_range *range, size_t size) | 
|  | { | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | static bool execmem_cache_free(void *ptr) | 
|  | { | 
|  | return false; | 
|  | } | 
|  | #endif /* CONFIG_ARCH_HAS_EXECMEM_ROX */ | 
|  |  | 
|  | void *execmem_alloc(enum execmem_type type, size_t size) | 
|  | { | 
|  | struct execmem_range *range = &execmem_info->ranges[type]; | 
|  | bool use_cache = range->flags & EXECMEM_ROX_CACHE; | 
|  | unsigned long vm_flags = VM_FLUSH_RESET_PERMS; | 
|  | pgprot_t pgprot = range->pgprot; | 
|  | void *p = NULL; | 
|  |  | 
|  | size = PAGE_ALIGN(size); | 
|  |  | 
|  | if (use_cache) | 
|  | p = execmem_cache_alloc(range, size); | 
|  | else | 
|  | p = execmem_vmalloc(range, size, pgprot, vm_flags); | 
|  |  | 
|  | return kasan_reset_tag(p); | 
|  | } | 
|  |  | 
|  | void *execmem_alloc_rw(enum execmem_type type, size_t size) | 
|  | { | 
|  | void *p __free(execmem) = execmem_alloc(type, size); | 
|  | int err; | 
|  |  | 
|  | if (!p) | 
|  | return NULL; | 
|  |  | 
|  | err = execmem_force_rw(p, size); | 
|  | if (err) | 
|  | return NULL; | 
|  |  | 
|  | return no_free_ptr(p); | 
|  | } | 
|  |  | 
|  | void execmem_free(void *ptr) | 
|  | { | 
|  | /* | 
|  | * This memory may be RO, and freeing RO memory in an interrupt is not | 
|  | * supported by vmalloc. | 
|  | */ | 
|  | WARN_ON(in_interrupt()); | 
|  |  | 
|  | if (!execmem_cache_free(ptr)) | 
|  | vfree(ptr); | 
|  | } | 
|  |  | 
|  | bool execmem_is_rox(enum execmem_type type) | 
|  | { | 
|  | return !!(execmem_info->ranges[type].flags & EXECMEM_ROX_CACHE); | 
|  | } | 
|  |  | 
|  | static bool execmem_validate(struct execmem_info *info) | 
|  | { | 
|  | struct execmem_range *r = &info->ranges[EXECMEM_DEFAULT]; | 
|  |  | 
|  | if (!r->alignment || !r->start || !r->end || !pgprot_val(r->pgprot)) { | 
|  | pr_crit("Invalid parameters for execmem allocator, module loading will fail"); | 
|  | return false; | 
|  | } | 
|  |  | 
|  | if (!IS_ENABLED(CONFIG_ARCH_HAS_EXECMEM_ROX)) { | 
|  | for (int i = EXECMEM_DEFAULT; i < EXECMEM_TYPE_MAX; i++) { | 
|  | r = &info->ranges[i]; | 
|  |  | 
|  | if (r->flags & EXECMEM_ROX_CACHE) { | 
|  | pr_warn_once("ROX cache is not supported\n"); | 
|  | r->flags &= ~EXECMEM_ROX_CACHE; | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | return true; | 
|  | } | 
|  |  | 
|  | static void execmem_init_missing(struct execmem_info *info) | 
|  | { | 
|  | struct execmem_range *default_range = &info->ranges[EXECMEM_DEFAULT]; | 
|  |  | 
|  | for (int i = EXECMEM_DEFAULT + 1; i < EXECMEM_TYPE_MAX; i++) { | 
|  | struct execmem_range *r = &info->ranges[i]; | 
|  |  | 
|  | if (!r->start) { | 
|  | if (i == EXECMEM_MODULE_DATA) | 
|  | r->pgprot = PAGE_KERNEL; | 
|  | else | 
|  | r->pgprot = default_range->pgprot; | 
|  | r->alignment = default_range->alignment; | 
|  | r->start = default_range->start; | 
|  | r->end = default_range->end; | 
|  | r->flags = default_range->flags; | 
|  | r->fallback_start = default_range->fallback_start; | 
|  | r->fallback_end = default_range->fallback_end; | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | struct execmem_info * __weak execmem_arch_setup(void) | 
|  | { | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | static void __init __execmem_init(void) | 
|  | { | 
|  | struct execmem_info *info = execmem_arch_setup(); | 
|  |  | 
|  | if (!info) { | 
|  | info = execmem_info = &default_execmem_info; | 
|  | info->ranges[EXECMEM_DEFAULT].start = VMALLOC_START; | 
|  | info->ranges[EXECMEM_DEFAULT].end = VMALLOC_END; | 
|  | info->ranges[EXECMEM_DEFAULT].pgprot = PAGE_KERNEL_EXEC; | 
|  | info->ranges[EXECMEM_DEFAULT].alignment = 1; | 
|  | } | 
|  |  | 
|  | if (!execmem_validate(info)) | 
|  | return; | 
|  |  | 
|  | execmem_init_missing(info); | 
|  |  | 
|  | execmem_info = info; | 
|  | } | 
|  |  | 
|  | #ifdef CONFIG_ARCH_WANTS_EXECMEM_LATE | 
|  | static int __init execmem_late_init(void) | 
|  | { | 
|  | __execmem_init(); | 
|  | return 0; | 
|  | } | 
|  | core_initcall(execmem_late_init); | 
|  | #else | 
|  | void __init execmem_init(void) | 
|  | { | 
|  | __execmem_init(); | 
|  | } | 
|  | #endif |