|  | // SPDX-License-Identifier: GPL-2.0-or-later | 
|  | /* | 
|  | * Copyright (C) 2021-2023 Oracle.  All Rights Reserved. | 
|  | * Author: Darrick J. Wong <djwong@kernel.org> | 
|  | */ | 
|  | #include "xfs.h" | 
|  | #include "xfs_fs.h" | 
|  | #include "xfs_shared.h" | 
|  | #include "xfs_format.h" | 
|  | #include "scrub/scrub.h" | 
|  | #include "scrub/xfile.h" | 
|  | #include "scrub/xfarray.h" | 
|  | #include "scrub/trace.h" | 
|  |  | 
|  | /* | 
|  | * Large Arrays of Fixed-Size Records | 
|  | * ================================== | 
|  | * | 
|  | * This memory array uses an xfile (which itself is a shmem file) to store | 
|  | * large numbers of fixed-size records in memory that can be paged out.  This | 
|  | * puts less stress on the memory reclaim algorithms during an online repair | 
|  | * because we don't have to pin so much memory.  However, array access is less | 
|  | * direct than would be in a regular memory array.  Access to the array is | 
|  | * performed via indexed load and store methods, and an append method is | 
|  | * provided for convenience.  Array elements can be unset, which sets them to | 
|  | * all zeroes.  Unset entries are skipped during iteration, though direct loads | 
|  | * will return a zeroed buffer.  Callers are responsible for concurrency | 
|  | * control. | 
|  | */ | 
|  |  | 
|  | /* | 
|  | * Pointer to scratch space.  Because we can't access the xfile data directly, | 
|  | * we allocate a small amount of memory on the end of the xfarray structure to | 
|  | * buffer array items when we need space to store values temporarily. | 
|  | */ | 
|  | static inline void *xfarray_scratch(struct xfarray *array) | 
|  | { | 
|  | return (array + 1); | 
|  | } | 
|  |  | 
|  | /* Compute array index given an xfile offset. */ | 
|  | static xfarray_idx_t | 
|  | xfarray_idx( | 
|  | struct xfarray	*array, | 
|  | loff_t		pos) | 
|  | { | 
|  | if (array->obj_size_log >= 0) | 
|  | return (xfarray_idx_t)pos >> array->obj_size_log; | 
|  |  | 
|  | return div_u64((xfarray_idx_t)pos, array->obj_size); | 
|  | } | 
|  |  | 
|  | /* Compute xfile offset of array element. */ | 
|  | static inline loff_t xfarray_pos(struct xfarray *array, xfarray_idx_t idx) | 
|  | { | 
|  | if (array->obj_size_log >= 0) | 
|  | return idx << array->obj_size_log; | 
|  |  | 
|  | return idx * array->obj_size; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Initialize a big memory array.  Array records cannot be larger than a | 
|  | * page, and the array cannot span more bytes than the page cache supports. | 
|  | * If @required_capacity is nonzero, the maximum array size will be set to this | 
|  | * quantity and the array creation will fail if the underlying storage cannot | 
|  | * support that many records. | 
|  | */ | 
|  | int | 
|  | xfarray_create( | 
|  | const char		*description, | 
|  | unsigned long long	required_capacity, | 
|  | size_t			obj_size, | 
|  | struct xfarray		**arrayp) | 
|  | { | 
|  | struct xfarray		*array; | 
|  | struct xfile		*xfile; | 
|  | int			error; | 
|  |  | 
|  | ASSERT(obj_size < PAGE_SIZE); | 
|  |  | 
|  | error = xfile_create(description, 0, &xfile); | 
|  | if (error) | 
|  | return error; | 
|  |  | 
|  | error = -ENOMEM; | 
|  | array = kzalloc(sizeof(struct xfarray) + obj_size, XCHK_GFP_FLAGS); | 
|  | if (!array) | 
|  | goto out_xfile; | 
|  |  | 
|  | array->xfile = xfile; | 
|  | array->obj_size = obj_size; | 
|  |  | 
|  | if (is_power_of_2(obj_size)) | 
|  | array->obj_size_log = ilog2(obj_size); | 
|  | else | 
|  | array->obj_size_log = -1; | 
|  |  | 
|  | array->max_nr = xfarray_idx(array, MAX_LFS_FILESIZE); | 
|  | trace_xfarray_create(array, required_capacity); | 
|  |  | 
|  | if (required_capacity > 0) { | 
|  | if (array->max_nr < required_capacity) { | 
|  | error = -ENOMEM; | 
|  | goto out_xfarray; | 
|  | } | 
|  | array->max_nr = required_capacity; | 
|  | } | 
|  |  | 
|  | *arrayp = array; | 
|  | return 0; | 
|  |  | 
|  | out_xfarray: | 
|  | kfree(array); | 
|  | out_xfile: | 
|  | xfile_destroy(xfile); | 
|  | return error; | 
|  | } | 
|  |  | 
|  | /* Destroy the array. */ | 
|  | void | 
|  | xfarray_destroy( | 
|  | struct xfarray	*array) | 
|  | { | 
|  | xfile_destroy(array->xfile); | 
|  | kfree(array); | 
|  | } | 
|  |  | 
|  | /* Load an element from the array. */ | 
|  | int | 
|  | xfarray_load( | 
|  | struct xfarray	*array, | 
|  | xfarray_idx_t	idx, | 
|  | void		*ptr) | 
|  | { | 
|  | if (idx >= array->nr) | 
|  | return -ENODATA; | 
|  |  | 
|  | return xfile_load(array->xfile, ptr, array->obj_size, | 
|  | xfarray_pos(array, idx)); | 
|  | } | 
|  |  | 
|  | /* Is this array element potentially unset? */ | 
|  | static inline bool | 
|  | xfarray_is_unset( | 
|  | struct xfarray	*array, | 
|  | loff_t		pos) | 
|  | { | 
|  | void		*temp = xfarray_scratch(array); | 
|  | int		error; | 
|  |  | 
|  | if (array->unset_slots == 0) | 
|  | return false; | 
|  |  | 
|  | error = xfile_load(array->xfile, temp, array->obj_size, pos); | 
|  | if (!error && xfarray_element_is_null(array, temp)) | 
|  | return true; | 
|  |  | 
|  | return false; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Unset an array element.  If @idx is the last element in the array, the | 
|  | * array will be truncated.  Otherwise, the entry will be zeroed. | 
|  | */ | 
|  | int | 
|  | xfarray_unset( | 
|  | struct xfarray	*array, | 
|  | xfarray_idx_t	idx) | 
|  | { | 
|  | void		*temp = xfarray_scratch(array); | 
|  | loff_t		pos = xfarray_pos(array, idx); | 
|  | int		error; | 
|  |  | 
|  | if (idx >= array->nr) | 
|  | return -ENODATA; | 
|  |  | 
|  | if (idx == array->nr - 1) { | 
|  | array->nr--; | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | if (xfarray_is_unset(array, pos)) | 
|  | return 0; | 
|  |  | 
|  | memset(temp, 0, array->obj_size); | 
|  | error = xfile_store(array->xfile, temp, array->obj_size, pos); | 
|  | if (error) | 
|  | return error; | 
|  |  | 
|  | array->unset_slots++; | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Store an element in the array.  The element must not be completely zeroed, | 
|  | * because those are considered unset sparse elements. | 
|  | */ | 
|  | int | 
|  | xfarray_store( | 
|  | struct xfarray	*array, | 
|  | xfarray_idx_t	idx, | 
|  | const void	*ptr) | 
|  | { | 
|  | int		ret; | 
|  |  | 
|  | if (idx >= array->max_nr) | 
|  | return -EFBIG; | 
|  |  | 
|  | ASSERT(!xfarray_element_is_null(array, ptr)); | 
|  |  | 
|  | ret = xfile_store(array->xfile, ptr, array->obj_size, | 
|  | xfarray_pos(array, idx)); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | array->nr = max(array->nr, idx + 1); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* Is this array element NULL? */ | 
|  | bool | 
|  | xfarray_element_is_null( | 
|  | struct xfarray	*array, | 
|  | const void	*ptr) | 
|  | { | 
|  | return !memchr_inv(ptr, 0, array->obj_size); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Store an element anywhere in the array that is unset.  If there are no | 
|  | * unset slots, append the element to the array. | 
|  | */ | 
|  | int | 
|  | xfarray_store_anywhere( | 
|  | struct xfarray	*array, | 
|  | const void	*ptr) | 
|  | { | 
|  | void		*temp = xfarray_scratch(array); | 
|  | loff_t		endpos = xfarray_pos(array, array->nr); | 
|  | loff_t		pos; | 
|  | int		error; | 
|  |  | 
|  | /* Find an unset slot to put it in. */ | 
|  | for (pos = 0; | 
|  | pos < endpos && array->unset_slots > 0; | 
|  | pos += array->obj_size) { | 
|  | error = xfile_load(array->xfile, temp, array->obj_size, | 
|  | pos); | 
|  | if (error || !xfarray_element_is_null(array, temp)) | 
|  | continue; | 
|  |  | 
|  | error = xfile_store(array->xfile, ptr, array->obj_size, | 
|  | pos); | 
|  | if (error) | 
|  | return error; | 
|  |  | 
|  | array->unset_slots--; | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* No unset slots found; attach it on the end. */ | 
|  | array->unset_slots = 0; | 
|  | return xfarray_append(array, ptr); | 
|  | } | 
|  |  | 
|  | /* Return length of array. */ | 
|  | uint64_t | 
|  | xfarray_length( | 
|  | struct xfarray	*array) | 
|  | { | 
|  | return array->nr; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Decide which array item we're going to read as part of an _iter_get. | 
|  | * @cur is the array index, and @pos is the file offset of that array index in | 
|  | * the backing xfile.  Returns ENODATA if we reach the end of the records. | 
|  | * | 
|  | * Reading from a hole in a sparse xfile causes page instantiation, so for | 
|  | * iterating a (possibly sparse) array we need to figure out if the cursor is | 
|  | * pointing at a totally uninitialized hole and move the cursor up if | 
|  | * necessary. | 
|  | */ | 
|  | static inline int | 
|  | xfarray_find_data( | 
|  | struct xfarray	*array, | 
|  | xfarray_idx_t	*cur, | 
|  | loff_t		*pos) | 
|  | { | 
|  | unsigned int	pgoff = offset_in_page(*pos); | 
|  | loff_t		end_pos = *pos + array->obj_size - 1; | 
|  | loff_t		new_pos; | 
|  |  | 
|  | /* | 
|  | * If the current array record is not adjacent to a page boundary, we | 
|  | * are in the middle of the page.  We do not need to move the cursor. | 
|  | */ | 
|  | if (pgoff != 0 && pgoff + array->obj_size - 1 < PAGE_SIZE) | 
|  | return 0; | 
|  |  | 
|  | /* | 
|  | * Call SEEK_DATA on the last byte in the record we're about to read. | 
|  | * If the record ends at (or crosses) the end of a page then we know | 
|  | * that the first byte of the record is backed by pages and don't need | 
|  | * to query it.  If instead the record begins at the start of the page | 
|  | * then we know that querying the last byte is just as good as querying | 
|  | * the first byte, since records cannot be larger than a page. | 
|  | * | 
|  | * If the call returns the same file offset, we know this record is | 
|  | * backed by real pages.  We do not need to move the cursor. | 
|  | */ | 
|  | new_pos = xfile_seek_data(array->xfile, end_pos); | 
|  | if (new_pos == -ENXIO) | 
|  | return -ENODATA; | 
|  | if (new_pos < 0) | 
|  | return new_pos; | 
|  | if (new_pos == end_pos) | 
|  | return 0; | 
|  |  | 
|  | /* | 
|  | * Otherwise, SEEK_DATA told us how far up to move the file pointer to | 
|  | * find more data.  Move the array index to the first record past the | 
|  | * byte offset we were given. | 
|  | */ | 
|  | new_pos = roundup_64(new_pos, array->obj_size); | 
|  | *cur = xfarray_idx(array, new_pos); | 
|  | *pos = xfarray_pos(array, *cur); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Starting at *idx, fetch the next non-null array entry and advance the index | 
|  | * to set up the next _load_next call.  Returns ENODATA if we reach the end of | 
|  | * the array.  Callers must set @*idx to XFARRAY_CURSOR_INIT before the first | 
|  | * call to this function. | 
|  | */ | 
|  | int | 
|  | xfarray_load_next( | 
|  | struct xfarray	*array, | 
|  | xfarray_idx_t	*idx, | 
|  | void		*rec) | 
|  | { | 
|  | xfarray_idx_t	cur = *idx; | 
|  | loff_t		pos = xfarray_pos(array, cur); | 
|  | int		error; | 
|  |  | 
|  | do { | 
|  | if (cur >= array->nr) | 
|  | return -ENODATA; | 
|  |  | 
|  | /* | 
|  | * Ask the backing store for the location of next possible | 
|  | * written record, then retrieve that record. | 
|  | */ | 
|  | error = xfarray_find_data(array, &cur, &pos); | 
|  | if (error) | 
|  | return error; | 
|  | error = xfarray_load(array, cur, rec); | 
|  | if (error) | 
|  | return error; | 
|  |  | 
|  | cur++; | 
|  | pos += array->obj_size; | 
|  | } while (xfarray_element_is_null(array, rec)); | 
|  |  | 
|  | *idx = cur; | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* Sorting functions */ | 
|  |  | 
|  | #ifdef DEBUG | 
|  | # define xfarray_sort_bump_loads(si)	do { (si)->loads++; } while (0) | 
|  | # define xfarray_sort_bump_stores(si)	do { (si)->stores++; } while (0) | 
|  | # define xfarray_sort_bump_compares(si)	do { (si)->compares++; } while (0) | 
|  | # define xfarray_sort_bump_heapsorts(si) do { (si)->heapsorts++; } while (0) | 
|  | #else | 
|  | # define xfarray_sort_bump_loads(si) | 
|  | # define xfarray_sort_bump_stores(si) | 
|  | # define xfarray_sort_bump_compares(si) | 
|  | # define xfarray_sort_bump_heapsorts(si) | 
|  | #endif /* DEBUG */ | 
|  |  | 
|  | /* Load an array element for sorting. */ | 
|  | static inline int | 
|  | xfarray_sort_load( | 
|  | struct xfarray_sortinfo	*si, | 
|  | xfarray_idx_t		idx, | 
|  | void			*ptr) | 
|  | { | 
|  | xfarray_sort_bump_loads(si); | 
|  | return xfarray_load(si->array, idx, ptr); | 
|  | } | 
|  |  | 
|  | /* Store an array element for sorting. */ | 
|  | static inline int | 
|  | xfarray_sort_store( | 
|  | struct xfarray_sortinfo	*si, | 
|  | xfarray_idx_t		idx, | 
|  | void			*ptr) | 
|  | { | 
|  | xfarray_sort_bump_stores(si); | 
|  | return xfarray_store(si->array, idx, ptr); | 
|  | } | 
|  |  | 
|  | /* Compare an array element for sorting. */ | 
|  | static inline int | 
|  | xfarray_sort_cmp( | 
|  | struct xfarray_sortinfo	*si, | 
|  | const void		*a, | 
|  | const void		*b) | 
|  | { | 
|  | xfarray_sort_bump_compares(si); | 
|  | return si->cmp_fn(a, b); | 
|  | } | 
|  |  | 
|  | /* Return a pointer to the low index stack for quicksort partitioning. */ | 
|  | static inline xfarray_idx_t *xfarray_sortinfo_lo(struct xfarray_sortinfo *si) | 
|  | { | 
|  | return (xfarray_idx_t *)(si + 1); | 
|  | } | 
|  |  | 
|  | /* Return a pointer to the high index stack for quicksort partitioning. */ | 
|  | static inline xfarray_idx_t *xfarray_sortinfo_hi(struct xfarray_sortinfo *si) | 
|  | { | 
|  | return xfarray_sortinfo_lo(si) + si->max_stack_depth; | 
|  | } | 
|  |  | 
|  | /* Size of each element in the quicksort pivot array. */ | 
|  | static inline size_t | 
|  | xfarray_pivot_rec_sz( | 
|  | struct xfarray		*array) | 
|  | { | 
|  | return round_up(array->obj_size, 8) + sizeof(xfarray_idx_t); | 
|  | } | 
|  |  | 
|  | /* Allocate memory to handle the sort. */ | 
|  | static inline int | 
|  | xfarray_sortinfo_alloc( | 
|  | struct xfarray		*array, | 
|  | xfarray_cmp_fn		cmp_fn, | 
|  | unsigned int		flags, | 
|  | struct xfarray_sortinfo	**infop) | 
|  | { | 
|  | struct xfarray_sortinfo	*si; | 
|  | size_t			nr_bytes = sizeof(struct xfarray_sortinfo); | 
|  | size_t			pivot_rec_sz = xfarray_pivot_rec_sz(array); | 
|  | int			max_stack_depth; | 
|  |  | 
|  | /* | 
|  | * The median-of-nine pivot algorithm doesn't work if a subset has | 
|  | * fewer than 9 items.  Make sure the in-memory sort will always take | 
|  | * over for subsets where this wouldn't be the case. | 
|  | */ | 
|  | BUILD_BUG_ON(XFARRAY_QSORT_PIVOT_NR >= XFARRAY_ISORT_NR); | 
|  |  | 
|  | /* | 
|  | * Tail-call recursion during the partitioning phase means that | 
|  | * quicksort will never recurse more than log2(nr) times.  We need one | 
|  | * extra level of stack to hold the initial parameters.  In-memory | 
|  | * sort will always take care of the last few levels of recursion for | 
|  | * us, so we can reduce the stack depth by that much. | 
|  | */ | 
|  | max_stack_depth = ilog2(array->nr) + 1 - (XFARRAY_ISORT_SHIFT - 1); | 
|  | if (max_stack_depth < 1) | 
|  | max_stack_depth = 1; | 
|  |  | 
|  | /* Each level of quicksort uses a lo and a hi index */ | 
|  | nr_bytes += max_stack_depth * sizeof(xfarray_idx_t) * 2; | 
|  |  | 
|  | /* Scratchpad for in-memory sort, or finding the pivot */ | 
|  | nr_bytes += max_t(size_t, | 
|  | (XFARRAY_QSORT_PIVOT_NR + 1) * pivot_rec_sz, | 
|  | XFARRAY_ISORT_NR * array->obj_size); | 
|  |  | 
|  | si = kvzalloc(nr_bytes, XCHK_GFP_FLAGS); | 
|  | if (!si) | 
|  | return -ENOMEM; | 
|  |  | 
|  | si->array = array; | 
|  | si->cmp_fn = cmp_fn; | 
|  | si->flags = flags; | 
|  | si->max_stack_depth = max_stack_depth; | 
|  | si->max_stack_used = 1; | 
|  |  | 
|  | xfarray_sortinfo_lo(si)[0] = 0; | 
|  | xfarray_sortinfo_hi(si)[0] = array->nr - 1; | 
|  | si->relax = INIT_XCHK_RELAX; | 
|  | if (flags & XFARRAY_SORT_KILLABLE) | 
|  | si->relax.interruptible = false; | 
|  |  | 
|  | trace_xfarray_sort(si, nr_bytes); | 
|  | *infop = si; | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* Should this sort be terminated by a fatal signal? */ | 
|  | static inline bool | 
|  | xfarray_sort_terminated( | 
|  | struct xfarray_sortinfo	*si, | 
|  | int			*error) | 
|  | { | 
|  | /* | 
|  | * If preemption is disabled, we need to yield to the scheduler every | 
|  | * few seconds so that we don't run afoul of the soft lockup watchdog | 
|  | * or RCU stall detector. | 
|  | */ | 
|  | if (xchk_maybe_relax(&si->relax)) { | 
|  | if (*error == 0) | 
|  | *error = -EINTR; | 
|  | return true; | 
|  | } | 
|  | return false; | 
|  | } | 
|  |  | 
|  | /* Do we want an in-memory sort? */ | 
|  | static inline bool | 
|  | xfarray_want_isort( | 
|  | struct xfarray_sortinfo *si, | 
|  | xfarray_idx_t		start, | 
|  | xfarray_idx_t		end) | 
|  | { | 
|  | /* | 
|  | * For array subsets that fit in the scratchpad, it's much faster to | 
|  | * use the kernel's heapsort than quicksort's stack machine. | 
|  | */ | 
|  | return (end - start) < XFARRAY_ISORT_NR; | 
|  | } | 
|  |  | 
|  | /* Return the scratch space within the sortinfo structure. */ | 
|  | static inline void *xfarray_sortinfo_isort_scratch(struct xfarray_sortinfo *si) | 
|  | { | 
|  | return xfarray_sortinfo_hi(si) + si->max_stack_depth; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Sort a small number of array records using scratchpad memory.  The records | 
|  | * need not be contiguous in the xfile's memory pages. | 
|  | */ | 
|  | STATIC int | 
|  | xfarray_isort( | 
|  | struct xfarray_sortinfo	*si, | 
|  | xfarray_idx_t		lo, | 
|  | xfarray_idx_t		hi) | 
|  | { | 
|  | void			*scratch = xfarray_sortinfo_isort_scratch(si); | 
|  | loff_t			lo_pos = xfarray_pos(si->array, lo); | 
|  | loff_t			len = xfarray_pos(si->array, hi - lo + 1); | 
|  | int			error; | 
|  |  | 
|  | trace_xfarray_isort(si, lo, hi); | 
|  |  | 
|  | xfarray_sort_bump_loads(si); | 
|  | error = xfile_load(si->array->xfile, scratch, len, lo_pos); | 
|  | if (error) | 
|  | return error; | 
|  |  | 
|  | xfarray_sort_bump_heapsorts(si); | 
|  | sort(scratch, hi - lo + 1, si->array->obj_size, si->cmp_fn, NULL); | 
|  |  | 
|  | xfarray_sort_bump_stores(si); | 
|  | return xfile_store(si->array->xfile, scratch, len, lo_pos); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Sort the records from lo to hi (inclusive) if they are all backed by the | 
|  | * same memory folio.  Returns 1 if it sorted, 0 if it did not, or a negative | 
|  | * errno. | 
|  | */ | 
|  | STATIC int | 
|  | xfarray_foliosort( | 
|  | struct xfarray_sortinfo	*si, | 
|  | xfarray_idx_t		lo, | 
|  | xfarray_idx_t		hi) | 
|  | { | 
|  | struct folio		*folio; | 
|  | void			*startp; | 
|  | loff_t			lo_pos = xfarray_pos(si->array, lo); | 
|  | uint64_t		len = xfarray_pos(si->array, hi - lo + 1); | 
|  |  | 
|  | /* No single folio could back this many records. */ | 
|  | if (len > XFILE_MAX_FOLIO_SIZE) | 
|  | return 0; | 
|  |  | 
|  | xfarray_sort_bump_loads(si); | 
|  | folio = xfile_get_folio(si->array->xfile, lo_pos, len, XFILE_ALLOC); | 
|  | if (IS_ERR(folio)) | 
|  | return PTR_ERR(folio); | 
|  | if (!folio) | 
|  | return 0; | 
|  |  | 
|  | trace_xfarray_foliosort(si, lo, hi); | 
|  |  | 
|  | xfarray_sort_bump_heapsorts(si); | 
|  | startp = folio_address(folio) + offset_in_folio(folio, lo_pos); | 
|  | sort(startp, hi - lo + 1, si->array->obj_size, si->cmp_fn, NULL); | 
|  |  | 
|  | xfarray_sort_bump_stores(si); | 
|  | xfile_put_folio(si->array->xfile, folio); | 
|  | return 1; | 
|  | } | 
|  |  | 
|  | /* Return a pointer to the xfarray pivot record within the sortinfo struct. */ | 
|  | static inline void *xfarray_sortinfo_pivot(struct xfarray_sortinfo *si) | 
|  | { | 
|  | return xfarray_sortinfo_hi(si) + si->max_stack_depth; | 
|  | } | 
|  |  | 
|  | /* Return a pointer to the start of the pivot array. */ | 
|  | static inline void * | 
|  | xfarray_sortinfo_pivot_array( | 
|  | struct xfarray_sortinfo	*si) | 
|  | { | 
|  | return xfarray_sortinfo_pivot(si) + si->array->obj_size; | 
|  | } | 
|  |  | 
|  | /* The xfarray record is stored at the start of each pivot array element. */ | 
|  | static inline void * | 
|  | xfarray_pivot_array_rec( | 
|  | void			*pa, | 
|  | size_t			pa_recsz, | 
|  | unsigned int		pa_idx) | 
|  | { | 
|  | return pa + (pa_recsz * pa_idx); | 
|  | } | 
|  |  | 
|  | /* The xfarray index is stored at the end of each pivot array element. */ | 
|  | static inline xfarray_idx_t * | 
|  | xfarray_pivot_array_idx( | 
|  | void			*pa, | 
|  | size_t			pa_recsz, | 
|  | unsigned int		pa_idx) | 
|  | { | 
|  | return xfarray_pivot_array_rec(pa, pa_recsz, pa_idx + 1) - | 
|  | sizeof(xfarray_idx_t); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Find a pivot value for quicksort partitioning, swap it with a[lo], and save | 
|  | * the cached pivot record for the next step. | 
|  | * | 
|  | * Load evenly-spaced records within the given range into memory, sort them, | 
|  | * and choose the pivot from the median record.  Using multiple points will | 
|  | * improve the quality of the pivot selection, and hopefully avoid the worst | 
|  | * quicksort behavior, since our array values are nearly always evenly sorted. | 
|  | */ | 
|  | STATIC int | 
|  | xfarray_qsort_pivot( | 
|  | struct xfarray_sortinfo	*si, | 
|  | xfarray_idx_t		lo, | 
|  | xfarray_idx_t		hi) | 
|  | { | 
|  | void			*pivot = xfarray_sortinfo_pivot(si); | 
|  | void			*parray = xfarray_sortinfo_pivot_array(si); | 
|  | void			*recp; | 
|  | xfarray_idx_t		*idxp; | 
|  | xfarray_idx_t		step = (hi - lo) / (XFARRAY_QSORT_PIVOT_NR - 1); | 
|  | size_t			pivot_rec_sz = xfarray_pivot_rec_sz(si->array); | 
|  | int			i, j; | 
|  | int			error; | 
|  |  | 
|  | ASSERT(step > 0); | 
|  |  | 
|  | /* | 
|  | * Load the xfarray indexes of the records we intend to sample into the | 
|  | * pivot array. | 
|  | */ | 
|  | idxp = xfarray_pivot_array_idx(parray, pivot_rec_sz, 0); | 
|  | *idxp = lo; | 
|  | for (i = 1; i < XFARRAY_QSORT_PIVOT_NR - 1; i++) { | 
|  | idxp = xfarray_pivot_array_idx(parray, pivot_rec_sz, i); | 
|  | *idxp = lo + (i * step); | 
|  | } | 
|  | idxp = xfarray_pivot_array_idx(parray, pivot_rec_sz, | 
|  | XFARRAY_QSORT_PIVOT_NR - 1); | 
|  | *idxp = hi; | 
|  |  | 
|  | /* Load the selected xfarray records into the pivot array. */ | 
|  | for (i = 0; i < XFARRAY_QSORT_PIVOT_NR; i++) { | 
|  | xfarray_idx_t	idx; | 
|  |  | 
|  | recp = xfarray_pivot_array_rec(parray, pivot_rec_sz, i); | 
|  | idxp = xfarray_pivot_array_idx(parray, pivot_rec_sz, i); | 
|  |  | 
|  | /* No unset records; load directly into the array. */ | 
|  | if (likely(si->array->unset_slots == 0)) { | 
|  | error = xfarray_sort_load(si, *idxp, recp); | 
|  | if (error) | 
|  | return error; | 
|  | continue; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Load non-null records into the scratchpad without changing | 
|  | * the xfarray_idx_t in the pivot array. | 
|  | */ | 
|  | idx = *idxp; | 
|  | xfarray_sort_bump_loads(si); | 
|  | error = xfarray_load_next(si->array, &idx, recp); | 
|  | if (error) | 
|  | return error; | 
|  | } | 
|  |  | 
|  | xfarray_sort_bump_heapsorts(si); | 
|  | sort(parray, XFARRAY_QSORT_PIVOT_NR, pivot_rec_sz, si->cmp_fn, NULL); | 
|  |  | 
|  | /* | 
|  | * We sorted the pivot array records (which includes the xfarray | 
|  | * indices) in xfarray record order.  The median element of the pivot | 
|  | * array contains the xfarray record that we will use as the pivot. | 
|  | * Copy that xfarray record to the designated space. | 
|  | */ | 
|  | recp = xfarray_pivot_array_rec(parray, pivot_rec_sz, | 
|  | XFARRAY_QSORT_PIVOT_NR / 2); | 
|  | memcpy(pivot, recp, si->array->obj_size); | 
|  |  | 
|  | /* If the pivot record we chose was already in a[lo] then we're done. */ | 
|  | idxp = xfarray_pivot_array_idx(parray, pivot_rec_sz, | 
|  | XFARRAY_QSORT_PIVOT_NR / 2); | 
|  | if (*idxp == lo) | 
|  | return 0; | 
|  |  | 
|  | /* | 
|  | * Find the cached copy of a[lo] in the pivot array so that we can swap | 
|  | * a[lo] and a[pivot]. | 
|  | */ | 
|  | for (i = 0, j = -1; i < XFARRAY_QSORT_PIVOT_NR; i++) { | 
|  | idxp = xfarray_pivot_array_idx(parray, pivot_rec_sz, i); | 
|  | if (*idxp == lo) | 
|  | j = i; | 
|  | } | 
|  | if (j < 0) { | 
|  | ASSERT(j >= 0); | 
|  | return -EFSCORRUPTED; | 
|  | } | 
|  |  | 
|  | /* Swap a[lo] and a[pivot]. */ | 
|  | error = xfarray_sort_store(si, lo, pivot); | 
|  | if (error) | 
|  | return error; | 
|  |  | 
|  | recp = xfarray_pivot_array_rec(parray, pivot_rec_sz, j); | 
|  | idxp = xfarray_pivot_array_idx(parray, pivot_rec_sz, | 
|  | XFARRAY_QSORT_PIVOT_NR / 2); | 
|  | return xfarray_sort_store(si, *idxp, recp); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Set up the pointers for the next iteration.  We push onto the stack all of | 
|  | * the unsorted values between a[lo + 1] and a[end[i]], and we tweak the | 
|  | * current stack frame to point to the unsorted values between a[beg[i]] and | 
|  | * a[lo] so that those values will be sorted when we pop the stack. | 
|  | */ | 
|  | static inline int | 
|  | xfarray_qsort_push( | 
|  | struct xfarray_sortinfo	*si, | 
|  | xfarray_idx_t		*si_lo, | 
|  | xfarray_idx_t		*si_hi, | 
|  | xfarray_idx_t		lo, | 
|  | xfarray_idx_t		hi) | 
|  | { | 
|  | /* Check for stack overflows */ | 
|  | if (si->stack_depth >= si->max_stack_depth - 1) { | 
|  | ASSERT(si->stack_depth < si->max_stack_depth - 1); | 
|  | return -EFSCORRUPTED; | 
|  | } | 
|  |  | 
|  | si->max_stack_used = max_t(uint8_t, si->max_stack_used, | 
|  | si->stack_depth + 2); | 
|  |  | 
|  | si_lo[si->stack_depth + 1] = lo + 1; | 
|  | si_hi[si->stack_depth + 1] = si_hi[si->stack_depth]; | 
|  | si_hi[si->stack_depth++] = lo - 1; | 
|  |  | 
|  | /* | 
|  | * Always start with the smaller of the two partitions to keep the | 
|  | * amount of recursion in check. | 
|  | */ | 
|  | if (si_hi[si->stack_depth]     - si_lo[si->stack_depth] > | 
|  | si_hi[si->stack_depth - 1] - si_lo[si->stack_depth - 1]) { | 
|  | swap(si_lo[si->stack_depth], si_lo[si->stack_depth - 1]); | 
|  | swap(si_hi[si->stack_depth], si_hi[si->stack_depth - 1]); | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static inline void | 
|  | xfarray_sort_scan_done( | 
|  | struct xfarray_sortinfo	*si) | 
|  | { | 
|  | if (si->folio) | 
|  | xfile_put_folio(si->array->xfile, si->folio); | 
|  | si->folio = NULL; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Cache the folio backing the start of the given array element.  If the array | 
|  | * element is contained entirely within the folio, return a pointer to the | 
|  | * cached folio.  Otherwise, load the element into the scratchpad and return a | 
|  | * pointer to the scratchpad. | 
|  | */ | 
|  | static inline int | 
|  | xfarray_sort_scan( | 
|  | struct xfarray_sortinfo	*si, | 
|  | xfarray_idx_t		idx, | 
|  | void			**ptrp) | 
|  | { | 
|  | loff_t			idx_pos = xfarray_pos(si->array, idx); | 
|  | int			error = 0; | 
|  |  | 
|  | if (xfarray_sort_terminated(si, &error)) | 
|  | return error; | 
|  |  | 
|  | trace_xfarray_sort_scan(si, idx); | 
|  |  | 
|  | /* If the cached folio doesn't cover this index, release it. */ | 
|  | if (si->folio && | 
|  | (idx < si->first_folio_idx || idx > si->last_folio_idx)) | 
|  | xfarray_sort_scan_done(si); | 
|  |  | 
|  | /* Grab the first folio that backs this array element. */ | 
|  | if (!si->folio) { | 
|  | struct folio	*folio; | 
|  | loff_t		next_pos; | 
|  |  | 
|  | folio = xfile_get_folio(si->array->xfile, idx_pos, | 
|  | si->array->obj_size, XFILE_ALLOC); | 
|  | if (IS_ERR(folio)) | 
|  | return PTR_ERR(folio); | 
|  | si->folio = folio; | 
|  |  | 
|  | si->first_folio_idx = xfarray_idx(si->array, | 
|  | folio_pos(si->folio) + si->array->obj_size - 1); | 
|  |  | 
|  | next_pos = folio_pos(si->folio) + folio_size(si->folio); | 
|  | si->last_folio_idx = xfarray_idx(si->array, next_pos - 1); | 
|  | if (xfarray_pos(si->array, si->last_folio_idx + 1) > next_pos) | 
|  | si->last_folio_idx--; | 
|  |  | 
|  | trace_xfarray_sort_scan(si, idx); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * If this folio still doesn't cover the desired element, it must cross | 
|  | * a folio boundary.  Read into the scratchpad and we're done. | 
|  | */ | 
|  | if (idx < si->first_folio_idx || idx > si->last_folio_idx) { | 
|  | void		*temp = xfarray_scratch(si->array); | 
|  |  | 
|  | error = xfile_load(si->array->xfile, temp, si->array->obj_size, | 
|  | idx_pos); | 
|  | if (error) | 
|  | return error; | 
|  |  | 
|  | *ptrp = temp; | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* Otherwise return a pointer to the array element in the folio. */ | 
|  | *ptrp = folio_address(si->folio) + offset_in_folio(si->folio, idx_pos); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Sort the array elements via quicksort.  This implementation incorporates | 
|  | * four optimizations discussed in Sedgewick: | 
|  | * | 
|  | * 1. Use an explicit stack of array indices to store the next array partition | 
|  | *    to sort.  This helps us to avoid recursion in the call stack, which is | 
|  | *    particularly expensive in the kernel. | 
|  | * | 
|  | * 2. For arrays with records in arbitrary or user-controlled order, choose the | 
|  | *    pivot element using a median-of-nine decision tree.  This reduces the | 
|  | *    probability of selecting a bad pivot value which causes worst case | 
|  | *    behavior (i.e. partition sizes of 1). | 
|  | * | 
|  | * 3. The smaller of the two sub-partitions is pushed onto the stack to start | 
|  | *    the next level of recursion, and the larger sub-partition replaces the | 
|  | *    current stack frame.  This guarantees that we won't need more than | 
|  | *    log2(nr) stack space. | 
|  | * | 
|  | * 4. For small sets, load the records into the scratchpad and run heapsort on | 
|  | *    them because that is very fast.  In the author's experience, this yields | 
|  | *    a ~10% reduction in runtime. | 
|  | * | 
|  | *    If a small set is contained entirely within a single xfile memory page, | 
|  | *    map the page directly and run heap sort directly on the xfile page | 
|  | *    instead of using the load/store interface.  This halves the runtime. | 
|  | * | 
|  | * 5. This optimization is specific to the implementation.  When converging lo | 
|  | *    and hi after selecting a pivot, we will try to retain the xfile memory | 
|  | *    page between load calls, which reduces run time by 50%. | 
|  | */ | 
|  |  | 
|  | /* | 
|  | * Due to the use of signed indices, we can only support up to 2^63 records. | 
|  | * Files can only grow to 2^63 bytes, so this is not much of a limitation. | 
|  | */ | 
|  | #define QSORT_MAX_RECS		(1ULL << 63) | 
|  |  | 
|  | int | 
|  | xfarray_sort( | 
|  | struct xfarray		*array, | 
|  | xfarray_cmp_fn		cmp_fn, | 
|  | unsigned int		flags) | 
|  | { | 
|  | struct xfarray_sortinfo	*si; | 
|  | xfarray_idx_t		*si_lo, *si_hi; | 
|  | void			*pivot; | 
|  | void			*scratch = xfarray_scratch(array); | 
|  | xfarray_idx_t		lo, hi; | 
|  | int			error = 0; | 
|  |  | 
|  | if (array->nr < 2) | 
|  | return 0; | 
|  | if (array->nr >= QSORT_MAX_RECS) | 
|  | return -E2BIG; | 
|  |  | 
|  | error = xfarray_sortinfo_alloc(array, cmp_fn, flags, &si); | 
|  | if (error) | 
|  | return error; | 
|  | si_lo = xfarray_sortinfo_lo(si); | 
|  | si_hi = xfarray_sortinfo_hi(si); | 
|  | pivot = xfarray_sortinfo_pivot(si); | 
|  |  | 
|  | while (si->stack_depth >= 0) { | 
|  | int		ret; | 
|  |  | 
|  | lo = si_lo[si->stack_depth]; | 
|  | hi = si_hi[si->stack_depth]; | 
|  |  | 
|  | trace_xfarray_qsort(si, lo, hi); | 
|  |  | 
|  | /* Nothing left in this partition to sort; pop stack. */ | 
|  | if (lo >= hi) { | 
|  | si->stack_depth--; | 
|  | continue; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * If directly mapping the folio and sorting can solve our | 
|  | * problems, we're done. | 
|  | */ | 
|  | ret = xfarray_foliosort(si, lo, hi); | 
|  | if (ret < 0) | 
|  | goto out_free; | 
|  | if (ret == 1) { | 
|  | si->stack_depth--; | 
|  | continue; | 
|  | } | 
|  |  | 
|  | /* If insertion sort can solve our problems, we're done. */ | 
|  | if (xfarray_want_isort(si, lo, hi)) { | 
|  | error = xfarray_isort(si, lo, hi); | 
|  | if (error) | 
|  | goto out_free; | 
|  | si->stack_depth--; | 
|  | continue; | 
|  | } | 
|  |  | 
|  | /* Pick a pivot, move it to a[lo] and stash it. */ | 
|  | error = xfarray_qsort_pivot(si, lo, hi); | 
|  | if (error) | 
|  | goto out_free; | 
|  |  | 
|  | /* | 
|  | * Rearrange a[lo..hi] such that everything smaller than the | 
|  | * pivot is on the left side of the range and everything larger | 
|  | * than the pivot is on the right side of the range. | 
|  | */ | 
|  | while (lo < hi) { | 
|  | void	*p; | 
|  |  | 
|  | /* | 
|  | * Decrement hi until it finds an a[hi] less than the | 
|  | * pivot value. | 
|  | */ | 
|  | error = xfarray_sort_scan(si, hi, &p); | 
|  | if (error) | 
|  | goto out_free; | 
|  | while (xfarray_sort_cmp(si, p, pivot) >= 0 && lo < hi) { | 
|  | hi--; | 
|  | error = xfarray_sort_scan(si, hi, &p); | 
|  | if (error) | 
|  | goto out_free; | 
|  | } | 
|  | if (p != scratch) | 
|  | memcpy(scratch, p, si->array->obj_size); | 
|  | xfarray_sort_scan_done(si); | 
|  | if (xfarray_sort_terminated(si, &error)) | 
|  | goto out_free; | 
|  |  | 
|  | /* Copy that item (a[hi]) to a[lo]. */ | 
|  | if (lo < hi) { | 
|  | error = xfarray_sort_store(si, lo++, scratch); | 
|  | if (error) | 
|  | goto out_free; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Increment lo until it finds an a[lo] greater than | 
|  | * the pivot value. | 
|  | */ | 
|  | error = xfarray_sort_scan(si, lo, &p); | 
|  | if (error) | 
|  | goto out_free; | 
|  | while (xfarray_sort_cmp(si, p, pivot) <= 0 && lo < hi) { | 
|  | lo++; | 
|  | error = xfarray_sort_scan(si, lo, &p); | 
|  | if (error) | 
|  | goto out_free; | 
|  | } | 
|  | if (p != scratch) | 
|  | memcpy(scratch, p, si->array->obj_size); | 
|  | xfarray_sort_scan_done(si); | 
|  | if (xfarray_sort_terminated(si, &error)) | 
|  | goto out_free; | 
|  |  | 
|  | /* Copy that item (a[lo]) to a[hi]. */ | 
|  | if (lo < hi) { | 
|  | error = xfarray_sort_store(si, hi--, scratch); | 
|  | if (error) | 
|  | goto out_free; | 
|  | } | 
|  |  | 
|  | if (xfarray_sort_terminated(si, &error)) | 
|  | goto out_free; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Put our pivot value in the correct place at a[lo].  All | 
|  | * values between a[beg[i]] and a[lo - 1] should be less than | 
|  | * the pivot; and all values between a[lo + 1] and a[end[i]-1] | 
|  | * should be greater than the pivot. | 
|  | */ | 
|  | error = xfarray_sort_store(si, lo, pivot); | 
|  | if (error) | 
|  | goto out_free; | 
|  |  | 
|  | /* Set up the stack frame to process the two partitions. */ | 
|  | error = xfarray_qsort_push(si, si_lo, si_hi, lo, hi); | 
|  | if (error) | 
|  | goto out_free; | 
|  |  | 
|  | if (xfarray_sort_terminated(si, &error)) | 
|  | goto out_free; | 
|  | } | 
|  |  | 
|  | out_free: | 
|  | trace_xfarray_sort_stats(si, error); | 
|  | xfarray_sort_scan_done(si); | 
|  | kvfree(si); | 
|  | return error; | 
|  | } | 
|  |  | 
|  | /* How many bytes is this array consuming? */ | 
|  | unsigned long long | 
|  | xfarray_bytes( | 
|  | struct xfarray		*array) | 
|  | { | 
|  | return xfile_bytes(array->xfile); | 
|  | } | 
|  |  | 
|  | /* Empty the entire array. */ | 
|  | void | 
|  | xfarray_truncate( | 
|  | struct xfarray	*array) | 
|  | { | 
|  | xfile_discard(array->xfile, 0, MAX_LFS_FILESIZE); | 
|  | array->nr = 0; | 
|  | } |