|  | // SPDX-License-Identifier: GPL-2.0 | 
|  | /* | 
|  | * Copyright (C) 2013 Fusion IO.  All rights reserved. | 
|  | */ | 
|  |  | 
|  | #include <linux/slab.h> | 
|  | #include "btrfs-tests.h" | 
|  | #include "../ctree.h" | 
|  | #include "../disk-io.h" | 
|  | #include "../free-space-cache.h" | 
|  | #include "../block-group.h" | 
|  |  | 
|  | #define BITS_PER_BITMAP		(PAGE_SIZE * 8UL) | 
|  |  | 
|  | /* | 
|  | * This test just does basic sanity checking, making sure we can add an extent | 
|  | * entry and remove space from either end and the middle, and make sure we can | 
|  | * remove space that covers adjacent extent entries. | 
|  | */ | 
|  | static int test_extents(struct btrfs_block_group *cache) | 
|  | { | 
|  | int ret = 0; | 
|  |  | 
|  | test_msg("running extent only tests"); | 
|  |  | 
|  | /* First just make sure we can remove an entire entry */ | 
|  | ret = btrfs_add_free_space(cache, 0, SZ_4M); | 
|  | if (ret) { | 
|  | test_err("error adding initial extents %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | ret = btrfs_remove_free_space(cache, 0, SZ_4M); | 
|  | if (ret) { | 
|  | test_err("error removing extent %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | if (test_check_exists(cache, 0, SZ_4M)) { | 
|  | test_err("full remove left some lingering space"); | 
|  | return -1; | 
|  | } | 
|  |  | 
|  | /* Ok edge and middle cases now */ | 
|  | ret = btrfs_add_free_space(cache, 0, SZ_4M); | 
|  | if (ret) { | 
|  | test_err("error adding half extent %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | ret = btrfs_remove_free_space(cache, 3 * SZ_1M, SZ_1M); | 
|  | if (ret) { | 
|  | test_err("error removing tail end %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | ret = btrfs_remove_free_space(cache, 0, SZ_1M); | 
|  | if (ret) { | 
|  | test_err("error removing front end %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | ret = btrfs_remove_free_space(cache, SZ_2M, 4096); | 
|  | if (ret) { | 
|  | test_err("error removing middle piece %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | if (test_check_exists(cache, 0, SZ_1M)) { | 
|  | test_err("still have space at the front"); | 
|  | return -1; | 
|  | } | 
|  |  | 
|  | if (test_check_exists(cache, SZ_2M, 4096)) { | 
|  | test_err("still have space in the middle"); | 
|  | return -1; | 
|  | } | 
|  |  | 
|  | if (test_check_exists(cache, 3 * SZ_1M, SZ_1M)) { | 
|  | test_err("still have space at the end"); | 
|  | return -1; | 
|  | } | 
|  |  | 
|  | /* Cleanup */ | 
|  | btrfs_remove_free_space_cache(cache); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int test_bitmaps(struct btrfs_block_group *cache, u32 sectorsize) | 
|  | { | 
|  | u64 next_bitmap_offset; | 
|  | int ret; | 
|  |  | 
|  | test_msg("running bitmap only tests"); | 
|  |  | 
|  | ret = test_add_free_space_entry(cache, 0, SZ_4M, 1); | 
|  | if (ret) { | 
|  | test_err("couldn't create a bitmap entry %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | ret = btrfs_remove_free_space(cache, 0, SZ_4M); | 
|  | if (ret) { | 
|  | test_err("error removing bitmap full range %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | if (test_check_exists(cache, 0, SZ_4M)) { | 
|  | test_err("left some space in bitmap"); | 
|  | return -1; | 
|  | } | 
|  |  | 
|  | ret = test_add_free_space_entry(cache, 0, SZ_4M, 1); | 
|  | if (ret) { | 
|  | test_err("couldn't add to our bitmap entry %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | ret = btrfs_remove_free_space(cache, SZ_1M, SZ_2M); | 
|  | if (ret) { | 
|  | test_err("couldn't remove middle chunk %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * The first bitmap we have starts at offset 0 so the next one is just | 
|  | * at the end of the first bitmap. | 
|  | */ | 
|  | next_bitmap_offset = (u64)(BITS_PER_BITMAP * sectorsize); | 
|  |  | 
|  | /* Test a bit straddling two bitmaps */ | 
|  | ret = test_add_free_space_entry(cache, next_bitmap_offset - SZ_2M, | 
|  | SZ_4M, 1); | 
|  | if (ret) { | 
|  | test_err("couldn't add space that straddles two bitmaps %d", | 
|  | ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | ret = btrfs_remove_free_space(cache, next_bitmap_offset - SZ_1M, SZ_2M); | 
|  | if (ret) { | 
|  | test_err("couldn't remove overlapping space %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | if (test_check_exists(cache, next_bitmap_offset - SZ_1M, SZ_2M)) { | 
|  | test_err("left some space when removing overlapping"); | 
|  | return -1; | 
|  | } | 
|  |  | 
|  | btrfs_remove_free_space_cache(cache); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* This is the high grade jackassery */ | 
|  | static int test_bitmaps_and_extents(struct btrfs_block_group *cache, | 
|  | u32 sectorsize) | 
|  | { | 
|  | u64 bitmap_offset = (u64)(BITS_PER_BITMAP * sectorsize); | 
|  | int ret; | 
|  |  | 
|  | test_msg("running bitmap and extent tests"); | 
|  |  | 
|  | /* | 
|  | * First let's do something simple, an extent at the same offset as the | 
|  | * bitmap, but the free space completely in the extent and then | 
|  | * completely in the bitmap. | 
|  | */ | 
|  | ret = test_add_free_space_entry(cache, SZ_4M, SZ_1M, 1); | 
|  | if (ret) { | 
|  | test_err("couldn't create bitmap entry %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | ret = test_add_free_space_entry(cache, 0, SZ_1M, 0); | 
|  | if (ret) { | 
|  | test_err("couldn't add extent entry %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | ret = btrfs_remove_free_space(cache, 0, SZ_1M); | 
|  | if (ret) { | 
|  | test_err("couldn't remove extent entry %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | if (test_check_exists(cache, 0, SZ_1M)) { | 
|  | test_err("left remnants after our remove"); | 
|  | return -1; | 
|  | } | 
|  |  | 
|  | /* Now to add back the extent entry and remove from the bitmap */ | 
|  | ret = test_add_free_space_entry(cache, 0, SZ_1M, 0); | 
|  | if (ret) { | 
|  | test_err("couldn't re-add extent entry %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | ret = btrfs_remove_free_space(cache, SZ_4M, SZ_1M); | 
|  | if (ret) { | 
|  | test_err("couldn't remove from bitmap %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | if (test_check_exists(cache, SZ_4M, SZ_1M)) { | 
|  | test_err("left remnants in the bitmap"); | 
|  | return -1; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Ok so a little more evil, extent entry and bitmap at the same offset, | 
|  | * removing an overlapping chunk. | 
|  | */ | 
|  | ret = test_add_free_space_entry(cache, SZ_1M, SZ_4M, 1); | 
|  | if (ret) { | 
|  | test_err("couldn't add to a bitmap %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | ret = btrfs_remove_free_space(cache, SZ_512K, 3 * SZ_1M); | 
|  | if (ret) { | 
|  | test_err("couldn't remove overlapping space %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | if (test_check_exists(cache, SZ_512K, 3 * SZ_1M)) { | 
|  | test_err("left over pieces after removing overlapping"); | 
|  | return -1; | 
|  | } | 
|  |  | 
|  | btrfs_remove_free_space_cache(cache); | 
|  |  | 
|  | /* Now with the extent entry offset into the bitmap */ | 
|  | ret = test_add_free_space_entry(cache, SZ_4M, SZ_4M, 1); | 
|  | if (ret) { | 
|  | test_err("couldn't add space to the bitmap %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | ret = test_add_free_space_entry(cache, SZ_2M, SZ_2M, 0); | 
|  | if (ret) { | 
|  | test_err("couldn't add extent to the cache %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | ret = btrfs_remove_free_space(cache, 3 * SZ_1M, SZ_4M); | 
|  | if (ret) { | 
|  | test_err("problem removing overlapping space %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | if (test_check_exists(cache, 3 * SZ_1M, SZ_4M)) { | 
|  | test_err("left something behind when removing space"); | 
|  | return -1; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * This has blown up in the past, the extent entry starts before the | 
|  | * bitmap entry, but we're trying to remove an offset that falls | 
|  | * completely within the bitmap range and is in both the extent entry | 
|  | * and the bitmap entry, looks like this | 
|  | * | 
|  | *   [ extent ] | 
|  | *      [ bitmap ] | 
|  | *        [ del ] | 
|  | */ | 
|  | btrfs_remove_free_space_cache(cache); | 
|  | ret = test_add_free_space_entry(cache, bitmap_offset + SZ_4M, SZ_4M, 1); | 
|  | if (ret) { | 
|  | test_err("couldn't add bitmap %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | ret = test_add_free_space_entry(cache, bitmap_offset - SZ_1M, | 
|  | 5 * SZ_1M, 0); | 
|  | if (ret) { | 
|  | test_err("couldn't add extent entry %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | ret = btrfs_remove_free_space(cache, bitmap_offset + SZ_1M, 5 * SZ_1M); | 
|  | if (ret) { | 
|  | test_err("failed to free our space %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | if (test_check_exists(cache, bitmap_offset + SZ_1M, 5 * SZ_1M)) { | 
|  | test_err("left stuff over"); | 
|  | return -1; | 
|  | } | 
|  |  | 
|  | btrfs_remove_free_space_cache(cache); | 
|  |  | 
|  | /* | 
|  | * This blew up before, we have part of the free space in a bitmap and | 
|  | * then the entirety of the rest of the space in an extent.  This used | 
|  | * to return -EAGAIN back from btrfs_remove_extent, make sure this | 
|  | * doesn't happen. | 
|  | */ | 
|  | ret = test_add_free_space_entry(cache, SZ_1M, SZ_2M, 1); | 
|  | if (ret) { | 
|  | test_err("couldn't add bitmap entry %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | ret = test_add_free_space_entry(cache, 3 * SZ_1M, SZ_1M, 0); | 
|  | if (ret) { | 
|  | test_err("couldn't add extent entry %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | ret = btrfs_remove_free_space(cache, SZ_1M, 3 * SZ_1M); | 
|  | if (ret) { | 
|  | test_err("error removing bitmap and extent overlapping %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | btrfs_remove_free_space_cache(cache); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* Used by test_steal_space_from_bitmap_to_extent(). */ | 
|  | static bool test_use_bitmap(struct btrfs_free_space_ctl *ctl, | 
|  | struct btrfs_free_space *info) | 
|  | { | 
|  | return ctl->free_extents > 0; | 
|  | } | 
|  |  | 
|  | /* Used by test_steal_space_from_bitmap_to_extent(). */ | 
|  | static int | 
|  | check_num_extents_and_bitmaps(const struct btrfs_block_group *cache, | 
|  | const int num_extents, | 
|  | const int num_bitmaps) | 
|  | { | 
|  | if (cache->free_space_ctl->free_extents != num_extents) { | 
|  | test_err( | 
|  | "incorrect # of extent entries in the cache: %d, expected %d", | 
|  | cache->free_space_ctl->free_extents, num_extents); | 
|  | return -EINVAL; | 
|  | } | 
|  | if (cache->free_space_ctl->total_bitmaps != num_bitmaps) { | 
|  | test_err( | 
|  | "incorrect # of extent entries in the cache: %d, expected %d", | 
|  | cache->free_space_ctl->total_bitmaps, num_bitmaps); | 
|  | return -EINVAL; | 
|  | } | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* Used by test_steal_space_from_bitmap_to_extent(). */ | 
|  | static int check_cache_empty(struct btrfs_block_group *cache) | 
|  | { | 
|  | u64 offset; | 
|  | u64 max_extent_size; | 
|  |  | 
|  | /* | 
|  | * Now lets confirm that there's absolutely no free space left to | 
|  | * allocate. | 
|  | */ | 
|  | if (cache->free_space_ctl->free_space != 0) { | 
|  | test_err("cache free space is not 0"); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | /* And any allocation request, no matter how small, should fail now. */ | 
|  | offset = btrfs_find_space_for_alloc(cache, 0, 4096, 0, | 
|  | &max_extent_size); | 
|  | if (offset != 0) { | 
|  | test_err("space allocation did not fail, returned offset: %llu", | 
|  | offset); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | /* And no extent nor bitmap entries in the cache anymore. */ | 
|  | return check_num_extents_and_bitmaps(cache, 0, 0); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Before we were able to steal free space from a bitmap entry to an extent | 
|  | * entry, we could end up with 2 entries representing a contiguous free space. | 
|  | * One would be an extent entry and the other a bitmap entry. Since in order | 
|  | * to allocate space to a caller we use only 1 entry, we couldn't return that | 
|  | * whole range to the caller if it was requested. This forced the caller to | 
|  | * either assume ENOSPC or perform several smaller space allocations, which | 
|  | * wasn't optimal as they could be spread all over the block group while under | 
|  | * concurrency (extra overhead and fragmentation). | 
|  | * | 
|  | * This stealing approach is beneficial, since we always prefer to allocate | 
|  | * from extent entries, both for clustered and non-clustered allocation | 
|  | * requests. | 
|  | */ | 
|  | static int | 
|  | test_steal_space_from_bitmap_to_extent(struct btrfs_block_group *cache, | 
|  | u32 sectorsize) | 
|  | { | 
|  | int ret; | 
|  | u64 offset; | 
|  | u64 max_extent_size; | 
|  | const struct btrfs_free_space_op test_free_space_ops = { | 
|  | .use_bitmap = test_use_bitmap, | 
|  | }; | 
|  | const struct btrfs_free_space_op *orig_free_space_ops; | 
|  |  | 
|  | test_msg("running space stealing from bitmap to extent tests"); | 
|  |  | 
|  | /* | 
|  | * For this test, we want to ensure we end up with an extent entry | 
|  | * immediately adjacent to a bitmap entry, where the bitmap starts | 
|  | * at an offset where the extent entry ends. We keep adding and | 
|  | * removing free space to reach into this state, but to get there | 
|  | * we need to reach a point where marking new free space doesn't | 
|  | * result in adding new extent entries or merging the new space | 
|  | * with existing extent entries - the space ends up being marked | 
|  | * in an existing bitmap that covers the new free space range. | 
|  | * | 
|  | * To get there, we need to reach the threshold defined set at | 
|  | * cache->free_space_ctl->extents_thresh, which currently is | 
|  | * 256 extents on a x86_64 system at least, and a few other | 
|  | * conditions (check free_space_cache.c). Instead of making the | 
|  | * test much longer and complicated, use a "use_bitmap" operation | 
|  | * that forces use of bitmaps as soon as we have at least 1 | 
|  | * extent entry. | 
|  | */ | 
|  | orig_free_space_ops = cache->free_space_ctl->op; | 
|  | cache->free_space_ctl->op = &test_free_space_ops; | 
|  |  | 
|  | /* | 
|  | * Extent entry covering free space range [128Mb - 256Kb, 128Mb - 128Kb[ | 
|  | */ | 
|  | ret = test_add_free_space_entry(cache, SZ_128M - SZ_256K, SZ_128K, 0); | 
|  | if (ret) { | 
|  | test_err("couldn't add extent entry %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* Bitmap entry covering free space range [128Mb + 512Kb, 256Mb[ */ | 
|  | ret = test_add_free_space_entry(cache, SZ_128M + SZ_512K, | 
|  | SZ_128M - SZ_512K, 1); | 
|  | if (ret) { | 
|  | test_err("couldn't add bitmap entry %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | ret = check_num_extents_and_bitmaps(cache, 2, 1); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | /* | 
|  | * Now make only the first 256Kb of the bitmap marked as free, so that | 
|  | * we end up with only the following ranges marked as free space: | 
|  | * | 
|  | * [128Mb - 256Kb, 128Mb - 128Kb[ | 
|  | * [128Mb + 512Kb, 128Mb + 768Kb[ | 
|  | */ | 
|  | ret = btrfs_remove_free_space(cache, | 
|  | SZ_128M + 768 * SZ_1K, | 
|  | SZ_128M - 768 * SZ_1K); | 
|  | if (ret) { | 
|  | test_err("failed to free part of bitmap space %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* Confirm that only those 2 ranges are marked as free. */ | 
|  | if (!test_check_exists(cache, SZ_128M - SZ_256K, SZ_128K)) { | 
|  | test_err("free space range missing"); | 
|  | return -ENOENT; | 
|  | } | 
|  | if (!test_check_exists(cache, SZ_128M + SZ_512K, SZ_256K)) { | 
|  | test_err("free space range missing"); | 
|  | return -ENOENT; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Confirm that the bitmap range [128Mb + 768Kb, 256Mb[ isn't marked | 
|  | * as free anymore. | 
|  | */ | 
|  | if (test_check_exists(cache, SZ_128M + 768 * SZ_1K, | 
|  | SZ_128M - 768 * SZ_1K)) { | 
|  | test_err("bitmap region not removed from space cache"); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Confirm that the region [128Mb + 256Kb, 128Mb + 512Kb[, which is | 
|  | * covered by the bitmap, isn't marked as free. | 
|  | */ | 
|  | if (test_check_exists(cache, SZ_128M + SZ_256K, SZ_256K)) { | 
|  | test_err("invalid bitmap region marked as free"); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Confirm that the region [128Mb, 128Mb + 256Kb[, which is covered | 
|  | * by the bitmap too, isn't marked as free either. | 
|  | */ | 
|  | if (test_check_exists(cache, SZ_128M, SZ_256K)) { | 
|  | test_err("invalid bitmap region marked as free"); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Now lets mark the region [128Mb, 128Mb + 512Kb[ as free too. But, | 
|  | * lets make sure the free space cache marks it as free in the bitmap, | 
|  | * and doesn't insert a new extent entry to represent this region. | 
|  | */ | 
|  | ret = btrfs_add_free_space(cache, SZ_128M, SZ_512K); | 
|  | if (ret) { | 
|  | test_err("error adding free space: %d", ret); | 
|  | return ret; | 
|  | } | 
|  | /* Confirm the region is marked as free. */ | 
|  | if (!test_check_exists(cache, SZ_128M, SZ_512K)) { | 
|  | test_err("bitmap region not marked as free"); | 
|  | return -ENOENT; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Confirm that no new extent entries or bitmap entries were added to | 
|  | * the cache after adding that free space region. | 
|  | */ | 
|  | ret = check_num_extents_and_bitmaps(cache, 2, 1); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | /* | 
|  | * Now lets add a small free space region to the right of the previous | 
|  | * one, which is not contiguous with it and is part of the bitmap too. | 
|  | * The goal is to test that the bitmap entry space stealing doesn't | 
|  | * steal this space region. | 
|  | */ | 
|  | ret = btrfs_add_free_space(cache, SZ_128M + SZ_16M, sectorsize); | 
|  | if (ret) { | 
|  | test_err("error adding free space: %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Confirm that no new extent entries or bitmap entries were added to | 
|  | * the cache after adding that free space region. | 
|  | */ | 
|  | ret = check_num_extents_and_bitmaps(cache, 2, 1); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | /* | 
|  | * Now mark the region [128Mb - 128Kb, 128Mb[ as free too. This will | 
|  | * expand the range covered by the existing extent entry that represents | 
|  | * the free space [128Mb - 256Kb, 128Mb - 128Kb[. | 
|  | */ | 
|  | ret = btrfs_add_free_space(cache, SZ_128M - SZ_128K, SZ_128K); | 
|  | if (ret) { | 
|  | test_err("error adding free space: %d", ret); | 
|  | return ret; | 
|  | } | 
|  | /* Confirm the region is marked as free. */ | 
|  | if (!test_check_exists(cache, SZ_128M - SZ_128K, SZ_128K)) { | 
|  | test_err("extent region not marked as free"); | 
|  | return -ENOENT; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Confirm that our extent entry didn't stole all free space from the | 
|  | * bitmap, because of the small 4Kb free space region. | 
|  | */ | 
|  | ret = check_num_extents_and_bitmaps(cache, 2, 1); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | /* | 
|  | * So now we have the range [128Mb - 256Kb, 128Mb + 768Kb[ as free | 
|  | * space. Without stealing bitmap free space into extent entry space, | 
|  | * we would have all this free space represented by 2 entries in the | 
|  | * cache: | 
|  | * | 
|  | * extent entry covering range: [128Mb - 256Kb, 128Mb[ | 
|  | * bitmap entry covering range: [128Mb, 128Mb + 768Kb[ | 
|  | * | 
|  | * Attempting to allocate the whole free space (1Mb) would fail, because | 
|  | * we can't allocate from multiple entries. | 
|  | * With the bitmap free space stealing, we get a single extent entry | 
|  | * that represents the 1Mb free space, and therefore we're able to | 
|  | * allocate the whole free space at once. | 
|  | */ | 
|  | if (!test_check_exists(cache, SZ_128M - SZ_256K, SZ_1M)) { | 
|  | test_err("expected region not marked as free"); | 
|  | return -ENOENT; | 
|  | } | 
|  |  | 
|  | if (cache->free_space_ctl->free_space != (SZ_1M + sectorsize)) { | 
|  | test_err("cache free space is not 1Mb + %u", sectorsize); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | offset = btrfs_find_space_for_alloc(cache, | 
|  | 0, SZ_1M, 0, | 
|  | &max_extent_size); | 
|  | if (offset != (SZ_128M - SZ_256K)) { | 
|  | test_err( | 
|  | "failed to allocate 1Mb from space cache, returned offset is: %llu", | 
|  | offset); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * All that remains is a sectorsize free space region in a bitmap. | 
|  | * Confirm. | 
|  | */ | 
|  | ret = check_num_extents_and_bitmaps(cache, 1, 1); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | if (cache->free_space_ctl->free_space != sectorsize) { | 
|  | test_err("cache free space is not %u", sectorsize); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | offset = btrfs_find_space_for_alloc(cache, | 
|  | 0, sectorsize, 0, | 
|  | &max_extent_size); | 
|  | if (offset != (SZ_128M + SZ_16M)) { | 
|  | test_err("failed to allocate %u, returned offset : %llu", | 
|  | sectorsize, offset); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | ret = check_cache_empty(cache); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | btrfs_remove_free_space_cache(cache); | 
|  |  | 
|  | /* | 
|  | * Now test a similar scenario, but where our extent entry is located | 
|  | * to the right of the bitmap entry, so that we can check that stealing | 
|  | * space from a bitmap to the front of an extent entry works. | 
|  | */ | 
|  |  | 
|  | /* | 
|  | * Extent entry covering free space range [128Mb + 128Kb, 128Mb + 256Kb[ | 
|  | */ | 
|  | ret = test_add_free_space_entry(cache, SZ_128M + SZ_128K, SZ_128K, 0); | 
|  | if (ret) { | 
|  | test_err("couldn't add extent entry %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* Bitmap entry covering free space range [0, 128Mb - 512Kb[ */ | 
|  | ret = test_add_free_space_entry(cache, 0, SZ_128M - SZ_512K, 1); | 
|  | if (ret) { | 
|  | test_err("couldn't add bitmap entry %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | ret = check_num_extents_and_bitmaps(cache, 2, 1); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | /* | 
|  | * Now make only the last 256Kb of the bitmap marked as free, so that | 
|  | * we end up with only the following ranges marked as free space: | 
|  | * | 
|  | * [128Mb + 128b, 128Mb + 256Kb[ | 
|  | * [128Mb - 768Kb, 128Mb - 512Kb[ | 
|  | */ | 
|  | ret = btrfs_remove_free_space(cache, 0, SZ_128M - 768 * SZ_1K); | 
|  | if (ret) { | 
|  | test_err("failed to free part of bitmap space %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* Confirm that only those 2 ranges are marked as free. */ | 
|  | if (!test_check_exists(cache, SZ_128M + SZ_128K, SZ_128K)) { | 
|  | test_err("free space range missing"); | 
|  | return -ENOENT; | 
|  | } | 
|  | if (!test_check_exists(cache, SZ_128M - 768 * SZ_1K, SZ_256K)) { | 
|  | test_err("free space range missing"); | 
|  | return -ENOENT; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Confirm that the bitmap range [0, 128Mb - 768Kb[ isn't marked | 
|  | * as free anymore. | 
|  | */ | 
|  | if (test_check_exists(cache, 0, SZ_128M - 768 * SZ_1K)) { | 
|  | test_err("bitmap region not removed from space cache"); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Confirm that the region [128Mb - 512Kb, 128Mb[, which is | 
|  | * covered by the bitmap, isn't marked as free. | 
|  | */ | 
|  | if (test_check_exists(cache, SZ_128M - SZ_512K, SZ_512K)) { | 
|  | test_err("invalid bitmap region marked as free"); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Now lets mark the region [128Mb - 512Kb, 128Mb[ as free too. But, | 
|  | * lets make sure the free space cache marks it as free in the bitmap, | 
|  | * and doesn't insert a new extent entry to represent this region. | 
|  | */ | 
|  | ret = btrfs_add_free_space(cache, SZ_128M - SZ_512K, SZ_512K); | 
|  | if (ret) { | 
|  | test_err("error adding free space: %d", ret); | 
|  | return ret; | 
|  | } | 
|  | /* Confirm the region is marked as free. */ | 
|  | if (!test_check_exists(cache, SZ_128M - SZ_512K, SZ_512K)) { | 
|  | test_err("bitmap region not marked as free"); | 
|  | return -ENOENT; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Confirm that no new extent entries or bitmap entries were added to | 
|  | * the cache after adding that free space region. | 
|  | */ | 
|  | ret = check_num_extents_and_bitmaps(cache, 2, 1); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | /* | 
|  | * Now lets add a small free space region to the left of the previous | 
|  | * one, which is not contiguous with it and is part of the bitmap too. | 
|  | * The goal is to test that the bitmap entry space stealing doesn't | 
|  | * steal this space region. | 
|  | */ | 
|  | ret = btrfs_add_free_space(cache, SZ_32M, 2 * sectorsize); | 
|  | if (ret) { | 
|  | test_err("error adding free space: %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Now mark the region [128Mb, 128Mb + 128Kb[ as free too. This will | 
|  | * expand the range covered by the existing extent entry that represents | 
|  | * the free space [128Mb + 128Kb, 128Mb + 256Kb[. | 
|  | */ | 
|  | ret = btrfs_add_free_space(cache, SZ_128M, SZ_128K); | 
|  | if (ret) { | 
|  | test_err("error adding free space: %d", ret); | 
|  | return ret; | 
|  | } | 
|  | /* Confirm the region is marked as free. */ | 
|  | if (!test_check_exists(cache, SZ_128M, SZ_128K)) { | 
|  | test_err("extent region not marked as free"); | 
|  | return -ENOENT; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Confirm that our extent entry didn't stole all free space from the | 
|  | * bitmap, because of the small 2 * sectorsize free space region. | 
|  | */ | 
|  | ret = check_num_extents_and_bitmaps(cache, 2, 1); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | /* | 
|  | * So now we have the range [128Mb - 768Kb, 128Mb + 256Kb[ as free | 
|  | * space. Without stealing bitmap free space into extent entry space, | 
|  | * we would have all this free space represented by 2 entries in the | 
|  | * cache: | 
|  | * | 
|  | * extent entry covering range: [128Mb, 128Mb + 256Kb[ | 
|  | * bitmap entry covering range: [128Mb - 768Kb, 128Mb[ | 
|  | * | 
|  | * Attempting to allocate the whole free space (1Mb) would fail, because | 
|  | * we can't allocate from multiple entries. | 
|  | * With the bitmap free space stealing, we get a single extent entry | 
|  | * that represents the 1Mb free space, and therefore we're able to | 
|  | * allocate the whole free space at once. | 
|  | */ | 
|  | if (!test_check_exists(cache, SZ_128M - 768 * SZ_1K, SZ_1M)) { | 
|  | test_err("expected region not marked as free"); | 
|  | return -ENOENT; | 
|  | } | 
|  |  | 
|  | if (cache->free_space_ctl->free_space != (SZ_1M + 2 * sectorsize)) { | 
|  | test_err("cache free space is not 1Mb + %u", 2 * sectorsize); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | offset = btrfs_find_space_for_alloc(cache, 0, SZ_1M, 0, | 
|  | &max_extent_size); | 
|  | if (offset != (SZ_128M - 768 * SZ_1K)) { | 
|  | test_err( | 
|  | "failed to allocate 1Mb from space cache, returned offset is: %llu", | 
|  | offset); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * All that remains is 2 * sectorsize free space region | 
|  | * in a bitmap. Confirm. | 
|  | */ | 
|  | ret = check_num_extents_and_bitmaps(cache, 1, 1); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | if (cache->free_space_ctl->free_space != 2 * sectorsize) { | 
|  | test_err("cache free space is not %u", 2 * sectorsize); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | offset = btrfs_find_space_for_alloc(cache, | 
|  | 0, 2 * sectorsize, 0, | 
|  | &max_extent_size); | 
|  | if (offset != SZ_32M) { | 
|  | test_err("failed to allocate %u, offset: %llu", | 
|  | 2 * sectorsize, offset); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | ret = check_cache_empty(cache); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | cache->free_space_ctl->op = orig_free_space_ops; | 
|  | btrfs_remove_free_space_cache(cache); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static bool bytes_index_use_bitmap(struct btrfs_free_space_ctl *ctl, | 
|  | struct btrfs_free_space *info) | 
|  | { | 
|  | return true; | 
|  | } | 
|  |  | 
|  | static int test_bytes_index(struct btrfs_block_group *cache, u32 sectorsize) | 
|  | { | 
|  | const struct btrfs_free_space_op test_free_space_ops = { | 
|  | .use_bitmap = bytes_index_use_bitmap, | 
|  | }; | 
|  | const struct btrfs_free_space_op *orig_free_space_ops; | 
|  | struct btrfs_free_space_ctl *ctl = cache->free_space_ctl; | 
|  | struct btrfs_free_space *entry; | 
|  | struct rb_node *node; | 
|  | u64 offset, max_extent_size, bytes; | 
|  | int ret, i; | 
|  |  | 
|  | test_msg("running bytes index tests"); | 
|  |  | 
|  | /* First just validate that it does everything in order. */ | 
|  | offset = 0; | 
|  | for (i = 0; i < 10; i++) { | 
|  | bytes = (i + 1) * SZ_1M; | 
|  | ret = test_add_free_space_entry(cache, offset, bytes, 0); | 
|  | if (ret) { | 
|  | test_err("couldn't add extent entry %d\n", ret); | 
|  | return ret; | 
|  | } | 
|  | offset += bytes + sectorsize; | 
|  | } | 
|  |  | 
|  | for (node = rb_first_cached(&ctl->free_space_bytes), i = 9; node; | 
|  | node = rb_next(node), i--) { | 
|  | entry = rb_entry(node, struct btrfs_free_space, bytes_index); | 
|  | bytes = (i + 1) * SZ_1M; | 
|  | if (entry->bytes != bytes) { | 
|  | test_err("invalid bytes index order, found %llu expected %llu", | 
|  | entry->bytes, bytes); | 
|  | return -EINVAL; | 
|  | } | 
|  | } | 
|  |  | 
|  | /* Now validate bitmaps do the correct thing. */ | 
|  | btrfs_remove_free_space_cache(cache); | 
|  | for (i = 0; i < 2; i++) { | 
|  | offset = i * BITS_PER_BITMAP * sectorsize; | 
|  | bytes = (i + 1) * SZ_1M; | 
|  | ret = test_add_free_space_entry(cache, offset, bytes, 1); | 
|  | if (ret) { | 
|  | test_err("couldn't add bitmap entry"); | 
|  | return ret; | 
|  | } | 
|  | } | 
|  |  | 
|  | for (node = rb_first_cached(&ctl->free_space_bytes), i = 1; node; | 
|  | node = rb_next(node), i--) { | 
|  | entry = rb_entry(node, struct btrfs_free_space, bytes_index); | 
|  | bytes = (i + 1) * SZ_1M; | 
|  | if (entry->bytes != bytes) { | 
|  | test_err("invalid bytes index order, found %llu expected %llu", | 
|  | entry->bytes, bytes); | 
|  | return -EINVAL; | 
|  | } | 
|  | } | 
|  |  | 
|  | /* Now validate bitmaps with different ->max_extent_size. */ | 
|  | btrfs_remove_free_space_cache(cache); | 
|  | orig_free_space_ops = cache->free_space_ctl->op; | 
|  | cache->free_space_ctl->op = &test_free_space_ops; | 
|  |  | 
|  | ret = test_add_free_space_entry(cache, 0, sectorsize, 1); | 
|  | if (ret) { | 
|  | test_err("couldn't add bitmap entry"); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | offset = BITS_PER_BITMAP * sectorsize; | 
|  | ret = test_add_free_space_entry(cache, offset, sectorsize, 1); | 
|  | if (ret) { | 
|  | test_err("couldn't add bitmap_entry"); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Now set a bunch of sectorsize extents in the first entry so it's | 
|  | * ->bytes is large. | 
|  | */ | 
|  | for (i = 2; i < 20; i += 2) { | 
|  | offset = sectorsize * i; | 
|  | ret = btrfs_add_free_space(cache, offset, sectorsize); | 
|  | if (ret) { | 
|  | test_err("error populating sparse bitmap %d", ret); | 
|  | return ret; | 
|  | } | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Now set a contiguous extent in the second bitmap so its | 
|  | * ->max_extent_size is larger than the first bitmaps. | 
|  | */ | 
|  | offset = (BITS_PER_BITMAP * sectorsize) + sectorsize; | 
|  | ret = btrfs_add_free_space(cache, offset, sectorsize); | 
|  | if (ret) { | 
|  | test_err("error adding contiguous extent %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Since we don't set ->max_extent_size unless we search everything | 
|  | * should be indexed on bytes. | 
|  | */ | 
|  | entry = rb_entry(rb_first_cached(&ctl->free_space_bytes), | 
|  | struct btrfs_free_space, bytes_index); | 
|  | if (entry->bytes != (10 * sectorsize)) { | 
|  | test_err("error, wrong entry in the first slot in bytes_index"); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | max_extent_size = 0; | 
|  | offset = btrfs_find_space_for_alloc(cache, cache->start, sectorsize * 3, | 
|  | 0, &max_extent_size); | 
|  | if (offset != 0) { | 
|  | test_err("found space to alloc even though we don't have enough space"); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | if (max_extent_size != (2 * sectorsize)) { | 
|  | test_err("got the wrong max_extent size %llu expected %llu", | 
|  | max_extent_size, (unsigned long long)(2 * sectorsize)); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * The search should have re-arranged the bytes index to use the | 
|  | * ->max_extent_size, validate it's now what we expect it to be. | 
|  | */ | 
|  | entry = rb_entry(rb_first_cached(&ctl->free_space_bytes), | 
|  | struct btrfs_free_space, bytes_index); | 
|  | if (entry->bytes != (2 * sectorsize)) { | 
|  | test_err("error, the bytes index wasn't recalculated properly"); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | /* Add another sectorsize to re-arrange the tree back to ->bytes. */ | 
|  | offset = (BITS_PER_BITMAP * sectorsize) - sectorsize; | 
|  | ret = btrfs_add_free_space(cache, offset, sectorsize); | 
|  | if (ret) { | 
|  | test_err("error adding extent to the sparse entry %d", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | entry = rb_entry(rb_first_cached(&ctl->free_space_bytes), | 
|  | struct btrfs_free_space, bytes_index); | 
|  | if (entry->bytes != (11 * sectorsize)) { | 
|  | test_err("error, wrong entry in the first slot in bytes_index"); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Now make sure we find our correct entry after searching that will | 
|  | * result in a re-arranging of the tree. | 
|  | */ | 
|  | max_extent_size = 0; | 
|  | offset = btrfs_find_space_for_alloc(cache, cache->start, sectorsize * 2, | 
|  | 0, &max_extent_size); | 
|  | if (offset != (BITS_PER_BITMAP * sectorsize)) { | 
|  | test_err("error, found %llu instead of %llu for our alloc", | 
|  | offset, | 
|  | (unsigned long long)(BITS_PER_BITMAP * sectorsize)); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | cache->free_space_ctl->op = orig_free_space_ops; | 
|  | btrfs_remove_free_space_cache(cache); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | int btrfs_test_free_space_cache(u32 sectorsize, u32 nodesize) | 
|  | { | 
|  | struct btrfs_fs_info *fs_info; | 
|  | struct btrfs_block_group *cache; | 
|  | struct btrfs_root *root = NULL; | 
|  | int ret = -ENOMEM; | 
|  |  | 
|  | test_msg("running btrfs free space cache tests"); | 
|  | fs_info = btrfs_alloc_dummy_fs_info(nodesize, sectorsize); | 
|  | if (!fs_info) { | 
|  | test_std_err(TEST_ALLOC_FS_INFO); | 
|  | return -ENOMEM; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * For ppc64 (with 64k page size), bytes per bitmap might be | 
|  | * larger than 1G.  To make bitmap test available in ppc64, | 
|  | * alloc dummy block group whose size cross bitmaps. | 
|  | */ | 
|  | cache = btrfs_alloc_dummy_block_group(fs_info, | 
|  | BITS_PER_BITMAP * sectorsize + PAGE_SIZE); | 
|  | if (!cache) { | 
|  | test_std_err(TEST_ALLOC_BLOCK_GROUP); | 
|  | btrfs_free_dummy_fs_info(fs_info); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | root = btrfs_alloc_dummy_root(fs_info); | 
|  | if (IS_ERR(root)) { | 
|  | test_std_err(TEST_ALLOC_ROOT); | 
|  | ret = PTR_ERR(root); | 
|  | goto out; | 
|  | } | 
|  |  | 
|  | root->root_key.objectid = BTRFS_EXTENT_TREE_OBJECTID; | 
|  | root->root_key.type = BTRFS_ROOT_ITEM_KEY; | 
|  | root->root_key.offset = 0; | 
|  | btrfs_global_root_insert(root); | 
|  |  | 
|  | ret = test_extents(cache); | 
|  | if (ret) | 
|  | goto out; | 
|  | ret = test_bitmaps(cache, sectorsize); | 
|  | if (ret) | 
|  | goto out; | 
|  | ret = test_bitmaps_and_extents(cache, sectorsize); | 
|  | if (ret) | 
|  | goto out; | 
|  |  | 
|  | ret = test_steal_space_from_bitmap_to_extent(cache, sectorsize); | 
|  | if (ret) | 
|  | goto out; | 
|  | ret = test_bytes_index(cache, sectorsize); | 
|  | out: | 
|  | btrfs_free_dummy_block_group(cache); | 
|  | btrfs_free_dummy_root(root); | 
|  | btrfs_free_dummy_fs_info(fs_info); | 
|  | return ret; | 
|  | } |