|  | // SPDX-License-Identifier: GPL-2.0-or-later | 
|  | /* | 
|  | * Squashfs - a compressed read only filesystem for Linux | 
|  | * | 
|  | * Copyright (c) 2002, 2003, 2004, 2005, 2006, 2007, 2008 | 
|  | * Phillip Lougher <phillip@squashfs.org.uk> | 
|  | * | 
|  | * block.c | 
|  | */ | 
|  |  | 
|  | /* | 
|  | * This file implements the low-level routines to read and decompress | 
|  | * datablocks and metadata blocks. | 
|  | */ | 
|  |  | 
|  | #include <linux/blkdev.h> | 
|  | #include <linux/fs.h> | 
|  | #include <linux/vfs.h> | 
|  | #include <linux/slab.h> | 
|  | #include <linux/pagemap.h> | 
|  | #include <linux/string.h> | 
|  | #include <linux/bio.h> | 
|  |  | 
|  | #include "squashfs_fs.h" | 
|  | #include "squashfs_fs_sb.h" | 
|  | #include "squashfs.h" | 
|  | #include "decompressor.h" | 
|  | #include "page_actor.h" | 
|  |  | 
|  | /* | 
|  | * Returns the amount of bytes copied to the page actor. | 
|  | */ | 
|  | static int copy_bio_to_actor(struct bio *bio, | 
|  | struct squashfs_page_actor *actor, | 
|  | int offset, int req_length) | 
|  | { | 
|  | void *actor_addr; | 
|  | struct bvec_iter_all iter_all = {}; | 
|  | struct bio_vec *bvec = bvec_init_iter_all(&iter_all); | 
|  | int copied_bytes = 0; | 
|  | int actor_offset = 0; | 
|  |  | 
|  | squashfs_actor_nobuff(actor); | 
|  | actor_addr = squashfs_first_page(actor); | 
|  |  | 
|  | if (WARN_ON_ONCE(!bio_next_segment(bio, &iter_all))) | 
|  | return 0; | 
|  |  | 
|  | while (copied_bytes < req_length) { | 
|  | int bytes_to_copy = min_t(int, bvec->bv_len - offset, | 
|  | PAGE_SIZE - actor_offset); | 
|  |  | 
|  | bytes_to_copy = min_t(int, bytes_to_copy, | 
|  | req_length - copied_bytes); | 
|  | if (!IS_ERR(actor_addr)) | 
|  | memcpy(actor_addr + actor_offset, bvec_virt(bvec) + | 
|  | offset, bytes_to_copy); | 
|  |  | 
|  | actor_offset += bytes_to_copy; | 
|  | copied_bytes += bytes_to_copy; | 
|  | offset += bytes_to_copy; | 
|  |  | 
|  | if (actor_offset >= PAGE_SIZE) { | 
|  | actor_addr = squashfs_next_page(actor); | 
|  | if (!actor_addr) | 
|  | break; | 
|  | actor_offset = 0; | 
|  | } | 
|  | if (offset >= bvec->bv_len) { | 
|  | if (!bio_next_segment(bio, &iter_all)) | 
|  | break; | 
|  | offset = 0; | 
|  | } | 
|  | } | 
|  | squashfs_finish_page(actor); | 
|  | return copied_bytes; | 
|  | } | 
|  |  | 
|  | static int squashfs_bio_read_cached(struct bio *fullbio, | 
|  | struct address_space *cache_mapping, u64 index, int length, | 
|  | u64 read_start, u64 read_end, int page_count) | 
|  | { | 
|  | struct folio *head_to_cache = NULL, *tail_to_cache = NULL; | 
|  | struct block_device *bdev = fullbio->bi_bdev; | 
|  | int start_idx = 0, end_idx = 0; | 
|  | struct folio_iter fi; | 
|  | struct bio *bio = NULL; | 
|  | int idx = 0; | 
|  | int err = 0; | 
|  | #ifdef CONFIG_SQUASHFS_COMP_CACHE_FULL | 
|  | struct folio **cache_folios = kmalloc_array(page_count, | 
|  | sizeof(*cache_folios), GFP_KERNEL | __GFP_ZERO); | 
|  | #endif | 
|  |  | 
|  | bio_for_each_folio_all(fi, fullbio) { | 
|  | struct folio *folio = fi.folio; | 
|  |  | 
|  | if (folio->mapping == cache_mapping) { | 
|  | idx++; | 
|  | continue; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * We only use this when the device block size is the same as | 
|  | * the page size, so read_start and read_end cover full pages. | 
|  | * | 
|  | * Compare these to the original required index and length to | 
|  | * only cache pages which were requested partially, since these | 
|  | * are the ones which are likely to be needed when reading | 
|  | * adjacent blocks. | 
|  | */ | 
|  | if (idx == 0 && index != read_start) | 
|  | head_to_cache = folio; | 
|  | else if (idx == page_count - 1 && index + length != read_end) | 
|  | tail_to_cache = folio; | 
|  | #ifdef CONFIG_SQUASHFS_COMP_CACHE_FULL | 
|  | /* Cache all pages in the BIO for repeated reads */ | 
|  | else if (cache_folios) | 
|  | cache_folios[idx] = folio; | 
|  | #endif | 
|  |  | 
|  | if (!bio || idx != end_idx) { | 
|  | struct bio *new = bio_alloc_clone(bdev, fullbio, | 
|  | GFP_NOIO, &fs_bio_set); | 
|  |  | 
|  | if (bio) { | 
|  | bio_trim(bio, start_idx * PAGE_SECTORS, | 
|  | (end_idx - start_idx) * PAGE_SECTORS); | 
|  | bio_chain(bio, new); | 
|  | submit_bio(bio); | 
|  | } | 
|  |  | 
|  | bio = new; | 
|  | start_idx = idx; | 
|  | } | 
|  |  | 
|  | idx++; | 
|  | end_idx = idx; | 
|  | } | 
|  |  | 
|  | if (bio) { | 
|  | bio_trim(bio, start_idx * PAGE_SECTORS, | 
|  | (end_idx - start_idx) * PAGE_SECTORS); | 
|  | err = submit_bio_wait(bio); | 
|  | bio_put(bio); | 
|  | } | 
|  |  | 
|  | if (err) | 
|  | return err; | 
|  |  | 
|  | if (head_to_cache) { | 
|  | int ret = filemap_add_folio(cache_mapping, head_to_cache, | 
|  | read_start >> PAGE_SHIFT, | 
|  | GFP_NOIO); | 
|  |  | 
|  | if (!ret) { | 
|  | folio_mark_uptodate(head_to_cache); | 
|  | folio_unlock(head_to_cache); | 
|  | } | 
|  |  | 
|  | } | 
|  |  | 
|  | if (tail_to_cache) { | 
|  | int ret = filemap_add_folio(cache_mapping, tail_to_cache, | 
|  | (read_end >> PAGE_SHIFT) - 1, | 
|  | GFP_NOIO); | 
|  |  | 
|  | if (!ret) { | 
|  | folio_mark_uptodate(tail_to_cache); | 
|  | folio_unlock(tail_to_cache); | 
|  | } | 
|  | } | 
|  |  | 
|  | #ifdef CONFIG_SQUASHFS_COMP_CACHE_FULL | 
|  | if (!cache_folios) | 
|  | goto out; | 
|  |  | 
|  | for (idx = 0; idx < page_count; idx++) { | 
|  | if (!cache_folios[idx]) | 
|  | continue; | 
|  | int ret = filemap_add_folio(cache_mapping, cache_folios[idx], | 
|  | (read_start >> PAGE_SHIFT) + idx, | 
|  | GFP_NOIO); | 
|  |  | 
|  | if (!ret) { | 
|  | folio_mark_uptodate(cache_folios[idx]); | 
|  | folio_unlock(cache_folios[idx]); | 
|  | } | 
|  | } | 
|  | kfree(cache_folios); | 
|  | out: | 
|  | #endif | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static struct page *squashfs_get_cache_page(struct address_space *mapping, | 
|  | pgoff_t index) | 
|  | { | 
|  | struct page *page; | 
|  |  | 
|  | if (!mapping) | 
|  | return NULL; | 
|  |  | 
|  | page = find_get_page(mapping, index); | 
|  | if (!page) | 
|  | return NULL; | 
|  |  | 
|  | if (!PageUptodate(page)) { | 
|  | put_page(page); | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | return page; | 
|  | } | 
|  |  | 
|  | static int squashfs_bio_read(struct super_block *sb, u64 index, int length, | 
|  | struct bio **biop, int *block_offset) | 
|  | { | 
|  | struct squashfs_sb_info *msblk = sb->s_fs_info; | 
|  | struct address_space *cache_mapping = msblk->cache_mapping; | 
|  | const u64 read_start = round_down(index, msblk->devblksize); | 
|  | const sector_t block = read_start >> msblk->devblksize_log2; | 
|  | const u64 read_end = round_up(index + length, msblk->devblksize); | 
|  | const sector_t block_end = read_end >> msblk->devblksize_log2; | 
|  | int offset = read_start - round_down(index, PAGE_SIZE); | 
|  | int total_len = (block_end - block) << msblk->devblksize_log2; | 
|  | const int page_count = DIV_ROUND_UP(total_len + offset, PAGE_SIZE); | 
|  | int error, i; | 
|  | struct bio *bio; | 
|  |  | 
|  | bio = bio_kmalloc(page_count, GFP_NOIO); | 
|  | if (!bio) | 
|  | return -ENOMEM; | 
|  | bio_init(bio, sb->s_bdev, bio->bi_inline_vecs, page_count, REQ_OP_READ); | 
|  | bio->bi_iter.bi_sector = block * (msblk->devblksize >> SECTOR_SHIFT); | 
|  |  | 
|  | for (i = 0; i < page_count; ++i) { | 
|  | unsigned int len = | 
|  | min_t(unsigned int, PAGE_SIZE - offset, total_len); | 
|  | pgoff_t index = (read_start >> PAGE_SHIFT) + i; | 
|  | struct page *page; | 
|  |  | 
|  | page = squashfs_get_cache_page(cache_mapping, index); | 
|  | if (!page) | 
|  | page = alloc_page(GFP_NOIO); | 
|  |  | 
|  | if (!page) { | 
|  | error = -ENOMEM; | 
|  | goto out_free_bio; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Use the __ version to avoid merging since we need each page | 
|  | * to be separate when we check for and avoid cached pages. | 
|  | */ | 
|  | __bio_add_page(bio, page, len, offset); | 
|  | offset = 0; | 
|  | total_len -= len; | 
|  | } | 
|  |  | 
|  | if (cache_mapping) | 
|  | error = squashfs_bio_read_cached(bio, cache_mapping, index, | 
|  | length, read_start, read_end, | 
|  | page_count); | 
|  | else | 
|  | error = submit_bio_wait(bio); | 
|  | if (error) | 
|  | goto out_free_bio; | 
|  |  | 
|  | *biop = bio; | 
|  | *block_offset = index & ((1 << msblk->devblksize_log2) - 1); | 
|  | return 0; | 
|  |  | 
|  | out_free_bio: | 
|  | bio_free_pages(bio); | 
|  | bio_uninit(bio); | 
|  | kfree(bio); | 
|  | return error; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Read and decompress a metadata block or datablock.  Length is non-zero | 
|  | * if a datablock is being read (the size is stored elsewhere in the | 
|  | * filesystem), otherwise the length is obtained from the first two bytes of | 
|  | * the metadata block.  A bit in the length field indicates if the block | 
|  | * is stored uncompressed in the filesystem (usually because compression | 
|  | * generated a larger block - this does occasionally happen with compression | 
|  | * algorithms). | 
|  | */ | 
|  | int squashfs_read_data(struct super_block *sb, u64 index, int length, | 
|  | u64 *next_index, struct squashfs_page_actor *output) | 
|  | { | 
|  | struct squashfs_sb_info *msblk = sb->s_fs_info; | 
|  | struct bio *bio = NULL; | 
|  | int compressed; | 
|  | int res; | 
|  | int offset; | 
|  |  | 
|  | if (length) { | 
|  | /* | 
|  | * Datablock. | 
|  | */ | 
|  | compressed = SQUASHFS_COMPRESSED_BLOCK(length); | 
|  | length = SQUASHFS_COMPRESSED_SIZE_BLOCK(length); | 
|  | TRACE("Block @ 0x%llx, %scompressed size %d, src size %d\n", | 
|  | index, compressed ? "" : "un", length, output->length); | 
|  | } else { | 
|  | /* | 
|  | * Metadata block. | 
|  | */ | 
|  | const u8 *data; | 
|  | struct bvec_iter_all iter_all = {}; | 
|  | struct bio_vec *bvec = bvec_init_iter_all(&iter_all); | 
|  |  | 
|  | if (index + 2 > msblk->bytes_used) { | 
|  | res = -EIO; | 
|  | goto out; | 
|  | } | 
|  | res = squashfs_bio_read(sb, index, 2, &bio, &offset); | 
|  | if (res) | 
|  | goto out; | 
|  |  | 
|  | if (WARN_ON_ONCE(!bio_next_segment(bio, &iter_all))) { | 
|  | res = -EIO; | 
|  | goto out_free_bio; | 
|  | } | 
|  | /* Extract the length of the metadata block */ | 
|  | data = bvec_virt(bvec); | 
|  | length = data[offset]; | 
|  | if (offset < bvec->bv_len - 1) { | 
|  | length |= data[offset + 1] << 8; | 
|  | } else { | 
|  | if (WARN_ON_ONCE(!bio_next_segment(bio, &iter_all))) { | 
|  | res = -EIO; | 
|  | goto out_free_bio; | 
|  | } | 
|  | data = bvec_virt(bvec); | 
|  | length |= data[0] << 8; | 
|  | } | 
|  | bio_free_pages(bio); | 
|  | bio_uninit(bio); | 
|  | kfree(bio); | 
|  |  | 
|  | compressed = SQUASHFS_COMPRESSED(length); | 
|  | length = SQUASHFS_COMPRESSED_SIZE(length); | 
|  | index += 2; | 
|  |  | 
|  | TRACE("Block @ 0x%llx, %scompressed size %d\n", index - 2, | 
|  | compressed ? "" : "un", length); | 
|  | } | 
|  | if (length <= 0 || length > output->length || | 
|  | (index + length) > msblk->bytes_used) { | 
|  | res = -EIO; | 
|  | goto out; | 
|  | } | 
|  |  | 
|  | if (next_index) | 
|  | *next_index = index + length; | 
|  |  | 
|  | res = squashfs_bio_read(sb, index, length, &bio, &offset); | 
|  | if (res) | 
|  | goto out; | 
|  |  | 
|  | if (compressed) { | 
|  | if (!msblk->stream) { | 
|  | res = -EIO; | 
|  | goto out_free_bio; | 
|  | } | 
|  | res = msblk->thread_ops->decompress(msblk, bio, offset, length, output); | 
|  | } else { | 
|  | res = copy_bio_to_actor(bio, output, offset, length); | 
|  | } | 
|  |  | 
|  | out_free_bio: | 
|  | bio_free_pages(bio); | 
|  | bio_uninit(bio); | 
|  | kfree(bio); | 
|  | out: | 
|  | if (res < 0) { | 
|  | ERROR("Failed to read block 0x%llx: %d\n", index, res); | 
|  | if (msblk->panic_on_errors) | 
|  | panic("squashfs read failed"); | 
|  | } | 
|  |  | 
|  | return res; | 
|  | } |