360 lines
		
	
	
		
			8.6 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			360 lines
		
	
	
		
			8.6 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0-or-later
 | |
| /*
 | |
|  * Squashfs - a compressed read only filesystem for Linux
 | |
|  *
 | |
|  * Copyright (c) 2002, 2003, 2004, 2005, 2006, 2007, 2008
 | |
|  * Phillip Lougher <phillip@squashfs.org.uk>
 | |
|  *
 | |
|  * block.c
 | |
|  */
 | |
| 
 | |
| /*
 | |
|  * This file implements the low-level routines to read and decompress
 | |
|  * datablocks and metadata blocks.
 | |
|  */
 | |
| 
 | |
| #include <linux/blkdev.h>
 | |
| #include <linux/fs.h>
 | |
| #include <linux/vfs.h>
 | |
| #include <linux/slab.h>
 | |
| #include <linux/pagemap.h>
 | |
| #include <linux/string.h>
 | |
| #include <linux/bio.h>
 | |
| 
 | |
| #include "squashfs_fs.h"
 | |
| #include "squashfs_fs_sb.h"
 | |
| #include "squashfs.h"
 | |
| #include "decompressor.h"
 | |
| #include "page_actor.h"
 | |
| 
 | |
| /*
 | |
|  * Returns the amount of bytes copied to the page actor.
 | |
|  */
 | |
| static int copy_bio_to_actor(struct bio *bio,
 | |
| 			     struct squashfs_page_actor *actor,
 | |
| 			     int offset, int req_length)
 | |
| {
 | |
| 	void *actor_addr;
 | |
| 	struct bvec_iter_all iter_all = {};
 | |
| 	struct bio_vec *bvec = bvec_init_iter_all(&iter_all);
 | |
| 	int copied_bytes = 0;
 | |
| 	int actor_offset = 0;
 | |
| 
 | |
| 	squashfs_actor_nobuff(actor);
 | |
| 	actor_addr = squashfs_first_page(actor);
 | |
| 
 | |
| 	if (WARN_ON_ONCE(!bio_next_segment(bio, &iter_all)))
 | |
| 		return 0;
 | |
| 
 | |
| 	while (copied_bytes < req_length) {
 | |
| 		int bytes_to_copy = min_t(int, bvec->bv_len - offset,
 | |
| 					  PAGE_SIZE - actor_offset);
 | |
| 
 | |
| 		bytes_to_copy = min_t(int, bytes_to_copy,
 | |
| 				      req_length - copied_bytes);
 | |
| 		if (!IS_ERR(actor_addr))
 | |
| 			memcpy(actor_addr + actor_offset, bvec_virt(bvec) +
 | |
| 					offset, bytes_to_copy);
 | |
| 
 | |
| 		actor_offset += bytes_to_copy;
 | |
| 		copied_bytes += bytes_to_copy;
 | |
| 		offset += bytes_to_copy;
 | |
| 
 | |
| 		if (actor_offset >= PAGE_SIZE) {
 | |
| 			actor_addr = squashfs_next_page(actor);
 | |
| 			if (!actor_addr)
 | |
| 				break;
 | |
| 			actor_offset = 0;
 | |
| 		}
 | |
| 		if (offset >= bvec->bv_len) {
 | |
| 			if (!bio_next_segment(bio, &iter_all))
 | |
| 				break;
 | |
| 			offset = 0;
 | |
| 		}
 | |
| 	}
 | |
| 	squashfs_finish_page(actor);
 | |
| 	return copied_bytes;
 | |
| }
 | |
| 
 | |
| static int squashfs_bio_read_cached(struct bio *fullbio,
 | |
| 		struct address_space *cache_mapping, u64 index, int length,
 | |
| 		u64 read_start, u64 read_end, int page_count)
 | |
| {
 | |
| 	struct page *head_to_cache = NULL, *tail_to_cache = NULL;
 | |
| 	struct block_device *bdev = fullbio->bi_bdev;
 | |
| 	int start_idx = 0, end_idx = 0;
 | |
| 	struct bvec_iter_all iter_all;
 | |
| 	struct bio *bio = NULL;
 | |
| 	struct bio_vec *bv;
 | |
| 	int idx = 0;
 | |
| 	int err = 0;
 | |
| 
 | |
| 	bio_for_each_segment_all(bv, fullbio, iter_all) {
 | |
| 		struct page *page = bv->bv_page;
 | |
| 
 | |
| 		if (page->mapping == cache_mapping) {
 | |
| 			idx++;
 | |
| 			continue;
 | |
| 		}
 | |
| 
 | |
| 		/*
 | |
| 		 * We only use this when the device block size is the same as
 | |
| 		 * the page size, so read_start and read_end cover full pages.
 | |
| 		 *
 | |
| 		 * Compare these to the original required index and length to
 | |
| 		 * only cache pages which were requested partially, since these
 | |
| 		 * are the ones which are likely to be needed when reading
 | |
| 		 * adjacent blocks.
 | |
| 		 */
 | |
| 		if (idx == 0 && index != read_start)
 | |
| 			head_to_cache = page;
 | |
| 		else if (idx == page_count - 1 && index + length != read_end)
 | |
| 			tail_to_cache = page;
 | |
| 
 | |
| 		if (!bio || idx != end_idx) {
 | |
| 			struct bio *new = bio_alloc_clone(bdev, fullbio,
 | |
| 							  GFP_NOIO, &fs_bio_set);
 | |
| 
 | |
| 			if (bio) {
 | |
| 				bio_trim(bio, start_idx * PAGE_SECTORS,
 | |
| 					 (end_idx - start_idx) * PAGE_SECTORS);
 | |
| 				bio_chain(bio, new);
 | |
| 				submit_bio(bio);
 | |
| 			}
 | |
| 
 | |
| 			bio = new;
 | |
| 			start_idx = idx;
 | |
| 		}
 | |
| 
 | |
| 		idx++;
 | |
| 		end_idx = idx;
 | |
| 	}
 | |
| 
 | |
| 	if (bio) {
 | |
| 		bio_trim(bio, start_idx * PAGE_SECTORS,
 | |
| 			 (end_idx - start_idx) * PAGE_SECTORS);
 | |
| 		err = submit_bio_wait(bio);
 | |
| 		bio_put(bio);
 | |
| 	}
 | |
| 
 | |
| 	if (err)
 | |
| 		return err;
 | |
| 
 | |
| 	if (head_to_cache) {
 | |
| 		int ret = add_to_page_cache_lru(head_to_cache, cache_mapping,
 | |
| 						read_start >> PAGE_SHIFT,
 | |
| 						GFP_NOIO);
 | |
| 
 | |
| 		if (!ret) {
 | |
| 			SetPageUptodate(head_to_cache);
 | |
| 			unlock_page(head_to_cache);
 | |
| 		}
 | |
| 
 | |
| 	}
 | |
| 
 | |
| 	if (tail_to_cache) {
 | |
| 		int ret = add_to_page_cache_lru(tail_to_cache, cache_mapping,
 | |
| 						(read_end >> PAGE_SHIFT) - 1,
 | |
| 						GFP_NOIO);
 | |
| 
 | |
| 		if (!ret) {
 | |
| 			SetPageUptodate(tail_to_cache);
 | |
| 			unlock_page(tail_to_cache);
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static struct page *squashfs_get_cache_page(struct address_space *mapping,
 | |
| 					    pgoff_t index)
 | |
| {
 | |
| 	struct page *page;
 | |
| 
 | |
| 	if (!mapping)
 | |
| 		return NULL;
 | |
| 
 | |
| 	page = find_get_page(mapping, index);
 | |
| 	if (!page)
 | |
| 		return NULL;
 | |
| 
 | |
| 	if (!PageUptodate(page)) {
 | |
| 		put_page(page);
 | |
| 		return NULL;
 | |
| 	}
 | |
| 
 | |
| 	return page;
 | |
| }
 | |
| 
 | |
| static int squashfs_bio_read(struct super_block *sb, u64 index, int length,
 | |
| 			     struct bio **biop, int *block_offset)
 | |
| {
 | |
| 	struct squashfs_sb_info *msblk = sb->s_fs_info;
 | |
| 	struct address_space *cache_mapping = msblk->cache_mapping;
 | |
| 	const u64 read_start = round_down(index, msblk->devblksize);
 | |
| 	const sector_t block = read_start >> msblk->devblksize_log2;
 | |
| 	const u64 read_end = round_up(index + length, msblk->devblksize);
 | |
| 	const sector_t block_end = read_end >> msblk->devblksize_log2;
 | |
| 	int offset = read_start - round_down(index, PAGE_SIZE);
 | |
| 	int total_len = (block_end - block) << msblk->devblksize_log2;
 | |
| 	const int page_count = DIV_ROUND_UP(total_len + offset, PAGE_SIZE);
 | |
| 	int error, i;
 | |
| 	struct bio *bio;
 | |
| 
 | |
| 	bio = bio_kmalloc(page_count, GFP_NOIO);
 | |
| 	if (!bio)
 | |
| 		return -ENOMEM;
 | |
| 	bio_init(bio, sb->s_bdev, bio->bi_inline_vecs, page_count, REQ_OP_READ);
 | |
| 	bio->bi_iter.bi_sector = block * (msblk->devblksize >> SECTOR_SHIFT);
 | |
| 
 | |
| 	for (i = 0; i < page_count; ++i) {
 | |
| 		unsigned int len =
 | |
| 			min_t(unsigned int, PAGE_SIZE - offset, total_len);
 | |
| 		pgoff_t index = (read_start >> PAGE_SHIFT) + i;
 | |
| 		struct page *page;
 | |
| 
 | |
| 		page = squashfs_get_cache_page(cache_mapping, index);
 | |
| 		if (!page)
 | |
| 			page = alloc_page(GFP_NOIO);
 | |
| 
 | |
| 		if (!page) {
 | |
| 			error = -ENOMEM;
 | |
| 			goto out_free_bio;
 | |
| 		}
 | |
| 
 | |
| 		/*
 | |
| 		 * Use the __ version to avoid merging since we need each page
 | |
| 		 * to be separate when we check for and avoid cached pages.
 | |
| 		 */
 | |
| 		__bio_add_page(bio, page, len, offset);
 | |
| 		offset = 0;
 | |
| 		total_len -= len;
 | |
| 	}
 | |
| 
 | |
| 	if (cache_mapping)
 | |
| 		error = squashfs_bio_read_cached(bio, cache_mapping, index,
 | |
| 						 length, read_start, read_end,
 | |
| 						 page_count);
 | |
| 	else
 | |
| 		error = submit_bio_wait(bio);
 | |
| 	if (error)
 | |
| 		goto out_free_bio;
 | |
| 
 | |
| 	*biop = bio;
 | |
| 	*block_offset = index & ((1 << msblk->devblksize_log2) - 1);
 | |
| 	return 0;
 | |
| 
 | |
| out_free_bio:
 | |
| 	bio_free_pages(bio);
 | |
| 	bio_uninit(bio);
 | |
| 	kfree(bio);
 | |
| 	return error;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Read and decompress a metadata block or datablock.  Length is non-zero
 | |
|  * if a datablock is being read (the size is stored elsewhere in the
 | |
|  * filesystem), otherwise the length is obtained from the first two bytes of
 | |
|  * the metadata block.  A bit in the length field indicates if the block
 | |
|  * is stored uncompressed in the filesystem (usually because compression
 | |
|  * generated a larger block - this does occasionally happen with compression
 | |
|  * algorithms).
 | |
|  */
 | |
| int squashfs_read_data(struct super_block *sb, u64 index, int length,
 | |
| 		       u64 *next_index, struct squashfs_page_actor *output)
 | |
| {
 | |
| 	struct squashfs_sb_info *msblk = sb->s_fs_info;
 | |
| 	struct bio *bio = NULL;
 | |
| 	int compressed;
 | |
| 	int res;
 | |
| 	int offset;
 | |
| 
 | |
| 	if (length) {
 | |
| 		/*
 | |
| 		 * Datablock.
 | |
| 		 */
 | |
| 		compressed = SQUASHFS_COMPRESSED_BLOCK(length);
 | |
| 		length = SQUASHFS_COMPRESSED_SIZE_BLOCK(length);
 | |
| 		TRACE("Block @ 0x%llx, %scompressed size %d, src size %d\n",
 | |
| 			index, compressed ? "" : "un", length, output->length);
 | |
| 	} else {
 | |
| 		/*
 | |
| 		 * Metadata block.
 | |
| 		 */
 | |
| 		const u8 *data;
 | |
| 		struct bvec_iter_all iter_all = {};
 | |
| 		struct bio_vec *bvec = bvec_init_iter_all(&iter_all);
 | |
| 
 | |
| 		if (index + 2 > msblk->bytes_used) {
 | |
| 			res = -EIO;
 | |
| 			goto out;
 | |
| 		}
 | |
| 		res = squashfs_bio_read(sb, index, 2, &bio, &offset);
 | |
| 		if (res)
 | |
| 			goto out;
 | |
| 
 | |
| 		if (WARN_ON_ONCE(!bio_next_segment(bio, &iter_all))) {
 | |
| 			res = -EIO;
 | |
| 			goto out_free_bio;
 | |
| 		}
 | |
| 		/* Extract the length of the metadata block */
 | |
| 		data = bvec_virt(bvec);
 | |
| 		length = data[offset];
 | |
| 		if (offset < bvec->bv_len - 1) {
 | |
| 			length |= data[offset + 1] << 8;
 | |
| 		} else {
 | |
| 			if (WARN_ON_ONCE(!bio_next_segment(bio, &iter_all))) {
 | |
| 				res = -EIO;
 | |
| 				goto out_free_bio;
 | |
| 			}
 | |
| 			data = bvec_virt(bvec);
 | |
| 			length |= data[0] << 8;
 | |
| 		}
 | |
| 		bio_free_pages(bio);
 | |
| 		bio_uninit(bio);
 | |
| 		kfree(bio);
 | |
| 
 | |
| 		compressed = SQUASHFS_COMPRESSED(length);
 | |
| 		length = SQUASHFS_COMPRESSED_SIZE(length);
 | |
| 		index += 2;
 | |
| 
 | |
| 		TRACE("Block @ 0x%llx, %scompressed size %d\n", index - 2,
 | |
| 		      compressed ? "" : "un", length);
 | |
| 	}
 | |
| 	if (length <= 0 || length > output->length ||
 | |
| 			(index + length) > msblk->bytes_used) {
 | |
| 		res = -EIO;
 | |
| 		goto out;
 | |
| 	}
 | |
| 
 | |
| 	if (next_index)
 | |
| 		*next_index = index + length;
 | |
| 
 | |
| 	res = squashfs_bio_read(sb, index, length, &bio, &offset);
 | |
| 	if (res)
 | |
| 		goto out;
 | |
| 
 | |
| 	if (compressed) {
 | |
| 		if (!msblk->stream) {
 | |
| 			res = -EIO;
 | |
| 			goto out_free_bio;
 | |
| 		}
 | |
| 		res = msblk->thread_ops->decompress(msblk, bio, offset, length, output);
 | |
| 	} else {
 | |
| 		res = copy_bio_to_actor(bio, output, offset, length);
 | |
| 	}
 | |
| 
 | |
| out_free_bio:
 | |
| 	bio_free_pages(bio);
 | |
| 	bio_uninit(bio);
 | |
| 	kfree(bio);
 | |
| out:
 | |
| 	if (res < 0) {
 | |
| 		ERROR("Failed to read block 0x%llx: %d\n", index, res);
 | |
| 		if (msblk->panic_on_errors)
 | |
| 			panic("squashfs read failed");
 | |
| 	}
 | |
| 
 | |
| 	return res;
 | |
| }
 |