277 lines
		
	
	
		
			7.1 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			277 lines
		
	
	
		
			7.1 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0
 | |
| /*
 | |
|  * FUSE inode io modes.
 | |
|  *
 | |
|  * Copyright (c) 2024 CTERA Networks.
 | |
|  */
 | |
| 
 | |
| #include "fuse_i.h"
 | |
| 
 | |
| #include <linux/kernel.h>
 | |
| #include <linux/sched.h>
 | |
| #include <linux/file.h>
 | |
| #include <linux/fs.h>
 | |
| 
 | |
| /*
 | |
|  * Return true if need to wait for new opens in caching mode.
 | |
|  */
 | |
| static inline bool fuse_is_io_cache_wait(struct fuse_inode *fi)
 | |
| {
 | |
| 	return READ_ONCE(fi->iocachectr) < 0 && !fuse_inode_backing(fi);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Called on cached file open() and on first mmap() of direct_io file.
 | |
|  * Takes cached_io inode mode reference to be dropped on file release.
 | |
|  *
 | |
|  * Blocks new parallel dio writes and waits for the in-progress parallel dio
 | |
|  * writes to complete.
 | |
|  */
 | |
| int fuse_file_cached_io_open(struct inode *inode, struct fuse_file *ff)
 | |
| {
 | |
| 	struct fuse_inode *fi = get_fuse_inode(inode);
 | |
| 
 | |
| 	/* There are no io modes if server does not implement open */
 | |
| 	if (!ff->args)
 | |
| 		return 0;
 | |
| 
 | |
| 	spin_lock(&fi->lock);
 | |
| 	/*
 | |
| 	 * Setting the bit advises new direct-io writes to use an exclusive
 | |
| 	 * lock - without it the wait below might be forever.
 | |
| 	 */
 | |
| 	while (fuse_is_io_cache_wait(fi)) {
 | |
| 		set_bit(FUSE_I_CACHE_IO_MODE, &fi->state);
 | |
| 		spin_unlock(&fi->lock);
 | |
| 		wait_event(fi->direct_io_waitq, !fuse_is_io_cache_wait(fi));
 | |
| 		spin_lock(&fi->lock);
 | |
| 	}
 | |
| 
 | |
| 	/*
 | |
| 	 * Check if inode entered passthrough io mode while waiting for parallel
 | |
| 	 * dio write completion.
 | |
| 	 */
 | |
| 	if (fuse_inode_backing(fi)) {
 | |
| 		clear_bit(FUSE_I_CACHE_IO_MODE, &fi->state);
 | |
| 		spin_unlock(&fi->lock);
 | |
| 		return -ETXTBSY;
 | |
| 	}
 | |
| 
 | |
| 	WARN_ON(ff->iomode == IOM_UNCACHED);
 | |
| 	if (ff->iomode == IOM_NONE) {
 | |
| 		ff->iomode = IOM_CACHED;
 | |
| 		if (fi->iocachectr == 0)
 | |
| 			set_bit(FUSE_I_CACHE_IO_MODE, &fi->state);
 | |
| 		fi->iocachectr++;
 | |
| 	}
 | |
| 	spin_unlock(&fi->lock);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static void fuse_file_cached_io_release(struct fuse_file *ff,
 | |
| 					struct fuse_inode *fi)
 | |
| {
 | |
| 	spin_lock(&fi->lock);
 | |
| 	WARN_ON(fi->iocachectr <= 0);
 | |
| 	WARN_ON(ff->iomode != IOM_CACHED);
 | |
| 	ff->iomode = IOM_NONE;
 | |
| 	fi->iocachectr--;
 | |
| 	if (fi->iocachectr == 0)
 | |
| 		clear_bit(FUSE_I_CACHE_IO_MODE, &fi->state);
 | |
| 	spin_unlock(&fi->lock);
 | |
| }
 | |
| 
 | |
| /* Start strictly uncached io mode where cache access is not allowed */
 | |
| int fuse_inode_uncached_io_start(struct fuse_inode *fi, struct fuse_backing *fb)
 | |
| {
 | |
| 	struct fuse_backing *oldfb;
 | |
| 	int err = 0;
 | |
| 
 | |
| 	spin_lock(&fi->lock);
 | |
| 	/* deny conflicting backing files on same fuse inode */
 | |
| 	oldfb = fuse_inode_backing(fi);
 | |
| 	if (fb && oldfb && oldfb != fb) {
 | |
| 		err = -EBUSY;
 | |
| 		goto unlock;
 | |
| 	}
 | |
| 	if (fi->iocachectr > 0) {
 | |
| 		err = -ETXTBSY;
 | |
| 		goto unlock;
 | |
| 	}
 | |
| 	fi->iocachectr--;
 | |
| 
 | |
| 	/* fuse inode holds a single refcount of backing file */
 | |
| 	if (fb && !oldfb) {
 | |
| 		oldfb = fuse_inode_backing_set(fi, fb);
 | |
| 		WARN_ON_ONCE(oldfb != NULL);
 | |
| 	} else {
 | |
| 		fuse_backing_put(fb);
 | |
| 	}
 | |
| unlock:
 | |
| 	spin_unlock(&fi->lock);
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| /* Takes uncached_io inode mode reference to be dropped on file release */
 | |
| static int fuse_file_uncached_io_open(struct inode *inode,
 | |
| 				      struct fuse_file *ff,
 | |
| 				      struct fuse_backing *fb)
 | |
| {
 | |
| 	struct fuse_inode *fi = get_fuse_inode(inode);
 | |
| 	int err;
 | |
| 
 | |
| 	err = fuse_inode_uncached_io_start(fi, fb);
 | |
| 	if (err)
 | |
| 		return err;
 | |
| 
 | |
| 	WARN_ON(ff->iomode != IOM_NONE);
 | |
| 	ff->iomode = IOM_UNCACHED;
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| void fuse_inode_uncached_io_end(struct fuse_inode *fi)
 | |
| {
 | |
| 	struct fuse_backing *oldfb = NULL;
 | |
| 
 | |
| 	spin_lock(&fi->lock);
 | |
| 	WARN_ON(fi->iocachectr >= 0);
 | |
| 	fi->iocachectr++;
 | |
| 	if (!fi->iocachectr) {
 | |
| 		wake_up(&fi->direct_io_waitq);
 | |
| 		oldfb = fuse_inode_backing_set(fi, NULL);
 | |
| 	}
 | |
| 	spin_unlock(&fi->lock);
 | |
| 	if (oldfb)
 | |
| 		fuse_backing_put(oldfb);
 | |
| }
 | |
| 
 | |
| /* Drop uncached_io reference from passthrough open */
 | |
| static void fuse_file_uncached_io_release(struct fuse_file *ff,
 | |
| 					  struct fuse_inode *fi)
 | |
| {
 | |
| 	WARN_ON(ff->iomode != IOM_UNCACHED);
 | |
| 	ff->iomode = IOM_NONE;
 | |
| 	fuse_inode_uncached_io_end(fi);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Open flags that are allowed in combination with FOPEN_PASSTHROUGH.
 | |
|  * A combination of FOPEN_PASSTHROUGH and FOPEN_DIRECT_IO means that read/write
 | |
|  * operations go directly to the server, but mmap is done on the backing file.
 | |
|  * FOPEN_PASSTHROUGH mode should not co-exist with any users of the fuse inode
 | |
|  * page cache, so FOPEN_KEEP_CACHE is a strange and undesired combination.
 | |
|  */
 | |
| #define FOPEN_PASSTHROUGH_MASK \
 | |
| 	(FOPEN_PASSTHROUGH | FOPEN_DIRECT_IO | FOPEN_PARALLEL_DIRECT_WRITES | \
 | |
| 	 FOPEN_NOFLUSH)
 | |
| 
 | |
| static int fuse_file_passthrough_open(struct inode *inode, struct file *file)
 | |
| {
 | |
| 	struct fuse_file *ff = file->private_data;
 | |
| 	struct fuse_conn *fc = get_fuse_conn(inode);
 | |
| 	struct fuse_backing *fb;
 | |
| 	int err;
 | |
| 
 | |
| 	/* Check allowed conditions for file open in passthrough mode */
 | |
| 	if (!IS_ENABLED(CONFIG_FUSE_PASSTHROUGH) || !fc->passthrough ||
 | |
| 	    (ff->open_flags & ~FOPEN_PASSTHROUGH_MASK))
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	fb = fuse_passthrough_open(file, inode,
 | |
| 				   ff->args->open_outarg.backing_id);
 | |
| 	if (IS_ERR(fb))
 | |
| 		return PTR_ERR(fb);
 | |
| 
 | |
| 	/* First passthrough file open denies caching inode io mode */
 | |
| 	err = fuse_file_uncached_io_open(inode, ff, fb);
 | |
| 	if (!err)
 | |
| 		return 0;
 | |
| 
 | |
| 	fuse_passthrough_release(ff, fb);
 | |
| 	fuse_backing_put(fb);
 | |
| 
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| /* Request access to submit new io to inode via open file */
 | |
| int fuse_file_io_open(struct file *file, struct inode *inode)
 | |
| {
 | |
| 	struct fuse_file *ff = file->private_data;
 | |
| 	struct fuse_inode *fi = get_fuse_inode(inode);
 | |
| 	int err;
 | |
| 
 | |
| 	/*
 | |
| 	 * io modes are not relevant with DAX and with server that does not
 | |
| 	 * implement open.
 | |
| 	 */
 | |
| 	if (FUSE_IS_DAX(inode) || !ff->args)
 | |
| 		return 0;
 | |
| 
 | |
| 	/*
 | |
| 	 * Server is expected to use FOPEN_PASSTHROUGH for all opens of an inode
 | |
| 	 * which is already open for passthrough.
 | |
| 	 */
 | |
| 	err = -EINVAL;
 | |
| 	if (fuse_inode_backing(fi) && !(ff->open_flags & FOPEN_PASSTHROUGH))
 | |
| 		goto fail;
 | |
| 
 | |
| 	/*
 | |
| 	 * FOPEN_PARALLEL_DIRECT_WRITES requires FOPEN_DIRECT_IO.
 | |
| 	 */
 | |
| 	if (!(ff->open_flags & FOPEN_DIRECT_IO))
 | |
| 		ff->open_flags &= ~FOPEN_PARALLEL_DIRECT_WRITES;
 | |
| 
 | |
| 	/*
 | |
| 	 * First passthrough file open denies caching inode io mode.
 | |
| 	 * First caching file open enters caching inode io mode.
 | |
| 	 *
 | |
| 	 * Note that if user opens a file open with O_DIRECT, but server did
 | |
| 	 * not specify FOPEN_DIRECT_IO, a later fcntl() could remove O_DIRECT,
 | |
| 	 * so we put the inode in caching mode to prevent parallel dio.
 | |
| 	 */
 | |
| 	if ((ff->open_flags & FOPEN_DIRECT_IO) &&
 | |
| 	    !(ff->open_flags & FOPEN_PASSTHROUGH))
 | |
| 		return 0;
 | |
| 
 | |
| 	if (ff->open_flags & FOPEN_PASSTHROUGH)
 | |
| 		err = fuse_file_passthrough_open(inode, file);
 | |
| 	else
 | |
| 		err = fuse_file_cached_io_open(inode, ff);
 | |
| 	if (err)
 | |
| 		goto fail;
 | |
| 
 | |
| 	return 0;
 | |
| 
 | |
| fail:
 | |
| 	pr_debug("failed to open file in requested io mode (open_flags=0x%x, err=%i).\n",
 | |
| 		 ff->open_flags, err);
 | |
| 	/*
 | |
| 	 * The file open mode determines the inode io mode.
 | |
| 	 * Using incorrect open mode is a server mistake, which results in
 | |
| 	 * user visible failure of open() with EIO error.
 | |
| 	 */
 | |
| 	return -EIO;
 | |
| }
 | |
| 
 | |
| /* No more pending io and no new io possible to inode via open/mmapped file */
 | |
| void fuse_file_io_release(struct fuse_file *ff, struct inode *inode)
 | |
| {
 | |
| 	struct fuse_inode *fi = get_fuse_inode(inode);
 | |
| 
 | |
| 	/*
 | |
| 	 * Last passthrough file close allows caching inode io mode.
 | |
| 	 * Last caching file close exits caching inode io mode.
 | |
| 	 */
 | |
| 	switch (ff->iomode) {
 | |
| 	case IOM_NONE:
 | |
| 		/* Nothing to do */
 | |
| 		break;
 | |
| 	case IOM_UNCACHED:
 | |
| 		fuse_file_uncached_io_release(ff, fi);
 | |
| 		break;
 | |
| 	case IOM_CACHED:
 | |
| 		fuse_file_cached_io_release(ff, fi);
 | |
| 		break;
 | |
| 	}
 | |
| }
 |