|  | /* SPDX-License-Identifier: GPL-2.0-or-later */ | 
|  | /* NFS filesystem cache interface definitions | 
|  | * | 
|  | * Copyright (C) 2008 Red Hat, Inc. All Rights Reserved. | 
|  | * Written by David Howells (dhowells@redhat.com) | 
|  | */ | 
|  |  | 
|  | #ifndef _NFS_FSCACHE_H | 
|  | #define _NFS_FSCACHE_H | 
|  |  | 
|  | #include <linux/swap.h> | 
|  | #include <linux/nfs_fs.h> | 
|  | #include <linux/nfs_mount.h> | 
|  | #include <linux/nfs4_mount.h> | 
|  | #include <linux/fscache.h> | 
|  | #include <linux/iversion.h> | 
|  |  | 
|  | #ifdef CONFIG_NFS_FSCACHE | 
|  |  | 
|  | /* | 
|  | * Definition of the auxiliary data attached to NFS inode storage objects | 
|  | * within the cache. | 
|  | * | 
|  | * The contents of this struct are recorded in the on-disk local cache in the | 
|  | * auxiliary data attached to the data storage object backing an inode.  This | 
|  | * permits coherency to be managed when a new inode binds to an already extant | 
|  | * cache object. | 
|  | */ | 
|  | struct nfs_fscache_inode_auxdata { | 
|  | s64	mtime_sec; | 
|  | s64	mtime_nsec; | 
|  | s64	ctime_sec; | 
|  | s64	ctime_nsec; | 
|  | u64	change_attr; | 
|  | }; | 
|  |  | 
|  | struct nfs_netfs_io_data { | 
|  | /* | 
|  | * NFS may split a netfs_io_subrequest into multiple RPCs, each | 
|  | * with their own read completion.  In netfs, we can only call | 
|  | * netfs_subreq_terminated() once for each subrequest.  Use the | 
|  | * refcount here to double as a marker of the last RPC completion, | 
|  | * and only call netfs via netfs_subreq_terminated() once. | 
|  | */ | 
|  | refcount_t			refcount; | 
|  | struct netfs_io_subrequest	*sreq; | 
|  |  | 
|  | /* | 
|  | * Final disposition of the netfs_io_subrequest, sent in | 
|  | * netfs_subreq_terminated() | 
|  | */ | 
|  | atomic64_t	transferred; | 
|  | int		error; | 
|  | }; | 
|  |  | 
|  | static inline void nfs_netfs_get(struct nfs_netfs_io_data *netfs) | 
|  | { | 
|  | refcount_inc(&netfs->refcount); | 
|  | } | 
|  |  | 
|  | static inline void nfs_netfs_put(struct nfs_netfs_io_data *netfs) | 
|  | { | 
|  | /* Only the last RPC completion should call netfs_subreq_terminated() */ | 
|  | if (!refcount_dec_and_test(&netfs->refcount)) | 
|  | return; | 
|  |  | 
|  | /* | 
|  | * The NFS pageio interface may read a complete page, even when netfs | 
|  | * only asked for a partial page.  Specifically, this may be seen when | 
|  | * one thread is truncating a file while another one is reading the last | 
|  | * page of the file. | 
|  | * Correct the final length here to be no larger than the netfs subrequest | 
|  | * length, and thus avoid netfs's "Subreq overread" warning message. | 
|  | */ | 
|  | netfs->sreq->transferred = min_t(s64, netfs->sreq->len, | 
|  | atomic64_read(&netfs->transferred)); | 
|  | netfs->sreq->error = netfs->error; | 
|  | netfs_read_subreq_terminated(netfs->sreq); | 
|  | kfree(netfs); | 
|  | } | 
|  | static inline void nfs_netfs_inode_init(struct nfs_inode *nfsi) | 
|  | { | 
|  | netfs_inode_init(&nfsi->netfs, &nfs_netfs_ops, false); | 
|  | } | 
|  | extern void nfs_netfs_initiate_read(struct nfs_pgio_header *hdr); | 
|  | extern void nfs_netfs_read_completion(struct nfs_pgio_header *hdr); | 
|  | extern int nfs_netfs_folio_unlock(struct folio *folio); | 
|  |  | 
|  | /* | 
|  | * fscache.c | 
|  | */ | 
|  | extern int nfs_fscache_get_super_cookie(struct super_block *, const char *, int); | 
|  | extern void nfs_fscache_release_super_cookie(struct super_block *); | 
|  |  | 
|  | extern void nfs_fscache_init_inode(struct inode *); | 
|  | extern void nfs_fscache_clear_inode(struct inode *); | 
|  | extern void nfs_fscache_open_file(struct inode *, struct file *); | 
|  | extern void nfs_fscache_release_file(struct inode *, struct file *); | 
|  | extern int nfs_netfs_readahead(struct readahead_control *ractl); | 
|  | extern int nfs_netfs_read_folio(struct file *file, struct folio *folio); | 
|  |  | 
|  | static inline bool nfs_fscache_release_folio(struct folio *folio, gfp_t gfp) | 
|  | { | 
|  | if (folio_test_private_2(folio)) { /* [DEPRECATED] */ | 
|  | if (current_is_kswapd() || !(gfp & __GFP_FS)) | 
|  | return false; | 
|  | folio_wait_private_2(folio); | 
|  | } | 
|  | fscache_note_page_release(netfs_i_cookie(netfs_inode(folio->mapping->host))); | 
|  | return true; | 
|  | } | 
|  |  | 
|  | static inline void nfs_fscache_update_auxdata(struct nfs_fscache_inode_auxdata *auxdata, | 
|  | struct inode *inode) | 
|  | { | 
|  | memset(auxdata, 0, sizeof(*auxdata)); | 
|  | auxdata->mtime_sec  = inode_get_mtime(inode).tv_sec; | 
|  | auxdata->mtime_nsec = inode_get_mtime(inode).tv_nsec; | 
|  | auxdata->ctime_sec  = inode_get_ctime(inode).tv_sec; | 
|  | auxdata->ctime_nsec = inode_get_ctime(inode).tv_nsec; | 
|  |  | 
|  | if (NFS_SERVER(inode)->nfs_client->rpc_ops->version == 4) | 
|  | auxdata->change_attr = inode_peek_iversion_raw(inode); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Invalidate the contents of fscache for this inode.  This will not sleep. | 
|  | */ | 
|  | static inline void nfs_fscache_invalidate(struct inode *inode, int flags) | 
|  | { | 
|  | struct nfs_fscache_inode_auxdata auxdata; | 
|  | struct fscache_cookie *cookie =  netfs_i_cookie(&NFS_I(inode)->netfs); | 
|  |  | 
|  | nfs_fscache_update_auxdata(&auxdata, inode); | 
|  | fscache_invalidate(cookie, &auxdata, i_size_read(inode), flags); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * indicate the client caching state as readable text | 
|  | */ | 
|  | static inline const char *nfs_server_fscache_state(struct nfs_server *server) | 
|  | { | 
|  | if (server->fscache) | 
|  | return "yes"; | 
|  | return "no "; | 
|  | } | 
|  |  | 
|  | static inline void nfs_netfs_set_pgio_header(struct nfs_pgio_header *hdr, | 
|  | struct nfs_pageio_descriptor *desc) | 
|  | { | 
|  | hdr->netfs = desc->pg_netfs; | 
|  | } | 
|  | static inline void nfs_netfs_set_pageio_descriptor(struct nfs_pageio_descriptor *desc, | 
|  | struct nfs_pgio_header *hdr) | 
|  | { | 
|  | desc->pg_netfs = hdr->netfs; | 
|  | } | 
|  | static inline void nfs_netfs_reset_pageio_descriptor(struct nfs_pageio_descriptor *desc) | 
|  | { | 
|  | desc->pg_netfs = NULL; | 
|  | } | 
|  | #else /* CONFIG_NFS_FSCACHE */ | 
|  | static inline void nfs_netfs_inode_init(struct nfs_inode *nfsi) {} | 
|  | static inline void nfs_netfs_initiate_read(struct nfs_pgio_header *hdr) {} | 
|  | static inline void nfs_netfs_read_completion(struct nfs_pgio_header *hdr) {} | 
|  | static inline int nfs_netfs_folio_unlock(struct folio *folio) | 
|  | { | 
|  | return 1; | 
|  | } | 
|  | static inline void nfs_fscache_release_super_cookie(struct super_block *sb) {} | 
|  |  | 
|  | static inline void nfs_fscache_init_inode(struct inode *inode) {} | 
|  | static inline void nfs_fscache_clear_inode(struct inode *inode) {} | 
|  | static inline void nfs_fscache_open_file(struct inode *inode, | 
|  | struct file *filp) {} | 
|  | static inline void nfs_fscache_release_file(struct inode *inode, struct file *file) {} | 
|  | static inline int nfs_netfs_readahead(struct readahead_control *ractl) | 
|  | { | 
|  | return -ENOBUFS; | 
|  | } | 
|  | static inline int nfs_netfs_read_folio(struct file *file, struct folio *folio) | 
|  | { | 
|  | return -ENOBUFS; | 
|  | } | 
|  |  | 
|  | static inline bool nfs_fscache_release_folio(struct folio *folio, gfp_t gfp) | 
|  | { | 
|  | return true; /* may release folio */ | 
|  | } | 
|  | static inline void nfs_fscache_invalidate(struct inode *inode, int flags) {} | 
|  |  | 
|  | static inline const char *nfs_server_fscache_state(struct nfs_server *server) | 
|  | { | 
|  | return "no "; | 
|  | } | 
|  | static inline void nfs_netfs_set_pgio_header(struct nfs_pgio_header *hdr, | 
|  | struct nfs_pageio_descriptor *desc) {} | 
|  | static inline void nfs_netfs_set_pageio_descriptor(struct nfs_pageio_descriptor *desc, | 
|  | struct nfs_pgio_header *hdr) {} | 
|  | static inline void nfs_netfs_reset_pageio_descriptor(struct nfs_pageio_descriptor *desc) {} | 
|  | #endif /* CONFIG_NFS_FSCACHE */ | 
|  | #endif /* _NFS_FSCACHE_H */ |