@@ -46,6 +46,7 @@ static int cachefiles_daemon_bind(struct cachefiles_cache *, char *);
static void cachefiles_daemon_unbind(struct cachefiles_cache *);
#ifdef CONFIG_CACHEFILES_ONDEMAND
static int cachefiles_ondemand_cinit(struct cachefiles_cache *, char *);
+static int cachefiles_ondemand_cread(struct cachefiles_cache *, char *);
#endif
static unsigned long cachefiles_open;
@@ -81,6 +82,7 @@ static const struct cachefiles_daemon_cmd cachefiles_daemon_cmds[] = {
{ "tag", cachefiles_daemon_tag },
#ifdef CONFIG_CACHEFILES_ONDEMAND
{ "cinit", cachefiles_ondemand_cinit },
+ { "cread", cachefiles_ondemand_cread },
#endif
{ "", NULL }
};
@@ -139,6 +141,9 @@ bool cachefiles_ondemand_daemon_bind(struct cachefiles_cache *cache, char *args)
static int cachefiles_ondemand_fd_release(struct inode *inode, struct file *file)
{
struct cachefiles_object *object = file->private_data;
+ struct cachefiles_cache *cache = object->volume->cache;
+ struct cachefiles_req *req;
+ unsigned long index;
/*
* Uninstall anon_fd to the cachefiles object, so that no further
@@ -146,6 +151,15 @@ static int cachefiles_ondemand_fd_release(struct inode *inode, struct file *file
*/
object->fd = -1;
+ /* complete all associated pending requests */
+ xa_for_each(&cache->reqs, index, req) {
+ if (req->object == object &&
+ req->msg.opcode == CACHEFILES_OP_READ) {
+ req->error = -EIO;
+ complete(&req->done);
+ }
+ }
+
cachefiles_put_object(object, cachefiles_obj_put_ondemand_fd);
return 0;
}
@@ -261,6 +275,36 @@ static int cachefiles_ondemand_cinit(struct cachefiles_cache *cache, char *args)
return ret;
}
+/*
+ * Read request completion
+ * - command: "cread <id>"
+ */
+static int cachefiles_ondemand_cread(struct cachefiles_cache *cache, char *args)
+{
+ struct cachefiles_req *req;
+ unsigned long id;
+ int ret;
+
+ if (!test_bit(CACHEFILES_ONDEMAND_MODE, &cache->flags))
+ return -EOPNOTSUPP;
+
+ if (!*args) {
+ pr_err("Empty id specified\n");
+ return -EINVAL;
+ }
+
+ ret = kstrtoul(args, 0, &id);
+ if (ret)
+ return ret;
+
+ req = xa_erase(&cache->reqs, id);
+ if (!req)
+ return -EINVAL;
+
+ complete(&req->done);
+ return 0;
+}
+
static int cachefiles_ondemand_get_fd(struct cachefiles_req *req)
{
struct cachefiles_init *init;
@@ -460,6 +504,60 @@ int cachefiles_ondemand_init_object(struct cachefiles_object *object)
return ret;
}
+static struct cachefiles_req *
+cachefiles_alloc_read_req(struct cachefiles_object *object,
+ loff_t pos, size_t len)
+{
+ struct cachefiles_req *req;
+ struct cachefiles_read *read;
+ int fd = object->fd;
+
+ /* Stop enqueuig request when daemon closes anon_fd prematurely. */
+ if (WARN_ON_ONCE(fd == -1))
+ return NULL;
+
+ req = cachefiles_alloc_req(object, CACHEFILES_OP_READ, sizeof(*read));
+ if (!req)
+ return NULL;
+
+ read = (void *)&req->msg.data;
+ read->off = pos;
+ read->len = len;
+ read->fd = fd;
+
+ return req;
+}
+
+int cachefiles_ondemand_read(struct cachefiles_object *object,
+ loff_t pos, size_t len)
+{
+ struct cachefiles_cache *cache = object->volume->cache;
+ struct cachefiles_req *req;
+ int ret;
+
+ ret = cachefiles_ondemand_check(cache);
+ if (ret)
+ return ret;
+
+ req = cachefiles_alloc_read_req(object, pos, len);
+ if (!req)
+ return -ENOMEM;
+
+ /*
+ * 1) Checking object->fd and 2) enqueuing request into xarray, is not
+ * atomic as a whole here. Thus similarly, when anon_fd is closed, it's
+ * possible that a new request may be enqueued into xarray, after
+ * associated requests in xarray have already been flushed. But it won't
+ * cause infinite hang since user daemon will still fetch and handle
+ * this request. And since the anon_fd has alrady been closed, any
+ * following file operation with this anon_fd will fail in this case.
+ */
+ ret = cachefiles_ondemand_send_req(cache, req);
+
+ kfree(req);
+ return ret;
+}
+
#else
static inline void cachefiles_ondemand_open(struct cachefiles_cache *cache) {}
static inline void cachefiles_ondemand_release(struct cachefiles_cache *cache) {}
@@ -167,6 +167,8 @@ extern const struct file_operations cachefiles_daemon_fops;
#ifdef CONFIG_CACHEFILES_ONDEMAND
extern int cachefiles_ondemand_init_object(struct cachefiles_object *object);
+extern int cachefiles_ondemand_read(struct cachefiles_object *object,
+ loff_t pos, size_t len);
#else
static inline
@@ -174,6 +176,12 @@ int cachefiles_ondemand_init_object(struct cachefiles_object *object)
{
return 0;
}
+
+static inline int cachefiles_ondemand_read(struct cachefiles_object *object,
+ loff_t pos, size_t len)
+{
+ return -EOPNOTSUPP;
+}
#endif
/*
@@ -95,6 +95,7 @@ static int cachefiles_read(struct netfs_cache_resources *cres,
file, file_inode(file)->i_ino, start_pos, len,
i_size_read(file_inode(file)));
+retry:
/* If the caller asked us to seek for data before doing the read, then
* we should do that now. If we find a gap, we fill it with zeros.
*/
@@ -119,6 +120,16 @@ static int cachefiles_read(struct netfs_cache_resources *cres,
if (read_hole == NETFS_READ_HOLE_FAIL)
goto presubmission_error;
+ if (read_hole == NETFS_READ_HOLE_ONDEMAND) {
+ if (!cachefiles_ondemand_read(object, off, len)) {
+ /* fail the read if no progress achieved */
+ read_hole = NETFS_READ_HOLE_FAIL;
+ goto retry;
+ }
+
+ goto presubmission_error;
+ }
+
iov_iter_zero(len, iter);
skipped = len;
ret = 0;
@@ -203,6 +203,7 @@ enum netfs_read_from_hole {
NETFS_READ_HOLE_IGNORE,
NETFS_READ_HOLE_CLEAR,
NETFS_READ_HOLE_FAIL,
+ NETFS_READ_HOLE_ONDEMAND,
};
/*
@@ -8,6 +8,7 @@
enum cachefiles_opcode {
CACHEFILES_OP_INIT,
+ CACHEFILES_OP_READ,
};
/*
@@ -38,4 +39,10 @@ enum cachefiles_init_flags {
#define CACHEFILES_INIT_FL_WANT_CACHE_SIZE (1 << CACHEFILES_INIT_WANT_CACHE_SIZE)
+struct cachefiles_read {
+ __u64 off;
+ __u64 len;
+ __u32 fd;
+};
+
#endif
Implement the data plane of on-demand read mode. A new NETFS_READ_HOLE_ONDEMAND flag is introduced to indicate that on-demand read should be done when a cache miss encountered. In this case, the read routine will send a READ request to user daemon, along with the anonymous fd and the file range that shall be read. Now user daemon is responsible for fetching data in the given file range, and then writing the fetched data into cache file with the given anonymous fd. After sending the READ request, the read routine will hang there, until the READ request is handled by user daemon. Then it will retry to read from the same file range. If a cache miss is encountered again on the same file range, the read routine will fail then. Signed-off-by: Jeffle Xu <jefflexu@linux.alibaba.com> --- fs/cachefiles/daemon.c | 98 +++++++++++++++++++++++++++++++++ fs/cachefiles/internal.h | 8 +++ fs/cachefiles/io.c | 11 ++++ include/linux/netfs.h | 1 + include/uapi/linux/cachefiles.h | 7 +++ 5 files changed, 125 insertions(+)