1a62a8ef9SStefan Hajnoczi // SPDX-License-Identifier: GPL-2.0 2a62a8ef9SStefan Hajnoczi /* 3a62a8ef9SStefan Hajnoczi * virtio-fs: Virtio Filesystem 4a62a8ef9SStefan Hajnoczi * Copyright (C) 2018 Red Hat, Inc. 5a62a8ef9SStefan Hajnoczi */ 6a62a8ef9SStefan Hajnoczi 7a62a8ef9SStefan Hajnoczi #include <linux/fs.h> 8*22f3787eSStefan Hajnoczi #include <linux/dax.h> 9*22f3787eSStefan Hajnoczi #include <linux/pci.h> 10*22f3787eSStefan Hajnoczi #include <linux/pfn_t.h> 11a62a8ef9SStefan Hajnoczi #include <linux/module.h> 12a62a8ef9SStefan Hajnoczi #include <linux/virtio.h> 13a62a8ef9SStefan Hajnoczi #include <linux/virtio_fs.h> 14a62a8ef9SStefan Hajnoczi #include <linux/delay.h> 15a62a8ef9SStefan Hajnoczi #include <linux/fs_context.h> 16a62a8ef9SStefan Hajnoczi #include <linux/highmem.h> 17*22f3787eSStefan Hajnoczi #include <linux/uio.h> 18a62a8ef9SStefan Hajnoczi #include "fuse_i.h" 19a62a8ef9SStefan Hajnoczi 20a62a8ef9SStefan Hajnoczi /* List of virtio-fs device instances and a lock for the list. Also provides 21a62a8ef9SStefan Hajnoczi * mutual exclusion in device removal and mounting path 22a62a8ef9SStefan Hajnoczi */ 23a62a8ef9SStefan Hajnoczi static DEFINE_MUTEX(virtio_fs_mutex); 24a62a8ef9SStefan Hajnoczi static LIST_HEAD(virtio_fs_instances); 25a62a8ef9SStefan Hajnoczi 26a62a8ef9SStefan Hajnoczi enum { 27a62a8ef9SStefan Hajnoczi VQ_HIPRIO, 28a62a8ef9SStefan Hajnoczi VQ_REQUEST 29a62a8ef9SStefan Hajnoczi }; 30a62a8ef9SStefan Hajnoczi 31b43b7e81SVivek Goyal #define VQ_NAME_LEN 24 32b43b7e81SVivek Goyal 33a62a8ef9SStefan Hajnoczi /* Per-virtqueue state */ 34a62a8ef9SStefan Hajnoczi struct virtio_fs_vq { 35a62a8ef9SStefan Hajnoczi spinlock_t lock; 36a62a8ef9SStefan Hajnoczi struct virtqueue *vq; /* protected by ->lock */ 37a62a8ef9SStefan Hajnoczi struct work_struct done_work; 38a62a8ef9SStefan Hajnoczi struct list_head queued_reqs; 3951fecdd2SVivek Goyal struct list_head end_reqs; /* End these requests */ 40a62a8ef9SStefan Hajnoczi struct delayed_work dispatch_work; 41a62a8ef9SStefan Hajnoczi struct fuse_dev *fud; 42a62a8ef9SStefan Hajnoczi bool connected; 43a62a8ef9SStefan Hajnoczi long in_flight; 44724c15a4SVivek Goyal struct completion in_flight_zero; /* No inflight requests */ 45b43b7e81SVivek Goyal char name[VQ_NAME_LEN]; 46a62a8ef9SStefan Hajnoczi } ____cacheline_aligned_in_smp; 47a62a8ef9SStefan Hajnoczi 48a62a8ef9SStefan Hajnoczi /* A virtio-fs device instance */ 49a62a8ef9SStefan Hajnoczi struct virtio_fs { 50a62a8ef9SStefan Hajnoczi struct kref refcount; 51a62a8ef9SStefan Hajnoczi struct list_head list; /* on virtio_fs_instances */ 52a62a8ef9SStefan Hajnoczi char *tag; 53a62a8ef9SStefan Hajnoczi struct virtio_fs_vq *vqs; 54a62a8ef9SStefan Hajnoczi unsigned int nvqs; /* number of virtqueues */ 55a62a8ef9SStefan Hajnoczi unsigned int num_request_queues; /* number of request queues */ 56*22f3787eSStefan Hajnoczi struct dax_device *dax_dev; 57*22f3787eSStefan Hajnoczi 58*22f3787eSStefan Hajnoczi /* DAX memory window where file contents are mapped */ 59*22f3787eSStefan Hajnoczi void *window_kaddr; 60*22f3787eSStefan Hajnoczi phys_addr_t window_phys_addr; 61*22f3787eSStefan Hajnoczi size_t window_len; 62a62a8ef9SStefan Hajnoczi }; 63a62a8ef9SStefan Hajnoczi 641efcf39eSVivek Goyal struct virtio_fs_forget_req { 65a62a8ef9SStefan Hajnoczi struct fuse_in_header ih; 66a62a8ef9SStefan Hajnoczi struct fuse_forget_in arg; 671efcf39eSVivek Goyal }; 681efcf39eSVivek Goyal 691efcf39eSVivek Goyal struct virtio_fs_forget { 70a62a8ef9SStefan Hajnoczi /* This request can be temporarily queued on virt queue */ 71a62a8ef9SStefan Hajnoczi struct list_head list; 721efcf39eSVivek Goyal struct virtio_fs_forget_req req; 73a62a8ef9SStefan Hajnoczi }; 74a62a8ef9SStefan Hajnoczi 75bb737bbeSVivek Goyal struct virtio_fs_req_work { 76bb737bbeSVivek Goyal struct fuse_req *req; 77bb737bbeSVivek Goyal struct virtio_fs_vq *fsvq; 78bb737bbeSVivek Goyal struct work_struct done_work; 79bb737bbeSVivek Goyal }; 80bb737bbeSVivek Goyal 81a9bfd9ddSVivek Goyal static int virtio_fs_enqueue_req(struct virtio_fs_vq *fsvq, 82a9bfd9ddSVivek Goyal struct fuse_req *req, bool in_flight); 83a9bfd9ddSVivek Goyal 84a62a8ef9SStefan Hajnoczi static inline struct virtio_fs_vq *vq_to_fsvq(struct virtqueue *vq) 85a62a8ef9SStefan Hajnoczi { 86a62a8ef9SStefan Hajnoczi struct virtio_fs *fs = vq->vdev->priv; 87a62a8ef9SStefan Hajnoczi 88a62a8ef9SStefan Hajnoczi return &fs->vqs[vq->index]; 89a62a8ef9SStefan Hajnoczi } 90a62a8ef9SStefan Hajnoczi 91a62a8ef9SStefan Hajnoczi static inline struct fuse_pqueue *vq_to_fpq(struct virtqueue *vq) 92a62a8ef9SStefan Hajnoczi { 93a62a8ef9SStefan Hajnoczi return &vq_to_fsvq(vq)->fud->pq; 94a62a8ef9SStefan Hajnoczi } 95a62a8ef9SStefan Hajnoczi 96c17ea009SVivek Goyal /* Should be called with fsvq->lock held. */ 97c17ea009SVivek Goyal static inline void inc_in_flight_req(struct virtio_fs_vq *fsvq) 98c17ea009SVivek Goyal { 99c17ea009SVivek Goyal fsvq->in_flight++; 100c17ea009SVivek Goyal } 101c17ea009SVivek Goyal 102c17ea009SVivek Goyal /* Should be called with fsvq->lock held. */ 103c17ea009SVivek Goyal static inline void dec_in_flight_req(struct virtio_fs_vq *fsvq) 104c17ea009SVivek Goyal { 105c17ea009SVivek Goyal WARN_ON(fsvq->in_flight <= 0); 106c17ea009SVivek Goyal fsvq->in_flight--; 107724c15a4SVivek Goyal if (!fsvq->in_flight) 108724c15a4SVivek Goyal complete(&fsvq->in_flight_zero); 109c17ea009SVivek Goyal } 110c17ea009SVivek Goyal 111a62a8ef9SStefan Hajnoczi static void release_virtio_fs_obj(struct kref *ref) 112a62a8ef9SStefan Hajnoczi { 113a62a8ef9SStefan Hajnoczi struct virtio_fs *vfs = container_of(ref, struct virtio_fs, refcount); 114a62a8ef9SStefan Hajnoczi 115a62a8ef9SStefan Hajnoczi kfree(vfs->vqs); 116a62a8ef9SStefan Hajnoczi kfree(vfs); 117a62a8ef9SStefan Hajnoczi } 118a62a8ef9SStefan Hajnoczi 119a62a8ef9SStefan Hajnoczi /* Make sure virtiofs_mutex is held */ 120a62a8ef9SStefan Hajnoczi static void virtio_fs_put(struct virtio_fs *fs) 121a62a8ef9SStefan Hajnoczi { 122a62a8ef9SStefan Hajnoczi kref_put(&fs->refcount, release_virtio_fs_obj); 123a62a8ef9SStefan Hajnoczi } 124a62a8ef9SStefan Hajnoczi 125a62a8ef9SStefan Hajnoczi static void virtio_fs_fiq_release(struct fuse_iqueue *fiq) 126a62a8ef9SStefan Hajnoczi { 127a62a8ef9SStefan Hajnoczi struct virtio_fs *vfs = fiq->priv; 128a62a8ef9SStefan Hajnoczi 129a62a8ef9SStefan Hajnoczi mutex_lock(&virtio_fs_mutex); 130a62a8ef9SStefan Hajnoczi virtio_fs_put(vfs); 131a62a8ef9SStefan Hajnoczi mutex_unlock(&virtio_fs_mutex); 132a62a8ef9SStefan Hajnoczi } 133a62a8ef9SStefan Hajnoczi 134a62a8ef9SStefan Hajnoczi static void virtio_fs_drain_queue(struct virtio_fs_vq *fsvq) 135a62a8ef9SStefan Hajnoczi { 136a62a8ef9SStefan Hajnoczi WARN_ON(fsvq->in_flight < 0); 137a62a8ef9SStefan Hajnoczi 138a62a8ef9SStefan Hajnoczi /* Wait for in flight requests to finish.*/ 139a62a8ef9SStefan Hajnoczi spin_lock(&fsvq->lock); 140724c15a4SVivek Goyal if (fsvq->in_flight) { 141724c15a4SVivek Goyal /* We are holding virtio_fs_mutex. There should not be any 142724c15a4SVivek Goyal * waiters waiting for completion. 143724c15a4SVivek Goyal */ 144724c15a4SVivek Goyal reinit_completion(&fsvq->in_flight_zero); 145a62a8ef9SStefan Hajnoczi spin_unlock(&fsvq->lock); 146724c15a4SVivek Goyal wait_for_completion(&fsvq->in_flight_zero); 147724c15a4SVivek Goyal } else { 148a62a8ef9SStefan Hajnoczi spin_unlock(&fsvq->lock); 149a62a8ef9SStefan Hajnoczi } 150a62a8ef9SStefan Hajnoczi 151a62a8ef9SStefan Hajnoczi flush_work(&fsvq->done_work); 152a62a8ef9SStefan Hajnoczi flush_delayed_work(&fsvq->dispatch_work); 153a62a8ef9SStefan Hajnoczi } 154a62a8ef9SStefan Hajnoczi 155724c15a4SVivek Goyal static void virtio_fs_drain_all_queues_locked(struct virtio_fs *fs) 156a62a8ef9SStefan Hajnoczi { 157a62a8ef9SStefan Hajnoczi struct virtio_fs_vq *fsvq; 158a62a8ef9SStefan Hajnoczi int i; 159a62a8ef9SStefan Hajnoczi 160a62a8ef9SStefan Hajnoczi for (i = 0; i < fs->nvqs; i++) { 161a62a8ef9SStefan Hajnoczi fsvq = &fs->vqs[i]; 162a62a8ef9SStefan Hajnoczi virtio_fs_drain_queue(fsvq); 163a62a8ef9SStefan Hajnoczi } 164a62a8ef9SStefan Hajnoczi } 165a62a8ef9SStefan Hajnoczi 166724c15a4SVivek Goyal static void virtio_fs_drain_all_queues(struct virtio_fs *fs) 167724c15a4SVivek Goyal { 168724c15a4SVivek Goyal /* Provides mutual exclusion between ->remove and ->kill_sb 169724c15a4SVivek Goyal * paths. We don't want both of these draining queue at the 170724c15a4SVivek Goyal * same time. Current completion logic reinits completion 171724c15a4SVivek Goyal * and that means there should not be any other thread 172724c15a4SVivek Goyal * doing reinit or waiting for completion already. 173724c15a4SVivek Goyal */ 174724c15a4SVivek Goyal mutex_lock(&virtio_fs_mutex); 175724c15a4SVivek Goyal virtio_fs_drain_all_queues_locked(fs); 176724c15a4SVivek Goyal mutex_unlock(&virtio_fs_mutex); 177724c15a4SVivek Goyal } 178724c15a4SVivek Goyal 179a62a8ef9SStefan Hajnoczi static void virtio_fs_start_all_queues(struct virtio_fs *fs) 180a62a8ef9SStefan Hajnoczi { 181a62a8ef9SStefan Hajnoczi struct virtio_fs_vq *fsvq; 182a62a8ef9SStefan Hajnoczi int i; 183a62a8ef9SStefan Hajnoczi 184a62a8ef9SStefan Hajnoczi for (i = 0; i < fs->nvqs; i++) { 185a62a8ef9SStefan Hajnoczi fsvq = &fs->vqs[i]; 186a62a8ef9SStefan Hajnoczi spin_lock(&fsvq->lock); 187a62a8ef9SStefan Hajnoczi fsvq->connected = true; 188a62a8ef9SStefan Hajnoczi spin_unlock(&fsvq->lock); 189a62a8ef9SStefan Hajnoczi } 190a62a8ef9SStefan Hajnoczi } 191a62a8ef9SStefan Hajnoczi 192a62a8ef9SStefan Hajnoczi /* Add a new instance to the list or return -EEXIST if tag name exists*/ 193a62a8ef9SStefan Hajnoczi static int virtio_fs_add_instance(struct virtio_fs *fs) 194a62a8ef9SStefan Hajnoczi { 195a62a8ef9SStefan Hajnoczi struct virtio_fs *fs2; 196a62a8ef9SStefan Hajnoczi bool duplicate = false; 197a62a8ef9SStefan Hajnoczi 198a62a8ef9SStefan Hajnoczi mutex_lock(&virtio_fs_mutex); 199a62a8ef9SStefan Hajnoczi 200a62a8ef9SStefan Hajnoczi list_for_each_entry(fs2, &virtio_fs_instances, list) { 201a62a8ef9SStefan Hajnoczi if (strcmp(fs->tag, fs2->tag) == 0) 202a62a8ef9SStefan Hajnoczi duplicate = true; 203a62a8ef9SStefan Hajnoczi } 204a62a8ef9SStefan Hajnoczi 205a62a8ef9SStefan Hajnoczi if (!duplicate) 206a62a8ef9SStefan Hajnoczi list_add_tail(&fs->list, &virtio_fs_instances); 207a62a8ef9SStefan Hajnoczi 208a62a8ef9SStefan Hajnoczi mutex_unlock(&virtio_fs_mutex); 209a62a8ef9SStefan Hajnoczi 210a62a8ef9SStefan Hajnoczi if (duplicate) 211a62a8ef9SStefan Hajnoczi return -EEXIST; 212a62a8ef9SStefan Hajnoczi return 0; 213a62a8ef9SStefan Hajnoczi } 214a62a8ef9SStefan Hajnoczi 215a62a8ef9SStefan Hajnoczi /* Return the virtio_fs with a given tag, or NULL */ 216a62a8ef9SStefan Hajnoczi static struct virtio_fs *virtio_fs_find_instance(const char *tag) 217a62a8ef9SStefan Hajnoczi { 218a62a8ef9SStefan Hajnoczi struct virtio_fs *fs; 219a62a8ef9SStefan Hajnoczi 220a62a8ef9SStefan Hajnoczi mutex_lock(&virtio_fs_mutex); 221a62a8ef9SStefan Hajnoczi 222a62a8ef9SStefan Hajnoczi list_for_each_entry(fs, &virtio_fs_instances, list) { 223a62a8ef9SStefan Hajnoczi if (strcmp(fs->tag, tag) == 0) { 224a62a8ef9SStefan Hajnoczi kref_get(&fs->refcount); 225a62a8ef9SStefan Hajnoczi goto found; 226a62a8ef9SStefan Hajnoczi } 227a62a8ef9SStefan Hajnoczi } 228a62a8ef9SStefan Hajnoczi 229a62a8ef9SStefan Hajnoczi fs = NULL; /* not found */ 230a62a8ef9SStefan Hajnoczi 231a62a8ef9SStefan Hajnoczi found: 232a62a8ef9SStefan Hajnoczi mutex_unlock(&virtio_fs_mutex); 233a62a8ef9SStefan Hajnoczi 234a62a8ef9SStefan Hajnoczi return fs; 235a62a8ef9SStefan Hajnoczi } 236a62a8ef9SStefan Hajnoczi 237a62a8ef9SStefan Hajnoczi static void virtio_fs_free_devs(struct virtio_fs *fs) 238a62a8ef9SStefan Hajnoczi { 239a62a8ef9SStefan Hajnoczi unsigned int i; 240a62a8ef9SStefan Hajnoczi 241a62a8ef9SStefan Hajnoczi for (i = 0; i < fs->nvqs; i++) { 242a62a8ef9SStefan Hajnoczi struct virtio_fs_vq *fsvq = &fs->vqs[i]; 243a62a8ef9SStefan Hajnoczi 244a62a8ef9SStefan Hajnoczi if (!fsvq->fud) 245a62a8ef9SStefan Hajnoczi continue; 246a62a8ef9SStefan Hajnoczi 247a62a8ef9SStefan Hajnoczi fuse_dev_free(fsvq->fud); 248a62a8ef9SStefan Hajnoczi fsvq->fud = NULL; 249a62a8ef9SStefan Hajnoczi } 250a62a8ef9SStefan Hajnoczi } 251a62a8ef9SStefan Hajnoczi 252a62a8ef9SStefan Hajnoczi /* Read filesystem name from virtio config into fs->tag (must kfree()). */ 253a62a8ef9SStefan Hajnoczi static int virtio_fs_read_tag(struct virtio_device *vdev, struct virtio_fs *fs) 254a62a8ef9SStefan Hajnoczi { 255a62a8ef9SStefan Hajnoczi char tag_buf[sizeof_field(struct virtio_fs_config, tag)]; 256a62a8ef9SStefan Hajnoczi char *end; 257a62a8ef9SStefan Hajnoczi size_t len; 258a62a8ef9SStefan Hajnoczi 259a62a8ef9SStefan Hajnoczi virtio_cread_bytes(vdev, offsetof(struct virtio_fs_config, tag), 260a62a8ef9SStefan Hajnoczi &tag_buf, sizeof(tag_buf)); 261a62a8ef9SStefan Hajnoczi end = memchr(tag_buf, '\0', sizeof(tag_buf)); 262a62a8ef9SStefan Hajnoczi if (end == tag_buf) 263a62a8ef9SStefan Hajnoczi return -EINVAL; /* empty tag */ 264a62a8ef9SStefan Hajnoczi if (!end) 265a62a8ef9SStefan Hajnoczi end = &tag_buf[sizeof(tag_buf)]; 266a62a8ef9SStefan Hajnoczi 267a62a8ef9SStefan Hajnoczi len = end - tag_buf; 268a62a8ef9SStefan Hajnoczi fs->tag = devm_kmalloc(&vdev->dev, len + 1, GFP_KERNEL); 269a62a8ef9SStefan Hajnoczi if (!fs->tag) 270a62a8ef9SStefan Hajnoczi return -ENOMEM; 271a62a8ef9SStefan Hajnoczi memcpy(fs->tag, tag_buf, len); 272a62a8ef9SStefan Hajnoczi fs->tag[len] = '\0'; 273a62a8ef9SStefan Hajnoczi return 0; 274a62a8ef9SStefan Hajnoczi } 275a62a8ef9SStefan Hajnoczi 276a62a8ef9SStefan Hajnoczi /* Work function for hiprio completion */ 277a62a8ef9SStefan Hajnoczi static void virtio_fs_hiprio_done_work(struct work_struct *work) 278a62a8ef9SStefan Hajnoczi { 279a62a8ef9SStefan Hajnoczi struct virtio_fs_vq *fsvq = container_of(work, struct virtio_fs_vq, 280a62a8ef9SStefan Hajnoczi done_work); 281a62a8ef9SStefan Hajnoczi struct virtqueue *vq = fsvq->vq; 282a62a8ef9SStefan Hajnoczi 283a62a8ef9SStefan Hajnoczi /* Free completed FUSE_FORGET requests */ 284a62a8ef9SStefan Hajnoczi spin_lock(&fsvq->lock); 285a62a8ef9SStefan Hajnoczi do { 286a62a8ef9SStefan Hajnoczi unsigned int len; 287a62a8ef9SStefan Hajnoczi void *req; 288a62a8ef9SStefan Hajnoczi 289a62a8ef9SStefan Hajnoczi virtqueue_disable_cb(vq); 290a62a8ef9SStefan Hajnoczi 291a62a8ef9SStefan Hajnoczi while ((req = virtqueue_get_buf(vq, &len)) != NULL) { 292a62a8ef9SStefan Hajnoczi kfree(req); 293c17ea009SVivek Goyal dec_in_flight_req(fsvq); 294a62a8ef9SStefan Hajnoczi } 295a62a8ef9SStefan Hajnoczi } while (!virtqueue_enable_cb(vq) && likely(!virtqueue_is_broken(vq))); 296a62a8ef9SStefan Hajnoczi spin_unlock(&fsvq->lock); 297a62a8ef9SStefan Hajnoczi } 298a62a8ef9SStefan Hajnoczi 29951fecdd2SVivek Goyal static void virtio_fs_request_dispatch_work(struct work_struct *work) 300a62a8ef9SStefan Hajnoczi { 30151fecdd2SVivek Goyal struct fuse_req *req; 30251fecdd2SVivek Goyal struct virtio_fs_vq *fsvq = container_of(work, struct virtio_fs_vq, 30351fecdd2SVivek Goyal dispatch_work.work); 30451fecdd2SVivek Goyal struct fuse_conn *fc = fsvq->fud->fc; 305a9bfd9ddSVivek Goyal int ret; 30651fecdd2SVivek Goyal 30751fecdd2SVivek Goyal pr_debug("virtio-fs: worker %s called.\n", __func__); 30851fecdd2SVivek Goyal while (1) { 30951fecdd2SVivek Goyal spin_lock(&fsvq->lock); 31051fecdd2SVivek Goyal req = list_first_entry_or_null(&fsvq->end_reqs, struct fuse_req, 31151fecdd2SVivek Goyal list); 31251fecdd2SVivek Goyal if (!req) { 31351fecdd2SVivek Goyal spin_unlock(&fsvq->lock); 314a9bfd9ddSVivek Goyal break; 31551fecdd2SVivek Goyal } 31651fecdd2SVivek Goyal 31751fecdd2SVivek Goyal list_del_init(&req->list); 31851fecdd2SVivek Goyal spin_unlock(&fsvq->lock); 31951fecdd2SVivek Goyal fuse_request_end(fc, req); 32051fecdd2SVivek Goyal } 321a9bfd9ddSVivek Goyal 322a9bfd9ddSVivek Goyal /* Dispatch pending requests */ 323a9bfd9ddSVivek Goyal while (1) { 324a9bfd9ddSVivek Goyal spin_lock(&fsvq->lock); 325a9bfd9ddSVivek Goyal req = list_first_entry_or_null(&fsvq->queued_reqs, 326a9bfd9ddSVivek Goyal struct fuse_req, list); 327a9bfd9ddSVivek Goyal if (!req) { 328a9bfd9ddSVivek Goyal spin_unlock(&fsvq->lock); 329a9bfd9ddSVivek Goyal return; 330a9bfd9ddSVivek Goyal } 331a9bfd9ddSVivek Goyal list_del_init(&req->list); 332a9bfd9ddSVivek Goyal spin_unlock(&fsvq->lock); 333a9bfd9ddSVivek Goyal 334a9bfd9ddSVivek Goyal ret = virtio_fs_enqueue_req(fsvq, req, true); 335a9bfd9ddSVivek Goyal if (ret < 0) { 336a9bfd9ddSVivek Goyal if (ret == -ENOMEM || ret == -ENOSPC) { 337a9bfd9ddSVivek Goyal spin_lock(&fsvq->lock); 338a9bfd9ddSVivek Goyal list_add_tail(&req->list, &fsvq->queued_reqs); 339a9bfd9ddSVivek Goyal schedule_delayed_work(&fsvq->dispatch_work, 340a9bfd9ddSVivek Goyal msecs_to_jiffies(1)); 341a9bfd9ddSVivek Goyal spin_unlock(&fsvq->lock); 342a9bfd9ddSVivek Goyal return; 343a9bfd9ddSVivek Goyal } 344a9bfd9ddSVivek Goyal req->out.h.error = ret; 345a9bfd9ddSVivek Goyal spin_lock(&fsvq->lock); 346a9bfd9ddSVivek Goyal dec_in_flight_req(fsvq); 347a9bfd9ddSVivek Goyal spin_unlock(&fsvq->lock); 348a9bfd9ddSVivek Goyal pr_err("virtio-fs: virtio_fs_enqueue_req() failed %d\n", 349a9bfd9ddSVivek Goyal ret); 350a9bfd9ddSVivek Goyal fuse_request_end(fc, req); 351a9bfd9ddSVivek Goyal } 352a9bfd9ddSVivek Goyal } 353a62a8ef9SStefan Hajnoczi } 354a62a8ef9SStefan Hajnoczi 35558ada94fSVivek Goyal /* 35658ada94fSVivek Goyal * Returns 1 if queue is full and sender should wait a bit before sending 35758ada94fSVivek Goyal * next request, 0 otherwise. 35858ada94fSVivek Goyal */ 35958ada94fSVivek Goyal static int send_forget_request(struct virtio_fs_vq *fsvq, 36058ada94fSVivek Goyal struct virtio_fs_forget *forget, 36158ada94fSVivek Goyal bool in_flight) 36258ada94fSVivek Goyal { 36358ada94fSVivek Goyal struct scatterlist sg; 36458ada94fSVivek Goyal struct virtqueue *vq; 36558ada94fSVivek Goyal int ret = 0; 36658ada94fSVivek Goyal bool notify; 3671efcf39eSVivek Goyal struct virtio_fs_forget_req *req = &forget->req; 36858ada94fSVivek Goyal 36958ada94fSVivek Goyal spin_lock(&fsvq->lock); 37058ada94fSVivek Goyal if (!fsvq->connected) { 37158ada94fSVivek Goyal if (in_flight) 37258ada94fSVivek Goyal dec_in_flight_req(fsvq); 37358ada94fSVivek Goyal kfree(forget); 37458ada94fSVivek Goyal goto out; 37558ada94fSVivek Goyal } 37658ada94fSVivek Goyal 3771efcf39eSVivek Goyal sg_init_one(&sg, req, sizeof(*req)); 37858ada94fSVivek Goyal vq = fsvq->vq; 37958ada94fSVivek Goyal dev_dbg(&vq->vdev->dev, "%s\n", __func__); 38058ada94fSVivek Goyal 38158ada94fSVivek Goyal ret = virtqueue_add_outbuf(vq, &sg, 1, forget, GFP_ATOMIC); 38258ada94fSVivek Goyal if (ret < 0) { 38358ada94fSVivek Goyal if (ret == -ENOMEM || ret == -ENOSPC) { 38458ada94fSVivek Goyal pr_debug("virtio-fs: Could not queue FORGET: err=%d. Will try later\n", 38558ada94fSVivek Goyal ret); 38658ada94fSVivek Goyal list_add_tail(&forget->list, &fsvq->queued_reqs); 38758ada94fSVivek Goyal schedule_delayed_work(&fsvq->dispatch_work, 38858ada94fSVivek Goyal msecs_to_jiffies(1)); 38958ada94fSVivek Goyal if (!in_flight) 39058ada94fSVivek Goyal inc_in_flight_req(fsvq); 39158ada94fSVivek Goyal /* Queue is full */ 39258ada94fSVivek Goyal ret = 1; 39358ada94fSVivek Goyal } else { 39458ada94fSVivek Goyal pr_debug("virtio-fs: Could not queue FORGET: err=%d. Dropping it.\n", 39558ada94fSVivek Goyal ret); 39658ada94fSVivek Goyal kfree(forget); 39758ada94fSVivek Goyal if (in_flight) 39858ada94fSVivek Goyal dec_in_flight_req(fsvq); 39958ada94fSVivek Goyal } 40058ada94fSVivek Goyal goto out; 40158ada94fSVivek Goyal } 40258ada94fSVivek Goyal 40358ada94fSVivek Goyal if (!in_flight) 40458ada94fSVivek Goyal inc_in_flight_req(fsvq); 40558ada94fSVivek Goyal notify = virtqueue_kick_prepare(vq); 40658ada94fSVivek Goyal spin_unlock(&fsvq->lock); 40758ada94fSVivek Goyal 40858ada94fSVivek Goyal if (notify) 40958ada94fSVivek Goyal virtqueue_notify(vq); 41058ada94fSVivek Goyal return ret; 41158ada94fSVivek Goyal out: 41258ada94fSVivek Goyal spin_unlock(&fsvq->lock); 41358ada94fSVivek Goyal return ret; 41458ada94fSVivek Goyal } 41558ada94fSVivek Goyal 416a62a8ef9SStefan Hajnoczi static void virtio_fs_hiprio_dispatch_work(struct work_struct *work) 417a62a8ef9SStefan Hajnoczi { 418a62a8ef9SStefan Hajnoczi struct virtio_fs_forget *forget; 419a62a8ef9SStefan Hajnoczi struct virtio_fs_vq *fsvq = container_of(work, struct virtio_fs_vq, 420a62a8ef9SStefan Hajnoczi dispatch_work.work); 421a62a8ef9SStefan Hajnoczi pr_debug("virtio-fs: worker %s called.\n", __func__); 422a62a8ef9SStefan Hajnoczi while (1) { 423a62a8ef9SStefan Hajnoczi spin_lock(&fsvq->lock); 424a62a8ef9SStefan Hajnoczi forget = list_first_entry_or_null(&fsvq->queued_reqs, 425a62a8ef9SStefan Hajnoczi struct virtio_fs_forget, list); 426a62a8ef9SStefan Hajnoczi if (!forget) { 427a62a8ef9SStefan Hajnoczi spin_unlock(&fsvq->lock); 428a62a8ef9SStefan Hajnoczi return; 429a62a8ef9SStefan Hajnoczi } 430a62a8ef9SStefan Hajnoczi 431a62a8ef9SStefan Hajnoczi list_del(&forget->list); 432a62a8ef9SStefan Hajnoczi spin_unlock(&fsvq->lock); 43358ada94fSVivek Goyal if (send_forget_request(fsvq, forget, true)) 434a62a8ef9SStefan Hajnoczi return; 435a62a8ef9SStefan Hajnoczi } 436a62a8ef9SStefan Hajnoczi } 437a62a8ef9SStefan Hajnoczi 438a62a8ef9SStefan Hajnoczi /* Allocate and copy args into req->argbuf */ 439a62a8ef9SStefan Hajnoczi static int copy_args_to_argbuf(struct fuse_req *req) 440a62a8ef9SStefan Hajnoczi { 441a62a8ef9SStefan Hajnoczi struct fuse_args *args = req->args; 442a62a8ef9SStefan Hajnoczi unsigned int offset = 0; 443a62a8ef9SStefan Hajnoczi unsigned int num_in; 444a62a8ef9SStefan Hajnoczi unsigned int num_out; 445a62a8ef9SStefan Hajnoczi unsigned int len; 446a62a8ef9SStefan Hajnoczi unsigned int i; 447a62a8ef9SStefan Hajnoczi 448a62a8ef9SStefan Hajnoczi num_in = args->in_numargs - args->in_pages; 449a62a8ef9SStefan Hajnoczi num_out = args->out_numargs - args->out_pages; 450a62a8ef9SStefan Hajnoczi len = fuse_len_args(num_in, (struct fuse_arg *) args->in_args) + 451a62a8ef9SStefan Hajnoczi fuse_len_args(num_out, args->out_args); 452a62a8ef9SStefan Hajnoczi 453a62a8ef9SStefan Hajnoczi req->argbuf = kmalloc(len, GFP_ATOMIC); 454a62a8ef9SStefan Hajnoczi if (!req->argbuf) 455a62a8ef9SStefan Hajnoczi return -ENOMEM; 456a62a8ef9SStefan Hajnoczi 457a62a8ef9SStefan Hajnoczi for (i = 0; i < num_in; i++) { 458a62a8ef9SStefan Hajnoczi memcpy(req->argbuf + offset, 459a62a8ef9SStefan Hajnoczi args->in_args[i].value, 460a62a8ef9SStefan Hajnoczi args->in_args[i].size); 461a62a8ef9SStefan Hajnoczi offset += args->in_args[i].size; 462a62a8ef9SStefan Hajnoczi } 463a62a8ef9SStefan Hajnoczi 464a62a8ef9SStefan Hajnoczi return 0; 465a62a8ef9SStefan Hajnoczi } 466a62a8ef9SStefan Hajnoczi 467a62a8ef9SStefan Hajnoczi /* Copy args out of and free req->argbuf */ 468a62a8ef9SStefan Hajnoczi static void copy_args_from_argbuf(struct fuse_args *args, struct fuse_req *req) 469a62a8ef9SStefan Hajnoczi { 470a62a8ef9SStefan Hajnoczi unsigned int remaining; 471a62a8ef9SStefan Hajnoczi unsigned int offset; 472a62a8ef9SStefan Hajnoczi unsigned int num_in; 473a62a8ef9SStefan Hajnoczi unsigned int num_out; 474a62a8ef9SStefan Hajnoczi unsigned int i; 475a62a8ef9SStefan Hajnoczi 476a62a8ef9SStefan Hajnoczi remaining = req->out.h.len - sizeof(req->out.h); 477a62a8ef9SStefan Hajnoczi num_in = args->in_numargs - args->in_pages; 478a62a8ef9SStefan Hajnoczi num_out = args->out_numargs - args->out_pages; 479a62a8ef9SStefan Hajnoczi offset = fuse_len_args(num_in, (struct fuse_arg *)args->in_args); 480a62a8ef9SStefan Hajnoczi 481a62a8ef9SStefan Hajnoczi for (i = 0; i < num_out; i++) { 482a62a8ef9SStefan Hajnoczi unsigned int argsize = args->out_args[i].size; 483a62a8ef9SStefan Hajnoczi 484a62a8ef9SStefan Hajnoczi if (args->out_argvar && 485a62a8ef9SStefan Hajnoczi i == args->out_numargs - 1 && 486a62a8ef9SStefan Hajnoczi argsize > remaining) { 487a62a8ef9SStefan Hajnoczi argsize = remaining; 488a62a8ef9SStefan Hajnoczi } 489a62a8ef9SStefan Hajnoczi 490a62a8ef9SStefan Hajnoczi memcpy(args->out_args[i].value, req->argbuf + offset, argsize); 491a62a8ef9SStefan Hajnoczi offset += argsize; 492a62a8ef9SStefan Hajnoczi 493a62a8ef9SStefan Hajnoczi if (i != args->out_numargs - 1) 494a62a8ef9SStefan Hajnoczi remaining -= argsize; 495a62a8ef9SStefan Hajnoczi } 496a62a8ef9SStefan Hajnoczi 497a62a8ef9SStefan Hajnoczi /* Store the actual size of the variable-length arg */ 498a62a8ef9SStefan Hajnoczi if (args->out_argvar) 499a62a8ef9SStefan Hajnoczi args->out_args[args->out_numargs - 1].size = remaining; 500a62a8ef9SStefan Hajnoczi 501a62a8ef9SStefan Hajnoczi kfree(req->argbuf); 502a62a8ef9SStefan Hajnoczi req->argbuf = NULL; 503a62a8ef9SStefan Hajnoczi } 504a62a8ef9SStefan Hajnoczi 505a62a8ef9SStefan Hajnoczi /* Work function for request completion */ 506bb737bbeSVivek Goyal static void virtio_fs_request_complete(struct fuse_req *req, 507bb737bbeSVivek Goyal struct virtio_fs_vq *fsvq) 508a62a8ef9SStefan Hajnoczi { 509a62a8ef9SStefan Hajnoczi struct fuse_pqueue *fpq = &fsvq->fud->pq; 510a62a8ef9SStefan Hajnoczi struct fuse_conn *fc = fsvq->fud->fc; 511a62a8ef9SStefan Hajnoczi struct fuse_args *args; 512bb737bbeSVivek Goyal struct fuse_args_pages *ap; 513a62a8ef9SStefan Hajnoczi unsigned int len, i, thislen; 514a62a8ef9SStefan Hajnoczi struct page *page; 515a62a8ef9SStefan Hajnoczi 516a62a8ef9SStefan Hajnoczi /* 517a62a8ef9SStefan Hajnoczi * TODO verify that server properly follows FUSE protocol 518a62a8ef9SStefan Hajnoczi * (oh.uniq, oh.len) 519a62a8ef9SStefan Hajnoczi */ 520a62a8ef9SStefan Hajnoczi args = req->args; 521a62a8ef9SStefan Hajnoczi copy_args_from_argbuf(args, req); 522a62a8ef9SStefan Hajnoczi 523a62a8ef9SStefan Hajnoczi if (args->out_pages && args->page_zeroing) { 524a62a8ef9SStefan Hajnoczi len = args->out_args[args->out_numargs - 1].size; 525a62a8ef9SStefan Hajnoczi ap = container_of(args, typeof(*ap), args); 526a62a8ef9SStefan Hajnoczi for (i = 0; i < ap->num_pages; i++) { 527a62a8ef9SStefan Hajnoczi thislen = ap->descs[i].length; 528a62a8ef9SStefan Hajnoczi if (len < thislen) { 529a62a8ef9SStefan Hajnoczi WARN_ON(ap->descs[i].offset); 530a62a8ef9SStefan Hajnoczi page = ap->pages[i]; 531a62a8ef9SStefan Hajnoczi zero_user_segment(page, len, thislen); 532a62a8ef9SStefan Hajnoczi len = 0; 533a62a8ef9SStefan Hajnoczi } else { 534a62a8ef9SStefan Hajnoczi len -= thislen; 535a62a8ef9SStefan Hajnoczi } 536a62a8ef9SStefan Hajnoczi } 537a62a8ef9SStefan Hajnoczi } 538a62a8ef9SStefan Hajnoczi 539a62a8ef9SStefan Hajnoczi spin_lock(&fpq->lock); 540a62a8ef9SStefan Hajnoczi clear_bit(FR_SENT, &req->flags); 541a62a8ef9SStefan Hajnoczi spin_unlock(&fpq->lock); 542a62a8ef9SStefan Hajnoczi 543a62a8ef9SStefan Hajnoczi fuse_request_end(fc, req); 544a62a8ef9SStefan Hajnoczi spin_lock(&fsvq->lock); 545c17ea009SVivek Goyal dec_in_flight_req(fsvq); 546a62a8ef9SStefan Hajnoczi spin_unlock(&fsvq->lock); 547a62a8ef9SStefan Hajnoczi } 548bb737bbeSVivek Goyal 549bb737bbeSVivek Goyal static void virtio_fs_complete_req_work(struct work_struct *work) 550bb737bbeSVivek Goyal { 551bb737bbeSVivek Goyal struct virtio_fs_req_work *w = 552bb737bbeSVivek Goyal container_of(work, typeof(*w), done_work); 553bb737bbeSVivek Goyal 554bb737bbeSVivek Goyal virtio_fs_request_complete(w->req, w->fsvq); 555bb737bbeSVivek Goyal kfree(w); 556bb737bbeSVivek Goyal } 557bb737bbeSVivek Goyal 558bb737bbeSVivek Goyal static void virtio_fs_requests_done_work(struct work_struct *work) 559bb737bbeSVivek Goyal { 560bb737bbeSVivek Goyal struct virtio_fs_vq *fsvq = container_of(work, struct virtio_fs_vq, 561bb737bbeSVivek Goyal done_work); 562bb737bbeSVivek Goyal struct fuse_pqueue *fpq = &fsvq->fud->pq; 563bb737bbeSVivek Goyal struct virtqueue *vq = fsvq->vq; 564bb737bbeSVivek Goyal struct fuse_req *req; 565bb737bbeSVivek Goyal struct fuse_req *next; 566bb737bbeSVivek Goyal unsigned int len; 567bb737bbeSVivek Goyal LIST_HEAD(reqs); 568bb737bbeSVivek Goyal 569bb737bbeSVivek Goyal /* Collect completed requests off the virtqueue */ 570bb737bbeSVivek Goyal spin_lock(&fsvq->lock); 571bb737bbeSVivek Goyal do { 572bb737bbeSVivek Goyal virtqueue_disable_cb(vq); 573bb737bbeSVivek Goyal 574bb737bbeSVivek Goyal while ((req = virtqueue_get_buf(vq, &len)) != NULL) { 575bb737bbeSVivek Goyal spin_lock(&fpq->lock); 576bb737bbeSVivek Goyal list_move_tail(&req->list, &reqs); 577bb737bbeSVivek Goyal spin_unlock(&fpq->lock); 578bb737bbeSVivek Goyal } 579bb737bbeSVivek Goyal } while (!virtqueue_enable_cb(vq) && likely(!virtqueue_is_broken(vq))); 580bb737bbeSVivek Goyal spin_unlock(&fsvq->lock); 581bb737bbeSVivek Goyal 582bb737bbeSVivek Goyal /* End requests */ 583bb737bbeSVivek Goyal list_for_each_entry_safe(req, next, &reqs, list) { 584bb737bbeSVivek Goyal list_del_init(&req->list); 585bb737bbeSVivek Goyal 586bb737bbeSVivek Goyal /* blocking async request completes in a worker context */ 587bb737bbeSVivek Goyal if (req->args->may_block) { 588bb737bbeSVivek Goyal struct virtio_fs_req_work *w; 589bb737bbeSVivek Goyal 590bb737bbeSVivek Goyal w = kzalloc(sizeof(*w), GFP_NOFS | __GFP_NOFAIL); 591bb737bbeSVivek Goyal INIT_WORK(&w->done_work, virtio_fs_complete_req_work); 592bb737bbeSVivek Goyal w->fsvq = fsvq; 593bb737bbeSVivek Goyal w->req = req; 594bb737bbeSVivek Goyal schedule_work(&w->done_work); 595bb737bbeSVivek Goyal } else { 596bb737bbeSVivek Goyal virtio_fs_request_complete(req, fsvq); 597bb737bbeSVivek Goyal } 598bb737bbeSVivek Goyal } 599a62a8ef9SStefan Hajnoczi } 600a62a8ef9SStefan Hajnoczi 601a62a8ef9SStefan Hajnoczi /* Virtqueue interrupt handler */ 602a62a8ef9SStefan Hajnoczi static void virtio_fs_vq_done(struct virtqueue *vq) 603a62a8ef9SStefan Hajnoczi { 604a62a8ef9SStefan Hajnoczi struct virtio_fs_vq *fsvq = vq_to_fsvq(vq); 605a62a8ef9SStefan Hajnoczi 606a62a8ef9SStefan Hajnoczi dev_dbg(&vq->vdev->dev, "%s %s\n", __func__, fsvq->name); 607a62a8ef9SStefan Hajnoczi 608a62a8ef9SStefan Hajnoczi schedule_work(&fsvq->done_work); 609a62a8ef9SStefan Hajnoczi } 610a62a8ef9SStefan Hajnoczi 611b43b7e81SVivek Goyal static void virtio_fs_init_vq(struct virtio_fs_vq *fsvq, char *name, 612b43b7e81SVivek Goyal int vq_type) 613b43b7e81SVivek Goyal { 614b43b7e81SVivek Goyal strncpy(fsvq->name, name, VQ_NAME_LEN); 615b43b7e81SVivek Goyal spin_lock_init(&fsvq->lock); 616b43b7e81SVivek Goyal INIT_LIST_HEAD(&fsvq->queued_reqs); 617b43b7e81SVivek Goyal INIT_LIST_HEAD(&fsvq->end_reqs); 618b43b7e81SVivek Goyal init_completion(&fsvq->in_flight_zero); 619b43b7e81SVivek Goyal 620b43b7e81SVivek Goyal if (vq_type == VQ_REQUEST) { 621b43b7e81SVivek Goyal INIT_WORK(&fsvq->done_work, virtio_fs_requests_done_work); 622b43b7e81SVivek Goyal INIT_DELAYED_WORK(&fsvq->dispatch_work, 623b43b7e81SVivek Goyal virtio_fs_request_dispatch_work); 624b43b7e81SVivek Goyal } else { 625b43b7e81SVivek Goyal INIT_WORK(&fsvq->done_work, virtio_fs_hiprio_done_work); 626b43b7e81SVivek Goyal INIT_DELAYED_WORK(&fsvq->dispatch_work, 627b43b7e81SVivek Goyal virtio_fs_hiprio_dispatch_work); 628b43b7e81SVivek Goyal } 629b43b7e81SVivek Goyal } 630b43b7e81SVivek Goyal 631a62a8ef9SStefan Hajnoczi /* Initialize virtqueues */ 632a62a8ef9SStefan Hajnoczi static int virtio_fs_setup_vqs(struct virtio_device *vdev, 633a62a8ef9SStefan Hajnoczi struct virtio_fs *fs) 634a62a8ef9SStefan Hajnoczi { 635a62a8ef9SStefan Hajnoczi struct virtqueue **vqs; 636a62a8ef9SStefan Hajnoczi vq_callback_t **callbacks; 637a62a8ef9SStefan Hajnoczi const char **names; 638a62a8ef9SStefan Hajnoczi unsigned int i; 639a62a8ef9SStefan Hajnoczi int ret = 0; 640a62a8ef9SStefan Hajnoczi 6412c0349ecSMichael S. Tsirkin virtio_cread_le(vdev, struct virtio_fs_config, num_request_queues, 642a62a8ef9SStefan Hajnoczi &fs->num_request_queues); 643a62a8ef9SStefan Hajnoczi if (fs->num_request_queues == 0) 644a62a8ef9SStefan Hajnoczi return -EINVAL; 645a62a8ef9SStefan Hajnoczi 646b43b7e81SVivek Goyal fs->nvqs = VQ_REQUEST + fs->num_request_queues; 647a62a8ef9SStefan Hajnoczi fs->vqs = kcalloc(fs->nvqs, sizeof(fs->vqs[VQ_HIPRIO]), GFP_KERNEL); 648a62a8ef9SStefan Hajnoczi if (!fs->vqs) 649a62a8ef9SStefan Hajnoczi return -ENOMEM; 650a62a8ef9SStefan Hajnoczi 651a62a8ef9SStefan Hajnoczi vqs = kmalloc_array(fs->nvqs, sizeof(vqs[VQ_HIPRIO]), GFP_KERNEL); 652a62a8ef9SStefan Hajnoczi callbacks = kmalloc_array(fs->nvqs, sizeof(callbacks[VQ_HIPRIO]), 653a62a8ef9SStefan Hajnoczi GFP_KERNEL); 654a62a8ef9SStefan Hajnoczi names = kmalloc_array(fs->nvqs, sizeof(names[VQ_HIPRIO]), GFP_KERNEL); 655a62a8ef9SStefan Hajnoczi if (!vqs || !callbacks || !names) { 656a62a8ef9SStefan Hajnoczi ret = -ENOMEM; 657a62a8ef9SStefan Hajnoczi goto out; 658a62a8ef9SStefan Hajnoczi } 659a62a8ef9SStefan Hajnoczi 660b43b7e81SVivek Goyal /* Initialize the hiprio/forget request virtqueue */ 661a62a8ef9SStefan Hajnoczi callbacks[VQ_HIPRIO] = virtio_fs_vq_done; 662b43b7e81SVivek Goyal virtio_fs_init_vq(&fs->vqs[VQ_HIPRIO], "hiprio", VQ_HIPRIO); 663a62a8ef9SStefan Hajnoczi names[VQ_HIPRIO] = fs->vqs[VQ_HIPRIO].name; 664a62a8ef9SStefan Hajnoczi 665a62a8ef9SStefan Hajnoczi /* Initialize the requests virtqueues */ 666a62a8ef9SStefan Hajnoczi for (i = VQ_REQUEST; i < fs->nvqs; i++) { 667b43b7e81SVivek Goyal char vq_name[VQ_NAME_LEN]; 668b43b7e81SVivek Goyal 669b43b7e81SVivek Goyal snprintf(vq_name, VQ_NAME_LEN, "requests.%u", i - VQ_REQUEST); 670b43b7e81SVivek Goyal virtio_fs_init_vq(&fs->vqs[i], vq_name, VQ_REQUEST); 671a62a8ef9SStefan Hajnoczi callbacks[i] = virtio_fs_vq_done; 672a62a8ef9SStefan Hajnoczi names[i] = fs->vqs[i].name; 673a62a8ef9SStefan Hajnoczi } 674a62a8ef9SStefan Hajnoczi 675a62a8ef9SStefan Hajnoczi ret = virtio_find_vqs(vdev, fs->nvqs, vqs, callbacks, names, NULL); 676a62a8ef9SStefan Hajnoczi if (ret < 0) 677a62a8ef9SStefan Hajnoczi goto out; 678a62a8ef9SStefan Hajnoczi 679a62a8ef9SStefan Hajnoczi for (i = 0; i < fs->nvqs; i++) 680a62a8ef9SStefan Hajnoczi fs->vqs[i].vq = vqs[i]; 681a62a8ef9SStefan Hajnoczi 682a62a8ef9SStefan Hajnoczi virtio_fs_start_all_queues(fs); 683a62a8ef9SStefan Hajnoczi out: 684a62a8ef9SStefan Hajnoczi kfree(names); 685a62a8ef9SStefan Hajnoczi kfree(callbacks); 686a62a8ef9SStefan Hajnoczi kfree(vqs); 687a62a8ef9SStefan Hajnoczi if (ret) 688a62a8ef9SStefan Hajnoczi kfree(fs->vqs); 689a62a8ef9SStefan Hajnoczi return ret; 690a62a8ef9SStefan Hajnoczi } 691a62a8ef9SStefan Hajnoczi 692a62a8ef9SStefan Hajnoczi /* Free virtqueues (device must already be reset) */ 693a62a8ef9SStefan Hajnoczi static void virtio_fs_cleanup_vqs(struct virtio_device *vdev, 694a62a8ef9SStefan Hajnoczi struct virtio_fs *fs) 695a62a8ef9SStefan Hajnoczi { 696a62a8ef9SStefan Hajnoczi vdev->config->del_vqs(vdev); 697a62a8ef9SStefan Hajnoczi } 698a62a8ef9SStefan Hajnoczi 699*22f3787eSStefan Hajnoczi /* Map a window offset to a page frame number. The window offset will have 700*22f3787eSStefan Hajnoczi * been produced by .iomap_begin(), which maps a file offset to a window 701*22f3787eSStefan Hajnoczi * offset. 702*22f3787eSStefan Hajnoczi */ 703*22f3787eSStefan Hajnoczi static long virtio_fs_direct_access(struct dax_device *dax_dev, pgoff_t pgoff, 704*22f3787eSStefan Hajnoczi long nr_pages, void **kaddr, pfn_t *pfn) 705*22f3787eSStefan Hajnoczi { 706*22f3787eSStefan Hajnoczi struct virtio_fs *fs = dax_get_private(dax_dev); 707*22f3787eSStefan Hajnoczi phys_addr_t offset = PFN_PHYS(pgoff); 708*22f3787eSStefan Hajnoczi size_t max_nr_pages = fs->window_len/PAGE_SIZE - pgoff; 709*22f3787eSStefan Hajnoczi 710*22f3787eSStefan Hajnoczi if (kaddr) 711*22f3787eSStefan Hajnoczi *kaddr = fs->window_kaddr + offset; 712*22f3787eSStefan Hajnoczi if (pfn) 713*22f3787eSStefan Hajnoczi *pfn = phys_to_pfn_t(fs->window_phys_addr + offset, 714*22f3787eSStefan Hajnoczi PFN_DEV | PFN_MAP); 715*22f3787eSStefan Hajnoczi return nr_pages > max_nr_pages ? max_nr_pages : nr_pages; 716*22f3787eSStefan Hajnoczi } 717*22f3787eSStefan Hajnoczi 718*22f3787eSStefan Hajnoczi static size_t virtio_fs_copy_from_iter(struct dax_device *dax_dev, 719*22f3787eSStefan Hajnoczi pgoff_t pgoff, void *addr, 720*22f3787eSStefan Hajnoczi size_t bytes, struct iov_iter *i) 721*22f3787eSStefan Hajnoczi { 722*22f3787eSStefan Hajnoczi return copy_from_iter(addr, bytes, i); 723*22f3787eSStefan Hajnoczi } 724*22f3787eSStefan Hajnoczi 725*22f3787eSStefan Hajnoczi static size_t virtio_fs_copy_to_iter(struct dax_device *dax_dev, 726*22f3787eSStefan Hajnoczi pgoff_t pgoff, void *addr, 727*22f3787eSStefan Hajnoczi size_t bytes, struct iov_iter *i) 728*22f3787eSStefan Hajnoczi { 729*22f3787eSStefan Hajnoczi return copy_to_iter(addr, bytes, i); 730*22f3787eSStefan Hajnoczi } 731*22f3787eSStefan Hajnoczi 732*22f3787eSStefan Hajnoczi static int virtio_fs_zero_page_range(struct dax_device *dax_dev, 733*22f3787eSStefan Hajnoczi pgoff_t pgoff, size_t nr_pages) 734*22f3787eSStefan Hajnoczi { 735*22f3787eSStefan Hajnoczi long rc; 736*22f3787eSStefan Hajnoczi void *kaddr; 737*22f3787eSStefan Hajnoczi 738*22f3787eSStefan Hajnoczi rc = dax_direct_access(dax_dev, pgoff, nr_pages, &kaddr, NULL); 739*22f3787eSStefan Hajnoczi if (rc < 0) 740*22f3787eSStefan Hajnoczi return rc; 741*22f3787eSStefan Hajnoczi memset(kaddr, 0, nr_pages << PAGE_SHIFT); 742*22f3787eSStefan Hajnoczi dax_flush(dax_dev, kaddr, nr_pages << PAGE_SHIFT); 743*22f3787eSStefan Hajnoczi return 0; 744*22f3787eSStefan Hajnoczi } 745*22f3787eSStefan Hajnoczi 746*22f3787eSStefan Hajnoczi static const struct dax_operations virtio_fs_dax_ops = { 747*22f3787eSStefan Hajnoczi .direct_access = virtio_fs_direct_access, 748*22f3787eSStefan Hajnoczi .copy_from_iter = virtio_fs_copy_from_iter, 749*22f3787eSStefan Hajnoczi .copy_to_iter = virtio_fs_copy_to_iter, 750*22f3787eSStefan Hajnoczi .zero_page_range = virtio_fs_zero_page_range, 751*22f3787eSStefan Hajnoczi }; 752*22f3787eSStefan Hajnoczi 753*22f3787eSStefan Hajnoczi static void virtio_fs_cleanup_dax(void *data) 754*22f3787eSStefan Hajnoczi { 755*22f3787eSStefan Hajnoczi struct dax_device *dax_dev = data; 756*22f3787eSStefan Hajnoczi 757*22f3787eSStefan Hajnoczi kill_dax(dax_dev); 758*22f3787eSStefan Hajnoczi put_dax(dax_dev); 759*22f3787eSStefan Hajnoczi } 760*22f3787eSStefan Hajnoczi 761*22f3787eSStefan Hajnoczi static int virtio_fs_setup_dax(struct virtio_device *vdev, struct virtio_fs *fs) 762*22f3787eSStefan Hajnoczi { 763*22f3787eSStefan Hajnoczi struct virtio_shm_region cache_reg; 764*22f3787eSStefan Hajnoczi struct dev_pagemap *pgmap; 765*22f3787eSStefan Hajnoczi bool have_cache; 766*22f3787eSStefan Hajnoczi 767*22f3787eSStefan Hajnoczi if (!IS_ENABLED(CONFIG_FUSE_DAX)) 768*22f3787eSStefan Hajnoczi return 0; 769*22f3787eSStefan Hajnoczi 770*22f3787eSStefan Hajnoczi /* Get cache region */ 771*22f3787eSStefan Hajnoczi have_cache = virtio_get_shm_region(vdev, &cache_reg, 772*22f3787eSStefan Hajnoczi (u8)VIRTIO_FS_SHMCAP_ID_CACHE); 773*22f3787eSStefan Hajnoczi if (!have_cache) { 774*22f3787eSStefan Hajnoczi dev_notice(&vdev->dev, "%s: No cache capability\n", __func__); 775*22f3787eSStefan Hajnoczi return 0; 776*22f3787eSStefan Hajnoczi } 777*22f3787eSStefan Hajnoczi 778*22f3787eSStefan Hajnoczi if (!devm_request_mem_region(&vdev->dev, cache_reg.addr, cache_reg.len, 779*22f3787eSStefan Hajnoczi dev_name(&vdev->dev))) { 780*22f3787eSStefan Hajnoczi dev_warn(&vdev->dev, "could not reserve region addr=0x%llx len=0x%llx\n", 781*22f3787eSStefan Hajnoczi cache_reg.addr, cache_reg.len); 782*22f3787eSStefan Hajnoczi return -EBUSY; 783*22f3787eSStefan Hajnoczi } 784*22f3787eSStefan Hajnoczi 785*22f3787eSStefan Hajnoczi dev_notice(&vdev->dev, "Cache len: 0x%llx @ 0x%llx\n", cache_reg.len, 786*22f3787eSStefan Hajnoczi cache_reg.addr); 787*22f3787eSStefan Hajnoczi 788*22f3787eSStefan Hajnoczi pgmap = devm_kzalloc(&vdev->dev, sizeof(*pgmap), GFP_KERNEL); 789*22f3787eSStefan Hajnoczi if (!pgmap) 790*22f3787eSStefan Hajnoczi return -ENOMEM; 791*22f3787eSStefan Hajnoczi 792*22f3787eSStefan Hajnoczi pgmap->type = MEMORY_DEVICE_FS_DAX; 793*22f3787eSStefan Hajnoczi 794*22f3787eSStefan Hajnoczi /* Ideally we would directly use the PCI BAR resource but 795*22f3787eSStefan Hajnoczi * devm_memremap_pages() wants its own copy in pgmap. So 796*22f3787eSStefan Hajnoczi * initialize a struct resource from scratch (only the start 797*22f3787eSStefan Hajnoczi * and end fields will be used). 798*22f3787eSStefan Hajnoczi */ 799*22f3787eSStefan Hajnoczi pgmap->res = (struct resource){ 800*22f3787eSStefan Hajnoczi .name = "virtio-fs dax window", 801*22f3787eSStefan Hajnoczi .start = (phys_addr_t) cache_reg.addr, 802*22f3787eSStefan Hajnoczi .end = (phys_addr_t) cache_reg.addr + cache_reg.len - 1, 803*22f3787eSStefan Hajnoczi }; 804*22f3787eSStefan Hajnoczi 805*22f3787eSStefan Hajnoczi fs->window_kaddr = devm_memremap_pages(&vdev->dev, pgmap); 806*22f3787eSStefan Hajnoczi if (IS_ERR(fs->window_kaddr)) 807*22f3787eSStefan Hajnoczi return PTR_ERR(fs->window_kaddr); 808*22f3787eSStefan Hajnoczi 809*22f3787eSStefan Hajnoczi fs->window_phys_addr = (phys_addr_t) cache_reg.addr; 810*22f3787eSStefan Hajnoczi fs->window_len = (phys_addr_t) cache_reg.len; 811*22f3787eSStefan Hajnoczi 812*22f3787eSStefan Hajnoczi dev_dbg(&vdev->dev, "%s: window kaddr 0x%px phys_addr 0x%llx len 0x%llx\n", 813*22f3787eSStefan Hajnoczi __func__, fs->window_kaddr, cache_reg.addr, cache_reg.len); 814*22f3787eSStefan Hajnoczi 815*22f3787eSStefan Hajnoczi fs->dax_dev = alloc_dax(fs, NULL, &virtio_fs_dax_ops, 0); 816*22f3787eSStefan Hajnoczi if (IS_ERR(fs->dax_dev)) 817*22f3787eSStefan Hajnoczi return PTR_ERR(fs->dax_dev); 818*22f3787eSStefan Hajnoczi 819*22f3787eSStefan Hajnoczi return devm_add_action_or_reset(&vdev->dev, virtio_fs_cleanup_dax, 820*22f3787eSStefan Hajnoczi fs->dax_dev); 821*22f3787eSStefan Hajnoczi } 822*22f3787eSStefan Hajnoczi 823a62a8ef9SStefan Hajnoczi static int virtio_fs_probe(struct virtio_device *vdev) 824a62a8ef9SStefan Hajnoczi { 825a62a8ef9SStefan Hajnoczi struct virtio_fs *fs; 826a62a8ef9SStefan Hajnoczi int ret; 827a62a8ef9SStefan Hajnoczi 828a62a8ef9SStefan Hajnoczi fs = kzalloc(sizeof(*fs), GFP_KERNEL); 829a62a8ef9SStefan Hajnoczi if (!fs) 830a62a8ef9SStefan Hajnoczi return -ENOMEM; 831a62a8ef9SStefan Hajnoczi kref_init(&fs->refcount); 832a62a8ef9SStefan Hajnoczi vdev->priv = fs; 833a62a8ef9SStefan Hajnoczi 834a62a8ef9SStefan Hajnoczi ret = virtio_fs_read_tag(vdev, fs); 835a62a8ef9SStefan Hajnoczi if (ret < 0) 836a62a8ef9SStefan Hajnoczi goto out; 837a62a8ef9SStefan Hajnoczi 838a62a8ef9SStefan Hajnoczi ret = virtio_fs_setup_vqs(vdev, fs); 839a62a8ef9SStefan Hajnoczi if (ret < 0) 840a62a8ef9SStefan Hajnoczi goto out; 841a62a8ef9SStefan Hajnoczi 842a62a8ef9SStefan Hajnoczi /* TODO vq affinity */ 843a62a8ef9SStefan Hajnoczi 844*22f3787eSStefan Hajnoczi ret = virtio_fs_setup_dax(vdev, fs); 845*22f3787eSStefan Hajnoczi if (ret < 0) 846*22f3787eSStefan Hajnoczi goto out_vqs; 847*22f3787eSStefan Hajnoczi 848a62a8ef9SStefan Hajnoczi /* Bring the device online in case the filesystem is mounted and 849a62a8ef9SStefan Hajnoczi * requests need to be sent before we return. 850a62a8ef9SStefan Hajnoczi */ 851a62a8ef9SStefan Hajnoczi virtio_device_ready(vdev); 852a62a8ef9SStefan Hajnoczi 853a62a8ef9SStefan Hajnoczi ret = virtio_fs_add_instance(fs); 854a62a8ef9SStefan Hajnoczi if (ret < 0) 855a62a8ef9SStefan Hajnoczi goto out_vqs; 856a62a8ef9SStefan Hajnoczi 857a62a8ef9SStefan Hajnoczi return 0; 858a62a8ef9SStefan Hajnoczi 859a62a8ef9SStefan Hajnoczi out_vqs: 860a62a8ef9SStefan Hajnoczi vdev->config->reset(vdev); 861a62a8ef9SStefan Hajnoczi virtio_fs_cleanup_vqs(vdev, fs); 862a62a8ef9SStefan Hajnoczi 863a62a8ef9SStefan Hajnoczi out: 864a62a8ef9SStefan Hajnoczi vdev->priv = NULL; 865a62a8ef9SStefan Hajnoczi kfree(fs); 866a62a8ef9SStefan Hajnoczi return ret; 867a62a8ef9SStefan Hajnoczi } 868a62a8ef9SStefan Hajnoczi 869a62a8ef9SStefan Hajnoczi static void virtio_fs_stop_all_queues(struct virtio_fs *fs) 870a62a8ef9SStefan Hajnoczi { 871a62a8ef9SStefan Hajnoczi struct virtio_fs_vq *fsvq; 872a62a8ef9SStefan Hajnoczi int i; 873a62a8ef9SStefan Hajnoczi 874a62a8ef9SStefan Hajnoczi for (i = 0; i < fs->nvqs; i++) { 875a62a8ef9SStefan Hajnoczi fsvq = &fs->vqs[i]; 876a62a8ef9SStefan Hajnoczi spin_lock(&fsvq->lock); 877a62a8ef9SStefan Hajnoczi fsvq->connected = false; 878a62a8ef9SStefan Hajnoczi spin_unlock(&fsvq->lock); 879a62a8ef9SStefan Hajnoczi } 880a62a8ef9SStefan Hajnoczi } 881a62a8ef9SStefan Hajnoczi 882a62a8ef9SStefan Hajnoczi static void virtio_fs_remove(struct virtio_device *vdev) 883a62a8ef9SStefan Hajnoczi { 884a62a8ef9SStefan Hajnoczi struct virtio_fs *fs = vdev->priv; 885a62a8ef9SStefan Hajnoczi 886a62a8ef9SStefan Hajnoczi mutex_lock(&virtio_fs_mutex); 887a62a8ef9SStefan Hajnoczi /* This device is going away. No one should get new reference */ 888a62a8ef9SStefan Hajnoczi list_del_init(&fs->list); 889a62a8ef9SStefan Hajnoczi virtio_fs_stop_all_queues(fs); 890724c15a4SVivek Goyal virtio_fs_drain_all_queues_locked(fs); 891a62a8ef9SStefan Hajnoczi vdev->config->reset(vdev); 892a62a8ef9SStefan Hajnoczi virtio_fs_cleanup_vqs(vdev, fs); 893a62a8ef9SStefan Hajnoczi 894a62a8ef9SStefan Hajnoczi vdev->priv = NULL; 895a62a8ef9SStefan Hajnoczi /* Put device reference on virtio_fs object */ 896a62a8ef9SStefan Hajnoczi virtio_fs_put(fs); 897a62a8ef9SStefan Hajnoczi mutex_unlock(&virtio_fs_mutex); 898a62a8ef9SStefan Hajnoczi } 899a62a8ef9SStefan Hajnoczi 900a62a8ef9SStefan Hajnoczi #ifdef CONFIG_PM_SLEEP 901a62a8ef9SStefan Hajnoczi static int virtio_fs_freeze(struct virtio_device *vdev) 902a62a8ef9SStefan Hajnoczi { 903a62a8ef9SStefan Hajnoczi /* TODO need to save state here */ 904a62a8ef9SStefan Hajnoczi pr_warn("virtio-fs: suspend/resume not yet supported\n"); 905a62a8ef9SStefan Hajnoczi return -EOPNOTSUPP; 906a62a8ef9SStefan Hajnoczi } 907a62a8ef9SStefan Hajnoczi 908a62a8ef9SStefan Hajnoczi static int virtio_fs_restore(struct virtio_device *vdev) 909a62a8ef9SStefan Hajnoczi { 910a62a8ef9SStefan Hajnoczi /* TODO need to restore state here */ 911a62a8ef9SStefan Hajnoczi return 0; 912a62a8ef9SStefan Hajnoczi } 913a62a8ef9SStefan Hajnoczi #endif /* CONFIG_PM_SLEEP */ 914a62a8ef9SStefan Hajnoczi 91500929447SYueHaibing static const struct virtio_device_id id_table[] = { 916a62a8ef9SStefan Hajnoczi { VIRTIO_ID_FS, VIRTIO_DEV_ANY_ID }, 917a62a8ef9SStefan Hajnoczi {}, 918a62a8ef9SStefan Hajnoczi }; 919a62a8ef9SStefan Hajnoczi 92000929447SYueHaibing static const unsigned int feature_table[] = {}; 921a62a8ef9SStefan Hajnoczi 922a62a8ef9SStefan Hajnoczi static struct virtio_driver virtio_fs_driver = { 923a62a8ef9SStefan Hajnoczi .driver.name = KBUILD_MODNAME, 924a62a8ef9SStefan Hajnoczi .driver.owner = THIS_MODULE, 925a62a8ef9SStefan Hajnoczi .id_table = id_table, 926a62a8ef9SStefan Hajnoczi .feature_table = feature_table, 927a62a8ef9SStefan Hajnoczi .feature_table_size = ARRAY_SIZE(feature_table), 928a62a8ef9SStefan Hajnoczi .probe = virtio_fs_probe, 929a62a8ef9SStefan Hajnoczi .remove = virtio_fs_remove, 930a62a8ef9SStefan Hajnoczi #ifdef CONFIG_PM_SLEEP 931a62a8ef9SStefan Hajnoczi .freeze = virtio_fs_freeze, 932a62a8ef9SStefan Hajnoczi .restore = virtio_fs_restore, 933a62a8ef9SStefan Hajnoczi #endif 934a62a8ef9SStefan Hajnoczi }; 935a62a8ef9SStefan Hajnoczi 936a62a8ef9SStefan Hajnoczi static void virtio_fs_wake_forget_and_unlock(struct fuse_iqueue *fiq) 937a62a8ef9SStefan Hajnoczi __releases(fiq->lock) 938a62a8ef9SStefan Hajnoczi { 939a62a8ef9SStefan Hajnoczi struct fuse_forget_link *link; 940a62a8ef9SStefan Hajnoczi struct virtio_fs_forget *forget; 9411efcf39eSVivek Goyal struct virtio_fs_forget_req *req; 942a62a8ef9SStefan Hajnoczi struct virtio_fs *fs; 943a62a8ef9SStefan Hajnoczi struct virtio_fs_vq *fsvq; 944a62a8ef9SStefan Hajnoczi u64 unique; 945a62a8ef9SStefan Hajnoczi 946a62a8ef9SStefan Hajnoczi link = fuse_dequeue_forget(fiq, 1, NULL); 947a62a8ef9SStefan Hajnoczi unique = fuse_get_unique(fiq); 948a62a8ef9SStefan Hajnoczi 949a62a8ef9SStefan Hajnoczi fs = fiq->priv; 950a62a8ef9SStefan Hajnoczi fsvq = &fs->vqs[VQ_HIPRIO]; 951a62a8ef9SStefan Hajnoczi spin_unlock(&fiq->lock); 952a62a8ef9SStefan Hajnoczi 953a62a8ef9SStefan Hajnoczi /* Allocate a buffer for the request */ 954a62a8ef9SStefan Hajnoczi forget = kmalloc(sizeof(*forget), GFP_NOFS | __GFP_NOFAIL); 9551efcf39eSVivek Goyal req = &forget->req; 956a62a8ef9SStefan Hajnoczi 9571efcf39eSVivek Goyal req->ih = (struct fuse_in_header){ 958a62a8ef9SStefan Hajnoczi .opcode = FUSE_FORGET, 959a62a8ef9SStefan Hajnoczi .nodeid = link->forget_one.nodeid, 960a62a8ef9SStefan Hajnoczi .unique = unique, 9611efcf39eSVivek Goyal .len = sizeof(*req), 962a62a8ef9SStefan Hajnoczi }; 9631efcf39eSVivek Goyal req->arg = (struct fuse_forget_in){ 964a62a8ef9SStefan Hajnoczi .nlookup = link->forget_one.nlookup, 965a62a8ef9SStefan Hajnoczi }; 966a62a8ef9SStefan Hajnoczi 96758ada94fSVivek Goyal send_forget_request(fsvq, forget, false); 968a62a8ef9SStefan Hajnoczi kfree(link); 969a62a8ef9SStefan Hajnoczi } 970a62a8ef9SStefan Hajnoczi 971a62a8ef9SStefan Hajnoczi static void virtio_fs_wake_interrupt_and_unlock(struct fuse_iqueue *fiq) 972a62a8ef9SStefan Hajnoczi __releases(fiq->lock) 973a62a8ef9SStefan Hajnoczi { 974a62a8ef9SStefan Hajnoczi /* 975a62a8ef9SStefan Hajnoczi * TODO interrupts. 976a62a8ef9SStefan Hajnoczi * 977a62a8ef9SStefan Hajnoczi * Normal fs operations on a local filesystems aren't interruptible. 978a62a8ef9SStefan Hajnoczi * Exceptions are blocking lock operations; for example fcntl(F_SETLKW) 979a62a8ef9SStefan Hajnoczi * with shared lock between host and guest. 980a62a8ef9SStefan Hajnoczi */ 981a62a8ef9SStefan Hajnoczi spin_unlock(&fiq->lock); 982a62a8ef9SStefan Hajnoczi } 983a62a8ef9SStefan Hajnoczi 984a62a8ef9SStefan Hajnoczi /* Return the number of scatter-gather list elements required */ 985a62a8ef9SStefan Hajnoczi static unsigned int sg_count_fuse_req(struct fuse_req *req) 986a62a8ef9SStefan Hajnoczi { 987a62a8ef9SStefan Hajnoczi struct fuse_args *args = req->args; 988a62a8ef9SStefan Hajnoczi struct fuse_args_pages *ap = container_of(args, typeof(*ap), args); 989a62a8ef9SStefan Hajnoczi unsigned int total_sgs = 1 /* fuse_in_header */; 990a62a8ef9SStefan Hajnoczi 991a62a8ef9SStefan Hajnoczi if (args->in_numargs - args->in_pages) 992a62a8ef9SStefan Hajnoczi total_sgs += 1; 993a62a8ef9SStefan Hajnoczi 994a62a8ef9SStefan Hajnoczi if (args->in_pages) 995a62a8ef9SStefan Hajnoczi total_sgs += ap->num_pages; 996a62a8ef9SStefan Hajnoczi 997a62a8ef9SStefan Hajnoczi if (!test_bit(FR_ISREPLY, &req->flags)) 998a62a8ef9SStefan Hajnoczi return total_sgs; 999a62a8ef9SStefan Hajnoczi 1000a62a8ef9SStefan Hajnoczi total_sgs += 1 /* fuse_out_header */; 1001a62a8ef9SStefan Hajnoczi 1002a62a8ef9SStefan Hajnoczi if (args->out_numargs - args->out_pages) 1003a62a8ef9SStefan Hajnoczi total_sgs += 1; 1004a62a8ef9SStefan Hajnoczi 1005a62a8ef9SStefan Hajnoczi if (args->out_pages) 1006a62a8ef9SStefan Hajnoczi total_sgs += ap->num_pages; 1007a62a8ef9SStefan Hajnoczi 1008a62a8ef9SStefan Hajnoczi return total_sgs; 1009a62a8ef9SStefan Hajnoczi } 1010a62a8ef9SStefan Hajnoczi 1011a62a8ef9SStefan Hajnoczi /* Add pages to scatter-gather list and return number of elements used */ 1012a62a8ef9SStefan Hajnoczi static unsigned int sg_init_fuse_pages(struct scatterlist *sg, 1013a62a8ef9SStefan Hajnoczi struct page **pages, 1014a62a8ef9SStefan Hajnoczi struct fuse_page_desc *page_descs, 1015a62a8ef9SStefan Hajnoczi unsigned int num_pages, 1016a62a8ef9SStefan Hajnoczi unsigned int total_len) 1017a62a8ef9SStefan Hajnoczi { 1018a62a8ef9SStefan Hajnoczi unsigned int i; 1019a62a8ef9SStefan Hajnoczi unsigned int this_len; 1020a62a8ef9SStefan Hajnoczi 1021a62a8ef9SStefan Hajnoczi for (i = 0; i < num_pages && total_len; i++) { 1022a62a8ef9SStefan Hajnoczi sg_init_table(&sg[i], 1); 1023a62a8ef9SStefan Hajnoczi this_len = min(page_descs[i].length, total_len); 1024a62a8ef9SStefan Hajnoczi sg_set_page(&sg[i], pages[i], this_len, page_descs[i].offset); 1025a62a8ef9SStefan Hajnoczi total_len -= this_len; 1026a62a8ef9SStefan Hajnoczi } 1027a62a8ef9SStefan Hajnoczi 1028a62a8ef9SStefan Hajnoczi return i; 1029a62a8ef9SStefan Hajnoczi } 1030a62a8ef9SStefan Hajnoczi 1031a62a8ef9SStefan Hajnoczi /* Add args to scatter-gather list and return number of elements used */ 1032a62a8ef9SStefan Hajnoczi static unsigned int sg_init_fuse_args(struct scatterlist *sg, 1033a62a8ef9SStefan Hajnoczi struct fuse_req *req, 1034a62a8ef9SStefan Hajnoczi struct fuse_arg *args, 1035a62a8ef9SStefan Hajnoczi unsigned int numargs, 1036a62a8ef9SStefan Hajnoczi bool argpages, 1037a62a8ef9SStefan Hajnoczi void *argbuf, 1038a62a8ef9SStefan Hajnoczi unsigned int *len_used) 1039a62a8ef9SStefan Hajnoczi { 1040a62a8ef9SStefan Hajnoczi struct fuse_args_pages *ap = container_of(req->args, typeof(*ap), args); 1041a62a8ef9SStefan Hajnoczi unsigned int total_sgs = 0; 1042a62a8ef9SStefan Hajnoczi unsigned int len; 1043a62a8ef9SStefan Hajnoczi 1044a62a8ef9SStefan Hajnoczi len = fuse_len_args(numargs - argpages, args); 1045a62a8ef9SStefan Hajnoczi if (len) 1046a62a8ef9SStefan Hajnoczi sg_init_one(&sg[total_sgs++], argbuf, len); 1047a62a8ef9SStefan Hajnoczi 1048a62a8ef9SStefan Hajnoczi if (argpages) 1049a62a8ef9SStefan Hajnoczi total_sgs += sg_init_fuse_pages(&sg[total_sgs], 1050a62a8ef9SStefan Hajnoczi ap->pages, ap->descs, 1051a62a8ef9SStefan Hajnoczi ap->num_pages, 1052a62a8ef9SStefan Hajnoczi args[numargs - 1].size); 1053a62a8ef9SStefan Hajnoczi 1054a62a8ef9SStefan Hajnoczi if (len_used) 1055a62a8ef9SStefan Hajnoczi *len_used = len; 1056a62a8ef9SStefan Hajnoczi 1057a62a8ef9SStefan Hajnoczi return total_sgs; 1058a62a8ef9SStefan Hajnoczi } 1059a62a8ef9SStefan Hajnoczi 1060a62a8ef9SStefan Hajnoczi /* Add a request to a virtqueue and kick the device */ 1061a62a8ef9SStefan Hajnoczi static int virtio_fs_enqueue_req(struct virtio_fs_vq *fsvq, 1062a9bfd9ddSVivek Goyal struct fuse_req *req, bool in_flight) 1063a62a8ef9SStefan Hajnoczi { 1064a62a8ef9SStefan Hajnoczi /* requests need at least 4 elements */ 1065a62a8ef9SStefan Hajnoczi struct scatterlist *stack_sgs[6]; 1066a62a8ef9SStefan Hajnoczi struct scatterlist stack_sg[ARRAY_SIZE(stack_sgs)]; 1067a62a8ef9SStefan Hajnoczi struct scatterlist **sgs = stack_sgs; 1068a62a8ef9SStefan Hajnoczi struct scatterlist *sg = stack_sg; 1069a62a8ef9SStefan Hajnoczi struct virtqueue *vq; 1070a62a8ef9SStefan Hajnoczi struct fuse_args *args = req->args; 1071a62a8ef9SStefan Hajnoczi unsigned int argbuf_used = 0; 1072a62a8ef9SStefan Hajnoczi unsigned int out_sgs = 0; 1073a62a8ef9SStefan Hajnoczi unsigned int in_sgs = 0; 1074a62a8ef9SStefan Hajnoczi unsigned int total_sgs; 1075a62a8ef9SStefan Hajnoczi unsigned int i; 1076a62a8ef9SStefan Hajnoczi int ret; 1077a62a8ef9SStefan Hajnoczi bool notify; 10785dbe190fSVivek Goyal struct fuse_pqueue *fpq; 1079a62a8ef9SStefan Hajnoczi 1080a62a8ef9SStefan Hajnoczi /* Does the sglist fit on the stack? */ 1081a62a8ef9SStefan Hajnoczi total_sgs = sg_count_fuse_req(req); 1082a62a8ef9SStefan Hajnoczi if (total_sgs > ARRAY_SIZE(stack_sgs)) { 1083a62a8ef9SStefan Hajnoczi sgs = kmalloc_array(total_sgs, sizeof(sgs[0]), GFP_ATOMIC); 1084a62a8ef9SStefan Hajnoczi sg = kmalloc_array(total_sgs, sizeof(sg[0]), GFP_ATOMIC); 1085a62a8ef9SStefan Hajnoczi if (!sgs || !sg) { 1086a62a8ef9SStefan Hajnoczi ret = -ENOMEM; 1087a62a8ef9SStefan Hajnoczi goto out; 1088a62a8ef9SStefan Hajnoczi } 1089a62a8ef9SStefan Hajnoczi } 1090a62a8ef9SStefan Hajnoczi 1091a62a8ef9SStefan Hajnoczi /* Use a bounce buffer since stack args cannot be mapped */ 1092a62a8ef9SStefan Hajnoczi ret = copy_args_to_argbuf(req); 1093a62a8ef9SStefan Hajnoczi if (ret < 0) 1094a62a8ef9SStefan Hajnoczi goto out; 1095a62a8ef9SStefan Hajnoczi 1096a62a8ef9SStefan Hajnoczi /* Request elements */ 1097a62a8ef9SStefan Hajnoczi sg_init_one(&sg[out_sgs++], &req->in.h, sizeof(req->in.h)); 1098a62a8ef9SStefan Hajnoczi out_sgs += sg_init_fuse_args(&sg[out_sgs], req, 1099a62a8ef9SStefan Hajnoczi (struct fuse_arg *)args->in_args, 1100a62a8ef9SStefan Hajnoczi args->in_numargs, args->in_pages, 1101a62a8ef9SStefan Hajnoczi req->argbuf, &argbuf_used); 1102a62a8ef9SStefan Hajnoczi 1103a62a8ef9SStefan Hajnoczi /* Reply elements */ 1104a62a8ef9SStefan Hajnoczi if (test_bit(FR_ISREPLY, &req->flags)) { 1105a62a8ef9SStefan Hajnoczi sg_init_one(&sg[out_sgs + in_sgs++], 1106a62a8ef9SStefan Hajnoczi &req->out.h, sizeof(req->out.h)); 1107a62a8ef9SStefan Hajnoczi in_sgs += sg_init_fuse_args(&sg[out_sgs + in_sgs], req, 1108a62a8ef9SStefan Hajnoczi args->out_args, args->out_numargs, 1109a62a8ef9SStefan Hajnoczi args->out_pages, 1110a62a8ef9SStefan Hajnoczi req->argbuf + argbuf_used, NULL); 1111a62a8ef9SStefan Hajnoczi } 1112a62a8ef9SStefan Hajnoczi 1113a62a8ef9SStefan Hajnoczi WARN_ON(out_sgs + in_sgs != total_sgs); 1114a62a8ef9SStefan Hajnoczi 1115a62a8ef9SStefan Hajnoczi for (i = 0; i < total_sgs; i++) 1116a62a8ef9SStefan Hajnoczi sgs[i] = &sg[i]; 1117a62a8ef9SStefan Hajnoczi 1118a62a8ef9SStefan Hajnoczi spin_lock(&fsvq->lock); 1119a62a8ef9SStefan Hajnoczi 1120a62a8ef9SStefan Hajnoczi if (!fsvq->connected) { 1121a62a8ef9SStefan Hajnoczi spin_unlock(&fsvq->lock); 1122a62a8ef9SStefan Hajnoczi ret = -ENOTCONN; 1123a62a8ef9SStefan Hajnoczi goto out; 1124a62a8ef9SStefan Hajnoczi } 1125a62a8ef9SStefan Hajnoczi 1126a62a8ef9SStefan Hajnoczi vq = fsvq->vq; 1127a62a8ef9SStefan Hajnoczi ret = virtqueue_add_sgs(vq, sgs, out_sgs, in_sgs, req, GFP_ATOMIC); 1128a62a8ef9SStefan Hajnoczi if (ret < 0) { 1129a62a8ef9SStefan Hajnoczi spin_unlock(&fsvq->lock); 1130a62a8ef9SStefan Hajnoczi goto out; 1131a62a8ef9SStefan Hajnoczi } 1132a62a8ef9SStefan Hajnoczi 11335dbe190fSVivek Goyal /* Request successfully sent. */ 11345dbe190fSVivek Goyal fpq = &fsvq->fud->pq; 11355dbe190fSVivek Goyal spin_lock(&fpq->lock); 11365dbe190fSVivek Goyal list_add_tail(&req->list, fpq->processing); 11375dbe190fSVivek Goyal spin_unlock(&fpq->lock); 11385dbe190fSVivek Goyal set_bit(FR_SENT, &req->flags); 11395dbe190fSVivek Goyal /* matches barrier in request_wait_answer() */ 11405dbe190fSVivek Goyal smp_mb__after_atomic(); 11415dbe190fSVivek Goyal 1142a9bfd9ddSVivek Goyal if (!in_flight) 1143c17ea009SVivek Goyal inc_in_flight_req(fsvq); 1144a62a8ef9SStefan Hajnoczi notify = virtqueue_kick_prepare(vq); 1145a62a8ef9SStefan Hajnoczi 1146a62a8ef9SStefan Hajnoczi spin_unlock(&fsvq->lock); 1147a62a8ef9SStefan Hajnoczi 1148a62a8ef9SStefan Hajnoczi if (notify) 1149a62a8ef9SStefan Hajnoczi virtqueue_notify(vq); 1150a62a8ef9SStefan Hajnoczi 1151a62a8ef9SStefan Hajnoczi out: 1152a62a8ef9SStefan Hajnoczi if (ret < 0 && req->argbuf) { 1153a62a8ef9SStefan Hajnoczi kfree(req->argbuf); 1154a62a8ef9SStefan Hajnoczi req->argbuf = NULL; 1155a62a8ef9SStefan Hajnoczi } 1156a62a8ef9SStefan Hajnoczi if (sgs != stack_sgs) { 1157a62a8ef9SStefan Hajnoczi kfree(sgs); 1158a62a8ef9SStefan Hajnoczi kfree(sg); 1159a62a8ef9SStefan Hajnoczi } 1160a62a8ef9SStefan Hajnoczi 1161a62a8ef9SStefan Hajnoczi return ret; 1162a62a8ef9SStefan Hajnoczi } 1163a62a8ef9SStefan Hajnoczi 1164a62a8ef9SStefan Hajnoczi static void virtio_fs_wake_pending_and_unlock(struct fuse_iqueue *fiq) 1165a62a8ef9SStefan Hajnoczi __releases(fiq->lock) 1166a62a8ef9SStefan Hajnoczi { 1167a62a8ef9SStefan Hajnoczi unsigned int queue_id = VQ_REQUEST; /* TODO multiqueue */ 1168a62a8ef9SStefan Hajnoczi struct virtio_fs *fs; 1169a62a8ef9SStefan Hajnoczi struct fuse_req *req; 117051fecdd2SVivek Goyal struct virtio_fs_vq *fsvq; 1171a62a8ef9SStefan Hajnoczi int ret; 1172a62a8ef9SStefan Hajnoczi 1173a62a8ef9SStefan Hajnoczi WARN_ON(list_empty(&fiq->pending)); 1174a62a8ef9SStefan Hajnoczi req = list_last_entry(&fiq->pending, struct fuse_req, list); 1175a62a8ef9SStefan Hajnoczi clear_bit(FR_PENDING, &req->flags); 1176a62a8ef9SStefan Hajnoczi list_del_init(&req->list); 1177a62a8ef9SStefan Hajnoczi WARN_ON(!list_empty(&fiq->pending)); 1178a62a8ef9SStefan Hajnoczi spin_unlock(&fiq->lock); 1179a62a8ef9SStefan Hajnoczi 1180a62a8ef9SStefan Hajnoczi fs = fiq->priv; 1181a62a8ef9SStefan Hajnoczi 1182a62a8ef9SStefan Hajnoczi pr_debug("%s: opcode %u unique %#llx nodeid %#llx in.len %u out.len %u\n", 1183a62a8ef9SStefan Hajnoczi __func__, req->in.h.opcode, req->in.h.unique, 1184a62a8ef9SStefan Hajnoczi req->in.h.nodeid, req->in.h.len, 1185a62a8ef9SStefan Hajnoczi fuse_len_args(req->args->out_numargs, req->args->out_args)); 1186a62a8ef9SStefan Hajnoczi 118751fecdd2SVivek Goyal fsvq = &fs->vqs[queue_id]; 1188a9bfd9ddSVivek Goyal ret = virtio_fs_enqueue_req(fsvq, req, false); 1189a62a8ef9SStefan Hajnoczi if (ret < 0) { 1190a62a8ef9SStefan Hajnoczi if (ret == -ENOMEM || ret == -ENOSPC) { 1191a9bfd9ddSVivek Goyal /* 1192a9bfd9ddSVivek Goyal * Virtqueue full. Retry submission from worker 1193a9bfd9ddSVivek Goyal * context as we might be holding fc->bg_lock. 1194a9bfd9ddSVivek Goyal */ 1195a9bfd9ddSVivek Goyal spin_lock(&fsvq->lock); 1196a9bfd9ddSVivek Goyal list_add_tail(&req->list, &fsvq->queued_reqs); 1197a9bfd9ddSVivek Goyal inc_in_flight_req(fsvq); 1198a9bfd9ddSVivek Goyal schedule_delayed_work(&fsvq->dispatch_work, 1199a9bfd9ddSVivek Goyal msecs_to_jiffies(1)); 1200a9bfd9ddSVivek Goyal spin_unlock(&fsvq->lock); 1201a9bfd9ddSVivek Goyal return; 1202a62a8ef9SStefan Hajnoczi } 1203a62a8ef9SStefan Hajnoczi req->out.h.error = ret; 1204a62a8ef9SStefan Hajnoczi pr_err("virtio-fs: virtio_fs_enqueue_req() failed %d\n", ret); 120551fecdd2SVivek Goyal 120651fecdd2SVivek Goyal /* Can't end request in submission context. Use a worker */ 120751fecdd2SVivek Goyal spin_lock(&fsvq->lock); 120851fecdd2SVivek Goyal list_add_tail(&req->list, &fsvq->end_reqs); 120951fecdd2SVivek Goyal schedule_delayed_work(&fsvq->dispatch_work, 0); 121051fecdd2SVivek Goyal spin_unlock(&fsvq->lock); 1211a62a8ef9SStefan Hajnoczi return; 1212a62a8ef9SStefan Hajnoczi } 1213a62a8ef9SStefan Hajnoczi } 1214a62a8ef9SStefan Hajnoczi 121500929447SYueHaibing static const struct fuse_iqueue_ops virtio_fs_fiq_ops = { 1216a62a8ef9SStefan Hajnoczi .wake_forget_and_unlock = virtio_fs_wake_forget_and_unlock, 1217a62a8ef9SStefan Hajnoczi .wake_interrupt_and_unlock = virtio_fs_wake_interrupt_and_unlock, 1218a62a8ef9SStefan Hajnoczi .wake_pending_and_unlock = virtio_fs_wake_pending_and_unlock, 1219a62a8ef9SStefan Hajnoczi .release = virtio_fs_fiq_release, 1220a62a8ef9SStefan Hajnoczi }; 1221a62a8ef9SStefan Hajnoczi 1222a62a8ef9SStefan Hajnoczi static int virtio_fs_fill_super(struct super_block *sb) 1223a62a8ef9SStefan Hajnoczi { 1224a62a8ef9SStefan Hajnoczi struct fuse_conn *fc = get_fuse_conn_super(sb); 1225a62a8ef9SStefan Hajnoczi struct virtio_fs *fs = fc->iq.priv; 1226a62a8ef9SStefan Hajnoczi unsigned int i; 1227a62a8ef9SStefan Hajnoczi int err; 1228a62a8ef9SStefan Hajnoczi struct fuse_fs_context ctx = { 1229a62a8ef9SStefan Hajnoczi .rootmode = S_IFDIR, 1230a62a8ef9SStefan Hajnoczi .default_permissions = 1, 1231a62a8ef9SStefan Hajnoczi .allow_other = 1, 1232a62a8ef9SStefan Hajnoczi .max_read = UINT_MAX, 1233a62a8ef9SStefan Hajnoczi .blksize = 512, 1234a62a8ef9SStefan Hajnoczi .destroy = true, 1235a62a8ef9SStefan Hajnoczi .no_control = true, 1236a62a8ef9SStefan Hajnoczi .no_force_umount = true, 1237a62a8ef9SStefan Hajnoczi }; 1238a62a8ef9SStefan Hajnoczi 1239a62a8ef9SStefan Hajnoczi mutex_lock(&virtio_fs_mutex); 1240a62a8ef9SStefan Hajnoczi 1241a62a8ef9SStefan Hajnoczi /* After holding mutex, make sure virtiofs device is still there. 1242a62a8ef9SStefan Hajnoczi * Though we are holding a reference to it, drive ->remove might 1243a62a8ef9SStefan Hajnoczi * still have cleaned up virtual queues. In that case bail out. 1244a62a8ef9SStefan Hajnoczi */ 1245a62a8ef9SStefan Hajnoczi err = -EINVAL; 1246a62a8ef9SStefan Hajnoczi if (list_empty(&fs->list)) { 1247a62a8ef9SStefan Hajnoczi pr_info("virtio-fs: tag <%s> not found\n", fs->tag); 1248a62a8ef9SStefan Hajnoczi goto err; 1249a62a8ef9SStefan Hajnoczi } 1250a62a8ef9SStefan Hajnoczi 1251a62a8ef9SStefan Hajnoczi err = -ENOMEM; 1252a62a8ef9SStefan Hajnoczi /* Allocate fuse_dev for hiprio and notification queues */ 12537fd3abfaSVivek Goyal for (i = 0; i < fs->nvqs; i++) { 1254a62a8ef9SStefan Hajnoczi struct virtio_fs_vq *fsvq = &fs->vqs[i]; 1255a62a8ef9SStefan Hajnoczi 1256a62a8ef9SStefan Hajnoczi fsvq->fud = fuse_dev_alloc(); 1257a62a8ef9SStefan Hajnoczi if (!fsvq->fud) 1258a62a8ef9SStefan Hajnoczi goto err_free_fuse_devs; 1259a62a8ef9SStefan Hajnoczi } 1260a62a8ef9SStefan Hajnoczi 12617fd3abfaSVivek Goyal /* virtiofs allocates and installs its own fuse devices */ 12627fd3abfaSVivek Goyal ctx.fudptr = NULL; 1263a62a8ef9SStefan Hajnoczi err = fuse_fill_super_common(sb, &ctx); 1264a62a8ef9SStefan Hajnoczi if (err < 0) 1265a62a8ef9SStefan Hajnoczi goto err_free_fuse_devs; 1266a62a8ef9SStefan Hajnoczi 1267a62a8ef9SStefan Hajnoczi for (i = 0; i < fs->nvqs; i++) { 1268a62a8ef9SStefan Hajnoczi struct virtio_fs_vq *fsvq = &fs->vqs[i]; 1269a62a8ef9SStefan Hajnoczi 1270a62a8ef9SStefan Hajnoczi fuse_dev_install(fsvq->fud, fc); 1271a62a8ef9SStefan Hajnoczi } 1272a62a8ef9SStefan Hajnoczi 1273a62a8ef9SStefan Hajnoczi /* Previous unmount will stop all queues. Start these again */ 1274a62a8ef9SStefan Hajnoczi virtio_fs_start_all_queues(fs); 1275a62a8ef9SStefan Hajnoczi fuse_send_init(fc); 1276a62a8ef9SStefan Hajnoczi mutex_unlock(&virtio_fs_mutex); 1277a62a8ef9SStefan Hajnoczi return 0; 1278a62a8ef9SStefan Hajnoczi 1279a62a8ef9SStefan Hajnoczi err_free_fuse_devs: 1280a62a8ef9SStefan Hajnoczi virtio_fs_free_devs(fs); 1281a62a8ef9SStefan Hajnoczi err: 1282a62a8ef9SStefan Hajnoczi mutex_unlock(&virtio_fs_mutex); 1283a62a8ef9SStefan Hajnoczi return err; 1284a62a8ef9SStefan Hajnoczi } 1285a62a8ef9SStefan Hajnoczi 1286a62a8ef9SStefan Hajnoczi static void virtio_kill_sb(struct super_block *sb) 1287a62a8ef9SStefan Hajnoczi { 1288a62a8ef9SStefan Hajnoczi struct fuse_conn *fc = get_fuse_conn_super(sb); 1289a62a8ef9SStefan Hajnoczi struct virtio_fs *vfs; 1290a62a8ef9SStefan Hajnoczi struct virtio_fs_vq *fsvq; 1291a62a8ef9SStefan Hajnoczi 1292a62a8ef9SStefan Hajnoczi /* If mount failed, we can still be called without any fc */ 1293a62a8ef9SStefan Hajnoczi if (!fc) 1294a62a8ef9SStefan Hajnoczi return fuse_kill_sb_anon(sb); 1295a62a8ef9SStefan Hajnoczi 1296a62a8ef9SStefan Hajnoczi vfs = fc->iq.priv; 1297a62a8ef9SStefan Hajnoczi fsvq = &vfs->vqs[VQ_HIPRIO]; 1298a62a8ef9SStefan Hajnoczi 1299a62a8ef9SStefan Hajnoczi /* Stop forget queue. Soon destroy will be sent */ 1300a62a8ef9SStefan Hajnoczi spin_lock(&fsvq->lock); 1301a62a8ef9SStefan Hajnoczi fsvq->connected = false; 1302a62a8ef9SStefan Hajnoczi spin_unlock(&fsvq->lock); 1303a62a8ef9SStefan Hajnoczi virtio_fs_drain_all_queues(vfs); 1304a62a8ef9SStefan Hajnoczi 1305a62a8ef9SStefan Hajnoczi fuse_kill_sb_anon(sb); 1306a62a8ef9SStefan Hajnoczi 1307a62a8ef9SStefan Hajnoczi /* fuse_kill_sb_anon() must have sent destroy. Stop all queues 1308a62a8ef9SStefan Hajnoczi * and drain one more time and free fuse devices. Freeing fuse 1309a62a8ef9SStefan Hajnoczi * devices will drop their reference on fuse_conn and that in 1310a62a8ef9SStefan Hajnoczi * turn will drop its reference on virtio_fs object. 1311a62a8ef9SStefan Hajnoczi */ 1312a62a8ef9SStefan Hajnoczi virtio_fs_stop_all_queues(vfs); 1313a62a8ef9SStefan Hajnoczi virtio_fs_drain_all_queues(vfs); 1314a62a8ef9SStefan Hajnoczi virtio_fs_free_devs(vfs); 1315a62a8ef9SStefan Hajnoczi } 1316a62a8ef9SStefan Hajnoczi 1317a62a8ef9SStefan Hajnoczi static int virtio_fs_test_super(struct super_block *sb, 1318a62a8ef9SStefan Hajnoczi struct fs_context *fsc) 1319a62a8ef9SStefan Hajnoczi { 1320a62a8ef9SStefan Hajnoczi struct fuse_conn *fc = fsc->s_fs_info; 1321a62a8ef9SStefan Hajnoczi 1322a62a8ef9SStefan Hajnoczi return fc->iq.priv == get_fuse_conn_super(sb)->iq.priv; 1323a62a8ef9SStefan Hajnoczi } 1324a62a8ef9SStefan Hajnoczi 1325a62a8ef9SStefan Hajnoczi static int virtio_fs_set_super(struct super_block *sb, 1326a62a8ef9SStefan Hajnoczi struct fs_context *fsc) 1327a62a8ef9SStefan Hajnoczi { 1328a62a8ef9SStefan Hajnoczi int err; 1329a62a8ef9SStefan Hajnoczi 1330a62a8ef9SStefan Hajnoczi err = get_anon_bdev(&sb->s_dev); 1331a62a8ef9SStefan Hajnoczi if (!err) 1332a62a8ef9SStefan Hajnoczi fuse_conn_get(fsc->s_fs_info); 1333a62a8ef9SStefan Hajnoczi 1334a62a8ef9SStefan Hajnoczi return err; 1335a62a8ef9SStefan Hajnoczi } 1336a62a8ef9SStefan Hajnoczi 1337a62a8ef9SStefan Hajnoczi static int virtio_fs_get_tree(struct fs_context *fsc) 1338a62a8ef9SStefan Hajnoczi { 1339a62a8ef9SStefan Hajnoczi struct virtio_fs *fs; 1340a62a8ef9SStefan Hajnoczi struct super_block *sb; 1341a62a8ef9SStefan Hajnoczi struct fuse_conn *fc; 1342a62a8ef9SStefan Hajnoczi int err; 1343a62a8ef9SStefan Hajnoczi 1344a62a8ef9SStefan Hajnoczi /* This gets a reference on virtio_fs object. This ptr gets installed 1345a62a8ef9SStefan Hajnoczi * in fc->iq->priv. Once fuse_conn is going away, it calls ->put() 1346a62a8ef9SStefan Hajnoczi * to drop the reference to this object. 1347a62a8ef9SStefan Hajnoczi */ 1348a62a8ef9SStefan Hajnoczi fs = virtio_fs_find_instance(fsc->source); 1349a62a8ef9SStefan Hajnoczi if (!fs) { 1350a62a8ef9SStefan Hajnoczi pr_info("virtio-fs: tag <%s> not found\n", fsc->source); 1351a62a8ef9SStefan Hajnoczi return -EINVAL; 1352a62a8ef9SStefan Hajnoczi } 1353a62a8ef9SStefan Hajnoczi 1354a62a8ef9SStefan Hajnoczi fc = kzalloc(sizeof(struct fuse_conn), GFP_KERNEL); 1355a62a8ef9SStefan Hajnoczi if (!fc) { 1356a62a8ef9SStefan Hajnoczi mutex_lock(&virtio_fs_mutex); 1357a62a8ef9SStefan Hajnoczi virtio_fs_put(fs); 1358a62a8ef9SStefan Hajnoczi mutex_unlock(&virtio_fs_mutex); 1359a62a8ef9SStefan Hajnoczi return -ENOMEM; 1360a62a8ef9SStefan Hajnoczi } 1361a62a8ef9SStefan Hajnoczi 1362a62a8ef9SStefan Hajnoczi fuse_conn_init(fc, get_user_ns(current_user_ns()), &virtio_fs_fiq_ops, 1363a62a8ef9SStefan Hajnoczi fs); 1364a62a8ef9SStefan Hajnoczi fc->release = fuse_free_conn; 1365a62a8ef9SStefan Hajnoczi fc->delete_stale = true; 1366a62a8ef9SStefan Hajnoczi 1367a62a8ef9SStefan Hajnoczi fsc->s_fs_info = fc; 1368a62a8ef9SStefan Hajnoczi sb = sget_fc(fsc, virtio_fs_test_super, virtio_fs_set_super); 1369a62a8ef9SStefan Hajnoczi fuse_conn_put(fc); 1370a62a8ef9SStefan Hajnoczi if (IS_ERR(sb)) 1371a62a8ef9SStefan Hajnoczi return PTR_ERR(sb); 1372a62a8ef9SStefan Hajnoczi 1373a62a8ef9SStefan Hajnoczi if (!sb->s_root) { 1374a62a8ef9SStefan Hajnoczi err = virtio_fs_fill_super(sb); 1375a62a8ef9SStefan Hajnoczi if (err) { 1376a62a8ef9SStefan Hajnoczi deactivate_locked_super(sb); 1377a62a8ef9SStefan Hajnoczi return err; 1378a62a8ef9SStefan Hajnoczi } 1379a62a8ef9SStefan Hajnoczi 1380a62a8ef9SStefan Hajnoczi sb->s_flags |= SB_ACTIVE; 1381a62a8ef9SStefan Hajnoczi } 1382a62a8ef9SStefan Hajnoczi 1383a62a8ef9SStefan Hajnoczi WARN_ON(fsc->root); 1384a62a8ef9SStefan Hajnoczi fsc->root = dget(sb->s_root); 1385a62a8ef9SStefan Hajnoczi return 0; 1386a62a8ef9SStefan Hajnoczi } 1387a62a8ef9SStefan Hajnoczi 1388a62a8ef9SStefan Hajnoczi static const struct fs_context_operations virtio_fs_context_ops = { 1389a62a8ef9SStefan Hajnoczi .get_tree = virtio_fs_get_tree, 1390a62a8ef9SStefan Hajnoczi }; 1391a62a8ef9SStefan Hajnoczi 1392a62a8ef9SStefan Hajnoczi static int virtio_fs_init_fs_context(struct fs_context *fsc) 1393a62a8ef9SStefan Hajnoczi { 1394a62a8ef9SStefan Hajnoczi fsc->ops = &virtio_fs_context_ops; 1395a62a8ef9SStefan Hajnoczi return 0; 1396a62a8ef9SStefan Hajnoczi } 1397a62a8ef9SStefan Hajnoczi 1398a62a8ef9SStefan Hajnoczi static struct file_system_type virtio_fs_type = { 1399a62a8ef9SStefan Hajnoczi .owner = THIS_MODULE, 1400a62a8ef9SStefan Hajnoczi .name = "virtiofs", 1401a62a8ef9SStefan Hajnoczi .init_fs_context = virtio_fs_init_fs_context, 1402a62a8ef9SStefan Hajnoczi .kill_sb = virtio_kill_sb, 1403a62a8ef9SStefan Hajnoczi }; 1404a62a8ef9SStefan Hajnoczi 1405a62a8ef9SStefan Hajnoczi static int __init virtio_fs_init(void) 1406a62a8ef9SStefan Hajnoczi { 1407a62a8ef9SStefan Hajnoczi int ret; 1408a62a8ef9SStefan Hajnoczi 1409a62a8ef9SStefan Hajnoczi ret = register_virtio_driver(&virtio_fs_driver); 1410a62a8ef9SStefan Hajnoczi if (ret < 0) 1411a62a8ef9SStefan Hajnoczi return ret; 1412a62a8ef9SStefan Hajnoczi 1413a62a8ef9SStefan Hajnoczi ret = register_filesystem(&virtio_fs_type); 1414a62a8ef9SStefan Hajnoczi if (ret < 0) { 1415a62a8ef9SStefan Hajnoczi unregister_virtio_driver(&virtio_fs_driver); 1416a62a8ef9SStefan Hajnoczi return ret; 1417a62a8ef9SStefan Hajnoczi } 1418a62a8ef9SStefan Hajnoczi 1419a62a8ef9SStefan Hajnoczi return 0; 1420a62a8ef9SStefan Hajnoczi } 1421a62a8ef9SStefan Hajnoczi module_init(virtio_fs_init); 1422a62a8ef9SStefan Hajnoczi 1423a62a8ef9SStefan Hajnoczi static void __exit virtio_fs_exit(void) 1424a62a8ef9SStefan Hajnoczi { 1425a62a8ef9SStefan Hajnoczi unregister_filesystem(&virtio_fs_type); 1426a62a8ef9SStefan Hajnoczi unregister_virtio_driver(&virtio_fs_driver); 1427a62a8ef9SStefan Hajnoczi } 1428a62a8ef9SStefan Hajnoczi module_exit(virtio_fs_exit); 1429a62a8ef9SStefan Hajnoczi 1430a62a8ef9SStefan Hajnoczi MODULE_AUTHOR("Stefan Hajnoczi <stefanha@redhat.com>"); 1431a62a8ef9SStefan Hajnoczi MODULE_DESCRIPTION("Virtio Filesystem"); 1432a62a8ef9SStefan Hajnoczi MODULE_LICENSE("GPL"); 1433a62a8ef9SStefan Hajnoczi MODULE_ALIAS_FS(KBUILD_MODNAME); 1434a62a8ef9SStefan Hajnoczi MODULE_DEVICE_TABLE(virtio, id_table); 1435