xref: /openbmc/linux/fs/fuse/virtio_fs.c (revision 22f3787e9d95e72d1f09795f294fb010e2998f43)
1a62a8ef9SStefan Hajnoczi // SPDX-License-Identifier: GPL-2.0
2a62a8ef9SStefan Hajnoczi /*
3a62a8ef9SStefan Hajnoczi  * virtio-fs: Virtio Filesystem
4a62a8ef9SStefan Hajnoczi  * Copyright (C) 2018 Red Hat, Inc.
5a62a8ef9SStefan Hajnoczi  */
6a62a8ef9SStefan Hajnoczi 
7a62a8ef9SStefan Hajnoczi #include <linux/fs.h>
8*22f3787eSStefan Hajnoczi #include <linux/dax.h>
9*22f3787eSStefan Hajnoczi #include <linux/pci.h>
10*22f3787eSStefan Hajnoczi #include <linux/pfn_t.h>
11a62a8ef9SStefan Hajnoczi #include <linux/module.h>
12a62a8ef9SStefan Hajnoczi #include <linux/virtio.h>
13a62a8ef9SStefan Hajnoczi #include <linux/virtio_fs.h>
14a62a8ef9SStefan Hajnoczi #include <linux/delay.h>
15a62a8ef9SStefan Hajnoczi #include <linux/fs_context.h>
16a62a8ef9SStefan Hajnoczi #include <linux/highmem.h>
17*22f3787eSStefan Hajnoczi #include <linux/uio.h>
18a62a8ef9SStefan Hajnoczi #include "fuse_i.h"
19a62a8ef9SStefan Hajnoczi 
20a62a8ef9SStefan Hajnoczi /* List of virtio-fs device instances and a lock for the list. Also provides
21a62a8ef9SStefan Hajnoczi  * mutual exclusion in device removal and mounting path
22a62a8ef9SStefan Hajnoczi  */
23a62a8ef9SStefan Hajnoczi static DEFINE_MUTEX(virtio_fs_mutex);
24a62a8ef9SStefan Hajnoczi static LIST_HEAD(virtio_fs_instances);
25a62a8ef9SStefan Hajnoczi 
26a62a8ef9SStefan Hajnoczi enum {
27a62a8ef9SStefan Hajnoczi 	VQ_HIPRIO,
28a62a8ef9SStefan Hajnoczi 	VQ_REQUEST
29a62a8ef9SStefan Hajnoczi };
30a62a8ef9SStefan Hajnoczi 
31b43b7e81SVivek Goyal #define VQ_NAME_LEN	24
32b43b7e81SVivek Goyal 
33a62a8ef9SStefan Hajnoczi /* Per-virtqueue state */
34a62a8ef9SStefan Hajnoczi struct virtio_fs_vq {
35a62a8ef9SStefan Hajnoczi 	spinlock_t lock;
36a62a8ef9SStefan Hajnoczi 	struct virtqueue *vq;     /* protected by ->lock */
37a62a8ef9SStefan Hajnoczi 	struct work_struct done_work;
38a62a8ef9SStefan Hajnoczi 	struct list_head queued_reqs;
3951fecdd2SVivek Goyal 	struct list_head end_reqs;	/* End these requests */
40a62a8ef9SStefan Hajnoczi 	struct delayed_work dispatch_work;
41a62a8ef9SStefan Hajnoczi 	struct fuse_dev *fud;
42a62a8ef9SStefan Hajnoczi 	bool connected;
43a62a8ef9SStefan Hajnoczi 	long in_flight;
44724c15a4SVivek Goyal 	struct completion in_flight_zero; /* No inflight requests */
45b43b7e81SVivek Goyal 	char name[VQ_NAME_LEN];
46a62a8ef9SStefan Hajnoczi } ____cacheline_aligned_in_smp;
47a62a8ef9SStefan Hajnoczi 
48a62a8ef9SStefan Hajnoczi /* A virtio-fs device instance */
49a62a8ef9SStefan Hajnoczi struct virtio_fs {
50a62a8ef9SStefan Hajnoczi 	struct kref refcount;
51a62a8ef9SStefan Hajnoczi 	struct list_head list;    /* on virtio_fs_instances */
52a62a8ef9SStefan Hajnoczi 	char *tag;
53a62a8ef9SStefan Hajnoczi 	struct virtio_fs_vq *vqs;
54a62a8ef9SStefan Hajnoczi 	unsigned int nvqs;               /* number of virtqueues */
55a62a8ef9SStefan Hajnoczi 	unsigned int num_request_queues; /* number of request queues */
56*22f3787eSStefan Hajnoczi 	struct dax_device *dax_dev;
57*22f3787eSStefan Hajnoczi 
58*22f3787eSStefan Hajnoczi 	/* DAX memory window where file contents are mapped */
59*22f3787eSStefan Hajnoczi 	void *window_kaddr;
60*22f3787eSStefan Hajnoczi 	phys_addr_t window_phys_addr;
61*22f3787eSStefan Hajnoczi 	size_t window_len;
62a62a8ef9SStefan Hajnoczi };
63a62a8ef9SStefan Hajnoczi 
641efcf39eSVivek Goyal struct virtio_fs_forget_req {
65a62a8ef9SStefan Hajnoczi 	struct fuse_in_header ih;
66a62a8ef9SStefan Hajnoczi 	struct fuse_forget_in arg;
671efcf39eSVivek Goyal };
681efcf39eSVivek Goyal 
691efcf39eSVivek Goyal struct virtio_fs_forget {
70a62a8ef9SStefan Hajnoczi 	/* This request can be temporarily queued on virt queue */
71a62a8ef9SStefan Hajnoczi 	struct list_head list;
721efcf39eSVivek Goyal 	struct virtio_fs_forget_req req;
73a62a8ef9SStefan Hajnoczi };
74a62a8ef9SStefan Hajnoczi 
75bb737bbeSVivek Goyal struct virtio_fs_req_work {
76bb737bbeSVivek Goyal 	struct fuse_req *req;
77bb737bbeSVivek Goyal 	struct virtio_fs_vq *fsvq;
78bb737bbeSVivek Goyal 	struct work_struct done_work;
79bb737bbeSVivek Goyal };
80bb737bbeSVivek Goyal 
81a9bfd9ddSVivek Goyal static int virtio_fs_enqueue_req(struct virtio_fs_vq *fsvq,
82a9bfd9ddSVivek Goyal 				 struct fuse_req *req, bool in_flight);
83a9bfd9ddSVivek Goyal 
84a62a8ef9SStefan Hajnoczi static inline struct virtio_fs_vq *vq_to_fsvq(struct virtqueue *vq)
85a62a8ef9SStefan Hajnoczi {
86a62a8ef9SStefan Hajnoczi 	struct virtio_fs *fs = vq->vdev->priv;
87a62a8ef9SStefan Hajnoczi 
88a62a8ef9SStefan Hajnoczi 	return &fs->vqs[vq->index];
89a62a8ef9SStefan Hajnoczi }
90a62a8ef9SStefan Hajnoczi 
91a62a8ef9SStefan Hajnoczi static inline struct fuse_pqueue *vq_to_fpq(struct virtqueue *vq)
92a62a8ef9SStefan Hajnoczi {
93a62a8ef9SStefan Hajnoczi 	return &vq_to_fsvq(vq)->fud->pq;
94a62a8ef9SStefan Hajnoczi }
95a62a8ef9SStefan Hajnoczi 
96c17ea009SVivek Goyal /* Should be called with fsvq->lock held. */
97c17ea009SVivek Goyal static inline void inc_in_flight_req(struct virtio_fs_vq *fsvq)
98c17ea009SVivek Goyal {
99c17ea009SVivek Goyal 	fsvq->in_flight++;
100c17ea009SVivek Goyal }
101c17ea009SVivek Goyal 
102c17ea009SVivek Goyal /* Should be called with fsvq->lock held. */
103c17ea009SVivek Goyal static inline void dec_in_flight_req(struct virtio_fs_vq *fsvq)
104c17ea009SVivek Goyal {
105c17ea009SVivek Goyal 	WARN_ON(fsvq->in_flight <= 0);
106c17ea009SVivek Goyal 	fsvq->in_flight--;
107724c15a4SVivek Goyal 	if (!fsvq->in_flight)
108724c15a4SVivek Goyal 		complete(&fsvq->in_flight_zero);
109c17ea009SVivek Goyal }
110c17ea009SVivek Goyal 
111a62a8ef9SStefan Hajnoczi static void release_virtio_fs_obj(struct kref *ref)
112a62a8ef9SStefan Hajnoczi {
113a62a8ef9SStefan Hajnoczi 	struct virtio_fs *vfs = container_of(ref, struct virtio_fs, refcount);
114a62a8ef9SStefan Hajnoczi 
115a62a8ef9SStefan Hajnoczi 	kfree(vfs->vqs);
116a62a8ef9SStefan Hajnoczi 	kfree(vfs);
117a62a8ef9SStefan Hajnoczi }
118a62a8ef9SStefan Hajnoczi 
119a62a8ef9SStefan Hajnoczi /* Make sure virtiofs_mutex is held */
120a62a8ef9SStefan Hajnoczi static void virtio_fs_put(struct virtio_fs *fs)
121a62a8ef9SStefan Hajnoczi {
122a62a8ef9SStefan Hajnoczi 	kref_put(&fs->refcount, release_virtio_fs_obj);
123a62a8ef9SStefan Hajnoczi }
124a62a8ef9SStefan Hajnoczi 
125a62a8ef9SStefan Hajnoczi static void virtio_fs_fiq_release(struct fuse_iqueue *fiq)
126a62a8ef9SStefan Hajnoczi {
127a62a8ef9SStefan Hajnoczi 	struct virtio_fs *vfs = fiq->priv;
128a62a8ef9SStefan Hajnoczi 
129a62a8ef9SStefan Hajnoczi 	mutex_lock(&virtio_fs_mutex);
130a62a8ef9SStefan Hajnoczi 	virtio_fs_put(vfs);
131a62a8ef9SStefan Hajnoczi 	mutex_unlock(&virtio_fs_mutex);
132a62a8ef9SStefan Hajnoczi }
133a62a8ef9SStefan Hajnoczi 
134a62a8ef9SStefan Hajnoczi static void virtio_fs_drain_queue(struct virtio_fs_vq *fsvq)
135a62a8ef9SStefan Hajnoczi {
136a62a8ef9SStefan Hajnoczi 	WARN_ON(fsvq->in_flight < 0);
137a62a8ef9SStefan Hajnoczi 
138a62a8ef9SStefan Hajnoczi 	/* Wait for in flight requests to finish.*/
139a62a8ef9SStefan Hajnoczi 	spin_lock(&fsvq->lock);
140724c15a4SVivek Goyal 	if (fsvq->in_flight) {
141724c15a4SVivek Goyal 		/* We are holding virtio_fs_mutex. There should not be any
142724c15a4SVivek Goyal 		 * waiters waiting for completion.
143724c15a4SVivek Goyal 		 */
144724c15a4SVivek Goyal 		reinit_completion(&fsvq->in_flight_zero);
145a62a8ef9SStefan Hajnoczi 		spin_unlock(&fsvq->lock);
146724c15a4SVivek Goyal 		wait_for_completion(&fsvq->in_flight_zero);
147724c15a4SVivek Goyal 	} else {
148a62a8ef9SStefan Hajnoczi 		spin_unlock(&fsvq->lock);
149a62a8ef9SStefan Hajnoczi 	}
150a62a8ef9SStefan Hajnoczi 
151a62a8ef9SStefan Hajnoczi 	flush_work(&fsvq->done_work);
152a62a8ef9SStefan Hajnoczi 	flush_delayed_work(&fsvq->dispatch_work);
153a62a8ef9SStefan Hajnoczi }
154a62a8ef9SStefan Hajnoczi 
155724c15a4SVivek Goyal static void virtio_fs_drain_all_queues_locked(struct virtio_fs *fs)
156a62a8ef9SStefan Hajnoczi {
157a62a8ef9SStefan Hajnoczi 	struct virtio_fs_vq *fsvq;
158a62a8ef9SStefan Hajnoczi 	int i;
159a62a8ef9SStefan Hajnoczi 
160a62a8ef9SStefan Hajnoczi 	for (i = 0; i < fs->nvqs; i++) {
161a62a8ef9SStefan Hajnoczi 		fsvq = &fs->vqs[i];
162a62a8ef9SStefan Hajnoczi 		virtio_fs_drain_queue(fsvq);
163a62a8ef9SStefan Hajnoczi 	}
164a62a8ef9SStefan Hajnoczi }
165a62a8ef9SStefan Hajnoczi 
166724c15a4SVivek Goyal static void virtio_fs_drain_all_queues(struct virtio_fs *fs)
167724c15a4SVivek Goyal {
168724c15a4SVivek Goyal 	/* Provides mutual exclusion between ->remove and ->kill_sb
169724c15a4SVivek Goyal 	 * paths. We don't want both of these draining queue at the
170724c15a4SVivek Goyal 	 * same time. Current completion logic reinits completion
171724c15a4SVivek Goyal 	 * and that means there should not be any other thread
172724c15a4SVivek Goyal 	 * doing reinit or waiting for completion already.
173724c15a4SVivek Goyal 	 */
174724c15a4SVivek Goyal 	mutex_lock(&virtio_fs_mutex);
175724c15a4SVivek Goyal 	virtio_fs_drain_all_queues_locked(fs);
176724c15a4SVivek Goyal 	mutex_unlock(&virtio_fs_mutex);
177724c15a4SVivek Goyal }
178724c15a4SVivek Goyal 
179a62a8ef9SStefan Hajnoczi static void virtio_fs_start_all_queues(struct virtio_fs *fs)
180a62a8ef9SStefan Hajnoczi {
181a62a8ef9SStefan Hajnoczi 	struct virtio_fs_vq *fsvq;
182a62a8ef9SStefan Hajnoczi 	int i;
183a62a8ef9SStefan Hajnoczi 
184a62a8ef9SStefan Hajnoczi 	for (i = 0; i < fs->nvqs; i++) {
185a62a8ef9SStefan Hajnoczi 		fsvq = &fs->vqs[i];
186a62a8ef9SStefan Hajnoczi 		spin_lock(&fsvq->lock);
187a62a8ef9SStefan Hajnoczi 		fsvq->connected = true;
188a62a8ef9SStefan Hajnoczi 		spin_unlock(&fsvq->lock);
189a62a8ef9SStefan Hajnoczi 	}
190a62a8ef9SStefan Hajnoczi }
191a62a8ef9SStefan Hajnoczi 
192a62a8ef9SStefan Hajnoczi /* Add a new instance to the list or return -EEXIST if tag name exists*/
193a62a8ef9SStefan Hajnoczi static int virtio_fs_add_instance(struct virtio_fs *fs)
194a62a8ef9SStefan Hajnoczi {
195a62a8ef9SStefan Hajnoczi 	struct virtio_fs *fs2;
196a62a8ef9SStefan Hajnoczi 	bool duplicate = false;
197a62a8ef9SStefan Hajnoczi 
198a62a8ef9SStefan Hajnoczi 	mutex_lock(&virtio_fs_mutex);
199a62a8ef9SStefan Hajnoczi 
200a62a8ef9SStefan Hajnoczi 	list_for_each_entry(fs2, &virtio_fs_instances, list) {
201a62a8ef9SStefan Hajnoczi 		if (strcmp(fs->tag, fs2->tag) == 0)
202a62a8ef9SStefan Hajnoczi 			duplicate = true;
203a62a8ef9SStefan Hajnoczi 	}
204a62a8ef9SStefan Hajnoczi 
205a62a8ef9SStefan Hajnoczi 	if (!duplicate)
206a62a8ef9SStefan Hajnoczi 		list_add_tail(&fs->list, &virtio_fs_instances);
207a62a8ef9SStefan Hajnoczi 
208a62a8ef9SStefan Hajnoczi 	mutex_unlock(&virtio_fs_mutex);
209a62a8ef9SStefan Hajnoczi 
210a62a8ef9SStefan Hajnoczi 	if (duplicate)
211a62a8ef9SStefan Hajnoczi 		return -EEXIST;
212a62a8ef9SStefan Hajnoczi 	return 0;
213a62a8ef9SStefan Hajnoczi }
214a62a8ef9SStefan Hajnoczi 
215a62a8ef9SStefan Hajnoczi /* Return the virtio_fs with a given tag, or NULL */
216a62a8ef9SStefan Hajnoczi static struct virtio_fs *virtio_fs_find_instance(const char *tag)
217a62a8ef9SStefan Hajnoczi {
218a62a8ef9SStefan Hajnoczi 	struct virtio_fs *fs;
219a62a8ef9SStefan Hajnoczi 
220a62a8ef9SStefan Hajnoczi 	mutex_lock(&virtio_fs_mutex);
221a62a8ef9SStefan Hajnoczi 
222a62a8ef9SStefan Hajnoczi 	list_for_each_entry(fs, &virtio_fs_instances, list) {
223a62a8ef9SStefan Hajnoczi 		if (strcmp(fs->tag, tag) == 0) {
224a62a8ef9SStefan Hajnoczi 			kref_get(&fs->refcount);
225a62a8ef9SStefan Hajnoczi 			goto found;
226a62a8ef9SStefan Hajnoczi 		}
227a62a8ef9SStefan Hajnoczi 	}
228a62a8ef9SStefan Hajnoczi 
229a62a8ef9SStefan Hajnoczi 	fs = NULL; /* not found */
230a62a8ef9SStefan Hajnoczi 
231a62a8ef9SStefan Hajnoczi found:
232a62a8ef9SStefan Hajnoczi 	mutex_unlock(&virtio_fs_mutex);
233a62a8ef9SStefan Hajnoczi 
234a62a8ef9SStefan Hajnoczi 	return fs;
235a62a8ef9SStefan Hajnoczi }
236a62a8ef9SStefan Hajnoczi 
237a62a8ef9SStefan Hajnoczi static void virtio_fs_free_devs(struct virtio_fs *fs)
238a62a8ef9SStefan Hajnoczi {
239a62a8ef9SStefan Hajnoczi 	unsigned int i;
240a62a8ef9SStefan Hajnoczi 
241a62a8ef9SStefan Hajnoczi 	for (i = 0; i < fs->nvqs; i++) {
242a62a8ef9SStefan Hajnoczi 		struct virtio_fs_vq *fsvq = &fs->vqs[i];
243a62a8ef9SStefan Hajnoczi 
244a62a8ef9SStefan Hajnoczi 		if (!fsvq->fud)
245a62a8ef9SStefan Hajnoczi 			continue;
246a62a8ef9SStefan Hajnoczi 
247a62a8ef9SStefan Hajnoczi 		fuse_dev_free(fsvq->fud);
248a62a8ef9SStefan Hajnoczi 		fsvq->fud = NULL;
249a62a8ef9SStefan Hajnoczi 	}
250a62a8ef9SStefan Hajnoczi }
251a62a8ef9SStefan Hajnoczi 
252a62a8ef9SStefan Hajnoczi /* Read filesystem name from virtio config into fs->tag (must kfree()). */
253a62a8ef9SStefan Hajnoczi static int virtio_fs_read_tag(struct virtio_device *vdev, struct virtio_fs *fs)
254a62a8ef9SStefan Hajnoczi {
255a62a8ef9SStefan Hajnoczi 	char tag_buf[sizeof_field(struct virtio_fs_config, tag)];
256a62a8ef9SStefan Hajnoczi 	char *end;
257a62a8ef9SStefan Hajnoczi 	size_t len;
258a62a8ef9SStefan Hajnoczi 
259a62a8ef9SStefan Hajnoczi 	virtio_cread_bytes(vdev, offsetof(struct virtio_fs_config, tag),
260a62a8ef9SStefan Hajnoczi 			   &tag_buf, sizeof(tag_buf));
261a62a8ef9SStefan Hajnoczi 	end = memchr(tag_buf, '\0', sizeof(tag_buf));
262a62a8ef9SStefan Hajnoczi 	if (end == tag_buf)
263a62a8ef9SStefan Hajnoczi 		return -EINVAL; /* empty tag */
264a62a8ef9SStefan Hajnoczi 	if (!end)
265a62a8ef9SStefan Hajnoczi 		end = &tag_buf[sizeof(tag_buf)];
266a62a8ef9SStefan Hajnoczi 
267a62a8ef9SStefan Hajnoczi 	len = end - tag_buf;
268a62a8ef9SStefan Hajnoczi 	fs->tag = devm_kmalloc(&vdev->dev, len + 1, GFP_KERNEL);
269a62a8ef9SStefan Hajnoczi 	if (!fs->tag)
270a62a8ef9SStefan Hajnoczi 		return -ENOMEM;
271a62a8ef9SStefan Hajnoczi 	memcpy(fs->tag, tag_buf, len);
272a62a8ef9SStefan Hajnoczi 	fs->tag[len] = '\0';
273a62a8ef9SStefan Hajnoczi 	return 0;
274a62a8ef9SStefan Hajnoczi }
275a62a8ef9SStefan Hajnoczi 
276a62a8ef9SStefan Hajnoczi /* Work function for hiprio completion */
277a62a8ef9SStefan Hajnoczi static void virtio_fs_hiprio_done_work(struct work_struct *work)
278a62a8ef9SStefan Hajnoczi {
279a62a8ef9SStefan Hajnoczi 	struct virtio_fs_vq *fsvq = container_of(work, struct virtio_fs_vq,
280a62a8ef9SStefan Hajnoczi 						 done_work);
281a62a8ef9SStefan Hajnoczi 	struct virtqueue *vq = fsvq->vq;
282a62a8ef9SStefan Hajnoczi 
283a62a8ef9SStefan Hajnoczi 	/* Free completed FUSE_FORGET requests */
284a62a8ef9SStefan Hajnoczi 	spin_lock(&fsvq->lock);
285a62a8ef9SStefan Hajnoczi 	do {
286a62a8ef9SStefan Hajnoczi 		unsigned int len;
287a62a8ef9SStefan Hajnoczi 		void *req;
288a62a8ef9SStefan Hajnoczi 
289a62a8ef9SStefan Hajnoczi 		virtqueue_disable_cb(vq);
290a62a8ef9SStefan Hajnoczi 
291a62a8ef9SStefan Hajnoczi 		while ((req = virtqueue_get_buf(vq, &len)) != NULL) {
292a62a8ef9SStefan Hajnoczi 			kfree(req);
293c17ea009SVivek Goyal 			dec_in_flight_req(fsvq);
294a62a8ef9SStefan Hajnoczi 		}
295a62a8ef9SStefan Hajnoczi 	} while (!virtqueue_enable_cb(vq) && likely(!virtqueue_is_broken(vq)));
296a62a8ef9SStefan Hajnoczi 	spin_unlock(&fsvq->lock);
297a62a8ef9SStefan Hajnoczi }
298a62a8ef9SStefan Hajnoczi 
29951fecdd2SVivek Goyal static void virtio_fs_request_dispatch_work(struct work_struct *work)
300a62a8ef9SStefan Hajnoczi {
30151fecdd2SVivek Goyal 	struct fuse_req *req;
30251fecdd2SVivek Goyal 	struct virtio_fs_vq *fsvq = container_of(work, struct virtio_fs_vq,
30351fecdd2SVivek Goyal 						 dispatch_work.work);
30451fecdd2SVivek Goyal 	struct fuse_conn *fc = fsvq->fud->fc;
305a9bfd9ddSVivek Goyal 	int ret;
30651fecdd2SVivek Goyal 
30751fecdd2SVivek Goyal 	pr_debug("virtio-fs: worker %s called.\n", __func__);
30851fecdd2SVivek Goyal 	while (1) {
30951fecdd2SVivek Goyal 		spin_lock(&fsvq->lock);
31051fecdd2SVivek Goyal 		req = list_first_entry_or_null(&fsvq->end_reqs, struct fuse_req,
31151fecdd2SVivek Goyal 					       list);
31251fecdd2SVivek Goyal 		if (!req) {
31351fecdd2SVivek Goyal 			spin_unlock(&fsvq->lock);
314a9bfd9ddSVivek Goyal 			break;
31551fecdd2SVivek Goyal 		}
31651fecdd2SVivek Goyal 
31751fecdd2SVivek Goyal 		list_del_init(&req->list);
31851fecdd2SVivek Goyal 		spin_unlock(&fsvq->lock);
31951fecdd2SVivek Goyal 		fuse_request_end(fc, req);
32051fecdd2SVivek Goyal 	}
321a9bfd9ddSVivek Goyal 
322a9bfd9ddSVivek Goyal 	/* Dispatch pending requests */
323a9bfd9ddSVivek Goyal 	while (1) {
324a9bfd9ddSVivek Goyal 		spin_lock(&fsvq->lock);
325a9bfd9ddSVivek Goyal 		req = list_first_entry_or_null(&fsvq->queued_reqs,
326a9bfd9ddSVivek Goyal 					       struct fuse_req, list);
327a9bfd9ddSVivek Goyal 		if (!req) {
328a9bfd9ddSVivek Goyal 			spin_unlock(&fsvq->lock);
329a9bfd9ddSVivek Goyal 			return;
330a9bfd9ddSVivek Goyal 		}
331a9bfd9ddSVivek Goyal 		list_del_init(&req->list);
332a9bfd9ddSVivek Goyal 		spin_unlock(&fsvq->lock);
333a9bfd9ddSVivek Goyal 
334a9bfd9ddSVivek Goyal 		ret = virtio_fs_enqueue_req(fsvq, req, true);
335a9bfd9ddSVivek Goyal 		if (ret < 0) {
336a9bfd9ddSVivek Goyal 			if (ret == -ENOMEM || ret == -ENOSPC) {
337a9bfd9ddSVivek Goyal 				spin_lock(&fsvq->lock);
338a9bfd9ddSVivek Goyal 				list_add_tail(&req->list, &fsvq->queued_reqs);
339a9bfd9ddSVivek Goyal 				schedule_delayed_work(&fsvq->dispatch_work,
340a9bfd9ddSVivek Goyal 						      msecs_to_jiffies(1));
341a9bfd9ddSVivek Goyal 				spin_unlock(&fsvq->lock);
342a9bfd9ddSVivek Goyal 				return;
343a9bfd9ddSVivek Goyal 			}
344a9bfd9ddSVivek Goyal 			req->out.h.error = ret;
345a9bfd9ddSVivek Goyal 			spin_lock(&fsvq->lock);
346a9bfd9ddSVivek Goyal 			dec_in_flight_req(fsvq);
347a9bfd9ddSVivek Goyal 			spin_unlock(&fsvq->lock);
348a9bfd9ddSVivek Goyal 			pr_err("virtio-fs: virtio_fs_enqueue_req() failed %d\n",
349a9bfd9ddSVivek Goyal 			       ret);
350a9bfd9ddSVivek Goyal 			fuse_request_end(fc, req);
351a9bfd9ddSVivek Goyal 		}
352a9bfd9ddSVivek Goyal 	}
353a62a8ef9SStefan Hajnoczi }
354a62a8ef9SStefan Hajnoczi 
35558ada94fSVivek Goyal /*
35658ada94fSVivek Goyal  * Returns 1 if queue is full and sender should wait a bit before sending
35758ada94fSVivek Goyal  * next request, 0 otherwise.
35858ada94fSVivek Goyal  */
35958ada94fSVivek Goyal static int send_forget_request(struct virtio_fs_vq *fsvq,
36058ada94fSVivek Goyal 			       struct virtio_fs_forget *forget,
36158ada94fSVivek Goyal 			       bool in_flight)
36258ada94fSVivek Goyal {
36358ada94fSVivek Goyal 	struct scatterlist sg;
36458ada94fSVivek Goyal 	struct virtqueue *vq;
36558ada94fSVivek Goyal 	int ret = 0;
36658ada94fSVivek Goyal 	bool notify;
3671efcf39eSVivek Goyal 	struct virtio_fs_forget_req *req = &forget->req;
36858ada94fSVivek Goyal 
36958ada94fSVivek Goyal 	spin_lock(&fsvq->lock);
37058ada94fSVivek Goyal 	if (!fsvq->connected) {
37158ada94fSVivek Goyal 		if (in_flight)
37258ada94fSVivek Goyal 			dec_in_flight_req(fsvq);
37358ada94fSVivek Goyal 		kfree(forget);
37458ada94fSVivek Goyal 		goto out;
37558ada94fSVivek Goyal 	}
37658ada94fSVivek Goyal 
3771efcf39eSVivek Goyal 	sg_init_one(&sg, req, sizeof(*req));
37858ada94fSVivek Goyal 	vq = fsvq->vq;
37958ada94fSVivek Goyal 	dev_dbg(&vq->vdev->dev, "%s\n", __func__);
38058ada94fSVivek Goyal 
38158ada94fSVivek Goyal 	ret = virtqueue_add_outbuf(vq, &sg, 1, forget, GFP_ATOMIC);
38258ada94fSVivek Goyal 	if (ret < 0) {
38358ada94fSVivek Goyal 		if (ret == -ENOMEM || ret == -ENOSPC) {
38458ada94fSVivek Goyal 			pr_debug("virtio-fs: Could not queue FORGET: err=%d. Will try later\n",
38558ada94fSVivek Goyal 				 ret);
38658ada94fSVivek Goyal 			list_add_tail(&forget->list, &fsvq->queued_reqs);
38758ada94fSVivek Goyal 			schedule_delayed_work(&fsvq->dispatch_work,
38858ada94fSVivek Goyal 					      msecs_to_jiffies(1));
38958ada94fSVivek Goyal 			if (!in_flight)
39058ada94fSVivek Goyal 				inc_in_flight_req(fsvq);
39158ada94fSVivek Goyal 			/* Queue is full */
39258ada94fSVivek Goyal 			ret = 1;
39358ada94fSVivek Goyal 		} else {
39458ada94fSVivek Goyal 			pr_debug("virtio-fs: Could not queue FORGET: err=%d. Dropping it.\n",
39558ada94fSVivek Goyal 				 ret);
39658ada94fSVivek Goyal 			kfree(forget);
39758ada94fSVivek Goyal 			if (in_flight)
39858ada94fSVivek Goyal 				dec_in_flight_req(fsvq);
39958ada94fSVivek Goyal 		}
40058ada94fSVivek Goyal 		goto out;
40158ada94fSVivek Goyal 	}
40258ada94fSVivek Goyal 
40358ada94fSVivek Goyal 	if (!in_flight)
40458ada94fSVivek Goyal 		inc_in_flight_req(fsvq);
40558ada94fSVivek Goyal 	notify = virtqueue_kick_prepare(vq);
40658ada94fSVivek Goyal 	spin_unlock(&fsvq->lock);
40758ada94fSVivek Goyal 
40858ada94fSVivek Goyal 	if (notify)
40958ada94fSVivek Goyal 		virtqueue_notify(vq);
41058ada94fSVivek Goyal 	return ret;
41158ada94fSVivek Goyal out:
41258ada94fSVivek Goyal 	spin_unlock(&fsvq->lock);
41358ada94fSVivek Goyal 	return ret;
41458ada94fSVivek Goyal }
41558ada94fSVivek Goyal 
416a62a8ef9SStefan Hajnoczi static void virtio_fs_hiprio_dispatch_work(struct work_struct *work)
417a62a8ef9SStefan Hajnoczi {
418a62a8ef9SStefan Hajnoczi 	struct virtio_fs_forget *forget;
419a62a8ef9SStefan Hajnoczi 	struct virtio_fs_vq *fsvq = container_of(work, struct virtio_fs_vq,
420a62a8ef9SStefan Hajnoczi 						 dispatch_work.work);
421a62a8ef9SStefan Hajnoczi 	pr_debug("virtio-fs: worker %s called.\n", __func__);
422a62a8ef9SStefan Hajnoczi 	while (1) {
423a62a8ef9SStefan Hajnoczi 		spin_lock(&fsvq->lock);
424a62a8ef9SStefan Hajnoczi 		forget = list_first_entry_or_null(&fsvq->queued_reqs,
425a62a8ef9SStefan Hajnoczi 					struct virtio_fs_forget, list);
426a62a8ef9SStefan Hajnoczi 		if (!forget) {
427a62a8ef9SStefan Hajnoczi 			spin_unlock(&fsvq->lock);
428a62a8ef9SStefan Hajnoczi 			return;
429a62a8ef9SStefan Hajnoczi 		}
430a62a8ef9SStefan Hajnoczi 
431a62a8ef9SStefan Hajnoczi 		list_del(&forget->list);
432a62a8ef9SStefan Hajnoczi 		spin_unlock(&fsvq->lock);
43358ada94fSVivek Goyal 		if (send_forget_request(fsvq, forget, true))
434a62a8ef9SStefan Hajnoczi 			return;
435a62a8ef9SStefan Hajnoczi 	}
436a62a8ef9SStefan Hajnoczi }
437a62a8ef9SStefan Hajnoczi 
438a62a8ef9SStefan Hajnoczi /* Allocate and copy args into req->argbuf */
439a62a8ef9SStefan Hajnoczi static int copy_args_to_argbuf(struct fuse_req *req)
440a62a8ef9SStefan Hajnoczi {
441a62a8ef9SStefan Hajnoczi 	struct fuse_args *args = req->args;
442a62a8ef9SStefan Hajnoczi 	unsigned int offset = 0;
443a62a8ef9SStefan Hajnoczi 	unsigned int num_in;
444a62a8ef9SStefan Hajnoczi 	unsigned int num_out;
445a62a8ef9SStefan Hajnoczi 	unsigned int len;
446a62a8ef9SStefan Hajnoczi 	unsigned int i;
447a62a8ef9SStefan Hajnoczi 
448a62a8ef9SStefan Hajnoczi 	num_in = args->in_numargs - args->in_pages;
449a62a8ef9SStefan Hajnoczi 	num_out = args->out_numargs - args->out_pages;
450a62a8ef9SStefan Hajnoczi 	len = fuse_len_args(num_in, (struct fuse_arg *) args->in_args) +
451a62a8ef9SStefan Hajnoczi 	      fuse_len_args(num_out, args->out_args);
452a62a8ef9SStefan Hajnoczi 
453a62a8ef9SStefan Hajnoczi 	req->argbuf = kmalloc(len, GFP_ATOMIC);
454a62a8ef9SStefan Hajnoczi 	if (!req->argbuf)
455a62a8ef9SStefan Hajnoczi 		return -ENOMEM;
456a62a8ef9SStefan Hajnoczi 
457a62a8ef9SStefan Hajnoczi 	for (i = 0; i < num_in; i++) {
458a62a8ef9SStefan Hajnoczi 		memcpy(req->argbuf + offset,
459a62a8ef9SStefan Hajnoczi 		       args->in_args[i].value,
460a62a8ef9SStefan Hajnoczi 		       args->in_args[i].size);
461a62a8ef9SStefan Hajnoczi 		offset += args->in_args[i].size;
462a62a8ef9SStefan Hajnoczi 	}
463a62a8ef9SStefan Hajnoczi 
464a62a8ef9SStefan Hajnoczi 	return 0;
465a62a8ef9SStefan Hajnoczi }
466a62a8ef9SStefan Hajnoczi 
467a62a8ef9SStefan Hajnoczi /* Copy args out of and free req->argbuf */
468a62a8ef9SStefan Hajnoczi static void copy_args_from_argbuf(struct fuse_args *args, struct fuse_req *req)
469a62a8ef9SStefan Hajnoczi {
470a62a8ef9SStefan Hajnoczi 	unsigned int remaining;
471a62a8ef9SStefan Hajnoczi 	unsigned int offset;
472a62a8ef9SStefan Hajnoczi 	unsigned int num_in;
473a62a8ef9SStefan Hajnoczi 	unsigned int num_out;
474a62a8ef9SStefan Hajnoczi 	unsigned int i;
475a62a8ef9SStefan Hajnoczi 
476a62a8ef9SStefan Hajnoczi 	remaining = req->out.h.len - sizeof(req->out.h);
477a62a8ef9SStefan Hajnoczi 	num_in = args->in_numargs - args->in_pages;
478a62a8ef9SStefan Hajnoczi 	num_out = args->out_numargs - args->out_pages;
479a62a8ef9SStefan Hajnoczi 	offset = fuse_len_args(num_in, (struct fuse_arg *)args->in_args);
480a62a8ef9SStefan Hajnoczi 
481a62a8ef9SStefan Hajnoczi 	for (i = 0; i < num_out; i++) {
482a62a8ef9SStefan Hajnoczi 		unsigned int argsize = args->out_args[i].size;
483a62a8ef9SStefan Hajnoczi 
484a62a8ef9SStefan Hajnoczi 		if (args->out_argvar &&
485a62a8ef9SStefan Hajnoczi 		    i == args->out_numargs - 1 &&
486a62a8ef9SStefan Hajnoczi 		    argsize > remaining) {
487a62a8ef9SStefan Hajnoczi 			argsize = remaining;
488a62a8ef9SStefan Hajnoczi 		}
489a62a8ef9SStefan Hajnoczi 
490a62a8ef9SStefan Hajnoczi 		memcpy(args->out_args[i].value, req->argbuf + offset, argsize);
491a62a8ef9SStefan Hajnoczi 		offset += argsize;
492a62a8ef9SStefan Hajnoczi 
493a62a8ef9SStefan Hajnoczi 		if (i != args->out_numargs - 1)
494a62a8ef9SStefan Hajnoczi 			remaining -= argsize;
495a62a8ef9SStefan Hajnoczi 	}
496a62a8ef9SStefan Hajnoczi 
497a62a8ef9SStefan Hajnoczi 	/* Store the actual size of the variable-length arg */
498a62a8ef9SStefan Hajnoczi 	if (args->out_argvar)
499a62a8ef9SStefan Hajnoczi 		args->out_args[args->out_numargs - 1].size = remaining;
500a62a8ef9SStefan Hajnoczi 
501a62a8ef9SStefan Hajnoczi 	kfree(req->argbuf);
502a62a8ef9SStefan Hajnoczi 	req->argbuf = NULL;
503a62a8ef9SStefan Hajnoczi }
504a62a8ef9SStefan Hajnoczi 
505a62a8ef9SStefan Hajnoczi /* Work function for request completion */
506bb737bbeSVivek Goyal static void virtio_fs_request_complete(struct fuse_req *req,
507bb737bbeSVivek Goyal 				       struct virtio_fs_vq *fsvq)
508a62a8ef9SStefan Hajnoczi {
509a62a8ef9SStefan Hajnoczi 	struct fuse_pqueue *fpq = &fsvq->fud->pq;
510a62a8ef9SStefan Hajnoczi 	struct fuse_conn *fc = fsvq->fud->fc;
511a62a8ef9SStefan Hajnoczi 	struct fuse_args *args;
512bb737bbeSVivek Goyal 	struct fuse_args_pages *ap;
513a62a8ef9SStefan Hajnoczi 	unsigned int len, i, thislen;
514a62a8ef9SStefan Hajnoczi 	struct page *page;
515a62a8ef9SStefan Hajnoczi 
516a62a8ef9SStefan Hajnoczi 	/*
517a62a8ef9SStefan Hajnoczi 	 * TODO verify that server properly follows FUSE protocol
518a62a8ef9SStefan Hajnoczi 	 * (oh.uniq, oh.len)
519a62a8ef9SStefan Hajnoczi 	 */
520a62a8ef9SStefan Hajnoczi 	args = req->args;
521a62a8ef9SStefan Hajnoczi 	copy_args_from_argbuf(args, req);
522a62a8ef9SStefan Hajnoczi 
523a62a8ef9SStefan Hajnoczi 	if (args->out_pages && args->page_zeroing) {
524a62a8ef9SStefan Hajnoczi 		len = args->out_args[args->out_numargs - 1].size;
525a62a8ef9SStefan Hajnoczi 		ap = container_of(args, typeof(*ap), args);
526a62a8ef9SStefan Hajnoczi 		for (i = 0; i < ap->num_pages; i++) {
527a62a8ef9SStefan Hajnoczi 			thislen = ap->descs[i].length;
528a62a8ef9SStefan Hajnoczi 			if (len < thislen) {
529a62a8ef9SStefan Hajnoczi 				WARN_ON(ap->descs[i].offset);
530a62a8ef9SStefan Hajnoczi 				page = ap->pages[i];
531a62a8ef9SStefan Hajnoczi 				zero_user_segment(page, len, thislen);
532a62a8ef9SStefan Hajnoczi 				len = 0;
533a62a8ef9SStefan Hajnoczi 			} else {
534a62a8ef9SStefan Hajnoczi 				len -= thislen;
535a62a8ef9SStefan Hajnoczi 			}
536a62a8ef9SStefan Hajnoczi 		}
537a62a8ef9SStefan Hajnoczi 	}
538a62a8ef9SStefan Hajnoczi 
539a62a8ef9SStefan Hajnoczi 	spin_lock(&fpq->lock);
540a62a8ef9SStefan Hajnoczi 	clear_bit(FR_SENT, &req->flags);
541a62a8ef9SStefan Hajnoczi 	spin_unlock(&fpq->lock);
542a62a8ef9SStefan Hajnoczi 
543a62a8ef9SStefan Hajnoczi 	fuse_request_end(fc, req);
544a62a8ef9SStefan Hajnoczi 	spin_lock(&fsvq->lock);
545c17ea009SVivek Goyal 	dec_in_flight_req(fsvq);
546a62a8ef9SStefan Hajnoczi 	spin_unlock(&fsvq->lock);
547a62a8ef9SStefan Hajnoczi }
548bb737bbeSVivek Goyal 
549bb737bbeSVivek Goyal static void virtio_fs_complete_req_work(struct work_struct *work)
550bb737bbeSVivek Goyal {
551bb737bbeSVivek Goyal 	struct virtio_fs_req_work *w =
552bb737bbeSVivek Goyal 		container_of(work, typeof(*w), done_work);
553bb737bbeSVivek Goyal 
554bb737bbeSVivek Goyal 	virtio_fs_request_complete(w->req, w->fsvq);
555bb737bbeSVivek Goyal 	kfree(w);
556bb737bbeSVivek Goyal }
557bb737bbeSVivek Goyal 
558bb737bbeSVivek Goyal static void virtio_fs_requests_done_work(struct work_struct *work)
559bb737bbeSVivek Goyal {
560bb737bbeSVivek Goyal 	struct virtio_fs_vq *fsvq = container_of(work, struct virtio_fs_vq,
561bb737bbeSVivek Goyal 						 done_work);
562bb737bbeSVivek Goyal 	struct fuse_pqueue *fpq = &fsvq->fud->pq;
563bb737bbeSVivek Goyal 	struct virtqueue *vq = fsvq->vq;
564bb737bbeSVivek Goyal 	struct fuse_req *req;
565bb737bbeSVivek Goyal 	struct fuse_req *next;
566bb737bbeSVivek Goyal 	unsigned int len;
567bb737bbeSVivek Goyal 	LIST_HEAD(reqs);
568bb737bbeSVivek Goyal 
569bb737bbeSVivek Goyal 	/* Collect completed requests off the virtqueue */
570bb737bbeSVivek Goyal 	spin_lock(&fsvq->lock);
571bb737bbeSVivek Goyal 	do {
572bb737bbeSVivek Goyal 		virtqueue_disable_cb(vq);
573bb737bbeSVivek Goyal 
574bb737bbeSVivek Goyal 		while ((req = virtqueue_get_buf(vq, &len)) != NULL) {
575bb737bbeSVivek Goyal 			spin_lock(&fpq->lock);
576bb737bbeSVivek Goyal 			list_move_tail(&req->list, &reqs);
577bb737bbeSVivek Goyal 			spin_unlock(&fpq->lock);
578bb737bbeSVivek Goyal 		}
579bb737bbeSVivek Goyal 	} while (!virtqueue_enable_cb(vq) && likely(!virtqueue_is_broken(vq)));
580bb737bbeSVivek Goyal 	spin_unlock(&fsvq->lock);
581bb737bbeSVivek Goyal 
582bb737bbeSVivek Goyal 	/* End requests */
583bb737bbeSVivek Goyal 	list_for_each_entry_safe(req, next, &reqs, list) {
584bb737bbeSVivek Goyal 		list_del_init(&req->list);
585bb737bbeSVivek Goyal 
586bb737bbeSVivek Goyal 		/* blocking async request completes in a worker context */
587bb737bbeSVivek Goyal 		if (req->args->may_block) {
588bb737bbeSVivek Goyal 			struct virtio_fs_req_work *w;
589bb737bbeSVivek Goyal 
590bb737bbeSVivek Goyal 			w = kzalloc(sizeof(*w), GFP_NOFS | __GFP_NOFAIL);
591bb737bbeSVivek Goyal 			INIT_WORK(&w->done_work, virtio_fs_complete_req_work);
592bb737bbeSVivek Goyal 			w->fsvq = fsvq;
593bb737bbeSVivek Goyal 			w->req = req;
594bb737bbeSVivek Goyal 			schedule_work(&w->done_work);
595bb737bbeSVivek Goyal 		} else {
596bb737bbeSVivek Goyal 			virtio_fs_request_complete(req, fsvq);
597bb737bbeSVivek Goyal 		}
598bb737bbeSVivek Goyal 	}
599a62a8ef9SStefan Hajnoczi }
600a62a8ef9SStefan Hajnoczi 
601a62a8ef9SStefan Hajnoczi /* Virtqueue interrupt handler */
602a62a8ef9SStefan Hajnoczi static void virtio_fs_vq_done(struct virtqueue *vq)
603a62a8ef9SStefan Hajnoczi {
604a62a8ef9SStefan Hajnoczi 	struct virtio_fs_vq *fsvq = vq_to_fsvq(vq);
605a62a8ef9SStefan Hajnoczi 
606a62a8ef9SStefan Hajnoczi 	dev_dbg(&vq->vdev->dev, "%s %s\n", __func__, fsvq->name);
607a62a8ef9SStefan Hajnoczi 
608a62a8ef9SStefan Hajnoczi 	schedule_work(&fsvq->done_work);
609a62a8ef9SStefan Hajnoczi }
610a62a8ef9SStefan Hajnoczi 
611b43b7e81SVivek Goyal static void virtio_fs_init_vq(struct virtio_fs_vq *fsvq, char *name,
612b43b7e81SVivek Goyal 			      int vq_type)
613b43b7e81SVivek Goyal {
614b43b7e81SVivek Goyal 	strncpy(fsvq->name, name, VQ_NAME_LEN);
615b43b7e81SVivek Goyal 	spin_lock_init(&fsvq->lock);
616b43b7e81SVivek Goyal 	INIT_LIST_HEAD(&fsvq->queued_reqs);
617b43b7e81SVivek Goyal 	INIT_LIST_HEAD(&fsvq->end_reqs);
618b43b7e81SVivek Goyal 	init_completion(&fsvq->in_flight_zero);
619b43b7e81SVivek Goyal 
620b43b7e81SVivek Goyal 	if (vq_type == VQ_REQUEST) {
621b43b7e81SVivek Goyal 		INIT_WORK(&fsvq->done_work, virtio_fs_requests_done_work);
622b43b7e81SVivek Goyal 		INIT_DELAYED_WORK(&fsvq->dispatch_work,
623b43b7e81SVivek Goyal 				  virtio_fs_request_dispatch_work);
624b43b7e81SVivek Goyal 	} else {
625b43b7e81SVivek Goyal 		INIT_WORK(&fsvq->done_work, virtio_fs_hiprio_done_work);
626b43b7e81SVivek Goyal 		INIT_DELAYED_WORK(&fsvq->dispatch_work,
627b43b7e81SVivek Goyal 				  virtio_fs_hiprio_dispatch_work);
628b43b7e81SVivek Goyal 	}
629b43b7e81SVivek Goyal }
630b43b7e81SVivek Goyal 
631a62a8ef9SStefan Hajnoczi /* Initialize virtqueues */
632a62a8ef9SStefan Hajnoczi static int virtio_fs_setup_vqs(struct virtio_device *vdev,
633a62a8ef9SStefan Hajnoczi 			       struct virtio_fs *fs)
634a62a8ef9SStefan Hajnoczi {
635a62a8ef9SStefan Hajnoczi 	struct virtqueue **vqs;
636a62a8ef9SStefan Hajnoczi 	vq_callback_t **callbacks;
637a62a8ef9SStefan Hajnoczi 	const char **names;
638a62a8ef9SStefan Hajnoczi 	unsigned int i;
639a62a8ef9SStefan Hajnoczi 	int ret = 0;
640a62a8ef9SStefan Hajnoczi 
6412c0349ecSMichael S. Tsirkin 	virtio_cread_le(vdev, struct virtio_fs_config, num_request_queues,
642a62a8ef9SStefan Hajnoczi 			&fs->num_request_queues);
643a62a8ef9SStefan Hajnoczi 	if (fs->num_request_queues == 0)
644a62a8ef9SStefan Hajnoczi 		return -EINVAL;
645a62a8ef9SStefan Hajnoczi 
646b43b7e81SVivek Goyal 	fs->nvqs = VQ_REQUEST + fs->num_request_queues;
647a62a8ef9SStefan Hajnoczi 	fs->vqs = kcalloc(fs->nvqs, sizeof(fs->vqs[VQ_HIPRIO]), GFP_KERNEL);
648a62a8ef9SStefan Hajnoczi 	if (!fs->vqs)
649a62a8ef9SStefan Hajnoczi 		return -ENOMEM;
650a62a8ef9SStefan Hajnoczi 
651a62a8ef9SStefan Hajnoczi 	vqs = kmalloc_array(fs->nvqs, sizeof(vqs[VQ_HIPRIO]), GFP_KERNEL);
652a62a8ef9SStefan Hajnoczi 	callbacks = kmalloc_array(fs->nvqs, sizeof(callbacks[VQ_HIPRIO]),
653a62a8ef9SStefan Hajnoczi 					GFP_KERNEL);
654a62a8ef9SStefan Hajnoczi 	names = kmalloc_array(fs->nvqs, sizeof(names[VQ_HIPRIO]), GFP_KERNEL);
655a62a8ef9SStefan Hajnoczi 	if (!vqs || !callbacks || !names) {
656a62a8ef9SStefan Hajnoczi 		ret = -ENOMEM;
657a62a8ef9SStefan Hajnoczi 		goto out;
658a62a8ef9SStefan Hajnoczi 	}
659a62a8ef9SStefan Hajnoczi 
660b43b7e81SVivek Goyal 	/* Initialize the hiprio/forget request virtqueue */
661a62a8ef9SStefan Hajnoczi 	callbacks[VQ_HIPRIO] = virtio_fs_vq_done;
662b43b7e81SVivek Goyal 	virtio_fs_init_vq(&fs->vqs[VQ_HIPRIO], "hiprio", VQ_HIPRIO);
663a62a8ef9SStefan Hajnoczi 	names[VQ_HIPRIO] = fs->vqs[VQ_HIPRIO].name;
664a62a8ef9SStefan Hajnoczi 
665a62a8ef9SStefan Hajnoczi 	/* Initialize the requests virtqueues */
666a62a8ef9SStefan Hajnoczi 	for (i = VQ_REQUEST; i < fs->nvqs; i++) {
667b43b7e81SVivek Goyal 		char vq_name[VQ_NAME_LEN];
668b43b7e81SVivek Goyal 
669b43b7e81SVivek Goyal 		snprintf(vq_name, VQ_NAME_LEN, "requests.%u", i - VQ_REQUEST);
670b43b7e81SVivek Goyal 		virtio_fs_init_vq(&fs->vqs[i], vq_name, VQ_REQUEST);
671a62a8ef9SStefan Hajnoczi 		callbacks[i] = virtio_fs_vq_done;
672a62a8ef9SStefan Hajnoczi 		names[i] = fs->vqs[i].name;
673a62a8ef9SStefan Hajnoczi 	}
674a62a8ef9SStefan Hajnoczi 
675a62a8ef9SStefan Hajnoczi 	ret = virtio_find_vqs(vdev, fs->nvqs, vqs, callbacks, names, NULL);
676a62a8ef9SStefan Hajnoczi 	if (ret < 0)
677a62a8ef9SStefan Hajnoczi 		goto out;
678a62a8ef9SStefan Hajnoczi 
679a62a8ef9SStefan Hajnoczi 	for (i = 0; i < fs->nvqs; i++)
680a62a8ef9SStefan Hajnoczi 		fs->vqs[i].vq = vqs[i];
681a62a8ef9SStefan Hajnoczi 
682a62a8ef9SStefan Hajnoczi 	virtio_fs_start_all_queues(fs);
683a62a8ef9SStefan Hajnoczi out:
684a62a8ef9SStefan Hajnoczi 	kfree(names);
685a62a8ef9SStefan Hajnoczi 	kfree(callbacks);
686a62a8ef9SStefan Hajnoczi 	kfree(vqs);
687a62a8ef9SStefan Hajnoczi 	if (ret)
688a62a8ef9SStefan Hajnoczi 		kfree(fs->vqs);
689a62a8ef9SStefan Hajnoczi 	return ret;
690a62a8ef9SStefan Hajnoczi }
691a62a8ef9SStefan Hajnoczi 
692a62a8ef9SStefan Hajnoczi /* Free virtqueues (device must already be reset) */
693a62a8ef9SStefan Hajnoczi static void virtio_fs_cleanup_vqs(struct virtio_device *vdev,
694a62a8ef9SStefan Hajnoczi 				  struct virtio_fs *fs)
695a62a8ef9SStefan Hajnoczi {
696a62a8ef9SStefan Hajnoczi 	vdev->config->del_vqs(vdev);
697a62a8ef9SStefan Hajnoczi }
698a62a8ef9SStefan Hajnoczi 
699*22f3787eSStefan Hajnoczi /* Map a window offset to a page frame number.  The window offset will have
700*22f3787eSStefan Hajnoczi  * been produced by .iomap_begin(), which maps a file offset to a window
701*22f3787eSStefan Hajnoczi  * offset.
702*22f3787eSStefan Hajnoczi  */
703*22f3787eSStefan Hajnoczi static long virtio_fs_direct_access(struct dax_device *dax_dev, pgoff_t pgoff,
704*22f3787eSStefan Hajnoczi 				    long nr_pages, void **kaddr, pfn_t *pfn)
705*22f3787eSStefan Hajnoczi {
706*22f3787eSStefan Hajnoczi 	struct virtio_fs *fs = dax_get_private(dax_dev);
707*22f3787eSStefan Hajnoczi 	phys_addr_t offset = PFN_PHYS(pgoff);
708*22f3787eSStefan Hajnoczi 	size_t max_nr_pages = fs->window_len/PAGE_SIZE - pgoff;
709*22f3787eSStefan Hajnoczi 
710*22f3787eSStefan Hajnoczi 	if (kaddr)
711*22f3787eSStefan Hajnoczi 		*kaddr = fs->window_kaddr + offset;
712*22f3787eSStefan Hajnoczi 	if (pfn)
713*22f3787eSStefan Hajnoczi 		*pfn = phys_to_pfn_t(fs->window_phys_addr + offset,
714*22f3787eSStefan Hajnoczi 					PFN_DEV | PFN_MAP);
715*22f3787eSStefan Hajnoczi 	return nr_pages > max_nr_pages ? max_nr_pages : nr_pages;
716*22f3787eSStefan Hajnoczi }
717*22f3787eSStefan Hajnoczi 
718*22f3787eSStefan Hajnoczi static size_t virtio_fs_copy_from_iter(struct dax_device *dax_dev,
719*22f3787eSStefan Hajnoczi 				       pgoff_t pgoff, void *addr,
720*22f3787eSStefan Hajnoczi 				       size_t bytes, struct iov_iter *i)
721*22f3787eSStefan Hajnoczi {
722*22f3787eSStefan Hajnoczi 	return copy_from_iter(addr, bytes, i);
723*22f3787eSStefan Hajnoczi }
724*22f3787eSStefan Hajnoczi 
725*22f3787eSStefan Hajnoczi static size_t virtio_fs_copy_to_iter(struct dax_device *dax_dev,
726*22f3787eSStefan Hajnoczi 				       pgoff_t pgoff, void *addr,
727*22f3787eSStefan Hajnoczi 				       size_t bytes, struct iov_iter *i)
728*22f3787eSStefan Hajnoczi {
729*22f3787eSStefan Hajnoczi 	return copy_to_iter(addr, bytes, i);
730*22f3787eSStefan Hajnoczi }
731*22f3787eSStefan Hajnoczi 
732*22f3787eSStefan Hajnoczi static int virtio_fs_zero_page_range(struct dax_device *dax_dev,
733*22f3787eSStefan Hajnoczi 				     pgoff_t pgoff, size_t nr_pages)
734*22f3787eSStefan Hajnoczi {
735*22f3787eSStefan Hajnoczi 	long rc;
736*22f3787eSStefan Hajnoczi 	void *kaddr;
737*22f3787eSStefan Hajnoczi 
738*22f3787eSStefan Hajnoczi 	rc = dax_direct_access(dax_dev, pgoff, nr_pages, &kaddr, NULL);
739*22f3787eSStefan Hajnoczi 	if (rc < 0)
740*22f3787eSStefan Hajnoczi 		return rc;
741*22f3787eSStefan Hajnoczi 	memset(kaddr, 0, nr_pages << PAGE_SHIFT);
742*22f3787eSStefan Hajnoczi 	dax_flush(dax_dev, kaddr, nr_pages << PAGE_SHIFT);
743*22f3787eSStefan Hajnoczi 	return 0;
744*22f3787eSStefan Hajnoczi }
745*22f3787eSStefan Hajnoczi 
746*22f3787eSStefan Hajnoczi static const struct dax_operations virtio_fs_dax_ops = {
747*22f3787eSStefan Hajnoczi 	.direct_access = virtio_fs_direct_access,
748*22f3787eSStefan Hajnoczi 	.copy_from_iter = virtio_fs_copy_from_iter,
749*22f3787eSStefan Hajnoczi 	.copy_to_iter = virtio_fs_copy_to_iter,
750*22f3787eSStefan Hajnoczi 	.zero_page_range = virtio_fs_zero_page_range,
751*22f3787eSStefan Hajnoczi };
752*22f3787eSStefan Hajnoczi 
753*22f3787eSStefan Hajnoczi static void virtio_fs_cleanup_dax(void *data)
754*22f3787eSStefan Hajnoczi {
755*22f3787eSStefan Hajnoczi 	struct dax_device *dax_dev = data;
756*22f3787eSStefan Hajnoczi 
757*22f3787eSStefan Hajnoczi 	kill_dax(dax_dev);
758*22f3787eSStefan Hajnoczi 	put_dax(dax_dev);
759*22f3787eSStefan Hajnoczi }
760*22f3787eSStefan Hajnoczi 
761*22f3787eSStefan Hajnoczi static int virtio_fs_setup_dax(struct virtio_device *vdev, struct virtio_fs *fs)
762*22f3787eSStefan Hajnoczi {
763*22f3787eSStefan Hajnoczi 	struct virtio_shm_region cache_reg;
764*22f3787eSStefan Hajnoczi 	struct dev_pagemap *pgmap;
765*22f3787eSStefan Hajnoczi 	bool have_cache;
766*22f3787eSStefan Hajnoczi 
767*22f3787eSStefan Hajnoczi 	if (!IS_ENABLED(CONFIG_FUSE_DAX))
768*22f3787eSStefan Hajnoczi 		return 0;
769*22f3787eSStefan Hajnoczi 
770*22f3787eSStefan Hajnoczi 	/* Get cache region */
771*22f3787eSStefan Hajnoczi 	have_cache = virtio_get_shm_region(vdev, &cache_reg,
772*22f3787eSStefan Hajnoczi 					   (u8)VIRTIO_FS_SHMCAP_ID_CACHE);
773*22f3787eSStefan Hajnoczi 	if (!have_cache) {
774*22f3787eSStefan Hajnoczi 		dev_notice(&vdev->dev, "%s: No cache capability\n", __func__);
775*22f3787eSStefan Hajnoczi 		return 0;
776*22f3787eSStefan Hajnoczi 	}
777*22f3787eSStefan Hajnoczi 
778*22f3787eSStefan Hajnoczi 	if (!devm_request_mem_region(&vdev->dev, cache_reg.addr, cache_reg.len,
779*22f3787eSStefan Hajnoczi 				     dev_name(&vdev->dev))) {
780*22f3787eSStefan Hajnoczi 		dev_warn(&vdev->dev, "could not reserve region addr=0x%llx len=0x%llx\n",
781*22f3787eSStefan Hajnoczi 			 cache_reg.addr, cache_reg.len);
782*22f3787eSStefan Hajnoczi 		return -EBUSY;
783*22f3787eSStefan Hajnoczi 	}
784*22f3787eSStefan Hajnoczi 
785*22f3787eSStefan Hajnoczi 	dev_notice(&vdev->dev, "Cache len: 0x%llx @ 0x%llx\n", cache_reg.len,
786*22f3787eSStefan Hajnoczi 		   cache_reg.addr);
787*22f3787eSStefan Hajnoczi 
788*22f3787eSStefan Hajnoczi 	pgmap = devm_kzalloc(&vdev->dev, sizeof(*pgmap), GFP_KERNEL);
789*22f3787eSStefan Hajnoczi 	if (!pgmap)
790*22f3787eSStefan Hajnoczi 		return -ENOMEM;
791*22f3787eSStefan Hajnoczi 
792*22f3787eSStefan Hajnoczi 	pgmap->type = MEMORY_DEVICE_FS_DAX;
793*22f3787eSStefan Hajnoczi 
794*22f3787eSStefan Hajnoczi 	/* Ideally we would directly use the PCI BAR resource but
795*22f3787eSStefan Hajnoczi 	 * devm_memremap_pages() wants its own copy in pgmap.  So
796*22f3787eSStefan Hajnoczi 	 * initialize a struct resource from scratch (only the start
797*22f3787eSStefan Hajnoczi 	 * and end fields will be used).
798*22f3787eSStefan Hajnoczi 	 */
799*22f3787eSStefan Hajnoczi 	pgmap->res = (struct resource){
800*22f3787eSStefan Hajnoczi 		.name = "virtio-fs dax window",
801*22f3787eSStefan Hajnoczi 		.start = (phys_addr_t) cache_reg.addr,
802*22f3787eSStefan Hajnoczi 		.end = (phys_addr_t) cache_reg.addr + cache_reg.len - 1,
803*22f3787eSStefan Hajnoczi 	};
804*22f3787eSStefan Hajnoczi 
805*22f3787eSStefan Hajnoczi 	fs->window_kaddr = devm_memremap_pages(&vdev->dev, pgmap);
806*22f3787eSStefan Hajnoczi 	if (IS_ERR(fs->window_kaddr))
807*22f3787eSStefan Hajnoczi 		return PTR_ERR(fs->window_kaddr);
808*22f3787eSStefan Hajnoczi 
809*22f3787eSStefan Hajnoczi 	fs->window_phys_addr = (phys_addr_t) cache_reg.addr;
810*22f3787eSStefan Hajnoczi 	fs->window_len = (phys_addr_t) cache_reg.len;
811*22f3787eSStefan Hajnoczi 
812*22f3787eSStefan Hajnoczi 	dev_dbg(&vdev->dev, "%s: window kaddr 0x%px phys_addr 0x%llx len 0x%llx\n",
813*22f3787eSStefan Hajnoczi 		__func__, fs->window_kaddr, cache_reg.addr, cache_reg.len);
814*22f3787eSStefan Hajnoczi 
815*22f3787eSStefan Hajnoczi 	fs->dax_dev = alloc_dax(fs, NULL, &virtio_fs_dax_ops, 0);
816*22f3787eSStefan Hajnoczi 	if (IS_ERR(fs->dax_dev))
817*22f3787eSStefan Hajnoczi 		return PTR_ERR(fs->dax_dev);
818*22f3787eSStefan Hajnoczi 
819*22f3787eSStefan Hajnoczi 	return devm_add_action_or_reset(&vdev->dev, virtio_fs_cleanup_dax,
820*22f3787eSStefan Hajnoczi 					fs->dax_dev);
821*22f3787eSStefan Hajnoczi }
822*22f3787eSStefan Hajnoczi 
823a62a8ef9SStefan Hajnoczi static int virtio_fs_probe(struct virtio_device *vdev)
824a62a8ef9SStefan Hajnoczi {
825a62a8ef9SStefan Hajnoczi 	struct virtio_fs *fs;
826a62a8ef9SStefan Hajnoczi 	int ret;
827a62a8ef9SStefan Hajnoczi 
828a62a8ef9SStefan Hajnoczi 	fs = kzalloc(sizeof(*fs), GFP_KERNEL);
829a62a8ef9SStefan Hajnoczi 	if (!fs)
830a62a8ef9SStefan Hajnoczi 		return -ENOMEM;
831a62a8ef9SStefan Hajnoczi 	kref_init(&fs->refcount);
832a62a8ef9SStefan Hajnoczi 	vdev->priv = fs;
833a62a8ef9SStefan Hajnoczi 
834a62a8ef9SStefan Hajnoczi 	ret = virtio_fs_read_tag(vdev, fs);
835a62a8ef9SStefan Hajnoczi 	if (ret < 0)
836a62a8ef9SStefan Hajnoczi 		goto out;
837a62a8ef9SStefan Hajnoczi 
838a62a8ef9SStefan Hajnoczi 	ret = virtio_fs_setup_vqs(vdev, fs);
839a62a8ef9SStefan Hajnoczi 	if (ret < 0)
840a62a8ef9SStefan Hajnoczi 		goto out;
841a62a8ef9SStefan Hajnoczi 
842a62a8ef9SStefan Hajnoczi 	/* TODO vq affinity */
843a62a8ef9SStefan Hajnoczi 
844*22f3787eSStefan Hajnoczi 	ret = virtio_fs_setup_dax(vdev, fs);
845*22f3787eSStefan Hajnoczi 	if (ret < 0)
846*22f3787eSStefan Hajnoczi 		goto out_vqs;
847*22f3787eSStefan Hajnoczi 
848a62a8ef9SStefan Hajnoczi 	/* Bring the device online in case the filesystem is mounted and
849a62a8ef9SStefan Hajnoczi 	 * requests need to be sent before we return.
850a62a8ef9SStefan Hajnoczi 	 */
851a62a8ef9SStefan Hajnoczi 	virtio_device_ready(vdev);
852a62a8ef9SStefan Hajnoczi 
853a62a8ef9SStefan Hajnoczi 	ret = virtio_fs_add_instance(fs);
854a62a8ef9SStefan Hajnoczi 	if (ret < 0)
855a62a8ef9SStefan Hajnoczi 		goto out_vqs;
856a62a8ef9SStefan Hajnoczi 
857a62a8ef9SStefan Hajnoczi 	return 0;
858a62a8ef9SStefan Hajnoczi 
859a62a8ef9SStefan Hajnoczi out_vqs:
860a62a8ef9SStefan Hajnoczi 	vdev->config->reset(vdev);
861a62a8ef9SStefan Hajnoczi 	virtio_fs_cleanup_vqs(vdev, fs);
862a62a8ef9SStefan Hajnoczi 
863a62a8ef9SStefan Hajnoczi out:
864a62a8ef9SStefan Hajnoczi 	vdev->priv = NULL;
865a62a8ef9SStefan Hajnoczi 	kfree(fs);
866a62a8ef9SStefan Hajnoczi 	return ret;
867a62a8ef9SStefan Hajnoczi }
868a62a8ef9SStefan Hajnoczi 
869a62a8ef9SStefan Hajnoczi static void virtio_fs_stop_all_queues(struct virtio_fs *fs)
870a62a8ef9SStefan Hajnoczi {
871a62a8ef9SStefan Hajnoczi 	struct virtio_fs_vq *fsvq;
872a62a8ef9SStefan Hajnoczi 	int i;
873a62a8ef9SStefan Hajnoczi 
874a62a8ef9SStefan Hajnoczi 	for (i = 0; i < fs->nvqs; i++) {
875a62a8ef9SStefan Hajnoczi 		fsvq = &fs->vqs[i];
876a62a8ef9SStefan Hajnoczi 		spin_lock(&fsvq->lock);
877a62a8ef9SStefan Hajnoczi 		fsvq->connected = false;
878a62a8ef9SStefan Hajnoczi 		spin_unlock(&fsvq->lock);
879a62a8ef9SStefan Hajnoczi 	}
880a62a8ef9SStefan Hajnoczi }
881a62a8ef9SStefan Hajnoczi 
882a62a8ef9SStefan Hajnoczi static void virtio_fs_remove(struct virtio_device *vdev)
883a62a8ef9SStefan Hajnoczi {
884a62a8ef9SStefan Hajnoczi 	struct virtio_fs *fs = vdev->priv;
885a62a8ef9SStefan Hajnoczi 
886a62a8ef9SStefan Hajnoczi 	mutex_lock(&virtio_fs_mutex);
887a62a8ef9SStefan Hajnoczi 	/* This device is going away. No one should get new reference */
888a62a8ef9SStefan Hajnoczi 	list_del_init(&fs->list);
889a62a8ef9SStefan Hajnoczi 	virtio_fs_stop_all_queues(fs);
890724c15a4SVivek Goyal 	virtio_fs_drain_all_queues_locked(fs);
891a62a8ef9SStefan Hajnoczi 	vdev->config->reset(vdev);
892a62a8ef9SStefan Hajnoczi 	virtio_fs_cleanup_vqs(vdev, fs);
893a62a8ef9SStefan Hajnoczi 
894a62a8ef9SStefan Hajnoczi 	vdev->priv = NULL;
895a62a8ef9SStefan Hajnoczi 	/* Put device reference on virtio_fs object */
896a62a8ef9SStefan Hajnoczi 	virtio_fs_put(fs);
897a62a8ef9SStefan Hajnoczi 	mutex_unlock(&virtio_fs_mutex);
898a62a8ef9SStefan Hajnoczi }
899a62a8ef9SStefan Hajnoczi 
900a62a8ef9SStefan Hajnoczi #ifdef CONFIG_PM_SLEEP
901a62a8ef9SStefan Hajnoczi static int virtio_fs_freeze(struct virtio_device *vdev)
902a62a8ef9SStefan Hajnoczi {
903a62a8ef9SStefan Hajnoczi 	/* TODO need to save state here */
904a62a8ef9SStefan Hajnoczi 	pr_warn("virtio-fs: suspend/resume not yet supported\n");
905a62a8ef9SStefan Hajnoczi 	return -EOPNOTSUPP;
906a62a8ef9SStefan Hajnoczi }
907a62a8ef9SStefan Hajnoczi 
908a62a8ef9SStefan Hajnoczi static int virtio_fs_restore(struct virtio_device *vdev)
909a62a8ef9SStefan Hajnoczi {
910a62a8ef9SStefan Hajnoczi 	 /* TODO need to restore state here */
911a62a8ef9SStefan Hajnoczi 	return 0;
912a62a8ef9SStefan Hajnoczi }
913a62a8ef9SStefan Hajnoczi #endif /* CONFIG_PM_SLEEP */
914a62a8ef9SStefan Hajnoczi 
91500929447SYueHaibing static const struct virtio_device_id id_table[] = {
916a62a8ef9SStefan Hajnoczi 	{ VIRTIO_ID_FS, VIRTIO_DEV_ANY_ID },
917a62a8ef9SStefan Hajnoczi 	{},
918a62a8ef9SStefan Hajnoczi };
919a62a8ef9SStefan Hajnoczi 
92000929447SYueHaibing static const unsigned int feature_table[] = {};
921a62a8ef9SStefan Hajnoczi 
922a62a8ef9SStefan Hajnoczi static struct virtio_driver virtio_fs_driver = {
923a62a8ef9SStefan Hajnoczi 	.driver.name		= KBUILD_MODNAME,
924a62a8ef9SStefan Hajnoczi 	.driver.owner		= THIS_MODULE,
925a62a8ef9SStefan Hajnoczi 	.id_table		= id_table,
926a62a8ef9SStefan Hajnoczi 	.feature_table		= feature_table,
927a62a8ef9SStefan Hajnoczi 	.feature_table_size	= ARRAY_SIZE(feature_table),
928a62a8ef9SStefan Hajnoczi 	.probe			= virtio_fs_probe,
929a62a8ef9SStefan Hajnoczi 	.remove			= virtio_fs_remove,
930a62a8ef9SStefan Hajnoczi #ifdef CONFIG_PM_SLEEP
931a62a8ef9SStefan Hajnoczi 	.freeze			= virtio_fs_freeze,
932a62a8ef9SStefan Hajnoczi 	.restore		= virtio_fs_restore,
933a62a8ef9SStefan Hajnoczi #endif
934a62a8ef9SStefan Hajnoczi };
935a62a8ef9SStefan Hajnoczi 
936a62a8ef9SStefan Hajnoczi static void virtio_fs_wake_forget_and_unlock(struct fuse_iqueue *fiq)
937a62a8ef9SStefan Hajnoczi __releases(fiq->lock)
938a62a8ef9SStefan Hajnoczi {
939a62a8ef9SStefan Hajnoczi 	struct fuse_forget_link *link;
940a62a8ef9SStefan Hajnoczi 	struct virtio_fs_forget *forget;
9411efcf39eSVivek Goyal 	struct virtio_fs_forget_req *req;
942a62a8ef9SStefan Hajnoczi 	struct virtio_fs *fs;
943a62a8ef9SStefan Hajnoczi 	struct virtio_fs_vq *fsvq;
944a62a8ef9SStefan Hajnoczi 	u64 unique;
945a62a8ef9SStefan Hajnoczi 
946a62a8ef9SStefan Hajnoczi 	link = fuse_dequeue_forget(fiq, 1, NULL);
947a62a8ef9SStefan Hajnoczi 	unique = fuse_get_unique(fiq);
948a62a8ef9SStefan Hajnoczi 
949a62a8ef9SStefan Hajnoczi 	fs = fiq->priv;
950a62a8ef9SStefan Hajnoczi 	fsvq = &fs->vqs[VQ_HIPRIO];
951a62a8ef9SStefan Hajnoczi 	spin_unlock(&fiq->lock);
952a62a8ef9SStefan Hajnoczi 
953a62a8ef9SStefan Hajnoczi 	/* Allocate a buffer for the request */
954a62a8ef9SStefan Hajnoczi 	forget = kmalloc(sizeof(*forget), GFP_NOFS | __GFP_NOFAIL);
9551efcf39eSVivek Goyal 	req = &forget->req;
956a62a8ef9SStefan Hajnoczi 
9571efcf39eSVivek Goyal 	req->ih = (struct fuse_in_header){
958a62a8ef9SStefan Hajnoczi 		.opcode = FUSE_FORGET,
959a62a8ef9SStefan Hajnoczi 		.nodeid = link->forget_one.nodeid,
960a62a8ef9SStefan Hajnoczi 		.unique = unique,
9611efcf39eSVivek Goyal 		.len = sizeof(*req),
962a62a8ef9SStefan Hajnoczi 	};
9631efcf39eSVivek Goyal 	req->arg = (struct fuse_forget_in){
964a62a8ef9SStefan Hajnoczi 		.nlookup = link->forget_one.nlookup,
965a62a8ef9SStefan Hajnoczi 	};
966a62a8ef9SStefan Hajnoczi 
96758ada94fSVivek Goyal 	send_forget_request(fsvq, forget, false);
968a62a8ef9SStefan Hajnoczi 	kfree(link);
969a62a8ef9SStefan Hajnoczi }
970a62a8ef9SStefan Hajnoczi 
971a62a8ef9SStefan Hajnoczi static void virtio_fs_wake_interrupt_and_unlock(struct fuse_iqueue *fiq)
972a62a8ef9SStefan Hajnoczi __releases(fiq->lock)
973a62a8ef9SStefan Hajnoczi {
974a62a8ef9SStefan Hajnoczi 	/*
975a62a8ef9SStefan Hajnoczi 	 * TODO interrupts.
976a62a8ef9SStefan Hajnoczi 	 *
977a62a8ef9SStefan Hajnoczi 	 * Normal fs operations on a local filesystems aren't interruptible.
978a62a8ef9SStefan Hajnoczi 	 * Exceptions are blocking lock operations; for example fcntl(F_SETLKW)
979a62a8ef9SStefan Hajnoczi 	 * with shared lock between host and guest.
980a62a8ef9SStefan Hajnoczi 	 */
981a62a8ef9SStefan Hajnoczi 	spin_unlock(&fiq->lock);
982a62a8ef9SStefan Hajnoczi }
983a62a8ef9SStefan Hajnoczi 
984a62a8ef9SStefan Hajnoczi /* Return the number of scatter-gather list elements required */
985a62a8ef9SStefan Hajnoczi static unsigned int sg_count_fuse_req(struct fuse_req *req)
986a62a8ef9SStefan Hajnoczi {
987a62a8ef9SStefan Hajnoczi 	struct fuse_args *args = req->args;
988a62a8ef9SStefan Hajnoczi 	struct fuse_args_pages *ap = container_of(args, typeof(*ap), args);
989a62a8ef9SStefan Hajnoczi 	unsigned int total_sgs = 1 /* fuse_in_header */;
990a62a8ef9SStefan Hajnoczi 
991a62a8ef9SStefan Hajnoczi 	if (args->in_numargs - args->in_pages)
992a62a8ef9SStefan Hajnoczi 		total_sgs += 1;
993a62a8ef9SStefan Hajnoczi 
994a62a8ef9SStefan Hajnoczi 	if (args->in_pages)
995a62a8ef9SStefan Hajnoczi 		total_sgs += ap->num_pages;
996a62a8ef9SStefan Hajnoczi 
997a62a8ef9SStefan Hajnoczi 	if (!test_bit(FR_ISREPLY, &req->flags))
998a62a8ef9SStefan Hajnoczi 		return total_sgs;
999a62a8ef9SStefan Hajnoczi 
1000a62a8ef9SStefan Hajnoczi 	total_sgs += 1 /* fuse_out_header */;
1001a62a8ef9SStefan Hajnoczi 
1002a62a8ef9SStefan Hajnoczi 	if (args->out_numargs - args->out_pages)
1003a62a8ef9SStefan Hajnoczi 		total_sgs += 1;
1004a62a8ef9SStefan Hajnoczi 
1005a62a8ef9SStefan Hajnoczi 	if (args->out_pages)
1006a62a8ef9SStefan Hajnoczi 		total_sgs += ap->num_pages;
1007a62a8ef9SStefan Hajnoczi 
1008a62a8ef9SStefan Hajnoczi 	return total_sgs;
1009a62a8ef9SStefan Hajnoczi }
1010a62a8ef9SStefan Hajnoczi 
1011a62a8ef9SStefan Hajnoczi /* Add pages to scatter-gather list and return number of elements used */
1012a62a8ef9SStefan Hajnoczi static unsigned int sg_init_fuse_pages(struct scatterlist *sg,
1013a62a8ef9SStefan Hajnoczi 				       struct page **pages,
1014a62a8ef9SStefan Hajnoczi 				       struct fuse_page_desc *page_descs,
1015a62a8ef9SStefan Hajnoczi 				       unsigned int num_pages,
1016a62a8ef9SStefan Hajnoczi 				       unsigned int total_len)
1017a62a8ef9SStefan Hajnoczi {
1018a62a8ef9SStefan Hajnoczi 	unsigned int i;
1019a62a8ef9SStefan Hajnoczi 	unsigned int this_len;
1020a62a8ef9SStefan Hajnoczi 
1021a62a8ef9SStefan Hajnoczi 	for (i = 0; i < num_pages && total_len; i++) {
1022a62a8ef9SStefan Hajnoczi 		sg_init_table(&sg[i], 1);
1023a62a8ef9SStefan Hajnoczi 		this_len =  min(page_descs[i].length, total_len);
1024a62a8ef9SStefan Hajnoczi 		sg_set_page(&sg[i], pages[i], this_len, page_descs[i].offset);
1025a62a8ef9SStefan Hajnoczi 		total_len -= this_len;
1026a62a8ef9SStefan Hajnoczi 	}
1027a62a8ef9SStefan Hajnoczi 
1028a62a8ef9SStefan Hajnoczi 	return i;
1029a62a8ef9SStefan Hajnoczi }
1030a62a8ef9SStefan Hajnoczi 
1031a62a8ef9SStefan Hajnoczi /* Add args to scatter-gather list and return number of elements used */
1032a62a8ef9SStefan Hajnoczi static unsigned int sg_init_fuse_args(struct scatterlist *sg,
1033a62a8ef9SStefan Hajnoczi 				      struct fuse_req *req,
1034a62a8ef9SStefan Hajnoczi 				      struct fuse_arg *args,
1035a62a8ef9SStefan Hajnoczi 				      unsigned int numargs,
1036a62a8ef9SStefan Hajnoczi 				      bool argpages,
1037a62a8ef9SStefan Hajnoczi 				      void *argbuf,
1038a62a8ef9SStefan Hajnoczi 				      unsigned int *len_used)
1039a62a8ef9SStefan Hajnoczi {
1040a62a8ef9SStefan Hajnoczi 	struct fuse_args_pages *ap = container_of(req->args, typeof(*ap), args);
1041a62a8ef9SStefan Hajnoczi 	unsigned int total_sgs = 0;
1042a62a8ef9SStefan Hajnoczi 	unsigned int len;
1043a62a8ef9SStefan Hajnoczi 
1044a62a8ef9SStefan Hajnoczi 	len = fuse_len_args(numargs - argpages, args);
1045a62a8ef9SStefan Hajnoczi 	if (len)
1046a62a8ef9SStefan Hajnoczi 		sg_init_one(&sg[total_sgs++], argbuf, len);
1047a62a8ef9SStefan Hajnoczi 
1048a62a8ef9SStefan Hajnoczi 	if (argpages)
1049a62a8ef9SStefan Hajnoczi 		total_sgs += sg_init_fuse_pages(&sg[total_sgs],
1050a62a8ef9SStefan Hajnoczi 						ap->pages, ap->descs,
1051a62a8ef9SStefan Hajnoczi 						ap->num_pages,
1052a62a8ef9SStefan Hajnoczi 						args[numargs - 1].size);
1053a62a8ef9SStefan Hajnoczi 
1054a62a8ef9SStefan Hajnoczi 	if (len_used)
1055a62a8ef9SStefan Hajnoczi 		*len_used = len;
1056a62a8ef9SStefan Hajnoczi 
1057a62a8ef9SStefan Hajnoczi 	return total_sgs;
1058a62a8ef9SStefan Hajnoczi }
1059a62a8ef9SStefan Hajnoczi 
1060a62a8ef9SStefan Hajnoczi /* Add a request to a virtqueue and kick the device */
1061a62a8ef9SStefan Hajnoczi static int virtio_fs_enqueue_req(struct virtio_fs_vq *fsvq,
1062a9bfd9ddSVivek Goyal 				 struct fuse_req *req, bool in_flight)
1063a62a8ef9SStefan Hajnoczi {
1064a62a8ef9SStefan Hajnoczi 	/* requests need at least 4 elements */
1065a62a8ef9SStefan Hajnoczi 	struct scatterlist *stack_sgs[6];
1066a62a8ef9SStefan Hajnoczi 	struct scatterlist stack_sg[ARRAY_SIZE(stack_sgs)];
1067a62a8ef9SStefan Hajnoczi 	struct scatterlist **sgs = stack_sgs;
1068a62a8ef9SStefan Hajnoczi 	struct scatterlist *sg = stack_sg;
1069a62a8ef9SStefan Hajnoczi 	struct virtqueue *vq;
1070a62a8ef9SStefan Hajnoczi 	struct fuse_args *args = req->args;
1071a62a8ef9SStefan Hajnoczi 	unsigned int argbuf_used = 0;
1072a62a8ef9SStefan Hajnoczi 	unsigned int out_sgs = 0;
1073a62a8ef9SStefan Hajnoczi 	unsigned int in_sgs = 0;
1074a62a8ef9SStefan Hajnoczi 	unsigned int total_sgs;
1075a62a8ef9SStefan Hajnoczi 	unsigned int i;
1076a62a8ef9SStefan Hajnoczi 	int ret;
1077a62a8ef9SStefan Hajnoczi 	bool notify;
10785dbe190fSVivek Goyal 	struct fuse_pqueue *fpq;
1079a62a8ef9SStefan Hajnoczi 
1080a62a8ef9SStefan Hajnoczi 	/* Does the sglist fit on the stack? */
1081a62a8ef9SStefan Hajnoczi 	total_sgs = sg_count_fuse_req(req);
1082a62a8ef9SStefan Hajnoczi 	if (total_sgs > ARRAY_SIZE(stack_sgs)) {
1083a62a8ef9SStefan Hajnoczi 		sgs = kmalloc_array(total_sgs, sizeof(sgs[0]), GFP_ATOMIC);
1084a62a8ef9SStefan Hajnoczi 		sg = kmalloc_array(total_sgs, sizeof(sg[0]), GFP_ATOMIC);
1085a62a8ef9SStefan Hajnoczi 		if (!sgs || !sg) {
1086a62a8ef9SStefan Hajnoczi 			ret = -ENOMEM;
1087a62a8ef9SStefan Hajnoczi 			goto out;
1088a62a8ef9SStefan Hajnoczi 		}
1089a62a8ef9SStefan Hajnoczi 	}
1090a62a8ef9SStefan Hajnoczi 
1091a62a8ef9SStefan Hajnoczi 	/* Use a bounce buffer since stack args cannot be mapped */
1092a62a8ef9SStefan Hajnoczi 	ret = copy_args_to_argbuf(req);
1093a62a8ef9SStefan Hajnoczi 	if (ret < 0)
1094a62a8ef9SStefan Hajnoczi 		goto out;
1095a62a8ef9SStefan Hajnoczi 
1096a62a8ef9SStefan Hajnoczi 	/* Request elements */
1097a62a8ef9SStefan Hajnoczi 	sg_init_one(&sg[out_sgs++], &req->in.h, sizeof(req->in.h));
1098a62a8ef9SStefan Hajnoczi 	out_sgs += sg_init_fuse_args(&sg[out_sgs], req,
1099a62a8ef9SStefan Hajnoczi 				     (struct fuse_arg *)args->in_args,
1100a62a8ef9SStefan Hajnoczi 				     args->in_numargs, args->in_pages,
1101a62a8ef9SStefan Hajnoczi 				     req->argbuf, &argbuf_used);
1102a62a8ef9SStefan Hajnoczi 
1103a62a8ef9SStefan Hajnoczi 	/* Reply elements */
1104a62a8ef9SStefan Hajnoczi 	if (test_bit(FR_ISREPLY, &req->flags)) {
1105a62a8ef9SStefan Hajnoczi 		sg_init_one(&sg[out_sgs + in_sgs++],
1106a62a8ef9SStefan Hajnoczi 			    &req->out.h, sizeof(req->out.h));
1107a62a8ef9SStefan Hajnoczi 		in_sgs += sg_init_fuse_args(&sg[out_sgs + in_sgs], req,
1108a62a8ef9SStefan Hajnoczi 					    args->out_args, args->out_numargs,
1109a62a8ef9SStefan Hajnoczi 					    args->out_pages,
1110a62a8ef9SStefan Hajnoczi 					    req->argbuf + argbuf_used, NULL);
1111a62a8ef9SStefan Hajnoczi 	}
1112a62a8ef9SStefan Hajnoczi 
1113a62a8ef9SStefan Hajnoczi 	WARN_ON(out_sgs + in_sgs != total_sgs);
1114a62a8ef9SStefan Hajnoczi 
1115a62a8ef9SStefan Hajnoczi 	for (i = 0; i < total_sgs; i++)
1116a62a8ef9SStefan Hajnoczi 		sgs[i] = &sg[i];
1117a62a8ef9SStefan Hajnoczi 
1118a62a8ef9SStefan Hajnoczi 	spin_lock(&fsvq->lock);
1119a62a8ef9SStefan Hajnoczi 
1120a62a8ef9SStefan Hajnoczi 	if (!fsvq->connected) {
1121a62a8ef9SStefan Hajnoczi 		spin_unlock(&fsvq->lock);
1122a62a8ef9SStefan Hajnoczi 		ret = -ENOTCONN;
1123a62a8ef9SStefan Hajnoczi 		goto out;
1124a62a8ef9SStefan Hajnoczi 	}
1125a62a8ef9SStefan Hajnoczi 
1126a62a8ef9SStefan Hajnoczi 	vq = fsvq->vq;
1127a62a8ef9SStefan Hajnoczi 	ret = virtqueue_add_sgs(vq, sgs, out_sgs, in_sgs, req, GFP_ATOMIC);
1128a62a8ef9SStefan Hajnoczi 	if (ret < 0) {
1129a62a8ef9SStefan Hajnoczi 		spin_unlock(&fsvq->lock);
1130a62a8ef9SStefan Hajnoczi 		goto out;
1131a62a8ef9SStefan Hajnoczi 	}
1132a62a8ef9SStefan Hajnoczi 
11335dbe190fSVivek Goyal 	/* Request successfully sent. */
11345dbe190fSVivek Goyal 	fpq = &fsvq->fud->pq;
11355dbe190fSVivek Goyal 	spin_lock(&fpq->lock);
11365dbe190fSVivek Goyal 	list_add_tail(&req->list, fpq->processing);
11375dbe190fSVivek Goyal 	spin_unlock(&fpq->lock);
11385dbe190fSVivek Goyal 	set_bit(FR_SENT, &req->flags);
11395dbe190fSVivek Goyal 	/* matches barrier in request_wait_answer() */
11405dbe190fSVivek Goyal 	smp_mb__after_atomic();
11415dbe190fSVivek Goyal 
1142a9bfd9ddSVivek Goyal 	if (!in_flight)
1143c17ea009SVivek Goyal 		inc_in_flight_req(fsvq);
1144a62a8ef9SStefan Hajnoczi 	notify = virtqueue_kick_prepare(vq);
1145a62a8ef9SStefan Hajnoczi 
1146a62a8ef9SStefan Hajnoczi 	spin_unlock(&fsvq->lock);
1147a62a8ef9SStefan Hajnoczi 
1148a62a8ef9SStefan Hajnoczi 	if (notify)
1149a62a8ef9SStefan Hajnoczi 		virtqueue_notify(vq);
1150a62a8ef9SStefan Hajnoczi 
1151a62a8ef9SStefan Hajnoczi out:
1152a62a8ef9SStefan Hajnoczi 	if (ret < 0 && req->argbuf) {
1153a62a8ef9SStefan Hajnoczi 		kfree(req->argbuf);
1154a62a8ef9SStefan Hajnoczi 		req->argbuf = NULL;
1155a62a8ef9SStefan Hajnoczi 	}
1156a62a8ef9SStefan Hajnoczi 	if (sgs != stack_sgs) {
1157a62a8ef9SStefan Hajnoczi 		kfree(sgs);
1158a62a8ef9SStefan Hajnoczi 		kfree(sg);
1159a62a8ef9SStefan Hajnoczi 	}
1160a62a8ef9SStefan Hajnoczi 
1161a62a8ef9SStefan Hajnoczi 	return ret;
1162a62a8ef9SStefan Hajnoczi }
1163a62a8ef9SStefan Hajnoczi 
1164a62a8ef9SStefan Hajnoczi static void virtio_fs_wake_pending_and_unlock(struct fuse_iqueue *fiq)
1165a62a8ef9SStefan Hajnoczi __releases(fiq->lock)
1166a62a8ef9SStefan Hajnoczi {
1167a62a8ef9SStefan Hajnoczi 	unsigned int queue_id = VQ_REQUEST; /* TODO multiqueue */
1168a62a8ef9SStefan Hajnoczi 	struct virtio_fs *fs;
1169a62a8ef9SStefan Hajnoczi 	struct fuse_req *req;
117051fecdd2SVivek Goyal 	struct virtio_fs_vq *fsvq;
1171a62a8ef9SStefan Hajnoczi 	int ret;
1172a62a8ef9SStefan Hajnoczi 
1173a62a8ef9SStefan Hajnoczi 	WARN_ON(list_empty(&fiq->pending));
1174a62a8ef9SStefan Hajnoczi 	req = list_last_entry(&fiq->pending, struct fuse_req, list);
1175a62a8ef9SStefan Hajnoczi 	clear_bit(FR_PENDING, &req->flags);
1176a62a8ef9SStefan Hajnoczi 	list_del_init(&req->list);
1177a62a8ef9SStefan Hajnoczi 	WARN_ON(!list_empty(&fiq->pending));
1178a62a8ef9SStefan Hajnoczi 	spin_unlock(&fiq->lock);
1179a62a8ef9SStefan Hajnoczi 
1180a62a8ef9SStefan Hajnoczi 	fs = fiq->priv;
1181a62a8ef9SStefan Hajnoczi 
1182a62a8ef9SStefan Hajnoczi 	pr_debug("%s: opcode %u unique %#llx nodeid %#llx in.len %u out.len %u\n",
1183a62a8ef9SStefan Hajnoczi 		  __func__, req->in.h.opcode, req->in.h.unique,
1184a62a8ef9SStefan Hajnoczi 		 req->in.h.nodeid, req->in.h.len,
1185a62a8ef9SStefan Hajnoczi 		 fuse_len_args(req->args->out_numargs, req->args->out_args));
1186a62a8ef9SStefan Hajnoczi 
118751fecdd2SVivek Goyal 	fsvq = &fs->vqs[queue_id];
1188a9bfd9ddSVivek Goyal 	ret = virtio_fs_enqueue_req(fsvq, req, false);
1189a62a8ef9SStefan Hajnoczi 	if (ret < 0) {
1190a62a8ef9SStefan Hajnoczi 		if (ret == -ENOMEM || ret == -ENOSPC) {
1191a9bfd9ddSVivek Goyal 			/*
1192a9bfd9ddSVivek Goyal 			 * Virtqueue full. Retry submission from worker
1193a9bfd9ddSVivek Goyal 			 * context as we might be holding fc->bg_lock.
1194a9bfd9ddSVivek Goyal 			 */
1195a9bfd9ddSVivek Goyal 			spin_lock(&fsvq->lock);
1196a9bfd9ddSVivek Goyal 			list_add_tail(&req->list, &fsvq->queued_reqs);
1197a9bfd9ddSVivek Goyal 			inc_in_flight_req(fsvq);
1198a9bfd9ddSVivek Goyal 			schedule_delayed_work(&fsvq->dispatch_work,
1199a9bfd9ddSVivek Goyal 						msecs_to_jiffies(1));
1200a9bfd9ddSVivek Goyal 			spin_unlock(&fsvq->lock);
1201a9bfd9ddSVivek Goyal 			return;
1202a62a8ef9SStefan Hajnoczi 		}
1203a62a8ef9SStefan Hajnoczi 		req->out.h.error = ret;
1204a62a8ef9SStefan Hajnoczi 		pr_err("virtio-fs: virtio_fs_enqueue_req() failed %d\n", ret);
120551fecdd2SVivek Goyal 
120651fecdd2SVivek Goyal 		/* Can't end request in submission context. Use a worker */
120751fecdd2SVivek Goyal 		spin_lock(&fsvq->lock);
120851fecdd2SVivek Goyal 		list_add_tail(&req->list, &fsvq->end_reqs);
120951fecdd2SVivek Goyal 		schedule_delayed_work(&fsvq->dispatch_work, 0);
121051fecdd2SVivek Goyal 		spin_unlock(&fsvq->lock);
1211a62a8ef9SStefan Hajnoczi 		return;
1212a62a8ef9SStefan Hajnoczi 	}
1213a62a8ef9SStefan Hajnoczi }
1214a62a8ef9SStefan Hajnoczi 
121500929447SYueHaibing static const struct fuse_iqueue_ops virtio_fs_fiq_ops = {
1216a62a8ef9SStefan Hajnoczi 	.wake_forget_and_unlock		= virtio_fs_wake_forget_and_unlock,
1217a62a8ef9SStefan Hajnoczi 	.wake_interrupt_and_unlock	= virtio_fs_wake_interrupt_and_unlock,
1218a62a8ef9SStefan Hajnoczi 	.wake_pending_and_unlock	= virtio_fs_wake_pending_and_unlock,
1219a62a8ef9SStefan Hajnoczi 	.release			= virtio_fs_fiq_release,
1220a62a8ef9SStefan Hajnoczi };
1221a62a8ef9SStefan Hajnoczi 
1222a62a8ef9SStefan Hajnoczi static int virtio_fs_fill_super(struct super_block *sb)
1223a62a8ef9SStefan Hajnoczi {
1224a62a8ef9SStefan Hajnoczi 	struct fuse_conn *fc = get_fuse_conn_super(sb);
1225a62a8ef9SStefan Hajnoczi 	struct virtio_fs *fs = fc->iq.priv;
1226a62a8ef9SStefan Hajnoczi 	unsigned int i;
1227a62a8ef9SStefan Hajnoczi 	int err;
1228a62a8ef9SStefan Hajnoczi 	struct fuse_fs_context ctx = {
1229a62a8ef9SStefan Hajnoczi 		.rootmode = S_IFDIR,
1230a62a8ef9SStefan Hajnoczi 		.default_permissions = 1,
1231a62a8ef9SStefan Hajnoczi 		.allow_other = 1,
1232a62a8ef9SStefan Hajnoczi 		.max_read = UINT_MAX,
1233a62a8ef9SStefan Hajnoczi 		.blksize = 512,
1234a62a8ef9SStefan Hajnoczi 		.destroy = true,
1235a62a8ef9SStefan Hajnoczi 		.no_control = true,
1236a62a8ef9SStefan Hajnoczi 		.no_force_umount = true,
1237a62a8ef9SStefan Hajnoczi 	};
1238a62a8ef9SStefan Hajnoczi 
1239a62a8ef9SStefan Hajnoczi 	mutex_lock(&virtio_fs_mutex);
1240a62a8ef9SStefan Hajnoczi 
1241a62a8ef9SStefan Hajnoczi 	/* After holding mutex, make sure virtiofs device is still there.
1242a62a8ef9SStefan Hajnoczi 	 * Though we are holding a reference to it, drive ->remove might
1243a62a8ef9SStefan Hajnoczi 	 * still have cleaned up virtual queues. In that case bail out.
1244a62a8ef9SStefan Hajnoczi 	 */
1245a62a8ef9SStefan Hajnoczi 	err = -EINVAL;
1246a62a8ef9SStefan Hajnoczi 	if (list_empty(&fs->list)) {
1247a62a8ef9SStefan Hajnoczi 		pr_info("virtio-fs: tag <%s> not found\n", fs->tag);
1248a62a8ef9SStefan Hajnoczi 		goto err;
1249a62a8ef9SStefan Hajnoczi 	}
1250a62a8ef9SStefan Hajnoczi 
1251a62a8ef9SStefan Hajnoczi 	err = -ENOMEM;
1252a62a8ef9SStefan Hajnoczi 	/* Allocate fuse_dev for hiprio and notification queues */
12537fd3abfaSVivek Goyal 	for (i = 0; i < fs->nvqs; i++) {
1254a62a8ef9SStefan Hajnoczi 		struct virtio_fs_vq *fsvq = &fs->vqs[i];
1255a62a8ef9SStefan Hajnoczi 
1256a62a8ef9SStefan Hajnoczi 		fsvq->fud = fuse_dev_alloc();
1257a62a8ef9SStefan Hajnoczi 		if (!fsvq->fud)
1258a62a8ef9SStefan Hajnoczi 			goto err_free_fuse_devs;
1259a62a8ef9SStefan Hajnoczi 	}
1260a62a8ef9SStefan Hajnoczi 
12617fd3abfaSVivek Goyal 	/* virtiofs allocates and installs its own fuse devices */
12627fd3abfaSVivek Goyal 	ctx.fudptr = NULL;
1263a62a8ef9SStefan Hajnoczi 	err = fuse_fill_super_common(sb, &ctx);
1264a62a8ef9SStefan Hajnoczi 	if (err < 0)
1265a62a8ef9SStefan Hajnoczi 		goto err_free_fuse_devs;
1266a62a8ef9SStefan Hajnoczi 
1267a62a8ef9SStefan Hajnoczi 	for (i = 0; i < fs->nvqs; i++) {
1268a62a8ef9SStefan Hajnoczi 		struct virtio_fs_vq *fsvq = &fs->vqs[i];
1269a62a8ef9SStefan Hajnoczi 
1270a62a8ef9SStefan Hajnoczi 		fuse_dev_install(fsvq->fud, fc);
1271a62a8ef9SStefan Hajnoczi 	}
1272a62a8ef9SStefan Hajnoczi 
1273a62a8ef9SStefan Hajnoczi 	/* Previous unmount will stop all queues. Start these again */
1274a62a8ef9SStefan Hajnoczi 	virtio_fs_start_all_queues(fs);
1275a62a8ef9SStefan Hajnoczi 	fuse_send_init(fc);
1276a62a8ef9SStefan Hajnoczi 	mutex_unlock(&virtio_fs_mutex);
1277a62a8ef9SStefan Hajnoczi 	return 0;
1278a62a8ef9SStefan Hajnoczi 
1279a62a8ef9SStefan Hajnoczi err_free_fuse_devs:
1280a62a8ef9SStefan Hajnoczi 	virtio_fs_free_devs(fs);
1281a62a8ef9SStefan Hajnoczi err:
1282a62a8ef9SStefan Hajnoczi 	mutex_unlock(&virtio_fs_mutex);
1283a62a8ef9SStefan Hajnoczi 	return err;
1284a62a8ef9SStefan Hajnoczi }
1285a62a8ef9SStefan Hajnoczi 
1286a62a8ef9SStefan Hajnoczi static void virtio_kill_sb(struct super_block *sb)
1287a62a8ef9SStefan Hajnoczi {
1288a62a8ef9SStefan Hajnoczi 	struct fuse_conn *fc = get_fuse_conn_super(sb);
1289a62a8ef9SStefan Hajnoczi 	struct virtio_fs *vfs;
1290a62a8ef9SStefan Hajnoczi 	struct virtio_fs_vq *fsvq;
1291a62a8ef9SStefan Hajnoczi 
1292a62a8ef9SStefan Hajnoczi 	/* If mount failed, we can still be called without any fc */
1293a62a8ef9SStefan Hajnoczi 	if (!fc)
1294a62a8ef9SStefan Hajnoczi 		return fuse_kill_sb_anon(sb);
1295a62a8ef9SStefan Hajnoczi 
1296a62a8ef9SStefan Hajnoczi 	vfs = fc->iq.priv;
1297a62a8ef9SStefan Hajnoczi 	fsvq = &vfs->vqs[VQ_HIPRIO];
1298a62a8ef9SStefan Hajnoczi 
1299a62a8ef9SStefan Hajnoczi 	/* Stop forget queue. Soon destroy will be sent */
1300a62a8ef9SStefan Hajnoczi 	spin_lock(&fsvq->lock);
1301a62a8ef9SStefan Hajnoczi 	fsvq->connected = false;
1302a62a8ef9SStefan Hajnoczi 	spin_unlock(&fsvq->lock);
1303a62a8ef9SStefan Hajnoczi 	virtio_fs_drain_all_queues(vfs);
1304a62a8ef9SStefan Hajnoczi 
1305a62a8ef9SStefan Hajnoczi 	fuse_kill_sb_anon(sb);
1306a62a8ef9SStefan Hajnoczi 
1307a62a8ef9SStefan Hajnoczi 	/* fuse_kill_sb_anon() must have sent destroy. Stop all queues
1308a62a8ef9SStefan Hajnoczi 	 * and drain one more time and free fuse devices. Freeing fuse
1309a62a8ef9SStefan Hajnoczi 	 * devices will drop their reference on fuse_conn and that in
1310a62a8ef9SStefan Hajnoczi 	 * turn will drop its reference on virtio_fs object.
1311a62a8ef9SStefan Hajnoczi 	 */
1312a62a8ef9SStefan Hajnoczi 	virtio_fs_stop_all_queues(vfs);
1313a62a8ef9SStefan Hajnoczi 	virtio_fs_drain_all_queues(vfs);
1314a62a8ef9SStefan Hajnoczi 	virtio_fs_free_devs(vfs);
1315a62a8ef9SStefan Hajnoczi }
1316a62a8ef9SStefan Hajnoczi 
1317a62a8ef9SStefan Hajnoczi static int virtio_fs_test_super(struct super_block *sb,
1318a62a8ef9SStefan Hajnoczi 				struct fs_context *fsc)
1319a62a8ef9SStefan Hajnoczi {
1320a62a8ef9SStefan Hajnoczi 	struct fuse_conn *fc = fsc->s_fs_info;
1321a62a8ef9SStefan Hajnoczi 
1322a62a8ef9SStefan Hajnoczi 	return fc->iq.priv == get_fuse_conn_super(sb)->iq.priv;
1323a62a8ef9SStefan Hajnoczi }
1324a62a8ef9SStefan Hajnoczi 
1325a62a8ef9SStefan Hajnoczi static int virtio_fs_set_super(struct super_block *sb,
1326a62a8ef9SStefan Hajnoczi 			       struct fs_context *fsc)
1327a62a8ef9SStefan Hajnoczi {
1328a62a8ef9SStefan Hajnoczi 	int err;
1329a62a8ef9SStefan Hajnoczi 
1330a62a8ef9SStefan Hajnoczi 	err = get_anon_bdev(&sb->s_dev);
1331a62a8ef9SStefan Hajnoczi 	if (!err)
1332a62a8ef9SStefan Hajnoczi 		fuse_conn_get(fsc->s_fs_info);
1333a62a8ef9SStefan Hajnoczi 
1334a62a8ef9SStefan Hajnoczi 	return err;
1335a62a8ef9SStefan Hajnoczi }
1336a62a8ef9SStefan Hajnoczi 
1337a62a8ef9SStefan Hajnoczi static int virtio_fs_get_tree(struct fs_context *fsc)
1338a62a8ef9SStefan Hajnoczi {
1339a62a8ef9SStefan Hajnoczi 	struct virtio_fs *fs;
1340a62a8ef9SStefan Hajnoczi 	struct super_block *sb;
1341a62a8ef9SStefan Hajnoczi 	struct fuse_conn *fc;
1342a62a8ef9SStefan Hajnoczi 	int err;
1343a62a8ef9SStefan Hajnoczi 
1344a62a8ef9SStefan Hajnoczi 	/* This gets a reference on virtio_fs object. This ptr gets installed
1345a62a8ef9SStefan Hajnoczi 	 * in fc->iq->priv. Once fuse_conn is going away, it calls ->put()
1346a62a8ef9SStefan Hajnoczi 	 * to drop the reference to this object.
1347a62a8ef9SStefan Hajnoczi 	 */
1348a62a8ef9SStefan Hajnoczi 	fs = virtio_fs_find_instance(fsc->source);
1349a62a8ef9SStefan Hajnoczi 	if (!fs) {
1350a62a8ef9SStefan Hajnoczi 		pr_info("virtio-fs: tag <%s> not found\n", fsc->source);
1351a62a8ef9SStefan Hajnoczi 		return -EINVAL;
1352a62a8ef9SStefan Hajnoczi 	}
1353a62a8ef9SStefan Hajnoczi 
1354a62a8ef9SStefan Hajnoczi 	fc = kzalloc(sizeof(struct fuse_conn), GFP_KERNEL);
1355a62a8ef9SStefan Hajnoczi 	if (!fc) {
1356a62a8ef9SStefan Hajnoczi 		mutex_lock(&virtio_fs_mutex);
1357a62a8ef9SStefan Hajnoczi 		virtio_fs_put(fs);
1358a62a8ef9SStefan Hajnoczi 		mutex_unlock(&virtio_fs_mutex);
1359a62a8ef9SStefan Hajnoczi 		return -ENOMEM;
1360a62a8ef9SStefan Hajnoczi 	}
1361a62a8ef9SStefan Hajnoczi 
1362a62a8ef9SStefan Hajnoczi 	fuse_conn_init(fc, get_user_ns(current_user_ns()), &virtio_fs_fiq_ops,
1363a62a8ef9SStefan Hajnoczi 		       fs);
1364a62a8ef9SStefan Hajnoczi 	fc->release = fuse_free_conn;
1365a62a8ef9SStefan Hajnoczi 	fc->delete_stale = true;
1366a62a8ef9SStefan Hajnoczi 
1367a62a8ef9SStefan Hajnoczi 	fsc->s_fs_info = fc;
1368a62a8ef9SStefan Hajnoczi 	sb = sget_fc(fsc, virtio_fs_test_super, virtio_fs_set_super);
1369a62a8ef9SStefan Hajnoczi 	fuse_conn_put(fc);
1370a62a8ef9SStefan Hajnoczi 	if (IS_ERR(sb))
1371a62a8ef9SStefan Hajnoczi 		return PTR_ERR(sb);
1372a62a8ef9SStefan Hajnoczi 
1373a62a8ef9SStefan Hajnoczi 	if (!sb->s_root) {
1374a62a8ef9SStefan Hajnoczi 		err = virtio_fs_fill_super(sb);
1375a62a8ef9SStefan Hajnoczi 		if (err) {
1376a62a8ef9SStefan Hajnoczi 			deactivate_locked_super(sb);
1377a62a8ef9SStefan Hajnoczi 			return err;
1378a62a8ef9SStefan Hajnoczi 		}
1379a62a8ef9SStefan Hajnoczi 
1380a62a8ef9SStefan Hajnoczi 		sb->s_flags |= SB_ACTIVE;
1381a62a8ef9SStefan Hajnoczi 	}
1382a62a8ef9SStefan Hajnoczi 
1383a62a8ef9SStefan Hajnoczi 	WARN_ON(fsc->root);
1384a62a8ef9SStefan Hajnoczi 	fsc->root = dget(sb->s_root);
1385a62a8ef9SStefan Hajnoczi 	return 0;
1386a62a8ef9SStefan Hajnoczi }
1387a62a8ef9SStefan Hajnoczi 
1388a62a8ef9SStefan Hajnoczi static const struct fs_context_operations virtio_fs_context_ops = {
1389a62a8ef9SStefan Hajnoczi 	.get_tree	= virtio_fs_get_tree,
1390a62a8ef9SStefan Hajnoczi };
1391a62a8ef9SStefan Hajnoczi 
1392a62a8ef9SStefan Hajnoczi static int virtio_fs_init_fs_context(struct fs_context *fsc)
1393a62a8ef9SStefan Hajnoczi {
1394a62a8ef9SStefan Hajnoczi 	fsc->ops = &virtio_fs_context_ops;
1395a62a8ef9SStefan Hajnoczi 	return 0;
1396a62a8ef9SStefan Hajnoczi }
1397a62a8ef9SStefan Hajnoczi 
1398a62a8ef9SStefan Hajnoczi static struct file_system_type virtio_fs_type = {
1399a62a8ef9SStefan Hajnoczi 	.owner		= THIS_MODULE,
1400a62a8ef9SStefan Hajnoczi 	.name		= "virtiofs",
1401a62a8ef9SStefan Hajnoczi 	.init_fs_context = virtio_fs_init_fs_context,
1402a62a8ef9SStefan Hajnoczi 	.kill_sb	= virtio_kill_sb,
1403a62a8ef9SStefan Hajnoczi };
1404a62a8ef9SStefan Hajnoczi 
1405a62a8ef9SStefan Hajnoczi static int __init virtio_fs_init(void)
1406a62a8ef9SStefan Hajnoczi {
1407a62a8ef9SStefan Hajnoczi 	int ret;
1408a62a8ef9SStefan Hajnoczi 
1409a62a8ef9SStefan Hajnoczi 	ret = register_virtio_driver(&virtio_fs_driver);
1410a62a8ef9SStefan Hajnoczi 	if (ret < 0)
1411a62a8ef9SStefan Hajnoczi 		return ret;
1412a62a8ef9SStefan Hajnoczi 
1413a62a8ef9SStefan Hajnoczi 	ret = register_filesystem(&virtio_fs_type);
1414a62a8ef9SStefan Hajnoczi 	if (ret < 0) {
1415a62a8ef9SStefan Hajnoczi 		unregister_virtio_driver(&virtio_fs_driver);
1416a62a8ef9SStefan Hajnoczi 		return ret;
1417a62a8ef9SStefan Hajnoczi 	}
1418a62a8ef9SStefan Hajnoczi 
1419a62a8ef9SStefan Hajnoczi 	return 0;
1420a62a8ef9SStefan Hajnoczi }
1421a62a8ef9SStefan Hajnoczi module_init(virtio_fs_init);
1422a62a8ef9SStefan Hajnoczi 
1423a62a8ef9SStefan Hajnoczi static void __exit virtio_fs_exit(void)
1424a62a8ef9SStefan Hajnoczi {
1425a62a8ef9SStefan Hajnoczi 	unregister_filesystem(&virtio_fs_type);
1426a62a8ef9SStefan Hajnoczi 	unregister_virtio_driver(&virtio_fs_driver);
1427a62a8ef9SStefan Hajnoczi }
1428a62a8ef9SStefan Hajnoczi module_exit(virtio_fs_exit);
1429a62a8ef9SStefan Hajnoczi 
1430a62a8ef9SStefan Hajnoczi MODULE_AUTHOR("Stefan Hajnoczi <stefanha@redhat.com>");
1431a62a8ef9SStefan Hajnoczi MODULE_DESCRIPTION("Virtio Filesystem");
1432a62a8ef9SStefan Hajnoczi MODULE_LICENSE("GPL");
1433a62a8ef9SStefan Hajnoczi MODULE_ALIAS_FS(KBUILD_MODNAME);
1434a62a8ef9SStefan Hajnoczi MODULE_DEVICE_TABLE(virtio, id_table);
1435