1*1f327613SThomas Gleixner // SPDX-License-Identifier: GPL-2.0-only 2b530cc79SEric Van Hensbergen /* 3fea511a6SEric Van Hensbergen * The Virtio 9p transport driver 4b530cc79SEric Van Hensbergen * 5e2735b77SEric Van Hensbergen * This is a block based transport driver based on the lguest block driver 6e2735b77SEric Van Hensbergen * code. 7b530cc79SEric Van Hensbergen * 8fea511a6SEric Van Hensbergen * Copyright (C) 2007, 2008 Eric Van Hensbergen, IBM Corporation 9b530cc79SEric Van Hensbergen * 10b530cc79SEric Van Hensbergen * Based on virtio console driver 11b530cc79SEric Van Hensbergen * Copyright (C) 2006, 2007 Rusty Russell, IBM Corporation 12b530cc79SEric Van Hensbergen */ 13b530cc79SEric Van Hensbergen 145d385153SJoe Perches #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt 155d385153SJoe Perches 16b530cc79SEric Van Hensbergen #include <linux/in.h> 17b530cc79SEric Van Hensbergen #include <linux/module.h> 18b530cc79SEric Van Hensbergen #include <linux/net.h> 19b530cc79SEric Van Hensbergen #include <linux/ipv6.h> 20b530cc79SEric Van Hensbergen #include <linux/errno.h> 21b530cc79SEric Van Hensbergen #include <linux/kernel.h> 22b530cc79SEric Van Hensbergen #include <linux/un.h> 23b530cc79SEric Van Hensbergen #include <linux/uaccess.h> 24b530cc79SEric Van Hensbergen #include <linux/inet.h> 25b530cc79SEric Van Hensbergen #include <linux/idr.h> 26b530cc79SEric Van Hensbergen #include <linux/file.h> 27b9cdc88dSWill Deacon #include <linux/highmem.h> 285a0e3ad6STejun Heo #include <linux/slab.h> 29b530cc79SEric Van Hensbergen #include <net/9p/9p.h> 30b530cc79SEric Van Hensbergen #include <linux/parser.h> 318b81ef58SEric Van Hensbergen #include <net/9p/client.h> 32b530cc79SEric Van Hensbergen #include <net/9p/transport.h> 33b530cc79SEric Van Hensbergen #include <linux/scatterlist.h> 3468da9ba4SVenkateswararao Jujjuri (JV) #include <linux/swap.h> 35b530cc79SEric Van Hensbergen #include <linux/virtio.h> 36b530cc79SEric Van Hensbergen #include <linux/virtio_9p.h> 374038866dSVenkateswararao Jujjuri (JV) #include "trans_common.h" 38b530cc79SEric Van Hensbergen 39e2735b77SEric Van Hensbergen #define VIRTQUEUE_NUM 128 40e2735b77SEric Van Hensbergen 41b530cc79SEric Van Hensbergen /* a single mutex to manage channel initialization and attachment */ 42c1549497SJosef 'Jeff' Sipek static DEFINE_MUTEX(virtio_9p_lock); 4368da9ba4SVenkateswararao Jujjuri (JV) static DECLARE_WAIT_QUEUE_HEAD(vp_wq); 4468da9ba4SVenkateswararao Jujjuri (JV) static atomic_t vp_pinned = ATOMIC_INIT(0); 45b530cc79SEric Van Hensbergen 46ee443996SEric Van Hensbergen /** 47ee443996SEric Van Hensbergen * struct virtio_chan - per-instance transport information 48ee443996SEric Van Hensbergen * @inuse: whether the channel is in use 49ee443996SEric Van Hensbergen * @lock: protects multiple elements within this structure 500e15597eSAbhishek Kulkarni * @client: client instance 51ee443996SEric Van Hensbergen * @vdev: virtio dev associated with this channel 52ee443996SEric Van Hensbergen * @vq: virtio queue associated with this channel 53ee443996SEric Van Hensbergen * @sg: scatter gather list which is used to pack a request (protected?) 54ee443996SEric Van Hensbergen * 55ee443996SEric Van Hensbergen * We keep all per-channel information in a structure. 56b530cc79SEric Van Hensbergen * This structure is allocated within the devices dev->mem space. 57b530cc79SEric Van Hensbergen * A pointer to the structure will get put in the transport private. 58ee443996SEric Van Hensbergen * 59b530cc79SEric Van Hensbergen */ 60ee443996SEric Van Hensbergen 6137c1209dSAneesh Kumar K.V struct virtio_chan { 62ee443996SEric Van Hensbergen bool inuse; 63b530cc79SEric Van Hensbergen 64e2735b77SEric Van Hensbergen spinlock_t lock; 65e2735b77SEric Van Hensbergen 66fea511a6SEric Van Hensbergen struct p9_client *client; 67b530cc79SEric Van Hensbergen struct virtio_device *vdev; 68e2735b77SEric Van Hensbergen struct virtqueue *vq; 6952f44e0dSVenkateswararao Jujjuri (JV) int ring_bufs_avail; 7052f44e0dSVenkateswararao Jujjuri (JV) wait_queue_head_t *vc_wq; 7168da9ba4SVenkateswararao Jujjuri (JV) /* This is global limit. Since we don't have a global structure, 7268da9ba4SVenkateswararao Jujjuri (JV) * will be placing it in each channel. 7368da9ba4SVenkateswararao Jujjuri (JV) */ 747293bfbaSZhang Yanfei unsigned long p9_max_pages; 75e2735b77SEric Van Hensbergen /* Scatterlist: can be too big for stack. */ 76e2735b77SEric Van Hensbergen struct scatterlist sg[VIRTQUEUE_NUM]; 7797ee9b02SAneesh Kumar K.V /* 78edcd9d97Spiaojun * tag name to identify a mount null terminated 7997ee9b02SAneesh Kumar K.V */ 8097ee9b02SAneesh Kumar K.V char *tag; 8197ee9b02SAneesh Kumar K.V 8237c1209dSAneesh Kumar K.V struct list_head chan_list; 8337c1209dSAneesh Kumar K.V }; 8437c1209dSAneesh Kumar K.V 8537c1209dSAneesh Kumar K.V static struct list_head virtio_chan_list; 86b530cc79SEric Van Hensbergen 87b530cc79SEric Van Hensbergen /* How many bytes left in this page. */ 88b530cc79SEric Van Hensbergen static unsigned int rest_of_page(void *data) 89b530cc79SEric Van Hensbergen { 90222e4adeSAl Viro return PAGE_SIZE - offset_in_page(data); 91b530cc79SEric Van Hensbergen } 92b530cc79SEric Van Hensbergen 93ee443996SEric Van Hensbergen /** 94ee443996SEric Van Hensbergen * p9_virtio_close - reclaim resources of a channel 950e15597eSAbhishek Kulkarni * @client: client instance 96ee443996SEric Van Hensbergen * 97ee443996SEric Van Hensbergen * This reclaims a channel by freeing its resources and 98ee443996SEric Van Hensbergen * reseting its inuse flag. 99ee443996SEric Van Hensbergen * 100ee443996SEric Van Hensbergen */ 101ee443996SEric Van Hensbergen 1028b81ef58SEric Van Hensbergen static void p9_virtio_close(struct p9_client *client) 103b530cc79SEric Van Hensbergen { 1048b81ef58SEric Van Hensbergen struct virtio_chan *chan = client->trans; 105b530cc79SEric Van Hensbergen 106c1549497SJosef 'Jeff' Sipek mutex_lock(&virtio_9p_lock); 107fb786100SAneesh Kumar K.V if (chan) 108b530cc79SEric Van Hensbergen chan->inuse = false; 109c1549497SJosef 'Jeff' Sipek mutex_unlock(&virtio_9p_lock); 110b530cc79SEric Van Hensbergen } 111b530cc79SEric Van Hensbergen 112ee443996SEric Van Hensbergen /** 113ee443996SEric Van Hensbergen * req_done - callback which signals activity from the server 114ee443996SEric Van Hensbergen * @vq: virtio queue activity was received on 115ee443996SEric Van Hensbergen * 116ee443996SEric Van Hensbergen * This notifies us that the server has triggered some activity 117ee443996SEric Van Hensbergen * on the virtio channel - most likely a response to request we 118ee443996SEric Van Hensbergen * sent. Figure out which requests now have responses and wake up 119ee443996SEric Van Hensbergen * those threads. 120ee443996SEric Van Hensbergen * 121ee443996SEric Van Hensbergen * Bugs: could do with some additional sanity checking, but appears to work. 122ee443996SEric Van Hensbergen * 123ee443996SEric Van Hensbergen */ 124ee443996SEric Van Hensbergen 125e2735b77SEric Van Hensbergen static void req_done(struct virtqueue *vq) 126b530cc79SEric Van Hensbergen { 127e2735b77SEric Van Hensbergen struct virtio_chan *chan = vq->vdev->priv; 128e2735b77SEric Van Hensbergen unsigned int len; 129e2735b77SEric Van Hensbergen struct p9_req_t *req; 13031934da8Sjiangyiwen bool need_wakeup = false; 131419b3956SVenkateswararao Jujjuri (JV) unsigned long flags; 132b530cc79SEric Van Hensbergen 1335d385153SJoe Perches p9_debug(P9_DEBUG_TRANS, ": request done\n"); 13491b8534fSEric Van Hensbergen 135419b3956SVenkateswararao Jujjuri (JV) spin_lock_irqsave(&chan->lock, flags); 13631934da8Sjiangyiwen while ((req = virtqueue_get_buf(chan->vq, &len)) != NULL) { 13731934da8Sjiangyiwen if (!chan->ring_bufs_avail) { 13852f44e0dSVenkateswararao Jujjuri (JV) chan->ring_bufs_avail = 1; 13931934da8Sjiangyiwen need_wakeup = true; 14031934da8Sjiangyiwen } 14131934da8Sjiangyiwen 142f984579aSTomas Bortoli if (len) { 143523adb6cSDominique Martinet req->rc.size = len; 1442b6e72edSDominique Martinet p9_client_cb(chan->client, req, REQ_STATUS_RCVD); 145e2735b77SEric Van Hensbergen } 146f984579aSTomas Bortoli } 14731934da8Sjiangyiwen spin_unlock_irqrestore(&chan->lock, flags); 14831934da8Sjiangyiwen /* Wakeup if anyone waiting for VirtIO ring space. */ 14931934da8Sjiangyiwen if (need_wakeup) 15031934da8Sjiangyiwen wake_up(chan->vc_wq); 151b530cc79SEric Van Hensbergen } 152b530cc79SEric Van Hensbergen 153ee443996SEric Van Hensbergen /** 154ee443996SEric Van Hensbergen * pack_sg_list - pack a scatter gather list from a linear buffer 155ee443996SEric Van Hensbergen * @sg: scatter/gather list to pack into 156ee443996SEric Van Hensbergen * @start: which segment of the sg_list to start at 157ee443996SEric Van Hensbergen * @limit: maximum segment to pack data to 158ee443996SEric Van Hensbergen * @data: data to pack into scatter/gather list 159ee443996SEric Van Hensbergen * @count: amount of data to pack into the scatter/gather list 160ee443996SEric Van Hensbergen * 161ee443996SEric Van Hensbergen * sg_lists have multiple segments of various sizes. This will pack 162ee443996SEric Van Hensbergen * arbitrary data into an existing scatter gather list, segmenting the 163ee443996SEric Van Hensbergen * data as necessary within constraints. 164ee443996SEric Van Hensbergen * 165ee443996SEric Van Hensbergen */ 166ee443996SEric Van Hensbergen 167abfa034eSAneesh Kumar K.V static int pack_sg_list(struct scatterlist *sg, int start, 168abfa034eSAneesh Kumar K.V int limit, char *data, int count) 169e2735b77SEric Van Hensbergen { 170e2735b77SEric Van Hensbergen int s; 171e2735b77SEric Van Hensbergen int index = start; 172e2735b77SEric Van Hensbergen 173e2735b77SEric Van Hensbergen while (count) { 174e2735b77SEric Van Hensbergen s = rest_of_page(data); 175e2735b77SEric Van Hensbergen if (s > count) 176e2735b77SEric Van Hensbergen s = count; 17723cba9cbSjiangyiwen BUG_ON(index >= limit); 1780b36f1adSRusty Russell /* Make sure we don't terminate early. */ 1790b36f1adSRusty Russell sg_unmark_end(&sg[index]); 180e2735b77SEric Van Hensbergen sg_set_buf(&sg[index++], data, s); 181e2735b77SEric Van Hensbergen count -= s; 182e2735b77SEric Van Hensbergen data += s; 183e2735b77SEric Van Hensbergen } 1840b36f1adSRusty Russell if (index-start) 1850b36f1adSRusty Russell sg_mark_end(&sg[index - 1]); 186e2735b77SEric Van Hensbergen return index-start; 187e2735b77SEric Van Hensbergen } 188e2735b77SEric Van Hensbergen 18991b8534fSEric Van Hensbergen /* We don't currently allow canceling of virtio requests */ 19091b8534fSEric Van Hensbergen static int p9_virtio_cancel(struct p9_client *client, struct p9_req_t *req) 19191b8534fSEric Van Hensbergen { 19291b8534fSEric Van Hensbergen return 1; 19391b8534fSEric Van Hensbergen } 19491b8534fSEric Van Hensbergen 195728356deSTomas Bortoli /* Reply won't come, so drop req ref */ 196728356deSTomas Bortoli static int p9_virtio_cancelled(struct p9_client *client, struct p9_req_t *req) 197728356deSTomas Bortoli { 198728356deSTomas Bortoli p9_req_put(req); 199728356deSTomas Bortoli return 0; 200728356deSTomas Bortoli } 201728356deSTomas Bortoli 202ee443996SEric Van Hensbergen /** 2034038866dSVenkateswararao Jujjuri (JV) * pack_sg_list_p - Just like pack_sg_list. Instead of taking a buffer, 2044038866dSVenkateswararao Jujjuri (JV) * this takes a list of pages. 2054038866dSVenkateswararao Jujjuri (JV) * @sg: scatter/gather list to pack into 2064038866dSVenkateswararao Jujjuri (JV) * @start: which segment of the sg_list to start at 2072c53040fSBen Hutchings * @pdata: a list of pages to add into sg. 208abfa034eSAneesh Kumar K.V * @nr_pages: number of pages to pack into the scatter/gather list 2094f3b35c1SAl Viro * @offs: amount of data in the beginning of first page _not_ to pack 2104038866dSVenkateswararao Jujjuri (JV) * @count: amount of data to pack into the scatter/gather list 2114038866dSVenkateswararao Jujjuri (JV) */ 2124038866dSVenkateswararao Jujjuri (JV) static int 213abfa034eSAneesh Kumar K.V pack_sg_list_p(struct scatterlist *sg, int start, int limit, 2144f3b35c1SAl Viro struct page **pdata, int nr_pages, size_t offs, int count) 2154038866dSVenkateswararao Jujjuri (JV) { 216abfa034eSAneesh Kumar K.V int i = 0, s; 2174f3b35c1SAl Viro int data_off = offs; 2184038866dSVenkateswararao Jujjuri (JV) int index = start; 2194038866dSVenkateswararao Jujjuri (JV) 220abfa034eSAneesh Kumar K.V BUG_ON(nr_pages > (limit - start)); 221abfa034eSAneesh Kumar K.V /* 222abfa034eSAneesh Kumar K.V * if the first page doesn't start at 223abfa034eSAneesh Kumar K.V * page boundary find the offset 224abfa034eSAneesh Kumar K.V */ 225abfa034eSAneesh Kumar K.V while (nr_pages) { 2264f3b35c1SAl Viro s = PAGE_SIZE - data_off; 227abfa034eSAneesh Kumar K.V if (s > count) 228abfa034eSAneesh Kumar K.V s = count; 22923cba9cbSjiangyiwen BUG_ON(index >= limit); 2300b36f1adSRusty Russell /* Make sure we don't terminate early. */ 2310b36f1adSRusty Russell sg_unmark_end(&sg[index]); 232abfa034eSAneesh Kumar K.V sg_set_page(&sg[index++], pdata[i++], s, data_off); 233abfa034eSAneesh Kumar K.V data_off = 0; 2344038866dSVenkateswararao Jujjuri (JV) count -= s; 235abfa034eSAneesh Kumar K.V nr_pages--; 2364038866dSVenkateswararao Jujjuri (JV) } 2370b36f1adSRusty Russell 2380b36f1adSRusty Russell if (index-start) 2390b36f1adSRusty Russell sg_mark_end(&sg[index - 1]); 2404038866dSVenkateswararao Jujjuri (JV) return index - start; 2414038866dSVenkateswararao Jujjuri (JV) } 2424038866dSVenkateswararao Jujjuri (JV) 2434038866dSVenkateswararao Jujjuri (JV) /** 24491b8534fSEric Van Hensbergen * p9_virtio_request - issue a request 2450e15597eSAbhishek Kulkarni * @client: client instance issuing the request 2460e15597eSAbhishek Kulkarni * @req: request to be issued 247ee443996SEric Van Hensbergen * 248ee443996SEric Van Hensbergen */ 249ee443996SEric Van Hensbergen 250e2735b77SEric Van Hensbergen static int 25191b8534fSEric Van Hensbergen p9_virtio_request(struct p9_client *client, struct p9_req_t *req) 252e2735b77SEric Van Hensbergen { 253abfa034eSAneesh Kumar K.V int err; 2540b36f1adSRusty Russell int in, out, out_sgs, in_sgs; 255419b3956SVenkateswararao Jujjuri (JV) unsigned long flags; 256abfa034eSAneesh Kumar K.V struct virtio_chan *chan = client->trans; 2570b36f1adSRusty Russell struct scatterlist *sgs[2]; 258e2735b77SEric Van Hensbergen 2595d385153SJoe Perches p9_debug(P9_DEBUG_TRANS, "9p debug: virtio request\n"); 260e2735b77SEric Van Hensbergen 261419b3956SVenkateswararao Jujjuri (JV) req->status = REQ_STATUS_SENT; 262abfa034eSAneesh Kumar K.V req_retry: 263419b3956SVenkateswararao Jujjuri (JV) spin_lock_irqsave(&chan->lock, flags); 2644038866dSVenkateswararao Jujjuri (JV) 2650b36f1adSRusty Russell out_sgs = in_sgs = 0; 2664038866dSVenkateswararao Jujjuri (JV) /* Handle out VirtIO ring buffers */ 267abfa034eSAneesh Kumar K.V out = pack_sg_list(chan->sg, 0, 268523adb6cSDominique Martinet VIRTQUEUE_NUM, req->tc.sdata, req->tc.size); 2690b36f1adSRusty Russell if (out) 2700b36f1adSRusty Russell sgs[out_sgs++] = chan->sg; 2714038866dSVenkateswararao Jujjuri (JV) 272abfa034eSAneesh Kumar K.V in = pack_sg_list(chan->sg, out, 273523adb6cSDominique Martinet VIRTQUEUE_NUM, req->rc.sdata, req->rc.capacity); 2740b36f1adSRusty Russell if (in) 2750b36f1adSRusty Russell sgs[out_sgs + in_sgs++] = chan->sg + out; 276e2735b77SEric Van Hensbergen 277474fe9f7SAl Viro err = virtqueue_add_sgs(chan->vq, sgs, out_sgs, in_sgs, req, 278f96fde41SRusty Russell GFP_ATOMIC); 279419b3956SVenkateswararao Jujjuri (JV) if (err < 0) { 28052f44e0dSVenkateswararao Jujjuri (JV) if (err == -ENOSPC) { 28152f44e0dSVenkateswararao Jujjuri (JV) chan->ring_bufs_avail = 0; 28252f44e0dSVenkateswararao Jujjuri (JV) spin_unlock_irqrestore(&chan->lock, flags); 2839523feacSTuomas Tynkkynen err = wait_event_killable(*chan->vc_wq, 28452f44e0dSVenkateswararao Jujjuri (JV) chan->ring_bufs_avail); 28552f44e0dSVenkateswararao Jujjuri (JV) if (err == -ERESTARTSYS) 28652f44e0dSVenkateswararao Jujjuri (JV) return err; 28752f44e0dSVenkateswararao Jujjuri (JV) 2885d385153SJoe Perches p9_debug(P9_DEBUG_TRANS, "Retry virtio request\n"); 289abfa034eSAneesh Kumar K.V goto req_retry; 290abfa034eSAneesh Kumar K.V } else { 291abfa034eSAneesh Kumar K.V spin_unlock_irqrestore(&chan->lock, flags); 2925d385153SJoe Perches p9_debug(P9_DEBUG_TRANS, 2930b36f1adSRusty Russell "virtio rpc add_sgs returned failure\n"); 294abfa034eSAneesh Kumar K.V return -EIO; 295abfa034eSAneesh Kumar K.V } 296abfa034eSAneesh Kumar K.V } 297abfa034eSAneesh Kumar K.V virtqueue_kick(chan->vq); 298abfa034eSAneesh Kumar K.V spin_unlock_irqrestore(&chan->lock, flags); 299abfa034eSAneesh Kumar K.V 3005d385153SJoe Perches p9_debug(P9_DEBUG_TRANS, "virtio request kicked\n"); 301abfa034eSAneesh Kumar K.V return 0; 302abfa034eSAneesh Kumar K.V } 303abfa034eSAneesh Kumar K.V 304abfa034eSAneesh Kumar K.V static int p9_get_mapped_pages(struct virtio_chan *chan, 3054f3b35c1SAl Viro struct page ***pages, 3064f3b35c1SAl Viro struct iov_iter *data, 3074f3b35c1SAl Viro int count, 3084f3b35c1SAl Viro size_t *offs, 3094f3b35c1SAl Viro int *need_drop) 310abfa034eSAneesh Kumar K.V { 3114f3b35c1SAl Viro int nr_pages; 312abfa034eSAneesh Kumar K.V int err; 3134f3b35c1SAl Viro 3144f3b35c1SAl Viro if (!iov_iter_count(data)) 3154f3b35c1SAl Viro return 0; 3164f3b35c1SAl Viro 3172cbfdf4dSMarc Zyngier if (!iov_iter_is_kvec(data)) { 3184f3b35c1SAl Viro int n; 319abfa034eSAneesh Kumar K.V /* 320abfa034eSAneesh Kumar K.V * We allow only p9_max_pages pinned. We wait for the 321abfa034eSAneesh Kumar K.V * Other zc request to finish here 322abfa034eSAneesh Kumar K.V */ 323abfa034eSAneesh Kumar K.V if (atomic_read(&vp_pinned) >= chan->p9_max_pages) { 3249523feacSTuomas Tynkkynen err = wait_event_killable(vp_wq, 325abfa034eSAneesh Kumar K.V (atomic_read(&vp_pinned) < chan->p9_max_pages)); 326abfa034eSAneesh Kumar K.V if (err == -ERESTARTSYS) 327abfa034eSAneesh Kumar K.V return err; 328abfa034eSAneesh Kumar K.V } 3294f3b35c1SAl Viro n = iov_iter_get_pages_alloc(data, pages, count, offs); 3304f3b35c1SAl Viro if (n < 0) 3314f3b35c1SAl Viro return n; 3324f3b35c1SAl Viro *need_drop = 1; 3334f3b35c1SAl Viro nr_pages = DIV_ROUND_UP(n + *offs, PAGE_SIZE); 334abfa034eSAneesh Kumar K.V atomic_add(nr_pages, &vp_pinned); 3354f3b35c1SAl Viro return n; 336abfa034eSAneesh Kumar K.V } else { 337abfa034eSAneesh Kumar K.V /* kernel buffer, no need to pin pages */ 3384f3b35c1SAl Viro int index; 3394f3b35c1SAl Viro size_t len; 3404f3b35c1SAl Viro void *p; 3414f3b35c1SAl Viro 3424f3b35c1SAl Viro /* we'd already checked that it's non-empty */ 3434f3b35c1SAl Viro while (1) { 3444f3b35c1SAl Viro len = iov_iter_single_seg_count(data); 3454f3b35c1SAl Viro if (likely(len)) { 3464f3b35c1SAl Viro p = data->kvec->iov_base + data->iov_offset; 3474f3b35c1SAl Viro break; 3484f3b35c1SAl Viro } 3494f3b35c1SAl Viro iov_iter_advance(data, 0); 3504f3b35c1SAl Viro } 3514f3b35c1SAl Viro if (len > count) 3524f3b35c1SAl Viro len = count; 3534f3b35c1SAl Viro 3544f3b35c1SAl Viro nr_pages = DIV_ROUND_UP((unsigned long)p + len, PAGE_SIZE) - 3554f3b35c1SAl Viro (unsigned long)p / PAGE_SIZE; 3564f3b35c1SAl Viro 3576da2ec56SKees Cook *pages = kmalloc_array(nr_pages, sizeof(struct page *), 3586da2ec56SKees Cook GFP_NOFS); 3594f3b35c1SAl Viro if (!*pages) 3604f3b35c1SAl Viro return -ENOMEM; 3614f3b35c1SAl Viro 3624f3b35c1SAl Viro *need_drop = 0; 363222e4adeSAl Viro p -= (*offs = offset_in_page(p)); 3644f3b35c1SAl Viro for (index = 0; index < nr_pages; index++) { 3654f3b35c1SAl Viro if (is_vmalloc_addr(p)) 3664f3b35c1SAl Viro (*pages)[index] = vmalloc_to_page(p); 367b6f52ae2SRichard Yao else 3684f3b35c1SAl Viro (*pages)[index] = kmap_to_page(p); 3694f3b35c1SAl Viro p += PAGE_SIZE; 370abfa034eSAneesh Kumar K.V } 3714f3b35c1SAl Viro return len; 372abfa034eSAneesh Kumar K.V } 373abfa034eSAneesh Kumar K.V } 374abfa034eSAneesh Kumar K.V 375abfa034eSAneesh Kumar K.V /** 376abfa034eSAneesh Kumar K.V * p9_virtio_zc_request - issue a zero copy request 377abfa034eSAneesh Kumar K.V * @client: client instance issuing the request 378abfa034eSAneesh Kumar K.V * @req: request to be issued 379c7ebbae7Spiaojun * @uidata: user buffer that should be used for zero copy read 380c7ebbae7Spiaojun * @uodata: user buffer that should be used for zero copy write 381abfa034eSAneesh Kumar K.V * @inlen: read buffer size 3824a026da9SSun Lianwen * @outlen: write buffer size 3834a026da9SSun Lianwen * @in_hdr_len: reader header size, This is the size of response protocol data 384abfa034eSAneesh Kumar K.V * 385abfa034eSAneesh Kumar K.V */ 386abfa034eSAneesh Kumar K.V static int 387abfa034eSAneesh Kumar K.V p9_virtio_zc_request(struct p9_client *client, struct p9_req_t *req, 3884f3b35c1SAl Viro struct iov_iter *uidata, struct iov_iter *uodata, 3894f3b35c1SAl Viro int inlen, int outlen, int in_hdr_len) 390abfa034eSAneesh Kumar K.V { 3910b36f1adSRusty Russell int in, out, err, out_sgs, in_sgs; 392abfa034eSAneesh Kumar K.V unsigned long flags; 393abfa034eSAneesh Kumar K.V int in_nr_pages = 0, out_nr_pages = 0; 394abfa034eSAneesh Kumar K.V struct page **in_pages = NULL, **out_pages = NULL; 395abfa034eSAneesh Kumar K.V struct virtio_chan *chan = client->trans; 3960b36f1adSRusty Russell struct scatterlist *sgs[4]; 3974f3b35c1SAl Viro size_t offs; 3984f3b35c1SAl Viro int need_drop = 0; 399728356deSTomas Bortoli int kicked = 0; 400abfa034eSAneesh Kumar K.V 4015d385153SJoe Perches p9_debug(P9_DEBUG_TRANS, "virtio request\n"); 402abfa034eSAneesh Kumar K.V 403abfa034eSAneesh Kumar K.V if (uodata) { 404d28c756cSChirantan Ekbote __le32 sz; 4054f3b35c1SAl Viro int n = p9_get_mapped_pages(chan, &out_pages, uodata, 4064f3b35c1SAl Viro outlen, &offs, &need_drop); 407728356deSTomas Bortoli if (n < 0) { 408728356deSTomas Bortoli err = n; 409728356deSTomas Bortoli goto err_out; 410728356deSTomas Bortoli } 4114f3b35c1SAl Viro out_nr_pages = DIV_ROUND_UP(n + offs, PAGE_SIZE); 4124f3b35c1SAl Viro if (n != outlen) { 4134f3b35c1SAl Viro __le32 v = cpu_to_le32(n); 414523adb6cSDominique Martinet memcpy(&req->tc.sdata[req->tc.size - 4], &v, 4); 4154f3b35c1SAl Viro outlen = n; 416abfa034eSAneesh Kumar K.V } 417d28c756cSChirantan Ekbote /* The size field of the message must include the length of the 418d28c756cSChirantan Ekbote * header and the length of the data. We didn't actually know 419d28c756cSChirantan Ekbote * the length of the data until this point so add it in now. 420d28c756cSChirantan Ekbote */ 421523adb6cSDominique Martinet sz = cpu_to_le32(req->tc.size + outlen); 422523adb6cSDominique Martinet memcpy(&req->tc.sdata[0], &sz, sizeof(sz)); 4234f3b35c1SAl Viro } else if (uidata) { 4244f3b35c1SAl Viro int n = p9_get_mapped_pages(chan, &in_pages, uidata, 4254f3b35c1SAl Viro inlen, &offs, &need_drop); 426728356deSTomas Bortoli if (n < 0) { 427728356deSTomas Bortoli err = n; 428728356deSTomas Bortoli goto err_out; 429728356deSTomas Bortoli } 4304f3b35c1SAl Viro in_nr_pages = DIV_ROUND_UP(n + offs, PAGE_SIZE); 4314f3b35c1SAl Viro if (n != inlen) { 4324f3b35c1SAl Viro __le32 v = cpu_to_le32(n); 433523adb6cSDominique Martinet memcpy(&req->tc.sdata[req->tc.size - 4], &v, 4); 4344f3b35c1SAl Viro inlen = n; 435abfa034eSAneesh Kumar K.V } 436abfa034eSAneesh Kumar K.V } 437abfa034eSAneesh Kumar K.V req->status = REQ_STATUS_SENT; 438abfa034eSAneesh Kumar K.V req_retry_pinned: 439abfa034eSAneesh Kumar K.V spin_lock_irqsave(&chan->lock, flags); 4400b36f1adSRusty Russell 4410b36f1adSRusty Russell out_sgs = in_sgs = 0; 4420b36f1adSRusty Russell 443abfa034eSAneesh Kumar K.V /* out data */ 444abfa034eSAneesh Kumar K.V out = pack_sg_list(chan->sg, 0, 445523adb6cSDominique Martinet VIRTQUEUE_NUM, req->tc.sdata, req->tc.size); 446abfa034eSAneesh Kumar K.V 4470b36f1adSRusty Russell if (out) 4480b36f1adSRusty Russell sgs[out_sgs++] = chan->sg; 4490b36f1adSRusty Russell 4500b36f1adSRusty Russell if (out_pages) { 4510b36f1adSRusty Russell sgs[out_sgs++] = chan->sg + out; 452abfa034eSAneesh Kumar K.V out += pack_sg_list_p(chan->sg, out, VIRTQUEUE_NUM, 4534f3b35c1SAl Viro out_pages, out_nr_pages, offs, outlen); 4540b36f1adSRusty Russell } 4550b36f1adSRusty Russell 456abfa034eSAneesh Kumar K.V /* 457abfa034eSAneesh Kumar K.V * Take care of in data 458abfa034eSAneesh Kumar K.V * For example TREAD have 11. 459abfa034eSAneesh Kumar K.V * 11 is the read/write header = PDU Header(7) + IO Size (4). 460abfa034eSAneesh Kumar K.V * Arrange in such a way that server places header in the 461abfa034eSAneesh Kumar K.V * alloced memory and payload onto the user buffer. 462abfa034eSAneesh Kumar K.V */ 463abfa034eSAneesh Kumar K.V in = pack_sg_list(chan->sg, out, 464523adb6cSDominique Martinet VIRTQUEUE_NUM, req->rc.sdata, in_hdr_len); 4650b36f1adSRusty Russell if (in) 4660b36f1adSRusty Russell sgs[out_sgs + in_sgs++] = chan->sg + out; 4670b36f1adSRusty Russell 4680b36f1adSRusty Russell if (in_pages) { 4690b36f1adSRusty Russell sgs[out_sgs + in_sgs++] = chan->sg + out + in; 470abfa034eSAneesh Kumar K.V in += pack_sg_list_p(chan->sg, out + in, VIRTQUEUE_NUM, 4714f3b35c1SAl Viro in_pages, in_nr_pages, offs, inlen); 4720b36f1adSRusty Russell } 473abfa034eSAneesh Kumar K.V 4740b36f1adSRusty Russell BUG_ON(out_sgs + in_sgs > ARRAY_SIZE(sgs)); 475474fe9f7SAl Viro err = virtqueue_add_sgs(chan->vq, sgs, out_sgs, in_sgs, req, 476f96fde41SRusty Russell GFP_ATOMIC); 477abfa034eSAneesh Kumar K.V if (err < 0) { 478abfa034eSAneesh Kumar K.V if (err == -ENOSPC) { 479abfa034eSAneesh Kumar K.V chan->ring_bufs_avail = 0; 480abfa034eSAneesh Kumar K.V spin_unlock_irqrestore(&chan->lock, flags); 4819523feacSTuomas Tynkkynen err = wait_event_killable(*chan->vc_wq, 482abfa034eSAneesh Kumar K.V chan->ring_bufs_avail); 483abfa034eSAneesh Kumar K.V if (err == -ERESTARTSYS) 484abfa034eSAneesh Kumar K.V goto err_out; 485abfa034eSAneesh Kumar K.V 4865d385153SJoe Perches p9_debug(P9_DEBUG_TRANS, "Retry virtio request\n"); 487316ad550SVenkateswararao Jujjuri (JV) goto req_retry_pinned; 48852f44e0dSVenkateswararao Jujjuri (JV) } else { 489419b3956SVenkateswararao Jujjuri (JV) spin_unlock_irqrestore(&chan->lock, flags); 4905d385153SJoe Perches p9_debug(P9_DEBUG_TRANS, 4910b36f1adSRusty Russell "virtio rpc add_sgs returned failure\n"); 492abfa034eSAneesh Kumar K.V err = -EIO; 493abfa034eSAneesh Kumar K.V goto err_out; 494e2735b77SEric Van Hensbergen } 49552f44e0dSVenkateswararao Jujjuri (JV) } 496dc3f5e68SMichael S. Tsirkin virtqueue_kick(chan->vq); 497419b3956SVenkateswararao Jujjuri (JV) spin_unlock_irqrestore(&chan->lock, flags); 498728356deSTomas Bortoli kicked = 1; 4995d385153SJoe Perches p9_debug(P9_DEBUG_TRANS, "virtio request kicked\n"); 5002557d0c5SMatthew Wilcox err = wait_event_killable(req->wq, req->status >= REQ_STATUS_RCVD); 501abfa034eSAneesh Kumar K.V /* 502abfa034eSAneesh Kumar K.V * Non kernel buffers are pinned, unpin them 503abfa034eSAneesh Kumar K.V */ 504abfa034eSAneesh Kumar K.V err_out: 5054f3b35c1SAl Viro if (need_drop) { 506abfa034eSAneesh Kumar K.V if (in_pages) { 507abfa034eSAneesh Kumar K.V p9_release_pages(in_pages, in_nr_pages); 508abfa034eSAneesh Kumar K.V atomic_sub(in_nr_pages, &vp_pinned); 509abfa034eSAneesh Kumar K.V } 510abfa034eSAneesh Kumar K.V if (out_pages) { 511abfa034eSAneesh Kumar K.V p9_release_pages(out_pages, out_nr_pages); 512abfa034eSAneesh Kumar K.V atomic_sub(out_nr_pages, &vp_pinned); 513abfa034eSAneesh Kumar K.V } 514abfa034eSAneesh Kumar K.V /* wakeup anybody waiting for slots to pin pages */ 515abfa034eSAneesh Kumar K.V wake_up(&vp_wq); 516abfa034eSAneesh Kumar K.V } 5171b8553c0SVegard Nossum kvfree(in_pages); 5181b8553c0SVegard Nossum kvfree(out_pages); 519728356deSTomas Bortoli if (!kicked) { 520728356deSTomas Bortoli /* reply won't come */ 521728356deSTomas Bortoli p9_req_put(req); 522728356deSTomas Bortoli } 523abfa034eSAneesh Kumar K.V return err; 524e2735b77SEric Van Hensbergen } 525e2735b77SEric Van Hensbergen 52686c84373SAneesh Kumar K.V static ssize_t p9_mount_tag_show(struct device *dev, 52786c84373SAneesh Kumar K.V struct device_attribute *attr, char *buf) 52886c84373SAneesh Kumar K.V { 52986c84373SAneesh Kumar K.V struct virtio_chan *chan; 53086c84373SAneesh Kumar K.V struct virtio_device *vdev; 531edcd9d97Spiaojun int tag_len; 53286c84373SAneesh Kumar K.V 53386c84373SAneesh Kumar K.V vdev = dev_to_virtio(dev); 53486c84373SAneesh Kumar K.V chan = vdev->priv; 535edcd9d97Spiaojun tag_len = strlen(chan->tag); 53686c84373SAneesh Kumar K.V 537edcd9d97Spiaojun memcpy(buf, chan->tag, tag_len + 1); 538179a5bc4SAndrey Ryabinin 539edcd9d97Spiaojun return tag_len + 1; 54086c84373SAneesh Kumar K.V } 54186c84373SAneesh Kumar K.V 54286c84373SAneesh Kumar K.V static DEVICE_ATTR(mount_tag, 0444, p9_mount_tag_show, NULL); 54386c84373SAneesh Kumar K.V 544ee443996SEric Van Hensbergen /** 545ee443996SEric Van Hensbergen * p9_virtio_probe - probe for existence of 9P virtio channels 546ee443996SEric Van Hensbergen * @vdev: virtio device to probe 547ee443996SEric Van Hensbergen * 54837c1209dSAneesh Kumar K.V * This probes for existing virtio channels. 549ee443996SEric Van Hensbergen * 550ee443996SEric Van Hensbergen */ 551ee443996SEric Van Hensbergen 552e2735b77SEric Van Hensbergen static int p9_virtio_probe(struct virtio_device *vdev) 553b530cc79SEric Van Hensbergen { 55497ee9b02SAneesh Kumar K.V __u16 tag_len; 55597ee9b02SAneesh Kumar K.V char *tag; 556b530cc79SEric Van Hensbergen int err; 557b530cc79SEric Van Hensbergen struct virtio_chan *chan; 558b530cc79SEric Van Hensbergen 5597754f53eSMichael S. Tsirkin if (!vdev->config->get) { 5607754f53eSMichael S. Tsirkin dev_err(&vdev->dev, "%s failure: config access disabled\n", 5617754f53eSMichael S. Tsirkin __func__); 5627754f53eSMichael S. Tsirkin return -EINVAL; 5637754f53eSMichael S. Tsirkin } 5647754f53eSMichael S. Tsirkin 56537c1209dSAneesh Kumar K.V chan = kmalloc(sizeof(struct virtio_chan), GFP_KERNEL); 56637c1209dSAneesh Kumar K.V if (!chan) { 5675d385153SJoe Perches pr_err("Failed to allocate virtio 9P channel\n"); 568b530cc79SEric Van Hensbergen err = -ENOMEM; 569b530cc79SEric Van Hensbergen goto fail; 570b530cc79SEric Van Hensbergen } 571b530cc79SEric Van Hensbergen 572e2735b77SEric Van Hensbergen chan->vdev = vdev; 573e2735b77SEric Van Hensbergen 574e2735b77SEric Van Hensbergen /* We expect one virtqueue, for requests. */ 575d2a7dddaSMichael S. Tsirkin chan->vq = virtio_find_single_vq(vdev, req_done, "requests"); 576e2735b77SEric Van Hensbergen if (IS_ERR(chan->vq)) { 577e2735b77SEric Van Hensbergen err = PTR_ERR(chan->vq); 57892aef467SJean-Philippe Brucker goto out_free_chan; 579b530cc79SEric Van Hensbergen } 580e2735b77SEric Van Hensbergen chan->vq->vdev->priv = chan; 581e2735b77SEric Van Hensbergen spin_lock_init(&chan->lock); 582b530cc79SEric Van Hensbergen 583e2735b77SEric Van Hensbergen sg_init_table(chan->sg, VIRTQUEUE_NUM); 584b530cc79SEric Van Hensbergen 585b530cc79SEric Van Hensbergen chan->inuse = false; 58697ee9b02SAneesh Kumar K.V if (virtio_has_feature(vdev, VIRTIO_9P_MOUNT_TAG)) { 587855e0c52SRusty Russell virtio_cread(vdev, struct virtio_9p_config, tag_len, &tag_len); 58897ee9b02SAneesh Kumar K.V } else { 58997ee9b02SAneesh Kumar K.V err = -EINVAL; 59097ee9b02SAneesh Kumar K.V goto out_free_vq; 59197ee9b02SAneesh Kumar K.V } 592edcd9d97Spiaojun tag = kzalloc(tag_len + 1, GFP_KERNEL); 59397ee9b02SAneesh Kumar K.V if (!tag) { 59497ee9b02SAneesh Kumar K.V err = -ENOMEM; 59597ee9b02SAneesh Kumar K.V goto out_free_vq; 59697ee9b02SAneesh Kumar K.V } 597855e0c52SRusty Russell 598855e0c52SRusty Russell virtio_cread_bytes(vdev, offsetof(struct virtio_9p_config, tag), 59997ee9b02SAneesh Kumar K.V tag, tag_len); 60097ee9b02SAneesh Kumar K.V chan->tag = tag; 60186c84373SAneesh Kumar K.V err = sysfs_create_file(&(vdev->dev.kobj), &dev_attr_mount_tag.attr); 60286c84373SAneesh Kumar K.V if (err) { 60352f44e0dSVenkateswararao Jujjuri (JV) goto out_free_tag; 60486c84373SAneesh Kumar K.V } 60552f44e0dSVenkateswararao Jujjuri (JV) chan->vc_wq = kmalloc(sizeof(wait_queue_head_t), GFP_KERNEL); 60652f44e0dSVenkateswararao Jujjuri (JV) if (!chan->vc_wq) { 60752f44e0dSVenkateswararao Jujjuri (JV) err = -ENOMEM; 60852f44e0dSVenkateswararao Jujjuri (JV) goto out_free_tag; 60952f44e0dSVenkateswararao Jujjuri (JV) } 61052f44e0dSVenkateswararao Jujjuri (JV) init_waitqueue_head(chan->vc_wq); 61152f44e0dSVenkateswararao Jujjuri (JV) chan->ring_bufs_avail = 1; 61268da9ba4SVenkateswararao Jujjuri (JV) /* Ceiling limit to avoid denial of service attacks */ 61368da9ba4SVenkateswararao Jujjuri (JV) chan->p9_max_pages = nr_free_buffer_pages()/4; 61452f44e0dSVenkateswararao Jujjuri (JV) 61564b4cc39SMichael S. Tsirkin virtio_device_ready(vdev); 61664b4cc39SMichael S. Tsirkin 61737c1209dSAneesh Kumar K.V mutex_lock(&virtio_9p_lock); 61837c1209dSAneesh Kumar K.V list_add_tail(&chan->chan_list, &virtio_chan_list); 61937c1209dSAneesh Kumar K.V mutex_unlock(&virtio_9p_lock); 620e0d6cb9cSMichael Marineau 621e0d6cb9cSMichael Marineau /* Let udev rules use the new mount_tag attribute. */ 622e0d6cb9cSMichael Marineau kobject_uevent(&(vdev->dev.kobj), KOBJ_CHANGE); 623e0d6cb9cSMichael Marineau 624b530cc79SEric Van Hensbergen return 0; 625b530cc79SEric Van Hensbergen 62652f44e0dSVenkateswararao Jujjuri (JV) out_free_tag: 62752f44e0dSVenkateswararao Jujjuri (JV) kfree(tag); 628e2735b77SEric Van Hensbergen out_free_vq: 629d2a7dddaSMichael S. Tsirkin vdev->config->del_vqs(vdev); 63092aef467SJean-Philippe Brucker out_free_chan: 63137c1209dSAneesh Kumar K.V kfree(chan); 632b530cc79SEric Van Hensbergen fail: 633b530cc79SEric Van Hensbergen return err; 634b530cc79SEric Van Hensbergen } 635b530cc79SEric Van Hensbergen 636ee443996SEric Van Hensbergen 637ee443996SEric Van Hensbergen /** 638ee443996SEric Van Hensbergen * p9_virtio_create - allocate a new virtio channel 6398b81ef58SEric Van Hensbergen * @client: client instance invoking this transport 640ee443996SEric Van Hensbergen * @devname: string identifying the channel to connect to (unused) 641ee443996SEric Van Hensbergen * @args: args passed from sys_mount() for per-transport options (unused) 642ee443996SEric Van Hensbergen * 643ee443996SEric Van Hensbergen * This sets up a transport channel for 9p communication. Right now 644b530cc79SEric Van Hensbergen * we only match the first available channel, but eventually we couldlook up 645b530cc79SEric Van Hensbergen * alternate channels by matching devname versus a virtio_config entry. 646b530cc79SEric Van Hensbergen * We use a simple reference count mechanism to ensure that only a single 647ee443996SEric Van Hensbergen * mount has a channel open at a time. 648ee443996SEric Van Hensbergen * 649ee443996SEric Van Hensbergen */ 650ee443996SEric Van Hensbergen 6518b81ef58SEric Van Hensbergen static int 6528b81ef58SEric Van Hensbergen p9_virtio_create(struct p9_client *client, const char *devname, char *args) 653b530cc79SEric Van Hensbergen { 65437c1209dSAneesh Kumar K.V struct virtio_chan *chan; 655c1a7c226SAneesh Kumar K.V int ret = -ENOENT; 65637c1209dSAneesh Kumar K.V int found = 0; 657b530cc79SEric Van Hensbergen 65810aa1452STomas Bortoli if (devname == NULL) 65910aa1452STomas Bortoli return -EINVAL; 66010aa1452STomas Bortoli 661c1549497SJosef 'Jeff' Sipek mutex_lock(&virtio_9p_lock); 66237c1209dSAneesh Kumar K.V list_for_each_entry(chan, &virtio_chan_list, chan_list) { 663edcd9d97Spiaojun if (!strcmp(devname, chan->tag)) { 664f75580c4SAneesh Kumar K.V if (!chan->inuse) { 665b530cc79SEric Van Hensbergen chan->inuse = true; 66637c1209dSAneesh Kumar K.V found = 1; 667b530cc79SEric Van Hensbergen break; 668f75580c4SAneesh Kumar K.V } 669c1a7c226SAneesh Kumar K.V ret = -EBUSY; 670f75580c4SAneesh Kumar K.V } 671b530cc79SEric Van Hensbergen } 672c1549497SJosef 'Jeff' Sipek mutex_unlock(&virtio_9p_lock); 673b530cc79SEric Van Hensbergen 67437c1209dSAneesh Kumar K.V if (!found) { 675c7c72c5aSAneesh Kumar K.V pr_err("no channels available for device %s\n", devname); 676c1a7c226SAneesh Kumar K.V return ret; 677b530cc79SEric Van Hensbergen } 678b530cc79SEric Van Hensbergen 6798b81ef58SEric Van Hensbergen client->trans = (void *)chan; 680562ada61SEric Van Hensbergen client->status = Connected; 681fea511a6SEric Van Hensbergen chan->client = client; 682b530cc79SEric Van Hensbergen 6838b81ef58SEric Van Hensbergen return 0; 684b530cc79SEric Van Hensbergen } 685b530cc79SEric Van Hensbergen 686ee443996SEric Van Hensbergen /** 687ee443996SEric Van Hensbergen * p9_virtio_remove - clean up resources associated with a virtio device 688ee443996SEric Van Hensbergen * @vdev: virtio device to remove 689ee443996SEric Van Hensbergen * 690ee443996SEric Van Hensbergen */ 691ee443996SEric Van Hensbergen 692f3933545SEric Van Hensbergen static void p9_virtio_remove(struct virtio_device *vdev) 693f3933545SEric Van Hensbergen { 694f3933545SEric Van Hensbergen struct virtio_chan *chan = vdev->priv; 6958051a2a5SMichael S. Tsirkin unsigned long warning_time; 69637c1209dSAneesh Kumar K.V 69737c1209dSAneesh Kumar K.V mutex_lock(&virtio_9p_lock); 6988051a2a5SMichael S. Tsirkin 6998051a2a5SMichael S. Tsirkin /* Remove self from list so we don't get new users. */ 70037c1209dSAneesh Kumar K.V list_del(&chan->chan_list); 7018051a2a5SMichael S. Tsirkin warning_time = jiffies; 7028051a2a5SMichael S. Tsirkin 7038051a2a5SMichael S. Tsirkin /* Wait for existing users to close. */ 7048051a2a5SMichael S. Tsirkin while (chan->inuse) { 70537c1209dSAneesh Kumar K.V mutex_unlock(&virtio_9p_lock); 7068051a2a5SMichael S. Tsirkin msleep(250); 7078051a2a5SMichael S. Tsirkin if (time_after(jiffies, warning_time + 10 * HZ)) { 7088051a2a5SMichael S. Tsirkin dev_emerg(&vdev->dev, 7098051a2a5SMichael S. Tsirkin "p9_virtio_remove: waiting for device in use.\n"); 7108051a2a5SMichael S. Tsirkin warning_time = jiffies; 7118051a2a5SMichael S. Tsirkin } 7128051a2a5SMichael S. Tsirkin mutex_lock(&virtio_9p_lock); 7138051a2a5SMichael S. Tsirkin } 7148051a2a5SMichael S. Tsirkin 7158051a2a5SMichael S. Tsirkin mutex_unlock(&virtio_9p_lock); 7168051a2a5SMichael S. Tsirkin 717ea52bf8eSPierre Morel vdev->config->reset(vdev); 7188051a2a5SMichael S. Tsirkin vdev->config->del_vqs(vdev); 7198051a2a5SMichael S. Tsirkin 72086c84373SAneesh Kumar K.V sysfs_remove_file(&(vdev->dev.kobj), &dev_attr_mount_tag.attr); 721e0d6cb9cSMichael Marineau kobject_uevent(&(vdev->dev.kobj), KOBJ_CHANGE); 72297ee9b02SAneesh Kumar K.V kfree(chan->tag); 72352f44e0dSVenkateswararao Jujjuri (JV) kfree(chan->vc_wq); 72437c1209dSAneesh Kumar K.V kfree(chan); 72537c1209dSAneesh Kumar K.V 726f3933545SEric Van Hensbergen } 727f3933545SEric Van Hensbergen 728b530cc79SEric Van Hensbergen static struct virtio_device_id id_table[] = { 729b530cc79SEric Van Hensbergen { VIRTIO_ID_9P, VIRTIO_DEV_ANY_ID }, 730b530cc79SEric Van Hensbergen { 0 }, 731b530cc79SEric Van Hensbergen }; 732b530cc79SEric Van Hensbergen 73397ee9b02SAneesh Kumar K.V static unsigned int features[] = { 73497ee9b02SAneesh Kumar K.V VIRTIO_9P_MOUNT_TAG, 73597ee9b02SAneesh Kumar K.V }; 73697ee9b02SAneesh Kumar K.V 737b530cc79SEric Van Hensbergen /* The standard "struct lguest_driver": */ 738b530cc79SEric Van Hensbergen static struct virtio_driver p9_virtio_drv = { 73997ee9b02SAneesh Kumar K.V .feature_table = features, 74097ee9b02SAneesh Kumar K.V .feature_table_size = ARRAY_SIZE(features), 741b530cc79SEric Van Hensbergen .driver.name = KBUILD_MODNAME, 742b530cc79SEric Van Hensbergen .driver.owner = THIS_MODULE, 743b530cc79SEric Van Hensbergen .id_table = id_table, 744b530cc79SEric Van Hensbergen .probe = p9_virtio_probe, 745f3933545SEric Van Hensbergen .remove = p9_virtio_remove, 746b530cc79SEric Van Hensbergen }; 747b530cc79SEric Van Hensbergen 748b530cc79SEric Van Hensbergen static struct p9_trans_module p9_virtio_trans = { 749b530cc79SEric Van Hensbergen .name = "virtio", 750b530cc79SEric Van Hensbergen .create = p9_virtio_create, 7518b81ef58SEric Van Hensbergen .close = p9_virtio_close, 75291b8534fSEric Van Hensbergen .request = p9_virtio_request, 753dc893e19SArnd Bergmann .zc_request = p9_virtio_zc_request, 75491b8534fSEric Van Hensbergen .cancel = p9_virtio_cancel, 755728356deSTomas Bortoli .cancelled = p9_virtio_cancelled, 756b49d8b5dSAneesh Kumar K.V /* 757b49d8b5dSAneesh Kumar K.V * We leave one entry for input and one entry for response 758b49d8b5dSAneesh Kumar K.V * headers. We also skip one more entry to accomodate, address 759b49d8b5dSAneesh Kumar K.V * that are not at page boundary, that can result in an extra 760b49d8b5dSAneesh Kumar K.V * page in zero copy. 761b49d8b5dSAneesh Kumar K.V */ 762b49d8b5dSAneesh Kumar K.V .maxsize = PAGE_SIZE * (VIRTQUEUE_NUM - 3), 763f94741fdSEric Van Hensbergen .def = 1, 76472029fe8STejun Heo .owner = THIS_MODULE, 765b530cc79SEric Van Hensbergen }; 766b530cc79SEric Van Hensbergen 767b530cc79SEric Van Hensbergen /* The standard init function */ 768b530cc79SEric Van Hensbergen static int __init p9_virtio_init(void) 769b530cc79SEric Van Hensbergen { 77037c1209dSAneesh Kumar K.V INIT_LIST_HEAD(&virtio_chan_list); 771b530cc79SEric Van Hensbergen 772b530cc79SEric Van Hensbergen v9fs_register_trans(&p9_virtio_trans); 773b530cc79SEric Van Hensbergen return register_virtio_driver(&p9_virtio_drv); 774b530cc79SEric Van Hensbergen } 775b530cc79SEric Van Hensbergen 776f3933545SEric Van Hensbergen static void __exit p9_virtio_cleanup(void) 777f3933545SEric Van Hensbergen { 778f3933545SEric Van Hensbergen unregister_virtio_driver(&p9_virtio_drv); 77972029fe8STejun Heo v9fs_unregister_trans(&p9_virtio_trans); 780f3933545SEric Van Hensbergen } 781f3933545SEric Van Hensbergen 782b530cc79SEric Van Hensbergen module_init(p9_virtio_init); 783f3933545SEric Van Hensbergen module_exit(p9_virtio_cleanup); 784b530cc79SEric Van Hensbergen 785b530cc79SEric Van Hensbergen MODULE_DEVICE_TABLE(virtio, id_table); 786b530cc79SEric Van Hensbergen MODULE_AUTHOR("Eric Van Hensbergen <ericvh@gmail.com>"); 787b530cc79SEric Van Hensbergen MODULE_DESCRIPTION("Virtio 9p Transport"); 788b530cc79SEric Van Hensbergen MODULE_LICENSE("GPL"); 789