xref: /openbmc/qemu/hw/9pfs/9p.c (revision 964d2a0c)
1 /*
2  * Virtio 9p backend
3  *
4  * Copyright IBM, Corp. 2010
5  *
6  * Authors:
7  *  Anthony Liguori   <aliguori@us.ibm.com>
8  *
9  * This work is licensed under the terms of the GNU GPL, version 2.  See
10  * the COPYING file in the top-level directory.
11  *
12  */
13 
14 /*
15  * Not so fast! You might want to read the 9p developer docs first:
16  * https://wiki.qemu.org/Documentation/9p
17  */
18 
19 #include "qemu/osdep.h"
20 #ifdef CONFIG_LINUX
21 #include <linux/limits.h>
22 #endif
23 #include <glib/gprintf.h>
24 #include "hw/virtio/virtio.h"
25 #include "qapi/error.h"
26 #include "qemu/error-report.h"
27 #include "qemu/iov.h"
28 #include "qemu/main-loop.h"
29 #include "qemu/sockets.h"
30 #include "virtio-9p.h"
31 #include "fsdev/qemu-fsdev.h"
32 #include "9p-xattr.h"
33 #include "9p-util.h"
34 #include "coth.h"
35 #include "trace.h"
36 #include "migration/blocker.h"
37 #include "qemu/xxhash.h"
38 #include <math.h>
39 
40 int open_fd_hw;
41 int total_open_fd;
42 static int open_fd_rc;
43 
44 enum {
45     Oread   = 0x00,
46     Owrite  = 0x01,
47     Ordwr   = 0x02,
48     Oexec   = 0x03,
49     Oexcl   = 0x04,
50     Otrunc  = 0x10,
51     Orexec  = 0x20,
52     Orclose = 0x40,
53     Oappend = 0x80,
54 };
55 
56 P9ARRAY_DEFINE_TYPE(V9fsPath, v9fs_path_free);
57 
58 static ssize_t pdu_marshal(V9fsPDU *pdu, size_t offset, const char *fmt, ...)
59 {
60     ssize_t ret;
61     va_list ap;
62 
63     va_start(ap, fmt);
64     ret = pdu->s->transport->pdu_vmarshal(pdu, offset, fmt, ap);
65     va_end(ap);
66 
67     return ret;
68 }
69 
70 static ssize_t pdu_unmarshal(V9fsPDU *pdu, size_t offset, const char *fmt, ...)
71 {
72     ssize_t ret;
73     va_list ap;
74 
75     va_start(ap, fmt);
76     ret = pdu->s->transport->pdu_vunmarshal(pdu, offset, fmt, ap);
77     va_end(ap);
78 
79     return ret;
80 }
81 
82 static int omode_to_uflags(int8_t mode)
83 {
84     int ret = 0;
85 
86     switch (mode & 3) {
87     case Oread:
88         ret = O_RDONLY;
89         break;
90     case Ordwr:
91         ret = O_RDWR;
92         break;
93     case Owrite:
94         ret = O_WRONLY;
95         break;
96     case Oexec:
97         ret = O_RDONLY;
98         break;
99     }
100 
101     if (mode & Otrunc) {
102         ret |= O_TRUNC;
103     }
104 
105     if (mode & Oappend) {
106         ret |= O_APPEND;
107     }
108 
109     if (mode & Oexcl) {
110         ret |= O_EXCL;
111     }
112 
113     return ret;
114 }
115 
116 typedef struct DotlOpenflagMap {
117     int dotl_flag;
118     int open_flag;
119 } DotlOpenflagMap;
120 
121 static int dotl_to_open_flags(int flags)
122 {
123     int i;
124     /*
125      * We have same bits for P9_DOTL_READONLY, P9_DOTL_WRONLY
126      * and P9_DOTL_NOACCESS
127      */
128     int oflags = flags & O_ACCMODE;
129 
130     DotlOpenflagMap dotl_oflag_map[] = {
131         { P9_DOTL_CREATE, O_CREAT },
132         { P9_DOTL_EXCL, O_EXCL },
133         { P9_DOTL_NOCTTY , O_NOCTTY },
134         { P9_DOTL_TRUNC, O_TRUNC },
135         { P9_DOTL_APPEND, O_APPEND },
136         { P9_DOTL_NONBLOCK, O_NONBLOCK } ,
137         { P9_DOTL_DSYNC, O_DSYNC },
138         { P9_DOTL_FASYNC, FASYNC },
139 #ifndef CONFIG_DARWIN
140         { P9_DOTL_NOATIME, O_NOATIME },
141         /*
142          *  On Darwin, we could map to F_NOCACHE, which is
143          *  similar, but doesn't quite have the same
144          *  semantics. However, we don't support O_DIRECT
145          *  even on linux at the moment, so we just ignore
146          *  it here.
147          */
148         { P9_DOTL_DIRECT, O_DIRECT },
149 #endif
150         { P9_DOTL_LARGEFILE, O_LARGEFILE },
151         { P9_DOTL_DIRECTORY, O_DIRECTORY },
152         { P9_DOTL_NOFOLLOW, O_NOFOLLOW },
153         { P9_DOTL_SYNC, O_SYNC },
154     };
155 
156     for (i = 0; i < ARRAY_SIZE(dotl_oflag_map); i++) {
157         if (flags & dotl_oflag_map[i].dotl_flag) {
158             oflags |= dotl_oflag_map[i].open_flag;
159         }
160     }
161 
162     return oflags;
163 }
164 
165 void cred_init(FsCred *credp)
166 {
167     credp->fc_uid = -1;
168     credp->fc_gid = -1;
169     credp->fc_mode = -1;
170     credp->fc_rdev = -1;
171 }
172 
173 static int get_dotl_openflags(V9fsState *s, int oflags)
174 {
175     int flags;
176     /*
177      * Filter the client open flags
178      */
179     flags = dotl_to_open_flags(oflags);
180     flags &= ~(O_NOCTTY | O_ASYNC | O_CREAT);
181 #ifndef CONFIG_DARWIN
182     /*
183      * Ignore direct disk access hint until the server supports it.
184      */
185     flags &= ~O_DIRECT;
186 #endif
187     return flags;
188 }
189 
190 void v9fs_path_init(V9fsPath *path)
191 {
192     path->data = NULL;
193     path->size = 0;
194 }
195 
196 void v9fs_path_free(V9fsPath *path)
197 {
198     g_free(path->data);
199     path->data = NULL;
200     path->size = 0;
201 }
202 
203 
204 void G_GNUC_PRINTF(2, 3)
205 v9fs_path_sprintf(V9fsPath *path, const char *fmt, ...)
206 {
207     va_list ap;
208 
209     v9fs_path_free(path);
210 
211     va_start(ap, fmt);
212     /* Bump the size for including terminating NULL */
213     path->size = g_vasprintf(&path->data, fmt, ap) + 1;
214     va_end(ap);
215 }
216 
217 void v9fs_path_copy(V9fsPath *dst, const V9fsPath *src)
218 {
219     v9fs_path_free(dst);
220     dst->size = src->size;
221     dst->data = g_memdup(src->data, src->size);
222 }
223 
224 int v9fs_name_to_path(V9fsState *s, V9fsPath *dirpath,
225                       const char *name, V9fsPath *path)
226 {
227     int err;
228     err = s->ops->name_to_path(&s->ctx, dirpath, name, path);
229     if (err < 0) {
230         err = -errno;
231     }
232     return err;
233 }
234 
235 /*
236  * Return TRUE if s1 is an ancestor of s2.
237  *
238  * E.g. "a/b" is an ancestor of "a/b/c" but not of "a/bc/d".
239  * As a special case, We treat s1 as ancestor of s2 if they are same!
240  */
241 static int v9fs_path_is_ancestor(V9fsPath *s1, V9fsPath *s2)
242 {
243     if (!strncmp(s1->data, s2->data, s1->size - 1)) {
244         if (s2->data[s1->size - 1] == '\0' || s2->data[s1->size - 1] == '/') {
245             return 1;
246         }
247     }
248     return 0;
249 }
250 
251 static size_t v9fs_string_size(V9fsString *str)
252 {
253     return str->size;
254 }
255 
256 /*
257  * returns 0 if fid got re-opened, 1 if not, < 0 on error
258  */
259 static int coroutine_fn v9fs_reopen_fid(V9fsPDU *pdu, V9fsFidState *f)
260 {
261     int err = 1;
262     if (f->fid_type == P9_FID_FILE) {
263         if (f->fs.fd == -1) {
264             do {
265                 err = v9fs_co_open(pdu, f, f->open_flags);
266             } while (err == -EINTR && !pdu->cancelled);
267         }
268     } else if (f->fid_type == P9_FID_DIR) {
269         if (f->fs.dir.stream == NULL) {
270             do {
271                 err = v9fs_co_opendir(pdu, f);
272             } while (err == -EINTR && !pdu->cancelled);
273         }
274     }
275     return err;
276 }
277 
278 static V9fsFidState *coroutine_fn get_fid(V9fsPDU *pdu, int32_t fid)
279 {
280     int err;
281     V9fsFidState *f;
282     V9fsState *s = pdu->s;
283 
284     f = g_hash_table_lookup(s->fids, GINT_TO_POINTER(fid));
285     if (f) {
286         BUG_ON(f->clunked);
287         /*
288          * Update the fid ref upfront so that
289          * we don't get reclaimed when we yield
290          * in open later.
291          */
292         f->ref++;
293         /*
294          * check whether we need to reopen the
295          * file. We might have closed the fd
296          * while trying to free up some file
297          * descriptors.
298          */
299         err = v9fs_reopen_fid(pdu, f);
300         if (err < 0) {
301             f->ref--;
302             return NULL;
303         }
304         /*
305          * Mark the fid as referenced so that the LRU
306          * reclaim won't close the file descriptor
307          */
308         f->flags |= FID_REFERENCED;
309         return f;
310     }
311     return NULL;
312 }
313 
314 static V9fsFidState *alloc_fid(V9fsState *s, int32_t fid)
315 {
316     V9fsFidState *f;
317 
318     f = g_hash_table_lookup(s->fids, GINT_TO_POINTER(fid));
319     if (f) {
320         /* If fid is already there return NULL */
321         BUG_ON(f->clunked);
322         return NULL;
323     }
324     f = g_new0(V9fsFidState, 1);
325     f->fid = fid;
326     f->fid_type = P9_FID_NONE;
327     f->ref = 1;
328     /*
329      * Mark the fid as referenced so that the LRU
330      * reclaim won't close the file descriptor
331      */
332     f->flags |= FID_REFERENCED;
333     g_hash_table_insert(s->fids, GINT_TO_POINTER(fid), f);
334 
335     v9fs_readdir_init(s->proto_version, &f->fs.dir);
336     v9fs_readdir_init(s->proto_version, &f->fs_reclaim.dir);
337 
338     return f;
339 }
340 
341 static int coroutine_fn v9fs_xattr_fid_clunk(V9fsPDU *pdu, V9fsFidState *fidp)
342 {
343     int retval = 0;
344 
345     if (fidp->fs.xattr.xattrwalk_fid) {
346         /* getxattr/listxattr fid */
347         goto free_value;
348     }
349     /*
350      * if this is fid for setxattr. clunk should
351      * result in setxattr localcall
352      */
353     if (fidp->fs.xattr.len != fidp->fs.xattr.copied_len) {
354         /* clunk after partial write */
355         retval = -EINVAL;
356         goto free_out;
357     }
358     if (fidp->fs.xattr.len) {
359         retval = v9fs_co_lsetxattr(pdu, &fidp->path, &fidp->fs.xattr.name,
360                                    fidp->fs.xattr.value,
361                                    fidp->fs.xattr.len,
362                                    fidp->fs.xattr.flags);
363     } else {
364         retval = v9fs_co_lremovexattr(pdu, &fidp->path, &fidp->fs.xattr.name);
365     }
366 free_out:
367     v9fs_string_free(&fidp->fs.xattr.name);
368 free_value:
369     g_free(fidp->fs.xattr.value);
370     return retval;
371 }
372 
373 static int coroutine_fn free_fid(V9fsPDU *pdu, V9fsFidState *fidp)
374 {
375     int retval = 0;
376 
377     if (fidp->fid_type == P9_FID_FILE) {
378         /* If we reclaimed the fd no need to close */
379         if (fidp->fs.fd != -1) {
380             retval = v9fs_co_close(pdu, &fidp->fs);
381         }
382     } else if (fidp->fid_type == P9_FID_DIR) {
383         if (fidp->fs.dir.stream != NULL) {
384             retval = v9fs_co_closedir(pdu, &fidp->fs);
385         }
386     } else if (fidp->fid_type == P9_FID_XATTR) {
387         retval = v9fs_xattr_fid_clunk(pdu, fidp);
388     }
389     v9fs_path_free(&fidp->path);
390     g_free(fidp);
391     return retval;
392 }
393 
394 static int coroutine_fn put_fid(V9fsPDU *pdu, V9fsFidState *fidp)
395 {
396     BUG_ON(!fidp->ref);
397     fidp->ref--;
398     /*
399      * Don't free the fid if it is in reclaim list
400      */
401     if (!fidp->ref && fidp->clunked) {
402         if (fidp->fid == pdu->s->root_fid) {
403             /*
404              * if the clunked fid is root fid then we
405              * have unmounted the fs on the client side.
406              * delete the migration blocker. Ideally, this
407              * should be hooked to transport close notification
408              */
409             migrate_del_blocker(&pdu->s->migration_blocker);
410         }
411         return free_fid(pdu, fidp);
412     }
413     return 0;
414 }
415 
416 static V9fsFidState *clunk_fid(V9fsState *s, int32_t fid)
417 {
418     V9fsFidState *fidp;
419 
420     /* TODO: Use g_hash_table_steal_extended() instead? */
421     fidp = g_hash_table_lookup(s->fids, GINT_TO_POINTER(fid));
422     if (fidp) {
423         g_hash_table_remove(s->fids, GINT_TO_POINTER(fid));
424         fidp->clunked = true;
425         return fidp;
426     }
427     return NULL;
428 }
429 
430 void coroutine_fn v9fs_reclaim_fd(V9fsPDU *pdu)
431 {
432     int reclaim_count = 0;
433     V9fsState *s = pdu->s;
434     V9fsFidState *f;
435     GHashTableIter iter;
436     gpointer fid;
437 
438     g_hash_table_iter_init(&iter, s->fids);
439 
440     QSLIST_HEAD(, V9fsFidState) reclaim_list =
441         QSLIST_HEAD_INITIALIZER(reclaim_list);
442 
443     while (g_hash_table_iter_next(&iter, &fid, (gpointer *) &f)) {
444         /*
445          * Unlink fids cannot be reclaimed. Check
446          * for them and skip them. Also skip fids
447          * currently being operated on.
448          */
449         if (f->ref || f->flags & FID_NON_RECLAIMABLE) {
450             continue;
451         }
452         /*
453          * if it is a recently referenced fid
454          * we leave the fid untouched and clear the
455          * reference bit. We come back to it later
456          * in the next iteration. (a simple LRU without
457          * moving list elements around)
458          */
459         if (f->flags & FID_REFERENCED) {
460             f->flags &= ~FID_REFERENCED;
461             continue;
462         }
463         /*
464          * Add fids to reclaim list.
465          */
466         if (f->fid_type == P9_FID_FILE) {
467             if (f->fs.fd != -1) {
468                 /*
469                  * Up the reference count so that
470                  * a clunk request won't free this fid
471                  */
472                 f->ref++;
473                 QSLIST_INSERT_HEAD(&reclaim_list, f, reclaim_next);
474                 f->fs_reclaim.fd = f->fs.fd;
475                 f->fs.fd = -1;
476                 reclaim_count++;
477             }
478         } else if (f->fid_type == P9_FID_DIR) {
479             if (f->fs.dir.stream != NULL) {
480                 /*
481                  * Up the reference count so that
482                  * a clunk request won't free this fid
483                  */
484                 f->ref++;
485                 QSLIST_INSERT_HEAD(&reclaim_list, f, reclaim_next);
486                 f->fs_reclaim.dir.stream = f->fs.dir.stream;
487                 f->fs.dir.stream = NULL;
488                 reclaim_count++;
489             }
490         }
491         if (reclaim_count >= open_fd_rc) {
492             break;
493         }
494     }
495     /*
496      * Now close the fid in reclaim list. Free them if they
497      * are already clunked.
498      */
499     while (!QSLIST_EMPTY(&reclaim_list)) {
500         f = QSLIST_FIRST(&reclaim_list);
501         QSLIST_REMOVE(&reclaim_list, f, V9fsFidState, reclaim_next);
502         if (f->fid_type == P9_FID_FILE) {
503             v9fs_co_close(pdu, &f->fs_reclaim);
504         } else if (f->fid_type == P9_FID_DIR) {
505             v9fs_co_closedir(pdu, &f->fs_reclaim);
506         }
507         /*
508          * Now drop the fid reference, free it
509          * if clunked.
510          */
511         put_fid(pdu, f);
512     }
513 }
514 
515 /*
516  * This is used when a path is removed from the directory tree. Any
517  * fids that still reference it must not be closed from then on, since
518  * they cannot be reopened.
519  */
520 static int coroutine_fn v9fs_mark_fids_unreclaim(V9fsPDU *pdu, V9fsPath *path)
521 {
522     int err = 0;
523     V9fsState *s = pdu->s;
524     V9fsFidState *fidp;
525     gpointer fid;
526     GHashTableIter iter;
527     /*
528      * The most common case is probably that we have exactly one
529      * fid for the given path, so preallocate exactly one.
530      */
531     g_autoptr(GArray) to_reopen = g_array_sized_new(FALSE, FALSE,
532             sizeof(V9fsFidState *), 1);
533     gint i;
534 
535     g_hash_table_iter_init(&iter, s->fids);
536 
537     /*
538      * We iterate over the fid table looking for the entries we need
539      * to reopen, and store them in to_reopen. This is because
540      * v9fs_reopen_fid() and put_fid() yield. This allows the fid table
541      * to be modified in the meantime, invalidating our iterator.
542      */
543     while (g_hash_table_iter_next(&iter, &fid, (gpointer *) &fidp)) {
544         if (fidp->path.size == path->size &&
545             !memcmp(fidp->path.data, path->data, path->size)) {
546             /*
547              * Ensure the fid survives a potential clunk request during
548              * v9fs_reopen_fid or put_fid.
549              */
550             fidp->ref++;
551             fidp->flags |= FID_NON_RECLAIMABLE;
552             g_array_append_val(to_reopen, fidp);
553         }
554     }
555 
556     for (i = 0; i < to_reopen->len; i++) {
557         fidp = g_array_index(to_reopen, V9fsFidState*, i);
558         /* reopen the file/dir if already closed */
559         err = v9fs_reopen_fid(pdu, fidp);
560         if (err < 0) {
561             break;
562         }
563     }
564 
565     for (i = 0; i < to_reopen->len; i++) {
566         put_fid(pdu, g_array_index(to_reopen, V9fsFidState*, i));
567     }
568     return err;
569 }
570 
571 static void coroutine_fn virtfs_reset(V9fsPDU *pdu)
572 {
573     V9fsState *s = pdu->s;
574     V9fsFidState *fidp;
575     GList *freeing;
576     /*
577      * Get a list of all the values (fid states) in the table, which
578      * we then...
579      */
580     g_autoptr(GList) fids = g_hash_table_get_values(s->fids);
581 
582     /* ... remove from the table, taking over ownership. */
583     g_hash_table_steal_all(s->fids);
584 
585     /*
586      * This allows us to release our references to them asynchronously without
587      * iterating over the hash table and risking iterator invalidation
588      * through concurrent modifications.
589      */
590     for (freeing = fids; freeing; freeing = freeing->next) {
591         fidp = freeing->data;
592         fidp->ref++;
593         fidp->clunked = true;
594         put_fid(pdu, fidp);
595     }
596 }
597 
598 #define P9_QID_TYPE_DIR         0x80
599 #define P9_QID_TYPE_SYMLINK     0x02
600 
601 #define P9_STAT_MODE_DIR        0x80000000
602 #define P9_STAT_MODE_APPEND     0x40000000
603 #define P9_STAT_MODE_EXCL       0x20000000
604 #define P9_STAT_MODE_MOUNT      0x10000000
605 #define P9_STAT_MODE_AUTH       0x08000000
606 #define P9_STAT_MODE_TMP        0x04000000
607 #define P9_STAT_MODE_SYMLINK    0x02000000
608 #define P9_STAT_MODE_LINK       0x01000000
609 #define P9_STAT_MODE_DEVICE     0x00800000
610 #define P9_STAT_MODE_NAMED_PIPE 0x00200000
611 #define P9_STAT_MODE_SOCKET     0x00100000
612 #define P9_STAT_MODE_SETUID     0x00080000
613 #define P9_STAT_MODE_SETGID     0x00040000
614 #define P9_STAT_MODE_SETVTX     0x00010000
615 
616 #define P9_STAT_MODE_TYPE_BITS (P9_STAT_MODE_DIR |          \
617                                 P9_STAT_MODE_SYMLINK |      \
618                                 P9_STAT_MODE_LINK |         \
619                                 P9_STAT_MODE_DEVICE |       \
620                                 P9_STAT_MODE_NAMED_PIPE |   \
621                                 P9_STAT_MODE_SOCKET)
622 
623 /* Mirrors all bits of a byte. So e.g. binary 10100000 would become 00000101. */
624 static inline uint8_t mirror8bit(uint8_t byte)
625 {
626     return (byte * 0x0202020202ULL & 0x010884422010ULL) % 1023;
627 }
628 
629 /* Same as mirror8bit() just for a 64 bit data type instead for a byte. */
630 static inline uint64_t mirror64bit(uint64_t value)
631 {
632     return ((uint64_t)mirror8bit(value         & 0xff) << 56) |
633            ((uint64_t)mirror8bit((value >> 8)  & 0xff) << 48) |
634            ((uint64_t)mirror8bit((value >> 16) & 0xff) << 40) |
635            ((uint64_t)mirror8bit((value >> 24) & 0xff) << 32) |
636            ((uint64_t)mirror8bit((value >> 32) & 0xff) << 24) |
637            ((uint64_t)mirror8bit((value >> 40) & 0xff) << 16) |
638            ((uint64_t)mirror8bit((value >> 48) & 0xff) << 8)  |
639            ((uint64_t)mirror8bit((value >> 56) & 0xff));
640 }
641 
642 /*
643  * Parameter k for the Exponential Golomb algorithm to be used.
644  *
645  * The smaller this value, the smaller the minimum bit count for the Exp.
646  * Golomb generated affixes will be (at lowest index) however for the
647  * price of having higher maximum bit count of generated affixes (at highest
648  * index). Likewise increasing this parameter yields in smaller maximum bit
649  * count for the price of having higher minimum bit count.
650  *
651  * In practice that means: a good value for k depends on the expected amount
652  * of devices to be exposed by one export. For a small amount of devices k
653  * should be small, for a large amount of devices k might be increased
654  * instead. The default of k=0 should be fine for most users though.
655  *
656  * IMPORTANT: In case this ever becomes a runtime parameter; the value of
657  * k should not change as long as guest is still running! Because that would
658  * cause completely different inode numbers to be generated on guest.
659  */
660 #define EXP_GOLOMB_K    0
661 
662 /**
663  * expGolombEncode() - Exponential Golomb algorithm for arbitrary k
664  *                     (including k=0).
665  *
666  * @n: natural number (or index) of the prefix to be generated
667  *     (1, 2, 3, ...)
668  * @k: parameter k of Exp. Golomb algorithm to be used
669  *     (see comment on EXP_GOLOMB_K macro for details about k)
670  * Return: prefix for given @n and @k
671  *
672  * The Exponential Golomb algorithm generates prefixes (NOT suffixes!)
673  * with growing length and with the mathematical property of being
674  * "prefix-free". The latter means the generated prefixes can be prepended
675  * in front of arbitrary numbers and the resulting concatenated numbers are
676  * guaranteed to be always unique.
677  *
678  * This is a minor adjustment to the original Exp. Golomb algorithm in the
679  * sense that lowest allowed index (@n) starts with 1, not with zero.
680  */
681 static VariLenAffix expGolombEncode(uint64_t n, int k)
682 {
683     const uint64_t value = n + (1 << k) - 1;
684     const int bits = (int) log2(value) + 1;
685     return (VariLenAffix) {
686         .type = AffixType_Prefix,
687         .value = value,
688         .bits = bits + MAX((bits - 1 - k), 0)
689     };
690 }
691 
692 /**
693  * invertAffix() - Converts a suffix into a prefix, or a prefix into a suffix.
694  * @affix: either suffix or prefix to be inverted
695  * Return: inversion of passed @affix
696  *
697  * Simply mirror all bits of the affix value, for the purpose to preserve
698  * respectively the mathematical "prefix-free" or "suffix-free" property
699  * after the conversion.
700  *
701  * If a passed prefix is suitable to create unique numbers, then the
702  * returned suffix is suitable to create unique numbers as well (and vice
703  * versa).
704  */
705 static VariLenAffix invertAffix(const VariLenAffix *affix)
706 {
707     return (VariLenAffix) {
708         .type =
709             (affix->type == AffixType_Suffix) ?
710                 AffixType_Prefix : AffixType_Suffix,
711         .value =
712             mirror64bit(affix->value) >>
713             ((sizeof(affix->value) * 8) - affix->bits),
714         .bits = affix->bits
715     };
716 }
717 
718 /**
719  * affixForIndex() - Generates suffix numbers with "suffix-free" property.
720  * @index: natural number (or index) of the suffix to be generated
721  *         (1, 2, 3, ...)
722  * Return: Suffix suitable to assemble unique number.
723  *
724  * This is just a wrapper function on top of the Exp. Golomb algorithm.
725  *
726  * Since the Exp. Golomb algorithm generates prefixes, but we need suffixes,
727  * this function converts the Exp. Golomb prefixes into appropriate suffixes
728  * which are still suitable for generating unique numbers.
729  */
730 static VariLenAffix affixForIndex(uint64_t index)
731 {
732     VariLenAffix prefix;
733     prefix = expGolombEncode(index, EXP_GOLOMB_K);
734     return invertAffix(&prefix); /* convert prefix to suffix */
735 }
736 
737 static uint32_t qpp_hash(QppEntry e)
738 {
739     return qemu_xxhash4(e.ino_prefix, e.dev);
740 }
741 
742 static uint32_t qpf_hash(QpfEntry e)
743 {
744     return qemu_xxhash4(e.ino, e.dev);
745 }
746 
747 static bool qpd_cmp_func(const void *obj, const void *userp)
748 {
749     const QpdEntry *e1 = obj, *e2 = userp;
750     return e1->dev == e2->dev;
751 }
752 
753 static bool qpp_cmp_func(const void *obj, const void *userp)
754 {
755     const QppEntry *e1 = obj, *e2 = userp;
756     return e1->dev == e2->dev && e1->ino_prefix == e2->ino_prefix;
757 }
758 
759 static bool qpf_cmp_func(const void *obj, const void *userp)
760 {
761     const QpfEntry *e1 = obj, *e2 = userp;
762     return e1->dev == e2->dev && e1->ino == e2->ino;
763 }
764 
765 static void qp_table_remove(void *p, uint32_t h, void *up)
766 {
767     g_free(p);
768 }
769 
770 static void qp_table_destroy(struct qht *ht)
771 {
772     if (!ht || !ht->map) {
773         return;
774     }
775     qht_iter(ht, qp_table_remove, NULL);
776     qht_destroy(ht);
777 }
778 
779 static void qpd_table_init(struct qht *ht)
780 {
781     qht_init(ht, qpd_cmp_func, 1, QHT_MODE_AUTO_RESIZE);
782 }
783 
784 static void qpp_table_init(struct qht *ht)
785 {
786     qht_init(ht, qpp_cmp_func, 1, QHT_MODE_AUTO_RESIZE);
787 }
788 
789 static void qpf_table_init(struct qht *ht)
790 {
791     qht_init(ht, qpf_cmp_func, 1 << 16, QHT_MODE_AUTO_RESIZE);
792 }
793 
794 /*
795  * Returns how many (high end) bits of inode numbers of the passed fs
796  * device shall be used (in combination with the device number) to
797  * generate hash values for qpp_table entries.
798  *
799  * This function is required if variable length suffixes are used for inode
800  * number mapping on guest level. Since a device may end up having multiple
801  * entries in qpp_table, each entry most probably with a different suffix
802  * length, we thus need this function in conjunction with qpd_table to
803  * "agree" about a fix amount of bits (per device) to be always used for
804  * generating hash values for the purpose of accessing qpp_table in order
805  * get consistent behaviour when accessing qpp_table.
806  */
807 static int qid_inode_prefix_hash_bits(V9fsPDU *pdu, dev_t dev)
808 {
809     QpdEntry lookup = {
810         .dev = dev
811     }, *val;
812     uint32_t hash = dev;
813     VariLenAffix affix;
814 
815     val = qht_lookup(&pdu->s->qpd_table, &lookup, hash);
816     if (!val) {
817         val = g_new0(QpdEntry, 1);
818         *val = lookup;
819         affix = affixForIndex(pdu->s->qp_affix_next);
820         val->prefix_bits = affix.bits;
821         qht_insert(&pdu->s->qpd_table, val, hash, NULL);
822         pdu->s->qp_ndevices++;
823     }
824     return val->prefix_bits;
825 }
826 
827 /*
828  * Slow / full mapping host inode nr -> guest inode nr.
829  *
830  * This function performs a slower and much more costly remapping of an
831  * original file inode number on host to an appropriate different inode
832  * number on guest. For every (dev, inode) combination on host a new
833  * sequential number is generated, cached and exposed as inode number on
834  * guest.
835  *
836  * This is just a "last resort" fallback solution if the much faster/cheaper
837  * qid_path_suffixmap() failed. In practice this slow / full mapping is not
838  * expected ever to be used at all though.
839  *
840  * See qid_path_suffixmap() for details
841  *
842  */
843 static int qid_path_fullmap(V9fsPDU *pdu, const struct stat *stbuf,
844                             uint64_t *path)
845 {
846     QpfEntry lookup = {
847         .dev = stbuf->st_dev,
848         .ino = stbuf->st_ino
849     }, *val;
850     uint32_t hash = qpf_hash(lookup);
851     VariLenAffix affix;
852 
853     val = qht_lookup(&pdu->s->qpf_table, &lookup, hash);
854 
855     if (!val) {
856         if (pdu->s->qp_fullpath_next == 0) {
857             /* no more files can be mapped :'( */
858             error_report_once(
859                 "9p: No more prefixes available for remapping inodes from "
860                 "host to guest."
861             );
862             return -ENFILE;
863         }
864 
865         val = g_new0(QpfEntry, 1);
866         *val = lookup;
867 
868         /* new unique inode and device combo */
869         affix = affixForIndex(
870             1ULL << (sizeof(pdu->s->qp_affix_next) * 8)
871         );
872         val->path = (pdu->s->qp_fullpath_next++ << affix.bits) | affix.value;
873         pdu->s->qp_fullpath_next &= ((1ULL << (64 - affix.bits)) - 1);
874         qht_insert(&pdu->s->qpf_table, val, hash, NULL);
875     }
876 
877     *path = val->path;
878     return 0;
879 }
880 
881 /*
882  * Quick mapping host inode nr -> guest inode nr.
883  *
884  * This function performs quick remapping of an original file inode number
885  * on host to an appropriate different inode number on guest. This remapping
886  * of inodes is required to avoid inode nr collisions on guest which would
887  * happen if the 9p export contains more than 1 exported file system (or
888  * more than 1 file system data set), because unlike on host level where the
889  * files would have different device nrs, all files exported by 9p would
890  * share the same device nr on guest (the device nr of the virtual 9p device
891  * that is).
892  *
893  * Inode remapping is performed by chopping off high end bits of the original
894  * inode number from host, shifting the result upwards and then assigning a
895  * generated suffix number for the low end bits, where the same suffix number
896  * will be shared by all inodes with the same device id AND the same high end
897  * bits that have been chopped off. That approach utilizes the fact that inode
898  * numbers very likely share the same high end bits (i.e. due to their common
899  * sequential generation by file systems) and hence we only have to generate
900  * and track a very limited amount of suffixes in practice due to that.
901  *
902  * We generate variable size suffixes for that purpose. The 1st generated
903  * suffix will only have 1 bit and hence we only need to chop off 1 bit from
904  * the original inode number. The subsequent suffixes being generated will
905  * grow in (bit) size subsequently, i.e. the 2nd and 3rd suffix being
906  * generated will have 3 bits and hence we have to chop off 3 bits from their
907  * original inodes, and so on. That approach of using variable length suffixes
908  * (i.e. over fixed size ones) utilizes the fact that in practice only a very
909  * limited amount of devices are shared by the same export (e.g. typically
910  * less than 2 dozen devices per 9p export), so in practice we need to chop
911  * off less bits than with fixed size prefixes and yet are flexible to add
912  * new devices at runtime below host's export directory at any time without
913  * having to reboot guest nor requiring to reconfigure guest for that. And due
914  * to the very limited amount of original high end bits that we chop off that
915  * way, the total amount of suffixes we need to generate is less than by using
916  * fixed size prefixes and hence it also improves performance of the inode
917  * remapping algorithm, and finally has the nice side effect that the inode
918  * numbers on guest will be much smaller & human friendly. ;-)
919  */
920 static int qid_path_suffixmap(V9fsPDU *pdu, const struct stat *stbuf,
921                               uint64_t *path)
922 {
923     const int ino_hash_bits = qid_inode_prefix_hash_bits(pdu, stbuf->st_dev);
924     QppEntry lookup = {
925         .dev = stbuf->st_dev,
926         .ino_prefix = (uint16_t) (stbuf->st_ino >> (64 - ino_hash_bits))
927     }, *val;
928     uint32_t hash = qpp_hash(lookup);
929 
930     val = qht_lookup(&pdu->s->qpp_table, &lookup, hash);
931 
932     if (!val) {
933         if (pdu->s->qp_affix_next == 0) {
934             /* we ran out of affixes */
935             warn_report_once(
936                 "9p: Potential degraded performance of inode remapping"
937             );
938             return -ENFILE;
939         }
940 
941         val = g_new0(QppEntry, 1);
942         *val = lookup;
943 
944         /* new unique inode affix and device combo */
945         val->qp_affix_index = pdu->s->qp_affix_next++;
946         val->qp_affix = affixForIndex(val->qp_affix_index);
947         qht_insert(&pdu->s->qpp_table, val, hash, NULL);
948     }
949     /* assuming generated affix to be suffix type, not prefix */
950     *path = (stbuf->st_ino << val->qp_affix.bits) | val->qp_affix.value;
951     return 0;
952 }
953 
954 static int stat_to_qid(V9fsPDU *pdu, const struct stat *stbuf, V9fsQID *qidp)
955 {
956     int err;
957     size_t size;
958 
959     if (pdu->s->ctx.export_flags & V9FS_REMAP_INODES) {
960         /* map inode+device to qid path (fast path) */
961         err = qid_path_suffixmap(pdu, stbuf, &qidp->path);
962         if (err == -ENFILE) {
963             /* fast path didn't work, fall back to full map */
964             err = qid_path_fullmap(pdu, stbuf, &qidp->path);
965         }
966         if (err) {
967             return err;
968         }
969     } else {
970         if (pdu->s->dev_id != stbuf->st_dev) {
971             if (pdu->s->ctx.export_flags & V9FS_FORBID_MULTIDEVS) {
972                 error_report_once(
973                     "9p: Multiple devices detected in same VirtFS export. "
974                     "Access of guest to additional devices is (partly) "
975                     "denied due to virtfs option 'multidevs=forbid' being "
976                     "effective."
977                 );
978                 return -ENODEV;
979             } else {
980                 warn_report_once(
981                     "9p: Multiple devices detected in same VirtFS export, "
982                     "which might lead to file ID collisions and severe "
983                     "misbehaviours on guest! You should either use a "
984                     "separate export for each device shared from host or "
985                     "use virtfs option 'multidevs=remap'!"
986                 );
987             }
988         }
989         memset(&qidp->path, 0, sizeof(qidp->path));
990         size = MIN(sizeof(stbuf->st_ino), sizeof(qidp->path));
991         memcpy(&qidp->path, &stbuf->st_ino, size);
992     }
993 
994     qidp->version = stbuf->st_mtime ^ (stbuf->st_size << 8);
995     qidp->type = 0;
996     if (S_ISDIR(stbuf->st_mode)) {
997         qidp->type |= P9_QID_TYPE_DIR;
998     }
999     if (S_ISLNK(stbuf->st_mode)) {
1000         qidp->type |= P9_QID_TYPE_SYMLINK;
1001     }
1002 
1003     return 0;
1004 }
1005 
1006 V9fsPDU *pdu_alloc(V9fsState *s)
1007 {
1008     V9fsPDU *pdu = NULL;
1009 
1010     if (!QLIST_EMPTY(&s->free_list)) {
1011         pdu = QLIST_FIRST(&s->free_list);
1012         QLIST_REMOVE(pdu, next);
1013         QLIST_INSERT_HEAD(&s->active_list, pdu, next);
1014     }
1015     return pdu;
1016 }
1017 
1018 void pdu_free(V9fsPDU *pdu)
1019 {
1020     V9fsState *s = pdu->s;
1021 
1022     g_assert(!pdu->cancelled);
1023     QLIST_REMOVE(pdu, next);
1024     QLIST_INSERT_HEAD(&s->free_list, pdu, next);
1025 }
1026 
1027 static void coroutine_fn pdu_complete(V9fsPDU *pdu, ssize_t len)
1028 {
1029     int8_t id = pdu->id + 1; /* Response */
1030     V9fsState *s = pdu->s;
1031     int ret;
1032 
1033     /*
1034      * The 9p spec requires that successfully cancelled pdus receive no reply.
1035      * Sending a reply would confuse clients because they would
1036      * assume that any EINTR is the actual result of the operation,
1037      * rather than a consequence of the cancellation. However, if
1038      * the operation completed (successfully or with an error other
1039      * than caused be cancellation), we do send out that reply, both
1040      * for efficiency and to avoid confusing the rest of the state machine
1041      * that assumes passing a non-error here will mean a successful
1042      * transmission of the reply.
1043      */
1044     bool discard = pdu->cancelled && len == -EINTR;
1045     if (discard) {
1046         trace_v9fs_rcancel(pdu->tag, pdu->id);
1047         pdu->size = 0;
1048         goto out_notify;
1049     }
1050 
1051     if (len < 0) {
1052         int err = -len;
1053         len = 7;
1054 
1055         if (s->proto_version != V9FS_PROTO_2000L) {
1056             V9fsString str;
1057 
1058             str.data = strerror(err);
1059             str.size = strlen(str.data);
1060 
1061             ret = pdu_marshal(pdu, len, "s", &str);
1062             if (ret < 0) {
1063                 goto out_notify;
1064             }
1065             len += ret;
1066             id = P9_RERROR;
1067         } else {
1068             err = errno_to_dotl(err);
1069         }
1070 
1071         ret = pdu_marshal(pdu, len, "d", err);
1072         if (ret < 0) {
1073             goto out_notify;
1074         }
1075         len += ret;
1076 
1077         if (s->proto_version == V9FS_PROTO_2000L) {
1078             id = P9_RLERROR;
1079         }
1080         trace_v9fs_rerror(pdu->tag, pdu->id, err); /* Trace ERROR */
1081     }
1082 
1083     /* fill out the header */
1084     if (pdu_marshal(pdu, 0, "dbw", (int32_t)len, id, pdu->tag) < 0) {
1085         goto out_notify;
1086     }
1087 
1088     /* keep these in sync */
1089     pdu->size = len;
1090     pdu->id = id;
1091 
1092 out_notify:
1093     pdu->s->transport->push_and_notify(pdu);
1094 
1095     /* Now wakeup anybody waiting in flush for this request */
1096     if (!qemu_co_queue_next(&pdu->complete)) {
1097         pdu_free(pdu);
1098     }
1099 }
1100 
1101 static mode_t v9mode_to_mode(uint32_t mode, V9fsString *extension)
1102 {
1103     mode_t ret;
1104 
1105     ret = mode & 0777;
1106     if (mode & P9_STAT_MODE_DIR) {
1107         ret |= S_IFDIR;
1108     }
1109 
1110     if (mode & P9_STAT_MODE_SYMLINK) {
1111         ret |= S_IFLNK;
1112     }
1113     if (mode & P9_STAT_MODE_SOCKET) {
1114         ret |= S_IFSOCK;
1115     }
1116     if (mode & P9_STAT_MODE_NAMED_PIPE) {
1117         ret |= S_IFIFO;
1118     }
1119     if (mode & P9_STAT_MODE_DEVICE) {
1120         if (extension->size && extension->data[0] == 'c') {
1121             ret |= S_IFCHR;
1122         } else {
1123             ret |= S_IFBLK;
1124         }
1125     }
1126 
1127     if (!(ret & ~0777)) {
1128         ret |= S_IFREG;
1129     }
1130 
1131     if (mode & P9_STAT_MODE_SETUID) {
1132         ret |= S_ISUID;
1133     }
1134     if (mode & P9_STAT_MODE_SETGID) {
1135         ret |= S_ISGID;
1136     }
1137     if (mode & P9_STAT_MODE_SETVTX) {
1138         ret |= S_ISVTX;
1139     }
1140 
1141     return ret;
1142 }
1143 
1144 static int donttouch_stat(V9fsStat *stat)
1145 {
1146     if (stat->type == -1 &&
1147         stat->dev == -1 &&
1148         stat->qid.type == 0xff &&
1149         stat->qid.version == (uint32_t) -1 &&
1150         stat->qid.path == (uint64_t) -1 &&
1151         stat->mode == -1 &&
1152         stat->atime == -1 &&
1153         stat->mtime == -1 &&
1154         stat->length == -1 &&
1155         !stat->name.size &&
1156         !stat->uid.size &&
1157         !stat->gid.size &&
1158         !stat->muid.size &&
1159         stat->n_uid == -1 &&
1160         stat->n_gid == -1 &&
1161         stat->n_muid == -1) {
1162         return 1;
1163     }
1164 
1165     return 0;
1166 }
1167 
1168 static void v9fs_stat_init(V9fsStat *stat)
1169 {
1170     v9fs_string_init(&stat->name);
1171     v9fs_string_init(&stat->uid);
1172     v9fs_string_init(&stat->gid);
1173     v9fs_string_init(&stat->muid);
1174     v9fs_string_init(&stat->extension);
1175 }
1176 
1177 static void v9fs_stat_free(V9fsStat *stat)
1178 {
1179     v9fs_string_free(&stat->name);
1180     v9fs_string_free(&stat->uid);
1181     v9fs_string_free(&stat->gid);
1182     v9fs_string_free(&stat->muid);
1183     v9fs_string_free(&stat->extension);
1184 }
1185 
1186 static uint32_t stat_to_v9mode(const struct stat *stbuf)
1187 {
1188     uint32_t mode;
1189 
1190     mode = stbuf->st_mode & 0777;
1191     if (S_ISDIR(stbuf->st_mode)) {
1192         mode |= P9_STAT_MODE_DIR;
1193     }
1194 
1195     if (S_ISLNK(stbuf->st_mode)) {
1196         mode |= P9_STAT_MODE_SYMLINK;
1197     }
1198 
1199     if (S_ISSOCK(stbuf->st_mode)) {
1200         mode |= P9_STAT_MODE_SOCKET;
1201     }
1202 
1203     if (S_ISFIFO(stbuf->st_mode)) {
1204         mode |= P9_STAT_MODE_NAMED_PIPE;
1205     }
1206 
1207     if (S_ISBLK(stbuf->st_mode) || S_ISCHR(stbuf->st_mode)) {
1208         mode |= P9_STAT_MODE_DEVICE;
1209     }
1210 
1211     if (stbuf->st_mode & S_ISUID) {
1212         mode |= P9_STAT_MODE_SETUID;
1213     }
1214 
1215     if (stbuf->st_mode & S_ISGID) {
1216         mode |= P9_STAT_MODE_SETGID;
1217     }
1218 
1219     if (stbuf->st_mode & S_ISVTX) {
1220         mode |= P9_STAT_MODE_SETVTX;
1221     }
1222 
1223     return mode;
1224 }
1225 
1226 static int coroutine_fn stat_to_v9stat(V9fsPDU *pdu, V9fsPath *path,
1227                                        const char *basename,
1228                                        const struct stat *stbuf,
1229                                        V9fsStat *v9stat)
1230 {
1231     int err;
1232 
1233     memset(v9stat, 0, sizeof(*v9stat));
1234 
1235     err = stat_to_qid(pdu, stbuf, &v9stat->qid);
1236     if (err < 0) {
1237         return err;
1238     }
1239     v9stat->mode = stat_to_v9mode(stbuf);
1240     v9stat->atime = stbuf->st_atime;
1241     v9stat->mtime = stbuf->st_mtime;
1242     v9stat->length = stbuf->st_size;
1243 
1244     v9fs_string_free(&v9stat->uid);
1245     v9fs_string_free(&v9stat->gid);
1246     v9fs_string_free(&v9stat->muid);
1247 
1248     v9stat->n_uid = stbuf->st_uid;
1249     v9stat->n_gid = stbuf->st_gid;
1250     v9stat->n_muid = 0;
1251 
1252     v9fs_string_free(&v9stat->extension);
1253 
1254     if (v9stat->mode & P9_STAT_MODE_SYMLINK) {
1255         err = v9fs_co_readlink(pdu, path, &v9stat->extension);
1256         if (err < 0) {
1257             return err;
1258         }
1259     } else if (v9stat->mode & P9_STAT_MODE_DEVICE) {
1260         v9fs_string_sprintf(&v9stat->extension, "%c %u %u",
1261                 S_ISCHR(stbuf->st_mode) ? 'c' : 'b',
1262                 major(stbuf->st_rdev), minor(stbuf->st_rdev));
1263     } else if (S_ISDIR(stbuf->st_mode) || S_ISREG(stbuf->st_mode)) {
1264         v9fs_string_sprintf(&v9stat->extension, "%s %lu",
1265                 "HARDLINKCOUNT", (unsigned long)stbuf->st_nlink);
1266     }
1267 
1268     v9fs_string_sprintf(&v9stat->name, "%s", basename);
1269 
1270     v9stat->size = 61 +
1271         v9fs_string_size(&v9stat->name) +
1272         v9fs_string_size(&v9stat->uid) +
1273         v9fs_string_size(&v9stat->gid) +
1274         v9fs_string_size(&v9stat->muid) +
1275         v9fs_string_size(&v9stat->extension);
1276     return 0;
1277 }
1278 
1279 #define P9_STATS_MODE          0x00000001ULL
1280 #define P9_STATS_NLINK         0x00000002ULL
1281 #define P9_STATS_UID           0x00000004ULL
1282 #define P9_STATS_GID           0x00000008ULL
1283 #define P9_STATS_RDEV          0x00000010ULL
1284 #define P9_STATS_ATIME         0x00000020ULL
1285 #define P9_STATS_MTIME         0x00000040ULL
1286 #define P9_STATS_CTIME         0x00000080ULL
1287 #define P9_STATS_INO           0x00000100ULL
1288 #define P9_STATS_SIZE          0x00000200ULL
1289 #define P9_STATS_BLOCKS        0x00000400ULL
1290 
1291 #define P9_STATS_BTIME         0x00000800ULL
1292 #define P9_STATS_GEN           0x00001000ULL
1293 #define P9_STATS_DATA_VERSION  0x00002000ULL
1294 
1295 #define P9_STATS_BASIC         0x000007ffULL /* Mask for fields up to BLOCKS */
1296 #define P9_STATS_ALL           0x00003fffULL /* Mask for All fields above */
1297 
1298 
1299 /**
1300  * blksize_to_iounit() - Block size exposed to 9p client.
1301  * Return: block size
1302  *
1303  * @pdu: 9p client request
1304  * @blksize: host filesystem's block size
1305  *
1306  * Convert host filesystem's block size into an appropriate block size for
1307  * 9p client (guest OS side). The value returned suggests an "optimum" block
1308  * size for 9p I/O, i.e. to maximize performance.
1309  */
1310 static int32_t blksize_to_iounit(const V9fsPDU *pdu, int32_t blksize)
1311 {
1312     int32_t iounit = 0;
1313     V9fsState *s = pdu->s;
1314 
1315     /*
1316      * iounit should be multiples of blksize (host filesystem block size)
1317      * as well as less than (client msize - P9_IOHDRSZ)
1318      */
1319     if (blksize) {
1320         iounit = QEMU_ALIGN_DOWN(s->msize - P9_IOHDRSZ, blksize);
1321     }
1322     if (!iounit) {
1323         iounit = s->msize - P9_IOHDRSZ;
1324     }
1325     return iounit;
1326 }
1327 
1328 static int32_t stat_to_iounit(const V9fsPDU *pdu, const struct stat *stbuf)
1329 {
1330     return blksize_to_iounit(pdu, stbuf->st_blksize);
1331 }
1332 
1333 static int stat_to_v9stat_dotl(V9fsPDU *pdu, const struct stat *stbuf,
1334                                 V9fsStatDotl *v9lstat)
1335 {
1336     memset(v9lstat, 0, sizeof(*v9lstat));
1337 
1338     v9lstat->st_mode = stbuf->st_mode;
1339     v9lstat->st_nlink = stbuf->st_nlink;
1340     v9lstat->st_uid = stbuf->st_uid;
1341     v9lstat->st_gid = stbuf->st_gid;
1342     v9lstat->st_rdev = host_dev_to_dotl_dev(stbuf->st_rdev);
1343     v9lstat->st_size = stbuf->st_size;
1344     v9lstat->st_blksize = stat_to_iounit(pdu, stbuf);
1345     v9lstat->st_blocks = stbuf->st_blocks;
1346     v9lstat->st_atime_sec = stbuf->st_atime;
1347     v9lstat->st_mtime_sec = stbuf->st_mtime;
1348     v9lstat->st_ctime_sec = stbuf->st_ctime;
1349 #ifdef CONFIG_DARWIN
1350     v9lstat->st_atime_nsec = stbuf->st_atimespec.tv_nsec;
1351     v9lstat->st_mtime_nsec = stbuf->st_mtimespec.tv_nsec;
1352     v9lstat->st_ctime_nsec = stbuf->st_ctimespec.tv_nsec;
1353 #else
1354     v9lstat->st_atime_nsec = stbuf->st_atim.tv_nsec;
1355     v9lstat->st_mtime_nsec = stbuf->st_mtim.tv_nsec;
1356     v9lstat->st_ctime_nsec = stbuf->st_ctim.tv_nsec;
1357 #endif
1358     /* Currently we only support BASIC fields in stat */
1359     v9lstat->st_result_mask = P9_STATS_BASIC;
1360 
1361     return stat_to_qid(pdu, stbuf, &v9lstat->qid);
1362 }
1363 
1364 static void print_sg(struct iovec *sg, int cnt)
1365 {
1366     int i;
1367 
1368     printf("sg[%d]: {", cnt);
1369     for (i = 0; i < cnt; i++) {
1370         if (i) {
1371             printf(", ");
1372         }
1373         printf("(%p, %zd)", sg[i].iov_base, sg[i].iov_len);
1374     }
1375     printf("}\n");
1376 }
1377 
1378 /* Will call this only for path name based fid */
1379 static void v9fs_fix_path(V9fsPath *dst, V9fsPath *src, int len)
1380 {
1381     V9fsPath str;
1382     v9fs_path_init(&str);
1383     v9fs_path_copy(&str, dst);
1384     v9fs_path_sprintf(dst, "%s%s", src->data, str.data + len);
1385     v9fs_path_free(&str);
1386 }
1387 
1388 static inline bool is_ro_export(FsContext *ctx)
1389 {
1390     return ctx->export_flags & V9FS_RDONLY;
1391 }
1392 
1393 static void coroutine_fn v9fs_version(void *opaque)
1394 {
1395     ssize_t err;
1396     V9fsPDU *pdu = opaque;
1397     V9fsState *s = pdu->s;
1398     V9fsString version;
1399     size_t offset = 7;
1400 
1401     v9fs_string_init(&version);
1402     err = pdu_unmarshal(pdu, offset, "ds", &s->msize, &version);
1403     if (err < 0) {
1404         goto out;
1405     }
1406     trace_v9fs_version(pdu->tag, pdu->id, s->msize, version.data);
1407 
1408     virtfs_reset(pdu);
1409 
1410     if (!strcmp(version.data, "9P2000.u")) {
1411         s->proto_version = V9FS_PROTO_2000U;
1412     } else if (!strcmp(version.data, "9P2000.L")) {
1413         s->proto_version = V9FS_PROTO_2000L;
1414     } else {
1415         v9fs_string_sprintf(&version, "unknown");
1416         /* skip min. msize check, reporting invalid version has priority */
1417         goto marshal;
1418     }
1419 
1420     if (s->msize < P9_MIN_MSIZE) {
1421         err = -EMSGSIZE;
1422         error_report(
1423             "9pfs: Client requested msize < minimum msize ("
1424             stringify(P9_MIN_MSIZE) ") supported by this server."
1425         );
1426         goto out;
1427     }
1428 
1429     /* 8192 is the default msize of Linux clients */
1430     if (s->msize <= 8192 && !(s->ctx.export_flags & V9FS_NO_PERF_WARN)) {
1431         warn_report_once(
1432             "9p: degraded performance: a reasonable high msize should be "
1433             "chosen on client/guest side (chosen msize is <= 8192). See "
1434             "https://wiki.qemu.org/Documentation/9psetup#msize for details."
1435         );
1436     }
1437 
1438 marshal:
1439     err = pdu_marshal(pdu, offset, "ds", s->msize, &version);
1440     if (err < 0) {
1441         goto out;
1442     }
1443     err += offset;
1444     trace_v9fs_version_return(pdu->tag, pdu->id, s->msize, version.data);
1445 out:
1446     pdu_complete(pdu, err);
1447     v9fs_string_free(&version);
1448 }
1449 
1450 static void coroutine_fn v9fs_attach(void *opaque)
1451 {
1452     V9fsPDU *pdu = opaque;
1453     V9fsState *s = pdu->s;
1454     int32_t fid, afid, n_uname;
1455     V9fsString uname, aname;
1456     V9fsFidState *fidp;
1457     size_t offset = 7;
1458     V9fsQID qid;
1459     ssize_t err;
1460     struct stat stbuf;
1461 
1462     v9fs_string_init(&uname);
1463     v9fs_string_init(&aname);
1464     err = pdu_unmarshal(pdu, offset, "ddssd", &fid,
1465                         &afid, &uname, &aname, &n_uname);
1466     if (err < 0) {
1467         goto out_nofid;
1468     }
1469     trace_v9fs_attach(pdu->tag, pdu->id, fid, afid, uname.data, aname.data);
1470 
1471     fidp = alloc_fid(s, fid);
1472     if (fidp == NULL) {
1473         err = -EINVAL;
1474         goto out_nofid;
1475     }
1476     fidp->uid = n_uname;
1477     err = v9fs_co_name_to_path(pdu, NULL, "/", &fidp->path);
1478     if (err < 0) {
1479         err = -EINVAL;
1480         clunk_fid(s, fid);
1481         goto out;
1482     }
1483     err = v9fs_co_lstat(pdu, &fidp->path, &stbuf);
1484     if (err < 0) {
1485         err = -EINVAL;
1486         clunk_fid(s, fid);
1487         goto out;
1488     }
1489     err = stat_to_qid(pdu, &stbuf, &qid);
1490     if (err < 0) {
1491         err = -EINVAL;
1492         clunk_fid(s, fid);
1493         goto out;
1494     }
1495 
1496     /*
1497      * disable migration if we haven't done already.
1498      * attach could get called multiple times for the same export.
1499      */
1500     if (!s->migration_blocker) {
1501         error_setg(&s->migration_blocker,
1502                    "Migration is disabled when VirtFS export path '%s' is mounted in the guest using mount_tag '%s'",
1503                    s->ctx.fs_root ? s->ctx.fs_root : "NULL", s->tag);
1504         err = migrate_add_blocker(&s->migration_blocker, NULL);
1505         if (err < 0) {
1506             clunk_fid(s, fid);
1507             goto out;
1508         }
1509         s->root_fid = fid;
1510     }
1511 
1512     err = pdu_marshal(pdu, offset, "Q", &qid);
1513     if (err < 0) {
1514         clunk_fid(s, fid);
1515         goto out;
1516     }
1517     err += offset;
1518 
1519     memcpy(&s->root_st, &stbuf, sizeof(stbuf));
1520     trace_v9fs_attach_return(pdu->tag, pdu->id,
1521                              qid.type, qid.version, qid.path);
1522 out:
1523     put_fid(pdu, fidp);
1524 out_nofid:
1525     pdu_complete(pdu, err);
1526     v9fs_string_free(&uname);
1527     v9fs_string_free(&aname);
1528 }
1529 
1530 static void coroutine_fn v9fs_stat(void *opaque)
1531 {
1532     int32_t fid;
1533     V9fsStat v9stat;
1534     ssize_t err = 0;
1535     size_t offset = 7;
1536     struct stat stbuf;
1537     V9fsFidState *fidp;
1538     V9fsPDU *pdu = opaque;
1539     char *basename;
1540 
1541     err = pdu_unmarshal(pdu, offset, "d", &fid);
1542     if (err < 0) {
1543         goto out_nofid;
1544     }
1545     trace_v9fs_stat(pdu->tag, pdu->id, fid);
1546 
1547     fidp = get_fid(pdu, fid);
1548     if (fidp == NULL) {
1549         err = -ENOENT;
1550         goto out_nofid;
1551     }
1552     err = v9fs_co_lstat(pdu, &fidp->path, &stbuf);
1553     if (err < 0) {
1554         goto out;
1555     }
1556     basename = g_path_get_basename(fidp->path.data);
1557     err = stat_to_v9stat(pdu, &fidp->path, basename, &stbuf, &v9stat);
1558     g_free(basename);
1559     if (err < 0) {
1560         goto out;
1561     }
1562     err = pdu_marshal(pdu, offset, "wS", 0, &v9stat);
1563     if (err < 0) {
1564         v9fs_stat_free(&v9stat);
1565         goto out;
1566     }
1567     trace_v9fs_stat_return(pdu->tag, pdu->id, v9stat.mode,
1568                            v9stat.atime, v9stat.mtime, v9stat.length);
1569     err += offset;
1570     v9fs_stat_free(&v9stat);
1571 out:
1572     put_fid(pdu, fidp);
1573 out_nofid:
1574     pdu_complete(pdu, err);
1575 }
1576 
1577 static void coroutine_fn v9fs_getattr(void *opaque)
1578 {
1579     int32_t fid;
1580     size_t offset = 7;
1581     ssize_t retval = 0;
1582     struct stat stbuf;
1583     V9fsFidState *fidp;
1584     uint64_t request_mask;
1585     V9fsStatDotl v9stat_dotl;
1586     V9fsPDU *pdu = opaque;
1587 
1588     retval = pdu_unmarshal(pdu, offset, "dq", &fid, &request_mask);
1589     if (retval < 0) {
1590         goto out_nofid;
1591     }
1592     trace_v9fs_getattr(pdu->tag, pdu->id, fid, request_mask);
1593 
1594     fidp = get_fid(pdu, fid);
1595     if (fidp == NULL) {
1596         retval = -ENOENT;
1597         goto out_nofid;
1598     }
1599     if ((fidp->fid_type == P9_FID_FILE && fidp->fs.fd != -1) ||
1600         (fidp->fid_type == P9_FID_DIR && fidp->fs.dir.stream))
1601     {
1602         retval = v9fs_co_fstat(pdu, fidp, &stbuf);
1603     } else {
1604         retval = v9fs_co_lstat(pdu, &fidp->path, &stbuf);
1605     }
1606     if (retval < 0) {
1607         goto out;
1608     }
1609     retval = stat_to_v9stat_dotl(pdu, &stbuf, &v9stat_dotl);
1610     if (retval < 0) {
1611         goto out;
1612     }
1613 
1614     /*  fill st_gen if requested and supported by underlying fs */
1615     if (request_mask & P9_STATS_GEN) {
1616         retval = v9fs_co_st_gen(pdu, &fidp->path, stbuf.st_mode, &v9stat_dotl);
1617         switch (retval) {
1618         case 0:
1619             /* we have valid st_gen: update result mask */
1620             v9stat_dotl.st_result_mask |= P9_STATS_GEN;
1621             break;
1622         case -EINTR:
1623             /* request cancelled, e.g. by Tflush */
1624             goto out;
1625         default:
1626             /* failed to get st_gen: not fatal, ignore */
1627             break;
1628         }
1629     }
1630     retval = pdu_marshal(pdu, offset, "A", &v9stat_dotl);
1631     if (retval < 0) {
1632         goto out;
1633     }
1634     retval += offset;
1635     trace_v9fs_getattr_return(pdu->tag, pdu->id, v9stat_dotl.st_result_mask,
1636                               v9stat_dotl.st_mode, v9stat_dotl.st_uid,
1637                               v9stat_dotl.st_gid);
1638 out:
1639     put_fid(pdu, fidp);
1640 out_nofid:
1641     pdu_complete(pdu, retval);
1642 }
1643 
1644 /* Attribute flags */
1645 #define P9_ATTR_MODE       (1 << 0)
1646 #define P9_ATTR_UID        (1 << 1)
1647 #define P9_ATTR_GID        (1 << 2)
1648 #define P9_ATTR_SIZE       (1 << 3)
1649 #define P9_ATTR_ATIME      (1 << 4)
1650 #define P9_ATTR_MTIME      (1 << 5)
1651 #define P9_ATTR_CTIME      (1 << 6)
1652 #define P9_ATTR_ATIME_SET  (1 << 7)
1653 #define P9_ATTR_MTIME_SET  (1 << 8)
1654 
1655 #define P9_ATTR_MASK    127
1656 
1657 static void coroutine_fn v9fs_setattr(void *opaque)
1658 {
1659     int err = 0;
1660     int32_t fid;
1661     V9fsFidState *fidp;
1662     size_t offset = 7;
1663     V9fsIattr v9iattr;
1664     V9fsPDU *pdu = opaque;
1665 
1666     err = pdu_unmarshal(pdu, offset, "dI", &fid, &v9iattr);
1667     if (err < 0) {
1668         goto out_nofid;
1669     }
1670 
1671     trace_v9fs_setattr(pdu->tag, pdu->id, fid,
1672                        v9iattr.valid, v9iattr.mode, v9iattr.uid, v9iattr.gid,
1673                        v9iattr.size, v9iattr.atime_sec, v9iattr.mtime_sec);
1674 
1675     fidp = get_fid(pdu, fid);
1676     if (fidp == NULL) {
1677         err = -EINVAL;
1678         goto out_nofid;
1679     }
1680     if (v9iattr.valid & P9_ATTR_MODE) {
1681         err = v9fs_co_chmod(pdu, &fidp->path, v9iattr.mode);
1682         if (err < 0) {
1683             goto out;
1684         }
1685     }
1686     if (v9iattr.valid & (P9_ATTR_ATIME | P9_ATTR_MTIME)) {
1687         struct timespec times[2];
1688         if (v9iattr.valid & P9_ATTR_ATIME) {
1689             if (v9iattr.valid & P9_ATTR_ATIME_SET) {
1690                 times[0].tv_sec = v9iattr.atime_sec;
1691                 times[0].tv_nsec = v9iattr.atime_nsec;
1692             } else {
1693                 times[0].tv_nsec = UTIME_NOW;
1694             }
1695         } else {
1696             times[0].tv_nsec = UTIME_OMIT;
1697         }
1698         if (v9iattr.valid & P9_ATTR_MTIME) {
1699             if (v9iattr.valid & P9_ATTR_MTIME_SET) {
1700                 times[1].tv_sec = v9iattr.mtime_sec;
1701                 times[1].tv_nsec = v9iattr.mtime_nsec;
1702             } else {
1703                 times[1].tv_nsec = UTIME_NOW;
1704             }
1705         } else {
1706             times[1].tv_nsec = UTIME_OMIT;
1707         }
1708         err = v9fs_co_utimensat(pdu, &fidp->path, times);
1709         if (err < 0) {
1710             goto out;
1711         }
1712     }
1713     /*
1714      * If the only valid entry in iattr is ctime we can call
1715      * chown(-1,-1) to update the ctime of the file
1716      */
1717     if ((v9iattr.valid & (P9_ATTR_UID | P9_ATTR_GID)) ||
1718         ((v9iattr.valid & P9_ATTR_CTIME)
1719          && !((v9iattr.valid & P9_ATTR_MASK) & ~P9_ATTR_CTIME))) {
1720         if (!(v9iattr.valid & P9_ATTR_UID)) {
1721             v9iattr.uid = -1;
1722         }
1723         if (!(v9iattr.valid & P9_ATTR_GID)) {
1724             v9iattr.gid = -1;
1725         }
1726         err = v9fs_co_chown(pdu, &fidp->path, v9iattr.uid,
1727                             v9iattr.gid);
1728         if (err < 0) {
1729             goto out;
1730         }
1731     }
1732     if (v9iattr.valid & (P9_ATTR_SIZE)) {
1733         err = v9fs_co_truncate(pdu, &fidp->path, v9iattr.size);
1734         if (err < 0) {
1735             goto out;
1736         }
1737     }
1738     err = offset;
1739     trace_v9fs_setattr_return(pdu->tag, pdu->id);
1740 out:
1741     put_fid(pdu, fidp);
1742 out_nofid:
1743     pdu_complete(pdu, err);
1744 }
1745 
1746 static int v9fs_walk_marshal(V9fsPDU *pdu, uint16_t nwnames, V9fsQID *qids)
1747 {
1748     int i;
1749     ssize_t err;
1750     size_t offset = 7;
1751 
1752     err = pdu_marshal(pdu, offset, "w", nwnames);
1753     if (err < 0) {
1754         return err;
1755     }
1756     offset += err;
1757     for (i = 0; i < nwnames; i++) {
1758         err = pdu_marshal(pdu, offset, "Q", &qids[i]);
1759         if (err < 0) {
1760             return err;
1761         }
1762         offset += err;
1763     }
1764     return offset;
1765 }
1766 
1767 static bool name_is_illegal(const char *name)
1768 {
1769     return !*name || strchr(name, '/') != NULL;
1770 }
1771 
1772 static bool same_stat_id(const struct stat *a, const struct stat *b)
1773 {
1774     return a->st_dev == b->st_dev && a->st_ino == b->st_ino;
1775 }
1776 
1777 static void coroutine_fn v9fs_walk(void *opaque)
1778 {
1779     int name_idx, nwalked;
1780     g_autofree V9fsQID *qids = NULL;
1781     int i, err = 0, any_err = 0;
1782     V9fsPath dpath, path;
1783     P9ARRAY_REF(V9fsPath) pathes = NULL;
1784     uint16_t nwnames;
1785     struct stat stbuf, fidst;
1786     g_autofree struct stat *stbufs = NULL;
1787     size_t offset = 7;
1788     int32_t fid, newfid;
1789     P9ARRAY_REF(V9fsString) wnames = NULL;
1790     V9fsFidState *fidp;
1791     V9fsFidState *newfidp = NULL;
1792     V9fsPDU *pdu = opaque;
1793     V9fsState *s = pdu->s;
1794     V9fsQID qid;
1795 
1796     err = pdu_unmarshal(pdu, offset, "ddw", &fid, &newfid, &nwnames);
1797     if (err < 0) {
1798         pdu_complete(pdu, err);
1799         return;
1800     }
1801     offset += err;
1802 
1803     trace_v9fs_walk(pdu->tag, pdu->id, fid, newfid, nwnames);
1804 
1805     if (nwnames > P9_MAXWELEM) {
1806         err = -EINVAL;
1807         goto out_nofid;
1808     }
1809     if (nwnames) {
1810         P9ARRAY_NEW(V9fsString, wnames, nwnames);
1811         qids   = g_new0(V9fsQID, nwnames);
1812         stbufs = g_new0(struct stat, nwnames);
1813         P9ARRAY_NEW(V9fsPath, pathes, nwnames);
1814         for (i = 0; i < nwnames; i++) {
1815             err = pdu_unmarshal(pdu, offset, "s", &wnames[i]);
1816             if (err < 0) {
1817                 goto out_nofid;
1818             }
1819             if (name_is_illegal(wnames[i].data)) {
1820                 err = -ENOENT;
1821                 goto out_nofid;
1822             }
1823             offset += err;
1824         }
1825     }
1826     fidp = get_fid(pdu, fid);
1827     if (fidp == NULL) {
1828         err = -ENOENT;
1829         goto out_nofid;
1830     }
1831 
1832     v9fs_path_init(&dpath);
1833     v9fs_path_init(&path);
1834     /*
1835      * Both dpath and path initially point to fidp.
1836      * Needed to handle request with nwnames == 0
1837      */
1838     v9fs_path_copy(&dpath, &fidp->path);
1839     v9fs_path_copy(&path, &fidp->path);
1840 
1841     /*
1842      * To keep latency (i.e. overall execution time for processing this
1843      * Twalk client request) as small as possible, run all the required fs
1844      * driver code altogether inside the following block.
1845      */
1846     v9fs_co_run_in_worker({
1847         nwalked = 0;
1848         if (v9fs_request_cancelled(pdu)) {
1849             any_err |= err = -EINTR;
1850             break;
1851         }
1852         err = s->ops->lstat(&s->ctx, &dpath, &fidst);
1853         if (err < 0) {
1854             any_err |= err = -errno;
1855             break;
1856         }
1857         stbuf = fidst;
1858         for (; nwalked < nwnames; nwalked++) {
1859             if (v9fs_request_cancelled(pdu)) {
1860                 any_err |= err = -EINTR;
1861                 break;
1862             }
1863             if (!same_stat_id(&pdu->s->root_st, &stbuf) ||
1864                 strcmp("..", wnames[nwalked].data))
1865             {
1866                 err = s->ops->name_to_path(&s->ctx, &dpath,
1867                                            wnames[nwalked].data,
1868                                            &pathes[nwalked]);
1869                 if (err < 0) {
1870                     any_err |= err = -errno;
1871                     break;
1872                 }
1873                 if (v9fs_request_cancelled(pdu)) {
1874                     any_err |= err = -EINTR;
1875                     break;
1876                 }
1877                 err = s->ops->lstat(&s->ctx, &pathes[nwalked], &stbuf);
1878                 if (err < 0) {
1879                     any_err |= err = -errno;
1880                     break;
1881                 }
1882                 stbufs[nwalked] = stbuf;
1883                 v9fs_path_copy(&dpath, &pathes[nwalked]);
1884             }
1885         }
1886     });
1887     /*
1888      * Handle all the rest of this Twalk request on main thread ...
1889      *
1890      * NOTE: -EINTR is an exception where we deviate from the protocol spec
1891      * and simply send a (R)Lerror response instead of bothering to assemble
1892      * a (deducted) Rwalk response; because -EINTR is always the result of a
1893      * Tflush request, so client would no longer wait for a response in this
1894      * case anyway.
1895      */
1896     if ((err < 0 && !nwalked) || err == -EINTR) {
1897         goto out;
1898     }
1899 
1900     any_err |= err = stat_to_qid(pdu, &fidst, &qid);
1901     if (err < 0 && !nwalked) {
1902         goto out;
1903     }
1904     stbuf = fidst;
1905 
1906     /* reset dpath and path */
1907     v9fs_path_copy(&dpath, &fidp->path);
1908     v9fs_path_copy(&path, &fidp->path);
1909 
1910     for (name_idx = 0; name_idx < nwalked; name_idx++) {
1911         if (!same_stat_id(&pdu->s->root_st, &stbuf) ||
1912             strcmp("..", wnames[name_idx].data))
1913         {
1914             stbuf = stbufs[name_idx];
1915             any_err |= err = stat_to_qid(pdu, &stbuf, &qid);
1916             if (err < 0) {
1917                 break;
1918             }
1919             v9fs_path_copy(&path, &pathes[name_idx]);
1920             v9fs_path_copy(&dpath, &path);
1921         }
1922         memcpy(&qids[name_idx], &qid, sizeof(qid));
1923     }
1924     if (any_err < 0) {
1925         if (!name_idx) {
1926             /* don't send any QIDs, send Rlerror instead */
1927             goto out;
1928         } else {
1929             /* send QIDs (not Rlerror), but fid MUST remain unaffected */
1930             goto send_qids;
1931         }
1932     }
1933     if (fid == newfid) {
1934         if (fidp->fid_type != P9_FID_NONE) {
1935             err = -EINVAL;
1936             goto out;
1937         }
1938         v9fs_path_write_lock(s);
1939         v9fs_path_copy(&fidp->path, &path);
1940         v9fs_path_unlock(s);
1941     } else {
1942         newfidp = alloc_fid(s, newfid);
1943         if (newfidp == NULL) {
1944             err = -EINVAL;
1945             goto out;
1946         }
1947         newfidp->uid = fidp->uid;
1948         v9fs_path_copy(&newfidp->path, &path);
1949     }
1950 send_qids:
1951     err = v9fs_walk_marshal(pdu, name_idx, qids);
1952     trace_v9fs_walk_return(pdu->tag, pdu->id, name_idx, qids);
1953 out:
1954     put_fid(pdu, fidp);
1955     if (newfidp) {
1956         put_fid(pdu, newfidp);
1957     }
1958     v9fs_path_free(&dpath);
1959     v9fs_path_free(&path);
1960 out_nofid:
1961     pdu_complete(pdu, err);
1962 }
1963 
1964 static int32_t coroutine_fn get_iounit(V9fsPDU *pdu, V9fsPath *path)
1965 {
1966     struct statfs stbuf;
1967     int err = v9fs_co_statfs(pdu, path, &stbuf);
1968 
1969     return blksize_to_iounit(pdu, (err >= 0) ? stbuf.f_bsize : 0);
1970 }
1971 
1972 static void coroutine_fn v9fs_open(void *opaque)
1973 {
1974     int flags;
1975     int32_t fid;
1976     int32_t mode;
1977     V9fsQID qid;
1978     int iounit = 0;
1979     ssize_t err = 0;
1980     size_t offset = 7;
1981     struct stat stbuf;
1982     V9fsFidState *fidp;
1983     V9fsPDU *pdu = opaque;
1984     V9fsState *s = pdu->s;
1985 
1986     if (s->proto_version == V9FS_PROTO_2000L) {
1987         err = pdu_unmarshal(pdu, offset, "dd", &fid, &mode);
1988     } else {
1989         uint8_t modebyte;
1990         err = pdu_unmarshal(pdu, offset, "db", &fid, &modebyte);
1991         mode = modebyte;
1992     }
1993     if (err < 0) {
1994         goto out_nofid;
1995     }
1996     trace_v9fs_open(pdu->tag, pdu->id, fid, mode);
1997 
1998     fidp = get_fid(pdu, fid);
1999     if (fidp == NULL) {
2000         err = -ENOENT;
2001         goto out_nofid;
2002     }
2003     if (fidp->fid_type != P9_FID_NONE) {
2004         err = -EINVAL;
2005         goto out;
2006     }
2007 
2008     err = v9fs_co_lstat(pdu, &fidp->path, &stbuf);
2009     if (err < 0) {
2010         goto out;
2011     }
2012     err = stat_to_qid(pdu, &stbuf, &qid);
2013     if (err < 0) {
2014         goto out;
2015     }
2016     if (S_ISDIR(stbuf.st_mode)) {
2017         err = v9fs_co_opendir(pdu, fidp);
2018         if (err < 0) {
2019             goto out;
2020         }
2021         fidp->fid_type = P9_FID_DIR;
2022         err = pdu_marshal(pdu, offset, "Qd", &qid, 0);
2023         if (err < 0) {
2024             goto out;
2025         }
2026         err += offset;
2027     } else {
2028         if (s->proto_version == V9FS_PROTO_2000L) {
2029             flags = get_dotl_openflags(s, mode);
2030         } else {
2031             flags = omode_to_uflags(mode);
2032         }
2033         if (is_ro_export(&s->ctx)) {
2034             if (mode & O_WRONLY || mode & O_RDWR ||
2035                 mode & O_APPEND || mode & O_TRUNC) {
2036                 err = -EROFS;
2037                 goto out;
2038             }
2039         }
2040         err = v9fs_co_open(pdu, fidp, flags);
2041         if (err < 0) {
2042             goto out;
2043         }
2044         fidp->fid_type = P9_FID_FILE;
2045         fidp->open_flags = flags;
2046         if (flags & O_EXCL) {
2047             /*
2048              * We let the host file system do O_EXCL check
2049              * We should not reclaim such fd
2050              */
2051             fidp->flags |= FID_NON_RECLAIMABLE;
2052         }
2053         iounit = get_iounit(pdu, &fidp->path);
2054         err = pdu_marshal(pdu, offset, "Qd", &qid, iounit);
2055         if (err < 0) {
2056             goto out;
2057         }
2058         err += offset;
2059     }
2060     trace_v9fs_open_return(pdu->tag, pdu->id,
2061                            qid.type, qid.version, qid.path, iounit);
2062 out:
2063     put_fid(pdu, fidp);
2064 out_nofid:
2065     pdu_complete(pdu, err);
2066 }
2067 
2068 static void coroutine_fn v9fs_lcreate(void *opaque)
2069 {
2070     int32_t dfid, flags, mode;
2071     gid_t gid;
2072     ssize_t err = 0;
2073     ssize_t offset = 7;
2074     V9fsString name;
2075     V9fsFidState *fidp;
2076     struct stat stbuf;
2077     V9fsQID qid;
2078     int32_t iounit;
2079     V9fsPDU *pdu = opaque;
2080 
2081     v9fs_string_init(&name);
2082     err = pdu_unmarshal(pdu, offset, "dsddd", &dfid,
2083                         &name, &flags, &mode, &gid);
2084     if (err < 0) {
2085         goto out_nofid;
2086     }
2087     trace_v9fs_lcreate(pdu->tag, pdu->id, dfid, flags, mode, gid);
2088 
2089     if (name_is_illegal(name.data)) {
2090         err = -ENOENT;
2091         goto out_nofid;
2092     }
2093 
2094     if (!strcmp(".", name.data) || !strcmp("..", name.data)) {
2095         err = -EEXIST;
2096         goto out_nofid;
2097     }
2098 
2099     fidp = get_fid(pdu, dfid);
2100     if (fidp == NULL) {
2101         err = -ENOENT;
2102         goto out_nofid;
2103     }
2104     if (fidp->fid_type != P9_FID_NONE) {
2105         err = -EINVAL;
2106         goto out;
2107     }
2108 
2109     flags = get_dotl_openflags(pdu->s, flags);
2110     err = v9fs_co_open2(pdu, fidp, &name, gid,
2111                         flags | O_CREAT, mode, &stbuf);
2112     if (err < 0) {
2113         goto out;
2114     }
2115     fidp->fid_type = P9_FID_FILE;
2116     fidp->open_flags = flags;
2117     if (flags & O_EXCL) {
2118         /*
2119          * We let the host file system do O_EXCL check
2120          * We should not reclaim such fd
2121          */
2122         fidp->flags |= FID_NON_RECLAIMABLE;
2123     }
2124     iounit =  get_iounit(pdu, &fidp->path);
2125     err = stat_to_qid(pdu, &stbuf, &qid);
2126     if (err < 0) {
2127         goto out;
2128     }
2129     err = pdu_marshal(pdu, offset, "Qd", &qid, iounit);
2130     if (err < 0) {
2131         goto out;
2132     }
2133     err += offset;
2134     trace_v9fs_lcreate_return(pdu->tag, pdu->id,
2135                               qid.type, qid.version, qid.path, iounit);
2136 out:
2137     put_fid(pdu, fidp);
2138 out_nofid:
2139     pdu_complete(pdu, err);
2140     v9fs_string_free(&name);
2141 }
2142 
2143 static void coroutine_fn v9fs_fsync(void *opaque)
2144 {
2145     int err;
2146     int32_t fid;
2147     int datasync;
2148     size_t offset = 7;
2149     V9fsFidState *fidp;
2150     V9fsPDU *pdu = opaque;
2151 
2152     err = pdu_unmarshal(pdu, offset, "dd", &fid, &datasync);
2153     if (err < 0) {
2154         goto out_nofid;
2155     }
2156     trace_v9fs_fsync(pdu->tag, pdu->id, fid, datasync);
2157 
2158     fidp = get_fid(pdu, fid);
2159     if (fidp == NULL) {
2160         err = -ENOENT;
2161         goto out_nofid;
2162     }
2163     err = v9fs_co_fsync(pdu, fidp, datasync);
2164     if (!err) {
2165         err = offset;
2166     }
2167     put_fid(pdu, fidp);
2168 out_nofid:
2169     pdu_complete(pdu, err);
2170 }
2171 
2172 static void coroutine_fn v9fs_clunk(void *opaque)
2173 {
2174     int err;
2175     int32_t fid;
2176     size_t offset = 7;
2177     V9fsFidState *fidp;
2178     V9fsPDU *pdu = opaque;
2179     V9fsState *s = pdu->s;
2180 
2181     err = pdu_unmarshal(pdu, offset, "d", &fid);
2182     if (err < 0) {
2183         goto out_nofid;
2184     }
2185     trace_v9fs_clunk(pdu->tag, pdu->id, fid);
2186 
2187     fidp = clunk_fid(s, fid);
2188     if (fidp == NULL) {
2189         err = -ENOENT;
2190         goto out_nofid;
2191     }
2192     /*
2193      * Bump the ref so that put_fid will
2194      * free the fid.
2195      */
2196     fidp->ref++;
2197     err = put_fid(pdu, fidp);
2198     if (!err) {
2199         err = offset;
2200     }
2201 out_nofid:
2202     pdu_complete(pdu, err);
2203 }
2204 
2205 /*
2206  * Create a QEMUIOVector for a sub-region of PDU iovecs
2207  *
2208  * @qiov:       uninitialized QEMUIOVector
2209  * @skip:       number of bytes to skip from beginning of PDU
2210  * @size:       number of bytes to include
2211  * @is_write:   true - write, false - read
2212  *
2213  * The resulting QEMUIOVector has heap-allocated iovecs and must be cleaned up
2214  * with qemu_iovec_destroy().
2215  */
2216 static void v9fs_init_qiov_from_pdu(QEMUIOVector *qiov, V9fsPDU *pdu,
2217                                     size_t skip, size_t size,
2218                                     bool is_write)
2219 {
2220     QEMUIOVector elem;
2221     struct iovec *iov;
2222     unsigned int niov;
2223 
2224     if (is_write) {
2225         pdu->s->transport->init_out_iov_from_pdu(pdu, &iov, &niov, size + skip);
2226     } else {
2227         pdu->s->transport->init_in_iov_from_pdu(pdu, &iov, &niov, size + skip);
2228     }
2229 
2230     qemu_iovec_init_external(&elem, iov, niov);
2231     qemu_iovec_init(qiov, niov);
2232     qemu_iovec_concat(qiov, &elem, skip, size);
2233 }
2234 
2235 static int v9fs_xattr_read(V9fsState *s, V9fsPDU *pdu, V9fsFidState *fidp,
2236                            uint64_t off, uint32_t max_count)
2237 {
2238     ssize_t err;
2239     size_t offset = 7;
2240     uint64_t read_count;
2241     QEMUIOVector qiov_full;
2242 
2243     if (fidp->fs.xattr.len < off) {
2244         read_count = 0;
2245     } else {
2246         read_count = fidp->fs.xattr.len - off;
2247     }
2248     if (read_count > max_count) {
2249         read_count = max_count;
2250     }
2251     err = pdu_marshal(pdu, offset, "d", read_count);
2252     if (err < 0) {
2253         return err;
2254     }
2255     offset += err;
2256 
2257     v9fs_init_qiov_from_pdu(&qiov_full, pdu, offset, read_count, false);
2258     err = v9fs_pack(qiov_full.iov, qiov_full.niov, 0,
2259                     ((char *)fidp->fs.xattr.value) + off,
2260                     read_count);
2261     qemu_iovec_destroy(&qiov_full);
2262     if (err < 0) {
2263         return err;
2264     }
2265     offset += err;
2266     return offset;
2267 }
2268 
2269 static int coroutine_fn v9fs_do_readdir_with_stat(V9fsPDU *pdu,
2270                                                   V9fsFidState *fidp,
2271                                                   uint32_t max_count)
2272 {
2273     V9fsPath path;
2274     V9fsStat v9stat;
2275     int len, err = 0;
2276     int32_t count = 0;
2277     struct stat stbuf;
2278     off_t saved_dir_pos;
2279     struct dirent *dent;
2280 
2281     /* save the directory position */
2282     saved_dir_pos = v9fs_co_telldir(pdu, fidp);
2283     if (saved_dir_pos < 0) {
2284         return saved_dir_pos;
2285     }
2286 
2287     while (1) {
2288         v9fs_path_init(&path);
2289 
2290         v9fs_readdir_lock(&fidp->fs.dir);
2291 
2292         err = v9fs_co_readdir(pdu, fidp, &dent);
2293         if (err || !dent) {
2294             break;
2295         }
2296         err = v9fs_co_name_to_path(pdu, &fidp->path, dent->d_name, &path);
2297         if (err < 0) {
2298             break;
2299         }
2300         err = v9fs_co_lstat(pdu, &path, &stbuf);
2301         if (err < 0) {
2302             break;
2303         }
2304         err = stat_to_v9stat(pdu, &path, dent->d_name, &stbuf, &v9stat);
2305         if (err < 0) {
2306             break;
2307         }
2308         if ((count + v9stat.size + 2) > max_count) {
2309             v9fs_readdir_unlock(&fidp->fs.dir);
2310 
2311             /* Ran out of buffer. Set dir back to old position and return */
2312             v9fs_co_seekdir(pdu, fidp, saved_dir_pos);
2313             v9fs_stat_free(&v9stat);
2314             v9fs_path_free(&path);
2315             return count;
2316         }
2317 
2318         /* 11 = 7 + 4 (7 = start offset, 4 = space for storing count) */
2319         len = pdu_marshal(pdu, 11 + count, "S", &v9stat);
2320 
2321         v9fs_readdir_unlock(&fidp->fs.dir);
2322 
2323         if (len < 0) {
2324             v9fs_co_seekdir(pdu, fidp, saved_dir_pos);
2325             v9fs_stat_free(&v9stat);
2326             v9fs_path_free(&path);
2327             return len;
2328         }
2329         count += len;
2330         v9fs_stat_free(&v9stat);
2331         v9fs_path_free(&path);
2332         saved_dir_pos = qemu_dirent_off(dent);
2333     }
2334 
2335     v9fs_readdir_unlock(&fidp->fs.dir);
2336 
2337     v9fs_path_free(&path);
2338     if (err < 0) {
2339         return err;
2340     }
2341     return count;
2342 }
2343 
2344 static void coroutine_fn v9fs_read(void *opaque)
2345 {
2346     int32_t fid;
2347     uint64_t off;
2348     ssize_t err = 0;
2349     int32_t count = 0;
2350     size_t offset = 7;
2351     uint32_t max_count;
2352     V9fsFidState *fidp;
2353     V9fsPDU *pdu = opaque;
2354     V9fsState *s = pdu->s;
2355 
2356     err = pdu_unmarshal(pdu, offset, "dqd", &fid, &off, &max_count);
2357     if (err < 0) {
2358         goto out_nofid;
2359     }
2360     trace_v9fs_read(pdu->tag, pdu->id, fid, off, max_count);
2361 
2362     fidp = get_fid(pdu, fid);
2363     if (fidp == NULL) {
2364         err = -EINVAL;
2365         goto out_nofid;
2366     }
2367     if (fidp->fid_type == P9_FID_DIR) {
2368         if (s->proto_version != V9FS_PROTO_2000U) {
2369             warn_report_once(
2370                 "9p: bad client: T_read request on directory only expected "
2371                 "with 9P2000.u protocol version"
2372             );
2373             err = -EOPNOTSUPP;
2374             goto out;
2375         }
2376         if (off == 0) {
2377             v9fs_co_rewinddir(pdu, fidp);
2378         }
2379         count = v9fs_do_readdir_with_stat(pdu, fidp, max_count);
2380         if (count < 0) {
2381             err = count;
2382             goto out;
2383         }
2384         err = pdu_marshal(pdu, offset, "d", count);
2385         if (err < 0) {
2386             goto out;
2387         }
2388         err += offset + count;
2389     } else if (fidp->fid_type == P9_FID_FILE) {
2390         QEMUIOVector qiov_full;
2391         QEMUIOVector qiov;
2392         int32_t len;
2393 
2394         v9fs_init_qiov_from_pdu(&qiov_full, pdu, offset + 4, max_count, false);
2395         qemu_iovec_init(&qiov, qiov_full.niov);
2396         do {
2397             qemu_iovec_reset(&qiov);
2398             qemu_iovec_concat(&qiov, &qiov_full, count, qiov_full.size - count);
2399             if (0) {
2400                 print_sg(qiov.iov, qiov.niov);
2401             }
2402             /* Loop in case of EINTR */
2403             do {
2404                 len = v9fs_co_preadv(pdu, fidp, qiov.iov, qiov.niov, off);
2405                 if (len >= 0) {
2406                     off   += len;
2407                     count += len;
2408                 }
2409             } while (len == -EINTR && !pdu->cancelled);
2410             if (len < 0) {
2411                 /* IO error return the error */
2412                 err = len;
2413                 goto out_free_iovec;
2414             }
2415         } while (count < max_count && len > 0);
2416         err = pdu_marshal(pdu, offset, "d", count);
2417         if (err < 0) {
2418             goto out_free_iovec;
2419         }
2420         err += offset + count;
2421 out_free_iovec:
2422         qemu_iovec_destroy(&qiov);
2423         qemu_iovec_destroy(&qiov_full);
2424     } else if (fidp->fid_type == P9_FID_XATTR) {
2425         err = v9fs_xattr_read(s, pdu, fidp, off, max_count);
2426     } else {
2427         err = -EINVAL;
2428     }
2429     trace_v9fs_read_return(pdu->tag, pdu->id, count, err);
2430 out:
2431     put_fid(pdu, fidp);
2432 out_nofid:
2433     pdu_complete(pdu, err);
2434 }
2435 
2436 /**
2437  * v9fs_readdir_response_size() - Returns size required in Rreaddir response
2438  * for the passed dirent @name.
2439  *
2440  * @name: directory entry's name (i.e. file name, directory name)
2441  * Return: required size in bytes
2442  */
2443 size_t v9fs_readdir_response_size(V9fsString *name)
2444 {
2445     /*
2446      * Size of each dirent on the wire: size of qid (13) + size of offset (8)
2447      * size of type (1) + size of name.size (2) + strlen(name.data)
2448      */
2449     return 24 + v9fs_string_size(name);
2450 }
2451 
2452 static void v9fs_free_dirents(struct V9fsDirEnt *e)
2453 {
2454     struct V9fsDirEnt *next = NULL;
2455 
2456     for (; e; e = next) {
2457         next = e->next;
2458         g_free(e->dent);
2459         g_free(e->st);
2460         g_free(e);
2461     }
2462 }
2463 
2464 static int coroutine_fn v9fs_do_readdir(V9fsPDU *pdu, V9fsFidState *fidp,
2465                                         off_t offset, int32_t max_count)
2466 {
2467     size_t size;
2468     V9fsQID qid;
2469     V9fsString name;
2470     int len, err = 0;
2471     int32_t count = 0;
2472     off_t off;
2473     struct dirent *dent;
2474     struct stat *st;
2475     struct V9fsDirEnt *entries = NULL;
2476 
2477     /*
2478      * inode remapping requires the device id, which in turn might be
2479      * different for different directory entries, so if inode remapping is
2480      * enabled we have to make a full stat for each directory entry
2481      */
2482     const bool dostat = pdu->s->ctx.export_flags & V9FS_REMAP_INODES;
2483 
2484     /*
2485      * Fetch all required directory entries altogether on a background IO
2486      * thread from fs driver. We don't want to do that for each entry
2487      * individually, because hopping between threads (this main IO thread
2488      * and background IO driver thread) would sum up to huge latencies.
2489      */
2490     count = v9fs_co_readdir_many(pdu, fidp, &entries, offset, max_count,
2491                                  dostat);
2492     if (count < 0) {
2493         err = count;
2494         count = 0;
2495         goto out;
2496     }
2497     count = 0;
2498 
2499     for (struct V9fsDirEnt *e = entries; e; e = e->next) {
2500         dent = e->dent;
2501 
2502         if (pdu->s->ctx.export_flags & V9FS_REMAP_INODES) {
2503             st = e->st;
2504             /* e->st should never be NULL, but just to be sure */
2505             if (!st) {
2506                 err = -1;
2507                 break;
2508             }
2509 
2510             /* remap inode */
2511             err = stat_to_qid(pdu, st, &qid);
2512             if (err < 0) {
2513                 break;
2514             }
2515         } else {
2516             /*
2517              * Fill up just the path field of qid because the client uses
2518              * only that. To fill the entire qid structure we will have
2519              * to stat each dirent found, which is expensive. For the
2520              * latter reason we don't call stat_to_qid() here. Only drawback
2521              * is that no multi-device export detection of stat_to_qid()
2522              * would be done and provided as error to the user here. But
2523              * user would get that error anyway when accessing those
2524              * files/dirs through other ways.
2525              */
2526             size = MIN(sizeof(dent->d_ino), sizeof(qid.path));
2527             memcpy(&qid.path, &dent->d_ino, size);
2528             /* Fill the other fields with dummy values */
2529             qid.type = 0;
2530             qid.version = 0;
2531         }
2532 
2533         off = qemu_dirent_off(dent);
2534         v9fs_string_init(&name);
2535         v9fs_string_sprintf(&name, "%s", dent->d_name);
2536 
2537         /* 11 = 7 + 4 (7 = start offset, 4 = space for storing count) */
2538         len = pdu_marshal(pdu, 11 + count, "Qqbs",
2539                           &qid, off,
2540                           dent->d_type, &name);
2541 
2542         v9fs_string_free(&name);
2543 
2544         if (len < 0) {
2545             err = len;
2546             break;
2547         }
2548 
2549         count += len;
2550     }
2551 
2552 out:
2553     v9fs_free_dirents(entries);
2554     if (err < 0) {
2555         return err;
2556     }
2557     return count;
2558 }
2559 
2560 static void coroutine_fn v9fs_readdir(void *opaque)
2561 {
2562     int32_t fid;
2563     V9fsFidState *fidp;
2564     ssize_t retval = 0;
2565     size_t offset = 7;
2566     uint64_t initial_offset;
2567     int32_t count;
2568     uint32_t max_count;
2569     V9fsPDU *pdu = opaque;
2570     V9fsState *s = pdu->s;
2571 
2572     retval = pdu_unmarshal(pdu, offset, "dqd", &fid,
2573                            &initial_offset, &max_count);
2574     if (retval < 0) {
2575         goto out_nofid;
2576     }
2577     trace_v9fs_readdir(pdu->tag, pdu->id, fid, initial_offset, max_count);
2578 
2579     /* Enough space for a R_readdir header: size[4] Rreaddir tag[2] count[4] */
2580     if (max_count > s->msize - 11) {
2581         max_count = s->msize - 11;
2582         warn_report_once(
2583             "9p: bad client: T_readdir with count > msize - 11"
2584         );
2585     }
2586 
2587     fidp = get_fid(pdu, fid);
2588     if (fidp == NULL) {
2589         retval = -EINVAL;
2590         goto out_nofid;
2591     }
2592     if (fidp->fid_type != P9_FID_DIR) {
2593         warn_report_once("9p: bad client: T_readdir on non-directory stream");
2594         retval = -ENOTDIR;
2595         goto out;
2596     }
2597     if (!fidp->fs.dir.stream) {
2598         retval = -EINVAL;
2599         goto out;
2600     }
2601     if (s->proto_version != V9FS_PROTO_2000L) {
2602         warn_report_once(
2603             "9p: bad client: T_readdir request only expected with 9P2000.L "
2604             "protocol version"
2605         );
2606         retval = -EOPNOTSUPP;
2607         goto out;
2608     }
2609     count = v9fs_do_readdir(pdu, fidp, (off_t) initial_offset, max_count);
2610     if (count < 0) {
2611         retval = count;
2612         goto out;
2613     }
2614     retval = pdu_marshal(pdu, offset, "d", count);
2615     if (retval < 0) {
2616         goto out;
2617     }
2618     retval += count + offset;
2619     trace_v9fs_readdir_return(pdu->tag, pdu->id, count, retval);
2620 out:
2621     put_fid(pdu, fidp);
2622 out_nofid:
2623     pdu_complete(pdu, retval);
2624 }
2625 
2626 static int v9fs_xattr_write(V9fsState *s, V9fsPDU *pdu, V9fsFidState *fidp,
2627                             uint64_t off, uint32_t count,
2628                             struct iovec *sg, int cnt)
2629 {
2630     int i, to_copy;
2631     ssize_t err = 0;
2632     uint64_t write_count;
2633     size_t offset = 7;
2634 
2635 
2636     if (fidp->fs.xattr.len < off) {
2637         return -ENOSPC;
2638     }
2639     write_count = fidp->fs.xattr.len - off;
2640     if (write_count > count) {
2641         write_count = count;
2642     }
2643     err = pdu_marshal(pdu, offset, "d", write_count);
2644     if (err < 0) {
2645         return err;
2646     }
2647     err += offset;
2648     fidp->fs.xattr.copied_len += write_count;
2649     /*
2650      * Now copy the content from sg list
2651      */
2652     for (i = 0; i < cnt; i++) {
2653         if (write_count > sg[i].iov_len) {
2654             to_copy = sg[i].iov_len;
2655         } else {
2656             to_copy = write_count;
2657         }
2658         memcpy((char *)fidp->fs.xattr.value + off, sg[i].iov_base, to_copy);
2659         /* updating vs->off since we are not using below */
2660         off += to_copy;
2661         write_count -= to_copy;
2662     }
2663 
2664     return err;
2665 }
2666 
2667 static void coroutine_fn v9fs_write(void *opaque)
2668 {
2669     ssize_t err;
2670     int32_t fid;
2671     uint64_t off;
2672     uint32_t count;
2673     int32_t len = 0;
2674     int32_t total = 0;
2675     size_t offset = 7;
2676     V9fsFidState *fidp;
2677     V9fsPDU *pdu = opaque;
2678     V9fsState *s = pdu->s;
2679     QEMUIOVector qiov_full;
2680     QEMUIOVector qiov;
2681 
2682     err = pdu_unmarshal(pdu, offset, "dqd", &fid, &off, &count);
2683     if (err < 0) {
2684         pdu_complete(pdu, err);
2685         return;
2686     }
2687     offset += err;
2688     v9fs_init_qiov_from_pdu(&qiov_full, pdu, offset, count, true);
2689     trace_v9fs_write(pdu->tag, pdu->id, fid, off, count, qiov_full.niov);
2690 
2691     fidp = get_fid(pdu, fid);
2692     if (fidp == NULL) {
2693         err = -EINVAL;
2694         goto out_nofid;
2695     }
2696     if (fidp->fid_type == P9_FID_FILE) {
2697         if (fidp->fs.fd == -1) {
2698             err = -EINVAL;
2699             goto out;
2700         }
2701     } else if (fidp->fid_type == P9_FID_XATTR) {
2702         /*
2703          * setxattr operation
2704          */
2705         err = v9fs_xattr_write(s, pdu, fidp, off, count,
2706                                qiov_full.iov, qiov_full.niov);
2707         goto out;
2708     } else {
2709         err = -EINVAL;
2710         goto out;
2711     }
2712     qemu_iovec_init(&qiov, qiov_full.niov);
2713     do {
2714         qemu_iovec_reset(&qiov);
2715         qemu_iovec_concat(&qiov, &qiov_full, total, qiov_full.size - total);
2716         if (0) {
2717             print_sg(qiov.iov, qiov.niov);
2718         }
2719         /* Loop in case of EINTR */
2720         do {
2721             len = v9fs_co_pwritev(pdu, fidp, qiov.iov, qiov.niov, off);
2722             if (len >= 0) {
2723                 off   += len;
2724                 total += len;
2725             }
2726         } while (len == -EINTR && !pdu->cancelled);
2727         if (len < 0) {
2728             /* IO error return the error */
2729             err = len;
2730             goto out_qiov;
2731         }
2732     } while (total < count && len > 0);
2733 
2734     offset = 7;
2735     err = pdu_marshal(pdu, offset, "d", total);
2736     if (err < 0) {
2737         goto out_qiov;
2738     }
2739     err += offset;
2740     trace_v9fs_write_return(pdu->tag, pdu->id, total, err);
2741 out_qiov:
2742     qemu_iovec_destroy(&qiov);
2743 out:
2744     put_fid(pdu, fidp);
2745 out_nofid:
2746     qemu_iovec_destroy(&qiov_full);
2747     pdu_complete(pdu, err);
2748 }
2749 
2750 static void coroutine_fn v9fs_create(void *opaque)
2751 {
2752     int32_t fid;
2753     int err = 0;
2754     size_t offset = 7;
2755     V9fsFidState *fidp;
2756     V9fsQID qid;
2757     int32_t perm;
2758     int8_t mode;
2759     V9fsPath path;
2760     struct stat stbuf;
2761     V9fsString name;
2762     V9fsString extension;
2763     int iounit;
2764     V9fsPDU *pdu = opaque;
2765     V9fsState *s = pdu->s;
2766 
2767     v9fs_path_init(&path);
2768     v9fs_string_init(&name);
2769     v9fs_string_init(&extension);
2770     err = pdu_unmarshal(pdu, offset, "dsdbs", &fid, &name,
2771                         &perm, &mode, &extension);
2772     if (err < 0) {
2773         goto out_nofid;
2774     }
2775     trace_v9fs_create(pdu->tag, pdu->id, fid, name.data, perm, mode);
2776 
2777     if (name_is_illegal(name.data)) {
2778         err = -ENOENT;
2779         goto out_nofid;
2780     }
2781 
2782     if (!strcmp(".", name.data) || !strcmp("..", name.data)) {
2783         err = -EEXIST;
2784         goto out_nofid;
2785     }
2786 
2787     fidp = get_fid(pdu, fid);
2788     if (fidp == NULL) {
2789         err = -EINVAL;
2790         goto out_nofid;
2791     }
2792     if (fidp->fid_type != P9_FID_NONE) {
2793         err = -EINVAL;
2794         goto out;
2795     }
2796     if (perm & P9_STAT_MODE_DIR) {
2797         err = v9fs_co_mkdir(pdu, fidp, &name, perm & 0777,
2798                             fidp->uid, -1, &stbuf);
2799         if (err < 0) {
2800             goto out;
2801         }
2802         err = v9fs_co_name_to_path(pdu, &fidp->path, name.data, &path);
2803         if (err < 0) {
2804             goto out;
2805         }
2806         v9fs_path_write_lock(s);
2807         v9fs_path_copy(&fidp->path, &path);
2808         v9fs_path_unlock(s);
2809         err = v9fs_co_opendir(pdu, fidp);
2810         if (err < 0) {
2811             goto out;
2812         }
2813         fidp->fid_type = P9_FID_DIR;
2814     } else if (perm & P9_STAT_MODE_SYMLINK) {
2815         err = v9fs_co_symlink(pdu, fidp, &name,
2816                               extension.data, -1 , &stbuf);
2817         if (err < 0) {
2818             goto out;
2819         }
2820         err = v9fs_co_name_to_path(pdu, &fidp->path, name.data, &path);
2821         if (err < 0) {
2822             goto out;
2823         }
2824         v9fs_path_write_lock(s);
2825         v9fs_path_copy(&fidp->path, &path);
2826         v9fs_path_unlock(s);
2827     } else if (perm & P9_STAT_MODE_LINK) {
2828         int32_t ofid = atoi(extension.data);
2829         V9fsFidState *ofidp = get_fid(pdu, ofid);
2830         if (ofidp == NULL) {
2831             err = -EINVAL;
2832             goto out;
2833         }
2834         err = v9fs_co_link(pdu, ofidp, fidp, &name);
2835         put_fid(pdu, ofidp);
2836         if (err < 0) {
2837             goto out;
2838         }
2839         err = v9fs_co_name_to_path(pdu, &fidp->path, name.data, &path);
2840         if (err < 0) {
2841             fidp->fid_type = P9_FID_NONE;
2842             goto out;
2843         }
2844         v9fs_path_write_lock(s);
2845         v9fs_path_copy(&fidp->path, &path);
2846         v9fs_path_unlock(s);
2847         err = v9fs_co_lstat(pdu, &fidp->path, &stbuf);
2848         if (err < 0) {
2849             fidp->fid_type = P9_FID_NONE;
2850             goto out;
2851         }
2852     } else if (perm & P9_STAT_MODE_DEVICE) {
2853         char ctype;
2854         uint32_t major, minor;
2855         mode_t nmode = 0;
2856 
2857         if (sscanf(extension.data, "%c %u %u", &ctype, &major, &minor) != 3) {
2858             err = -errno;
2859             goto out;
2860         }
2861 
2862         switch (ctype) {
2863         case 'c':
2864             nmode = S_IFCHR;
2865             break;
2866         case 'b':
2867             nmode = S_IFBLK;
2868             break;
2869         default:
2870             err = -EIO;
2871             goto out;
2872         }
2873 
2874         nmode |= perm & 0777;
2875         err = v9fs_co_mknod(pdu, fidp, &name, fidp->uid, -1,
2876                             makedev(major, minor), nmode, &stbuf);
2877         if (err < 0) {
2878             goto out;
2879         }
2880         err = v9fs_co_name_to_path(pdu, &fidp->path, name.data, &path);
2881         if (err < 0) {
2882             goto out;
2883         }
2884         v9fs_path_write_lock(s);
2885         v9fs_path_copy(&fidp->path, &path);
2886         v9fs_path_unlock(s);
2887     } else if (perm & P9_STAT_MODE_NAMED_PIPE) {
2888         err = v9fs_co_mknod(pdu, fidp, &name, fidp->uid, -1,
2889                             0, S_IFIFO | (perm & 0777), &stbuf);
2890         if (err < 0) {
2891             goto out;
2892         }
2893         err = v9fs_co_name_to_path(pdu, &fidp->path, name.data, &path);
2894         if (err < 0) {
2895             goto out;
2896         }
2897         v9fs_path_write_lock(s);
2898         v9fs_path_copy(&fidp->path, &path);
2899         v9fs_path_unlock(s);
2900     } else if (perm & P9_STAT_MODE_SOCKET) {
2901         err = v9fs_co_mknod(pdu, fidp, &name, fidp->uid, -1,
2902                             0, S_IFSOCK | (perm & 0777), &stbuf);
2903         if (err < 0) {
2904             goto out;
2905         }
2906         err = v9fs_co_name_to_path(pdu, &fidp->path, name.data, &path);
2907         if (err < 0) {
2908             goto out;
2909         }
2910         v9fs_path_write_lock(s);
2911         v9fs_path_copy(&fidp->path, &path);
2912         v9fs_path_unlock(s);
2913     } else {
2914         err = v9fs_co_open2(pdu, fidp, &name, -1,
2915                             omode_to_uflags(mode) | O_CREAT, perm, &stbuf);
2916         if (err < 0) {
2917             goto out;
2918         }
2919         fidp->fid_type = P9_FID_FILE;
2920         fidp->open_flags = omode_to_uflags(mode);
2921         if (fidp->open_flags & O_EXCL) {
2922             /*
2923              * We let the host file system do O_EXCL check
2924              * We should not reclaim such fd
2925              */
2926             fidp->flags |= FID_NON_RECLAIMABLE;
2927         }
2928     }
2929     iounit = get_iounit(pdu, &fidp->path);
2930     err = stat_to_qid(pdu, &stbuf, &qid);
2931     if (err < 0) {
2932         goto out;
2933     }
2934     err = pdu_marshal(pdu, offset, "Qd", &qid, iounit);
2935     if (err < 0) {
2936         goto out;
2937     }
2938     err += offset;
2939     trace_v9fs_create_return(pdu->tag, pdu->id,
2940                              qid.type, qid.version, qid.path, iounit);
2941 out:
2942     put_fid(pdu, fidp);
2943 out_nofid:
2944    pdu_complete(pdu, err);
2945    v9fs_string_free(&name);
2946    v9fs_string_free(&extension);
2947    v9fs_path_free(&path);
2948 }
2949 
2950 static void coroutine_fn v9fs_symlink(void *opaque)
2951 {
2952     V9fsPDU *pdu = opaque;
2953     V9fsString name;
2954     V9fsString symname;
2955     V9fsFidState *dfidp;
2956     V9fsQID qid;
2957     struct stat stbuf;
2958     int32_t dfid;
2959     int err = 0;
2960     gid_t gid;
2961     size_t offset = 7;
2962 
2963     v9fs_string_init(&name);
2964     v9fs_string_init(&symname);
2965     err = pdu_unmarshal(pdu, offset, "dssd", &dfid, &name, &symname, &gid);
2966     if (err < 0) {
2967         goto out_nofid;
2968     }
2969     trace_v9fs_symlink(pdu->tag, pdu->id, dfid, name.data, symname.data, gid);
2970 
2971     if (name_is_illegal(name.data)) {
2972         err = -ENOENT;
2973         goto out_nofid;
2974     }
2975 
2976     if (!strcmp(".", name.data) || !strcmp("..", name.data)) {
2977         err = -EEXIST;
2978         goto out_nofid;
2979     }
2980 
2981     dfidp = get_fid(pdu, dfid);
2982     if (dfidp == NULL) {
2983         err = -EINVAL;
2984         goto out_nofid;
2985     }
2986     err = v9fs_co_symlink(pdu, dfidp, &name, symname.data, gid, &stbuf);
2987     if (err < 0) {
2988         goto out;
2989     }
2990     err = stat_to_qid(pdu, &stbuf, &qid);
2991     if (err < 0) {
2992         goto out;
2993     }
2994     err =  pdu_marshal(pdu, offset, "Q", &qid);
2995     if (err < 0) {
2996         goto out;
2997     }
2998     err += offset;
2999     trace_v9fs_symlink_return(pdu->tag, pdu->id,
3000                               qid.type, qid.version, qid.path);
3001 out:
3002     put_fid(pdu, dfidp);
3003 out_nofid:
3004     pdu_complete(pdu, err);
3005     v9fs_string_free(&name);
3006     v9fs_string_free(&symname);
3007 }
3008 
3009 static void coroutine_fn v9fs_flush(void *opaque)
3010 {
3011     ssize_t err;
3012     int16_t tag;
3013     size_t offset = 7;
3014     V9fsPDU *cancel_pdu = NULL;
3015     V9fsPDU *pdu = opaque;
3016     V9fsState *s = pdu->s;
3017 
3018     err = pdu_unmarshal(pdu, offset, "w", &tag);
3019     if (err < 0) {
3020         pdu_complete(pdu, err);
3021         return;
3022     }
3023     trace_v9fs_flush(pdu->tag, pdu->id, tag);
3024 
3025     if (pdu->tag == tag) {
3026         warn_report("the guest sent a self-referencing 9P flush request");
3027     } else {
3028         QLIST_FOREACH(cancel_pdu, &s->active_list, next) {
3029             if (cancel_pdu->tag == tag) {
3030                 break;
3031             }
3032         }
3033     }
3034     if (cancel_pdu) {
3035         cancel_pdu->cancelled = 1;
3036         /*
3037          * Wait for pdu to complete.
3038          */
3039         qemu_co_queue_wait(&cancel_pdu->complete, NULL);
3040         if (!qemu_co_queue_next(&cancel_pdu->complete)) {
3041             cancel_pdu->cancelled = 0;
3042             pdu_free(cancel_pdu);
3043         }
3044     }
3045     pdu_complete(pdu, 7);
3046 }
3047 
3048 static void coroutine_fn v9fs_link(void *opaque)
3049 {
3050     V9fsPDU *pdu = opaque;
3051     int32_t dfid, oldfid;
3052     V9fsFidState *dfidp, *oldfidp;
3053     V9fsString name;
3054     size_t offset = 7;
3055     int err = 0;
3056 
3057     v9fs_string_init(&name);
3058     err = pdu_unmarshal(pdu, offset, "dds", &dfid, &oldfid, &name);
3059     if (err < 0) {
3060         goto out_nofid;
3061     }
3062     trace_v9fs_link(pdu->tag, pdu->id, dfid, oldfid, name.data);
3063 
3064     if (name_is_illegal(name.data)) {
3065         err = -ENOENT;
3066         goto out_nofid;
3067     }
3068 
3069     if (!strcmp(".", name.data) || !strcmp("..", name.data)) {
3070         err = -EEXIST;
3071         goto out_nofid;
3072     }
3073 
3074     dfidp = get_fid(pdu, dfid);
3075     if (dfidp == NULL) {
3076         err = -ENOENT;
3077         goto out_nofid;
3078     }
3079 
3080     oldfidp = get_fid(pdu, oldfid);
3081     if (oldfidp == NULL) {
3082         err = -ENOENT;
3083         goto out;
3084     }
3085     err = v9fs_co_link(pdu, oldfidp, dfidp, &name);
3086     if (!err) {
3087         err = offset;
3088     }
3089     put_fid(pdu, oldfidp);
3090 out:
3091     put_fid(pdu, dfidp);
3092 out_nofid:
3093     v9fs_string_free(&name);
3094     pdu_complete(pdu, err);
3095 }
3096 
3097 /* Only works with path name based fid */
3098 static void coroutine_fn v9fs_remove(void *opaque)
3099 {
3100     int32_t fid;
3101     int err = 0;
3102     size_t offset = 7;
3103     V9fsFidState *fidp;
3104     V9fsPDU *pdu = opaque;
3105 
3106     err = pdu_unmarshal(pdu, offset, "d", &fid);
3107     if (err < 0) {
3108         goto out_nofid;
3109     }
3110     trace_v9fs_remove(pdu->tag, pdu->id, fid);
3111 
3112     fidp = get_fid(pdu, fid);
3113     if (fidp == NULL) {
3114         err = -EINVAL;
3115         goto out_nofid;
3116     }
3117     /* if fs driver is not path based, return EOPNOTSUPP */
3118     if (!(pdu->s->ctx.export_flags & V9FS_PATHNAME_FSCONTEXT)) {
3119         err = -EOPNOTSUPP;
3120         goto out_err;
3121     }
3122     /*
3123      * IF the file is unlinked, we cannot reopen
3124      * the file later. So don't reclaim fd
3125      */
3126     err = v9fs_mark_fids_unreclaim(pdu, &fidp->path);
3127     if (err < 0) {
3128         goto out_err;
3129     }
3130     err = v9fs_co_remove(pdu, &fidp->path);
3131     if (!err) {
3132         err = offset;
3133     }
3134 out_err:
3135     /* For TREMOVE we need to clunk the fid even on failed remove */
3136     clunk_fid(pdu->s, fidp->fid);
3137     put_fid(pdu, fidp);
3138 out_nofid:
3139     pdu_complete(pdu, err);
3140 }
3141 
3142 static void coroutine_fn v9fs_unlinkat(void *opaque)
3143 {
3144     int err = 0;
3145     V9fsString name;
3146     int32_t dfid, flags, rflags = 0;
3147     size_t offset = 7;
3148     V9fsPath path;
3149     V9fsFidState *dfidp;
3150     V9fsPDU *pdu = opaque;
3151 
3152     v9fs_string_init(&name);
3153     err = pdu_unmarshal(pdu, offset, "dsd", &dfid, &name, &flags);
3154     if (err < 0) {
3155         goto out_nofid;
3156     }
3157 
3158     if (name_is_illegal(name.data)) {
3159         err = -ENOENT;
3160         goto out_nofid;
3161     }
3162 
3163     if (!strcmp(".", name.data)) {
3164         err = -EINVAL;
3165         goto out_nofid;
3166     }
3167 
3168     if (!strcmp("..", name.data)) {
3169         err = -ENOTEMPTY;
3170         goto out_nofid;
3171     }
3172 
3173     if (flags & ~P9_DOTL_AT_REMOVEDIR) {
3174         err = -EINVAL;
3175         goto out_nofid;
3176     }
3177 
3178     if (flags & P9_DOTL_AT_REMOVEDIR) {
3179         rflags |= AT_REMOVEDIR;
3180     }
3181 
3182     dfidp = get_fid(pdu, dfid);
3183     if (dfidp == NULL) {
3184         err = -EINVAL;
3185         goto out_nofid;
3186     }
3187     /*
3188      * IF the file is unlinked, we cannot reopen
3189      * the file later. So don't reclaim fd
3190      */
3191     v9fs_path_init(&path);
3192     err = v9fs_co_name_to_path(pdu, &dfidp->path, name.data, &path);
3193     if (err < 0) {
3194         goto out_err;
3195     }
3196     err = v9fs_mark_fids_unreclaim(pdu, &path);
3197     if (err < 0) {
3198         goto out_err;
3199     }
3200     err = v9fs_co_unlinkat(pdu, &dfidp->path, &name, rflags);
3201     if (!err) {
3202         err = offset;
3203     }
3204 out_err:
3205     put_fid(pdu, dfidp);
3206     v9fs_path_free(&path);
3207 out_nofid:
3208     pdu_complete(pdu, err);
3209     v9fs_string_free(&name);
3210 }
3211 
3212 
3213 /* Only works with path name based fid */
3214 static int coroutine_fn v9fs_complete_rename(V9fsPDU *pdu, V9fsFidState *fidp,
3215                                              int32_t newdirfid,
3216                                              V9fsString *name)
3217 {
3218     int err = 0;
3219     V9fsPath new_path;
3220     V9fsFidState *tfidp;
3221     V9fsState *s = pdu->s;
3222     V9fsFidState *dirfidp = NULL;
3223     GHashTableIter iter;
3224     gpointer fid;
3225 
3226     v9fs_path_init(&new_path);
3227     if (newdirfid != -1) {
3228         dirfidp = get_fid(pdu, newdirfid);
3229         if (dirfidp == NULL) {
3230             return -ENOENT;
3231         }
3232         if (fidp->fid_type != P9_FID_NONE) {
3233             err = -EINVAL;
3234             goto out;
3235         }
3236         err = v9fs_co_name_to_path(pdu, &dirfidp->path, name->data, &new_path);
3237         if (err < 0) {
3238             goto out;
3239         }
3240     } else {
3241         char *dir_name = g_path_get_dirname(fidp->path.data);
3242         V9fsPath dir_path;
3243 
3244         v9fs_path_init(&dir_path);
3245         v9fs_path_sprintf(&dir_path, "%s", dir_name);
3246         g_free(dir_name);
3247 
3248         err = v9fs_co_name_to_path(pdu, &dir_path, name->data, &new_path);
3249         v9fs_path_free(&dir_path);
3250         if (err < 0) {
3251             goto out;
3252         }
3253     }
3254     err = v9fs_co_rename(pdu, &fidp->path, &new_path);
3255     if (err < 0) {
3256         goto out;
3257     }
3258 
3259     /*
3260      * Fixup fid's pointing to the old name to
3261      * start pointing to the new name
3262      */
3263     g_hash_table_iter_init(&iter, s->fids);
3264     while (g_hash_table_iter_next(&iter, &fid, (gpointer *) &tfidp)) {
3265         if (v9fs_path_is_ancestor(&fidp->path, &tfidp->path)) {
3266             /* replace the name */
3267             v9fs_fix_path(&tfidp->path, &new_path, strlen(fidp->path.data));
3268         }
3269     }
3270 out:
3271     if (dirfidp) {
3272         put_fid(pdu, dirfidp);
3273     }
3274     v9fs_path_free(&new_path);
3275     return err;
3276 }
3277 
3278 /* Only works with path name based fid */
3279 static void coroutine_fn v9fs_rename(void *opaque)
3280 {
3281     int32_t fid;
3282     ssize_t err = 0;
3283     size_t offset = 7;
3284     V9fsString name;
3285     int32_t newdirfid;
3286     V9fsFidState *fidp;
3287     V9fsPDU *pdu = opaque;
3288     V9fsState *s = pdu->s;
3289 
3290     v9fs_string_init(&name);
3291     err = pdu_unmarshal(pdu, offset, "dds", &fid, &newdirfid, &name);
3292     if (err < 0) {
3293         goto out_nofid;
3294     }
3295 
3296     if (name_is_illegal(name.data)) {
3297         err = -ENOENT;
3298         goto out_nofid;
3299     }
3300 
3301     if (!strcmp(".", name.data) || !strcmp("..", name.data)) {
3302         err = -EISDIR;
3303         goto out_nofid;
3304     }
3305 
3306     fidp = get_fid(pdu, fid);
3307     if (fidp == NULL) {
3308         err = -ENOENT;
3309         goto out_nofid;
3310     }
3311     if (fidp->fid_type != P9_FID_NONE) {
3312         err = -EINVAL;
3313         goto out;
3314     }
3315     /* if fs driver is not path based, return EOPNOTSUPP */
3316     if (!(pdu->s->ctx.export_flags & V9FS_PATHNAME_FSCONTEXT)) {
3317         err = -EOPNOTSUPP;
3318         goto out;
3319     }
3320     v9fs_path_write_lock(s);
3321     err = v9fs_complete_rename(pdu, fidp, newdirfid, &name);
3322     v9fs_path_unlock(s);
3323     if (!err) {
3324         err = offset;
3325     }
3326 out:
3327     put_fid(pdu, fidp);
3328 out_nofid:
3329     pdu_complete(pdu, err);
3330     v9fs_string_free(&name);
3331 }
3332 
3333 static int coroutine_fn v9fs_fix_fid_paths(V9fsPDU *pdu, V9fsPath *olddir,
3334                                            V9fsString *old_name,
3335                                            V9fsPath *newdir,
3336                                            V9fsString *new_name)
3337 {
3338     V9fsFidState *tfidp;
3339     V9fsPath oldpath, newpath;
3340     V9fsState *s = pdu->s;
3341     int err;
3342     GHashTableIter iter;
3343     gpointer fid;
3344 
3345     v9fs_path_init(&oldpath);
3346     v9fs_path_init(&newpath);
3347     err = v9fs_co_name_to_path(pdu, olddir, old_name->data, &oldpath);
3348     if (err < 0) {
3349         goto out;
3350     }
3351     err = v9fs_co_name_to_path(pdu, newdir, new_name->data, &newpath);
3352     if (err < 0) {
3353         goto out;
3354     }
3355 
3356     /*
3357      * Fixup fid's pointing to the old name to
3358      * start pointing to the new name
3359      */
3360     g_hash_table_iter_init(&iter, s->fids);
3361     while (g_hash_table_iter_next(&iter, &fid, (gpointer *) &tfidp)) {
3362         if (v9fs_path_is_ancestor(&oldpath, &tfidp->path)) {
3363             /* replace the name */
3364             v9fs_fix_path(&tfidp->path, &newpath, strlen(oldpath.data));
3365         }
3366     }
3367 out:
3368     v9fs_path_free(&oldpath);
3369     v9fs_path_free(&newpath);
3370     return err;
3371 }
3372 
3373 static int coroutine_fn v9fs_complete_renameat(V9fsPDU *pdu, int32_t olddirfid,
3374                                                V9fsString *old_name,
3375                                                int32_t newdirfid,
3376                                                V9fsString *new_name)
3377 {
3378     int err = 0;
3379     V9fsState *s = pdu->s;
3380     V9fsFidState *newdirfidp = NULL, *olddirfidp = NULL;
3381 
3382     olddirfidp = get_fid(pdu, olddirfid);
3383     if (olddirfidp == NULL) {
3384         err = -ENOENT;
3385         goto out;
3386     }
3387     if (newdirfid != -1) {
3388         newdirfidp = get_fid(pdu, newdirfid);
3389         if (newdirfidp == NULL) {
3390             err = -ENOENT;
3391             goto out;
3392         }
3393     } else {
3394         newdirfidp = get_fid(pdu, olddirfid);
3395     }
3396 
3397     err = v9fs_co_renameat(pdu, &olddirfidp->path, old_name,
3398                            &newdirfidp->path, new_name);
3399     if (err < 0) {
3400         goto out;
3401     }
3402     if (s->ctx.export_flags & V9FS_PATHNAME_FSCONTEXT) {
3403         /* Only for path based fid  we need to do the below fixup */
3404         err = v9fs_fix_fid_paths(pdu, &olddirfidp->path, old_name,
3405                                  &newdirfidp->path, new_name);
3406     }
3407 out:
3408     if (olddirfidp) {
3409         put_fid(pdu, olddirfidp);
3410     }
3411     if (newdirfidp) {
3412         put_fid(pdu, newdirfidp);
3413     }
3414     return err;
3415 }
3416 
3417 static void coroutine_fn v9fs_renameat(void *opaque)
3418 {
3419     ssize_t err = 0;
3420     size_t offset = 7;
3421     V9fsPDU *pdu = opaque;
3422     V9fsState *s = pdu->s;
3423     int32_t olddirfid, newdirfid;
3424     V9fsString old_name, new_name;
3425 
3426     v9fs_string_init(&old_name);
3427     v9fs_string_init(&new_name);
3428     err = pdu_unmarshal(pdu, offset, "dsds", &olddirfid,
3429                         &old_name, &newdirfid, &new_name);
3430     if (err < 0) {
3431         goto out_err;
3432     }
3433 
3434     if (name_is_illegal(old_name.data) || name_is_illegal(new_name.data)) {
3435         err = -ENOENT;
3436         goto out_err;
3437     }
3438 
3439     if (!strcmp(".", old_name.data) || !strcmp("..", old_name.data) ||
3440         !strcmp(".", new_name.data) || !strcmp("..", new_name.data)) {
3441         err = -EISDIR;
3442         goto out_err;
3443     }
3444 
3445     v9fs_path_write_lock(s);
3446     err = v9fs_complete_renameat(pdu, olddirfid,
3447                                  &old_name, newdirfid, &new_name);
3448     v9fs_path_unlock(s);
3449     if (!err) {
3450         err = offset;
3451     }
3452 
3453 out_err:
3454     pdu_complete(pdu, err);
3455     v9fs_string_free(&old_name);
3456     v9fs_string_free(&new_name);
3457 }
3458 
3459 static void coroutine_fn v9fs_wstat(void *opaque)
3460 {
3461     int32_t fid;
3462     int err = 0;
3463     int16_t unused;
3464     V9fsStat v9stat;
3465     size_t offset = 7;
3466     struct stat stbuf;
3467     V9fsFidState *fidp;
3468     V9fsPDU *pdu = opaque;
3469     V9fsState *s = pdu->s;
3470 
3471     v9fs_stat_init(&v9stat);
3472     err = pdu_unmarshal(pdu, offset, "dwS", &fid, &unused, &v9stat);
3473     if (err < 0) {
3474         goto out_nofid;
3475     }
3476     trace_v9fs_wstat(pdu->tag, pdu->id, fid,
3477                      v9stat.mode, v9stat.atime, v9stat.mtime);
3478 
3479     fidp = get_fid(pdu, fid);
3480     if (fidp == NULL) {
3481         err = -EINVAL;
3482         goto out_nofid;
3483     }
3484     /* do we need to sync the file? */
3485     if (donttouch_stat(&v9stat)) {
3486         err = v9fs_co_fsync(pdu, fidp, 0);
3487         goto out;
3488     }
3489     if (v9stat.mode != -1) {
3490         uint32_t v9_mode;
3491         err = v9fs_co_lstat(pdu, &fidp->path, &stbuf);
3492         if (err < 0) {
3493             goto out;
3494         }
3495         v9_mode = stat_to_v9mode(&stbuf);
3496         if ((v9stat.mode & P9_STAT_MODE_TYPE_BITS) !=
3497             (v9_mode & P9_STAT_MODE_TYPE_BITS)) {
3498             /* Attempting to change the type */
3499             err = -EIO;
3500             goto out;
3501         }
3502         err = v9fs_co_chmod(pdu, &fidp->path,
3503                             v9mode_to_mode(v9stat.mode,
3504                                            &v9stat.extension));
3505         if (err < 0) {
3506             goto out;
3507         }
3508     }
3509     if (v9stat.mtime != -1 || v9stat.atime != -1) {
3510         struct timespec times[2];
3511         if (v9stat.atime != -1) {
3512             times[0].tv_sec = v9stat.atime;
3513             times[0].tv_nsec = 0;
3514         } else {
3515             times[0].tv_nsec = UTIME_OMIT;
3516         }
3517         if (v9stat.mtime != -1) {
3518             times[1].tv_sec = v9stat.mtime;
3519             times[1].tv_nsec = 0;
3520         } else {
3521             times[1].tv_nsec = UTIME_OMIT;
3522         }
3523         err = v9fs_co_utimensat(pdu, &fidp->path, times);
3524         if (err < 0) {
3525             goto out;
3526         }
3527     }
3528     if (v9stat.n_gid != -1 || v9stat.n_uid != -1) {
3529         err = v9fs_co_chown(pdu, &fidp->path, v9stat.n_uid, v9stat.n_gid);
3530         if (err < 0) {
3531             goto out;
3532         }
3533     }
3534     if (v9stat.name.size != 0) {
3535         v9fs_path_write_lock(s);
3536         err = v9fs_complete_rename(pdu, fidp, -1, &v9stat.name);
3537         v9fs_path_unlock(s);
3538         if (err < 0) {
3539             goto out;
3540         }
3541     }
3542     if (v9stat.length != -1) {
3543         err = v9fs_co_truncate(pdu, &fidp->path, v9stat.length);
3544         if (err < 0) {
3545             goto out;
3546         }
3547     }
3548     err = offset;
3549 out:
3550     put_fid(pdu, fidp);
3551 out_nofid:
3552     v9fs_stat_free(&v9stat);
3553     pdu_complete(pdu, err);
3554 }
3555 
3556 static int v9fs_fill_statfs(V9fsState *s, V9fsPDU *pdu, struct statfs *stbuf)
3557 {
3558     uint32_t f_type;
3559     uint32_t f_bsize;
3560     uint64_t f_blocks;
3561     uint64_t f_bfree;
3562     uint64_t f_bavail;
3563     uint64_t f_files;
3564     uint64_t f_ffree;
3565     uint64_t fsid_val;
3566     uint32_t f_namelen;
3567     size_t offset = 7;
3568     int32_t bsize_factor;
3569 
3570     /*
3571      * compute bsize factor based on host file system block size
3572      * and client msize
3573      */
3574     bsize_factor = (s->msize - P9_IOHDRSZ) / stbuf->f_bsize;
3575     if (!bsize_factor) {
3576         bsize_factor = 1;
3577     }
3578     f_type  = stbuf->f_type;
3579     f_bsize = stbuf->f_bsize;
3580     f_bsize *= bsize_factor;
3581     /*
3582      * f_bsize is adjusted(multiplied) by bsize factor, so we need to
3583      * adjust(divide) the number of blocks, free blocks and available
3584      * blocks by bsize factor
3585      */
3586     f_blocks = stbuf->f_blocks / bsize_factor;
3587     f_bfree  = stbuf->f_bfree / bsize_factor;
3588     f_bavail = stbuf->f_bavail / bsize_factor;
3589     f_files  = stbuf->f_files;
3590     f_ffree  = stbuf->f_ffree;
3591 #ifdef CONFIG_DARWIN
3592     fsid_val = (unsigned int)stbuf->f_fsid.val[0] |
3593                (unsigned long long)stbuf->f_fsid.val[1] << 32;
3594     f_namelen = NAME_MAX;
3595 #else
3596     fsid_val = (unsigned int) stbuf->f_fsid.__val[0] |
3597                (unsigned long long)stbuf->f_fsid.__val[1] << 32;
3598     f_namelen = stbuf->f_namelen;
3599 #endif
3600 
3601     return pdu_marshal(pdu, offset, "ddqqqqqqd",
3602                        f_type, f_bsize, f_blocks, f_bfree,
3603                        f_bavail, f_files, f_ffree,
3604                        fsid_val, f_namelen);
3605 }
3606 
3607 static void coroutine_fn v9fs_statfs(void *opaque)
3608 {
3609     int32_t fid;
3610     ssize_t retval = 0;
3611     size_t offset = 7;
3612     V9fsFidState *fidp;
3613     struct statfs stbuf;
3614     V9fsPDU *pdu = opaque;
3615     V9fsState *s = pdu->s;
3616 
3617     retval = pdu_unmarshal(pdu, offset, "d", &fid);
3618     if (retval < 0) {
3619         goto out_nofid;
3620     }
3621     fidp = get_fid(pdu, fid);
3622     if (fidp == NULL) {
3623         retval = -ENOENT;
3624         goto out_nofid;
3625     }
3626     retval = v9fs_co_statfs(pdu, &fidp->path, &stbuf);
3627     if (retval < 0) {
3628         goto out;
3629     }
3630     retval = v9fs_fill_statfs(s, pdu, &stbuf);
3631     if (retval < 0) {
3632         goto out;
3633     }
3634     retval += offset;
3635 out:
3636     put_fid(pdu, fidp);
3637 out_nofid:
3638     pdu_complete(pdu, retval);
3639 }
3640 
3641 static void coroutine_fn v9fs_mknod(void *opaque)
3642 {
3643 
3644     int mode;
3645     gid_t gid;
3646     int32_t fid;
3647     V9fsQID qid;
3648     int err = 0;
3649     int major, minor;
3650     size_t offset = 7;
3651     V9fsString name;
3652     struct stat stbuf;
3653     V9fsFidState *fidp;
3654     V9fsPDU *pdu = opaque;
3655 
3656     v9fs_string_init(&name);
3657     err = pdu_unmarshal(pdu, offset, "dsdddd", &fid, &name, &mode,
3658                         &major, &minor, &gid);
3659     if (err < 0) {
3660         goto out_nofid;
3661     }
3662     trace_v9fs_mknod(pdu->tag, pdu->id, fid, mode, major, minor);
3663 
3664     if (name_is_illegal(name.data)) {
3665         err = -ENOENT;
3666         goto out_nofid;
3667     }
3668 
3669     if (!strcmp(".", name.data) || !strcmp("..", name.data)) {
3670         err = -EEXIST;
3671         goto out_nofid;
3672     }
3673 
3674     fidp = get_fid(pdu, fid);
3675     if (fidp == NULL) {
3676         err = -ENOENT;
3677         goto out_nofid;
3678     }
3679     err = v9fs_co_mknod(pdu, fidp, &name, fidp->uid, gid,
3680                         makedev(major, minor), mode, &stbuf);
3681     if (err < 0) {
3682         goto out;
3683     }
3684     err = stat_to_qid(pdu, &stbuf, &qid);
3685     if (err < 0) {
3686         goto out;
3687     }
3688     err = pdu_marshal(pdu, offset, "Q", &qid);
3689     if (err < 0) {
3690         goto out;
3691     }
3692     err += offset;
3693     trace_v9fs_mknod_return(pdu->tag, pdu->id,
3694                             qid.type, qid.version, qid.path);
3695 out:
3696     put_fid(pdu, fidp);
3697 out_nofid:
3698     pdu_complete(pdu, err);
3699     v9fs_string_free(&name);
3700 }
3701 
3702 /*
3703  * Implement posix byte range locking code
3704  * Server side handling of locking code is very simple, because 9p server in
3705  * QEMU can handle only one client. And most of the lock handling
3706  * (like conflict, merging) etc is done by the VFS layer itself, so no need to
3707  * do any thing in * qemu 9p server side lock code path.
3708  * So when a TLOCK request comes, always return success
3709  */
3710 static void coroutine_fn v9fs_lock(void *opaque)
3711 {
3712     V9fsFlock flock;
3713     size_t offset = 7;
3714     struct stat stbuf;
3715     V9fsFidState *fidp;
3716     int32_t fid, err = 0;
3717     V9fsPDU *pdu = opaque;
3718 
3719     v9fs_string_init(&flock.client_id);
3720     err = pdu_unmarshal(pdu, offset, "dbdqqds", &fid, &flock.type,
3721                         &flock.flags, &flock.start, &flock.length,
3722                         &flock.proc_id, &flock.client_id);
3723     if (err < 0) {
3724         goto out_nofid;
3725     }
3726     trace_v9fs_lock(pdu->tag, pdu->id, fid,
3727                     flock.type, flock.start, flock.length);
3728 
3729 
3730     /* We support only block flag now (that too ignored currently) */
3731     if (flock.flags & ~P9_LOCK_FLAGS_BLOCK) {
3732         err = -EINVAL;
3733         goto out_nofid;
3734     }
3735     fidp = get_fid(pdu, fid);
3736     if (fidp == NULL) {
3737         err = -ENOENT;
3738         goto out_nofid;
3739     }
3740     err = v9fs_co_fstat(pdu, fidp, &stbuf);
3741     if (err < 0) {
3742         goto out;
3743     }
3744     err = pdu_marshal(pdu, offset, "b", P9_LOCK_SUCCESS);
3745     if (err < 0) {
3746         goto out;
3747     }
3748     err += offset;
3749     trace_v9fs_lock_return(pdu->tag, pdu->id, P9_LOCK_SUCCESS);
3750 out:
3751     put_fid(pdu, fidp);
3752 out_nofid:
3753     pdu_complete(pdu, err);
3754     v9fs_string_free(&flock.client_id);
3755 }
3756 
3757 /*
3758  * When a TGETLOCK request comes, always return success because all lock
3759  * handling is done by client's VFS layer.
3760  */
3761 static void coroutine_fn v9fs_getlock(void *opaque)
3762 {
3763     size_t offset = 7;
3764     struct stat stbuf;
3765     V9fsFidState *fidp;
3766     V9fsGetlock glock;
3767     int32_t fid, err = 0;
3768     V9fsPDU *pdu = opaque;
3769 
3770     v9fs_string_init(&glock.client_id);
3771     err = pdu_unmarshal(pdu, offset, "dbqqds", &fid, &glock.type,
3772                         &glock.start, &glock.length, &glock.proc_id,
3773                         &glock.client_id);
3774     if (err < 0) {
3775         goto out_nofid;
3776     }
3777     trace_v9fs_getlock(pdu->tag, pdu->id, fid,
3778                        glock.type, glock.start, glock.length);
3779 
3780     fidp = get_fid(pdu, fid);
3781     if (fidp == NULL) {
3782         err = -ENOENT;
3783         goto out_nofid;
3784     }
3785     err = v9fs_co_fstat(pdu, fidp, &stbuf);
3786     if (err < 0) {
3787         goto out;
3788     }
3789     glock.type = P9_LOCK_TYPE_UNLCK;
3790     err = pdu_marshal(pdu, offset, "bqqds", glock.type,
3791                           glock.start, glock.length, glock.proc_id,
3792                           &glock.client_id);
3793     if (err < 0) {
3794         goto out;
3795     }
3796     err += offset;
3797     trace_v9fs_getlock_return(pdu->tag, pdu->id, glock.type, glock.start,
3798                               glock.length, glock.proc_id);
3799 out:
3800     put_fid(pdu, fidp);
3801 out_nofid:
3802     pdu_complete(pdu, err);
3803     v9fs_string_free(&glock.client_id);
3804 }
3805 
3806 static void coroutine_fn v9fs_mkdir(void *opaque)
3807 {
3808     V9fsPDU *pdu = opaque;
3809     size_t offset = 7;
3810     int32_t fid;
3811     struct stat stbuf;
3812     V9fsQID qid;
3813     V9fsString name;
3814     V9fsFidState *fidp;
3815     gid_t gid;
3816     int mode;
3817     int err = 0;
3818 
3819     v9fs_string_init(&name);
3820     err = pdu_unmarshal(pdu, offset, "dsdd", &fid, &name, &mode, &gid);
3821     if (err < 0) {
3822         goto out_nofid;
3823     }
3824     trace_v9fs_mkdir(pdu->tag, pdu->id, fid, name.data, mode, gid);
3825 
3826     if (name_is_illegal(name.data)) {
3827         err = -ENOENT;
3828         goto out_nofid;
3829     }
3830 
3831     if (!strcmp(".", name.data) || !strcmp("..", name.data)) {
3832         err = -EEXIST;
3833         goto out_nofid;
3834     }
3835 
3836     fidp = get_fid(pdu, fid);
3837     if (fidp == NULL) {
3838         err = -ENOENT;
3839         goto out_nofid;
3840     }
3841     err = v9fs_co_mkdir(pdu, fidp, &name, mode, fidp->uid, gid, &stbuf);
3842     if (err < 0) {
3843         goto out;
3844     }
3845     err = stat_to_qid(pdu, &stbuf, &qid);
3846     if (err < 0) {
3847         goto out;
3848     }
3849     err = pdu_marshal(pdu, offset, "Q", &qid);
3850     if (err < 0) {
3851         goto out;
3852     }
3853     err += offset;
3854     trace_v9fs_mkdir_return(pdu->tag, pdu->id,
3855                             qid.type, qid.version, qid.path, err);
3856 out:
3857     put_fid(pdu, fidp);
3858 out_nofid:
3859     pdu_complete(pdu, err);
3860     v9fs_string_free(&name);
3861 }
3862 
3863 static void coroutine_fn v9fs_xattrwalk(void *opaque)
3864 {
3865     int64_t size;
3866     V9fsString name;
3867     ssize_t err = 0;
3868     size_t offset = 7;
3869     int32_t fid, newfid;
3870     V9fsFidState *file_fidp;
3871     V9fsFidState *xattr_fidp = NULL;
3872     V9fsPDU *pdu = opaque;
3873     V9fsState *s = pdu->s;
3874 
3875     v9fs_string_init(&name);
3876     err = pdu_unmarshal(pdu, offset, "dds", &fid, &newfid, &name);
3877     if (err < 0) {
3878         goto out_nofid;
3879     }
3880     trace_v9fs_xattrwalk(pdu->tag, pdu->id, fid, newfid, name.data);
3881 
3882     file_fidp = get_fid(pdu, fid);
3883     if (file_fidp == NULL) {
3884         err = -ENOENT;
3885         goto out_nofid;
3886     }
3887     xattr_fidp = alloc_fid(s, newfid);
3888     if (xattr_fidp == NULL) {
3889         err = -EINVAL;
3890         goto out;
3891     }
3892     v9fs_path_copy(&xattr_fidp->path, &file_fidp->path);
3893     if (!v9fs_string_size(&name)) {
3894         /*
3895          * listxattr request. Get the size first
3896          */
3897         size = v9fs_co_llistxattr(pdu, &xattr_fidp->path, NULL, 0);
3898         if (size < 0) {
3899             err = size;
3900             clunk_fid(s, xattr_fidp->fid);
3901             goto out;
3902         }
3903         /*
3904          * Read the xattr value
3905          */
3906         xattr_fidp->fs.xattr.len = size;
3907         xattr_fidp->fid_type = P9_FID_XATTR;
3908         xattr_fidp->fs.xattr.xattrwalk_fid = true;
3909         xattr_fidp->fs.xattr.value = g_malloc0(size);
3910         if (size) {
3911             err = v9fs_co_llistxattr(pdu, &xattr_fidp->path,
3912                                      xattr_fidp->fs.xattr.value,
3913                                      xattr_fidp->fs.xattr.len);
3914             if (err < 0) {
3915                 clunk_fid(s, xattr_fidp->fid);
3916                 goto out;
3917             }
3918         }
3919         err = pdu_marshal(pdu, offset, "q", size);
3920         if (err < 0) {
3921             goto out;
3922         }
3923         err += offset;
3924     } else {
3925         /*
3926          * specific xattr fid. We check for xattr
3927          * presence also collect the xattr size
3928          */
3929         size = v9fs_co_lgetxattr(pdu, &xattr_fidp->path,
3930                                  &name, NULL, 0);
3931         if (size < 0) {
3932             err = size;
3933             clunk_fid(s, xattr_fidp->fid);
3934             goto out;
3935         }
3936         /*
3937          * Read the xattr value
3938          */
3939         xattr_fidp->fs.xattr.len = size;
3940         xattr_fidp->fid_type = P9_FID_XATTR;
3941         xattr_fidp->fs.xattr.xattrwalk_fid = true;
3942         xattr_fidp->fs.xattr.value = g_malloc0(size);
3943         if (size) {
3944             err = v9fs_co_lgetxattr(pdu, &xattr_fidp->path,
3945                                     &name, xattr_fidp->fs.xattr.value,
3946                                     xattr_fidp->fs.xattr.len);
3947             if (err < 0) {
3948                 clunk_fid(s, xattr_fidp->fid);
3949                 goto out;
3950             }
3951         }
3952         err = pdu_marshal(pdu, offset, "q", size);
3953         if (err < 0) {
3954             goto out;
3955         }
3956         err += offset;
3957     }
3958     trace_v9fs_xattrwalk_return(pdu->tag, pdu->id, size);
3959 out:
3960     put_fid(pdu, file_fidp);
3961     if (xattr_fidp) {
3962         put_fid(pdu, xattr_fidp);
3963     }
3964 out_nofid:
3965     pdu_complete(pdu, err);
3966     v9fs_string_free(&name);
3967 }
3968 
3969 #if defined(CONFIG_LINUX)
3970 /* Currently, only Linux has XATTR_SIZE_MAX */
3971 #define P9_XATTR_SIZE_MAX XATTR_SIZE_MAX
3972 #elif defined(CONFIG_DARWIN)
3973 /*
3974  * Darwin doesn't seem to define a maximum xattr size in its user
3975  * space header, so manually configure it across platforms as 64k.
3976  *
3977  * Having no limit at all can lead to QEMU crashing during large g_malloc()
3978  * calls. Because QEMU does not currently support macOS guests, the below
3979  * preliminary solution only works due to its being a reflection of the limit of
3980  * Linux guests.
3981  */
3982 #define P9_XATTR_SIZE_MAX 65536
3983 #else
3984 #error Missing definition for P9_XATTR_SIZE_MAX for this host system
3985 #endif
3986 
3987 static void coroutine_fn v9fs_xattrcreate(void *opaque)
3988 {
3989     int flags, rflags = 0;
3990     int32_t fid;
3991     uint64_t size;
3992     ssize_t err = 0;
3993     V9fsString name;
3994     size_t offset = 7;
3995     V9fsFidState *file_fidp;
3996     V9fsFidState *xattr_fidp;
3997     V9fsPDU *pdu = opaque;
3998 
3999     v9fs_string_init(&name);
4000     err = pdu_unmarshal(pdu, offset, "dsqd", &fid, &name, &size, &flags);
4001     if (err < 0) {
4002         goto out_nofid;
4003     }
4004     trace_v9fs_xattrcreate(pdu->tag, pdu->id, fid, name.data, size, flags);
4005 
4006     if (flags & ~(P9_XATTR_CREATE | P9_XATTR_REPLACE)) {
4007         err = -EINVAL;
4008         goto out_nofid;
4009     }
4010 
4011     if (flags & P9_XATTR_CREATE) {
4012         rflags |= XATTR_CREATE;
4013     }
4014 
4015     if (flags & P9_XATTR_REPLACE) {
4016         rflags |= XATTR_REPLACE;
4017     }
4018 
4019     if (size > P9_XATTR_SIZE_MAX) {
4020         err = -E2BIG;
4021         goto out_nofid;
4022     }
4023 
4024     file_fidp = get_fid(pdu, fid);
4025     if (file_fidp == NULL) {
4026         err = -EINVAL;
4027         goto out_nofid;
4028     }
4029     if (file_fidp->fid_type != P9_FID_NONE) {
4030         err = -EINVAL;
4031         goto out_put_fid;
4032     }
4033 
4034     /* Make the file fid point to xattr */
4035     xattr_fidp = file_fidp;
4036     xattr_fidp->fid_type = P9_FID_XATTR;
4037     xattr_fidp->fs.xattr.copied_len = 0;
4038     xattr_fidp->fs.xattr.xattrwalk_fid = false;
4039     xattr_fidp->fs.xattr.len = size;
4040     xattr_fidp->fs.xattr.flags = rflags;
4041     v9fs_string_init(&xattr_fidp->fs.xattr.name);
4042     v9fs_string_copy(&xattr_fidp->fs.xattr.name, &name);
4043     xattr_fidp->fs.xattr.value = g_malloc0(size);
4044     err = offset;
4045 out_put_fid:
4046     put_fid(pdu, file_fidp);
4047 out_nofid:
4048     pdu_complete(pdu, err);
4049     v9fs_string_free(&name);
4050 }
4051 
4052 static void coroutine_fn v9fs_readlink(void *opaque)
4053 {
4054     V9fsPDU *pdu = opaque;
4055     size_t offset = 7;
4056     V9fsString target;
4057     int32_t fid;
4058     int err = 0;
4059     V9fsFidState *fidp;
4060 
4061     err = pdu_unmarshal(pdu, offset, "d", &fid);
4062     if (err < 0) {
4063         goto out_nofid;
4064     }
4065     trace_v9fs_readlink(pdu->tag, pdu->id, fid);
4066     fidp = get_fid(pdu, fid);
4067     if (fidp == NULL) {
4068         err = -ENOENT;
4069         goto out_nofid;
4070     }
4071 
4072     v9fs_string_init(&target);
4073     err = v9fs_co_readlink(pdu, &fidp->path, &target);
4074     if (err < 0) {
4075         goto out;
4076     }
4077     err = pdu_marshal(pdu, offset, "s", &target);
4078     if (err < 0) {
4079         v9fs_string_free(&target);
4080         goto out;
4081     }
4082     err += offset;
4083     trace_v9fs_readlink_return(pdu->tag, pdu->id, target.data);
4084     v9fs_string_free(&target);
4085 out:
4086     put_fid(pdu, fidp);
4087 out_nofid:
4088     pdu_complete(pdu, err);
4089 }
4090 
4091 static CoroutineEntry *pdu_co_handlers[] = {
4092     [P9_TREADDIR] = v9fs_readdir,
4093     [P9_TSTATFS] = v9fs_statfs,
4094     [P9_TGETATTR] = v9fs_getattr,
4095     [P9_TSETATTR] = v9fs_setattr,
4096     [P9_TXATTRWALK] = v9fs_xattrwalk,
4097     [P9_TXATTRCREATE] = v9fs_xattrcreate,
4098     [P9_TMKNOD] = v9fs_mknod,
4099     [P9_TRENAME] = v9fs_rename,
4100     [P9_TLOCK] = v9fs_lock,
4101     [P9_TGETLOCK] = v9fs_getlock,
4102     [P9_TRENAMEAT] = v9fs_renameat,
4103     [P9_TREADLINK] = v9fs_readlink,
4104     [P9_TUNLINKAT] = v9fs_unlinkat,
4105     [P9_TMKDIR] = v9fs_mkdir,
4106     [P9_TVERSION] = v9fs_version,
4107     [P9_TLOPEN] = v9fs_open,
4108     [P9_TATTACH] = v9fs_attach,
4109     [P9_TSTAT] = v9fs_stat,
4110     [P9_TWALK] = v9fs_walk,
4111     [P9_TCLUNK] = v9fs_clunk,
4112     [P9_TFSYNC] = v9fs_fsync,
4113     [P9_TOPEN] = v9fs_open,
4114     [P9_TREAD] = v9fs_read,
4115 #if 0
4116     [P9_TAUTH] = v9fs_auth,
4117 #endif
4118     [P9_TFLUSH] = v9fs_flush,
4119     [P9_TLINK] = v9fs_link,
4120     [P9_TSYMLINK] = v9fs_symlink,
4121     [P9_TCREATE] = v9fs_create,
4122     [P9_TLCREATE] = v9fs_lcreate,
4123     [P9_TWRITE] = v9fs_write,
4124     [P9_TWSTAT] = v9fs_wstat,
4125     [P9_TREMOVE] = v9fs_remove,
4126 };
4127 
4128 static void coroutine_fn v9fs_op_not_supp(void *opaque)
4129 {
4130     V9fsPDU *pdu = opaque;
4131     pdu_complete(pdu, -EOPNOTSUPP);
4132 }
4133 
4134 static void coroutine_fn v9fs_fs_ro(void *opaque)
4135 {
4136     V9fsPDU *pdu = opaque;
4137     pdu_complete(pdu, -EROFS);
4138 }
4139 
4140 static inline bool is_read_only_op(V9fsPDU *pdu)
4141 {
4142     switch (pdu->id) {
4143     case P9_TREADDIR:
4144     case P9_TSTATFS:
4145     case P9_TGETATTR:
4146     case P9_TXATTRWALK:
4147     case P9_TLOCK:
4148     case P9_TGETLOCK:
4149     case P9_TREADLINK:
4150     case P9_TVERSION:
4151     case P9_TLOPEN:
4152     case P9_TATTACH:
4153     case P9_TSTAT:
4154     case P9_TWALK:
4155     case P9_TCLUNK:
4156     case P9_TFSYNC:
4157     case P9_TOPEN:
4158     case P9_TREAD:
4159     case P9_TAUTH:
4160     case P9_TFLUSH:
4161         return 1;
4162     default:
4163         return 0;
4164     }
4165 }
4166 
4167 void pdu_submit(V9fsPDU *pdu, P9MsgHeader *hdr)
4168 {
4169     Coroutine *co;
4170     CoroutineEntry *handler;
4171     V9fsState *s = pdu->s;
4172 
4173     pdu->size = le32_to_cpu(hdr->size_le);
4174     pdu->id = hdr->id;
4175     pdu->tag = le16_to_cpu(hdr->tag_le);
4176 
4177     if (pdu->id >= ARRAY_SIZE(pdu_co_handlers) ||
4178         (pdu_co_handlers[pdu->id] == NULL)) {
4179         handler = v9fs_op_not_supp;
4180     } else if (is_ro_export(&s->ctx) && !is_read_only_op(pdu)) {
4181         handler = v9fs_fs_ro;
4182     } else {
4183         handler = pdu_co_handlers[pdu->id];
4184     }
4185 
4186     qemu_co_queue_init(&pdu->complete);
4187     co = qemu_coroutine_create(handler, pdu);
4188     qemu_coroutine_enter(co);
4189 }
4190 
4191 /* Returns 0 on success, 1 on failure. */
4192 int v9fs_device_realize_common(V9fsState *s, const V9fsTransport *t,
4193                                Error **errp)
4194 {
4195     ERRP_GUARD();
4196     int i, len;
4197     struct stat stat;
4198     FsDriverEntry *fse;
4199     V9fsPath path;
4200     int rc = 1;
4201 
4202     assert(!s->transport);
4203     s->transport = t;
4204 
4205     /* initialize pdu allocator */
4206     QLIST_INIT(&s->free_list);
4207     QLIST_INIT(&s->active_list);
4208     for (i = 0; i < MAX_REQ; i++) {
4209         QLIST_INSERT_HEAD(&s->free_list, &s->pdus[i], next);
4210         s->pdus[i].s = s;
4211         s->pdus[i].idx = i;
4212     }
4213 
4214     v9fs_path_init(&path);
4215 
4216     fse = get_fsdev_fsentry(s->fsconf.fsdev_id);
4217 
4218     if (!fse) {
4219         /* We don't have a fsdev identified by fsdev_id */
4220         error_setg(errp, "9pfs device couldn't find fsdev with the "
4221                    "id = %s",
4222                    s->fsconf.fsdev_id ? s->fsconf.fsdev_id : "NULL");
4223         goto out;
4224     }
4225 
4226     if (!s->fsconf.tag) {
4227         /* we haven't specified a mount_tag */
4228         error_setg(errp, "fsdev with id %s needs mount_tag arguments",
4229                    s->fsconf.fsdev_id);
4230         goto out;
4231     }
4232 
4233     s->ctx.export_flags = fse->export_flags;
4234     s->ctx.fs_root = g_strdup(fse->path);
4235     s->ctx.exops.get_st_gen = NULL;
4236     len = strlen(s->fsconf.tag);
4237     if (len > MAX_TAG_LEN - 1) {
4238         error_setg(errp, "mount tag '%s' (%d bytes) is longer than "
4239                    "maximum (%d bytes)", s->fsconf.tag, len, MAX_TAG_LEN - 1);
4240         goto out;
4241     }
4242 
4243     s->tag = g_strdup(s->fsconf.tag);
4244     s->ctx.uid = -1;
4245 
4246     s->ops = fse->ops;
4247 
4248     s->ctx.fmode = fse->fmode;
4249     s->ctx.dmode = fse->dmode;
4250 
4251     s->fids = g_hash_table_new(NULL, NULL);
4252     qemu_co_rwlock_init(&s->rename_lock);
4253 
4254     if (s->ops->init(&s->ctx, errp) < 0) {
4255         error_prepend(errp, "cannot initialize fsdev '%s': ",
4256                       s->fsconf.fsdev_id);
4257         goto out;
4258     }
4259 
4260     /*
4261      * Check details of export path, We need to use fs driver
4262      * call back to do that. Since we are in the init path, we don't
4263      * use co-routines here.
4264      */
4265     if (s->ops->name_to_path(&s->ctx, NULL, "/", &path) < 0) {
4266         error_setg(errp,
4267                    "error in converting name to path %s", strerror(errno));
4268         goto out;
4269     }
4270     if (s->ops->lstat(&s->ctx, &path, &stat)) {
4271         error_setg(errp, "share path %s does not exist", fse->path);
4272         goto out;
4273     } else if (!S_ISDIR(stat.st_mode)) {
4274         error_setg(errp, "share path %s is not a directory", fse->path);
4275         goto out;
4276     }
4277 
4278     s->dev_id = stat.st_dev;
4279 
4280     /* init inode remapping : */
4281     /* hash table for variable length inode suffixes */
4282     qpd_table_init(&s->qpd_table);
4283     /* hash table for slow/full inode remapping (most users won't need it) */
4284     qpf_table_init(&s->qpf_table);
4285     /* hash table for quick inode remapping */
4286     qpp_table_init(&s->qpp_table);
4287     s->qp_ndevices = 0;
4288     s->qp_affix_next = 1; /* reserve 0 to detect overflow */
4289     s->qp_fullpath_next = 1;
4290 
4291     s->ctx.fst = &fse->fst;
4292     fsdev_throttle_init(s->ctx.fst);
4293 
4294     rc = 0;
4295 out:
4296     if (rc) {
4297         v9fs_device_unrealize_common(s);
4298     }
4299     v9fs_path_free(&path);
4300     return rc;
4301 }
4302 
4303 void v9fs_device_unrealize_common(V9fsState *s)
4304 {
4305     if (s->ops && s->ops->cleanup) {
4306         s->ops->cleanup(&s->ctx);
4307     }
4308     if (s->ctx.fst) {
4309         fsdev_throttle_cleanup(s->ctx.fst);
4310     }
4311     if (s->fids) {
4312         g_hash_table_destroy(s->fids);
4313         s->fids = NULL;
4314     }
4315     g_free(s->tag);
4316     qp_table_destroy(&s->qpd_table);
4317     qp_table_destroy(&s->qpp_table);
4318     qp_table_destroy(&s->qpf_table);
4319     g_free(s->ctx.fs_root);
4320 }
4321 
4322 typedef struct VirtfsCoResetData {
4323     V9fsPDU pdu;
4324     bool done;
4325 } VirtfsCoResetData;
4326 
4327 static void coroutine_fn virtfs_co_reset(void *opaque)
4328 {
4329     VirtfsCoResetData *data = opaque;
4330 
4331     virtfs_reset(&data->pdu);
4332     data->done = true;
4333 }
4334 
4335 void v9fs_reset(V9fsState *s)
4336 {
4337     VirtfsCoResetData data = { .pdu = { .s = s }, .done = false };
4338     Coroutine *co;
4339 
4340     while (!QLIST_EMPTY(&s->active_list)) {
4341         aio_poll(qemu_get_aio_context(), true);
4342     }
4343 
4344     co = qemu_coroutine_create(virtfs_co_reset, &data);
4345     qemu_coroutine_enter(co);
4346 
4347     while (!data.done) {
4348         aio_poll(qemu_get_aio_context(), true);
4349     }
4350 }
4351 
4352 static void __attribute__((__constructor__)) v9fs_set_fd_limit(void)
4353 {
4354     struct rlimit rlim;
4355     if (getrlimit(RLIMIT_NOFILE, &rlim) < 0) {
4356         error_report("Failed to get the resource limit");
4357         exit(1);
4358     }
4359     open_fd_hw = rlim.rlim_cur - MIN(400, rlim.rlim_cur / 3);
4360     open_fd_rc = rlim.rlim_cur / 2;
4361 }
4362