xref: /openbmc/qemu/linux-user/mmap.c (revision 0dacec87)
1 /*
2  *  mmap support for qemu
3  *
4  *  Copyright (c) 2003 Fabrice Bellard
5  *
6  *  This program is free software; you can redistribute it and/or modify
7  *  it under the terms of the GNU General Public License as published by
8  *  the Free Software Foundation; either version 2 of the License, or
9  *  (at your option) any later version.
10  *
11  *  This program is distributed in the hope that it will be useful,
12  *  but WITHOUT ANY WARRANTY; without even the implied warranty of
13  *  MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
14  *  GNU General Public License for more details.
15  *
16  *  You should have received a copy of the GNU General Public License
17  *  along with this program; if not, see <http://www.gnu.org/licenses/>.
18  */
19 #include "qemu/osdep.h"
20 
21 #include "qemu.h"
22 #include "qemu-common.h"
23 #include "translate-all.h"
24 
25 //#define DEBUG_MMAP
26 
27 static pthread_mutex_t mmap_mutex = PTHREAD_MUTEX_INITIALIZER;
28 static __thread int mmap_lock_count;
29 
30 void mmap_lock(void)
31 {
32     if (mmap_lock_count++ == 0) {
33         pthread_mutex_lock(&mmap_mutex);
34     }
35 }
36 
37 void mmap_unlock(void)
38 {
39     if (--mmap_lock_count == 0) {
40         pthread_mutex_unlock(&mmap_mutex);
41     }
42 }
43 
44 bool have_mmap_lock(void)
45 {
46     return mmap_lock_count > 0 ? true : false;
47 }
48 
49 /* Grab lock to make sure things are in a consistent state after fork().  */
50 void mmap_fork_start(void)
51 {
52     if (mmap_lock_count)
53         abort();
54     pthread_mutex_lock(&mmap_mutex);
55 }
56 
57 void mmap_fork_end(int child)
58 {
59     if (child)
60         pthread_mutex_init(&mmap_mutex, NULL);
61     else
62         pthread_mutex_unlock(&mmap_mutex);
63 }
64 
65 /* NOTE: all the constants are the HOST ones, but addresses are target. */
66 int target_mprotect(abi_ulong start, abi_ulong len, int prot)
67 {
68     abi_ulong end, host_start, host_end, addr;
69     int prot1, ret;
70 
71 #ifdef DEBUG_MMAP
72     printf("mprotect: start=0x" TARGET_ABI_FMT_lx
73            "len=0x" TARGET_ABI_FMT_lx " prot=%c%c%c\n", start, len,
74            prot & PROT_READ ? 'r' : '-',
75            prot & PROT_WRITE ? 'w' : '-',
76            prot & PROT_EXEC ? 'x' : '-');
77 #endif
78 
79     if ((start & ~TARGET_PAGE_MASK) != 0)
80         return -TARGET_EINVAL;
81     len = TARGET_PAGE_ALIGN(len);
82     end = start + len;
83     if (!guest_range_valid(start, len)) {
84         return -TARGET_ENOMEM;
85     }
86     prot &= PROT_READ | PROT_WRITE | PROT_EXEC;
87     if (len == 0)
88         return 0;
89 
90     mmap_lock();
91     host_start = start & qemu_host_page_mask;
92     host_end = HOST_PAGE_ALIGN(end);
93     if (start > host_start) {
94         /* handle host page containing start */
95         prot1 = prot;
96         for(addr = host_start; addr < start; addr += TARGET_PAGE_SIZE) {
97             prot1 |= page_get_flags(addr);
98         }
99         if (host_end == host_start + qemu_host_page_size) {
100             for(addr = end; addr < host_end; addr += TARGET_PAGE_SIZE) {
101                 prot1 |= page_get_flags(addr);
102             }
103             end = host_end;
104         }
105         ret = mprotect(g2h(host_start), qemu_host_page_size, prot1 & PAGE_BITS);
106         if (ret != 0)
107             goto error;
108         host_start += qemu_host_page_size;
109     }
110     if (end < host_end) {
111         prot1 = prot;
112         for(addr = end; addr < host_end; addr += TARGET_PAGE_SIZE) {
113             prot1 |= page_get_flags(addr);
114         }
115         ret = mprotect(g2h(host_end - qemu_host_page_size), qemu_host_page_size,
116                        prot1 & PAGE_BITS);
117         if (ret != 0)
118             goto error;
119         host_end -= qemu_host_page_size;
120     }
121 
122     /* handle the pages in the middle */
123     if (host_start < host_end) {
124         ret = mprotect(g2h(host_start), host_end - host_start, prot);
125         if (ret != 0)
126             goto error;
127     }
128     page_set_flags(start, start + len, prot | PAGE_VALID);
129     mmap_unlock();
130     return 0;
131 error:
132     mmap_unlock();
133     return ret;
134 }
135 
136 /* map an incomplete host page */
137 static int mmap_frag(abi_ulong real_start,
138                      abi_ulong start, abi_ulong end,
139                      int prot, int flags, int fd, abi_ulong offset)
140 {
141     abi_ulong real_end, addr;
142     void *host_start;
143     int prot1, prot_new;
144 
145     real_end = real_start + qemu_host_page_size;
146     host_start = g2h(real_start);
147 
148     /* get the protection of the target pages outside the mapping */
149     prot1 = 0;
150     for(addr = real_start; addr < real_end; addr++) {
151         if (addr < start || addr >= end)
152             prot1 |= page_get_flags(addr);
153     }
154 
155     if (prot1 == 0) {
156         /* no page was there, so we allocate one */
157         void *p = mmap(host_start, qemu_host_page_size, prot,
158                        flags | MAP_ANONYMOUS, -1, 0);
159         if (p == MAP_FAILED)
160             return -1;
161         prot1 = prot;
162     }
163     prot1 &= PAGE_BITS;
164 
165     prot_new = prot | prot1;
166     if (!(flags & MAP_ANONYMOUS)) {
167         /* msync() won't work here, so we return an error if write is
168            possible while it is a shared mapping */
169         if ((flags & MAP_TYPE) == MAP_SHARED &&
170             (prot & PROT_WRITE))
171             return -1;
172 
173         /* adjust protection to be able to read */
174         if (!(prot1 & PROT_WRITE))
175             mprotect(host_start, qemu_host_page_size, prot1 | PROT_WRITE);
176 
177         /* read the corresponding file data */
178         if (pread(fd, g2h(start), end - start, offset) == -1)
179             return -1;
180 
181         /* put final protection */
182         if (prot_new != (prot1 | PROT_WRITE))
183             mprotect(host_start, qemu_host_page_size, prot_new);
184     } else {
185         if (prot_new != prot1) {
186             mprotect(host_start, qemu_host_page_size, prot_new);
187         }
188         if (prot_new & PROT_WRITE) {
189             memset(g2h(start), 0, end - start);
190         }
191     }
192     return 0;
193 }
194 
195 #if HOST_LONG_BITS == 64 && TARGET_ABI_BITS == 64
196 # define TASK_UNMAPPED_BASE  (1ul << 38)
197 #else
198 # define TASK_UNMAPPED_BASE  0x40000000
199 #endif
200 abi_ulong mmap_next_start = TASK_UNMAPPED_BASE;
201 
202 unsigned long last_brk;
203 
204 /* Subroutine of mmap_find_vma, used when we have pre-allocated a chunk
205    of guest address space.  */
206 static abi_ulong mmap_find_vma_reserved(abi_ulong start, abi_ulong size)
207 {
208     abi_ulong addr;
209     abi_ulong end_addr;
210     int prot;
211     int looped = 0;
212 
213     if (size > reserved_va) {
214         return (abi_ulong)-1;
215     }
216 
217     size = HOST_PAGE_ALIGN(size);
218     end_addr = start + size;
219     if (end_addr > reserved_va) {
220         end_addr = reserved_va;
221     }
222     addr = end_addr - qemu_host_page_size;
223 
224     while (1) {
225         if (addr > end_addr) {
226             if (looped) {
227                 return (abi_ulong)-1;
228             }
229             end_addr = reserved_va;
230             addr = end_addr - qemu_host_page_size;
231             looped = 1;
232             continue;
233         }
234         prot = page_get_flags(addr);
235         if (prot) {
236             end_addr = addr;
237         }
238         if (addr && addr + size == end_addr) {
239             break;
240         }
241         addr -= qemu_host_page_size;
242     }
243 
244     if (start == mmap_next_start) {
245         mmap_next_start = addr;
246     }
247 
248     return addr;
249 }
250 
251 /*
252  * Find and reserve a free memory area of size 'size'. The search
253  * starts at 'start'.
254  * It must be called with mmap_lock() held.
255  * Return -1 if error.
256  */
257 abi_ulong mmap_find_vma(abi_ulong start, abi_ulong size)
258 {
259     void *ptr, *prev;
260     abi_ulong addr;
261     int wrapped, repeat;
262 
263     /* If 'start' == 0, then a default start address is used. */
264     if (start == 0) {
265         start = mmap_next_start;
266     } else {
267         start &= qemu_host_page_mask;
268     }
269 
270     size = HOST_PAGE_ALIGN(size);
271 
272     if (reserved_va) {
273         return mmap_find_vma_reserved(start, size);
274     }
275 
276     addr = start;
277     wrapped = repeat = 0;
278     prev = 0;
279 
280     for (;; prev = ptr) {
281         /*
282          * Reserve needed memory area to avoid a race.
283          * It should be discarded using:
284          *  - mmap() with MAP_FIXED flag
285          *  - mremap() with MREMAP_FIXED flag
286          *  - shmat() with SHM_REMAP flag
287          */
288         ptr = mmap(g2h(addr), size, PROT_NONE,
289                    MAP_ANONYMOUS|MAP_PRIVATE|MAP_NORESERVE, -1, 0);
290 
291         /* ENOMEM, if host address space has no memory */
292         if (ptr == MAP_FAILED) {
293             return (abi_ulong)-1;
294         }
295 
296         /* Count the number of sequential returns of the same address.
297            This is used to modify the search algorithm below.  */
298         repeat = (ptr == prev ? repeat + 1 : 0);
299 
300         if (h2g_valid(ptr + size - 1)) {
301             addr = h2g(ptr);
302 
303             if ((addr & ~TARGET_PAGE_MASK) == 0) {
304                 /* Success.  */
305                 if (start == mmap_next_start && addr >= TASK_UNMAPPED_BASE) {
306                     mmap_next_start = addr + size;
307                 }
308                 return addr;
309             }
310 
311             /* The address is not properly aligned for the target.  */
312             switch (repeat) {
313             case 0:
314                 /* Assume the result that the kernel gave us is the
315                    first with enough free space, so start again at the
316                    next higher target page.  */
317                 addr = TARGET_PAGE_ALIGN(addr);
318                 break;
319             case 1:
320                 /* Sometimes the kernel decides to perform the allocation
321                    at the top end of memory instead.  */
322                 addr &= TARGET_PAGE_MASK;
323                 break;
324             case 2:
325                 /* Start over at low memory.  */
326                 addr = 0;
327                 break;
328             default:
329                 /* Fail.  This unaligned block must the last.  */
330                 addr = -1;
331                 break;
332             }
333         } else {
334             /* Since the result the kernel gave didn't fit, start
335                again at low memory.  If any repetition, fail.  */
336             addr = (repeat ? -1 : 0);
337         }
338 
339         /* Unmap and try again.  */
340         munmap(ptr, size);
341 
342         /* ENOMEM if we checked the whole of the target address space.  */
343         if (addr == (abi_ulong)-1) {
344             return (abi_ulong)-1;
345         } else if (addr == 0) {
346             if (wrapped) {
347                 return (abi_ulong)-1;
348             }
349             wrapped = 1;
350             /* Don't actually use 0 when wrapping, instead indicate
351                that we'd truly like an allocation in low memory.  */
352             addr = (mmap_min_addr > TARGET_PAGE_SIZE
353                      ? TARGET_PAGE_ALIGN(mmap_min_addr)
354                      : TARGET_PAGE_SIZE);
355         } else if (wrapped && addr >= start) {
356             return (abi_ulong)-1;
357         }
358     }
359 }
360 
361 /* NOTE: all the constants are the HOST ones */
362 abi_long target_mmap(abi_ulong start, abi_ulong len, int prot,
363                      int flags, int fd, abi_ulong offset)
364 {
365     abi_ulong ret, end, real_start, real_end, retaddr, host_offset, host_len;
366 
367     mmap_lock();
368 #ifdef DEBUG_MMAP
369     {
370         printf("mmap: start=0x" TARGET_ABI_FMT_lx
371                " len=0x" TARGET_ABI_FMT_lx " prot=%c%c%c flags=",
372                start, len,
373                prot & PROT_READ ? 'r' : '-',
374                prot & PROT_WRITE ? 'w' : '-',
375                prot & PROT_EXEC ? 'x' : '-');
376         if (flags & MAP_FIXED)
377             printf("MAP_FIXED ");
378         if (flags & MAP_ANONYMOUS)
379             printf("MAP_ANON ");
380         switch(flags & MAP_TYPE) {
381         case MAP_PRIVATE:
382             printf("MAP_PRIVATE ");
383             break;
384         case MAP_SHARED:
385             printf("MAP_SHARED ");
386             break;
387         default:
388             printf("[MAP_TYPE=0x%x] ", flags & MAP_TYPE);
389             break;
390         }
391         printf("fd=%d offset=" TARGET_ABI_FMT_lx "\n", fd, offset);
392     }
393 #endif
394 
395     if (offset & ~TARGET_PAGE_MASK) {
396         errno = EINVAL;
397         goto fail;
398     }
399 
400     len = TARGET_PAGE_ALIGN(len);
401     if (len == 0)
402         goto the_end;
403     real_start = start & qemu_host_page_mask;
404     host_offset = offset & qemu_host_page_mask;
405 
406     /* If the user is asking for the kernel to find a location, do that
407        before we truncate the length for mapping files below.  */
408     if (!(flags & MAP_FIXED)) {
409         host_len = len + offset - host_offset;
410         host_len = HOST_PAGE_ALIGN(host_len);
411         start = mmap_find_vma(real_start, host_len);
412         if (start == (abi_ulong)-1) {
413             errno = ENOMEM;
414             goto fail;
415         }
416     }
417 
418     /* When mapping files into a memory area larger than the file, accesses
419        to pages beyond the file size will cause a SIGBUS.
420 
421        For example, if mmaping a file of 100 bytes on a host with 4K pages
422        emulating a target with 8K pages, the target expects to be able to
423        access the first 8K. But the host will trap us on any access beyond
424        4K.
425 
426        When emulating a target with a larger page-size than the hosts, we
427        may need to truncate file maps at EOF and add extra anonymous pages
428        up to the targets page boundary.  */
429 
430     if ((qemu_real_host_page_size < qemu_host_page_size) &&
431         !(flags & MAP_ANONYMOUS)) {
432         struct stat sb;
433 
434        if (fstat (fd, &sb) == -1)
435            goto fail;
436 
437        /* Are we trying to create a map beyond EOF?.  */
438        if (offset + len > sb.st_size) {
439            /* If so, truncate the file map at eof aligned with
440               the hosts real pagesize. Additional anonymous maps
441               will be created beyond EOF.  */
442            len = REAL_HOST_PAGE_ALIGN(sb.st_size - offset);
443        }
444     }
445 
446     if (!(flags & MAP_FIXED)) {
447         unsigned long host_start;
448         void *p;
449 
450         host_len = len + offset - host_offset;
451         host_len = HOST_PAGE_ALIGN(host_len);
452 
453         /* Note: we prefer to control the mapping address. It is
454            especially important if qemu_host_page_size >
455            qemu_real_host_page_size */
456         p = mmap(g2h(start), host_len, prot,
457                  flags | MAP_FIXED | MAP_ANONYMOUS, -1, 0);
458         if (p == MAP_FAILED)
459             goto fail;
460         /* update start so that it points to the file position at 'offset' */
461         host_start = (unsigned long)p;
462         if (!(flags & MAP_ANONYMOUS)) {
463             p = mmap(g2h(start), len, prot,
464                      flags | MAP_FIXED, fd, host_offset);
465             if (p == MAP_FAILED) {
466                 munmap(g2h(start), host_len);
467                 goto fail;
468             }
469             host_start += offset - host_offset;
470         }
471         start = h2g(host_start);
472     } else {
473         if (start & ~TARGET_PAGE_MASK) {
474             errno = EINVAL;
475             goto fail;
476         }
477         end = start + len;
478         real_end = HOST_PAGE_ALIGN(end);
479 
480 	/*
481 	 * Test if requested memory area fits target address space
482 	 * It can fail only on 64-bit host with 32-bit target.
483 	 * On any other target/host host mmap() handles this error correctly.
484 	 */
485         if (!guest_range_valid(start, len)) {
486             errno = ENOMEM;
487             goto fail;
488         }
489 
490         /* worst case: we cannot map the file because the offset is not
491            aligned, so we read it */
492         if (!(flags & MAP_ANONYMOUS) &&
493             (offset & ~qemu_host_page_mask) != (start & ~qemu_host_page_mask)) {
494             /* msync() won't work here, so we return an error if write is
495                possible while it is a shared mapping */
496             if ((flags & MAP_TYPE) == MAP_SHARED &&
497                 (prot & PROT_WRITE)) {
498                 errno = EINVAL;
499                 goto fail;
500             }
501             retaddr = target_mmap(start, len, prot | PROT_WRITE,
502                                   MAP_FIXED | MAP_PRIVATE | MAP_ANONYMOUS,
503                                   -1, 0);
504             if (retaddr == -1)
505                 goto fail;
506             if (pread(fd, g2h(start), len, offset) == -1)
507                 goto fail;
508             if (!(prot & PROT_WRITE)) {
509                 ret = target_mprotect(start, len, prot);
510                 assert(ret == 0);
511             }
512             goto the_end;
513         }
514 
515         /* handle the start of the mapping */
516         if (start > real_start) {
517             if (real_end == real_start + qemu_host_page_size) {
518                 /* one single host page */
519                 ret = mmap_frag(real_start, start, end,
520                                 prot, flags, fd, offset);
521                 if (ret == -1)
522                     goto fail;
523                 goto the_end1;
524             }
525             ret = mmap_frag(real_start, start, real_start + qemu_host_page_size,
526                             prot, flags, fd, offset);
527             if (ret == -1)
528                 goto fail;
529             real_start += qemu_host_page_size;
530         }
531         /* handle the end of the mapping */
532         if (end < real_end) {
533             ret = mmap_frag(real_end - qemu_host_page_size,
534                             real_end - qemu_host_page_size, end,
535                             prot, flags, fd,
536                             offset + real_end - qemu_host_page_size - start);
537             if (ret == -1)
538                 goto fail;
539             real_end -= qemu_host_page_size;
540         }
541 
542         /* map the middle (easier) */
543         if (real_start < real_end) {
544             void *p;
545             unsigned long offset1;
546             if (flags & MAP_ANONYMOUS)
547                 offset1 = 0;
548             else
549                 offset1 = offset + real_start - start;
550             p = mmap(g2h(real_start), real_end - real_start,
551                      prot, flags, fd, offset1);
552             if (p == MAP_FAILED)
553                 goto fail;
554         }
555     }
556  the_end1:
557     page_set_flags(start, start + len, prot | PAGE_VALID);
558  the_end:
559 #ifdef DEBUG_MMAP
560     printf("ret=0x" TARGET_ABI_FMT_lx "\n", start);
561     page_dump(stdout);
562     printf("\n");
563 #endif
564     tb_invalidate_phys_range(start, start + len);
565     mmap_unlock();
566     return start;
567 fail:
568     mmap_unlock();
569     return -1;
570 }
571 
572 static void mmap_reserve(abi_ulong start, abi_ulong size)
573 {
574     abi_ulong real_start;
575     abi_ulong real_end;
576     abi_ulong addr;
577     abi_ulong end;
578     int prot;
579 
580     real_start = start & qemu_host_page_mask;
581     real_end = HOST_PAGE_ALIGN(start + size);
582     end = start + size;
583     if (start > real_start) {
584         /* handle host page containing start */
585         prot = 0;
586         for (addr = real_start; addr < start; addr += TARGET_PAGE_SIZE) {
587             prot |= page_get_flags(addr);
588         }
589         if (real_end == real_start + qemu_host_page_size) {
590             for (addr = end; addr < real_end; addr += TARGET_PAGE_SIZE) {
591                 prot |= page_get_flags(addr);
592             }
593             end = real_end;
594         }
595         if (prot != 0)
596             real_start += qemu_host_page_size;
597     }
598     if (end < real_end) {
599         prot = 0;
600         for (addr = end; addr < real_end; addr += TARGET_PAGE_SIZE) {
601             prot |= page_get_flags(addr);
602         }
603         if (prot != 0)
604             real_end -= qemu_host_page_size;
605     }
606     if (real_start != real_end) {
607         mmap(g2h(real_start), real_end - real_start, PROT_NONE,
608                  MAP_FIXED | MAP_ANONYMOUS | MAP_PRIVATE | MAP_NORESERVE,
609                  -1, 0);
610     }
611 }
612 
613 int target_munmap(abi_ulong start, abi_ulong len)
614 {
615     abi_ulong end, real_start, real_end, addr;
616     int prot, ret;
617 
618 #ifdef DEBUG_MMAP
619     printf("munmap: start=0x" TARGET_ABI_FMT_lx " len=0x"
620            TARGET_ABI_FMT_lx "\n",
621            start, len);
622 #endif
623     if (start & ~TARGET_PAGE_MASK)
624         return -TARGET_EINVAL;
625     len = TARGET_PAGE_ALIGN(len);
626     if (len == 0 || !guest_range_valid(start, len)) {
627         return -TARGET_EINVAL;
628     }
629 
630     mmap_lock();
631     end = start + len;
632     real_start = start & qemu_host_page_mask;
633     real_end = HOST_PAGE_ALIGN(end);
634 
635     if (start > real_start) {
636         /* handle host page containing start */
637         prot = 0;
638         for(addr = real_start; addr < start; addr += TARGET_PAGE_SIZE) {
639             prot |= page_get_flags(addr);
640         }
641         if (real_end == real_start + qemu_host_page_size) {
642             for(addr = end; addr < real_end; addr += TARGET_PAGE_SIZE) {
643                 prot |= page_get_flags(addr);
644             }
645             end = real_end;
646         }
647         if (prot != 0)
648             real_start += qemu_host_page_size;
649     }
650     if (end < real_end) {
651         prot = 0;
652         for(addr = end; addr < real_end; addr += TARGET_PAGE_SIZE) {
653             prot |= page_get_flags(addr);
654         }
655         if (prot != 0)
656             real_end -= qemu_host_page_size;
657     }
658 
659     ret = 0;
660     /* unmap what we can */
661     if (real_start < real_end) {
662         if (reserved_va) {
663             mmap_reserve(real_start, real_end - real_start);
664         } else {
665             ret = munmap(g2h(real_start), real_end - real_start);
666         }
667     }
668 
669     if (ret == 0) {
670         page_set_flags(start, start + len, 0);
671         tb_invalidate_phys_range(start, start + len);
672     }
673     mmap_unlock();
674     return ret;
675 }
676 
677 abi_long target_mremap(abi_ulong old_addr, abi_ulong old_size,
678                        abi_ulong new_size, unsigned long flags,
679                        abi_ulong new_addr)
680 {
681     int prot;
682     void *host_addr;
683 
684     if (!guest_range_valid(old_addr, old_size) ||
685         ((flags & MREMAP_FIXED) &&
686          !guest_range_valid(new_addr, new_size))) {
687         errno = ENOMEM;
688         return -1;
689     }
690 
691     mmap_lock();
692 
693     if (flags & MREMAP_FIXED) {
694         host_addr = mremap(g2h(old_addr), old_size, new_size,
695                            flags, g2h(new_addr));
696 
697         if (reserved_va && host_addr != MAP_FAILED) {
698             /* If new and old addresses overlap then the above mremap will
699                already have failed with EINVAL.  */
700             mmap_reserve(old_addr, old_size);
701         }
702     } else if (flags & MREMAP_MAYMOVE) {
703         abi_ulong mmap_start;
704 
705         mmap_start = mmap_find_vma(0, new_size);
706 
707         if (mmap_start == -1) {
708             errno = ENOMEM;
709             host_addr = MAP_FAILED;
710         } else {
711             host_addr = mremap(g2h(old_addr), old_size, new_size,
712                                flags | MREMAP_FIXED, g2h(mmap_start));
713             if (reserved_va) {
714                 mmap_reserve(old_addr, old_size);
715             }
716         }
717     } else {
718         int prot = 0;
719         if (reserved_va && old_size < new_size) {
720             abi_ulong addr;
721             for (addr = old_addr + old_size;
722                  addr < old_addr + new_size;
723                  addr++) {
724                 prot |= page_get_flags(addr);
725             }
726         }
727         if (prot == 0) {
728             host_addr = mremap(g2h(old_addr), old_size, new_size, flags);
729             if (host_addr != MAP_FAILED && reserved_va && old_size > new_size) {
730                 mmap_reserve(old_addr + old_size, new_size - old_size);
731             }
732         } else {
733             errno = ENOMEM;
734             host_addr = MAP_FAILED;
735         }
736         /* Check if address fits target address space */
737         if ((unsigned long)host_addr + new_size > (abi_ulong)-1) {
738             /* Revert mremap() changes */
739             host_addr = mremap(g2h(old_addr), new_size, old_size, flags);
740             errno = ENOMEM;
741             host_addr = MAP_FAILED;
742         }
743     }
744 
745     if (host_addr == MAP_FAILED) {
746         new_addr = -1;
747     } else {
748         new_addr = h2g(host_addr);
749         prot = page_get_flags(old_addr);
750         page_set_flags(old_addr, old_addr + old_size, 0);
751         page_set_flags(new_addr, new_addr + new_size, prot | PAGE_VALID);
752     }
753     tb_invalidate_phys_range(new_addr, new_addr + new_size);
754     mmap_unlock();
755     return new_addr;
756 }
757