1 /* 2 * Linux syscalls 3 * 4 * Copyright (c) 2003 Fabrice Bellard 5 * 6 * This program is free software; you can redistribute it and/or modify 7 * it under the terms of the GNU General Public License as published by 8 * the Free Software Foundation; either version 2 of the License, or 9 * (at your option) any later version. 10 * 11 * This program is distributed in the hope that it will be useful, 12 * but WITHOUT ANY WARRANTY; without even the implied warranty of 13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 14 * GNU General Public License for more details. 15 * 16 * You should have received a copy of the GNU General Public License 17 * along with this program; if not, see <http://www.gnu.org/licenses/>. 18 */ 19 #define _ATFILE_SOURCE 20 #include "qemu/osdep.h" 21 #include "qemu/cutils.h" 22 #include "qemu/path.h" 23 #include <elf.h> 24 #include <endian.h> 25 #include <grp.h> 26 #include <sys/ipc.h> 27 #include <sys/msg.h> 28 #include <sys/wait.h> 29 #include <sys/mount.h> 30 #include <sys/file.h> 31 #include <sys/fsuid.h> 32 #include <sys/personality.h> 33 #include <sys/prctl.h> 34 #include <sys/resource.h> 35 #include <sys/swap.h> 36 #include <linux/capability.h> 37 #include <sched.h> 38 #include <sys/timex.h> 39 #include <sys/socket.h> 40 #include <sys/un.h> 41 #include <sys/uio.h> 42 #include <poll.h> 43 #include <sys/times.h> 44 #include <sys/shm.h> 45 #include <sys/sem.h> 46 #include <sys/statfs.h> 47 #include <utime.h> 48 #include <sys/sysinfo.h> 49 #include <sys/signalfd.h> 50 //#include <sys/user.h> 51 #include <netinet/ip.h> 52 #include <netinet/tcp.h> 53 #include <linux/wireless.h> 54 #include <linux/icmp.h> 55 #include <linux/icmpv6.h> 56 #include <linux/errqueue.h> 57 #include <linux/random.h> 58 #include "qemu-common.h" 59 #ifdef CONFIG_TIMERFD 60 #include <sys/timerfd.h> 61 #endif 62 #ifdef TARGET_GPROF 63 #include <sys/gmon.h> 64 #endif 65 #ifdef CONFIG_EVENTFD 66 #include <sys/eventfd.h> 67 #endif 68 #ifdef CONFIG_EPOLL 69 #include <sys/epoll.h> 70 #endif 71 #ifdef CONFIG_ATTR 72 #include "qemu/xattr.h" 73 #endif 74 #ifdef CONFIG_SENDFILE 75 #include <sys/sendfile.h> 76 #endif 77 78 #define termios host_termios 79 #define winsize host_winsize 80 #define termio host_termio 81 #define sgttyb host_sgttyb /* same as target */ 82 #define tchars host_tchars /* same as target */ 83 #define ltchars host_ltchars /* same as target */ 84 85 #include <linux/termios.h> 86 #include <linux/unistd.h> 87 #include <linux/cdrom.h> 88 #include <linux/hdreg.h> 89 #include <linux/soundcard.h> 90 #include <linux/kd.h> 91 #include <linux/mtio.h> 92 #include <linux/fs.h> 93 #if defined(CONFIG_FIEMAP) 94 #include <linux/fiemap.h> 95 #endif 96 #include <linux/fb.h> 97 #if defined(CONFIG_USBFS) 98 #include <linux/usbdevice_fs.h> 99 #include <linux/usb/ch9.h> 100 #endif 101 #include <linux/vt.h> 102 #include <linux/dm-ioctl.h> 103 #include <linux/reboot.h> 104 #include <linux/route.h> 105 #include <linux/filter.h> 106 #include <linux/blkpg.h> 107 #include <netpacket/packet.h> 108 #include <linux/netlink.h> 109 #include "linux_loop.h" 110 #include "uname.h" 111 112 #include "qemu.h" 113 #include "fd-trans.h" 114 115 #ifndef CLONE_IO 116 #define CLONE_IO 0x80000000 /* Clone io context */ 117 #endif 118 119 /* We can't directly call the host clone syscall, because this will 120 * badly confuse libc (breaking mutexes, for example). So we must 121 * divide clone flags into: 122 * * flag combinations that look like pthread_create() 123 * * flag combinations that look like fork() 124 * * flags we can implement within QEMU itself 125 * * flags we can't support and will return an error for 126 */ 127 /* For thread creation, all these flags must be present; for 128 * fork, none must be present. 129 */ 130 #define CLONE_THREAD_FLAGS \ 131 (CLONE_VM | CLONE_FS | CLONE_FILES | \ 132 CLONE_SIGHAND | CLONE_THREAD | CLONE_SYSVSEM) 133 134 /* These flags are ignored: 135 * CLONE_DETACHED is now ignored by the kernel; 136 * CLONE_IO is just an optimisation hint to the I/O scheduler 137 */ 138 #define CLONE_IGNORED_FLAGS \ 139 (CLONE_DETACHED | CLONE_IO) 140 141 /* Flags for fork which we can implement within QEMU itself */ 142 #define CLONE_OPTIONAL_FORK_FLAGS \ 143 (CLONE_SETTLS | CLONE_PARENT_SETTID | \ 144 CLONE_CHILD_CLEARTID | CLONE_CHILD_SETTID) 145 146 /* Flags for thread creation which we can implement within QEMU itself */ 147 #define CLONE_OPTIONAL_THREAD_FLAGS \ 148 (CLONE_SETTLS | CLONE_PARENT_SETTID | \ 149 CLONE_CHILD_CLEARTID | CLONE_CHILD_SETTID | CLONE_PARENT) 150 151 #define CLONE_INVALID_FORK_FLAGS \ 152 (~(CSIGNAL | CLONE_OPTIONAL_FORK_FLAGS | CLONE_IGNORED_FLAGS)) 153 154 #define CLONE_INVALID_THREAD_FLAGS \ 155 (~(CSIGNAL | CLONE_THREAD_FLAGS | CLONE_OPTIONAL_THREAD_FLAGS | \ 156 CLONE_IGNORED_FLAGS)) 157 158 /* CLONE_VFORK is special cased early in do_fork(). The other flag bits 159 * have almost all been allocated. We cannot support any of 160 * CLONE_NEWNS, CLONE_NEWCGROUP, CLONE_NEWUTS, CLONE_NEWIPC, 161 * CLONE_NEWUSER, CLONE_NEWPID, CLONE_NEWNET, CLONE_PTRACE, CLONE_UNTRACED. 162 * The checks against the invalid thread masks above will catch these. 163 * (The one remaining unallocated bit is 0x1000 which used to be CLONE_PID.) 164 */ 165 166 /* Define DEBUG_ERESTARTSYS to force every syscall to be restarted 167 * once. This exercises the codepaths for restart. 168 */ 169 //#define DEBUG_ERESTARTSYS 170 171 //#include <linux/msdos_fs.h> 172 #define VFAT_IOCTL_READDIR_BOTH _IOR('r', 1, struct linux_dirent [2]) 173 #define VFAT_IOCTL_READDIR_SHORT _IOR('r', 2, struct linux_dirent [2]) 174 175 #undef _syscall0 176 #undef _syscall1 177 #undef _syscall2 178 #undef _syscall3 179 #undef _syscall4 180 #undef _syscall5 181 #undef _syscall6 182 183 #define _syscall0(type,name) \ 184 static type name (void) \ 185 { \ 186 return syscall(__NR_##name); \ 187 } 188 189 #define _syscall1(type,name,type1,arg1) \ 190 static type name (type1 arg1) \ 191 { \ 192 return syscall(__NR_##name, arg1); \ 193 } 194 195 #define _syscall2(type,name,type1,arg1,type2,arg2) \ 196 static type name (type1 arg1,type2 arg2) \ 197 { \ 198 return syscall(__NR_##name, arg1, arg2); \ 199 } 200 201 #define _syscall3(type,name,type1,arg1,type2,arg2,type3,arg3) \ 202 static type name (type1 arg1,type2 arg2,type3 arg3) \ 203 { \ 204 return syscall(__NR_##name, arg1, arg2, arg3); \ 205 } 206 207 #define _syscall4(type,name,type1,arg1,type2,arg2,type3,arg3,type4,arg4) \ 208 static type name (type1 arg1,type2 arg2,type3 arg3,type4 arg4) \ 209 { \ 210 return syscall(__NR_##name, arg1, arg2, arg3, arg4); \ 211 } 212 213 #define _syscall5(type,name,type1,arg1,type2,arg2,type3,arg3,type4,arg4, \ 214 type5,arg5) \ 215 static type name (type1 arg1,type2 arg2,type3 arg3,type4 arg4,type5 arg5) \ 216 { \ 217 return syscall(__NR_##name, arg1, arg2, arg3, arg4, arg5); \ 218 } 219 220 221 #define _syscall6(type,name,type1,arg1,type2,arg2,type3,arg3,type4,arg4, \ 222 type5,arg5,type6,arg6) \ 223 static type name (type1 arg1,type2 arg2,type3 arg3,type4 arg4,type5 arg5, \ 224 type6 arg6) \ 225 { \ 226 return syscall(__NR_##name, arg1, arg2, arg3, arg4, arg5, arg6); \ 227 } 228 229 230 #define __NR_sys_uname __NR_uname 231 #define __NR_sys_getcwd1 __NR_getcwd 232 #define __NR_sys_getdents __NR_getdents 233 #define __NR_sys_getdents64 __NR_getdents64 234 #define __NR_sys_getpriority __NR_getpriority 235 #define __NR_sys_rt_sigqueueinfo __NR_rt_sigqueueinfo 236 #define __NR_sys_rt_tgsigqueueinfo __NR_rt_tgsigqueueinfo 237 #define __NR_sys_syslog __NR_syslog 238 #define __NR_sys_futex __NR_futex 239 #define __NR_sys_inotify_init __NR_inotify_init 240 #define __NR_sys_inotify_add_watch __NR_inotify_add_watch 241 #define __NR_sys_inotify_rm_watch __NR_inotify_rm_watch 242 243 #if defined(__alpha__) || defined(__x86_64__) || defined(__s390x__) 244 #define __NR__llseek __NR_lseek 245 #endif 246 247 /* Newer kernel ports have llseek() instead of _llseek() */ 248 #if defined(TARGET_NR_llseek) && !defined(TARGET_NR__llseek) 249 #define TARGET_NR__llseek TARGET_NR_llseek 250 #endif 251 252 #ifdef __NR_gettid 253 _syscall0(int, gettid) 254 #else 255 /* This is a replacement for the host gettid() and must return a host 256 errno. */ 257 static int gettid(void) { 258 return -ENOSYS; 259 } 260 #endif 261 262 /* For the 64-bit guest on 32-bit host case we must emulate 263 * getdents using getdents64, because otherwise the host 264 * might hand us back more dirent records than we can fit 265 * into the guest buffer after structure format conversion. 266 * Otherwise we emulate getdents with getdents if the host has it. 267 */ 268 #if defined(__NR_getdents) && HOST_LONG_BITS >= TARGET_ABI_BITS 269 #define EMULATE_GETDENTS_WITH_GETDENTS 270 #endif 271 272 #if defined(TARGET_NR_getdents) && defined(EMULATE_GETDENTS_WITH_GETDENTS) 273 _syscall3(int, sys_getdents, uint, fd, struct linux_dirent *, dirp, uint, count); 274 #endif 275 #if (defined(TARGET_NR_getdents) && \ 276 !defined(EMULATE_GETDENTS_WITH_GETDENTS)) || \ 277 (defined(TARGET_NR_getdents64) && defined(__NR_getdents64)) 278 _syscall3(int, sys_getdents64, uint, fd, struct linux_dirent64 *, dirp, uint, count); 279 #endif 280 #if defined(TARGET_NR__llseek) && defined(__NR_llseek) 281 _syscall5(int, _llseek, uint, fd, ulong, hi, ulong, lo, 282 loff_t *, res, uint, wh); 283 #endif 284 _syscall3(int, sys_rt_sigqueueinfo, pid_t, pid, int, sig, siginfo_t *, uinfo) 285 _syscall4(int, sys_rt_tgsigqueueinfo, pid_t, pid, pid_t, tid, int, sig, 286 siginfo_t *, uinfo) 287 _syscall3(int,sys_syslog,int,type,char*,bufp,int,len) 288 #ifdef __NR_exit_group 289 _syscall1(int,exit_group,int,error_code) 290 #endif 291 #if defined(TARGET_NR_set_tid_address) && defined(__NR_set_tid_address) 292 _syscall1(int,set_tid_address,int *,tidptr) 293 #endif 294 #if defined(TARGET_NR_futex) && defined(__NR_futex) 295 _syscall6(int,sys_futex,int *,uaddr,int,op,int,val, 296 const struct timespec *,timeout,int *,uaddr2,int,val3) 297 #endif 298 #define __NR_sys_sched_getaffinity __NR_sched_getaffinity 299 _syscall3(int, sys_sched_getaffinity, pid_t, pid, unsigned int, len, 300 unsigned long *, user_mask_ptr); 301 #define __NR_sys_sched_setaffinity __NR_sched_setaffinity 302 _syscall3(int, sys_sched_setaffinity, pid_t, pid, unsigned int, len, 303 unsigned long *, user_mask_ptr); 304 #define __NR_sys_getcpu __NR_getcpu 305 _syscall3(int, sys_getcpu, unsigned *, cpu, unsigned *, node, void *, tcache); 306 _syscall4(int, reboot, int, magic1, int, magic2, unsigned int, cmd, 307 void *, arg); 308 _syscall2(int, capget, struct __user_cap_header_struct *, header, 309 struct __user_cap_data_struct *, data); 310 _syscall2(int, capset, struct __user_cap_header_struct *, header, 311 struct __user_cap_data_struct *, data); 312 #if defined(TARGET_NR_ioprio_get) && defined(__NR_ioprio_get) 313 _syscall2(int, ioprio_get, int, which, int, who) 314 #endif 315 #if defined(TARGET_NR_ioprio_set) && defined(__NR_ioprio_set) 316 _syscall3(int, ioprio_set, int, which, int, who, int, ioprio) 317 #endif 318 #if defined(TARGET_NR_getrandom) && defined(__NR_getrandom) 319 _syscall3(int, getrandom, void *, buf, size_t, buflen, unsigned int, flags) 320 #endif 321 322 #if defined(TARGET_NR_kcmp) && defined(__NR_kcmp) 323 _syscall5(int, kcmp, pid_t, pid1, pid_t, pid2, int, type, 324 unsigned long, idx1, unsigned long, idx2) 325 #endif 326 327 static bitmask_transtbl fcntl_flags_tbl[] = { 328 { TARGET_O_ACCMODE, TARGET_O_WRONLY, O_ACCMODE, O_WRONLY, }, 329 { TARGET_O_ACCMODE, TARGET_O_RDWR, O_ACCMODE, O_RDWR, }, 330 { TARGET_O_CREAT, TARGET_O_CREAT, O_CREAT, O_CREAT, }, 331 { TARGET_O_EXCL, TARGET_O_EXCL, O_EXCL, O_EXCL, }, 332 { TARGET_O_NOCTTY, TARGET_O_NOCTTY, O_NOCTTY, O_NOCTTY, }, 333 { TARGET_O_TRUNC, TARGET_O_TRUNC, O_TRUNC, O_TRUNC, }, 334 { TARGET_O_APPEND, TARGET_O_APPEND, O_APPEND, O_APPEND, }, 335 { TARGET_O_NONBLOCK, TARGET_O_NONBLOCK, O_NONBLOCK, O_NONBLOCK, }, 336 { TARGET_O_SYNC, TARGET_O_DSYNC, O_SYNC, O_DSYNC, }, 337 { TARGET_O_SYNC, TARGET_O_SYNC, O_SYNC, O_SYNC, }, 338 { TARGET_FASYNC, TARGET_FASYNC, FASYNC, FASYNC, }, 339 { TARGET_O_DIRECTORY, TARGET_O_DIRECTORY, O_DIRECTORY, O_DIRECTORY, }, 340 { TARGET_O_NOFOLLOW, TARGET_O_NOFOLLOW, O_NOFOLLOW, O_NOFOLLOW, }, 341 #if defined(O_DIRECT) 342 { TARGET_O_DIRECT, TARGET_O_DIRECT, O_DIRECT, O_DIRECT, }, 343 #endif 344 #if defined(O_NOATIME) 345 { TARGET_O_NOATIME, TARGET_O_NOATIME, O_NOATIME, O_NOATIME }, 346 #endif 347 #if defined(O_CLOEXEC) 348 { TARGET_O_CLOEXEC, TARGET_O_CLOEXEC, O_CLOEXEC, O_CLOEXEC }, 349 #endif 350 #if defined(O_PATH) 351 { TARGET_O_PATH, TARGET_O_PATH, O_PATH, O_PATH }, 352 #endif 353 #if defined(O_TMPFILE) 354 { TARGET_O_TMPFILE, TARGET_O_TMPFILE, O_TMPFILE, O_TMPFILE }, 355 #endif 356 /* Don't terminate the list prematurely on 64-bit host+guest. */ 357 #if TARGET_O_LARGEFILE != 0 || O_LARGEFILE != 0 358 { TARGET_O_LARGEFILE, TARGET_O_LARGEFILE, O_LARGEFILE, O_LARGEFILE, }, 359 #endif 360 { 0, 0, 0, 0 } 361 }; 362 363 static int sys_getcwd1(char *buf, size_t size) 364 { 365 if (getcwd(buf, size) == NULL) { 366 /* getcwd() sets errno */ 367 return (-1); 368 } 369 return strlen(buf)+1; 370 } 371 372 #ifdef TARGET_NR_utimensat 373 #if defined(__NR_utimensat) 374 #define __NR_sys_utimensat __NR_utimensat 375 _syscall4(int,sys_utimensat,int,dirfd,const char *,pathname, 376 const struct timespec *,tsp,int,flags) 377 #else 378 static int sys_utimensat(int dirfd, const char *pathname, 379 const struct timespec times[2], int flags) 380 { 381 errno = ENOSYS; 382 return -1; 383 } 384 #endif 385 #endif /* TARGET_NR_utimensat */ 386 387 #ifdef TARGET_NR_renameat2 388 #if defined(__NR_renameat2) 389 #define __NR_sys_renameat2 __NR_renameat2 390 _syscall5(int, sys_renameat2, int, oldfd, const char *, old, int, newfd, 391 const char *, new, unsigned int, flags) 392 #else 393 static int sys_renameat2(int oldfd, const char *old, 394 int newfd, const char *new, int flags) 395 { 396 if (flags == 0) { 397 return renameat(oldfd, old, newfd, new); 398 } 399 errno = ENOSYS; 400 return -1; 401 } 402 #endif 403 #endif /* TARGET_NR_renameat2 */ 404 405 #ifdef CONFIG_INOTIFY 406 #include <sys/inotify.h> 407 408 #if defined(TARGET_NR_inotify_init) && defined(__NR_inotify_init) 409 static int sys_inotify_init(void) 410 { 411 return (inotify_init()); 412 } 413 #endif 414 #if defined(TARGET_NR_inotify_add_watch) && defined(__NR_inotify_add_watch) 415 static int sys_inotify_add_watch(int fd,const char *pathname, int32_t mask) 416 { 417 return (inotify_add_watch(fd, pathname, mask)); 418 } 419 #endif 420 #if defined(TARGET_NR_inotify_rm_watch) && defined(__NR_inotify_rm_watch) 421 static int sys_inotify_rm_watch(int fd, int32_t wd) 422 { 423 return (inotify_rm_watch(fd, wd)); 424 } 425 #endif 426 #ifdef CONFIG_INOTIFY1 427 #if defined(TARGET_NR_inotify_init1) && defined(__NR_inotify_init1) 428 static int sys_inotify_init1(int flags) 429 { 430 return (inotify_init1(flags)); 431 } 432 #endif 433 #endif 434 #else 435 /* Userspace can usually survive runtime without inotify */ 436 #undef TARGET_NR_inotify_init 437 #undef TARGET_NR_inotify_init1 438 #undef TARGET_NR_inotify_add_watch 439 #undef TARGET_NR_inotify_rm_watch 440 #endif /* CONFIG_INOTIFY */ 441 442 #if defined(TARGET_NR_prlimit64) 443 #ifndef __NR_prlimit64 444 # define __NR_prlimit64 -1 445 #endif 446 #define __NR_sys_prlimit64 __NR_prlimit64 447 /* The glibc rlimit structure may not be that used by the underlying syscall */ 448 struct host_rlimit64 { 449 uint64_t rlim_cur; 450 uint64_t rlim_max; 451 }; 452 _syscall4(int, sys_prlimit64, pid_t, pid, int, resource, 453 const struct host_rlimit64 *, new_limit, 454 struct host_rlimit64 *, old_limit) 455 #endif 456 457 458 #if defined(TARGET_NR_timer_create) 459 /* Maxiumum of 32 active POSIX timers allowed at any one time. */ 460 static timer_t g_posix_timers[32] = { 0, } ; 461 462 static inline int next_free_host_timer(void) 463 { 464 int k ; 465 /* FIXME: Does finding the next free slot require a lock? */ 466 for (k = 0; k < ARRAY_SIZE(g_posix_timers); k++) { 467 if (g_posix_timers[k] == 0) { 468 g_posix_timers[k] = (timer_t) 1; 469 return k; 470 } 471 } 472 return -1; 473 } 474 #endif 475 476 /* ARM EABI and MIPS expect 64bit types aligned even on pairs or registers */ 477 #ifdef TARGET_ARM 478 static inline int regpairs_aligned(void *cpu_env, int num) 479 { 480 return ((((CPUARMState *)cpu_env)->eabi) == 1) ; 481 } 482 #elif defined(TARGET_MIPS) && (TARGET_ABI_BITS == 32) 483 static inline int regpairs_aligned(void *cpu_env, int num) { return 1; } 484 #elif defined(TARGET_PPC) && !defined(TARGET_PPC64) 485 /* SysV AVI for PPC32 expects 64bit parameters to be passed on odd/even pairs 486 * of registers which translates to the same as ARM/MIPS, because we start with 487 * r3 as arg1 */ 488 static inline int regpairs_aligned(void *cpu_env, int num) { return 1; } 489 #elif defined(TARGET_SH4) 490 /* SH4 doesn't align register pairs, except for p{read,write}64 */ 491 static inline int regpairs_aligned(void *cpu_env, int num) 492 { 493 switch (num) { 494 case TARGET_NR_pread64: 495 case TARGET_NR_pwrite64: 496 return 1; 497 498 default: 499 return 0; 500 } 501 } 502 #elif defined(TARGET_XTENSA) 503 static inline int regpairs_aligned(void *cpu_env, int num) { return 1; } 504 #else 505 static inline int regpairs_aligned(void *cpu_env, int num) { return 0; } 506 #endif 507 508 #define ERRNO_TABLE_SIZE 1200 509 510 /* target_to_host_errno_table[] is initialized from 511 * host_to_target_errno_table[] in syscall_init(). */ 512 static uint16_t target_to_host_errno_table[ERRNO_TABLE_SIZE] = { 513 }; 514 515 /* 516 * This list is the union of errno values overridden in asm-<arch>/errno.h 517 * minus the errnos that are not actually generic to all archs. 518 */ 519 static uint16_t host_to_target_errno_table[ERRNO_TABLE_SIZE] = { 520 [EAGAIN] = TARGET_EAGAIN, 521 [EIDRM] = TARGET_EIDRM, 522 [ECHRNG] = TARGET_ECHRNG, 523 [EL2NSYNC] = TARGET_EL2NSYNC, 524 [EL3HLT] = TARGET_EL3HLT, 525 [EL3RST] = TARGET_EL3RST, 526 [ELNRNG] = TARGET_ELNRNG, 527 [EUNATCH] = TARGET_EUNATCH, 528 [ENOCSI] = TARGET_ENOCSI, 529 [EL2HLT] = TARGET_EL2HLT, 530 [EDEADLK] = TARGET_EDEADLK, 531 [ENOLCK] = TARGET_ENOLCK, 532 [EBADE] = TARGET_EBADE, 533 [EBADR] = TARGET_EBADR, 534 [EXFULL] = TARGET_EXFULL, 535 [ENOANO] = TARGET_ENOANO, 536 [EBADRQC] = TARGET_EBADRQC, 537 [EBADSLT] = TARGET_EBADSLT, 538 [EBFONT] = TARGET_EBFONT, 539 [ENOSTR] = TARGET_ENOSTR, 540 [ENODATA] = TARGET_ENODATA, 541 [ETIME] = TARGET_ETIME, 542 [ENOSR] = TARGET_ENOSR, 543 [ENONET] = TARGET_ENONET, 544 [ENOPKG] = TARGET_ENOPKG, 545 [EREMOTE] = TARGET_EREMOTE, 546 [ENOLINK] = TARGET_ENOLINK, 547 [EADV] = TARGET_EADV, 548 [ESRMNT] = TARGET_ESRMNT, 549 [ECOMM] = TARGET_ECOMM, 550 [EPROTO] = TARGET_EPROTO, 551 [EDOTDOT] = TARGET_EDOTDOT, 552 [EMULTIHOP] = TARGET_EMULTIHOP, 553 [EBADMSG] = TARGET_EBADMSG, 554 [ENAMETOOLONG] = TARGET_ENAMETOOLONG, 555 [EOVERFLOW] = TARGET_EOVERFLOW, 556 [ENOTUNIQ] = TARGET_ENOTUNIQ, 557 [EBADFD] = TARGET_EBADFD, 558 [EREMCHG] = TARGET_EREMCHG, 559 [ELIBACC] = TARGET_ELIBACC, 560 [ELIBBAD] = TARGET_ELIBBAD, 561 [ELIBSCN] = TARGET_ELIBSCN, 562 [ELIBMAX] = TARGET_ELIBMAX, 563 [ELIBEXEC] = TARGET_ELIBEXEC, 564 [EILSEQ] = TARGET_EILSEQ, 565 [ENOSYS] = TARGET_ENOSYS, 566 [ELOOP] = TARGET_ELOOP, 567 [ERESTART] = TARGET_ERESTART, 568 [ESTRPIPE] = TARGET_ESTRPIPE, 569 [ENOTEMPTY] = TARGET_ENOTEMPTY, 570 [EUSERS] = TARGET_EUSERS, 571 [ENOTSOCK] = TARGET_ENOTSOCK, 572 [EDESTADDRREQ] = TARGET_EDESTADDRREQ, 573 [EMSGSIZE] = TARGET_EMSGSIZE, 574 [EPROTOTYPE] = TARGET_EPROTOTYPE, 575 [ENOPROTOOPT] = TARGET_ENOPROTOOPT, 576 [EPROTONOSUPPORT] = TARGET_EPROTONOSUPPORT, 577 [ESOCKTNOSUPPORT] = TARGET_ESOCKTNOSUPPORT, 578 [EOPNOTSUPP] = TARGET_EOPNOTSUPP, 579 [EPFNOSUPPORT] = TARGET_EPFNOSUPPORT, 580 [EAFNOSUPPORT] = TARGET_EAFNOSUPPORT, 581 [EADDRINUSE] = TARGET_EADDRINUSE, 582 [EADDRNOTAVAIL] = TARGET_EADDRNOTAVAIL, 583 [ENETDOWN] = TARGET_ENETDOWN, 584 [ENETUNREACH] = TARGET_ENETUNREACH, 585 [ENETRESET] = TARGET_ENETRESET, 586 [ECONNABORTED] = TARGET_ECONNABORTED, 587 [ECONNRESET] = TARGET_ECONNRESET, 588 [ENOBUFS] = TARGET_ENOBUFS, 589 [EISCONN] = TARGET_EISCONN, 590 [ENOTCONN] = TARGET_ENOTCONN, 591 [EUCLEAN] = TARGET_EUCLEAN, 592 [ENOTNAM] = TARGET_ENOTNAM, 593 [ENAVAIL] = TARGET_ENAVAIL, 594 [EISNAM] = TARGET_EISNAM, 595 [EREMOTEIO] = TARGET_EREMOTEIO, 596 [EDQUOT] = TARGET_EDQUOT, 597 [ESHUTDOWN] = TARGET_ESHUTDOWN, 598 [ETOOMANYREFS] = TARGET_ETOOMANYREFS, 599 [ETIMEDOUT] = TARGET_ETIMEDOUT, 600 [ECONNREFUSED] = TARGET_ECONNREFUSED, 601 [EHOSTDOWN] = TARGET_EHOSTDOWN, 602 [EHOSTUNREACH] = TARGET_EHOSTUNREACH, 603 [EALREADY] = TARGET_EALREADY, 604 [EINPROGRESS] = TARGET_EINPROGRESS, 605 [ESTALE] = TARGET_ESTALE, 606 [ECANCELED] = TARGET_ECANCELED, 607 [ENOMEDIUM] = TARGET_ENOMEDIUM, 608 [EMEDIUMTYPE] = TARGET_EMEDIUMTYPE, 609 #ifdef ENOKEY 610 [ENOKEY] = TARGET_ENOKEY, 611 #endif 612 #ifdef EKEYEXPIRED 613 [EKEYEXPIRED] = TARGET_EKEYEXPIRED, 614 #endif 615 #ifdef EKEYREVOKED 616 [EKEYREVOKED] = TARGET_EKEYREVOKED, 617 #endif 618 #ifdef EKEYREJECTED 619 [EKEYREJECTED] = TARGET_EKEYREJECTED, 620 #endif 621 #ifdef EOWNERDEAD 622 [EOWNERDEAD] = TARGET_EOWNERDEAD, 623 #endif 624 #ifdef ENOTRECOVERABLE 625 [ENOTRECOVERABLE] = TARGET_ENOTRECOVERABLE, 626 #endif 627 #ifdef ENOMSG 628 [ENOMSG] = TARGET_ENOMSG, 629 #endif 630 #ifdef ERKFILL 631 [ERFKILL] = TARGET_ERFKILL, 632 #endif 633 #ifdef EHWPOISON 634 [EHWPOISON] = TARGET_EHWPOISON, 635 #endif 636 }; 637 638 static inline int host_to_target_errno(int err) 639 { 640 if (err >= 0 && err < ERRNO_TABLE_SIZE && 641 host_to_target_errno_table[err]) { 642 return host_to_target_errno_table[err]; 643 } 644 return err; 645 } 646 647 static inline int target_to_host_errno(int err) 648 { 649 if (err >= 0 && err < ERRNO_TABLE_SIZE && 650 target_to_host_errno_table[err]) { 651 return target_to_host_errno_table[err]; 652 } 653 return err; 654 } 655 656 static inline abi_long get_errno(abi_long ret) 657 { 658 if (ret == -1) 659 return -host_to_target_errno(errno); 660 else 661 return ret; 662 } 663 664 const char *target_strerror(int err) 665 { 666 if (err == TARGET_ERESTARTSYS) { 667 return "To be restarted"; 668 } 669 if (err == TARGET_QEMU_ESIGRETURN) { 670 return "Successful exit from sigreturn"; 671 } 672 673 if ((err >= ERRNO_TABLE_SIZE) || (err < 0)) { 674 return NULL; 675 } 676 return strerror(target_to_host_errno(err)); 677 } 678 679 #define safe_syscall0(type, name) \ 680 static type safe_##name(void) \ 681 { \ 682 return safe_syscall(__NR_##name); \ 683 } 684 685 #define safe_syscall1(type, name, type1, arg1) \ 686 static type safe_##name(type1 arg1) \ 687 { \ 688 return safe_syscall(__NR_##name, arg1); \ 689 } 690 691 #define safe_syscall2(type, name, type1, arg1, type2, arg2) \ 692 static type safe_##name(type1 arg1, type2 arg2) \ 693 { \ 694 return safe_syscall(__NR_##name, arg1, arg2); \ 695 } 696 697 #define safe_syscall3(type, name, type1, arg1, type2, arg2, type3, arg3) \ 698 static type safe_##name(type1 arg1, type2 arg2, type3 arg3) \ 699 { \ 700 return safe_syscall(__NR_##name, arg1, arg2, arg3); \ 701 } 702 703 #define safe_syscall4(type, name, type1, arg1, type2, arg2, type3, arg3, \ 704 type4, arg4) \ 705 static type safe_##name(type1 arg1, type2 arg2, type3 arg3, type4 arg4) \ 706 { \ 707 return safe_syscall(__NR_##name, arg1, arg2, arg3, arg4); \ 708 } 709 710 #define safe_syscall5(type, name, type1, arg1, type2, arg2, type3, arg3, \ 711 type4, arg4, type5, arg5) \ 712 static type safe_##name(type1 arg1, type2 arg2, type3 arg3, type4 arg4, \ 713 type5 arg5) \ 714 { \ 715 return safe_syscall(__NR_##name, arg1, arg2, arg3, arg4, arg5); \ 716 } 717 718 #define safe_syscall6(type, name, type1, arg1, type2, arg2, type3, arg3, \ 719 type4, arg4, type5, arg5, type6, arg6) \ 720 static type safe_##name(type1 arg1, type2 arg2, type3 arg3, type4 arg4, \ 721 type5 arg5, type6 arg6) \ 722 { \ 723 return safe_syscall(__NR_##name, arg1, arg2, arg3, arg4, arg5, arg6); \ 724 } 725 726 safe_syscall3(ssize_t, read, int, fd, void *, buff, size_t, count) 727 safe_syscall3(ssize_t, write, int, fd, const void *, buff, size_t, count) 728 safe_syscall4(int, openat, int, dirfd, const char *, pathname, \ 729 int, flags, mode_t, mode) 730 safe_syscall4(pid_t, wait4, pid_t, pid, int *, status, int, options, \ 731 struct rusage *, rusage) 732 safe_syscall5(int, waitid, idtype_t, idtype, id_t, id, siginfo_t *, infop, \ 733 int, options, struct rusage *, rusage) 734 safe_syscall3(int, execve, const char *, filename, char **, argv, char **, envp) 735 safe_syscall6(int, pselect6, int, nfds, fd_set *, readfds, fd_set *, writefds, \ 736 fd_set *, exceptfds, struct timespec *, timeout, void *, sig) 737 safe_syscall5(int, ppoll, struct pollfd *, ufds, unsigned int, nfds, 738 struct timespec *, tsp, const sigset_t *, sigmask, 739 size_t, sigsetsize) 740 safe_syscall6(int, epoll_pwait, int, epfd, struct epoll_event *, events, 741 int, maxevents, int, timeout, const sigset_t *, sigmask, 742 size_t, sigsetsize) 743 safe_syscall6(int,futex,int *,uaddr,int,op,int,val, \ 744 const struct timespec *,timeout,int *,uaddr2,int,val3) 745 safe_syscall2(int, rt_sigsuspend, sigset_t *, newset, size_t, sigsetsize) 746 safe_syscall2(int, kill, pid_t, pid, int, sig) 747 safe_syscall2(int, tkill, int, tid, int, sig) 748 safe_syscall3(int, tgkill, int, tgid, int, pid, int, sig) 749 safe_syscall3(ssize_t, readv, int, fd, const struct iovec *, iov, int, iovcnt) 750 safe_syscall3(ssize_t, writev, int, fd, const struct iovec *, iov, int, iovcnt) 751 safe_syscall5(ssize_t, preadv, int, fd, const struct iovec *, iov, int, iovcnt, 752 unsigned long, pos_l, unsigned long, pos_h) 753 safe_syscall5(ssize_t, pwritev, int, fd, const struct iovec *, iov, int, iovcnt, 754 unsigned long, pos_l, unsigned long, pos_h) 755 safe_syscall3(int, connect, int, fd, const struct sockaddr *, addr, 756 socklen_t, addrlen) 757 safe_syscall6(ssize_t, sendto, int, fd, const void *, buf, size_t, len, 758 int, flags, const struct sockaddr *, addr, socklen_t, addrlen) 759 safe_syscall6(ssize_t, recvfrom, int, fd, void *, buf, size_t, len, 760 int, flags, struct sockaddr *, addr, socklen_t *, addrlen) 761 safe_syscall3(ssize_t, sendmsg, int, fd, const struct msghdr *, msg, int, flags) 762 safe_syscall3(ssize_t, recvmsg, int, fd, struct msghdr *, msg, int, flags) 763 safe_syscall2(int, flock, int, fd, int, operation) 764 safe_syscall4(int, rt_sigtimedwait, const sigset_t *, these, siginfo_t *, uinfo, 765 const struct timespec *, uts, size_t, sigsetsize) 766 safe_syscall4(int, accept4, int, fd, struct sockaddr *, addr, socklen_t *, len, 767 int, flags) 768 safe_syscall2(int, nanosleep, const struct timespec *, req, 769 struct timespec *, rem) 770 #ifdef TARGET_NR_clock_nanosleep 771 safe_syscall4(int, clock_nanosleep, const clockid_t, clock, int, flags, 772 const struct timespec *, req, struct timespec *, rem) 773 #endif 774 #ifdef __NR_msgsnd 775 safe_syscall4(int, msgsnd, int, msgid, const void *, msgp, size_t, sz, 776 int, flags) 777 safe_syscall5(int, msgrcv, int, msgid, void *, msgp, size_t, sz, 778 long, msgtype, int, flags) 779 safe_syscall4(int, semtimedop, int, semid, struct sembuf *, tsops, 780 unsigned, nsops, const struct timespec *, timeout) 781 #else 782 /* This host kernel architecture uses a single ipc syscall; fake up 783 * wrappers for the sub-operations to hide this implementation detail. 784 * Annoyingly we can't include linux/ipc.h to get the constant definitions 785 * for the call parameter because some structs in there conflict with the 786 * sys/ipc.h ones. So we just define them here, and rely on them being 787 * the same for all host architectures. 788 */ 789 #define Q_SEMTIMEDOP 4 790 #define Q_MSGSND 11 791 #define Q_MSGRCV 12 792 #define Q_IPCCALL(VERSION, OP) ((VERSION) << 16 | (OP)) 793 794 safe_syscall6(int, ipc, int, call, long, first, long, second, long, third, 795 void *, ptr, long, fifth) 796 static int safe_msgsnd(int msgid, const void *msgp, size_t sz, int flags) 797 { 798 return safe_ipc(Q_IPCCALL(0, Q_MSGSND), msgid, sz, flags, (void *)msgp, 0); 799 } 800 static int safe_msgrcv(int msgid, void *msgp, size_t sz, long type, int flags) 801 { 802 return safe_ipc(Q_IPCCALL(1, Q_MSGRCV), msgid, sz, flags, msgp, type); 803 } 804 static int safe_semtimedop(int semid, struct sembuf *tsops, unsigned nsops, 805 const struct timespec *timeout) 806 { 807 return safe_ipc(Q_IPCCALL(0, Q_SEMTIMEDOP), semid, nsops, 0, tsops, 808 (long)timeout); 809 } 810 #endif 811 #if defined(TARGET_NR_mq_open) && defined(__NR_mq_open) 812 safe_syscall5(int, mq_timedsend, int, mqdes, const char *, msg_ptr, 813 size_t, len, unsigned, prio, const struct timespec *, timeout) 814 safe_syscall5(int, mq_timedreceive, int, mqdes, char *, msg_ptr, 815 size_t, len, unsigned *, prio, const struct timespec *, timeout) 816 #endif 817 /* We do ioctl like this rather than via safe_syscall3 to preserve the 818 * "third argument might be integer or pointer or not present" behaviour of 819 * the libc function. 820 */ 821 #define safe_ioctl(...) safe_syscall(__NR_ioctl, __VA_ARGS__) 822 /* Similarly for fcntl. Note that callers must always: 823 * pass the F_GETLK64 etc constants rather than the unsuffixed F_GETLK 824 * use the flock64 struct rather than unsuffixed flock 825 * This will then work and use a 64-bit offset for both 32-bit and 64-bit hosts. 826 */ 827 #ifdef __NR_fcntl64 828 #define safe_fcntl(...) safe_syscall(__NR_fcntl64, __VA_ARGS__) 829 #else 830 #define safe_fcntl(...) safe_syscall(__NR_fcntl, __VA_ARGS__) 831 #endif 832 833 static inline int host_to_target_sock_type(int host_type) 834 { 835 int target_type; 836 837 switch (host_type & 0xf /* SOCK_TYPE_MASK */) { 838 case SOCK_DGRAM: 839 target_type = TARGET_SOCK_DGRAM; 840 break; 841 case SOCK_STREAM: 842 target_type = TARGET_SOCK_STREAM; 843 break; 844 default: 845 target_type = host_type & 0xf /* SOCK_TYPE_MASK */; 846 break; 847 } 848 849 #if defined(SOCK_CLOEXEC) 850 if (host_type & SOCK_CLOEXEC) { 851 target_type |= TARGET_SOCK_CLOEXEC; 852 } 853 #endif 854 855 #if defined(SOCK_NONBLOCK) 856 if (host_type & SOCK_NONBLOCK) { 857 target_type |= TARGET_SOCK_NONBLOCK; 858 } 859 #endif 860 861 return target_type; 862 } 863 864 static abi_ulong target_brk; 865 static abi_ulong target_original_brk; 866 static abi_ulong brk_page; 867 868 void target_set_brk(abi_ulong new_brk) 869 { 870 target_original_brk = target_brk = HOST_PAGE_ALIGN(new_brk); 871 brk_page = HOST_PAGE_ALIGN(target_brk); 872 } 873 874 //#define DEBUGF_BRK(message, args...) do { fprintf(stderr, (message), ## args); } while (0) 875 #define DEBUGF_BRK(message, args...) 876 877 /* do_brk() must return target values and target errnos. */ 878 abi_long do_brk(abi_ulong new_brk) 879 { 880 abi_long mapped_addr; 881 abi_ulong new_alloc_size; 882 883 DEBUGF_BRK("do_brk(" TARGET_ABI_FMT_lx ") -> ", new_brk); 884 885 if (!new_brk) { 886 DEBUGF_BRK(TARGET_ABI_FMT_lx " (!new_brk)\n", target_brk); 887 return target_brk; 888 } 889 if (new_brk < target_original_brk) { 890 DEBUGF_BRK(TARGET_ABI_FMT_lx " (new_brk < target_original_brk)\n", 891 target_brk); 892 return target_brk; 893 } 894 895 /* If the new brk is less than the highest page reserved to the 896 * target heap allocation, set it and we're almost done... */ 897 if (new_brk <= brk_page) { 898 /* Heap contents are initialized to zero, as for anonymous 899 * mapped pages. */ 900 if (new_brk > target_brk) { 901 memset(g2h(target_brk), 0, new_brk - target_brk); 902 } 903 target_brk = new_brk; 904 DEBUGF_BRK(TARGET_ABI_FMT_lx " (new_brk <= brk_page)\n", target_brk); 905 return target_brk; 906 } 907 908 /* We need to allocate more memory after the brk... Note that 909 * we don't use MAP_FIXED because that will map over the top of 910 * any existing mapping (like the one with the host libc or qemu 911 * itself); instead we treat "mapped but at wrong address" as 912 * a failure and unmap again. 913 */ 914 new_alloc_size = HOST_PAGE_ALIGN(new_brk - brk_page); 915 mapped_addr = get_errno(target_mmap(brk_page, new_alloc_size, 916 PROT_READ|PROT_WRITE, 917 MAP_ANON|MAP_PRIVATE, 0, 0)); 918 919 if (mapped_addr == brk_page) { 920 /* Heap contents are initialized to zero, as for anonymous 921 * mapped pages. Technically the new pages are already 922 * initialized to zero since they *are* anonymous mapped 923 * pages, however we have to take care with the contents that 924 * come from the remaining part of the previous page: it may 925 * contains garbage data due to a previous heap usage (grown 926 * then shrunken). */ 927 memset(g2h(target_brk), 0, brk_page - target_brk); 928 929 target_brk = new_brk; 930 brk_page = HOST_PAGE_ALIGN(target_brk); 931 DEBUGF_BRK(TARGET_ABI_FMT_lx " (mapped_addr == brk_page)\n", 932 target_brk); 933 return target_brk; 934 } else if (mapped_addr != -1) { 935 /* Mapped but at wrong address, meaning there wasn't actually 936 * enough space for this brk. 937 */ 938 target_munmap(mapped_addr, new_alloc_size); 939 mapped_addr = -1; 940 DEBUGF_BRK(TARGET_ABI_FMT_lx " (mapped_addr != -1)\n", target_brk); 941 } 942 else { 943 DEBUGF_BRK(TARGET_ABI_FMT_lx " (otherwise)\n", target_brk); 944 } 945 946 #if defined(TARGET_ALPHA) 947 /* We (partially) emulate OSF/1 on Alpha, which requires we 948 return a proper errno, not an unchanged brk value. */ 949 return -TARGET_ENOMEM; 950 #endif 951 /* For everything else, return the previous break. */ 952 return target_brk; 953 } 954 955 static inline abi_long copy_from_user_fdset(fd_set *fds, 956 abi_ulong target_fds_addr, 957 int n) 958 { 959 int i, nw, j, k; 960 abi_ulong b, *target_fds; 961 962 nw = DIV_ROUND_UP(n, TARGET_ABI_BITS); 963 if (!(target_fds = lock_user(VERIFY_READ, 964 target_fds_addr, 965 sizeof(abi_ulong) * nw, 966 1))) 967 return -TARGET_EFAULT; 968 969 FD_ZERO(fds); 970 k = 0; 971 for (i = 0; i < nw; i++) { 972 /* grab the abi_ulong */ 973 __get_user(b, &target_fds[i]); 974 for (j = 0; j < TARGET_ABI_BITS; j++) { 975 /* check the bit inside the abi_ulong */ 976 if ((b >> j) & 1) 977 FD_SET(k, fds); 978 k++; 979 } 980 } 981 982 unlock_user(target_fds, target_fds_addr, 0); 983 984 return 0; 985 } 986 987 static inline abi_ulong copy_from_user_fdset_ptr(fd_set *fds, fd_set **fds_ptr, 988 abi_ulong target_fds_addr, 989 int n) 990 { 991 if (target_fds_addr) { 992 if (copy_from_user_fdset(fds, target_fds_addr, n)) 993 return -TARGET_EFAULT; 994 *fds_ptr = fds; 995 } else { 996 *fds_ptr = NULL; 997 } 998 return 0; 999 } 1000 1001 static inline abi_long copy_to_user_fdset(abi_ulong target_fds_addr, 1002 const fd_set *fds, 1003 int n) 1004 { 1005 int i, nw, j, k; 1006 abi_long v; 1007 abi_ulong *target_fds; 1008 1009 nw = DIV_ROUND_UP(n, TARGET_ABI_BITS); 1010 if (!(target_fds = lock_user(VERIFY_WRITE, 1011 target_fds_addr, 1012 sizeof(abi_ulong) * nw, 1013 0))) 1014 return -TARGET_EFAULT; 1015 1016 k = 0; 1017 for (i = 0; i < nw; i++) { 1018 v = 0; 1019 for (j = 0; j < TARGET_ABI_BITS; j++) { 1020 v |= ((abi_ulong)(FD_ISSET(k, fds) != 0) << j); 1021 k++; 1022 } 1023 __put_user(v, &target_fds[i]); 1024 } 1025 1026 unlock_user(target_fds, target_fds_addr, sizeof(abi_ulong) * nw); 1027 1028 return 0; 1029 } 1030 1031 #if defined(__alpha__) 1032 #define HOST_HZ 1024 1033 #else 1034 #define HOST_HZ 100 1035 #endif 1036 1037 static inline abi_long host_to_target_clock_t(long ticks) 1038 { 1039 #if HOST_HZ == TARGET_HZ 1040 return ticks; 1041 #else 1042 return ((int64_t)ticks * TARGET_HZ) / HOST_HZ; 1043 #endif 1044 } 1045 1046 static inline abi_long host_to_target_rusage(abi_ulong target_addr, 1047 const struct rusage *rusage) 1048 { 1049 struct target_rusage *target_rusage; 1050 1051 if (!lock_user_struct(VERIFY_WRITE, target_rusage, target_addr, 0)) 1052 return -TARGET_EFAULT; 1053 target_rusage->ru_utime.tv_sec = tswapal(rusage->ru_utime.tv_sec); 1054 target_rusage->ru_utime.tv_usec = tswapal(rusage->ru_utime.tv_usec); 1055 target_rusage->ru_stime.tv_sec = tswapal(rusage->ru_stime.tv_sec); 1056 target_rusage->ru_stime.tv_usec = tswapal(rusage->ru_stime.tv_usec); 1057 target_rusage->ru_maxrss = tswapal(rusage->ru_maxrss); 1058 target_rusage->ru_ixrss = tswapal(rusage->ru_ixrss); 1059 target_rusage->ru_idrss = tswapal(rusage->ru_idrss); 1060 target_rusage->ru_isrss = tswapal(rusage->ru_isrss); 1061 target_rusage->ru_minflt = tswapal(rusage->ru_minflt); 1062 target_rusage->ru_majflt = tswapal(rusage->ru_majflt); 1063 target_rusage->ru_nswap = tswapal(rusage->ru_nswap); 1064 target_rusage->ru_inblock = tswapal(rusage->ru_inblock); 1065 target_rusage->ru_oublock = tswapal(rusage->ru_oublock); 1066 target_rusage->ru_msgsnd = tswapal(rusage->ru_msgsnd); 1067 target_rusage->ru_msgrcv = tswapal(rusage->ru_msgrcv); 1068 target_rusage->ru_nsignals = tswapal(rusage->ru_nsignals); 1069 target_rusage->ru_nvcsw = tswapal(rusage->ru_nvcsw); 1070 target_rusage->ru_nivcsw = tswapal(rusage->ru_nivcsw); 1071 unlock_user_struct(target_rusage, target_addr, 1); 1072 1073 return 0; 1074 } 1075 1076 static inline rlim_t target_to_host_rlim(abi_ulong target_rlim) 1077 { 1078 abi_ulong target_rlim_swap; 1079 rlim_t result; 1080 1081 target_rlim_swap = tswapal(target_rlim); 1082 if (target_rlim_swap == TARGET_RLIM_INFINITY) 1083 return RLIM_INFINITY; 1084 1085 result = target_rlim_swap; 1086 if (target_rlim_swap != (rlim_t)result) 1087 return RLIM_INFINITY; 1088 1089 return result; 1090 } 1091 1092 static inline abi_ulong host_to_target_rlim(rlim_t rlim) 1093 { 1094 abi_ulong target_rlim_swap; 1095 abi_ulong result; 1096 1097 if (rlim == RLIM_INFINITY || rlim != (abi_long)rlim) 1098 target_rlim_swap = TARGET_RLIM_INFINITY; 1099 else 1100 target_rlim_swap = rlim; 1101 result = tswapal(target_rlim_swap); 1102 1103 return result; 1104 } 1105 1106 static inline int target_to_host_resource(int code) 1107 { 1108 switch (code) { 1109 case TARGET_RLIMIT_AS: 1110 return RLIMIT_AS; 1111 case TARGET_RLIMIT_CORE: 1112 return RLIMIT_CORE; 1113 case TARGET_RLIMIT_CPU: 1114 return RLIMIT_CPU; 1115 case TARGET_RLIMIT_DATA: 1116 return RLIMIT_DATA; 1117 case TARGET_RLIMIT_FSIZE: 1118 return RLIMIT_FSIZE; 1119 case TARGET_RLIMIT_LOCKS: 1120 return RLIMIT_LOCKS; 1121 case TARGET_RLIMIT_MEMLOCK: 1122 return RLIMIT_MEMLOCK; 1123 case TARGET_RLIMIT_MSGQUEUE: 1124 return RLIMIT_MSGQUEUE; 1125 case TARGET_RLIMIT_NICE: 1126 return RLIMIT_NICE; 1127 case TARGET_RLIMIT_NOFILE: 1128 return RLIMIT_NOFILE; 1129 case TARGET_RLIMIT_NPROC: 1130 return RLIMIT_NPROC; 1131 case TARGET_RLIMIT_RSS: 1132 return RLIMIT_RSS; 1133 case TARGET_RLIMIT_RTPRIO: 1134 return RLIMIT_RTPRIO; 1135 case TARGET_RLIMIT_SIGPENDING: 1136 return RLIMIT_SIGPENDING; 1137 case TARGET_RLIMIT_STACK: 1138 return RLIMIT_STACK; 1139 default: 1140 return code; 1141 } 1142 } 1143 1144 static inline abi_long copy_from_user_timeval(struct timeval *tv, 1145 abi_ulong target_tv_addr) 1146 { 1147 struct target_timeval *target_tv; 1148 1149 if (!lock_user_struct(VERIFY_READ, target_tv, target_tv_addr, 1)) 1150 return -TARGET_EFAULT; 1151 1152 __get_user(tv->tv_sec, &target_tv->tv_sec); 1153 __get_user(tv->tv_usec, &target_tv->tv_usec); 1154 1155 unlock_user_struct(target_tv, target_tv_addr, 0); 1156 1157 return 0; 1158 } 1159 1160 static inline abi_long copy_to_user_timeval(abi_ulong target_tv_addr, 1161 const struct timeval *tv) 1162 { 1163 struct target_timeval *target_tv; 1164 1165 if (!lock_user_struct(VERIFY_WRITE, target_tv, target_tv_addr, 0)) 1166 return -TARGET_EFAULT; 1167 1168 __put_user(tv->tv_sec, &target_tv->tv_sec); 1169 __put_user(tv->tv_usec, &target_tv->tv_usec); 1170 1171 unlock_user_struct(target_tv, target_tv_addr, 1); 1172 1173 return 0; 1174 } 1175 1176 static inline abi_long copy_from_user_timezone(struct timezone *tz, 1177 abi_ulong target_tz_addr) 1178 { 1179 struct target_timezone *target_tz; 1180 1181 if (!lock_user_struct(VERIFY_READ, target_tz, target_tz_addr, 1)) { 1182 return -TARGET_EFAULT; 1183 } 1184 1185 __get_user(tz->tz_minuteswest, &target_tz->tz_minuteswest); 1186 __get_user(tz->tz_dsttime, &target_tz->tz_dsttime); 1187 1188 unlock_user_struct(target_tz, target_tz_addr, 0); 1189 1190 return 0; 1191 } 1192 1193 #if defined(TARGET_NR_mq_open) && defined(__NR_mq_open) 1194 #include <mqueue.h> 1195 1196 static inline abi_long copy_from_user_mq_attr(struct mq_attr *attr, 1197 abi_ulong target_mq_attr_addr) 1198 { 1199 struct target_mq_attr *target_mq_attr; 1200 1201 if (!lock_user_struct(VERIFY_READ, target_mq_attr, 1202 target_mq_attr_addr, 1)) 1203 return -TARGET_EFAULT; 1204 1205 __get_user(attr->mq_flags, &target_mq_attr->mq_flags); 1206 __get_user(attr->mq_maxmsg, &target_mq_attr->mq_maxmsg); 1207 __get_user(attr->mq_msgsize, &target_mq_attr->mq_msgsize); 1208 __get_user(attr->mq_curmsgs, &target_mq_attr->mq_curmsgs); 1209 1210 unlock_user_struct(target_mq_attr, target_mq_attr_addr, 0); 1211 1212 return 0; 1213 } 1214 1215 static inline abi_long copy_to_user_mq_attr(abi_ulong target_mq_attr_addr, 1216 const struct mq_attr *attr) 1217 { 1218 struct target_mq_attr *target_mq_attr; 1219 1220 if (!lock_user_struct(VERIFY_WRITE, target_mq_attr, 1221 target_mq_attr_addr, 0)) 1222 return -TARGET_EFAULT; 1223 1224 __put_user(attr->mq_flags, &target_mq_attr->mq_flags); 1225 __put_user(attr->mq_maxmsg, &target_mq_attr->mq_maxmsg); 1226 __put_user(attr->mq_msgsize, &target_mq_attr->mq_msgsize); 1227 __put_user(attr->mq_curmsgs, &target_mq_attr->mq_curmsgs); 1228 1229 unlock_user_struct(target_mq_attr, target_mq_attr_addr, 1); 1230 1231 return 0; 1232 } 1233 #endif 1234 1235 #if defined(TARGET_NR_select) || defined(TARGET_NR__newselect) 1236 /* do_select() must return target values and target errnos. */ 1237 static abi_long do_select(int n, 1238 abi_ulong rfd_addr, abi_ulong wfd_addr, 1239 abi_ulong efd_addr, abi_ulong target_tv_addr) 1240 { 1241 fd_set rfds, wfds, efds; 1242 fd_set *rfds_ptr, *wfds_ptr, *efds_ptr; 1243 struct timeval tv; 1244 struct timespec ts, *ts_ptr; 1245 abi_long ret; 1246 1247 ret = copy_from_user_fdset_ptr(&rfds, &rfds_ptr, rfd_addr, n); 1248 if (ret) { 1249 return ret; 1250 } 1251 ret = copy_from_user_fdset_ptr(&wfds, &wfds_ptr, wfd_addr, n); 1252 if (ret) { 1253 return ret; 1254 } 1255 ret = copy_from_user_fdset_ptr(&efds, &efds_ptr, efd_addr, n); 1256 if (ret) { 1257 return ret; 1258 } 1259 1260 if (target_tv_addr) { 1261 if (copy_from_user_timeval(&tv, target_tv_addr)) 1262 return -TARGET_EFAULT; 1263 ts.tv_sec = tv.tv_sec; 1264 ts.tv_nsec = tv.tv_usec * 1000; 1265 ts_ptr = &ts; 1266 } else { 1267 ts_ptr = NULL; 1268 } 1269 1270 ret = get_errno(safe_pselect6(n, rfds_ptr, wfds_ptr, efds_ptr, 1271 ts_ptr, NULL)); 1272 1273 if (!is_error(ret)) { 1274 if (rfd_addr && copy_to_user_fdset(rfd_addr, &rfds, n)) 1275 return -TARGET_EFAULT; 1276 if (wfd_addr && copy_to_user_fdset(wfd_addr, &wfds, n)) 1277 return -TARGET_EFAULT; 1278 if (efd_addr && copy_to_user_fdset(efd_addr, &efds, n)) 1279 return -TARGET_EFAULT; 1280 1281 if (target_tv_addr) { 1282 tv.tv_sec = ts.tv_sec; 1283 tv.tv_usec = ts.tv_nsec / 1000; 1284 if (copy_to_user_timeval(target_tv_addr, &tv)) { 1285 return -TARGET_EFAULT; 1286 } 1287 } 1288 } 1289 1290 return ret; 1291 } 1292 1293 #if defined(TARGET_WANT_OLD_SYS_SELECT) 1294 static abi_long do_old_select(abi_ulong arg1) 1295 { 1296 struct target_sel_arg_struct *sel; 1297 abi_ulong inp, outp, exp, tvp; 1298 long nsel; 1299 1300 if (!lock_user_struct(VERIFY_READ, sel, arg1, 1)) { 1301 return -TARGET_EFAULT; 1302 } 1303 1304 nsel = tswapal(sel->n); 1305 inp = tswapal(sel->inp); 1306 outp = tswapal(sel->outp); 1307 exp = tswapal(sel->exp); 1308 tvp = tswapal(sel->tvp); 1309 1310 unlock_user_struct(sel, arg1, 0); 1311 1312 return do_select(nsel, inp, outp, exp, tvp); 1313 } 1314 #endif 1315 #endif 1316 1317 static abi_long do_pipe2(int host_pipe[], int flags) 1318 { 1319 #ifdef CONFIG_PIPE2 1320 return pipe2(host_pipe, flags); 1321 #else 1322 return -ENOSYS; 1323 #endif 1324 } 1325 1326 static abi_long do_pipe(void *cpu_env, abi_ulong pipedes, 1327 int flags, int is_pipe2) 1328 { 1329 int host_pipe[2]; 1330 abi_long ret; 1331 ret = flags ? do_pipe2(host_pipe, flags) : pipe(host_pipe); 1332 1333 if (is_error(ret)) 1334 return get_errno(ret); 1335 1336 /* Several targets have special calling conventions for the original 1337 pipe syscall, but didn't replicate this into the pipe2 syscall. */ 1338 if (!is_pipe2) { 1339 #if defined(TARGET_ALPHA) 1340 ((CPUAlphaState *)cpu_env)->ir[IR_A4] = host_pipe[1]; 1341 return host_pipe[0]; 1342 #elif defined(TARGET_MIPS) 1343 ((CPUMIPSState*)cpu_env)->active_tc.gpr[3] = host_pipe[1]; 1344 return host_pipe[0]; 1345 #elif defined(TARGET_SH4) 1346 ((CPUSH4State*)cpu_env)->gregs[1] = host_pipe[1]; 1347 return host_pipe[0]; 1348 #elif defined(TARGET_SPARC) 1349 ((CPUSPARCState*)cpu_env)->regwptr[1] = host_pipe[1]; 1350 return host_pipe[0]; 1351 #endif 1352 } 1353 1354 if (put_user_s32(host_pipe[0], pipedes) 1355 || put_user_s32(host_pipe[1], pipedes + sizeof(host_pipe[0]))) 1356 return -TARGET_EFAULT; 1357 return get_errno(ret); 1358 } 1359 1360 static inline abi_long target_to_host_ip_mreq(struct ip_mreqn *mreqn, 1361 abi_ulong target_addr, 1362 socklen_t len) 1363 { 1364 struct target_ip_mreqn *target_smreqn; 1365 1366 target_smreqn = lock_user(VERIFY_READ, target_addr, len, 1); 1367 if (!target_smreqn) 1368 return -TARGET_EFAULT; 1369 mreqn->imr_multiaddr.s_addr = target_smreqn->imr_multiaddr.s_addr; 1370 mreqn->imr_address.s_addr = target_smreqn->imr_address.s_addr; 1371 if (len == sizeof(struct target_ip_mreqn)) 1372 mreqn->imr_ifindex = tswapal(target_smreqn->imr_ifindex); 1373 unlock_user(target_smreqn, target_addr, 0); 1374 1375 return 0; 1376 } 1377 1378 static inline abi_long target_to_host_sockaddr(int fd, struct sockaddr *addr, 1379 abi_ulong target_addr, 1380 socklen_t len) 1381 { 1382 const socklen_t unix_maxlen = sizeof (struct sockaddr_un); 1383 sa_family_t sa_family; 1384 struct target_sockaddr *target_saddr; 1385 1386 if (fd_trans_target_to_host_addr(fd)) { 1387 return fd_trans_target_to_host_addr(fd)(addr, target_addr, len); 1388 } 1389 1390 target_saddr = lock_user(VERIFY_READ, target_addr, len, 1); 1391 if (!target_saddr) 1392 return -TARGET_EFAULT; 1393 1394 sa_family = tswap16(target_saddr->sa_family); 1395 1396 /* Oops. The caller might send a incomplete sun_path; sun_path 1397 * must be terminated by \0 (see the manual page), but 1398 * unfortunately it is quite common to specify sockaddr_un 1399 * length as "strlen(x->sun_path)" while it should be 1400 * "strlen(...) + 1". We'll fix that here if needed. 1401 * Linux kernel has a similar feature. 1402 */ 1403 1404 if (sa_family == AF_UNIX) { 1405 if (len < unix_maxlen && len > 0) { 1406 char *cp = (char*)target_saddr; 1407 1408 if ( cp[len-1] && !cp[len] ) 1409 len++; 1410 } 1411 if (len > unix_maxlen) 1412 len = unix_maxlen; 1413 } 1414 1415 memcpy(addr, target_saddr, len); 1416 addr->sa_family = sa_family; 1417 if (sa_family == AF_NETLINK) { 1418 struct sockaddr_nl *nladdr; 1419 1420 nladdr = (struct sockaddr_nl *)addr; 1421 nladdr->nl_pid = tswap32(nladdr->nl_pid); 1422 nladdr->nl_groups = tswap32(nladdr->nl_groups); 1423 } else if (sa_family == AF_PACKET) { 1424 struct target_sockaddr_ll *lladdr; 1425 1426 lladdr = (struct target_sockaddr_ll *)addr; 1427 lladdr->sll_ifindex = tswap32(lladdr->sll_ifindex); 1428 lladdr->sll_hatype = tswap16(lladdr->sll_hatype); 1429 } 1430 unlock_user(target_saddr, target_addr, 0); 1431 1432 return 0; 1433 } 1434 1435 static inline abi_long host_to_target_sockaddr(abi_ulong target_addr, 1436 struct sockaddr *addr, 1437 socklen_t len) 1438 { 1439 struct target_sockaddr *target_saddr; 1440 1441 if (len == 0) { 1442 return 0; 1443 } 1444 assert(addr); 1445 1446 target_saddr = lock_user(VERIFY_WRITE, target_addr, len, 0); 1447 if (!target_saddr) 1448 return -TARGET_EFAULT; 1449 memcpy(target_saddr, addr, len); 1450 if (len >= offsetof(struct target_sockaddr, sa_family) + 1451 sizeof(target_saddr->sa_family)) { 1452 target_saddr->sa_family = tswap16(addr->sa_family); 1453 } 1454 if (addr->sa_family == AF_NETLINK && len >= sizeof(struct sockaddr_nl)) { 1455 struct sockaddr_nl *target_nl = (struct sockaddr_nl *)target_saddr; 1456 target_nl->nl_pid = tswap32(target_nl->nl_pid); 1457 target_nl->nl_groups = tswap32(target_nl->nl_groups); 1458 } else if (addr->sa_family == AF_PACKET) { 1459 struct sockaddr_ll *target_ll = (struct sockaddr_ll *)target_saddr; 1460 target_ll->sll_ifindex = tswap32(target_ll->sll_ifindex); 1461 target_ll->sll_hatype = tswap16(target_ll->sll_hatype); 1462 } else if (addr->sa_family == AF_INET6 && 1463 len >= sizeof(struct target_sockaddr_in6)) { 1464 struct target_sockaddr_in6 *target_in6 = 1465 (struct target_sockaddr_in6 *)target_saddr; 1466 target_in6->sin6_scope_id = tswap16(target_in6->sin6_scope_id); 1467 } 1468 unlock_user(target_saddr, target_addr, len); 1469 1470 return 0; 1471 } 1472 1473 static inline abi_long target_to_host_cmsg(struct msghdr *msgh, 1474 struct target_msghdr *target_msgh) 1475 { 1476 struct cmsghdr *cmsg = CMSG_FIRSTHDR(msgh); 1477 abi_long msg_controllen; 1478 abi_ulong target_cmsg_addr; 1479 struct target_cmsghdr *target_cmsg, *target_cmsg_start; 1480 socklen_t space = 0; 1481 1482 msg_controllen = tswapal(target_msgh->msg_controllen); 1483 if (msg_controllen < sizeof (struct target_cmsghdr)) 1484 goto the_end; 1485 target_cmsg_addr = tswapal(target_msgh->msg_control); 1486 target_cmsg = lock_user(VERIFY_READ, target_cmsg_addr, msg_controllen, 1); 1487 target_cmsg_start = target_cmsg; 1488 if (!target_cmsg) 1489 return -TARGET_EFAULT; 1490 1491 while (cmsg && target_cmsg) { 1492 void *data = CMSG_DATA(cmsg); 1493 void *target_data = TARGET_CMSG_DATA(target_cmsg); 1494 1495 int len = tswapal(target_cmsg->cmsg_len) 1496 - sizeof(struct target_cmsghdr); 1497 1498 space += CMSG_SPACE(len); 1499 if (space > msgh->msg_controllen) { 1500 space -= CMSG_SPACE(len); 1501 /* This is a QEMU bug, since we allocated the payload 1502 * area ourselves (unlike overflow in host-to-target 1503 * conversion, which is just the guest giving us a buffer 1504 * that's too small). It can't happen for the payload types 1505 * we currently support; if it becomes an issue in future 1506 * we would need to improve our allocation strategy to 1507 * something more intelligent than "twice the size of the 1508 * target buffer we're reading from". 1509 */ 1510 gemu_log("Host cmsg overflow\n"); 1511 break; 1512 } 1513 1514 if (tswap32(target_cmsg->cmsg_level) == TARGET_SOL_SOCKET) { 1515 cmsg->cmsg_level = SOL_SOCKET; 1516 } else { 1517 cmsg->cmsg_level = tswap32(target_cmsg->cmsg_level); 1518 } 1519 cmsg->cmsg_type = tswap32(target_cmsg->cmsg_type); 1520 cmsg->cmsg_len = CMSG_LEN(len); 1521 1522 if (cmsg->cmsg_level == SOL_SOCKET && cmsg->cmsg_type == SCM_RIGHTS) { 1523 int *fd = (int *)data; 1524 int *target_fd = (int *)target_data; 1525 int i, numfds = len / sizeof(int); 1526 1527 for (i = 0; i < numfds; i++) { 1528 __get_user(fd[i], target_fd + i); 1529 } 1530 } else if (cmsg->cmsg_level == SOL_SOCKET 1531 && cmsg->cmsg_type == SCM_CREDENTIALS) { 1532 struct ucred *cred = (struct ucred *)data; 1533 struct target_ucred *target_cred = 1534 (struct target_ucred *)target_data; 1535 1536 __get_user(cred->pid, &target_cred->pid); 1537 __get_user(cred->uid, &target_cred->uid); 1538 __get_user(cred->gid, &target_cred->gid); 1539 } else { 1540 gemu_log("Unsupported ancillary data: %d/%d\n", 1541 cmsg->cmsg_level, cmsg->cmsg_type); 1542 memcpy(data, target_data, len); 1543 } 1544 1545 cmsg = CMSG_NXTHDR(msgh, cmsg); 1546 target_cmsg = TARGET_CMSG_NXTHDR(target_msgh, target_cmsg, 1547 target_cmsg_start); 1548 } 1549 unlock_user(target_cmsg, target_cmsg_addr, 0); 1550 the_end: 1551 msgh->msg_controllen = space; 1552 return 0; 1553 } 1554 1555 static inline abi_long host_to_target_cmsg(struct target_msghdr *target_msgh, 1556 struct msghdr *msgh) 1557 { 1558 struct cmsghdr *cmsg = CMSG_FIRSTHDR(msgh); 1559 abi_long msg_controllen; 1560 abi_ulong target_cmsg_addr; 1561 struct target_cmsghdr *target_cmsg, *target_cmsg_start; 1562 socklen_t space = 0; 1563 1564 msg_controllen = tswapal(target_msgh->msg_controllen); 1565 if (msg_controllen < sizeof (struct target_cmsghdr)) 1566 goto the_end; 1567 target_cmsg_addr = tswapal(target_msgh->msg_control); 1568 target_cmsg = lock_user(VERIFY_WRITE, target_cmsg_addr, msg_controllen, 0); 1569 target_cmsg_start = target_cmsg; 1570 if (!target_cmsg) 1571 return -TARGET_EFAULT; 1572 1573 while (cmsg && target_cmsg) { 1574 void *data = CMSG_DATA(cmsg); 1575 void *target_data = TARGET_CMSG_DATA(target_cmsg); 1576 1577 int len = cmsg->cmsg_len - sizeof(struct cmsghdr); 1578 int tgt_len, tgt_space; 1579 1580 /* We never copy a half-header but may copy half-data; 1581 * this is Linux's behaviour in put_cmsg(). Note that 1582 * truncation here is a guest problem (which we report 1583 * to the guest via the CTRUNC bit), unlike truncation 1584 * in target_to_host_cmsg, which is a QEMU bug. 1585 */ 1586 if (msg_controllen < sizeof(struct target_cmsghdr)) { 1587 target_msgh->msg_flags |= tswap32(MSG_CTRUNC); 1588 break; 1589 } 1590 1591 if (cmsg->cmsg_level == SOL_SOCKET) { 1592 target_cmsg->cmsg_level = tswap32(TARGET_SOL_SOCKET); 1593 } else { 1594 target_cmsg->cmsg_level = tswap32(cmsg->cmsg_level); 1595 } 1596 target_cmsg->cmsg_type = tswap32(cmsg->cmsg_type); 1597 1598 /* Payload types which need a different size of payload on 1599 * the target must adjust tgt_len here. 1600 */ 1601 tgt_len = len; 1602 switch (cmsg->cmsg_level) { 1603 case SOL_SOCKET: 1604 switch (cmsg->cmsg_type) { 1605 case SO_TIMESTAMP: 1606 tgt_len = sizeof(struct target_timeval); 1607 break; 1608 default: 1609 break; 1610 } 1611 break; 1612 default: 1613 break; 1614 } 1615 1616 if (msg_controllen < TARGET_CMSG_LEN(tgt_len)) { 1617 target_msgh->msg_flags |= tswap32(MSG_CTRUNC); 1618 tgt_len = msg_controllen - sizeof(struct target_cmsghdr); 1619 } 1620 1621 /* We must now copy-and-convert len bytes of payload 1622 * into tgt_len bytes of destination space. Bear in mind 1623 * that in both source and destination we may be dealing 1624 * with a truncated value! 1625 */ 1626 switch (cmsg->cmsg_level) { 1627 case SOL_SOCKET: 1628 switch (cmsg->cmsg_type) { 1629 case SCM_RIGHTS: 1630 { 1631 int *fd = (int *)data; 1632 int *target_fd = (int *)target_data; 1633 int i, numfds = tgt_len / sizeof(int); 1634 1635 for (i = 0; i < numfds; i++) { 1636 __put_user(fd[i], target_fd + i); 1637 } 1638 break; 1639 } 1640 case SO_TIMESTAMP: 1641 { 1642 struct timeval *tv = (struct timeval *)data; 1643 struct target_timeval *target_tv = 1644 (struct target_timeval *)target_data; 1645 1646 if (len != sizeof(struct timeval) || 1647 tgt_len != sizeof(struct target_timeval)) { 1648 goto unimplemented; 1649 } 1650 1651 /* copy struct timeval to target */ 1652 __put_user(tv->tv_sec, &target_tv->tv_sec); 1653 __put_user(tv->tv_usec, &target_tv->tv_usec); 1654 break; 1655 } 1656 case SCM_CREDENTIALS: 1657 { 1658 struct ucred *cred = (struct ucred *)data; 1659 struct target_ucred *target_cred = 1660 (struct target_ucred *)target_data; 1661 1662 __put_user(cred->pid, &target_cred->pid); 1663 __put_user(cred->uid, &target_cred->uid); 1664 __put_user(cred->gid, &target_cred->gid); 1665 break; 1666 } 1667 default: 1668 goto unimplemented; 1669 } 1670 break; 1671 1672 case SOL_IP: 1673 switch (cmsg->cmsg_type) { 1674 case IP_TTL: 1675 { 1676 uint32_t *v = (uint32_t *)data; 1677 uint32_t *t_int = (uint32_t *)target_data; 1678 1679 if (len != sizeof(uint32_t) || 1680 tgt_len != sizeof(uint32_t)) { 1681 goto unimplemented; 1682 } 1683 __put_user(*v, t_int); 1684 break; 1685 } 1686 case IP_RECVERR: 1687 { 1688 struct errhdr_t { 1689 struct sock_extended_err ee; 1690 struct sockaddr_in offender; 1691 }; 1692 struct errhdr_t *errh = (struct errhdr_t *)data; 1693 struct errhdr_t *target_errh = 1694 (struct errhdr_t *)target_data; 1695 1696 if (len != sizeof(struct errhdr_t) || 1697 tgt_len != sizeof(struct errhdr_t)) { 1698 goto unimplemented; 1699 } 1700 __put_user(errh->ee.ee_errno, &target_errh->ee.ee_errno); 1701 __put_user(errh->ee.ee_origin, &target_errh->ee.ee_origin); 1702 __put_user(errh->ee.ee_type, &target_errh->ee.ee_type); 1703 __put_user(errh->ee.ee_code, &target_errh->ee.ee_code); 1704 __put_user(errh->ee.ee_pad, &target_errh->ee.ee_pad); 1705 __put_user(errh->ee.ee_info, &target_errh->ee.ee_info); 1706 __put_user(errh->ee.ee_data, &target_errh->ee.ee_data); 1707 host_to_target_sockaddr((unsigned long) &target_errh->offender, 1708 (void *) &errh->offender, sizeof(errh->offender)); 1709 break; 1710 } 1711 default: 1712 goto unimplemented; 1713 } 1714 break; 1715 1716 case SOL_IPV6: 1717 switch (cmsg->cmsg_type) { 1718 case IPV6_HOPLIMIT: 1719 { 1720 uint32_t *v = (uint32_t *)data; 1721 uint32_t *t_int = (uint32_t *)target_data; 1722 1723 if (len != sizeof(uint32_t) || 1724 tgt_len != sizeof(uint32_t)) { 1725 goto unimplemented; 1726 } 1727 __put_user(*v, t_int); 1728 break; 1729 } 1730 case IPV6_RECVERR: 1731 { 1732 struct errhdr6_t { 1733 struct sock_extended_err ee; 1734 struct sockaddr_in6 offender; 1735 }; 1736 struct errhdr6_t *errh = (struct errhdr6_t *)data; 1737 struct errhdr6_t *target_errh = 1738 (struct errhdr6_t *)target_data; 1739 1740 if (len != sizeof(struct errhdr6_t) || 1741 tgt_len != sizeof(struct errhdr6_t)) { 1742 goto unimplemented; 1743 } 1744 __put_user(errh->ee.ee_errno, &target_errh->ee.ee_errno); 1745 __put_user(errh->ee.ee_origin, &target_errh->ee.ee_origin); 1746 __put_user(errh->ee.ee_type, &target_errh->ee.ee_type); 1747 __put_user(errh->ee.ee_code, &target_errh->ee.ee_code); 1748 __put_user(errh->ee.ee_pad, &target_errh->ee.ee_pad); 1749 __put_user(errh->ee.ee_info, &target_errh->ee.ee_info); 1750 __put_user(errh->ee.ee_data, &target_errh->ee.ee_data); 1751 host_to_target_sockaddr((unsigned long) &target_errh->offender, 1752 (void *) &errh->offender, sizeof(errh->offender)); 1753 break; 1754 } 1755 default: 1756 goto unimplemented; 1757 } 1758 break; 1759 1760 default: 1761 unimplemented: 1762 gemu_log("Unsupported ancillary data: %d/%d\n", 1763 cmsg->cmsg_level, cmsg->cmsg_type); 1764 memcpy(target_data, data, MIN(len, tgt_len)); 1765 if (tgt_len > len) { 1766 memset(target_data + len, 0, tgt_len - len); 1767 } 1768 } 1769 1770 target_cmsg->cmsg_len = tswapal(TARGET_CMSG_LEN(tgt_len)); 1771 tgt_space = TARGET_CMSG_SPACE(tgt_len); 1772 if (msg_controllen < tgt_space) { 1773 tgt_space = msg_controllen; 1774 } 1775 msg_controllen -= tgt_space; 1776 space += tgt_space; 1777 cmsg = CMSG_NXTHDR(msgh, cmsg); 1778 target_cmsg = TARGET_CMSG_NXTHDR(target_msgh, target_cmsg, 1779 target_cmsg_start); 1780 } 1781 unlock_user(target_cmsg, target_cmsg_addr, space); 1782 the_end: 1783 target_msgh->msg_controllen = tswapal(space); 1784 return 0; 1785 } 1786 1787 /* do_setsockopt() Must return target values and target errnos. */ 1788 static abi_long do_setsockopt(int sockfd, int level, int optname, 1789 abi_ulong optval_addr, socklen_t optlen) 1790 { 1791 abi_long ret; 1792 int val; 1793 struct ip_mreqn *ip_mreq; 1794 struct ip_mreq_source *ip_mreq_source; 1795 1796 switch(level) { 1797 case SOL_TCP: 1798 /* TCP options all take an 'int' value. */ 1799 if (optlen < sizeof(uint32_t)) 1800 return -TARGET_EINVAL; 1801 1802 if (get_user_u32(val, optval_addr)) 1803 return -TARGET_EFAULT; 1804 ret = get_errno(setsockopt(sockfd, level, optname, &val, sizeof(val))); 1805 break; 1806 case SOL_IP: 1807 switch(optname) { 1808 case IP_TOS: 1809 case IP_TTL: 1810 case IP_HDRINCL: 1811 case IP_ROUTER_ALERT: 1812 case IP_RECVOPTS: 1813 case IP_RETOPTS: 1814 case IP_PKTINFO: 1815 case IP_MTU_DISCOVER: 1816 case IP_RECVERR: 1817 case IP_RECVTTL: 1818 case IP_RECVTOS: 1819 #ifdef IP_FREEBIND 1820 case IP_FREEBIND: 1821 #endif 1822 case IP_MULTICAST_TTL: 1823 case IP_MULTICAST_LOOP: 1824 val = 0; 1825 if (optlen >= sizeof(uint32_t)) { 1826 if (get_user_u32(val, optval_addr)) 1827 return -TARGET_EFAULT; 1828 } else if (optlen >= 1) { 1829 if (get_user_u8(val, optval_addr)) 1830 return -TARGET_EFAULT; 1831 } 1832 ret = get_errno(setsockopt(sockfd, level, optname, &val, sizeof(val))); 1833 break; 1834 case IP_ADD_MEMBERSHIP: 1835 case IP_DROP_MEMBERSHIP: 1836 if (optlen < sizeof (struct target_ip_mreq) || 1837 optlen > sizeof (struct target_ip_mreqn)) 1838 return -TARGET_EINVAL; 1839 1840 ip_mreq = (struct ip_mreqn *) alloca(optlen); 1841 target_to_host_ip_mreq(ip_mreq, optval_addr, optlen); 1842 ret = get_errno(setsockopt(sockfd, level, optname, ip_mreq, optlen)); 1843 break; 1844 1845 case IP_BLOCK_SOURCE: 1846 case IP_UNBLOCK_SOURCE: 1847 case IP_ADD_SOURCE_MEMBERSHIP: 1848 case IP_DROP_SOURCE_MEMBERSHIP: 1849 if (optlen != sizeof (struct target_ip_mreq_source)) 1850 return -TARGET_EINVAL; 1851 1852 ip_mreq_source = lock_user(VERIFY_READ, optval_addr, optlen, 1); 1853 ret = get_errno(setsockopt(sockfd, level, optname, ip_mreq_source, optlen)); 1854 unlock_user (ip_mreq_source, optval_addr, 0); 1855 break; 1856 1857 default: 1858 goto unimplemented; 1859 } 1860 break; 1861 case SOL_IPV6: 1862 switch (optname) { 1863 case IPV6_MTU_DISCOVER: 1864 case IPV6_MTU: 1865 case IPV6_V6ONLY: 1866 case IPV6_RECVPKTINFO: 1867 case IPV6_UNICAST_HOPS: 1868 case IPV6_MULTICAST_HOPS: 1869 case IPV6_MULTICAST_LOOP: 1870 case IPV6_RECVERR: 1871 case IPV6_RECVHOPLIMIT: 1872 case IPV6_2292HOPLIMIT: 1873 case IPV6_CHECKSUM: 1874 val = 0; 1875 if (optlen < sizeof(uint32_t)) { 1876 return -TARGET_EINVAL; 1877 } 1878 if (get_user_u32(val, optval_addr)) { 1879 return -TARGET_EFAULT; 1880 } 1881 ret = get_errno(setsockopt(sockfd, level, optname, 1882 &val, sizeof(val))); 1883 break; 1884 case IPV6_PKTINFO: 1885 { 1886 struct in6_pktinfo pki; 1887 1888 if (optlen < sizeof(pki)) { 1889 return -TARGET_EINVAL; 1890 } 1891 1892 if (copy_from_user(&pki, optval_addr, sizeof(pki))) { 1893 return -TARGET_EFAULT; 1894 } 1895 1896 pki.ipi6_ifindex = tswap32(pki.ipi6_ifindex); 1897 1898 ret = get_errno(setsockopt(sockfd, level, optname, 1899 &pki, sizeof(pki))); 1900 break; 1901 } 1902 default: 1903 goto unimplemented; 1904 } 1905 break; 1906 case SOL_ICMPV6: 1907 switch (optname) { 1908 case ICMPV6_FILTER: 1909 { 1910 struct icmp6_filter icmp6f; 1911 1912 if (optlen > sizeof(icmp6f)) { 1913 optlen = sizeof(icmp6f); 1914 } 1915 1916 if (copy_from_user(&icmp6f, optval_addr, optlen)) { 1917 return -TARGET_EFAULT; 1918 } 1919 1920 for (val = 0; val < 8; val++) { 1921 icmp6f.data[val] = tswap32(icmp6f.data[val]); 1922 } 1923 1924 ret = get_errno(setsockopt(sockfd, level, optname, 1925 &icmp6f, optlen)); 1926 break; 1927 } 1928 default: 1929 goto unimplemented; 1930 } 1931 break; 1932 case SOL_RAW: 1933 switch (optname) { 1934 case ICMP_FILTER: 1935 case IPV6_CHECKSUM: 1936 /* those take an u32 value */ 1937 if (optlen < sizeof(uint32_t)) { 1938 return -TARGET_EINVAL; 1939 } 1940 1941 if (get_user_u32(val, optval_addr)) { 1942 return -TARGET_EFAULT; 1943 } 1944 ret = get_errno(setsockopt(sockfd, level, optname, 1945 &val, sizeof(val))); 1946 break; 1947 1948 default: 1949 goto unimplemented; 1950 } 1951 break; 1952 case TARGET_SOL_SOCKET: 1953 switch (optname) { 1954 case TARGET_SO_RCVTIMEO: 1955 { 1956 struct timeval tv; 1957 1958 optname = SO_RCVTIMEO; 1959 1960 set_timeout: 1961 if (optlen != sizeof(struct target_timeval)) { 1962 return -TARGET_EINVAL; 1963 } 1964 1965 if (copy_from_user_timeval(&tv, optval_addr)) { 1966 return -TARGET_EFAULT; 1967 } 1968 1969 ret = get_errno(setsockopt(sockfd, SOL_SOCKET, optname, 1970 &tv, sizeof(tv))); 1971 return ret; 1972 } 1973 case TARGET_SO_SNDTIMEO: 1974 optname = SO_SNDTIMEO; 1975 goto set_timeout; 1976 case TARGET_SO_ATTACH_FILTER: 1977 { 1978 struct target_sock_fprog *tfprog; 1979 struct target_sock_filter *tfilter; 1980 struct sock_fprog fprog; 1981 struct sock_filter *filter; 1982 int i; 1983 1984 if (optlen != sizeof(*tfprog)) { 1985 return -TARGET_EINVAL; 1986 } 1987 if (!lock_user_struct(VERIFY_READ, tfprog, optval_addr, 0)) { 1988 return -TARGET_EFAULT; 1989 } 1990 if (!lock_user_struct(VERIFY_READ, tfilter, 1991 tswapal(tfprog->filter), 0)) { 1992 unlock_user_struct(tfprog, optval_addr, 1); 1993 return -TARGET_EFAULT; 1994 } 1995 1996 fprog.len = tswap16(tfprog->len); 1997 filter = g_try_new(struct sock_filter, fprog.len); 1998 if (filter == NULL) { 1999 unlock_user_struct(tfilter, tfprog->filter, 1); 2000 unlock_user_struct(tfprog, optval_addr, 1); 2001 return -TARGET_ENOMEM; 2002 } 2003 for (i = 0; i < fprog.len; i++) { 2004 filter[i].code = tswap16(tfilter[i].code); 2005 filter[i].jt = tfilter[i].jt; 2006 filter[i].jf = tfilter[i].jf; 2007 filter[i].k = tswap32(tfilter[i].k); 2008 } 2009 fprog.filter = filter; 2010 2011 ret = get_errno(setsockopt(sockfd, SOL_SOCKET, 2012 SO_ATTACH_FILTER, &fprog, sizeof(fprog))); 2013 g_free(filter); 2014 2015 unlock_user_struct(tfilter, tfprog->filter, 1); 2016 unlock_user_struct(tfprog, optval_addr, 1); 2017 return ret; 2018 } 2019 case TARGET_SO_BINDTODEVICE: 2020 { 2021 char *dev_ifname, *addr_ifname; 2022 2023 if (optlen > IFNAMSIZ - 1) { 2024 optlen = IFNAMSIZ - 1; 2025 } 2026 dev_ifname = lock_user(VERIFY_READ, optval_addr, optlen, 1); 2027 if (!dev_ifname) { 2028 return -TARGET_EFAULT; 2029 } 2030 optname = SO_BINDTODEVICE; 2031 addr_ifname = alloca(IFNAMSIZ); 2032 memcpy(addr_ifname, dev_ifname, optlen); 2033 addr_ifname[optlen] = 0; 2034 ret = get_errno(setsockopt(sockfd, SOL_SOCKET, optname, 2035 addr_ifname, optlen)); 2036 unlock_user (dev_ifname, optval_addr, 0); 2037 return ret; 2038 } 2039 case TARGET_SO_LINGER: 2040 { 2041 struct linger lg; 2042 struct target_linger *tlg; 2043 2044 if (optlen != sizeof(struct target_linger)) { 2045 return -TARGET_EINVAL; 2046 } 2047 if (!lock_user_struct(VERIFY_READ, tlg, optval_addr, 1)) { 2048 return -TARGET_EFAULT; 2049 } 2050 __get_user(lg.l_onoff, &tlg->l_onoff); 2051 __get_user(lg.l_linger, &tlg->l_linger); 2052 ret = get_errno(setsockopt(sockfd, SOL_SOCKET, SO_LINGER, 2053 &lg, sizeof(lg))); 2054 unlock_user_struct(tlg, optval_addr, 0); 2055 return ret; 2056 } 2057 /* Options with 'int' argument. */ 2058 case TARGET_SO_DEBUG: 2059 optname = SO_DEBUG; 2060 break; 2061 case TARGET_SO_REUSEADDR: 2062 optname = SO_REUSEADDR; 2063 break; 2064 #ifdef SO_REUSEPORT 2065 case TARGET_SO_REUSEPORT: 2066 optname = SO_REUSEPORT; 2067 break; 2068 #endif 2069 case TARGET_SO_TYPE: 2070 optname = SO_TYPE; 2071 break; 2072 case TARGET_SO_ERROR: 2073 optname = SO_ERROR; 2074 break; 2075 case TARGET_SO_DONTROUTE: 2076 optname = SO_DONTROUTE; 2077 break; 2078 case TARGET_SO_BROADCAST: 2079 optname = SO_BROADCAST; 2080 break; 2081 case TARGET_SO_SNDBUF: 2082 optname = SO_SNDBUF; 2083 break; 2084 case TARGET_SO_SNDBUFFORCE: 2085 optname = SO_SNDBUFFORCE; 2086 break; 2087 case TARGET_SO_RCVBUF: 2088 optname = SO_RCVBUF; 2089 break; 2090 case TARGET_SO_RCVBUFFORCE: 2091 optname = SO_RCVBUFFORCE; 2092 break; 2093 case TARGET_SO_KEEPALIVE: 2094 optname = SO_KEEPALIVE; 2095 break; 2096 case TARGET_SO_OOBINLINE: 2097 optname = SO_OOBINLINE; 2098 break; 2099 case TARGET_SO_NO_CHECK: 2100 optname = SO_NO_CHECK; 2101 break; 2102 case TARGET_SO_PRIORITY: 2103 optname = SO_PRIORITY; 2104 break; 2105 #ifdef SO_BSDCOMPAT 2106 case TARGET_SO_BSDCOMPAT: 2107 optname = SO_BSDCOMPAT; 2108 break; 2109 #endif 2110 case TARGET_SO_PASSCRED: 2111 optname = SO_PASSCRED; 2112 break; 2113 case TARGET_SO_PASSSEC: 2114 optname = SO_PASSSEC; 2115 break; 2116 case TARGET_SO_TIMESTAMP: 2117 optname = SO_TIMESTAMP; 2118 break; 2119 case TARGET_SO_RCVLOWAT: 2120 optname = SO_RCVLOWAT; 2121 break; 2122 default: 2123 goto unimplemented; 2124 } 2125 if (optlen < sizeof(uint32_t)) 2126 return -TARGET_EINVAL; 2127 2128 if (get_user_u32(val, optval_addr)) 2129 return -TARGET_EFAULT; 2130 ret = get_errno(setsockopt(sockfd, SOL_SOCKET, optname, &val, sizeof(val))); 2131 break; 2132 default: 2133 unimplemented: 2134 gemu_log("Unsupported setsockopt level=%d optname=%d\n", level, optname); 2135 ret = -TARGET_ENOPROTOOPT; 2136 } 2137 return ret; 2138 } 2139 2140 /* do_getsockopt() Must return target values and target errnos. */ 2141 static abi_long do_getsockopt(int sockfd, int level, int optname, 2142 abi_ulong optval_addr, abi_ulong optlen) 2143 { 2144 abi_long ret; 2145 int len, val; 2146 socklen_t lv; 2147 2148 switch(level) { 2149 case TARGET_SOL_SOCKET: 2150 level = SOL_SOCKET; 2151 switch (optname) { 2152 /* These don't just return a single integer */ 2153 case TARGET_SO_RCVTIMEO: 2154 case TARGET_SO_SNDTIMEO: 2155 case TARGET_SO_PEERNAME: 2156 goto unimplemented; 2157 case TARGET_SO_PEERCRED: { 2158 struct ucred cr; 2159 socklen_t crlen; 2160 struct target_ucred *tcr; 2161 2162 if (get_user_u32(len, optlen)) { 2163 return -TARGET_EFAULT; 2164 } 2165 if (len < 0) { 2166 return -TARGET_EINVAL; 2167 } 2168 2169 crlen = sizeof(cr); 2170 ret = get_errno(getsockopt(sockfd, level, SO_PEERCRED, 2171 &cr, &crlen)); 2172 if (ret < 0) { 2173 return ret; 2174 } 2175 if (len > crlen) { 2176 len = crlen; 2177 } 2178 if (!lock_user_struct(VERIFY_WRITE, tcr, optval_addr, 0)) { 2179 return -TARGET_EFAULT; 2180 } 2181 __put_user(cr.pid, &tcr->pid); 2182 __put_user(cr.uid, &tcr->uid); 2183 __put_user(cr.gid, &tcr->gid); 2184 unlock_user_struct(tcr, optval_addr, 1); 2185 if (put_user_u32(len, optlen)) { 2186 return -TARGET_EFAULT; 2187 } 2188 break; 2189 } 2190 case TARGET_SO_LINGER: 2191 { 2192 struct linger lg; 2193 socklen_t lglen; 2194 struct target_linger *tlg; 2195 2196 if (get_user_u32(len, optlen)) { 2197 return -TARGET_EFAULT; 2198 } 2199 if (len < 0) { 2200 return -TARGET_EINVAL; 2201 } 2202 2203 lglen = sizeof(lg); 2204 ret = get_errno(getsockopt(sockfd, level, SO_LINGER, 2205 &lg, &lglen)); 2206 if (ret < 0) { 2207 return ret; 2208 } 2209 if (len > lglen) { 2210 len = lglen; 2211 } 2212 if (!lock_user_struct(VERIFY_WRITE, tlg, optval_addr, 0)) { 2213 return -TARGET_EFAULT; 2214 } 2215 __put_user(lg.l_onoff, &tlg->l_onoff); 2216 __put_user(lg.l_linger, &tlg->l_linger); 2217 unlock_user_struct(tlg, optval_addr, 1); 2218 if (put_user_u32(len, optlen)) { 2219 return -TARGET_EFAULT; 2220 } 2221 break; 2222 } 2223 /* Options with 'int' argument. */ 2224 case TARGET_SO_DEBUG: 2225 optname = SO_DEBUG; 2226 goto int_case; 2227 case TARGET_SO_REUSEADDR: 2228 optname = SO_REUSEADDR; 2229 goto int_case; 2230 #ifdef SO_REUSEPORT 2231 case TARGET_SO_REUSEPORT: 2232 optname = SO_REUSEPORT; 2233 goto int_case; 2234 #endif 2235 case TARGET_SO_TYPE: 2236 optname = SO_TYPE; 2237 goto int_case; 2238 case TARGET_SO_ERROR: 2239 optname = SO_ERROR; 2240 goto int_case; 2241 case TARGET_SO_DONTROUTE: 2242 optname = SO_DONTROUTE; 2243 goto int_case; 2244 case TARGET_SO_BROADCAST: 2245 optname = SO_BROADCAST; 2246 goto int_case; 2247 case TARGET_SO_SNDBUF: 2248 optname = SO_SNDBUF; 2249 goto int_case; 2250 case TARGET_SO_RCVBUF: 2251 optname = SO_RCVBUF; 2252 goto int_case; 2253 case TARGET_SO_KEEPALIVE: 2254 optname = SO_KEEPALIVE; 2255 goto int_case; 2256 case TARGET_SO_OOBINLINE: 2257 optname = SO_OOBINLINE; 2258 goto int_case; 2259 case TARGET_SO_NO_CHECK: 2260 optname = SO_NO_CHECK; 2261 goto int_case; 2262 case TARGET_SO_PRIORITY: 2263 optname = SO_PRIORITY; 2264 goto int_case; 2265 #ifdef SO_BSDCOMPAT 2266 case TARGET_SO_BSDCOMPAT: 2267 optname = SO_BSDCOMPAT; 2268 goto int_case; 2269 #endif 2270 case TARGET_SO_PASSCRED: 2271 optname = SO_PASSCRED; 2272 goto int_case; 2273 case TARGET_SO_TIMESTAMP: 2274 optname = SO_TIMESTAMP; 2275 goto int_case; 2276 case TARGET_SO_RCVLOWAT: 2277 optname = SO_RCVLOWAT; 2278 goto int_case; 2279 case TARGET_SO_ACCEPTCONN: 2280 optname = SO_ACCEPTCONN; 2281 goto int_case; 2282 default: 2283 goto int_case; 2284 } 2285 break; 2286 case SOL_TCP: 2287 /* TCP options all take an 'int' value. */ 2288 int_case: 2289 if (get_user_u32(len, optlen)) 2290 return -TARGET_EFAULT; 2291 if (len < 0) 2292 return -TARGET_EINVAL; 2293 lv = sizeof(lv); 2294 ret = get_errno(getsockopt(sockfd, level, optname, &val, &lv)); 2295 if (ret < 0) 2296 return ret; 2297 if (optname == SO_TYPE) { 2298 val = host_to_target_sock_type(val); 2299 } 2300 if (len > lv) 2301 len = lv; 2302 if (len == 4) { 2303 if (put_user_u32(val, optval_addr)) 2304 return -TARGET_EFAULT; 2305 } else { 2306 if (put_user_u8(val, optval_addr)) 2307 return -TARGET_EFAULT; 2308 } 2309 if (put_user_u32(len, optlen)) 2310 return -TARGET_EFAULT; 2311 break; 2312 case SOL_IP: 2313 switch(optname) { 2314 case IP_TOS: 2315 case IP_TTL: 2316 case IP_HDRINCL: 2317 case IP_ROUTER_ALERT: 2318 case IP_RECVOPTS: 2319 case IP_RETOPTS: 2320 case IP_PKTINFO: 2321 case IP_MTU_DISCOVER: 2322 case IP_RECVERR: 2323 case IP_RECVTOS: 2324 #ifdef IP_FREEBIND 2325 case IP_FREEBIND: 2326 #endif 2327 case IP_MULTICAST_TTL: 2328 case IP_MULTICAST_LOOP: 2329 if (get_user_u32(len, optlen)) 2330 return -TARGET_EFAULT; 2331 if (len < 0) 2332 return -TARGET_EINVAL; 2333 lv = sizeof(lv); 2334 ret = get_errno(getsockopt(sockfd, level, optname, &val, &lv)); 2335 if (ret < 0) 2336 return ret; 2337 if (len < sizeof(int) && len > 0 && val >= 0 && val < 255) { 2338 len = 1; 2339 if (put_user_u32(len, optlen) 2340 || put_user_u8(val, optval_addr)) 2341 return -TARGET_EFAULT; 2342 } else { 2343 if (len > sizeof(int)) 2344 len = sizeof(int); 2345 if (put_user_u32(len, optlen) 2346 || put_user_u32(val, optval_addr)) 2347 return -TARGET_EFAULT; 2348 } 2349 break; 2350 default: 2351 ret = -TARGET_ENOPROTOOPT; 2352 break; 2353 } 2354 break; 2355 case SOL_IPV6: 2356 switch (optname) { 2357 case IPV6_MTU_DISCOVER: 2358 case IPV6_MTU: 2359 case IPV6_V6ONLY: 2360 case IPV6_RECVPKTINFO: 2361 case IPV6_UNICAST_HOPS: 2362 case IPV6_MULTICAST_HOPS: 2363 case IPV6_MULTICAST_LOOP: 2364 case IPV6_RECVERR: 2365 case IPV6_RECVHOPLIMIT: 2366 case IPV6_2292HOPLIMIT: 2367 case IPV6_CHECKSUM: 2368 if (get_user_u32(len, optlen)) 2369 return -TARGET_EFAULT; 2370 if (len < 0) 2371 return -TARGET_EINVAL; 2372 lv = sizeof(lv); 2373 ret = get_errno(getsockopt(sockfd, level, optname, &val, &lv)); 2374 if (ret < 0) 2375 return ret; 2376 if (len < sizeof(int) && len > 0 && val >= 0 && val < 255) { 2377 len = 1; 2378 if (put_user_u32(len, optlen) 2379 || put_user_u8(val, optval_addr)) 2380 return -TARGET_EFAULT; 2381 } else { 2382 if (len > sizeof(int)) 2383 len = sizeof(int); 2384 if (put_user_u32(len, optlen) 2385 || put_user_u32(val, optval_addr)) 2386 return -TARGET_EFAULT; 2387 } 2388 break; 2389 default: 2390 ret = -TARGET_ENOPROTOOPT; 2391 break; 2392 } 2393 break; 2394 default: 2395 unimplemented: 2396 gemu_log("getsockopt level=%d optname=%d not yet supported\n", 2397 level, optname); 2398 ret = -TARGET_EOPNOTSUPP; 2399 break; 2400 } 2401 return ret; 2402 } 2403 2404 /* Convert target low/high pair representing file offset into the host 2405 * low/high pair. This function doesn't handle offsets bigger than 64 bits 2406 * as the kernel doesn't handle them either. 2407 */ 2408 static void target_to_host_low_high(abi_ulong tlow, 2409 abi_ulong thigh, 2410 unsigned long *hlow, 2411 unsigned long *hhigh) 2412 { 2413 uint64_t off = tlow | 2414 ((unsigned long long)thigh << TARGET_LONG_BITS / 2) << 2415 TARGET_LONG_BITS / 2; 2416 2417 *hlow = off; 2418 *hhigh = (off >> HOST_LONG_BITS / 2) >> HOST_LONG_BITS / 2; 2419 } 2420 2421 static struct iovec *lock_iovec(int type, abi_ulong target_addr, 2422 abi_ulong count, int copy) 2423 { 2424 struct target_iovec *target_vec; 2425 struct iovec *vec; 2426 abi_ulong total_len, max_len; 2427 int i; 2428 int err = 0; 2429 bool bad_address = false; 2430 2431 if (count == 0) { 2432 errno = 0; 2433 return NULL; 2434 } 2435 if (count > IOV_MAX) { 2436 errno = EINVAL; 2437 return NULL; 2438 } 2439 2440 vec = g_try_new0(struct iovec, count); 2441 if (vec == NULL) { 2442 errno = ENOMEM; 2443 return NULL; 2444 } 2445 2446 target_vec = lock_user(VERIFY_READ, target_addr, 2447 count * sizeof(struct target_iovec), 1); 2448 if (target_vec == NULL) { 2449 err = EFAULT; 2450 goto fail2; 2451 } 2452 2453 /* ??? If host page size > target page size, this will result in a 2454 value larger than what we can actually support. */ 2455 max_len = 0x7fffffff & TARGET_PAGE_MASK; 2456 total_len = 0; 2457 2458 for (i = 0; i < count; i++) { 2459 abi_ulong base = tswapal(target_vec[i].iov_base); 2460 abi_long len = tswapal(target_vec[i].iov_len); 2461 2462 if (len < 0) { 2463 err = EINVAL; 2464 goto fail; 2465 } else if (len == 0) { 2466 /* Zero length pointer is ignored. */ 2467 vec[i].iov_base = 0; 2468 } else { 2469 vec[i].iov_base = lock_user(type, base, len, copy); 2470 /* If the first buffer pointer is bad, this is a fault. But 2471 * subsequent bad buffers will result in a partial write; this 2472 * is realized by filling the vector with null pointers and 2473 * zero lengths. */ 2474 if (!vec[i].iov_base) { 2475 if (i == 0) { 2476 err = EFAULT; 2477 goto fail; 2478 } else { 2479 bad_address = true; 2480 } 2481 } 2482 if (bad_address) { 2483 len = 0; 2484 } 2485 if (len > max_len - total_len) { 2486 len = max_len - total_len; 2487 } 2488 } 2489 vec[i].iov_len = len; 2490 total_len += len; 2491 } 2492 2493 unlock_user(target_vec, target_addr, 0); 2494 return vec; 2495 2496 fail: 2497 while (--i >= 0) { 2498 if (tswapal(target_vec[i].iov_len) > 0) { 2499 unlock_user(vec[i].iov_base, tswapal(target_vec[i].iov_base), 0); 2500 } 2501 } 2502 unlock_user(target_vec, target_addr, 0); 2503 fail2: 2504 g_free(vec); 2505 errno = err; 2506 return NULL; 2507 } 2508 2509 static void unlock_iovec(struct iovec *vec, abi_ulong target_addr, 2510 abi_ulong count, int copy) 2511 { 2512 struct target_iovec *target_vec; 2513 int i; 2514 2515 target_vec = lock_user(VERIFY_READ, target_addr, 2516 count * sizeof(struct target_iovec), 1); 2517 if (target_vec) { 2518 for (i = 0; i < count; i++) { 2519 abi_ulong base = tswapal(target_vec[i].iov_base); 2520 abi_long len = tswapal(target_vec[i].iov_len); 2521 if (len < 0) { 2522 break; 2523 } 2524 unlock_user(vec[i].iov_base, base, copy ? vec[i].iov_len : 0); 2525 } 2526 unlock_user(target_vec, target_addr, 0); 2527 } 2528 2529 g_free(vec); 2530 } 2531 2532 static inline int target_to_host_sock_type(int *type) 2533 { 2534 int host_type = 0; 2535 int target_type = *type; 2536 2537 switch (target_type & TARGET_SOCK_TYPE_MASK) { 2538 case TARGET_SOCK_DGRAM: 2539 host_type = SOCK_DGRAM; 2540 break; 2541 case TARGET_SOCK_STREAM: 2542 host_type = SOCK_STREAM; 2543 break; 2544 default: 2545 host_type = target_type & TARGET_SOCK_TYPE_MASK; 2546 break; 2547 } 2548 if (target_type & TARGET_SOCK_CLOEXEC) { 2549 #if defined(SOCK_CLOEXEC) 2550 host_type |= SOCK_CLOEXEC; 2551 #else 2552 return -TARGET_EINVAL; 2553 #endif 2554 } 2555 if (target_type & TARGET_SOCK_NONBLOCK) { 2556 #if defined(SOCK_NONBLOCK) 2557 host_type |= SOCK_NONBLOCK; 2558 #elif !defined(O_NONBLOCK) 2559 return -TARGET_EINVAL; 2560 #endif 2561 } 2562 *type = host_type; 2563 return 0; 2564 } 2565 2566 /* Try to emulate socket type flags after socket creation. */ 2567 static int sock_flags_fixup(int fd, int target_type) 2568 { 2569 #if !defined(SOCK_NONBLOCK) && defined(O_NONBLOCK) 2570 if (target_type & TARGET_SOCK_NONBLOCK) { 2571 int flags = fcntl(fd, F_GETFL); 2572 if (fcntl(fd, F_SETFL, O_NONBLOCK | flags) == -1) { 2573 close(fd); 2574 return -TARGET_EINVAL; 2575 } 2576 } 2577 #endif 2578 return fd; 2579 } 2580 2581 /* do_socket() Must return target values and target errnos. */ 2582 static abi_long do_socket(int domain, int type, int protocol) 2583 { 2584 int target_type = type; 2585 int ret; 2586 2587 ret = target_to_host_sock_type(&type); 2588 if (ret) { 2589 return ret; 2590 } 2591 2592 if (domain == PF_NETLINK && !( 2593 #ifdef CONFIG_RTNETLINK 2594 protocol == NETLINK_ROUTE || 2595 #endif 2596 protocol == NETLINK_KOBJECT_UEVENT || 2597 protocol == NETLINK_AUDIT)) { 2598 return -EPFNOSUPPORT; 2599 } 2600 2601 if (domain == AF_PACKET || 2602 (domain == AF_INET && type == SOCK_PACKET)) { 2603 protocol = tswap16(protocol); 2604 } 2605 2606 ret = get_errno(socket(domain, type, protocol)); 2607 if (ret >= 0) { 2608 ret = sock_flags_fixup(ret, target_type); 2609 if (type == SOCK_PACKET) { 2610 /* Manage an obsolete case : 2611 * if socket type is SOCK_PACKET, bind by name 2612 */ 2613 fd_trans_register(ret, &target_packet_trans); 2614 } else if (domain == PF_NETLINK) { 2615 switch (protocol) { 2616 #ifdef CONFIG_RTNETLINK 2617 case NETLINK_ROUTE: 2618 fd_trans_register(ret, &target_netlink_route_trans); 2619 break; 2620 #endif 2621 case NETLINK_KOBJECT_UEVENT: 2622 /* nothing to do: messages are strings */ 2623 break; 2624 case NETLINK_AUDIT: 2625 fd_trans_register(ret, &target_netlink_audit_trans); 2626 break; 2627 default: 2628 g_assert_not_reached(); 2629 } 2630 } 2631 } 2632 return ret; 2633 } 2634 2635 /* do_bind() Must return target values and target errnos. */ 2636 static abi_long do_bind(int sockfd, abi_ulong target_addr, 2637 socklen_t addrlen) 2638 { 2639 void *addr; 2640 abi_long ret; 2641 2642 if ((int)addrlen < 0) { 2643 return -TARGET_EINVAL; 2644 } 2645 2646 addr = alloca(addrlen+1); 2647 2648 ret = target_to_host_sockaddr(sockfd, addr, target_addr, addrlen); 2649 if (ret) 2650 return ret; 2651 2652 return get_errno(bind(sockfd, addr, addrlen)); 2653 } 2654 2655 /* do_connect() Must return target values and target errnos. */ 2656 static abi_long do_connect(int sockfd, abi_ulong target_addr, 2657 socklen_t addrlen) 2658 { 2659 void *addr; 2660 abi_long ret; 2661 2662 if ((int)addrlen < 0) { 2663 return -TARGET_EINVAL; 2664 } 2665 2666 addr = alloca(addrlen+1); 2667 2668 ret = target_to_host_sockaddr(sockfd, addr, target_addr, addrlen); 2669 if (ret) 2670 return ret; 2671 2672 return get_errno(safe_connect(sockfd, addr, addrlen)); 2673 } 2674 2675 /* do_sendrecvmsg_locked() Must return target values and target errnos. */ 2676 static abi_long do_sendrecvmsg_locked(int fd, struct target_msghdr *msgp, 2677 int flags, int send) 2678 { 2679 abi_long ret, len; 2680 struct msghdr msg; 2681 abi_ulong count; 2682 struct iovec *vec; 2683 abi_ulong target_vec; 2684 2685 if (msgp->msg_name) { 2686 msg.msg_namelen = tswap32(msgp->msg_namelen); 2687 msg.msg_name = alloca(msg.msg_namelen+1); 2688 ret = target_to_host_sockaddr(fd, msg.msg_name, 2689 tswapal(msgp->msg_name), 2690 msg.msg_namelen); 2691 if (ret == -TARGET_EFAULT) { 2692 /* For connected sockets msg_name and msg_namelen must 2693 * be ignored, so returning EFAULT immediately is wrong. 2694 * Instead, pass a bad msg_name to the host kernel, and 2695 * let it decide whether to return EFAULT or not. 2696 */ 2697 msg.msg_name = (void *)-1; 2698 } else if (ret) { 2699 goto out2; 2700 } 2701 } else { 2702 msg.msg_name = NULL; 2703 msg.msg_namelen = 0; 2704 } 2705 msg.msg_controllen = 2 * tswapal(msgp->msg_controllen); 2706 msg.msg_control = alloca(msg.msg_controllen); 2707 memset(msg.msg_control, 0, msg.msg_controllen); 2708 2709 msg.msg_flags = tswap32(msgp->msg_flags); 2710 2711 count = tswapal(msgp->msg_iovlen); 2712 target_vec = tswapal(msgp->msg_iov); 2713 2714 if (count > IOV_MAX) { 2715 /* sendrcvmsg returns a different errno for this condition than 2716 * readv/writev, so we must catch it here before lock_iovec() does. 2717 */ 2718 ret = -TARGET_EMSGSIZE; 2719 goto out2; 2720 } 2721 2722 vec = lock_iovec(send ? VERIFY_READ : VERIFY_WRITE, 2723 target_vec, count, send); 2724 if (vec == NULL) { 2725 ret = -host_to_target_errno(errno); 2726 goto out2; 2727 } 2728 msg.msg_iovlen = count; 2729 msg.msg_iov = vec; 2730 2731 if (send) { 2732 if (fd_trans_target_to_host_data(fd)) { 2733 void *host_msg; 2734 2735 host_msg = g_malloc(msg.msg_iov->iov_len); 2736 memcpy(host_msg, msg.msg_iov->iov_base, msg.msg_iov->iov_len); 2737 ret = fd_trans_target_to_host_data(fd)(host_msg, 2738 msg.msg_iov->iov_len); 2739 if (ret >= 0) { 2740 msg.msg_iov->iov_base = host_msg; 2741 ret = get_errno(safe_sendmsg(fd, &msg, flags)); 2742 } 2743 g_free(host_msg); 2744 } else { 2745 ret = target_to_host_cmsg(&msg, msgp); 2746 if (ret == 0) { 2747 ret = get_errno(safe_sendmsg(fd, &msg, flags)); 2748 } 2749 } 2750 } else { 2751 ret = get_errno(safe_recvmsg(fd, &msg, flags)); 2752 if (!is_error(ret)) { 2753 len = ret; 2754 if (fd_trans_host_to_target_data(fd)) { 2755 ret = fd_trans_host_to_target_data(fd)(msg.msg_iov->iov_base, 2756 MIN(msg.msg_iov->iov_len, len)); 2757 } else { 2758 ret = host_to_target_cmsg(msgp, &msg); 2759 } 2760 if (!is_error(ret)) { 2761 msgp->msg_namelen = tswap32(msg.msg_namelen); 2762 msgp->msg_flags = tswap32(msg.msg_flags); 2763 if (msg.msg_name != NULL && msg.msg_name != (void *)-1) { 2764 ret = host_to_target_sockaddr(tswapal(msgp->msg_name), 2765 msg.msg_name, msg.msg_namelen); 2766 if (ret) { 2767 goto out; 2768 } 2769 } 2770 2771 ret = len; 2772 } 2773 } 2774 } 2775 2776 out: 2777 unlock_iovec(vec, target_vec, count, !send); 2778 out2: 2779 return ret; 2780 } 2781 2782 static abi_long do_sendrecvmsg(int fd, abi_ulong target_msg, 2783 int flags, int send) 2784 { 2785 abi_long ret; 2786 struct target_msghdr *msgp; 2787 2788 if (!lock_user_struct(send ? VERIFY_READ : VERIFY_WRITE, 2789 msgp, 2790 target_msg, 2791 send ? 1 : 0)) { 2792 return -TARGET_EFAULT; 2793 } 2794 ret = do_sendrecvmsg_locked(fd, msgp, flags, send); 2795 unlock_user_struct(msgp, target_msg, send ? 0 : 1); 2796 return ret; 2797 } 2798 2799 /* We don't rely on the C library to have sendmmsg/recvmmsg support, 2800 * so it might not have this *mmsg-specific flag either. 2801 */ 2802 #ifndef MSG_WAITFORONE 2803 #define MSG_WAITFORONE 0x10000 2804 #endif 2805 2806 static abi_long do_sendrecvmmsg(int fd, abi_ulong target_msgvec, 2807 unsigned int vlen, unsigned int flags, 2808 int send) 2809 { 2810 struct target_mmsghdr *mmsgp; 2811 abi_long ret = 0; 2812 int i; 2813 2814 if (vlen > UIO_MAXIOV) { 2815 vlen = UIO_MAXIOV; 2816 } 2817 2818 mmsgp = lock_user(VERIFY_WRITE, target_msgvec, sizeof(*mmsgp) * vlen, 1); 2819 if (!mmsgp) { 2820 return -TARGET_EFAULT; 2821 } 2822 2823 for (i = 0; i < vlen; i++) { 2824 ret = do_sendrecvmsg_locked(fd, &mmsgp[i].msg_hdr, flags, send); 2825 if (is_error(ret)) { 2826 break; 2827 } 2828 mmsgp[i].msg_len = tswap32(ret); 2829 /* MSG_WAITFORONE turns on MSG_DONTWAIT after one packet */ 2830 if (flags & MSG_WAITFORONE) { 2831 flags |= MSG_DONTWAIT; 2832 } 2833 } 2834 2835 unlock_user(mmsgp, target_msgvec, sizeof(*mmsgp) * i); 2836 2837 /* Return number of datagrams sent if we sent any at all; 2838 * otherwise return the error. 2839 */ 2840 if (i) { 2841 return i; 2842 } 2843 return ret; 2844 } 2845 2846 /* do_accept4() Must return target values and target errnos. */ 2847 static abi_long do_accept4(int fd, abi_ulong target_addr, 2848 abi_ulong target_addrlen_addr, int flags) 2849 { 2850 socklen_t addrlen, ret_addrlen; 2851 void *addr; 2852 abi_long ret; 2853 int host_flags; 2854 2855 host_flags = target_to_host_bitmask(flags, fcntl_flags_tbl); 2856 2857 if (target_addr == 0) { 2858 return get_errno(safe_accept4(fd, NULL, NULL, host_flags)); 2859 } 2860 2861 /* linux returns EINVAL if addrlen pointer is invalid */ 2862 if (get_user_u32(addrlen, target_addrlen_addr)) 2863 return -TARGET_EINVAL; 2864 2865 if ((int)addrlen < 0) { 2866 return -TARGET_EINVAL; 2867 } 2868 2869 if (!access_ok(VERIFY_WRITE, target_addr, addrlen)) 2870 return -TARGET_EINVAL; 2871 2872 addr = alloca(addrlen); 2873 2874 ret_addrlen = addrlen; 2875 ret = get_errno(safe_accept4(fd, addr, &ret_addrlen, host_flags)); 2876 if (!is_error(ret)) { 2877 host_to_target_sockaddr(target_addr, addr, MIN(addrlen, ret_addrlen)); 2878 if (put_user_u32(ret_addrlen, target_addrlen_addr)) { 2879 ret = -TARGET_EFAULT; 2880 } 2881 } 2882 return ret; 2883 } 2884 2885 /* do_getpeername() Must return target values and target errnos. */ 2886 static abi_long do_getpeername(int fd, abi_ulong target_addr, 2887 abi_ulong target_addrlen_addr) 2888 { 2889 socklen_t addrlen, ret_addrlen; 2890 void *addr; 2891 abi_long ret; 2892 2893 if (get_user_u32(addrlen, target_addrlen_addr)) 2894 return -TARGET_EFAULT; 2895 2896 if ((int)addrlen < 0) { 2897 return -TARGET_EINVAL; 2898 } 2899 2900 if (!access_ok(VERIFY_WRITE, target_addr, addrlen)) 2901 return -TARGET_EFAULT; 2902 2903 addr = alloca(addrlen); 2904 2905 ret_addrlen = addrlen; 2906 ret = get_errno(getpeername(fd, addr, &ret_addrlen)); 2907 if (!is_error(ret)) { 2908 host_to_target_sockaddr(target_addr, addr, MIN(addrlen, ret_addrlen)); 2909 if (put_user_u32(ret_addrlen, target_addrlen_addr)) { 2910 ret = -TARGET_EFAULT; 2911 } 2912 } 2913 return ret; 2914 } 2915 2916 /* do_getsockname() Must return target values and target errnos. */ 2917 static abi_long do_getsockname(int fd, abi_ulong target_addr, 2918 abi_ulong target_addrlen_addr) 2919 { 2920 socklen_t addrlen, ret_addrlen; 2921 void *addr; 2922 abi_long ret; 2923 2924 if (get_user_u32(addrlen, target_addrlen_addr)) 2925 return -TARGET_EFAULT; 2926 2927 if ((int)addrlen < 0) { 2928 return -TARGET_EINVAL; 2929 } 2930 2931 if (!access_ok(VERIFY_WRITE, target_addr, addrlen)) 2932 return -TARGET_EFAULT; 2933 2934 addr = alloca(addrlen); 2935 2936 ret_addrlen = addrlen; 2937 ret = get_errno(getsockname(fd, addr, &ret_addrlen)); 2938 if (!is_error(ret)) { 2939 host_to_target_sockaddr(target_addr, addr, MIN(addrlen, ret_addrlen)); 2940 if (put_user_u32(ret_addrlen, target_addrlen_addr)) { 2941 ret = -TARGET_EFAULT; 2942 } 2943 } 2944 return ret; 2945 } 2946 2947 /* do_socketpair() Must return target values and target errnos. */ 2948 static abi_long do_socketpair(int domain, int type, int protocol, 2949 abi_ulong target_tab_addr) 2950 { 2951 int tab[2]; 2952 abi_long ret; 2953 2954 target_to_host_sock_type(&type); 2955 2956 ret = get_errno(socketpair(domain, type, protocol, tab)); 2957 if (!is_error(ret)) { 2958 if (put_user_s32(tab[0], target_tab_addr) 2959 || put_user_s32(tab[1], target_tab_addr + sizeof(tab[0]))) 2960 ret = -TARGET_EFAULT; 2961 } 2962 return ret; 2963 } 2964 2965 /* do_sendto() Must return target values and target errnos. */ 2966 static abi_long do_sendto(int fd, abi_ulong msg, size_t len, int flags, 2967 abi_ulong target_addr, socklen_t addrlen) 2968 { 2969 void *addr; 2970 void *host_msg; 2971 void *copy_msg = NULL; 2972 abi_long ret; 2973 2974 if ((int)addrlen < 0) { 2975 return -TARGET_EINVAL; 2976 } 2977 2978 host_msg = lock_user(VERIFY_READ, msg, len, 1); 2979 if (!host_msg) 2980 return -TARGET_EFAULT; 2981 if (fd_trans_target_to_host_data(fd)) { 2982 copy_msg = host_msg; 2983 host_msg = g_malloc(len); 2984 memcpy(host_msg, copy_msg, len); 2985 ret = fd_trans_target_to_host_data(fd)(host_msg, len); 2986 if (ret < 0) { 2987 goto fail; 2988 } 2989 } 2990 if (target_addr) { 2991 addr = alloca(addrlen+1); 2992 ret = target_to_host_sockaddr(fd, addr, target_addr, addrlen); 2993 if (ret) { 2994 goto fail; 2995 } 2996 ret = get_errno(safe_sendto(fd, host_msg, len, flags, addr, addrlen)); 2997 } else { 2998 ret = get_errno(safe_sendto(fd, host_msg, len, flags, NULL, 0)); 2999 } 3000 fail: 3001 if (copy_msg) { 3002 g_free(host_msg); 3003 host_msg = copy_msg; 3004 } 3005 unlock_user(host_msg, msg, 0); 3006 return ret; 3007 } 3008 3009 /* do_recvfrom() Must return target values and target errnos. */ 3010 static abi_long do_recvfrom(int fd, abi_ulong msg, size_t len, int flags, 3011 abi_ulong target_addr, 3012 abi_ulong target_addrlen) 3013 { 3014 socklen_t addrlen, ret_addrlen; 3015 void *addr; 3016 void *host_msg; 3017 abi_long ret; 3018 3019 host_msg = lock_user(VERIFY_WRITE, msg, len, 0); 3020 if (!host_msg) 3021 return -TARGET_EFAULT; 3022 if (target_addr) { 3023 if (get_user_u32(addrlen, target_addrlen)) { 3024 ret = -TARGET_EFAULT; 3025 goto fail; 3026 } 3027 if ((int)addrlen < 0) { 3028 ret = -TARGET_EINVAL; 3029 goto fail; 3030 } 3031 addr = alloca(addrlen); 3032 ret_addrlen = addrlen; 3033 ret = get_errno(safe_recvfrom(fd, host_msg, len, flags, 3034 addr, &ret_addrlen)); 3035 } else { 3036 addr = NULL; /* To keep compiler quiet. */ 3037 addrlen = 0; /* To keep compiler quiet. */ 3038 ret = get_errno(safe_recvfrom(fd, host_msg, len, flags, NULL, 0)); 3039 } 3040 if (!is_error(ret)) { 3041 if (fd_trans_host_to_target_data(fd)) { 3042 abi_long trans; 3043 trans = fd_trans_host_to_target_data(fd)(host_msg, MIN(ret, len)); 3044 if (is_error(trans)) { 3045 ret = trans; 3046 goto fail; 3047 } 3048 } 3049 if (target_addr) { 3050 host_to_target_sockaddr(target_addr, addr, 3051 MIN(addrlen, ret_addrlen)); 3052 if (put_user_u32(ret_addrlen, target_addrlen)) { 3053 ret = -TARGET_EFAULT; 3054 goto fail; 3055 } 3056 } 3057 unlock_user(host_msg, msg, len); 3058 } else { 3059 fail: 3060 unlock_user(host_msg, msg, 0); 3061 } 3062 return ret; 3063 } 3064 3065 #ifdef TARGET_NR_socketcall 3066 /* do_socketcall() must return target values and target errnos. */ 3067 static abi_long do_socketcall(int num, abi_ulong vptr) 3068 { 3069 static const unsigned nargs[] = { /* number of arguments per operation */ 3070 [TARGET_SYS_SOCKET] = 3, /* domain, type, protocol */ 3071 [TARGET_SYS_BIND] = 3, /* fd, addr, addrlen */ 3072 [TARGET_SYS_CONNECT] = 3, /* fd, addr, addrlen */ 3073 [TARGET_SYS_LISTEN] = 2, /* fd, backlog */ 3074 [TARGET_SYS_ACCEPT] = 3, /* fd, addr, addrlen */ 3075 [TARGET_SYS_GETSOCKNAME] = 3, /* fd, addr, addrlen */ 3076 [TARGET_SYS_GETPEERNAME] = 3, /* fd, addr, addrlen */ 3077 [TARGET_SYS_SOCKETPAIR] = 4, /* domain, type, protocol, tab */ 3078 [TARGET_SYS_SEND] = 4, /* fd, msg, len, flags */ 3079 [TARGET_SYS_RECV] = 4, /* fd, msg, len, flags */ 3080 [TARGET_SYS_SENDTO] = 6, /* fd, msg, len, flags, addr, addrlen */ 3081 [TARGET_SYS_RECVFROM] = 6, /* fd, msg, len, flags, addr, addrlen */ 3082 [TARGET_SYS_SHUTDOWN] = 2, /* fd, how */ 3083 [TARGET_SYS_SETSOCKOPT] = 5, /* fd, level, optname, optval, optlen */ 3084 [TARGET_SYS_GETSOCKOPT] = 5, /* fd, level, optname, optval, optlen */ 3085 [TARGET_SYS_SENDMSG] = 3, /* fd, msg, flags */ 3086 [TARGET_SYS_RECVMSG] = 3, /* fd, msg, flags */ 3087 [TARGET_SYS_ACCEPT4] = 4, /* fd, addr, addrlen, flags */ 3088 [TARGET_SYS_RECVMMSG] = 4, /* fd, msgvec, vlen, flags */ 3089 [TARGET_SYS_SENDMMSG] = 4, /* fd, msgvec, vlen, flags */ 3090 }; 3091 abi_long a[6]; /* max 6 args */ 3092 unsigned i; 3093 3094 /* check the range of the first argument num */ 3095 /* (TARGET_SYS_SENDMMSG is the highest among TARGET_SYS_xxx) */ 3096 if (num < 1 || num > TARGET_SYS_SENDMMSG) { 3097 return -TARGET_EINVAL; 3098 } 3099 /* ensure we have space for args */ 3100 if (nargs[num] > ARRAY_SIZE(a)) { 3101 return -TARGET_EINVAL; 3102 } 3103 /* collect the arguments in a[] according to nargs[] */ 3104 for (i = 0; i < nargs[num]; ++i) { 3105 if (get_user_ual(a[i], vptr + i * sizeof(abi_long)) != 0) { 3106 return -TARGET_EFAULT; 3107 } 3108 } 3109 /* now when we have the args, invoke the appropriate underlying function */ 3110 switch (num) { 3111 case TARGET_SYS_SOCKET: /* domain, type, protocol */ 3112 return do_socket(a[0], a[1], a[2]); 3113 case TARGET_SYS_BIND: /* sockfd, addr, addrlen */ 3114 return do_bind(a[0], a[1], a[2]); 3115 case TARGET_SYS_CONNECT: /* sockfd, addr, addrlen */ 3116 return do_connect(a[0], a[1], a[2]); 3117 case TARGET_SYS_LISTEN: /* sockfd, backlog */ 3118 return get_errno(listen(a[0], a[1])); 3119 case TARGET_SYS_ACCEPT: /* sockfd, addr, addrlen */ 3120 return do_accept4(a[0], a[1], a[2], 0); 3121 case TARGET_SYS_GETSOCKNAME: /* sockfd, addr, addrlen */ 3122 return do_getsockname(a[0], a[1], a[2]); 3123 case TARGET_SYS_GETPEERNAME: /* sockfd, addr, addrlen */ 3124 return do_getpeername(a[0], a[1], a[2]); 3125 case TARGET_SYS_SOCKETPAIR: /* domain, type, protocol, tab */ 3126 return do_socketpair(a[0], a[1], a[2], a[3]); 3127 case TARGET_SYS_SEND: /* sockfd, msg, len, flags */ 3128 return do_sendto(a[0], a[1], a[2], a[3], 0, 0); 3129 case TARGET_SYS_RECV: /* sockfd, msg, len, flags */ 3130 return do_recvfrom(a[0], a[1], a[2], a[3], 0, 0); 3131 case TARGET_SYS_SENDTO: /* sockfd, msg, len, flags, addr, addrlen */ 3132 return do_sendto(a[0], a[1], a[2], a[3], a[4], a[5]); 3133 case TARGET_SYS_RECVFROM: /* sockfd, msg, len, flags, addr, addrlen */ 3134 return do_recvfrom(a[0], a[1], a[2], a[3], a[4], a[5]); 3135 case TARGET_SYS_SHUTDOWN: /* sockfd, how */ 3136 return get_errno(shutdown(a[0], a[1])); 3137 case TARGET_SYS_SETSOCKOPT: /* sockfd, level, optname, optval, optlen */ 3138 return do_setsockopt(a[0], a[1], a[2], a[3], a[4]); 3139 case TARGET_SYS_GETSOCKOPT: /* sockfd, level, optname, optval, optlen */ 3140 return do_getsockopt(a[0], a[1], a[2], a[3], a[4]); 3141 case TARGET_SYS_SENDMSG: /* sockfd, msg, flags */ 3142 return do_sendrecvmsg(a[0], a[1], a[2], 1); 3143 case TARGET_SYS_RECVMSG: /* sockfd, msg, flags */ 3144 return do_sendrecvmsg(a[0], a[1], a[2], 0); 3145 case TARGET_SYS_ACCEPT4: /* sockfd, addr, addrlen, flags */ 3146 return do_accept4(a[0], a[1], a[2], a[3]); 3147 case TARGET_SYS_RECVMMSG: /* sockfd, msgvec, vlen, flags */ 3148 return do_sendrecvmmsg(a[0], a[1], a[2], a[3], 0); 3149 case TARGET_SYS_SENDMMSG: /* sockfd, msgvec, vlen, flags */ 3150 return do_sendrecvmmsg(a[0], a[1], a[2], a[3], 1); 3151 default: 3152 gemu_log("Unsupported socketcall: %d\n", num); 3153 return -TARGET_EINVAL; 3154 } 3155 } 3156 #endif 3157 3158 #define N_SHM_REGIONS 32 3159 3160 static struct shm_region { 3161 abi_ulong start; 3162 abi_ulong size; 3163 bool in_use; 3164 } shm_regions[N_SHM_REGIONS]; 3165 3166 #ifndef TARGET_SEMID64_DS 3167 /* asm-generic version of this struct */ 3168 struct target_semid64_ds 3169 { 3170 struct target_ipc_perm sem_perm; 3171 abi_ulong sem_otime; 3172 #if TARGET_ABI_BITS == 32 3173 abi_ulong __unused1; 3174 #endif 3175 abi_ulong sem_ctime; 3176 #if TARGET_ABI_BITS == 32 3177 abi_ulong __unused2; 3178 #endif 3179 abi_ulong sem_nsems; 3180 abi_ulong __unused3; 3181 abi_ulong __unused4; 3182 }; 3183 #endif 3184 3185 static inline abi_long target_to_host_ipc_perm(struct ipc_perm *host_ip, 3186 abi_ulong target_addr) 3187 { 3188 struct target_ipc_perm *target_ip; 3189 struct target_semid64_ds *target_sd; 3190 3191 if (!lock_user_struct(VERIFY_READ, target_sd, target_addr, 1)) 3192 return -TARGET_EFAULT; 3193 target_ip = &(target_sd->sem_perm); 3194 host_ip->__key = tswap32(target_ip->__key); 3195 host_ip->uid = tswap32(target_ip->uid); 3196 host_ip->gid = tswap32(target_ip->gid); 3197 host_ip->cuid = tswap32(target_ip->cuid); 3198 host_ip->cgid = tswap32(target_ip->cgid); 3199 #if defined(TARGET_ALPHA) || defined(TARGET_MIPS) || defined(TARGET_PPC) 3200 host_ip->mode = tswap32(target_ip->mode); 3201 #else 3202 host_ip->mode = tswap16(target_ip->mode); 3203 #endif 3204 #if defined(TARGET_PPC) 3205 host_ip->__seq = tswap32(target_ip->__seq); 3206 #else 3207 host_ip->__seq = tswap16(target_ip->__seq); 3208 #endif 3209 unlock_user_struct(target_sd, target_addr, 0); 3210 return 0; 3211 } 3212 3213 static inline abi_long host_to_target_ipc_perm(abi_ulong target_addr, 3214 struct ipc_perm *host_ip) 3215 { 3216 struct target_ipc_perm *target_ip; 3217 struct target_semid64_ds *target_sd; 3218 3219 if (!lock_user_struct(VERIFY_WRITE, target_sd, target_addr, 0)) 3220 return -TARGET_EFAULT; 3221 target_ip = &(target_sd->sem_perm); 3222 target_ip->__key = tswap32(host_ip->__key); 3223 target_ip->uid = tswap32(host_ip->uid); 3224 target_ip->gid = tswap32(host_ip->gid); 3225 target_ip->cuid = tswap32(host_ip->cuid); 3226 target_ip->cgid = tswap32(host_ip->cgid); 3227 #if defined(TARGET_ALPHA) || defined(TARGET_MIPS) || defined(TARGET_PPC) 3228 target_ip->mode = tswap32(host_ip->mode); 3229 #else 3230 target_ip->mode = tswap16(host_ip->mode); 3231 #endif 3232 #if defined(TARGET_PPC) 3233 target_ip->__seq = tswap32(host_ip->__seq); 3234 #else 3235 target_ip->__seq = tswap16(host_ip->__seq); 3236 #endif 3237 unlock_user_struct(target_sd, target_addr, 1); 3238 return 0; 3239 } 3240 3241 static inline abi_long target_to_host_semid_ds(struct semid_ds *host_sd, 3242 abi_ulong target_addr) 3243 { 3244 struct target_semid64_ds *target_sd; 3245 3246 if (!lock_user_struct(VERIFY_READ, target_sd, target_addr, 1)) 3247 return -TARGET_EFAULT; 3248 if (target_to_host_ipc_perm(&(host_sd->sem_perm),target_addr)) 3249 return -TARGET_EFAULT; 3250 host_sd->sem_nsems = tswapal(target_sd->sem_nsems); 3251 host_sd->sem_otime = tswapal(target_sd->sem_otime); 3252 host_sd->sem_ctime = tswapal(target_sd->sem_ctime); 3253 unlock_user_struct(target_sd, target_addr, 0); 3254 return 0; 3255 } 3256 3257 static inline abi_long host_to_target_semid_ds(abi_ulong target_addr, 3258 struct semid_ds *host_sd) 3259 { 3260 struct target_semid64_ds *target_sd; 3261 3262 if (!lock_user_struct(VERIFY_WRITE, target_sd, target_addr, 0)) 3263 return -TARGET_EFAULT; 3264 if (host_to_target_ipc_perm(target_addr,&(host_sd->sem_perm))) 3265 return -TARGET_EFAULT; 3266 target_sd->sem_nsems = tswapal(host_sd->sem_nsems); 3267 target_sd->sem_otime = tswapal(host_sd->sem_otime); 3268 target_sd->sem_ctime = tswapal(host_sd->sem_ctime); 3269 unlock_user_struct(target_sd, target_addr, 1); 3270 return 0; 3271 } 3272 3273 struct target_seminfo { 3274 int semmap; 3275 int semmni; 3276 int semmns; 3277 int semmnu; 3278 int semmsl; 3279 int semopm; 3280 int semume; 3281 int semusz; 3282 int semvmx; 3283 int semaem; 3284 }; 3285 3286 static inline abi_long host_to_target_seminfo(abi_ulong target_addr, 3287 struct seminfo *host_seminfo) 3288 { 3289 struct target_seminfo *target_seminfo; 3290 if (!lock_user_struct(VERIFY_WRITE, target_seminfo, target_addr, 0)) 3291 return -TARGET_EFAULT; 3292 __put_user(host_seminfo->semmap, &target_seminfo->semmap); 3293 __put_user(host_seminfo->semmni, &target_seminfo->semmni); 3294 __put_user(host_seminfo->semmns, &target_seminfo->semmns); 3295 __put_user(host_seminfo->semmnu, &target_seminfo->semmnu); 3296 __put_user(host_seminfo->semmsl, &target_seminfo->semmsl); 3297 __put_user(host_seminfo->semopm, &target_seminfo->semopm); 3298 __put_user(host_seminfo->semume, &target_seminfo->semume); 3299 __put_user(host_seminfo->semusz, &target_seminfo->semusz); 3300 __put_user(host_seminfo->semvmx, &target_seminfo->semvmx); 3301 __put_user(host_seminfo->semaem, &target_seminfo->semaem); 3302 unlock_user_struct(target_seminfo, target_addr, 1); 3303 return 0; 3304 } 3305 3306 union semun { 3307 int val; 3308 struct semid_ds *buf; 3309 unsigned short *array; 3310 struct seminfo *__buf; 3311 }; 3312 3313 union target_semun { 3314 int val; 3315 abi_ulong buf; 3316 abi_ulong array; 3317 abi_ulong __buf; 3318 }; 3319 3320 static inline abi_long target_to_host_semarray(int semid, unsigned short **host_array, 3321 abi_ulong target_addr) 3322 { 3323 int nsems; 3324 unsigned short *array; 3325 union semun semun; 3326 struct semid_ds semid_ds; 3327 int i, ret; 3328 3329 semun.buf = &semid_ds; 3330 3331 ret = semctl(semid, 0, IPC_STAT, semun); 3332 if (ret == -1) 3333 return get_errno(ret); 3334 3335 nsems = semid_ds.sem_nsems; 3336 3337 *host_array = g_try_new(unsigned short, nsems); 3338 if (!*host_array) { 3339 return -TARGET_ENOMEM; 3340 } 3341 array = lock_user(VERIFY_READ, target_addr, 3342 nsems*sizeof(unsigned short), 1); 3343 if (!array) { 3344 g_free(*host_array); 3345 return -TARGET_EFAULT; 3346 } 3347 3348 for(i=0; i<nsems; i++) { 3349 __get_user((*host_array)[i], &array[i]); 3350 } 3351 unlock_user(array, target_addr, 0); 3352 3353 return 0; 3354 } 3355 3356 static inline abi_long host_to_target_semarray(int semid, abi_ulong target_addr, 3357 unsigned short **host_array) 3358 { 3359 int nsems; 3360 unsigned short *array; 3361 union semun semun; 3362 struct semid_ds semid_ds; 3363 int i, ret; 3364 3365 semun.buf = &semid_ds; 3366 3367 ret = semctl(semid, 0, IPC_STAT, semun); 3368 if (ret == -1) 3369 return get_errno(ret); 3370 3371 nsems = semid_ds.sem_nsems; 3372 3373 array = lock_user(VERIFY_WRITE, target_addr, 3374 nsems*sizeof(unsigned short), 0); 3375 if (!array) 3376 return -TARGET_EFAULT; 3377 3378 for(i=0; i<nsems; i++) { 3379 __put_user((*host_array)[i], &array[i]); 3380 } 3381 g_free(*host_array); 3382 unlock_user(array, target_addr, 1); 3383 3384 return 0; 3385 } 3386 3387 static inline abi_long do_semctl(int semid, int semnum, int cmd, 3388 abi_ulong target_arg) 3389 { 3390 union target_semun target_su = { .buf = target_arg }; 3391 union semun arg; 3392 struct semid_ds dsarg; 3393 unsigned short *array = NULL; 3394 struct seminfo seminfo; 3395 abi_long ret = -TARGET_EINVAL; 3396 abi_long err; 3397 cmd &= 0xff; 3398 3399 switch( cmd ) { 3400 case GETVAL: 3401 case SETVAL: 3402 /* In 64 bit cross-endian situations, we will erroneously pick up 3403 * the wrong half of the union for the "val" element. To rectify 3404 * this, the entire 8-byte structure is byteswapped, followed by 3405 * a swap of the 4 byte val field. In other cases, the data is 3406 * already in proper host byte order. */ 3407 if (sizeof(target_su.val) != (sizeof(target_su.buf))) { 3408 target_su.buf = tswapal(target_su.buf); 3409 arg.val = tswap32(target_su.val); 3410 } else { 3411 arg.val = target_su.val; 3412 } 3413 ret = get_errno(semctl(semid, semnum, cmd, arg)); 3414 break; 3415 case GETALL: 3416 case SETALL: 3417 err = target_to_host_semarray(semid, &array, target_su.array); 3418 if (err) 3419 return err; 3420 arg.array = array; 3421 ret = get_errno(semctl(semid, semnum, cmd, arg)); 3422 err = host_to_target_semarray(semid, target_su.array, &array); 3423 if (err) 3424 return err; 3425 break; 3426 case IPC_STAT: 3427 case IPC_SET: 3428 case SEM_STAT: 3429 err = target_to_host_semid_ds(&dsarg, target_su.buf); 3430 if (err) 3431 return err; 3432 arg.buf = &dsarg; 3433 ret = get_errno(semctl(semid, semnum, cmd, arg)); 3434 err = host_to_target_semid_ds(target_su.buf, &dsarg); 3435 if (err) 3436 return err; 3437 break; 3438 case IPC_INFO: 3439 case SEM_INFO: 3440 arg.__buf = &seminfo; 3441 ret = get_errno(semctl(semid, semnum, cmd, arg)); 3442 err = host_to_target_seminfo(target_su.__buf, &seminfo); 3443 if (err) 3444 return err; 3445 break; 3446 case IPC_RMID: 3447 case GETPID: 3448 case GETNCNT: 3449 case GETZCNT: 3450 ret = get_errno(semctl(semid, semnum, cmd, NULL)); 3451 break; 3452 } 3453 3454 return ret; 3455 } 3456 3457 struct target_sembuf { 3458 unsigned short sem_num; 3459 short sem_op; 3460 short sem_flg; 3461 }; 3462 3463 static inline abi_long target_to_host_sembuf(struct sembuf *host_sembuf, 3464 abi_ulong target_addr, 3465 unsigned nsops) 3466 { 3467 struct target_sembuf *target_sembuf; 3468 int i; 3469 3470 target_sembuf = lock_user(VERIFY_READ, target_addr, 3471 nsops*sizeof(struct target_sembuf), 1); 3472 if (!target_sembuf) 3473 return -TARGET_EFAULT; 3474 3475 for(i=0; i<nsops; i++) { 3476 __get_user(host_sembuf[i].sem_num, &target_sembuf[i].sem_num); 3477 __get_user(host_sembuf[i].sem_op, &target_sembuf[i].sem_op); 3478 __get_user(host_sembuf[i].sem_flg, &target_sembuf[i].sem_flg); 3479 } 3480 3481 unlock_user(target_sembuf, target_addr, 0); 3482 3483 return 0; 3484 } 3485 3486 static inline abi_long do_semop(int semid, abi_long ptr, unsigned nsops) 3487 { 3488 struct sembuf sops[nsops]; 3489 3490 if (target_to_host_sembuf(sops, ptr, nsops)) 3491 return -TARGET_EFAULT; 3492 3493 return get_errno(safe_semtimedop(semid, sops, nsops, NULL)); 3494 } 3495 3496 struct target_msqid_ds 3497 { 3498 struct target_ipc_perm msg_perm; 3499 abi_ulong msg_stime; 3500 #if TARGET_ABI_BITS == 32 3501 abi_ulong __unused1; 3502 #endif 3503 abi_ulong msg_rtime; 3504 #if TARGET_ABI_BITS == 32 3505 abi_ulong __unused2; 3506 #endif 3507 abi_ulong msg_ctime; 3508 #if TARGET_ABI_BITS == 32 3509 abi_ulong __unused3; 3510 #endif 3511 abi_ulong __msg_cbytes; 3512 abi_ulong msg_qnum; 3513 abi_ulong msg_qbytes; 3514 abi_ulong msg_lspid; 3515 abi_ulong msg_lrpid; 3516 abi_ulong __unused4; 3517 abi_ulong __unused5; 3518 }; 3519 3520 static inline abi_long target_to_host_msqid_ds(struct msqid_ds *host_md, 3521 abi_ulong target_addr) 3522 { 3523 struct target_msqid_ds *target_md; 3524 3525 if (!lock_user_struct(VERIFY_READ, target_md, target_addr, 1)) 3526 return -TARGET_EFAULT; 3527 if (target_to_host_ipc_perm(&(host_md->msg_perm),target_addr)) 3528 return -TARGET_EFAULT; 3529 host_md->msg_stime = tswapal(target_md->msg_stime); 3530 host_md->msg_rtime = tswapal(target_md->msg_rtime); 3531 host_md->msg_ctime = tswapal(target_md->msg_ctime); 3532 host_md->__msg_cbytes = tswapal(target_md->__msg_cbytes); 3533 host_md->msg_qnum = tswapal(target_md->msg_qnum); 3534 host_md->msg_qbytes = tswapal(target_md->msg_qbytes); 3535 host_md->msg_lspid = tswapal(target_md->msg_lspid); 3536 host_md->msg_lrpid = tswapal(target_md->msg_lrpid); 3537 unlock_user_struct(target_md, target_addr, 0); 3538 return 0; 3539 } 3540 3541 static inline abi_long host_to_target_msqid_ds(abi_ulong target_addr, 3542 struct msqid_ds *host_md) 3543 { 3544 struct target_msqid_ds *target_md; 3545 3546 if (!lock_user_struct(VERIFY_WRITE, target_md, target_addr, 0)) 3547 return -TARGET_EFAULT; 3548 if (host_to_target_ipc_perm(target_addr,&(host_md->msg_perm))) 3549 return -TARGET_EFAULT; 3550 target_md->msg_stime = tswapal(host_md->msg_stime); 3551 target_md->msg_rtime = tswapal(host_md->msg_rtime); 3552 target_md->msg_ctime = tswapal(host_md->msg_ctime); 3553 target_md->__msg_cbytes = tswapal(host_md->__msg_cbytes); 3554 target_md->msg_qnum = tswapal(host_md->msg_qnum); 3555 target_md->msg_qbytes = tswapal(host_md->msg_qbytes); 3556 target_md->msg_lspid = tswapal(host_md->msg_lspid); 3557 target_md->msg_lrpid = tswapal(host_md->msg_lrpid); 3558 unlock_user_struct(target_md, target_addr, 1); 3559 return 0; 3560 } 3561 3562 struct target_msginfo { 3563 int msgpool; 3564 int msgmap; 3565 int msgmax; 3566 int msgmnb; 3567 int msgmni; 3568 int msgssz; 3569 int msgtql; 3570 unsigned short int msgseg; 3571 }; 3572 3573 static inline abi_long host_to_target_msginfo(abi_ulong target_addr, 3574 struct msginfo *host_msginfo) 3575 { 3576 struct target_msginfo *target_msginfo; 3577 if (!lock_user_struct(VERIFY_WRITE, target_msginfo, target_addr, 0)) 3578 return -TARGET_EFAULT; 3579 __put_user(host_msginfo->msgpool, &target_msginfo->msgpool); 3580 __put_user(host_msginfo->msgmap, &target_msginfo->msgmap); 3581 __put_user(host_msginfo->msgmax, &target_msginfo->msgmax); 3582 __put_user(host_msginfo->msgmnb, &target_msginfo->msgmnb); 3583 __put_user(host_msginfo->msgmni, &target_msginfo->msgmni); 3584 __put_user(host_msginfo->msgssz, &target_msginfo->msgssz); 3585 __put_user(host_msginfo->msgtql, &target_msginfo->msgtql); 3586 __put_user(host_msginfo->msgseg, &target_msginfo->msgseg); 3587 unlock_user_struct(target_msginfo, target_addr, 1); 3588 return 0; 3589 } 3590 3591 static inline abi_long do_msgctl(int msgid, int cmd, abi_long ptr) 3592 { 3593 struct msqid_ds dsarg; 3594 struct msginfo msginfo; 3595 abi_long ret = -TARGET_EINVAL; 3596 3597 cmd &= 0xff; 3598 3599 switch (cmd) { 3600 case IPC_STAT: 3601 case IPC_SET: 3602 case MSG_STAT: 3603 if (target_to_host_msqid_ds(&dsarg,ptr)) 3604 return -TARGET_EFAULT; 3605 ret = get_errno(msgctl(msgid, cmd, &dsarg)); 3606 if (host_to_target_msqid_ds(ptr,&dsarg)) 3607 return -TARGET_EFAULT; 3608 break; 3609 case IPC_RMID: 3610 ret = get_errno(msgctl(msgid, cmd, NULL)); 3611 break; 3612 case IPC_INFO: 3613 case MSG_INFO: 3614 ret = get_errno(msgctl(msgid, cmd, (struct msqid_ds *)&msginfo)); 3615 if (host_to_target_msginfo(ptr, &msginfo)) 3616 return -TARGET_EFAULT; 3617 break; 3618 } 3619 3620 return ret; 3621 } 3622 3623 struct target_msgbuf { 3624 abi_long mtype; 3625 char mtext[1]; 3626 }; 3627 3628 static inline abi_long do_msgsnd(int msqid, abi_long msgp, 3629 ssize_t msgsz, int msgflg) 3630 { 3631 struct target_msgbuf *target_mb; 3632 struct msgbuf *host_mb; 3633 abi_long ret = 0; 3634 3635 if (msgsz < 0) { 3636 return -TARGET_EINVAL; 3637 } 3638 3639 if (!lock_user_struct(VERIFY_READ, target_mb, msgp, 0)) 3640 return -TARGET_EFAULT; 3641 host_mb = g_try_malloc(msgsz + sizeof(long)); 3642 if (!host_mb) { 3643 unlock_user_struct(target_mb, msgp, 0); 3644 return -TARGET_ENOMEM; 3645 } 3646 host_mb->mtype = (abi_long) tswapal(target_mb->mtype); 3647 memcpy(host_mb->mtext, target_mb->mtext, msgsz); 3648 ret = get_errno(safe_msgsnd(msqid, host_mb, msgsz, msgflg)); 3649 g_free(host_mb); 3650 unlock_user_struct(target_mb, msgp, 0); 3651 3652 return ret; 3653 } 3654 3655 static inline abi_long do_msgrcv(int msqid, abi_long msgp, 3656 ssize_t msgsz, abi_long msgtyp, 3657 int msgflg) 3658 { 3659 struct target_msgbuf *target_mb; 3660 char *target_mtext; 3661 struct msgbuf *host_mb; 3662 abi_long ret = 0; 3663 3664 if (msgsz < 0) { 3665 return -TARGET_EINVAL; 3666 } 3667 3668 if (!lock_user_struct(VERIFY_WRITE, target_mb, msgp, 0)) 3669 return -TARGET_EFAULT; 3670 3671 host_mb = g_try_malloc(msgsz + sizeof(long)); 3672 if (!host_mb) { 3673 ret = -TARGET_ENOMEM; 3674 goto end; 3675 } 3676 ret = get_errno(safe_msgrcv(msqid, host_mb, msgsz, msgtyp, msgflg)); 3677 3678 if (ret > 0) { 3679 abi_ulong target_mtext_addr = msgp + sizeof(abi_ulong); 3680 target_mtext = lock_user(VERIFY_WRITE, target_mtext_addr, ret, 0); 3681 if (!target_mtext) { 3682 ret = -TARGET_EFAULT; 3683 goto end; 3684 } 3685 memcpy(target_mb->mtext, host_mb->mtext, ret); 3686 unlock_user(target_mtext, target_mtext_addr, ret); 3687 } 3688 3689 target_mb->mtype = tswapal(host_mb->mtype); 3690 3691 end: 3692 if (target_mb) 3693 unlock_user_struct(target_mb, msgp, 1); 3694 g_free(host_mb); 3695 return ret; 3696 } 3697 3698 static inline abi_long target_to_host_shmid_ds(struct shmid_ds *host_sd, 3699 abi_ulong target_addr) 3700 { 3701 struct target_shmid_ds *target_sd; 3702 3703 if (!lock_user_struct(VERIFY_READ, target_sd, target_addr, 1)) 3704 return -TARGET_EFAULT; 3705 if (target_to_host_ipc_perm(&(host_sd->shm_perm), target_addr)) 3706 return -TARGET_EFAULT; 3707 __get_user(host_sd->shm_segsz, &target_sd->shm_segsz); 3708 __get_user(host_sd->shm_atime, &target_sd->shm_atime); 3709 __get_user(host_sd->shm_dtime, &target_sd->shm_dtime); 3710 __get_user(host_sd->shm_ctime, &target_sd->shm_ctime); 3711 __get_user(host_sd->shm_cpid, &target_sd->shm_cpid); 3712 __get_user(host_sd->shm_lpid, &target_sd->shm_lpid); 3713 __get_user(host_sd->shm_nattch, &target_sd->shm_nattch); 3714 unlock_user_struct(target_sd, target_addr, 0); 3715 return 0; 3716 } 3717 3718 static inline abi_long host_to_target_shmid_ds(abi_ulong target_addr, 3719 struct shmid_ds *host_sd) 3720 { 3721 struct target_shmid_ds *target_sd; 3722 3723 if (!lock_user_struct(VERIFY_WRITE, target_sd, target_addr, 0)) 3724 return -TARGET_EFAULT; 3725 if (host_to_target_ipc_perm(target_addr, &(host_sd->shm_perm))) 3726 return -TARGET_EFAULT; 3727 __put_user(host_sd->shm_segsz, &target_sd->shm_segsz); 3728 __put_user(host_sd->shm_atime, &target_sd->shm_atime); 3729 __put_user(host_sd->shm_dtime, &target_sd->shm_dtime); 3730 __put_user(host_sd->shm_ctime, &target_sd->shm_ctime); 3731 __put_user(host_sd->shm_cpid, &target_sd->shm_cpid); 3732 __put_user(host_sd->shm_lpid, &target_sd->shm_lpid); 3733 __put_user(host_sd->shm_nattch, &target_sd->shm_nattch); 3734 unlock_user_struct(target_sd, target_addr, 1); 3735 return 0; 3736 } 3737 3738 struct target_shminfo { 3739 abi_ulong shmmax; 3740 abi_ulong shmmin; 3741 abi_ulong shmmni; 3742 abi_ulong shmseg; 3743 abi_ulong shmall; 3744 }; 3745 3746 static inline abi_long host_to_target_shminfo(abi_ulong target_addr, 3747 struct shminfo *host_shminfo) 3748 { 3749 struct target_shminfo *target_shminfo; 3750 if (!lock_user_struct(VERIFY_WRITE, target_shminfo, target_addr, 0)) 3751 return -TARGET_EFAULT; 3752 __put_user(host_shminfo->shmmax, &target_shminfo->shmmax); 3753 __put_user(host_shminfo->shmmin, &target_shminfo->shmmin); 3754 __put_user(host_shminfo->shmmni, &target_shminfo->shmmni); 3755 __put_user(host_shminfo->shmseg, &target_shminfo->shmseg); 3756 __put_user(host_shminfo->shmall, &target_shminfo->shmall); 3757 unlock_user_struct(target_shminfo, target_addr, 1); 3758 return 0; 3759 } 3760 3761 struct target_shm_info { 3762 int used_ids; 3763 abi_ulong shm_tot; 3764 abi_ulong shm_rss; 3765 abi_ulong shm_swp; 3766 abi_ulong swap_attempts; 3767 abi_ulong swap_successes; 3768 }; 3769 3770 static inline abi_long host_to_target_shm_info(abi_ulong target_addr, 3771 struct shm_info *host_shm_info) 3772 { 3773 struct target_shm_info *target_shm_info; 3774 if (!lock_user_struct(VERIFY_WRITE, target_shm_info, target_addr, 0)) 3775 return -TARGET_EFAULT; 3776 __put_user(host_shm_info->used_ids, &target_shm_info->used_ids); 3777 __put_user(host_shm_info->shm_tot, &target_shm_info->shm_tot); 3778 __put_user(host_shm_info->shm_rss, &target_shm_info->shm_rss); 3779 __put_user(host_shm_info->shm_swp, &target_shm_info->shm_swp); 3780 __put_user(host_shm_info->swap_attempts, &target_shm_info->swap_attempts); 3781 __put_user(host_shm_info->swap_successes, &target_shm_info->swap_successes); 3782 unlock_user_struct(target_shm_info, target_addr, 1); 3783 return 0; 3784 } 3785 3786 static inline abi_long do_shmctl(int shmid, int cmd, abi_long buf) 3787 { 3788 struct shmid_ds dsarg; 3789 struct shminfo shminfo; 3790 struct shm_info shm_info; 3791 abi_long ret = -TARGET_EINVAL; 3792 3793 cmd &= 0xff; 3794 3795 switch(cmd) { 3796 case IPC_STAT: 3797 case IPC_SET: 3798 case SHM_STAT: 3799 if (target_to_host_shmid_ds(&dsarg, buf)) 3800 return -TARGET_EFAULT; 3801 ret = get_errno(shmctl(shmid, cmd, &dsarg)); 3802 if (host_to_target_shmid_ds(buf, &dsarg)) 3803 return -TARGET_EFAULT; 3804 break; 3805 case IPC_INFO: 3806 ret = get_errno(shmctl(shmid, cmd, (struct shmid_ds *)&shminfo)); 3807 if (host_to_target_shminfo(buf, &shminfo)) 3808 return -TARGET_EFAULT; 3809 break; 3810 case SHM_INFO: 3811 ret = get_errno(shmctl(shmid, cmd, (struct shmid_ds *)&shm_info)); 3812 if (host_to_target_shm_info(buf, &shm_info)) 3813 return -TARGET_EFAULT; 3814 break; 3815 case IPC_RMID: 3816 case SHM_LOCK: 3817 case SHM_UNLOCK: 3818 ret = get_errno(shmctl(shmid, cmd, NULL)); 3819 break; 3820 } 3821 3822 return ret; 3823 } 3824 3825 #ifndef TARGET_FORCE_SHMLBA 3826 /* For most architectures, SHMLBA is the same as the page size; 3827 * some architectures have larger values, in which case they should 3828 * define TARGET_FORCE_SHMLBA and provide a target_shmlba() function. 3829 * This corresponds to the kernel arch code defining __ARCH_FORCE_SHMLBA 3830 * and defining its own value for SHMLBA. 3831 * 3832 * The kernel also permits SHMLBA to be set by the architecture to a 3833 * value larger than the page size without setting __ARCH_FORCE_SHMLBA; 3834 * this means that addresses are rounded to the large size if 3835 * SHM_RND is set but addresses not aligned to that size are not rejected 3836 * as long as they are at least page-aligned. Since the only architecture 3837 * which uses this is ia64 this code doesn't provide for that oddity. 3838 */ 3839 static inline abi_ulong target_shmlba(CPUArchState *cpu_env) 3840 { 3841 return TARGET_PAGE_SIZE; 3842 } 3843 #endif 3844 3845 static inline abi_ulong do_shmat(CPUArchState *cpu_env, 3846 int shmid, abi_ulong shmaddr, int shmflg) 3847 { 3848 abi_long raddr; 3849 void *host_raddr; 3850 struct shmid_ds shm_info; 3851 int i,ret; 3852 abi_ulong shmlba; 3853 3854 /* find out the length of the shared memory segment */ 3855 ret = get_errno(shmctl(shmid, IPC_STAT, &shm_info)); 3856 if (is_error(ret)) { 3857 /* can't get length, bail out */ 3858 return ret; 3859 } 3860 3861 shmlba = target_shmlba(cpu_env); 3862 3863 if (shmaddr & (shmlba - 1)) { 3864 if (shmflg & SHM_RND) { 3865 shmaddr &= ~(shmlba - 1); 3866 } else { 3867 return -TARGET_EINVAL; 3868 } 3869 } 3870 if (!guest_range_valid(shmaddr, shm_info.shm_segsz)) { 3871 return -TARGET_EINVAL; 3872 } 3873 3874 mmap_lock(); 3875 3876 if (shmaddr) 3877 host_raddr = shmat(shmid, (void *)g2h(shmaddr), shmflg); 3878 else { 3879 abi_ulong mmap_start; 3880 3881 mmap_start = mmap_find_vma(0, shm_info.shm_segsz); 3882 3883 if (mmap_start == -1) { 3884 errno = ENOMEM; 3885 host_raddr = (void *)-1; 3886 } else 3887 host_raddr = shmat(shmid, g2h(mmap_start), shmflg | SHM_REMAP); 3888 } 3889 3890 if (host_raddr == (void *)-1) { 3891 mmap_unlock(); 3892 return get_errno((long)host_raddr); 3893 } 3894 raddr=h2g((unsigned long)host_raddr); 3895 3896 page_set_flags(raddr, raddr + shm_info.shm_segsz, 3897 PAGE_VALID | PAGE_READ | 3898 ((shmflg & SHM_RDONLY)? 0 : PAGE_WRITE)); 3899 3900 for (i = 0; i < N_SHM_REGIONS; i++) { 3901 if (!shm_regions[i].in_use) { 3902 shm_regions[i].in_use = true; 3903 shm_regions[i].start = raddr; 3904 shm_regions[i].size = shm_info.shm_segsz; 3905 break; 3906 } 3907 } 3908 3909 mmap_unlock(); 3910 return raddr; 3911 3912 } 3913 3914 static inline abi_long do_shmdt(abi_ulong shmaddr) 3915 { 3916 int i; 3917 abi_long rv; 3918 3919 mmap_lock(); 3920 3921 for (i = 0; i < N_SHM_REGIONS; ++i) { 3922 if (shm_regions[i].in_use && shm_regions[i].start == shmaddr) { 3923 shm_regions[i].in_use = false; 3924 page_set_flags(shmaddr, shmaddr + shm_regions[i].size, 0); 3925 break; 3926 } 3927 } 3928 rv = get_errno(shmdt(g2h(shmaddr))); 3929 3930 mmap_unlock(); 3931 3932 return rv; 3933 } 3934 3935 #ifdef TARGET_NR_ipc 3936 /* ??? This only works with linear mappings. */ 3937 /* do_ipc() must return target values and target errnos. */ 3938 static abi_long do_ipc(CPUArchState *cpu_env, 3939 unsigned int call, abi_long first, 3940 abi_long second, abi_long third, 3941 abi_long ptr, abi_long fifth) 3942 { 3943 int version; 3944 abi_long ret = 0; 3945 3946 version = call >> 16; 3947 call &= 0xffff; 3948 3949 switch (call) { 3950 case IPCOP_semop: 3951 ret = do_semop(first, ptr, second); 3952 break; 3953 3954 case IPCOP_semget: 3955 ret = get_errno(semget(first, second, third)); 3956 break; 3957 3958 case IPCOP_semctl: { 3959 /* The semun argument to semctl is passed by value, so dereference the 3960 * ptr argument. */ 3961 abi_ulong atptr; 3962 get_user_ual(atptr, ptr); 3963 ret = do_semctl(first, second, third, atptr); 3964 break; 3965 } 3966 3967 case IPCOP_msgget: 3968 ret = get_errno(msgget(first, second)); 3969 break; 3970 3971 case IPCOP_msgsnd: 3972 ret = do_msgsnd(first, ptr, second, third); 3973 break; 3974 3975 case IPCOP_msgctl: 3976 ret = do_msgctl(first, second, ptr); 3977 break; 3978 3979 case IPCOP_msgrcv: 3980 switch (version) { 3981 case 0: 3982 { 3983 struct target_ipc_kludge { 3984 abi_long msgp; 3985 abi_long msgtyp; 3986 } *tmp; 3987 3988 if (!lock_user_struct(VERIFY_READ, tmp, ptr, 1)) { 3989 ret = -TARGET_EFAULT; 3990 break; 3991 } 3992 3993 ret = do_msgrcv(first, tswapal(tmp->msgp), second, tswapal(tmp->msgtyp), third); 3994 3995 unlock_user_struct(tmp, ptr, 0); 3996 break; 3997 } 3998 default: 3999 ret = do_msgrcv(first, ptr, second, fifth, third); 4000 } 4001 break; 4002 4003 case IPCOP_shmat: 4004 switch (version) { 4005 default: 4006 { 4007 abi_ulong raddr; 4008 raddr = do_shmat(cpu_env, first, ptr, second); 4009 if (is_error(raddr)) 4010 return get_errno(raddr); 4011 if (put_user_ual(raddr, third)) 4012 return -TARGET_EFAULT; 4013 break; 4014 } 4015 case 1: 4016 ret = -TARGET_EINVAL; 4017 break; 4018 } 4019 break; 4020 case IPCOP_shmdt: 4021 ret = do_shmdt(ptr); 4022 break; 4023 4024 case IPCOP_shmget: 4025 /* IPC_* flag values are the same on all linux platforms */ 4026 ret = get_errno(shmget(first, second, third)); 4027 break; 4028 4029 /* IPC_* and SHM_* command values are the same on all linux platforms */ 4030 case IPCOP_shmctl: 4031 ret = do_shmctl(first, second, ptr); 4032 break; 4033 default: 4034 gemu_log("Unsupported ipc call: %d (version %d)\n", call, version); 4035 ret = -TARGET_ENOSYS; 4036 break; 4037 } 4038 return ret; 4039 } 4040 #endif 4041 4042 /* kernel structure types definitions */ 4043 4044 #define STRUCT(name, ...) STRUCT_ ## name, 4045 #define STRUCT_SPECIAL(name) STRUCT_ ## name, 4046 enum { 4047 #include "syscall_types.h" 4048 STRUCT_MAX 4049 }; 4050 #undef STRUCT 4051 #undef STRUCT_SPECIAL 4052 4053 #define STRUCT(name, ...) static const argtype struct_ ## name ## _def[] = { __VA_ARGS__, TYPE_NULL }; 4054 #define STRUCT_SPECIAL(name) 4055 #include "syscall_types.h" 4056 #undef STRUCT 4057 #undef STRUCT_SPECIAL 4058 4059 typedef struct IOCTLEntry IOCTLEntry; 4060 4061 typedef abi_long do_ioctl_fn(const IOCTLEntry *ie, uint8_t *buf_temp, 4062 int fd, int cmd, abi_long arg); 4063 4064 struct IOCTLEntry { 4065 int target_cmd; 4066 unsigned int host_cmd; 4067 const char *name; 4068 int access; 4069 do_ioctl_fn *do_ioctl; 4070 const argtype arg_type[5]; 4071 }; 4072 4073 #define IOC_R 0x0001 4074 #define IOC_W 0x0002 4075 #define IOC_RW (IOC_R | IOC_W) 4076 4077 #define MAX_STRUCT_SIZE 4096 4078 4079 #ifdef CONFIG_FIEMAP 4080 /* So fiemap access checks don't overflow on 32 bit systems. 4081 * This is very slightly smaller than the limit imposed by 4082 * the underlying kernel. 4083 */ 4084 #define FIEMAP_MAX_EXTENTS ((UINT_MAX - sizeof(struct fiemap)) \ 4085 / sizeof(struct fiemap_extent)) 4086 4087 static abi_long do_ioctl_fs_ioc_fiemap(const IOCTLEntry *ie, uint8_t *buf_temp, 4088 int fd, int cmd, abi_long arg) 4089 { 4090 /* The parameter for this ioctl is a struct fiemap followed 4091 * by an array of struct fiemap_extent whose size is set 4092 * in fiemap->fm_extent_count. The array is filled in by the 4093 * ioctl. 4094 */ 4095 int target_size_in, target_size_out; 4096 struct fiemap *fm; 4097 const argtype *arg_type = ie->arg_type; 4098 const argtype extent_arg_type[] = { MK_STRUCT(STRUCT_fiemap_extent) }; 4099 void *argptr, *p; 4100 abi_long ret; 4101 int i, extent_size = thunk_type_size(extent_arg_type, 0); 4102 uint32_t outbufsz; 4103 int free_fm = 0; 4104 4105 assert(arg_type[0] == TYPE_PTR); 4106 assert(ie->access == IOC_RW); 4107 arg_type++; 4108 target_size_in = thunk_type_size(arg_type, 0); 4109 argptr = lock_user(VERIFY_READ, arg, target_size_in, 1); 4110 if (!argptr) { 4111 return -TARGET_EFAULT; 4112 } 4113 thunk_convert(buf_temp, argptr, arg_type, THUNK_HOST); 4114 unlock_user(argptr, arg, 0); 4115 fm = (struct fiemap *)buf_temp; 4116 if (fm->fm_extent_count > FIEMAP_MAX_EXTENTS) { 4117 return -TARGET_EINVAL; 4118 } 4119 4120 outbufsz = sizeof (*fm) + 4121 (sizeof(struct fiemap_extent) * fm->fm_extent_count); 4122 4123 if (outbufsz > MAX_STRUCT_SIZE) { 4124 /* We can't fit all the extents into the fixed size buffer. 4125 * Allocate one that is large enough and use it instead. 4126 */ 4127 fm = g_try_malloc(outbufsz); 4128 if (!fm) { 4129 return -TARGET_ENOMEM; 4130 } 4131 memcpy(fm, buf_temp, sizeof(struct fiemap)); 4132 free_fm = 1; 4133 } 4134 ret = get_errno(safe_ioctl(fd, ie->host_cmd, fm)); 4135 if (!is_error(ret)) { 4136 target_size_out = target_size_in; 4137 /* An extent_count of 0 means we were only counting the extents 4138 * so there are no structs to copy 4139 */ 4140 if (fm->fm_extent_count != 0) { 4141 target_size_out += fm->fm_mapped_extents * extent_size; 4142 } 4143 argptr = lock_user(VERIFY_WRITE, arg, target_size_out, 0); 4144 if (!argptr) { 4145 ret = -TARGET_EFAULT; 4146 } else { 4147 /* Convert the struct fiemap */ 4148 thunk_convert(argptr, fm, arg_type, THUNK_TARGET); 4149 if (fm->fm_extent_count != 0) { 4150 p = argptr + target_size_in; 4151 /* ...and then all the struct fiemap_extents */ 4152 for (i = 0; i < fm->fm_mapped_extents; i++) { 4153 thunk_convert(p, &fm->fm_extents[i], extent_arg_type, 4154 THUNK_TARGET); 4155 p += extent_size; 4156 } 4157 } 4158 unlock_user(argptr, arg, target_size_out); 4159 } 4160 } 4161 if (free_fm) { 4162 g_free(fm); 4163 } 4164 return ret; 4165 } 4166 #endif 4167 4168 static abi_long do_ioctl_ifconf(const IOCTLEntry *ie, uint8_t *buf_temp, 4169 int fd, int cmd, abi_long arg) 4170 { 4171 const argtype *arg_type = ie->arg_type; 4172 int target_size; 4173 void *argptr; 4174 int ret; 4175 struct ifconf *host_ifconf; 4176 uint32_t outbufsz; 4177 const argtype ifreq_arg_type[] = { MK_STRUCT(STRUCT_sockaddr_ifreq) }; 4178 int target_ifreq_size; 4179 int nb_ifreq; 4180 int free_buf = 0; 4181 int i; 4182 int target_ifc_len; 4183 abi_long target_ifc_buf; 4184 int host_ifc_len; 4185 char *host_ifc_buf; 4186 4187 assert(arg_type[0] == TYPE_PTR); 4188 assert(ie->access == IOC_RW); 4189 4190 arg_type++; 4191 target_size = thunk_type_size(arg_type, 0); 4192 4193 argptr = lock_user(VERIFY_READ, arg, target_size, 1); 4194 if (!argptr) 4195 return -TARGET_EFAULT; 4196 thunk_convert(buf_temp, argptr, arg_type, THUNK_HOST); 4197 unlock_user(argptr, arg, 0); 4198 4199 host_ifconf = (struct ifconf *)(unsigned long)buf_temp; 4200 target_ifc_buf = (abi_long)(unsigned long)host_ifconf->ifc_buf; 4201 target_ifreq_size = thunk_type_size(ifreq_arg_type, 0); 4202 4203 if (target_ifc_buf != 0) { 4204 target_ifc_len = host_ifconf->ifc_len; 4205 nb_ifreq = target_ifc_len / target_ifreq_size; 4206 host_ifc_len = nb_ifreq * sizeof(struct ifreq); 4207 4208 outbufsz = sizeof(*host_ifconf) + host_ifc_len; 4209 if (outbufsz > MAX_STRUCT_SIZE) { 4210 /* 4211 * We can't fit all the extents into the fixed size buffer. 4212 * Allocate one that is large enough and use it instead. 4213 */ 4214 host_ifconf = malloc(outbufsz); 4215 if (!host_ifconf) { 4216 return -TARGET_ENOMEM; 4217 } 4218 memcpy(host_ifconf, buf_temp, sizeof(*host_ifconf)); 4219 free_buf = 1; 4220 } 4221 host_ifc_buf = (char *)host_ifconf + sizeof(*host_ifconf); 4222 4223 host_ifconf->ifc_len = host_ifc_len; 4224 } else { 4225 host_ifc_buf = NULL; 4226 } 4227 host_ifconf->ifc_buf = host_ifc_buf; 4228 4229 ret = get_errno(safe_ioctl(fd, ie->host_cmd, host_ifconf)); 4230 if (!is_error(ret)) { 4231 /* convert host ifc_len to target ifc_len */ 4232 4233 nb_ifreq = host_ifconf->ifc_len / sizeof(struct ifreq); 4234 target_ifc_len = nb_ifreq * target_ifreq_size; 4235 host_ifconf->ifc_len = target_ifc_len; 4236 4237 /* restore target ifc_buf */ 4238 4239 host_ifconf->ifc_buf = (char *)(unsigned long)target_ifc_buf; 4240 4241 /* copy struct ifconf to target user */ 4242 4243 argptr = lock_user(VERIFY_WRITE, arg, target_size, 0); 4244 if (!argptr) 4245 return -TARGET_EFAULT; 4246 thunk_convert(argptr, host_ifconf, arg_type, THUNK_TARGET); 4247 unlock_user(argptr, arg, target_size); 4248 4249 if (target_ifc_buf != 0) { 4250 /* copy ifreq[] to target user */ 4251 argptr = lock_user(VERIFY_WRITE, target_ifc_buf, target_ifc_len, 0); 4252 for (i = 0; i < nb_ifreq ; i++) { 4253 thunk_convert(argptr + i * target_ifreq_size, 4254 host_ifc_buf + i * sizeof(struct ifreq), 4255 ifreq_arg_type, THUNK_TARGET); 4256 } 4257 unlock_user(argptr, target_ifc_buf, target_ifc_len); 4258 } 4259 } 4260 4261 if (free_buf) { 4262 free(host_ifconf); 4263 } 4264 4265 return ret; 4266 } 4267 4268 #if defined(CONFIG_USBFS) 4269 #if HOST_LONG_BITS > 64 4270 #error USBDEVFS thunks do not support >64 bit hosts yet. 4271 #endif 4272 struct live_urb { 4273 uint64_t target_urb_adr; 4274 uint64_t target_buf_adr; 4275 char *target_buf_ptr; 4276 struct usbdevfs_urb host_urb; 4277 }; 4278 4279 static GHashTable *usbdevfs_urb_hashtable(void) 4280 { 4281 static GHashTable *urb_hashtable; 4282 4283 if (!urb_hashtable) { 4284 urb_hashtable = g_hash_table_new(g_int64_hash, g_int64_equal); 4285 } 4286 return urb_hashtable; 4287 } 4288 4289 static void urb_hashtable_insert(struct live_urb *urb) 4290 { 4291 GHashTable *urb_hashtable = usbdevfs_urb_hashtable(); 4292 g_hash_table_insert(urb_hashtable, urb, urb); 4293 } 4294 4295 static struct live_urb *urb_hashtable_lookup(uint64_t target_urb_adr) 4296 { 4297 GHashTable *urb_hashtable = usbdevfs_urb_hashtable(); 4298 return g_hash_table_lookup(urb_hashtable, &target_urb_adr); 4299 } 4300 4301 static void urb_hashtable_remove(struct live_urb *urb) 4302 { 4303 GHashTable *urb_hashtable = usbdevfs_urb_hashtable(); 4304 g_hash_table_remove(urb_hashtable, urb); 4305 } 4306 4307 static abi_long 4308 do_ioctl_usbdevfs_reapurb(const IOCTLEntry *ie, uint8_t *buf_temp, 4309 int fd, int cmd, abi_long arg) 4310 { 4311 const argtype usbfsurb_arg_type[] = { MK_STRUCT(STRUCT_usbdevfs_urb) }; 4312 const argtype ptrvoid_arg_type[] = { TYPE_PTRVOID, 0, 0 }; 4313 struct live_urb *lurb; 4314 void *argptr; 4315 uint64_t hurb; 4316 int target_size; 4317 uintptr_t target_urb_adr; 4318 abi_long ret; 4319 4320 target_size = thunk_type_size(usbfsurb_arg_type, THUNK_TARGET); 4321 4322 memset(buf_temp, 0, sizeof(uint64_t)); 4323 ret = get_errno(safe_ioctl(fd, ie->host_cmd, buf_temp)); 4324 if (is_error(ret)) { 4325 return ret; 4326 } 4327 4328 memcpy(&hurb, buf_temp, sizeof(uint64_t)); 4329 lurb = (void *)((uintptr_t)hurb - offsetof(struct live_urb, host_urb)); 4330 if (!lurb->target_urb_adr) { 4331 return -TARGET_EFAULT; 4332 } 4333 urb_hashtable_remove(lurb); 4334 unlock_user(lurb->target_buf_ptr, lurb->target_buf_adr, 4335 lurb->host_urb.buffer_length); 4336 lurb->target_buf_ptr = NULL; 4337 4338 /* restore the guest buffer pointer */ 4339 lurb->host_urb.buffer = (void *)(uintptr_t)lurb->target_buf_adr; 4340 4341 /* update the guest urb struct */ 4342 argptr = lock_user(VERIFY_WRITE, lurb->target_urb_adr, target_size, 0); 4343 if (!argptr) { 4344 g_free(lurb); 4345 return -TARGET_EFAULT; 4346 } 4347 thunk_convert(argptr, &lurb->host_urb, usbfsurb_arg_type, THUNK_TARGET); 4348 unlock_user(argptr, lurb->target_urb_adr, target_size); 4349 4350 target_size = thunk_type_size(ptrvoid_arg_type, THUNK_TARGET); 4351 /* write back the urb handle */ 4352 argptr = lock_user(VERIFY_WRITE, arg, target_size, 0); 4353 if (!argptr) { 4354 g_free(lurb); 4355 return -TARGET_EFAULT; 4356 } 4357 4358 /* GHashTable uses 64-bit keys but thunk_convert expects uintptr_t */ 4359 target_urb_adr = lurb->target_urb_adr; 4360 thunk_convert(argptr, &target_urb_adr, ptrvoid_arg_type, THUNK_TARGET); 4361 unlock_user(argptr, arg, target_size); 4362 4363 g_free(lurb); 4364 return ret; 4365 } 4366 4367 static abi_long 4368 do_ioctl_usbdevfs_discardurb(const IOCTLEntry *ie, 4369 uint8_t *buf_temp __attribute__((unused)), 4370 int fd, int cmd, abi_long arg) 4371 { 4372 struct live_urb *lurb; 4373 4374 /* map target address back to host URB with metadata. */ 4375 lurb = urb_hashtable_lookup(arg); 4376 if (!lurb) { 4377 return -TARGET_EFAULT; 4378 } 4379 return get_errno(safe_ioctl(fd, ie->host_cmd, &lurb->host_urb)); 4380 } 4381 4382 static abi_long 4383 do_ioctl_usbdevfs_submiturb(const IOCTLEntry *ie, uint8_t *buf_temp, 4384 int fd, int cmd, abi_long arg) 4385 { 4386 const argtype *arg_type = ie->arg_type; 4387 int target_size; 4388 abi_long ret; 4389 void *argptr; 4390 int rw_dir; 4391 struct live_urb *lurb; 4392 4393 /* 4394 * each submitted URB needs to map to a unique ID for the 4395 * kernel, and that unique ID needs to be a pointer to 4396 * host memory. hence, we need to malloc for each URB. 4397 * isochronous transfers have a variable length struct. 4398 */ 4399 arg_type++; 4400 target_size = thunk_type_size(arg_type, THUNK_TARGET); 4401 4402 /* construct host copy of urb and metadata */ 4403 lurb = g_try_malloc0(sizeof(struct live_urb)); 4404 if (!lurb) { 4405 return -TARGET_ENOMEM; 4406 } 4407 4408 argptr = lock_user(VERIFY_READ, arg, target_size, 1); 4409 if (!argptr) { 4410 g_free(lurb); 4411 return -TARGET_EFAULT; 4412 } 4413 thunk_convert(&lurb->host_urb, argptr, arg_type, THUNK_HOST); 4414 unlock_user(argptr, arg, 0); 4415 4416 lurb->target_urb_adr = arg; 4417 lurb->target_buf_adr = (uintptr_t)lurb->host_urb.buffer; 4418 4419 /* buffer space used depends on endpoint type so lock the entire buffer */ 4420 /* control type urbs should check the buffer contents for true direction */ 4421 rw_dir = lurb->host_urb.endpoint & USB_DIR_IN ? VERIFY_WRITE : VERIFY_READ; 4422 lurb->target_buf_ptr = lock_user(rw_dir, lurb->target_buf_adr, 4423 lurb->host_urb.buffer_length, 1); 4424 if (lurb->target_buf_ptr == NULL) { 4425 g_free(lurb); 4426 return -TARGET_EFAULT; 4427 } 4428 4429 /* update buffer pointer in host copy */ 4430 lurb->host_urb.buffer = lurb->target_buf_ptr; 4431 4432 ret = get_errno(safe_ioctl(fd, ie->host_cmd, &lurb->host_urb)); 4433 if (is_error(ret)) { 4434 unlock_user(lurb->target_buf_ptr, lurb->target_buf_adr, 0); 4435 g_free(lurb); 4436 } else { 4437 urb_hashtable_insert(lurb); 4438 } 4439 4440 return ret; 4441 } 4442 #endif /* CONFIG_USBFS */ 4443 4444 static abi_long do_ioctl_dm(const IOCTLEntry *ie, uint8_t *buf_temp, int fd, 4445 int cmd, abi_long arg) 4446 { 4447 void *argptr; 4448 struct dm_ioctl *host_dm; 4449 abi_long guest_data; 4450 uint32_t guest_data_size; 4451 int target_size; 4452 const argtype *arg_type = ie->arg_type; 4453 abi_long ret; 4454 void *big_buf = NULL; 4455 char *host_data; 4456 4457 arg_type++; 4458 target_size = thunk_type_size(arg_type, 0); 4459 argptr = lock_user(VERIFY_READ, arg, target_size, 1); 4460 if (!argptr) { 4461 ret = -TARGET_EFAULT; 4462 goto out; 4463 } 4464 thunk_convert(buf_temp, argptr, arg_type, THUNK_HOST); 4465 unlock_user(argptr, arg, 0); 4466 4467 /* buf_temp is too small, so fetch things into a bigger buffer */ 4468 big_buf = g_malloc0(((struct dm_ioctl*)buf_temp)->data_size * 2); 4469 memcpy(big_buf, buf_temp, target_size); 4470 buf_temp = big_buf; 4471 host_dm = big_buf; 4472 4473 guest_data = arg + host_dm->data_start; 4474 if ((guest_data - arg) < 0) { 4475 ret = -TARGET_EINVAL; 4476 goto out; 4477 } 4478 guest_data_size = host_dm->data_size - host_dm->data_start; 4479 host_data = (char*)host_dm + host_dm->data_start; 4480 4481 argptr = lock_user(VERIFY_READ, guest_data, guest_data_size, 1); 4482 if (!argptr) { 4483 ret = -TARGET_EFAULT; 4484 goto out; 4485 } 4486 4487 switch (ie->host_cmd) { 4488 case DM_REMOVE_ALL: 4489 case DM_LIST_DEVICES: 4490 case DM_DEV_CREATE: 4491 case DM_DEV_REMOVE: 4492 case DM_DEV_SUSPEND: 4493 case DM_DEV_STATUS: 4494 case DM_DEV_WAIT: 4495 case DM_TABLE_STATUS: 4496 case DM_TABLE_CLEAR: 4497 case DM_TABLE_DEPS: 4498 case DM_LIST_VERSIONS: 4499 /* no input data */ 4500 break; 4501 case DM_DEV_RENAME: 4502 case DM_DEV_SET_GEOMETRY: 4503 /* data contains only strings */ 4504 memcpy(host_data, argptr, guest_data_size); 4505 break; 4506 case DM_TARGET_MSG: 4507 memcpy(host_data, argptr, guest_data_size); 4508 *(uint64_t*)host_data = tswap64(*(uint64_t*)argptr); 4509 break; 4510 case DM_TABLE_LOAD: 4511 { 4512 void *gspec = argptr; 4513 void *cur_data = host_data; 4514 const argtype arg_type[] = { MK_STRUCT(STRUCT_dm_target_spec) }; 4515 int spec_size = thunk_type_size(arg_type, 0); 4516 int i; 4517 4518 for (i = 0; i < host_dm->target_count; i++) { 4519 struct dm_target_spec *spec = cur_data; 4520 uint32_t next; 4521 int slen; 4522 4523 thunk_convert(spec, gspec, arg_type, THUNK_HOST); 4524 slen = strlen((char*)gspec + spec_size) + 1; 4525 next = spec->next; 4526 spec->next = sizeof(*spec) + slen; 4527 strcpy((char*)&spec[1], gspec + spec_size); 4528 gspec += next; 4529 cur_data += spec->next; 4530 } 4531 break; 4532 } 4533 default: 4534 ret = -TARGET_EINVAL; 4535 unlock_user(argptr, guest_data, 0); 4536 goto out; 4537 } 4538 unlock_user(argptr, guest_data, 0); 4539 4540 ret = get_errno(safe_ioctl(fd, ie->host_cmd, buf_temp)); 4541 if (!is_error(ret)) { 4542 guest_data = arg + host_dm->data_start; 4543 guest_data_size = host_dm->data_size - host_dm->data_start; 4544 argptr = lock_user(VERIFY_WRITE, guest_data, guest_data_size, 0); 4545 switch (ie->host_cmd) { 4546 case DM_REMOVE_ALL: 4547 case DM_DEV_CREATE: 4548 case DM_DEV_REMOVE: 4549 case DM_DEV_RENAME: 4550 case DM_DEV_SUSPEND: 4551 case DM_DEV_STATUS: 4552 case DM_TABLE_LOAD: 4553 case DM_TABLE_CLEAR: 4554 case DM_TARGET_MSG: 4555 case DM_DEV_SET_GEOMETRY: 4556 /* no return data */ 4557 break; 4558 case DM_LIST_DEVICES: 4559 { 4560 struct dm_name_list *nl = (void*)host_dm + host_dm->data_start; 4561 uint32_t remaining_data = guest_data_size; 4562 void *cur_data = argptr; 4563 const argtype arg_type[] = { MK_STRUCT(STRUCT_dm_name_list) }; 4564 int nl_size = 12; /* can't use thunk_size due to alignment */ 4565 4566 while (1) { 4567 uint32_t next = nl->next; 4568 if (next) { 4569 nl->next = nl_size + (strlen(nl->name) + 1); 4570 } 4571 if (remaining_data < nl->next) { 4572 host_dm->flags |= DM_BUFFER_FULL_FLAG; 4573 break; 4574 } 4575 thunk_convert(cur_data, nl, arg_type, THUNK_TARGET); 4576 strcpy(cur_data + nl_size, nl->name); 4577 cur_data += nl->next; 4578 remaining_data -= nl->next; 4579 if (!next) { 4580 break; 4581 } 4582 nl = (void*)nl + next; 4583 } 4584 break; 4585 } 4586 case DM_DEV_WAIT: 4587 case DM_TABLE_STATUS: 4588 { 4589 struct dm_target_spec *spec = (void*)host_dm + host_dm->data_start; 4590 void *cur_data = argptr; 4591 const argtype arg_type[] = { MK_STRUCT(STRUCT_dm_target_spec) }; 4592 int spec_size = thunk_type_size(arg_type, 0); 4593 int i; 4594 4595 for (i = 0; i < host_dm->target_count; i++) { 4596 uint32_t next = spec->next; 4597 int slen = strlen((char*)&spec[1]) + 1; 4598 spec->next = (cur_data - argptr) + spec_size + slen; 4599 if (guest_data_size < spec->next) { 4600 host_dm->flags |= DM_BUFFER_FULL_FLAG; 4601 break; 4602 } 4603 thunk_convert(cur_data, spec, arg_type, THUNK_TARGET); 4604 strcpy(cur_data + spec_size, (char*)&spec[1]); 4605 cur_data = argptr + spec->next; 4606 spec = (void*)host_dm + host_dm->data_start + next; 4607 } 4608 break; 4609 } 4610 case DM_TABLE_DEPS: 4611 { 4612 void *hdata = (void*)host_dm + host_dm->data_start; 4613 int count = *(uint32_t*)hdata; 4614 uint64_t *hdev = hdata + 8; 4615 uint64_t *gdev = argptr + 8; 4616 int i; 4617 4618 *(uint32_t*)argptr = tswap32(count); 4619 for (i = 0; i < count; i++) { 4620 *gdev = tswap64(*hdev); 4621 gdev++; 4622 hdev++; 4623 } 4624 break; 4625 } 4626 case DM_LIST_VERSIONS: 4627 { 4628 struct dm_target_versions *vers = (void*)host_dm + host_dm->data_start; 4629 uint32_t remaining_data = guest_data_size; 4630 void *cur_data = argptr; 4631 const argtype arg_type[] = { MK_STRUCT(STRUCT_dm_target_versions) }; 4632 int vers_size = thunk_type_size(arg_type, 0); 4633 4634 while (1) { 4635 uint32_t next = vers->next; 4636 if (next) { 4637 vers->next = vers_size + (strlen(vers->name) + 1); 4638 } 4639 if (remaining_data < vers->next) { 4640 host_dm->flags |= DM_BUFFER_FULL_FLAG; 4641 break; 4642 } 4643 thunk_convert(cur_data, vers, arg_type, THUNK_TARGET); 4644 strcpy(cur_data + vers_size, vers->name); 4645 cur_data += vers->next; 4646 remaining_data -= vers->next; 4647 if (!next) { 4648 break; 4649 } 4650 vers = (void*)vers + next; 4651 } 4652 break; 4653 } 4654 default: 4655 unlock_user(argptr, guest_data, 0); 4656 ret = -TARGET_EINVAL; 4657 goto out; 4658 } 4659 unlock_user(argptr, guest_data, guest_data_size); 4660 4661 argptr = lock_user(VERIFY_WRITE, arg, target_size, 0); 4662 if (!argptr) { 4663 ret = -TARGET_EFAULT; 4664 goto out; 4665 } 4666 thunk_convert(argptr, buf_temp, arg_type, THUNK_TARGET); 4667 unlock_user(argptr, arg, target_size); 4668 } 4669 out: 4670 g_free(big_buf); 4671 return ret; 4672 } 4673 4674 static abi_long do_ioctl_blkpg(const IOCTLEntry *ie, uint8_t *buf_temp, int fd, 4675 int cmd, abi_long arg) 4676 { 4677 void *argptr; 4678 int target_size; 4679 const argtype *arg_type = ie->arg_type; 4680 const argtype part_arg_type[] = { MK_STRUCT(STRUCT_blkpg_partition) }; 4681 abi_long ret; 4682 4683 struct blkpg_ioctl_arg *host_blkpg = (void*)buf_temp; 4684 struct blkpg_partition host_part; 4685 4686 /* Read and convert blkpg */ 4687 arg_type++; 4688 target_size = thunk_type_size(arg_type, 0); 4689 argptr = lock_user(VERIFY_READ, arg, target_size, 1); 4690 if (!argptr) { 4691 ret = -TARGET_EFAULT; 4692 goto out; 4693 } 4694 thunk_convert(buf_temp, argptr, arg_type, THUNK_HOST); 4695 unlock_user(argptr, arg, 0); 4696 4697 switch (host_blkpg->op) { 4698 case BLKPG_ADD_PARTITION: 4699 case BLKPG_DEL_PARTITION: 4700 /* payload is struct blkpg_partition */ 4701 break; 4702 default: 4703 /* Unknown opcode */ 4704 ret = -TARGET_EINVAL; 4705 goto out; 4706 } 4707 4708 /* Read and convert blkpg->data */ 4709 arg = (abi_long)(uintptr_t)host_blkpg->data; 4710 target_size = thunk_type_size(part_arg_type, 0); 4711 argptr = lock_user(VERIFY_READ, arg, target_size, 1); 4712 if (!argptr) { 4713 ret = -TARGET_EFAULT; 4714 goto out; 4715 } 4716 thunk_convert(&host_part, argptr, part_arg_type, THUNK_HOST); 4717 unlock_user(argptr, arg, 0); 4718 4719 /* Swizzle the data pointer to our local copy and call! */ 4720 host_blkpg->data = &host_part; 4721 ret = get_errno(safe_ioctl(fd, ie->host_cmd, host_blkpg)); 4722 4723 out: 4724 return ret; 4725 } 4726 4727 static abi_long do_ioctl_rt(const IOCTLEntry *ie, uint8_t *buf_temp, 4728 int fd, int cmd, abi_long arg) 4729 { 4730 const argtype *arg_type = ie->arg_type; 4731 const StructEntry *se; 4732 const argtype *field_types; 4733 const int *dst_offsets, *src_offsets; 4734 int target_size; 4735 void *argptr; 4736 abi_ulong *target_rt_dev_ptr = NULL; 4737 unsigned long *host_rt_dev_ptr = NULL; 4738 abi_long ret; 4739 int i; 4740 4741 assert(ie->access == IOC_W); 4742 assert(*arg_type == TYPE_PTR); 4743 arg_type++; 4744 assert(*arg_type == TYPE_STRUCT); 4745 target_size = thunk_type_size(arg_type, 0); 4746 argptr = lock_user(VERIFY_READ, arg, target_size, 1); 4747 if (!argptr) { 4748 return -TARGET_EFAULT; 4749 } 4750 arg_type++; 4751 assert(*arg_type == (int)STRUCT_rtentry); 4752 se = struct_entries + *arg_type++; 4753 assert(se->convert[0] == NULL); 4754 /* convert struct here to be able to catch rt_dev string */ 4755 field_types = se->field_types; 4756 dst_offsets = se->field_offsets[THUNK_HOST]; 4757 src_offsets = se->field_offsets[THUNK_TARGET]; 4758 for (i = 0; i < se->nb_fields; i++) { 4759 if (dst_offsets[i] == offsetof(struct rtentry, rt_dev)) { 4760 assert(*field_types == TYPE_PTRVOID); 4761 target_rt_dev_ptr = (abi_ulong *)(argptr + src_offsets[i]); 4762 host_rt_dev_ptr = (unsigned long *)(buf_temp + dst_offsets[i]); 4763 if (*target_rt_dev_ptr != 0) { 4764 *host_rt_dev_ptr = (unsigned long)lock_user_string( 4765 tswapal(*target_rt_dev_ptr)); 4766 if (!*host_rt_dev_ptr) { 4767 unlock_user(argptr, arg, 0); 4768 return -TARGET_EFAULT; 4769 } 4770 } else { 4771 *host_rt_dev_ptr = 0; 4772 } 4773 field_types++; 4774 continue; 4775 } 4776 field_types = thunk_convert(buf_temp + dst_offsets[i], 4777 argptr + src_offsets[i], 4778 field_types, THUNK_HOST); 4779 } 4780 unlock_user(argptr, arg, 0); 4781 4782 ret = get_errno(safe_ioctl(fd, ie->host_cmd, buf_temp)); 4783 4784 assert(host_rt_dev_ptr != NULL); 4785 assert(target_rt_dev_ptr != NULL); 4786 if (*host_rt_dev_ptr != 0) { 4787 unlock_user((void *)*host_rt_dev_ptr, 4788 *target_rt_dev_ptr, 0); 4789 } 4790 return ret; 4791 } 4792 4793 static abi_long do_ioctl_kdsigaccept(const IOCTLEntry *ie, uint8_t *buf_temp, 4794 int fd, int cmd, abi_long arg) 4795 { 4796 int sig = target_to_host_signal(arg); 4797 return get_errno(safe_ioctl(fd, ie->host_cmd, sig)); 4798 } 4799 4800 #ifdef TIOCGPTPEER 4801 static abi_long do_ioctl_tiocgptpeer(const IOCTLEntry *ie, uint8_t *buf_temp, 4802 int fd, int cmd, abi_long arg) 4803 { 4804 int flags = target_to_host_bitmask(arg, fcntl_flags_tbl); 4805 return get_errno(safe_ioctl(fd, ie->host_cmd, flags)); 4806 } 4807 #endif 4808 4809 static IOCTLEntry ioctl_entries[] = { 4810 #define IOCTL(cmd, access, ...) \ 4811 { TARGET_ ## cmd, cmd, #cmd, access, 0, { __VA_ARGS__ } }, 4812 #define IOCTL_SPECIAL(cmd, access, dofn, ...) \ 4813 { TARGET_ ## cmd, cmd, #cmd, access, dofn, { __VA_ARGS__ } }, 4814 #define IOCTL_IGNORE(cmd) \ 4815 { TARGET_ ## cmd, 0, #cmd }, 4816 #include "ioctls.h" 4817 { 0, 0, }, 4818 }; 4819 4820 /* ??? Implement proper locking for ioctls. */ 4821 /* do_ioctl() Must return target values and target errnos. */ 4822 static abi_long do_ioctl(int fd, int cmd, abi_long arg) 4823 { 4824 const IOCTLEntry *ie; 4825 const argtype *arg_type; 4826 abi_long ret; 4827 uint8_t buf_temp[MAX_STRUCT_SIZE]; 4828 int target_size; 4829 void *argptr; 4830 4831 ie = ioctl_entries; 4832 for(;;) { 4833 if (ie->target_cmd == 0) { 4834 gemu_log("Unsupported ioctl: cmd=0x%04lx\n", (long)cmd); 4835 return -TARGET_ENOSYS; 4836 } 4837 if (ie->target_cmd == cmd) 4838 break; 4839 ie++; 4840 } 4841 arg_type = ie->arg_type; 4842 if (ie->do_ioctl) { 4843 return ie->do_ioctl(ie, buf_temp, fd, cmd, arg); 4844 } else if (!ie->host_cmd) { 4845 /* Some architectures define BSD ioctls in their headers 4846 that are not implemented in Linux. */ 4847 return -TARGET_ENOSYS; 4848 } 4849 4850 switch(arg_type[0]) { 4851 case TYPE_NULL: 4852 /* no argument */ 4853 ret = get_errno(safe_ioctl(fd, ie->host_cmd)); 4854 break; 4855 case TYPE_PTRVOID: 4856 case TYPE_INT: 4857 ret = get_errno(safe_ioctl(fd, ie->host_cmd, arg)); 4858 break; 4859 case TYPE_PTR: 4860 arg_type++; 4861 target_size = thunk_type_size(arg_type, 0); 4862 switch(ie->access) { 4863 case IOC_R: 4864 ret = get_errno(safe_ioctl(fd, ie->host_cmd, buf_temp)); 4865 if (!is_error(ret)) { 4866 argptr = lock_user(VERIFY_WRITE, arg, target_size, 0); 4867 if (!argptr) 4868 return -TARGET_EFAULT; 4869 thunk_convert(argptr, buf_temp, arg_type, THUNK_TARGET); 4870 unlock_user(argptr, arg, target_size); 4871 } 4872 break; 4873 case IOC_W: 4874 argptr = lock_user(VERIFY_READ, arg, target_size, 1); 4875 if (!argptr) 4876 return -TARGET_EFAULT; 4877 thunk_convert(buf_temp, argptr, arg_type, THUNK_HOST); 4878 unlock_user(argptr, arg, 0); 4879 ret = get_errno(safe_ioctl(fd, ie->host_cmd, buf_temp)); 4880 break; 4881 default: 4882 case IOC_RW: 4883 argptr = lock_user(VERIFY_READ, arg, target_size, 1); 4884 if (!argptr) 4885 return -TARGET_EFAULT; 4886 thunk_convert(buf_temp, argptr, arg_type, THUNK_HOST); 4887 unlock_user(argptr, arg, 0); 4888 ret = get_errno(safe_ioctl(fd, ie->host_cmd, buf_temp)); 4889 if (!is_error(ret)) { 4890 argptr = lock_user(VERIFY_WRITE, arg, target_size, 0); 4891 if (!argptr) 4892 return -TARGET_EFAULT; 4893 thunk_convert(argptr, buf_temp, arg_type, THUNK_TARGET); 4894 unlock_user(argptr, arg, target_size); 4895 } 4896 break; 4897 } 4898 break; 4899 default: 4900 gemu_log("Unsupported ioctl type: cmd=0x%04lx type=%d\n", 4901 (long)cmd, arg_type[0]); 4902 ret = -TARGET_ENOSYS; 4903 break; 4904 } 4905 return ret; 4906 } 4907 4908 static const bitmask_transtbl iflag_tbl[] = { 4909 { TARGET_IGNBRK, TARGET_IGNBRK, IGNBRK, IGNBRK }, 4910 { TARGET_BRKINT, TARGET_BRKINT, BRKINT, BRKINT }, 4911 { TARGET_IGNPAR, TARGET_IGNPAR, IGNPAR, IGNPAR }, 4912 { TARGET_PARMRK, TARGET_PARMRK, PARMRK, PARMRK }, 4913 { TARGET_INPCK, TARGET_INPCK, INPCK, INPCK }, 4914 { TARGET_ISTRIP, TARGET_ISTRIP, ISTRIP, ISTRIP }, 4915 { TARGET_INLCR, TARGET_INLCR, INLCR, INLCR }, 4916 { TARGET_IGNCR, TARGET_IGNCR, IGNCR, IGNCR }, 4917 { TARGET_ICRNL, TARGET_ICRNL, ICRNL, ICRNL }, 4918 { TARGET_IUCLC, TARGET_IUCLC, IUCLC, IUCLC }, 4919 { TARGET_IXON, TARGET_IXON, IXON, IXON }, 4920 { TARGET_IXANY, TARGET_IXANY, IXANY, IXANY }, 4921 { TARGET_IXOFF, TARGET_IXOFF, IXOFF, IXOFF }, 4922 { TARGET_IMAXBEL, TARGET_IMAXBEL, IMAXBEL, IMAXBEL }, 4923 { 0, 0, 0, 0 } 4924 }; 4925 4926 static const bitmask_transtbl oflag_tbl[] = { 4927 { TARGET_OPOST, TARGET_OPOST, OPOST, OPOST }, 4928 { TARGET_OLCUC, TARGET_OLCUC, OLCUC, OLCUC }, 4929 { TARGET_ONLCR, TARGET_ONLCR, ONLCR, ONLCR }, 4930 { TARGET_OCRNL, TARGET_OCRNL, OCRNL, OCRNL }, 4931 { TARGET_ONOCR, TARGET_ONOCR, ONOCR, ONOCR }, 4932 { TARGET_ONLRET, TARGET_ONLRET, ONLRET, ONLRET }, 4933 { TARGET_OFILL, TARGET_OFILL, OFILL, OFILL }, 4934 { TARGET_OFDEL, TARGET_OFDEL, OFDEL, OFDEL }, 4935 { TARGET_NLDLY, TARGET_NL0, NLDLY, NL0 }, 4936 { TARGET_NLDLY, TARGET_NL1, NLDLY, NL1 }, 4937 { TARGET_CRDLY, TARGET_CR0, CRDLY, CR0 }, 4938 { TARGET_CRDLY, TARGET_CR1, CRDLY, CR1 }, 4939 { TARGET_CRDLY, TARGET_CR2, CRDLY, CR2 }, 4940 { TARGET_CRDLY, TARGET_CR3, CRDLY, CR3 }, 4941 { TARGET_TABDLY, TARGET_TAB0, TABDLY, TAB0 }, 4942 { TARGET_TABDLY, TARGET_TAB1, TABDLY, TAB1 }, 4943 { TARGET_TABDLY, TARGET_TAB2, TABDLY, TAB2 }, 4944 { TARGET_TABDLY, TARGET_TAB3, TABDLY, TAB3 }, 4945 { TARGET_BSDLY, TARGET_BS0, BSDLY, BS0 }, 4946 { TARGET_BSDLY, TARGET_BS1, BSDLY, BS1 }, 4947 { TARGET_VTDLY, TARGET_VT0, VTDLY, VT0 }, 4948 { TARGET_VTDLY, TARGET_VT1, VTDLY, VT1 }, 4949 { TARGET_FFDLY, TARGET_FF0, FFDLY, FF0 }, 4950 { TARGET_FFDLY, TARGET_FF1, FFDLY, FF1 }, 4951 { 0, 0, 0, 0 } 4952 }; 4953 4954 static const bitmask_transtbl cflag_tbl[] = { 4955 { TARGET_CBAUD, TARGET_B0, CBAUD, B0 }, 4956 { TARGET_CBAUD, TARGET_B50, CBAUD, B50 }, 4957 { TARGET_CBAUD, TARGET_B75, CBAUD, B75 }, 4958 { TARGET_CBAUD, TARGET_B110, CBAUD, B110 }, 4959 { TARGET_CBAUD, TARGET_B134, CBAUD, B134 }, 4960 { TARGET_CBAUD, TARGET_B150, CBAUD, B150 }, 4961 { TARGET_CBAUD, TARGET_B200, CBAUD, B200 }, 4962 { TARGET_CBAUD, TARGET_B300, CBAUD, B300 }, 4963 { TARGET_CBAUD, TARGET_B600, CBAUD, B600 }, 4964 { TARGET_CBAUD, TARGET_B1200, CBAUD, B1200 }, 4965 { TARGET_CBAUD, TARGET_B1800, CBAUD, B1800 }, 4966 { TARGET_CBAUD, TARGET_B2400, CBAUD, B2400 }, 4967 { TARGET_CBAUD, TARGET_B4800, CBAUD, B4800 }, 4968 { TARGET_CBAUD, TARGET_B9600, CBAUD, B9600 }, 4969 { TARGET_CBAUD, TARGET_B19200, CBAUD, B19200 }, 4970 { TARGET_CBAUD, TARGET_B38400, CBAUD, B38400 }, 4971 { TARGET_CBAUD, TARGET_B57600, CBAUD, B57600 }, 4972 { TARGET_CBAUD, TARGET_B115200, CBAUD, B115200 }, 4973 { TARGET_CBAUD, TARGET_B230400, CBAUD, B230400 }, 4974 { TARGET_CBAUD, TARGET_B460800, CBAUD, B460800 }, 4975 { TARGET_CSIZE, TARGET_CS5, CSIZE, CS5 }, 4976 { TARGET_CSIZE, TARGET_CS6, CSIZE, CS6 }, 4977 { TARGET_CSIZE, TARGET_CS7, CSIZE, CS7 }, 4978 { TARGET_CSIZE, TARGET_CS8, CSIZE, CS8 }, 4979 { TARGET_CSTOPB, TARGET_CSTOPB, CSTOPB, CSTOPB }, 4980 { TARGET_CREAD, TARGET_CREAD, CREAD, CREAD }, 4981 { TARGET_PARENB, TARGET_PARENB, PARENB, PARENB }, 4982 { TARGET_PARODD, TARGET_PARODD, PARODD, PARODD }, 4983 { TARGET_HUPCL, TARGET_HUPCL, HUPCL, HUPCL }, 4984 { TARGET_CLOCAL, TARGET_CLOCAL, CLOCAL, CLOCAL }, 4985 { TARGET_CRTSCTS, TARGET_CRTSCTS, CRTSCTS, CRTSCTS }, 4986 { 0, 0, 0, 0 } 4987 }; 4988 4989 static const bitmask_transtbl lflag_tbl[] = { 4990 { TARGET_ISIG, TARGET_ISIG, ISIG, ISIG }, 4991 { TARGET_ICANON, TARGET_ICANON, ICANON, ICANON }, 4992 { TARGET_XCASE, TARGET_XCASE, XCASE, XCASE }, 4993 { TARGET_ECHO, TARGET_ECHO, ECHO, ECHO }, 4994 { TARGET_ECHOE, TARGET_ECHOE, ECHOE, ECHOE }, 4995 { TARGET_ECHOK, TARGET_ECHOK, ECHOK, ECHOK }, 4996 { TARGET_ECHONL, TARGET_ECHONL, ECHONL, ECHONL }, 4997 { TARGET_NOFLSH, TARGET_NOFLSH, NOFLSH, NOFLSH }, 4998 { TARGET_TOSTOP, TARGET_TOSTOP, TOSTOP, TOSTOP }, 4999 { TARGET_ECHOCTL, TARGET_ECHOCTL, ECHOCTL, ECHOCTL }, 5000 { TARGET_ECHOPRT, TARGET_ECHOPRT, ECHOPRT, ECHOPRT }, 5001 { TARGET_ECHOKE, TARGET_ECHOKE, ECHOKE, ECHOKE }, 5002 { TARGET_FLUSHO, TARGET_FLUSHO, FLUSHO, FLUSHO }, 5003 { TARGET_PENDIN, TARGET_PENDIN, PENDIN, PENDIN }, 5004 { TARGET_IEXTEN, TARGET_IEXTEN, IEXTEN, IEXTEN }, 5005 { 0, 0, 0, 0 } 5006 }; 5007 5008 static void target_to_host_termios (void *dst, const void *src) 5009 { 5010 struct host_termios *host = dst; 5011 const struct target_termios *target = src; 5012 5013 host->c_iflag = 5014 target_to_host_bitmask(tswap32(target->c_iflag), iflag_tbl); 5015 host->c_oflag = 5016 target_to_host_bitmask(tswap32(target->c_oflag), oflag_tbl); 5017 host->c_cflag = 5018 target_to_host_bitmask(tswap32(target->c_cflag), cflag_tbl); 5019 host->c_lflag = 5020 target_to_host_bitmask(tswap32(target->c_lflag), lflag_tbl); 5021 host->c_line = target->c_line; 5022 5023 memset(host->c_cc, 0, sizeof(host->c_cc)); 5024 host->c_cc[VINTR] = target->c_cc[TARGET_VINTR]; 5025 host->c_cc[VQUIT] = target->c_cc[TARGET_VQUIT]; 5026 host->c_cc[VERASE] = target->c_cc[TARGET_VERASE]; 5027 host->c_cc[VKILL] = target->c_cc[TARGET_VKILL]; 5028 host->c_cc[VEOF] = target->c_cc[TARGET_VEOF]; 5029 host->c_cc[VTIME] = target->c_cc[TARGET_VTIME]; 5030 host->c_cc[VMIN] = target->c_cc[TARGET_VMIN]; 5031 host->c_cc[VSWTC] = target->c_cc[TARGET_VSWTC]; 5032 host->c_cc[VSTART] = target->c_cc[TARGET_VSTART]; 5033 host->c_cc[VSTOP] = target->c_cc[TARGET_VSTOP]; 5034 host->c_cc[VSUSP] = target->c_cc[TARGET_VSUSP]; 5035 host->c_cc[VEOL] = target->c_cc[TARGET_VEOL]; 5036 host->c_cc[VREPRINT] = target->c_cc[TARGET_VREPRINT]; 5037 host->c_cc[VDISCARD] = target->c_cc[TARGET_VDISCARD]; 5038 host->c_cc[VWERASE] = target->c_cc[TARGET_VWERASE]; 5039 host->c_cc[VLNEXT] = target->c_cc[TARGET_VLNEXT]; 5040 host->c_cc[VEOL2] = target->c_cc[TARGET_VEOL2]; 5041 } 5042 5043 static void host_to_target_termios (void *dst, const void *src) 5044 { 5045 struct target_termios *target = dst; 5046 const struct host_termios *host = src; 5047 5048 target->c_iflag = 5049 tswap32(host_to_target_bitmask(host->c_iflag, iflag_tbl)); 5050 target->c_oflag = 5051 tswap32(host_to_target_bitmask(host->c_oflag, oflag_tbl)); 5052 target->c_cflag = 5053 tswap32(host_to_target_bitmask(host->c_cflag, cflag_tbl)); 5054 target->c_lflag = 5055 tswap32(host_to_target_bitmask(host->c_lflag, lflag_tbl)); 5056 target->c_line = host->c_line; 5057 5058 memset(target->c_cc, 0, sizeof(target->c_cc)); 5059 target->c_cc[TARGET_VINTR] = host->c_cc[VINTR]; 5060 target->c_cc[TARGET_VQUIT] = host->c_cc[VQUIT]; 5061 target->c_cc[TARGET_VERASE] = host->c_cc[VERASE]; 5062 target->c_cc[TARGET_VKILL] = host->c_cc[VKILL]; 5063 target->c_cc[TARGET_VEOF] = host->c_cc[VEOF]; 5064 target->c_cc[TARGET_VTIME] = host->c_cc[VTIME]; 5065 target->c_cc[TARGET_VMIN] = host->c_cc[VMIN]; 5066 target->c_cc[TARGET_VSWTC] = host->c_cc[VSWTC]; 5067 target->c_cc[TARGET_VSTART] = host->c_cc[VSTART]; 5068 target->c_cc[TARGET_VSTOP] = host->c_cc[VSTOP]; 5069 target->c_cc[TARGET_VSUSP] = host->c_cc[VSUSP]; 5070 target->c_cc[TARGET_VEOL] = host->c_cc[VEOL]; 5071 target->c_cc[TARGET_VREPRINT] = host->c_cc[VREPRINT]; 5072 target->c_cc[TARGET_VDISCARD] = host->c_cc[VDISCARD]; 5073 target->c_cc[TARGET_VWERASE] = host->c_cc[VWERASE]; 5074 target->c_cc[TARGET_VLNEXT] = host->c_cc[VLNEXT]; 5075 target->c_cc[TARGET_VEOL2] = host->c_cc[VEOL2]; 5076 } 5077 5078 static const StructEntry struct_termios_def = { 5079 .convert = { host_to_target_termios, target_to_host_termios }, 5080 .size = { sizeof(struct target_termios), sizeof(struct host_termios) }, 5081 .align = { __alignof__(struct target_termios), __alignof__(struct host_termios) }, 5082 }; 5083 5084 static bitmask_transtbl mmap_flags_tbl[] = { 5085 { TARGET_MAP_SHARED, TARGET_MAP_SHARED, MAP_SHARED, MAP_SHARED }, 5086 { TARGET_MAP_PRIVATE, TARGET_MAP_PRIVATE, MAP_PRIVATE, MAP_PRIVATE }, 5087 { TARGET_MAP_FIXED, TARGET_MAP_FIXED, MAP_FIXED, MAP_FIXED }, 5088 { TARGET_MAP_ANONYMOUS, TARGET_MAP_ANONYMOUS, 5089 MAP_ANONYMOUS, MAP_ANONYMOUS }, 5090 { TARGET_MAP_GROWSDOWN, TARGET_MAP_GROWSDOWN, 5091 MAP_GROWSDOWN, MAP_GROWSDOWN }, 5092 { TARGET_MAP_DENYWRITE, TARGET_MAP_DENYWRITE, 5093 MAP_DENYWRITE, MAP_DENYWRITE }, 5094 { TARGET_MAP_EXECUTABLE, TARGET_MAP_EXECUTABLE, 5095 MAP_EXECUTABLE, MAP_EXECUTABLE }, 5096 { TARGET_MAP_LOCKED, TARGET_MAP_LOCKED, MAP_LOCKED, MAP_LOCKED }, 5097 { TARGET_MAP_NORESERVE, TARGET_MAP_NORESERVE, 5098 MAP_NORESERVE, MAP_NORESERVE }, 5099 { TARGET_MAP_HUGETLB, TARGET_MAP_HUGETLB, MAP_HUGETLB, MAP_HUGETLB }, 5100 /* MAP_STACK had been ignored by the kernel for quite some time. 5101 Recognize it for the target insofar as we do not want to pass 5102 it through to the host. */ 5103 { TARGET_MAP_STACK, TARGET_MAP_STACK, 0, 0 }, 5104 { 0, 0, 0, 0 } 5105 }; 5106 5107 #if defined(TARGET_I386) 5108 5109 /* NOTE: there is really one LDT for all the threads */ 5110 static uint8_t *ldt_table; 5111 5112 static abi_long read_ldt(abi_ulong ptr, unsigned long bytecount) 5113 { 5114 int size; 5115 void *p; 5116 5117 if (!ldt_table) 5118 return 0; 5119 size = TARGET_LDT_ENTRIES * TARGET_LDT_ENTRY_SIZE; 5120 if (size > bytecount) 5121 size = bytecount; 5122 p = lock_user(VERIFY_WRITE, ptr, size, 0); 5123 if (!p) 5124 return -TARGET_EFAULT; 5125 /* ??? Should this by byteswapped? */ 5126 memcpy(p, ldt_table, size); 5127 unlock_user(p, ptr, size); 5128 return size; 5129 } 5130 5131 /* XXX: add locking support */ 5132 static abi_long write_ldt(CPUX86State *env, 5133 abi_ulong ptr, unsigned long bytecount, int oldmode) 5134 { 5135 struct target_modify_ldt_ldt_s ldt_info; 5136 struct target_modify_ldt_ldt_s *target_ldt_info; 5137 int seg_32bit, contents, read_exec_only, limit_in_pages; 5138 int seg_not_present, useable, lm; 5139 uint32_t *lp, entry_1, entry_2; 5140 5141 if (bytecount != sizeof(ldt_info)) 5142 return -TARGET_EINVAL; 5143 if (!lock_user_struct(VERIFY_READ, target_ldt_info, ptr, 1)) 5144 return -TARGET_EFAULT; 5145 ldt_info.entry_number = tswap32(target_ldt_info->entry_number); 5146 ldt_info.base_addr = tswapal(target_ldt_info->base_addr); 5147 ldt_info.limit = tswap32(target_ldt_info->limit); 5148 ldt_info.flags = tswap32(target_ldt_info->flags); 5149 unlock_user_struct(target_ldt_info, ptr, 0); 5150 5151 if (ldt_info.entry_number >= TARGET_LDT_ENTRIES) 5152 return -TARGET_EINVAL; 5153 seg_32bit = ldt_info.flags & 1; 5154 contents = (ldt_info.flags >> 1) & 3; 5155 read_exec_only = (ldt_info.flags >> 3) & 1; 5156 limit_in_pages = (ldt_info.flags >> 4) & 1; 5157 seg_not_present = (ldt_info.flags >> 5) & 1; 5158 useable = (ldt_info.flags >> 6) & 1; 5159 #ifdef TARGET_ABI32 5160 lm = 0; 5161 #else 5162 lm = (ldt_info.flags >> 7) & 1; 5163 #endif 5164 if (contents == 3) { 5165 if (oldmode) 5166 return -TARGET_EINVAL; 5167 if (seg_not_present == 0) 5168 return -TARGET_EINVAL; 5169 } 5170 /* allocate the LDT */ 5171 if (!ldt_table) { 5172 env->ldt.base = target_mmap(0, 5173 TARGET_LDT_ENTRIES * TARGET_LDT_ENTRY_SIZE, 5174 PROT_READ|PROT_WRITE, 5175 MAP_ANONYMOUS|MAP_PRIVATE, -1, 0); 5176 if (env->ldt.base == -1) 5177 return -TARGET_ENOMEM; 5178 memset(g2h(env->ldt.base), 0, 5179 TARGET_LDT_ENTRIES * TARGET_LDT_ENTRY_SIZE); 5180 env->ldt.limit = 0xffff; 5181 ldt_table = g2h(env->ldt.base); 5182 } 5183 5184 /* NOTE: same code as Linux kernel */ 5185 /* Allow LDTs to be cleared by the user. */ 5186 if (ldt_info.base_addr == 0 && ldt_info.limit == 0) { 5187 if (oldmode || 5188 (contents == 0 && 5189 read_exec_only == 1 && 5190 seg_32bit == 0 && 5191 limit_in_pages == 0 && 5192 seg_not_present == 1 && 5193 useable == 0 )) { 5194 entry_1 = 0; 5195 entry_2 = 0; 5196 goto install; 5197 } 5198 } 5199 5200 entry_1 = ((ldt_info.base_addr & 0x0000ffff) << 16) | 5201 (ldt_info.limit & 0x0ffff); 5202 entry_2 = (ldt_info.base_addr & 0xff000000) | 5203 ((ldt_info.base_addr & 0x00ff0000) >> 16) | 5204 (ldt_info.limit & 0xf0000) | 5205 ((read_exec_only ^ 1) << 9) | 5206 (contents << 10) | 5207 ((seg_not_present ^ 1) << 15) | 5208 (seg_32bit << 22) | 5209 (limit_in_pages << 23) | 5210 (lm << 21) | 5211 0x7000; 5212 if (!oldmode) 5213 entry_2 |= (useable << 20); 5214 5215 /* Install the new entry ... */ 5216 install: 5217 lp = (uint32_t *)(ldt_table + (ldt_info.entry_number << 3)); 5218 lp[0] = tswap32(entry_1); 5219 lp[1] = tswap32(entry_2); 5220 return 0; 5221 } 5222 5223 /* specific and weird i386 syscalls */ 5224 static abi_long do_modify_ldt(CPUX86State *env, int func, abi_ulong ptr, 5225 unsigned long bytecount) 5226 { 5227 abi_long ret; 5228 5229 switch (func) { 5230 case 0: 5231 ret = read_ldt(ptr, bytecount); 5232 break; 5233 case 1: 5234 ret = write_ldt(env, ptr, bytecount, 1); 5235 break; 5236 case 0x11: 5237 ret = write_ldt(env, ptr, bytecount, 0); 5238 break; 5239 default: 5240 ret = -TARGET_ENOSYS; 5241 break; 5242 } 5243 return ret; 5244 } 5245 5246 #if defined(TARGET_I386) && defined(TARGET_ABI32) 5247 abi_long do_set_thread_area(CPUX86State *env, abi_ulong ptr) 5248 { 5249 uint64_t *gdt_table = g2h(env->gdt.base); 5250 struct target_modify_ldt_ldt_s ldt_info; 5251 struct target_modify_ldt_ldt_s *target_ldt_info; 5252 int seg_32bit, contents, read_exec_only, limit_in_pages; 5253 int seg_not_present, useable, lm; 5254 uint32_t *lp, entry_1, entry_2; 5255 int i; 5256 5257 lock_user_struct(VERIFY_WRITE, target_ldt_info, ptr, 1); 5258 if (!target_ldt_info) 5259 return -TARGET_EFAULT; 5260 ldt_info.entry_number = tswap32(target_ldt_info->entry_number); 5261 ldt_info.base_addr = tswapal(target_ldt_info->base_addr); 5262 ldt_info.limit = tswap32(target_ldt_info->limit); 5263 ldt_info.flags = tswap32(target_ldt_info->flags); 5264 if (ldt_info.entry_number == -1) { 5265 for (i=TARGET_GDT_ENTRY_TLS_MIN; i<=TARGET_GDT_ENTRY_TLS_MAX; i++) { 5266 if (gdt_table[i] == 0) { 5267 ldt_info.entry_number = i; 5268 target_ldt_info->entry_number = tswap32(i); 5269 break; 5270 } 5271 } 5272 } 5273 unlock_user_struct(target_ldt_info, ptr, 1); 5274 5275 if (ldt_info.entry_number < TARGET_GDT_ENTRY_TLS_MIN || 5276 ldt_info.entry_number > TARGET_GDT_ENTRY_TLS_MAX) 5277 return -TARGET_EINVAL; 5278 seg_32bit = ldt_info.flags & 1; 5279 contents = (ldt_info.flags >> 1) & 3; 5280 read_exec_only = (ldt_info.flags >> 3) & 1; 5281 limit_in_pages = (ldt_info.flags >> 4) & 1; 5282 seg_not_present = (ldt_info.flags >> 5) & 1; 5283 useable = (ldt_info.flags >> 6) & 1; 5284 #ifdef TARGET_ABI32 5285 lm = 0; 5286 #else 5287 lm = (ldt_info.flags >> 7) & 1; 5288 #endif 5289 5290 if (contents == 3) { 5291 if (seg_not_present == 0) 5292 return -TARGET_EINVAL; 5293 } 5294 5295 /* NOTE: same code as Linux kernel */ 5296 /* Allow LDTs to be cleared by the user. */ 5297 if (ldt_info.base_addr == 0 && ldt_info.limit == 0) { 5298 if ((contents == 0 && 5299 read_exec_only == 1 && 5300 seg_32bit == 0 && 5301 limit_in_pages == 0 && 5302 seg_not_present == 1 && 5303 useable == 0 )) { 5304 entry_1 = 0; 5305 entry_2 = 0; 5306 goto install; 5307 } 5308 } 5309 5310 entry_1 = ((ldt_info.base_addr & 0x0000ffff) << 16) | 5311 (ldt_info.limit & 0x0ffff); 5312 entry_2 = (ldt_info.base_addr & 0xff000000) | 5313 ((ldt_info.base_addr & 0x00ff0000) >> 16) | 5314 (ldt_info.limit & 0xf0000) | 5315 ((read_exec_only ^ 1) << 9) | 5316 (contents << 10) | 5317 ((seg_not_present ^ 1) << 15) | 5318 (seg_32bit << 22) | 5319 (limit_in_pages << 23) | 5320 (useable << 20) | 5321 (lm << 21) | 5322 0x7000; 5323 5324 /* Install the new entry ... */ 5325 install: 5326 lp = (uint32_t *)(gdt_table + ldt_info.entry_number); 5327 lp[0] = tswap32(entry_1); 5328 lp[1] = tswap32(entry_2); 5329 return 0; 5330 } 5331 5332 static abi_long do_get_thread_area(CPUX86State *env, abi_ulong ptr) 5333 { 5334 struct target_modify_ldt_ldt_s *target_ldt_info; 5335 uint64_t *gdt_table = g2h(env->gdt.base); 5336 uint32_t base_addr, limit, flags; 5337 int seg_32bit, contents, read_exec_only, limit_in_pages, idx; 5338 int seg_not_present, useable, lm; 5339 uint32_t *lp, entry_1, entry_2; 5340 5341 lock_user_struct(VERIFY_WRITE, target_ldt_info, ptr, 1); 5342 if (!target_ldt_info) 5343 return -TARGET_EFAULT; 5344 idx = tswap32(target_ldt_info->entry_number); 5345 if (idx < TARGET_GDT_ENTRY_TLS_MIN || 5346 idx > TARGET_GDT_ENTRY_TLS_MAX) { 5347 unlock_user_struct(target_ldt_info, ptr, 1); 5348 return -TARGET_EINVAL; 5349 } 5350 lp = (uint32_t *)(gdt_table + idx); 5351 entry_1 = tswap32(lp[0]); 5352 entry_2 = tswap32(lp[1]); 5353 5354 read_exec_only = ((entry_2 >> 9) & 1) ^ 1; 5355 contents = (entry_2 >> 10) & 3; 5356 seg_not_present = ((entry_2 >> 15) & 1) ^ 1; 5357 seg_32bit = (entry_2 >> 22) & 1; 5358 limit_in_pages = (entry_2 >> 23) & 1; 5359 useable = (entry_2 >> 20) & 1; 5360 #ifdef TARGET_ABI32 5361 lm = 0; 5362 #else 5363 lm = (entry_2 >> 21) & 1; 5364 #endif 5365 flags = (seg_32bit << 0) | (contents << 1) | 5366 (read_exec_only << 3) | (limit_in_pages << 4) | 5367 (seg_not_present << 5) | (useable << 6) | (lm << 7); 5368 limit = (entry_1 & 0xffff) | (entry_2 & 0xf0000); 5369 base_addr = (entry_1 >> 16) | 5370 (entry_2 & 0xff000000) | 5371 ((entry_2 & 0xff) << 16); 5372 target_ldt_info->base_addr = tswapal(base_addr); 5373 target_ldt_info->limit = tswap32(limit); 5374 target_ldt_info->flags = tswap32(flags); 5375 unlock_user_struct(target_ldt_info, ptr, 1); 5376 return 0; 5377 } 5378 #endif /* TARGET_I386 && TARGET_ABI32 */ 5379 5380 #ifndef TARGET_ABI32 5381 abi_long do_arch_prctl(CPUX86State *env, int code, abi_ulong addr) 5382 { 5383 abi_long ret = 0; 5384 abi_ulong val; 5385 int idx; 5386 5387 switch(code) { 5388 case TARGET_ARCH_SET_GS: 5389 case TARGET_ARCH_SET_FS: 5390 if (code == TARGET_ARCH_SET_GS) 5391 idx = R_GS; 5392 else 5393 idx = R_FS; 5394 cpu_x86_load_seg(env, idx, 0); 5395 env->segs[idx].base = addr; 5396 break; 5397 case TARGET_ARCH_GET_GS: 5398 case TARGET_ARCH_GET_FS: 5399 if (code == TARGET_ARCH_GET_GS) 5400 idx = R_GS; 5401 else 5402 idx = R_FS; 5403 val = env->segs[idx].base; 5404 if (put_user(val, addr, abi_ulong)) 5405 ret = -TARGET_EFAULT; 5406 break; 5407 default: 5408 ret = -TARGET_EINVAL; 5409 break; 5410 } 5411 return ret; 5412 } 5413 #endif 5414 5415 #endif /* defined(TARGET_I386) */ 5416 5417 #define NEW_STACK_SIZE 0x40000 5418 5419 5420 static pthread_mutex_t clone_lock = PTHREAD_MUTEX_INITIALIZER; 5421 typedef struct { 5422 CPUArchState *env; 5423 pthread_mutex_t mutex; 5424 pthread_cond_t cond; 5425 pthread_t thread; 5426 uint32_t tid; 5427 abi_ulong child_tidptr; 5428 abi_ulong parent_tidptr; 5429 sigset_t sigmask; 5430 } new_thread_info; 5431 5432 static void *clone_func(void *arg) 5433 { 5434 new_thread_info *info = arg; 5435 CPUArchState *env; 5436 CPUState *cpu; 5437 TaskState *ts; 5438 5439 rcu_register_thread(); 5440 tcg_register_thread(); 5441 env = info->env; 5442 cpu = ENV_GET_CPU(env); 5443 thread_cpu = cpu; 5444 ts = (TaskState *)cpu->opaque; 5445 info->tid = gettid(); 5446 task_settid(ts); 5447 if (info->child_tidptr) 5448 put_user_u32(info->tid, info->child_tidptr); 5449 if (info->parent_tidptr) 5450 put_user_u32(info->tid, info->parent_tidptr); 5451 /* Enable signals. */ 5452 sigprocmask(SIG_SETMASK, &info->sigmask, NULL); 5453 /* Signal to the parent that we're ready. */ 5454 pthread_mutex_lock(&info->mutex); 5455 pthread_cond_broadcast(&info->cond); 5456 pthread_mutex_unlock(&info->mutex); 5457 /* Wait until the parent has finished initializing the tls state. */ 5458 pthread_mutex_lock(&clone_lock); 5459 pthread_mutex_unlock(&clone_lock); 5460 cpu_loop(env); 5461 /* never exits */ 5462 return NULL; 5463 } 5464 5465 /* do_fork() Must return host values and target errnos (unlike most 5466 do_*() functions). */ 5467 static int do_fork(CPUArchState *env, unsigned int flags, abi_ulong newsp, 5468 abi_ulong parent_tidptr, target_ulong newtls, 5469 abi_ulong child_tidptr) 5470 { 5471 CPUState *cpu = ENV_GET_CPU(env); 5472 int ret; 5473 TaskState *ts; 5474 CPUState *new_cpu; 5475 CPUArchState *new_env; 5476 sigset_t sigmask; 5477 5478 flags &= ~CLONE_IGNORED_FLAGS; 5479 5480 /* Emulate vfork() with fork() */ 5481 if (flags & CLONE_VFORK) 5482 flags &= ~(CLONE_VFORK | CLONE_VM); 5483 5484 if (flags & CLONE_VM) { 5485 TaskState *parent_ts = (TaskState *)cpu->opaque; 5486 new_thread_info info; 5487 pthread_attr_t attr; 5488 5489 if (((flags & CLONE_THREAD_FLAGS) != CLONE_THREAD_FLAGS) || 5490 (flags & CLONE_INVALID_THREAD_FLAGS)) { 5491 return -TARGET_EINVAL; 5492 } 5493 5494 ts = g_new0(TaskState, 1); 5495 init_task_state(ts); 5496 5497 /* Grab a mutex so that thread setup appears atomic. */ 5498 pthread_mutex_lock(&clone_lock); 5499 5500 /* we create a new CPU instance. */ 5501 new_env = cpu_copy(env); 5502 /* Init regs that differ from the parent. */ 5503 cpu_clone_regs(new_env, newsp); 5504 new_cpu = ENV_GET_CPU(new_env); 5505 new_cpu->opaque = ts; 5506 ts->bprm = parent_ts->bprm; 5507 ts->info = parent_ts->info; 5508 ts->signal_mask = parent_ts->signal_mask; 5509 5510 if (flags & CLONE_CHILD_CLEARTID) { 5511 ts->child_tidptr = child_tidptr; 5512 } 5513 5514 if (flags & CLONE_SETTLS) { 5515 cpu_set_tls (new_env, newtls); 5516 } 5517 5518 memset(&info, 0, sizeof(info)); 5519 pthread_mutex_init(&info.mutex, NULL); 5520 pthread_mutex_lock(&info.mutex); 5521 pthread_cond_init(&info.cond, NULL); 5522 info.env = new_env; 5523 if (flags & CLONE_CHILD_SETTID) { 5524 info.child_tidptr = child_tidptr; 5525 } 5526 if (flags & CLONE_PARENT_SETTID) { 5527 info.parent_tidptr = parent_tidptr; 5528 } 5529 5530 ret = pthread_attr_init(&attr); 5531 ret = pthread_attr_setstacksize(&attr, NEW_STACK_SIZE); 5532 ret = pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_DETACHED); 5533 /* It is not safe to deliver signals until the child has finished 5534 initializing, so temporarily block all signals. */ 5535 sigfillset(&sigmask); 5536 sigprocmask(SIG_BLOCK, &sigmask, &info.sigmask); 5537 5538 /* If this is our first additional thread, we need to ensure we 5539 * generate code for parallel execution and flush old translations. 5540 */ 5541 if (!parallel_cpus) { 5542 parallel_cpus = true; 5543 tb_flush(cpu); 5544 } 5545 5546 ret = pthread_create(&info.thread, &attr, clone_func, &info); 5547 /* TODO: Free new CPU state if thread creation failed. */ 5548 5549 sigprocmask(SIG_SETMASK, &info.sigmask, NULL); 5550 pthread_attr_destroy(&attr); 5551 if (ret == 0) { 5552 /* Wait for the child to initialize. */ 5553 pthread_cond_wait(&info.cond, &info.mutex); 5554 ret = info.tid; 5555 } else { 5556 ret = -1; 5557 } 5558 pthread_mutex_unlock(&info.mutex); 5559 pthread_cond_destroy(&info.cond); 5560 pthread_mutex_destroy(&info.mutex); 5561 pthread_mutex_unlock(&clone_lock); 5562 } else { 5563 /* if no CLONE_VM, we consider it is a fork */ 5564 if (flags & CLONE_INVALID_FORK_FLAGS) { 5565 return -TARGET_EINVAL; 5566 } 5567 5568 /* We can't support custom termination signals */ 5569 if ((flags & CSIGNAL) != TARGET_SIGCHLD) { 5570 return -TARGET_EINVAL; 5571 } 5572 5573 if (block_signals()) { 5574 return -TARGET_ERESTARTSYS; 5575 } 5576 5577 fork_start(); 5578 ret = fork(); 5579 if (ret == 0) { 5580 /* Child Process. */ 5581 cpu_clone_regs(env, newsp); 5582 fork_end(1); 5583 /* There is a race condition here. The parent process could 5584 theoretically read the TID in the child process before the child 5585 tid is set. This would require using either ptrace 5586 (not implemented) or having *_tidptr to point at a shared memory 5587 mapping. We can't repeat the spinlock hack used above because 5588 the child process gets its own copy of the lock. */ 5589 if (flags & CLONE_CHILD_SETTID) 5590 put_user_u32(gettid(), child_tidptr); 5591 if (flags & CLONE_PARENT_SETTID) 5592 put_user_u32(gettid(), parent_tidptr); 5593 ts = (TaskState *)cpu->opaque; 5594 if (flags & CLONE_SETTLS) 5595 cpu_set_tls (env, newtls); 5596 if (flags & CLONE_CHILD_CLEARTID) 5597 ts->child_tidptr = child_tidptr; 5598 } else { 5599 fork_end(0); 5600 } 5601 } 5602 return ret; 5603 } 5604 5605 /* warning : doesn't handle linux specific flags... */ 5606 static int target_to_host_fcntl_cmd(int cmd) 5607 { 5608 int ret; 5609 5610 switch(cmd) { 5611 case TARGET_F_DUPFD: 5612 case TARGET_F_GETFD: 5613 case TARGET_F_SETFD: 5614 case TARGET_F_GETFL: 5615 case TARGET_F_SETFL: 5616 ret = cmd; 5617 break; 5618 case TARGET_F_GETLK: 5619 ret = F_GETLK64; 5620 break; 5621 case TARGET_F_SETLK: 5622 ret = F_SETLK64; 5623 break; 5624 case TARGET_F_SETLKW: 5625 ret = F_SETLKW64; 5626 break; 5627 case TARGET_F_GETOWN: 5628 ret = F_GETOWN; 5629 break; 5630 case TARGET_F_SETOWN: 5631 ret = F_SETOWN; 5632 break; 5633 case TARGET_F_GETSIG: 5634 ret = F_GETSIG; 5635 break; 5636 case TARGET_F_SETSIG: 5637 ret = F_SETSIG; 5638 break; 5639 #if TARGET_ABI_BITS == 32 5640 case TARGET_F_GETLK64: 5641 ret = F_GETLK64; 5642 break; 5643 case TARGET_F_SETLK64: 5644 ret = F_SETLK64; 5645 break; 5646 case TARGET_F_SETLKW64: 5647 ret = F_SETLKW64; 5648 break; 5649 #endif 5650 case TARGET_F_SETLEASE: 5651 ret = F_SETLEASE; 5652 break; 5653 case TARGET_F_GETLEASE: 5654 ret = F_GETLEASE; 5655 break; 5656 #ifdef F_DUPFD_CLOEXEC 5657 case TARGET_F_DUPFD_CLOEXEC: 5658 ret = F_DUPFD_CLOEXEC; 5659 break; 5660 #endif 5661 case TARGET_F_NOTIFY: 5662 ret = F_NOTIFY; 5663 break; 5664 #ifdef F_GETOWN_EX 5665 case TARGET_F_GETOWN_EX: 5666 ret = F_GETOWN_EX; 5667 break; 5668 #endif 5669 #ifdef F_SETOWN_EX 5670 case TARGET_F_SETOWN_EX: 5671 ret = F_SETOWN_EX; 5672 break; 5673 #endif 5674 #ifdef F_SETPIPE_SZ 5675 case TARGET_F_SETPIPE_SZ: 5676 ret = F_SETPIPE_SZ; 5677 break; 5678 case TARGET_F_GETPIPE_SZ: 5679 ret = F_GETPIPE_SZ; 5680 break; 5681 #endif 5682 default: 5683 ret = -TARGET_EINVAL; 5684 break; 5685 } 5686 5687 #if defined(__powerpc64__) 5688 /* On PPC64, glibc headers has the F_*LK* defined to 12, 13 and 14 and 5689 * is not supported by kernel. The glibc fcntl call actually adjusts 5690 * them to 5, 6 and 7 before making the syscall(). Since we make the 5691 * syscall directly, adjust to what is supported by the kernel. 5692 */ 5693 if (ret >= F_GETLK64 && ret <= F_SETLKW64) { 5694 ret -= F_GETLK64 - 5; 5695 } 5696 #endif 5697 5698 return ret; 5699 } 5700 5701 #define FLOCK_TRANSTBL \ 5702 switch (type) { \ 5703 TRANSTBL_CONVERT(F_RDLCK); \ 5704 TRANSTBL_CONVERT(F_WRLCK); \ 5705 TRANSTBL_CONVERT(F_UNLCK); \ 5706 TRANSTBL_CONVERT(F_EXLCK); \ 5707 TRANSTBL_CONVERT(F_SHLCK); \ 5708 } 5709 5710 static int target_to_host_flock(int type) 5711 { 5712 #define TRANSTBL_CONVERT(a) case TARGET_##a: return a 5713 FLOCK_TRANSTBL 5714 #undef TRANSTBL_CONVERT 5715 return -TARGET_EINVAL; 5716 } 5717 5718 static int host_to_target_flock(int type) 5719 { 5720 #define TRANSTBL_CONVERT(a) case a: return TARGET_##a 5721 FLOCK_TRANSTBL 5722 #undef TRANSTBL_CONVERT 5723 /* if we don't know how to convert the value coming 5724 * from the host we copy to the target field as-is 5725 */ 5726 return type; 5727 } 5728 5729 static inline abi_long copy_from_user_flock(struct flock64 *fl, 5730 abi_ulong target_flock_addr) 5731 { 5732 struct target_flock *target_fl; 5733 int l_type; 5734 5735 if (!lock_user_struct(VERIFY_READ, target_fl, target_flock_addr, 1)) { 5736 return -TARGET_EFAULT; 5737 } 5738 5739 __get_user(l_type, &target_fl->l_type); 5740 l_type = target_to_host_flock(l_type); 5741 if (l_type < 0) { 5742 return l_type; 5743 } 5744 fl->l_type = l_type; 5745 __get_user(fl->l_whence, &target_fl->l_whence); 5746 __get_user(fl->l_start, &target_fl->l_start); 5747 __get_user(fl->l_len, &target_fl->l_len); 5748 __get_user(fl->l_pid, &target_fl->l_pid); 5749 unlock_user_struct(target_fl, target_flock_addr, 0); 5750 return 0; 5751 } 5752 5753 static inline abi_long copy_to_user_flock(abi_ulong target_flock_addr, 5754 const struct flock64 *fl) 5755 { 5756 struct target_flock *target_fl; 5757 short l_type; 5758 5759 if (!lock_user_struct(VERIFY_WRITE, target_fl, target_flock_addr, 0)) { 5760 return -TARGET_EFAULT; 5761 } 5762 5763 l_type = host_to_target_flock(fl->l_type); 5764 __put_user(l_type, &target_fl->l_type); 5765 __put_user(fl->l_whence, &target_fl->l_whence); 5766 __put_user(fl->l_start, &target_fl->l_start); 5767 __put_user(fl->l_len, &target_fl->l_len); 5768 __put_user(fl->l_pid, &target_fl->l_pid); 5769 unlock_user_struct(target_fl, target_flock_addr, 1); 5770 return 0; 5771 } 5772 5773 typedef abi_long from_flock64_fn(struct flock64 *fl, abi_ulong target_addr); 5774 typedef abi_long to_flock64_fn(abi_ulong target_addr, const struct flock64 *fl); 5775 5776 #if defined(TARGET_ARM) && TARGET_ABI_BITS == 32 5777 static inline abi_long copy_from_user_oabi_flock64(struct flock64 *fl, 5778 abi_ulong target_flock_addr) 5779 { 5780 struct target_oabi_flock64 *target_fl; 5781 int l_type; 5782 5783 if (!lock_user_struct(VERIFY_READ, target_fl, target_flock_addr, 1)) { 5784 return -TARGET_EFAULT; 5785 } 5786 5787 __get_user(l_type, &target_fl->l_type); 5788 l_type = target_to_host_flock(l_type); 5789 if (l_type < 0) { 5790 return l_type; 5791 } 5792 fl->l_type = l_type; 5793 __get_user(fl->l_whence, &target_fl->l_whence); 5794 __get_user(fl->l_start, &target_fl->l_start); 5795 __get_user(fl->l_len, &target_fl->l_len); 5796 __get_user(fl->l_pid, &target_fl->l_pid); 5797 unlock_user_struct(target_fl, target_flock_addr, 0); 5798 return 0; 5799 } 5800 5801 static inline abi_long copy_to_user_oabi_flock64(abi_ulong target_flock_addr, 5802 const struct flock64 *fl) 5803 { 5804 struct target_oabi_flock64 *target_fl; 5805 short l_type; 5806 5807 if (!lock_user_struct(VERIFY_WRITE, target_fl, target_flock_addr, 0)) { 5808 return -TARGET_EFAULT; 5809 } 5810 5811 l_type = host_to_target_flock(fl->l_type); 5812 __put_user(l_type, &target_fl->l_type); 5813 __put_user(fl->l_whence, &target_fl->l_whence); 5814 __put_user(fl->l_start, &target_fl->l_start); 5815 __put_user(fl->l_len, &target_fl->l_len); 5816 __put_user(fl->l_pid, &target_fl->l_pid); 5817 unlock_user_struct(target_fl, target_flock_addr, 1); 5818 return 0; 5819 } 5820 #endif 5821 5822 static inline abi_long copy_from_user_flock64(struct flock64 *fl, 5823 abi_ulong target_flock_addr) 5824 { 5825 struct target_flock64 *target_fl; 5826 int l_type; 5827 5828 if (!lock_user_struct(VERIFY_READ, target_fl, target_flock_addr, 1)) { 5829 return -TARGET_EFAULT; 5830 } 5831 5832 __get_user(l_type, &target_fl->l_type); 5833 l_type = target_to_host_flock(l_type); 5834 if (l_type < 0) { 5835 return l_type; 5836 } 5837 fl->l_type = l_type; 5838 __get_user(fl->l_whence, &target_fl->l_whence); 5839 __get_user(fl->l_start, &target_fl->l_start); 5840 __get_user(fl->l_len, &target_fl->l_len); 5841 __get_user(fl->l_pid, &target_fl->l_pid); 5842 unlock_user_struct(target_fl, target_flock_addr, 0); 5843 return 0; 5844 } 5845 5846 static inline abi_long copy_to_user_flock64(abi_ulong target_flock_addr, 5847 const struct flock64 *fl) 5848 { 5849 struct target_flock64 *target_fl; 5850 short l_type; 5851 5852 if (!lock_user_struct(VERIFY_WRITE, target_fl, target_flock_addr, 0)) { 5853 return -TARGET_EFAULT; 5854 } 5855 5856 l_type = host_to_target_flock(fl->l_type); 5857 __put_user(l_type, &target_fl->l_type); 5858 __put_user(fl->l_whence, &target_fl->l_whence); 5859 __put_user(fl->l_start, &target_fl->l_start); 5860 __put_user(fl->l_len, &target_fl->l_len); 5861 __put_user(fl->l_pid, &target_fl->l_pid); 5862 unlock_user_struct(target_fl, target_flock_addr, 1); 5863 return 0; 5864 } 5865 5866 static abi_long do_fcntl(int fd, int cmd, abi_ulong arg) 5867 { 5868 struct flock64 fl64; 5869 #ifdef F_GETOWN_EX 5870 struct f_owner_ex fox; 5871 struct target_f_owner_ex *target_fox; 5872 #endif 5873 abi_long ret; 5874 int host_cmd = target_to_host_fcntl_cmd(cmd); 5875 5876 if (host_cmd == -TARGET_EINVAL) 5877 return host_cmd; 5878 5879 switch(cmd) { 5880 case TARGET_F_GETLK: 5881 ret = copy_from_user_flock(&fl64, arg); 5882 if (ret) { 5883 return ret; 5884 } 5885 ret = get_errno(safe_fcntl(fd, host_cmd, &fl64)); 5886 if (ret == 0) { 5887 ret = copy_to_user_flock(arg, &fl64); 5888 } 5889 break; 5890 5891 case TARGET_F_SETLK: 5892 case TARGET_F_SETLKW: 5893 ret = copy_from_user_flock(&fl64, arg); 5894 if (ret) { 5895 return ret; 5896 } 5897 ret = get_errno(safe_fcntl(fd, host_cmd, &fl64)); 5898 break; 5899 5900 case TARGET_F_GETLK64: 5901 ret = copy_from_user_flock64(&fl64, arg); 5902 if (ret) { 5903 return ret; 5904 } 5905 ret = get_errno(safe_fcntl(fd, host_cmd, &fl64)); 5906 if (ret == 0) { 5907 ret = copy_to_user_flock64(arg, &fl64); 5908 } 5909 break; 5910 case TARGET_F_SETLK64: 5911 case TARGET_F_SETLKW64: 5912 ret = copy_from_user_flock64(&fl64, arg); 5913 if (ret) { 5914 return ret; 5915 } 5916 ret = get_errno(safe_fcntl(fd, host_cmd, &fl64)); 5917 break; 5918 5919 case TARGET_F_GETFL: 5920 ret = get_errno(safe_fcntl(fd, host_cmd, arg)); 5921 if (ret >= 0) { 5922 ret = host_to_target_bitmask(ret, fcntl_flags_tbl); 5923 } 5924 break; 5925 5926 case TARGET_F_SETFL: 5927 ret = get_errno(safe_fcntl(fd, host_cmd, 5928 target_to_host_bitmask(arg, 5929 fcntl_flags_tbl))); 5930 break; 5931 5932 #ifdef F_GETOWN_EX 5933 case TARGET_F_GETOWN_EX: 5934 ret = get_errno(safe_fcntl(fd, host_cmd, &fox)); 5935 if (ret >= 0) { 5936 if (!lock_user_struct(VERIFY_WRITE, target_fox, arg, 0)) 5937 return -TARGET_EFAULT; 5938 target_fox->type = tswap32(fox.type); 5939 target_fox->pid = tswap32(fox.pid); 5940 unlock_user_struct(target_fox, arg, 1); 5941 } 5942 break; 5943 #endif 5944 5945 #ifdef F_SETOWN_EX 5946 case TARGET_F_SETOWN_EX: 5947 if (!lock_user_struct(VERIFY_READ, target_fox, arg, 1)) 5948 return -TARGET_EFAULT; 5949 fox.type = tswap32(target_fox->type); 5950 fox.pid = tswap32(target_fox->pid); 5951 unlock_user_struct(target_fox, arg, 0); 5952 ret = get_errno(safe_fcntl(fd, host_cmd, &fox)); 5953 break; 5954 #endif 5955 5956 case TARGET_F_SETOWN: 5957 case TARGET_F_GETOWN: 5958 case TARGET_F_SETSIG: 5959 case TARGET_F_GETSIG: 5960 case TARGET_F_SETLEASE: 5961 case TARGET_F_GETLEASE: 5962 case TARGET_F_SETPIPE_SZ: 5963 case TARGET_F_GETPIPE_SZ: 5964 ret = get_errno(safe_fcntl(fd, host_cmd, arg)); 5965 break; 5966 5967 default: 5968 ret = get_errno(safe_fcntl(fd, cmd, arg)); 5969 break; 5970 } 5971 return ret; 5972 } 5973 5974 #ifdef USE_UID16 5975 5976 static inline int high2lowuid(int uid) 5977 { 5978 if (uid > 65535) 5979 return 65534; 5980 else 5981 return uid; 5982 } 5983 5984 static inline int high2lowgid(int gid) 5985 { 5986 if (gid > 65535) 5987 return 65534; 5988 else 5989 return gid; 5990 } 5991 5992 static inline int low2highuid(int uid) 5993 { 5994 if ((int16_t)uid == -1) 5995 return -1; 5996 else 5997 return uid; 5998 } 5999 6000 static inline int low2highgid(int gid) 6001 { 6002 if ((int16_t)gid == -1) 6003 return -1; 6004 else 6005 return gid; 6006 } 6007 static inline int tswapid(int id) 6008 { 6009 return tswap16(id); 6010 } 6011 6012 #define put_user_id(x, gaddr) put_user_u16(x, gaddr) 6013 6014 #else /* !USE_UID16 */ 6015 static inline int high2lowuid(int uid) 6016 { 6017 return uid; 6018 } 6019 static inline int high2lowgid(int gid) 6020 { 6021 return gid; 6022 } 6023 static inline int low2highuid(int uid) 6024 { 6025 return uid; 6026 } 6027 static inline int low2highgid(int gid) 6028 { 6029 return gid; 6030 } 6031 static inline int tswapid(int id) 6032 { 6033 return tswap32(id); 6034 } 6035 6036 #define put_user_id(x, gaddr) put_user_u32(x, gaddr) 6037 6038 #endif /* USE_UID16 */ 6039 6040 /* We must do direct syscalls for setting UID/GID, because we want to 6041 * implement the Linux system call semantics of "change only for this thread", 6042 * not the libc/POSIX semantics of "change for all threads in process". 6043 * (See http://ewontfix.com/17/ for more details.) 6044 * We use the 32-bit version of the syscalls if present; if it is not 6045 * then either the host architecture supports 32-bit UIDs natively with 6046 * the standard syscall, or the 16-bit UID is the best we can do. 6047 */ 6048 #ifdef __NR_setuid32 6049 #define __NR_sys_setuid __NR_setuid32 6050 #else 6051 #define __NR_sys_setuid __NR_setuid 6052 #endif 6053 #ifdef __NR_setgid32 6054 #define __NR_sys_setgid __NR_setgid32 6055 #else 6056 #define __NR_sys_setgid __NR_setgid 6057 #endif 6058 #ifdef __NR_setresuid32 6059 #define __NR_sys_setresuid __NR_setresuid32 6060 #else 6061 #define __NR_sys_setresuid __NR_setresuid 6062 #endif 6063 #ifdef __NR_setresgid32 6064 #define __NR_sys_setresgid __NR_setresgid32 6065 #else 6066 #define __NR_sys_setresgid __NR_setresgid 6067 #endif 6068 6069 _syscall1(int, sys_setuid, uid_t, uid) 6070 _syscall1(int, sys_setgid, gid_t, gid) 6071 _syscall3(int, sys_setresuid, uid_t, ruid, uid_t, euid, uid_t, suid) 6072 _syscall3(int, sys_setresgid, gid_t, rgid, gid_t, egid, gid_t, sgid) 6073 6074 void syscall_init(void) 6075 { 6076 IOCTLEntry *ie; 6077 const argtype *arg_type; 6078 int size; 6079 int i; 6080 6081 thunk_init(STRUCT_MAX); 6082 6083 #define STRUCT(name, ...) thunk_register_struct(STRUCT_ ## name, #name, struct_ ## name ## _def); 6084 #define STRUCT_SPECIAL(name) thunk_register_struct_direct(STRUCT_ ## name, #name, &struct_ ## name ## _def); 6085 #include "syscall_types.h" 6086 #undef STRUCT 6087 #undef STRUCT_SPECIAL 6088 6089 /* Build target_to_host_errno_table[] table from 6090 * host_to_target_errno_table[]. */ 6091 for (i = 0; i < ERRNO_TABLE_SIZE; i++) { 6092 target_to_host_errno_table[host_to_target_errno_table[i]] = i; 6093 } 6094 6095 /* we patch the ioctl size if necessary. We rely on the fact that 6096 no ioctl has all the bits at '1' in the size field */ 6097 ie = ioctl_entries; 6098 while (ie->target_cmd != 0) { 6099 if (((ie->target_cmd >> TARGET_IOC_SIZESHIFT) & TARGET_IOC_SIZEMASK) == 6100 TARGET_IOC_SIZEMASK) { 6101 arg_type = ie->arg_type; 6102 if (arg_type[0] != TYPE_PTR) { 6103 fprintf(stderr, "cannot patch size for ioctl 0x%x\n", 6104 ie->target_cmd); 6105 exit(1); 6106 } 6107 arg_type++; 6108 size = thunk_type_size(arg_type, 0); 6109 ie->target_cmd = (ie->target_cmd & 6110 ~(TARGET_IOC_SIZEMASK << TARGET_IOC_SIZESHIFT)) | 6111 (size << TARGET_IOC_SIZESHIFT); 6112 } 6113 6114 /* automatic consistency check if same arch */ 6115 #if (defined(__i386__) && defined(TARGET_I386) && defined(TARGET_ABI32)) || \ 6116 (defined(__x86_64__) && defined(TARGET_X86_64)) 6117 if (unlikely(ie->target_cmd != ie->host_cmd)) { 6118 fprintf(stderr, "ERROR: ioctl(%s): target=0x%x host=0x%x\n", 6119 ie->name, ie->target_cmd, ie->host_cmd); 6120 } 6121 #endif 6122 ie++; 6123 } 6124 } 6125 6126 #if TARGET_ABI_BITS == 32 6127 static inline uint64_t target_offset64(uint32_t word0, uint32_t word1) 6128 { 6129 #ifdef TARGET_WORDS_BIGENDIAN 6130 return ((uint64_t)word0 << 32) | word1; 6131 #else 6132 return ((uint64_t)word1 << 32) | word0; 6133 #endif 6134 } 6135 #else /* TARGET_ABI_BITS == 32 */ 6136 static inline uint64_t target_offset64(uint64_t word0, uint64_t word1) 6137 { 6138 return word0; 6139 } 6140 #endif /* TARGET_ABI_BITS != 32 */ 6141 6142 #ifdef TARGET_NR_truncate64 6143 static inline abi_long target_truncate64(void *cpu_env, const char *arg1, 6144 abi_long arg2, 6145 abi_long arg3, 6146 abi_long arg4) 6147 { 6148 if (regpairs_aligned(cpu_env, TARGET_NR_truncate64)) { 6149 arg2 = arg3; 6150 arg3 = arg4; 6151 } 6152 return get_errno(truncate64(arg1, target_offset64(arg2, arg3))); 6153 } 6154 #endif 6155 6156 #ifdef TARGET_NR_ftruncate64 6157 static inline abi_long target_ftruncate64(void *cpu_env, abi_long arg1, 6158 abi_long arg2, 6159 abi_long arg3, 6160 abi_long arg4) 6161 { 6162 if (regpairs_aligned(cpu_env, TARGET_NR_ftruncate64)) { 6163 arg2 = arg3; 6164 arg3 = arg4; 6165 } 6166 return get_errno(ftruncate64(arg1, target_offset64(arg2, arg3))); 6167 } 6168 #endif 6169 6170 static inline abi_long target_to_host_timespec(struct timespec *host_ts, 6171 abi_ulong target_addr) 6172 { 6173 struct target_timespec *target_ts; 6174 6175 if (!lock_user_struct(VERIFY_READ, target_ts, target_addr, 1)) 6176 return -TARGET_EFAULT; 6177 __get_user(host_ts->tv_sec, &target_ts->tv_sec); 6178 __get_user(host_ts->tv_nsec, &target_ts->tv_nsec); 6179 unlock_user_struct(target_ts, target_addr, 0); 6180 return 0; 6181 } 6182 6183 static inline abi_long host_to_target_timespec(abi_ulong target_addr, 6184 struct timespec *host_ts) 6185 { 6186 struct target_timespec *target_ts; 6187 6188 if (!lock_user_struct(VERIFY_WRITE, target_ts, target_addr, 0)) 6189 return -TARGET_EFAULT; 6190 __put_user(host_ts->tv_sec, &target_ts->tv_sec); 6191 __put_user(host_ts->tv_nsec, &target_ts->tv_nsec); 6192 unlock_user_struct(target_ts, target_addr, 1); 6193 return 0; 6194 } 6195 6196 static inline abi_long target_to_host_itimerspec(struct itimerspec *host_itspec, 6197 abi_ulong target_addr) 6198 { 6199 struct target_itimerspec *target_itspec; 6200 6201 if (!lock_user_struct(VERIFY_READ, target_itspec, target_addr, 1)) { 6202 return -TARGET_EFAULT; 6203 } 6204 6205 host_itspec->it_interval.tv_sec = 6206 tswapal(target_itspec->it_interval.tv_sec); 6207 host_itspec->it_interval.tv_nsec = 6208 tswapal(target_itspec->it_interval.tv_nsec); 6209 host_itspec->it_value.tv_sec = tswapal(target_itspec->it_value.tv_sec); 6210 host_itspec->it_value.tv_nsec = tswapal(target_itspec->it_value.tv_nsec); 6211 6212 unlock_user_struct(target_itspec, target_addr, 1); 6213 return 0; 6214 } 6215 6216 static inline abi_long host_to_target_itimerspec(abi_ulong target_addr, 6217 struct itimerspec *host_its) 6218 { 6219 struct target_itimerspec *target_itspec; 6220 6221 if (!lock_user_struct(VERIFY_WRITE, target_itspec, target_addr, 0)) { 6222 return -TARGET_EFAULT; 6223 } 6224 6225 target_itspec->it_interval.tv_sec = tswapal(host_its->it_interval.tv_sec); 6226 target_itspec->it_interval.tv_nsec = tswapal(host_its->it_interval.tv_nsec); 6227 6228 target_itspec->it_value.tv_sec = tswapal(host_its->it_value.tv_sec); 6229 target_itspec->it_value.tv_nsec = tswapal(host_its->it_value.tv_nsec); 6230 6231 unlock_user_struct(target_itspec, target_addr, 0); 6232 return 0; 6233 } 6234 6235 static inline abi_long target_to_host_timex(struct timex *host_tx, 6236 abi_long target_addr) 6237 { 6238 struct target_timex *target_tx; 6239 6240 if (!lock_user_struct(VERIFY_READ, target_tx, target_addr, 1)) { 6241 return -TARGET_EFAULT; 6242 } 6243 6244 __get_user(host_tx->modes, &target_tx->modes); 6245 __get_user(host_tx->offset, &target_tx->offset); 6246 __get_user(host_tx->freq, &target_tx->freq); 6247 __get_user(host_tx->maxerror, &target_tx->maxerror); 6248 __get_user(host_tx->esterror, &target_tx->esterror); 6249 __get_user(host_tx->status, &target_tx->status); 6250 __get_user(host_tx->constant, &target_tx->constant); 6251 __get_user(host_tx->precision, &target_tx->precision); 6252 __get_user(host_tx->tolerance, &target_tx->tolerance); 6253 __get_user(host_tx->time.tv_sec, &target_tx->time.tv_sec); 6254 __get_user(host_tx->time.tv_usec, &target_tx->time.tv_usec); 6255 __get_user(host_tx->tick, &target_tx->tick); 6256 __get_user(host_tx->ppsfreq, &target_tx->ppsfreq); 6257 __get_user(host_tx->jitter, &target_tx->jitter); 6258 __get_user(host_tx->shift, &target_tx->shift); 6259 __get_user(host_tx->stabil, &target_tx->stabil); 6260 __get_user(host_tx->jitcnt, &target_tx->jitcnt); 6261 __get_user(host_tx->calcnt, &target_tx->calcnt); 6262 __get_user(host_tx->errcnt, &target_tx->errcnt); 6263 __get_user(host_tx->stbcnt, &target_tx->stbcnt); 6264 __get_user(host_tx->tai, &target_tx->tai); 6265 6266 unlock_user_struct(target_tx, target_addr, 0); 6267 return 0; 6268 } 6269 6270 static inline abi_long host_to_target_timex(abi_long target_addr, 6271 struct timex *host_tx) 6272 { 6273 struct target_timex *target_tx; 6274 6275 if (!lock_user_struct(VERIFY_WRITE, target_tx, target_addr, 0)) { 6276 return -TARGET_EFAULT; 6277 } 6278 6279 __put_user(host_tx->modes, &target_tx->modes); 6280 __put_user(host_tx->offset, &target_tx->offset); 6281 __put_user(host_tx->freq, &target_tx->freq); 6282 __put_user(host_tx->maxerror, &target_tx->maxerror); 6283 __put_user(host_tx->esterror, &target_tx->esterror); 6284 __put_user(host_tx->status, &target_tx->status); 6285 __put_user(host_tx->constant, &target_tx->constant); 6286 __put_user(host_tx->precision, &target_tx->precision); 6287 __put_user(host_tx->tolerance, &target_tx->tolerance); 6288 __put_user(host_tx->time.tv_sec, &target_tx->time.tv_sec); 6289 __put_user(host_tx->time.tv_usec, &target_tx->time.tv_usec); 6290 __put_user(host_tx->tick, &target_tx->tick); 6291 __put_user(host_tx->ppsfreq, &target_tx->ppsfreq); 6292 __put_user(host_tx->jitter, &target_tx->jitter); 6293 __put_user(host_tx->shift, &target_tx->shift); 6294 __put_user(host_tx->stabil, &target_tx->stabil); 6295 __put_user(host_tx->jitcnt, &target_tx->jitcnt); 6296 __put_user(host_tx->calcnt, &target_tx->calcnt); 6297 __put_user(host_tx->errcnt, &target_tx->errcnt); 6298 __put_user(host_tx->stbcnt, &target_tx->stbcnt); 6299 __put_user(host_tx->tai, &target_tx->tai); 6300 6301 unlock_user_struct(target_tx, target_addr, 1); 6302 return 0; 6303 } 6304 6305 6306 static inline abi_long target_to_host_sigevent(struct sigevent *host_sevp, 6307 abi_ulong target_addr) 6308 { 6309 struct target_sigevent *target_sevp; 6310 6311 if (!lock_user_struct(VERIFY_READ, target_sevp, target_addr, 1)) { 6312 return -TARGET_EFAULT; 6313 } 6314 6315 /* This union is awkward on 64 bit systems because it has a 32 bit 6316 * integer and a pointer in it; we follow the conversion approach 6317 * used for handling sigval types in signal.c so the guest should get 6318 * the correct value back even if we did a 64 bit byteswap and it's 6319 * using the 32 bit integer. 6320 */ 6321 host_sevp->sigev_value.sival_ptr = 6322 (void *)(uintptr_t)tswapal(target_sevp->sigev_value.sival_ptr); 6323 host_sevp->sigev_signo = 6324 target_to_host_signal(tswap32(target_sevp->sigev_signo)); 6325 host_sevp->sigev_notify = tswap32(target_sevp->sigev_notify); 6326 host_sevp->_sigev_un._tid = tswap32(target_sevp->_sigev_un._tid); 6327 6328 unlock_user_struct(target_sevp, target_addr, 1); 6329 return 0; 6330 } 6331 6332 #if defined(TARGET_NR_mlockall) 6333 static inline int target_to_host_mlockall_arg(int arg) 6334 { 6335 int result = 0; 6336 6337 if (arg & TARGET_MLOCKALL_MCL_CURRENT) { 6338 result |= MCL_CURRENT; 6339 } 6340 if (arg & TARGET_MLOCKALL_MCL_FUTURE) { 6341 result |= MCL_FUTURE; 6342 } 6343 return result; 6344 } 6345 #endif 6346 6347 #if (defined(TARGET_NR_stat64) || defined(TARGET_NR_lstat64) || \ 6348 defined(TARGET_NR_fstat64) || defined(TARGET_NR_fstatat64) || \ 6349 defined(TARGET_NR_newfstatat)) 6350 static inline abi_long host_to_target_stat64(void *cpu_env, 6351 abi_ulong target_addr, 6352 struct stat *host_st) 6353 { 6354 #if defined(TARGET_ARM) && defined(TARGET_ABI32) 6355 if (((CPUARMState *)cpu_env)->eabi) { 6356 struct target_eabi_stat64 *target_st; 6357 6358 if (!lock_user_struct(VERIFY_WRITE, target_st, target_addr, 0)) 6359 return -TARGET_EFAULT; 6360 memset(target_st, 0, sizeof(struct target_eabi_stat64)); 6361 __put_user(host_st->st_dev, &target_st->st_dev); 6362 __put_user(host_st->st_ino, &target_st->st_ino); 6363 #ifdef TARGET_STAT64_HAS_BROKEN_ST_INO 6364 __put_user(host_st->st_ino, &target_st->__st_ino); 6365 #endif 6366 __put_user(host_st->st_mode, &target_st->st_mode); 6367 __put_user(host_st->st_nlink, &target_st->st_nlink); 6368 __put_user(host_st->st_uid, &target_st->st_uid); 6369 __put_user(host_st->st_gid, &target_st->st_gid); 6370 __put_user(host_st->st_rdev, &target_st->st_rdev); 6371 __put_user(host_st->st_size, &target_st->st_size); 6372 __put_user(host_st->st_blksize, &target_st->st_blksize); 6373 __put_user(host_st->st_blocks, &target_st->st_blocks); 6374 __put_user(host_st->st_atime, &target_st->target_st_atime); 6375 __put_user(host_st->st_mtime, &target_st->target_st_mtime); 6376 __put_user(host_st->st_ctime, &target_st->target_st_ctime); 6377 unlock_user_struct(target_st, target_addr, 1); 6378 } else 6379 #endif 6380 { 6381 #if defined(TARGET_HAS_STRUCT_STAT64) 6382 struct target_stat64 *target_st; 6383 #else 6384 struct target_stat *target_st; 6385 #endif 6386 6387 if (!lock_user_struct(VERIFY_WRITE, target_st, target_addr, 0)) 6388 return -TARGET_EFAULT; 6389 memset(target_st, 0, sizeof(*target_st)); 6390 __put_user(host_st->st_dev, &target_st->st_dev); 6391 __put_user(host_st->st_ino, &target_st->st_ino); 6392 #ifdef TARGET_STAT64_HAS_BROKEN_ST_INO 6393 __put_user(host_st->st_ino, &target_st->__st_ino); 6394 #endif 6395 __put_user(host_st->st_mode, &target_st->st_mode); 6396 __put_user(host_st->st_nlink, &target_st->st_nlink); 6397 __put_user(host_st->st_uid, &target_st->st_uid); 6398 __put_user(host_st->st_gid, &target_st->st_gid); 6399 __put_user(host_st->st_rdev, &target_st->st_rdev); 6400 /* XXX: better use of kernel struct */ 6401 __put_user(host_st->st_size, &target_st->st_size); 6402 __put_user(host_st->st_blksize, &target_st->st_blksize); 6403 __put_user(host_st->st_blocks, &target_st->st_blocks); 6404 __put_user(host_st->st_atime, &target_st->target_st_atime); 6405 __put_user(host_st->st_mtime, &target_st->target_st_mtime); 6406 __put_user(host_st->st_ctime, &target_st->target_st_ctime); 6407 unlock_user_struct(target_st, target_addr, 1); 6408 } 6409 6410 return 0; 6411 } 6412 #endif 6413 6414 /* ??? Using host futex calls even when target atomic operations 6415 are not really atomic probably breaks things. However implementing 6416 futexes locally would make futexes shared between multiple processes 6417 tricky. However they're probably useless because guest atomic 6418 operations won't work either. */ 6419 static int do_futex(target_ulong uaddr, int op, int val, target_ulong timeout, 6420 target_ulong uaddr2, int val3) 6421 { 6422 struct timespec ts, *pts; 6423 int base_op; 6424 6425 /* ??? We assume FUTEX_* constants are the same on both host 6426 and target. */ 6427 #ifdef FUTEX_CMD_MASK 6428 base_op = op & FUTEX_CMD_MASK; 6429 #else 6430 base_op = op; 6431 #endif 6432 switch (base_op) { 6433 case FUTEX_WAIT: 6434 case FUTEX_WAIT_BITSET: 6435 if (timeout) { 6436 pts = &ts; 6437 target_to_host_timespec(pts, timeout); 6438 } else { 6439 pts = NULL; 6440 } 6441 return get_errno(safe_futex(g2h(uaddr), op, tswap32(val), 6442 pts, NULL, val3)); 6443 case FUTEX_WAKE: 6444 return get_errno(safe_futex(g2h(uaddr), op, val, NULL, NULL, 0)); 6445 case FUTEX_FD: 6446 return get_errno(safe_futex(g2h(uaddr), op, val, NULL, NULL, 0)); 6447 case FUTEX_REQUEUE: 6448 case FUTEX_CMP_REQUEUE: 6449 case FUTEX_WAKE_OP: 6450 /* For FUTEX_REQUEUE, FUTEX_CMP_REQUEUE, and FUTEX_WAKE_OP, the 6451 TIMEOUT parameter is interpreted as a uint32_t by the kernel. 6452 But the prototype takes a `struct timespec *'; insert casts 6453 to satisfy the compiler. We do not need to tswap TIMEOUT 6454 since it's not compared to guest memory. */ 6455 pts = (struct timespec *)(uintptr_t) timeout; 6456 return get_errno(safe_futex(g2h(uaddr), op, val, pts, 6457 g2h(uaddr2), 6458 (base_op == FUTEX_CMP_REQUEUE 6459 ? tswap32(val3) 6460 : val3))); 6461 default: 6462 return -TARGET_ENOSYS; 6463 } 6464 } 6465 #if defined(TARGET_NR_name_to_handle_at) && defined(CONFIG_OPEN_BY_HANDLE) 6466 static abi_long do_name_to_handle_at(abi_long dirfd, abi_long pathname, 6467 abi_long handle, abi_long mount_id, 6468 abi_long flags) 6469 { 6470 struct file_handle *target_fh; 6471 struct file_handle *fh; 6472 int mid = 0; 6473 abi_long ret; 6474 char *name; 6475 unsigned int size, total_size; 6476 6477 if (get_user_s32(size, handle)) { 6478 return -TARGET_EFAULT; 6479 } 6480 6481 name = lock_user_string(pathname); 6482 if (!name) { 6483 return -TARGET_EFAULT; 6484 } 6485 6486 total_size = sizeof(struct file_handle) + size; 6487 target_fh = lock_user(VERIFY_WRITE, handle, total_size, 0); 6488 if (!target_fh) { 6489 unlock_user(name, pathname, 0); 6490 return -TARGET_EFAULT; 6491 } 6492 6493 fh = g_malloc0(total_size); 6494 fh->handle_bytes = size; 6495 6496 ret = get_errno(name_to_handle_at(dirfd, path(name), fh, &mid, flags)); 6497 unlock_user(name, pathname, 0); 6498 6499 /* man name_to_handle_at(2): 6500 * Other than the use of the handle_bytes field, the caller should treat 6501 * the file_handle structure as an opaque data type 6502 */ 6503 6504 memcpy(target_fh, fh, total_size); 6505 target_fh->handle_bytes = tswap32(fh->handle_bytes); 6506 target_fh->handle_type = tswap32(fh->handle_type); 6507 g_free(fh); 6508 unlock_user(target_fh, handle, total_size); 6509 6510 if (put_user_s32(mid, mount_id)) { 6511 return -TARGET_EFAULT; 6512 } 6513 6514 return ret; 6515 6516 } 6517 #endif 6518 6519 #if defined(TARGET_NR_open_by_handle_at) && defined(CONFIG_OPEN_BY_HANDLE) 6520 static abi_long do_open_by_handle_at(abi_long mount_fd, abi_long handle, 6521 abi_long flags) 6522 { 6523 struct file_handle *target_fh; 6524 struct file_handle *fh; 6525 unsigned int size, total_size; 6526 abi_long ret; 6527 6528 if (get_user_s32(size, handle)) { 6529 return -TARGET_EFAULT; 6530 } 6531 6532 total_size = sizeof(struct file_handle) + size; 6533 target_fh = lock_user(VERIFY_READ, handle, total_size, 1); 6534 if (!target_fh) { 6535 return -TARGET_EFAULT; 6536 } 6537 6538 fh = g_memdup(target_fh, total_size); 6539 fh->handle_bytes = size; 6540 fh->handle_type = tswap32(target_fh->handle_type); 6541 6542 ret = get_errno(open_by_handle_at(mount_fd, fh, 6543 target_to_host_bitmask(flags, fcntl_flags_tbl))); 6544 6545 g_free(fh); 6546 6547 unlock_user(target_fh, handle, total_size); 6548 6549 return ret; 6550 } 6551 #endif 6552 6553 #if defined(TARGET_NR_signalfd) || defined(TARGET_NR_signalfd4) 6554 6555 static abi_long do_signalfd4(int fd, abi_long mask, int flags) 6556 { 6557 int host_flags; 6558 target_sigset_t *target_mask; 6559 sigset_t host_mask; 6560 abi_long ret; 6561 6562 if (flags & ~(TARGET_O_NONBLOCK | TARGET_O_CLOEXEC)) { 6563 return -TARGET_EINVAL; 6564 } 6565 if (!lock_user_struct(VERIFY_READ, target_mask, mask, 1)) { 6566 return -TARGET_EFAULT; 6567 } 6568 6569 target_to_host_sigset(&host_mask, target_mask); 6570 6571 host_flags = target_to_host_bitmask(flags, fcntl_flags_tbl); 6572 6573 ret = get_errno(signalfd(fd, &host_mask, host_flags)); 6574 if (ret >= 0) { 6575 fd_trans_register(ret, &target_signalfd_trans); 6576 } 6577 6578 unlock_user_struct(target_mask, mask, 0); 6579 6580 return ret; 6581 } 6582 #endif 6583 6584 /* Map host to target signal numbers for the wait family of syscalls. 6585 Assume all other status bits are the same. */ 6586 int host_to_target_waitstatus(int status) 6587 { 6588 if (WIFSIGNALED(status)) { 6589 return host_to_target_signal(WTERMSIG(status)) | (status & ~0x7f); 6590 } 6591 if (WIFSTOPPED(status)) { 6592 return (host_to_target_signal(WSTOPSIG(status)) << 8) 6593 | (status & 0xff); 6594 } 6595 return status; 6596 } 6597 6598 static int open_self_cmdline(void *cpu_env, int fd) 6599 { 6600 CPUState *cpu = ENV_GET_CPU((CPUArchState *)cpu_env); 6601 struct linux_binprm *bprm = ((TaskState *)cpu->opaque)->bprm; 6602 int i; 6603 6604 for (i = 0; i < bprm->argc; i++) { 6605 size_t len = strlen(bprm->argv[i]) + 1; 6606 6607 if (write(fd, bprm->argv[i], len) != len) { 6608 return -1; 6609 } 6610 } 6611 6612 return 0; 6613 } 6614 6615 static int open_self_maps(void *cpu_env, int fd) 6616 { 6617 CPUState *cpu = ENV_GET_CPU((CPUArchState *)cpu_env); 6618 TaskState *ts = cpu->opaque; 6619 FILE *fp; 6620 char *line = NULL; 6621 size_t len = 0; 6622 ssize_t read; 6623 6624 fp = fopen("/proc/self/maps", "r"); 6625 if (fp == NULL) { 6626 return -1; 6627 } 6628 6629 while ((read = getline(&line, &len, fp)) != -1) { 6630 int fields, dev_maj, dev_min, inode; 6631 uint64_t min, max, offset; 6632 char flag_r, flag_w, flag_x, flag_p; 6633 char path[512] = ""; 6634 fields = sscanf(line, "%"PRIx64"-%"PRIx64" %c%c%c%c %"PRIx64" %x:%x %d" 6635 " %512s", &min, &max, &flag_r, &flag_w, &flag_x, 6636 &flag_p, &offset, &dev_maj, &dev_min, &inode, path); 6637 6638 if ((fields < 10) || (fields > 11)) { 6639 continue; 6640 } 6641 if (h2g_valid(min)) { 6642 int flags = page_get_flags(h2g(min)); 6643 max = h2g_valid(max - 1) ? max : (uintptr_t)g2h(GUEST_ADDR_MAX) + 1; 6644 if (page_check_range(h2g(min), max - min, flags) == -1) { 6645 continue; 6646 } 6647 if (h2g(min) == ts->info->stack_limit) { 6648 pstrcpy(path, sizeof(path), " [stack]"); 6649 } 6650 dprintf(fd, TARGET_ABI_FMT_ptr "-" TARGET_ABI_FMT_ptr 6651 " %c%c%c%c %08" PRIx64 " %02x:%02x %d %s%s\n", 6652 h2g(min), h2g(max - 1) + 1, flag_r, flag_w, 6653 flag_x, flag_p, offset, dev_maj, dev_min, inode, 6654 path[0] ? " " : "", path); 6655 } 6656 } 6657 6658 free(line); 6659 fclose(fp); 6660 6661 return 0; 6662 } 6663 6664 static int open_self_stat(void *cpu_env, int fd) 6665 { 6666 CPUState *cpu = ENV_GET_CPU((CPUArchState *)cpu_env); 6667 TaskState *ts = cpu->opaque; 6668 abi_ulong start_stack = ts->info->start_stack; 6669 int i; 6670 6671 for (i = 0; i < 44; i++) { 6672 char buf[128]; 6673 int len; 6674 uint64_t val = 0; 6675 6676 if (i == 0) { 6677 /* pid */ 6678 val = getpid(); 6679 snprintf(buf, sizeof(buf), "%"PRId64 " ", val); 6680 } else if (i == 1) { 6681 /* app name */ 6682 snprintf(buf, sizeof(buf), "(%s) ", ts->bprm->argv[0]); 6683 } else if (i == 27) { 6684 /* stack bottom */ 6685 val = start_stack; 6686 snprintf(buf, sizeof(buf), "%"PRId64 " ", val); 6687 } else { 6688 /* for the rest, there is MasterCard */ 6689 snprintf(buf, sizeof(buf), "0%c", i == 43 ? '\n' : ' '); 6690 } 6691 6692 len = strlen(buf); 6693 if (write(fd, buf, len) != len) { 6694 return -1; 6695 } 6696 } 6697 6698 return 0; 6699 } 6700 6701 static int open_self_auxv(void *cpu_env, int fd) 6702 { 6703 CPUState *cpu = ENV_GET_CPU((CPUArchState *)cpu_env); 6704 TaskState *ts = cpu->opaque; 6705 abi_ulong auxv = ts->info->saved_auxv; 6706 abi_ulong len = ts->info->auxv_len; 6707 char *ptr; 6708 6709 /* 6710 * Auxiliary vector is stored in target process stack. 6711 * read in whole auxv vector and copy it to file 6712 */ 6713 ptr = lock_user(VERIFY_READ, auxv, len, 0); 6714 if (ptr != NULL) { 6715 while (len > 0) { 6716 ssize_t r; 6717 r = write(fd, ptr, len); 6718 if (r <= 0) { 6719 break; 6720 } 6721 len -= r; 6722 ptr += r; 6723 } 6724 lseek(fd, 0, SEEK_SET); 6725 unlock_user(ptr, auxv, len); 6726 } 6727 6728 return 0; 6729 } 6730 6731 static int is_proc_myself(const char *filename, const char *entry) 6732 { 6733 if (!strncmp(filename, "/proc/", strlen("/proc/"))) { 6734 filename += strlen("/proc/"); 6735 if (!strncmp(filename, "self/", strlen("self/"))) { 6736 filename += strlen("self/"); 6737 } else if (*filename >= '1' && *filename <= '9') { 6738 char myself[80]; 6739 snprintf(myself, sizeof(myself), "%d/", getpid()); 6740 if (!strncmp(filename, myself, strlen(myself))) { 6741 filename += strlen(myself); 6742 } else { 6743 return 0; 6744 } 6745 } else { 6746 return 0; 6747 } 6748 if (!strcmp(filename, entry)) { 6749 return 1; 6750 } 6751 } 6752 return 0; 6753 } 6754 6755 #if defined(HOST_WORDS_BIGENDIAN) != defined(TARGET_WORDS_BIGENDIAN) 6756 static int is_proc(const char *filename, const char *entry) 6757 { 6758 return strcmp(filename, entry) == 0; 6759 } 6760 6761 static int open_net_route(void *cpu_env, int fd) 6762 { 6763 FILE *fp; 6764 char *line = NULL; 6765 size_t len = 0; 6766 ssize_t read; 6767 6768 fp = fopen("/proc/net/route", "r"); 6769 if (fp == NULL) { 6770 return -1; 6771 } 6772 6773 /* read header */ 6774 6775 read = getline(&line, &len, fp); 6776 dprintf(fd, "%s", line); 6777 6778 /* read routes */ 6779 6780 while ((read = getline(&line, &len, fp)) != -1) { 6781 char iface[16]; 6782 uint32_t dest, gw, mask; 6783 unsigned int flags, refcnt, use, metric, mtu, window, irtt; 6784 int fields; 6785 6786 fields = sscanf(line, 6787 "%s\t%08x\t%08x\t%04x\t%d\t%d\t%d\t%08x\t%d\t%u\t%u\n", 6788 iface, &dest, &gw, &flags, &refcnt, &use, &metric, 6789 &mask, &mtu, &window, &irtt); 6790 if (fields != 11) { 6791 continue; 6792 } 6793 dprintf(fd, "%s\t%08x\t%08x\t%04x\t%d\t%d\t%d\t%08x\t%d\t%u\t%u\n", 6794 iface, tswap32(dest), tswap32(gw), flags, refcnt, use, 6795 metric, tswap32(mask), mtu, window, irtt); 6796 } 6797 6798 free(line); 6799 fclose(fp); 6800 6801 return 0; 6802 } 6803 #endif 6804 6805 static int do_openat(void *cpu_env, int dirfd, const char *pathname, int flags, mode_t mode) 6806 { 6807 struct fake_open { 6808 const char *filename; 6809 int (*fill)(void *cpu_env, int fd); 6810 int (*cmp)(const char *s1, const char *s2); 6811 }; 6812 const struct fake_open *fake_open; 6813 static const struct fake_open fakes[] = { 6814 { "maps", open_self_maps, is_proc_myself }, 6815 { "stat", open_self_stat, is_proc_myself }, 6816 { "auxv", open_self_auxv, is_proc_myself }, 6817 { "cmdline", open_self_cmdline, is_proc_myself }, 6818 #if defined(HOST_WORDS_BIGENDIAN) != defined(TARGET_WORDS_BIGENDIAN) 6819 { "/proc/net/route", open_net_route, is_proc }, 6820 #endif 6821 { NULL, NULL, NULL } 6822 }; 6823 6824 if (is_proc_myself(pathname, "exe")) { 6825 int execfd = qemu_getauxval(AT_EXECFD); 6826 return execfd ? execfd : safe_openat(dirfd, exec_path, flags, mode); 6827 } 6828 6829 for (fake_open = fakes; fake_open->filename; fake_open++) { 6830 if (fake_open->cmp(pathname, fake_open->filename)) { 6831 break; 6832 } 6833 } 6834 6835 if (fake_open->filename) { 6836 const char *tmpdir; 6837 char filename[PATH_MAX]; 6838 int fd, r; 6839 6840 /* create temporary file to map stat to */ 6841 tmpdir = getenv("TMPDIR"); 6842 if (!tmpdir) 6843 tmpdir = "/tmp"; 6844 snprintf(filename, sizeof(filename), "%s/qemu-open.XXXXXX", tmpdir); 6845 fd = mkstemp(filename); 6846 if (fd < 0) { 6847 return fd; 6848 } 6849 unlink(filename); 6850 6851 if ((r = fake_open->fill(cpu_env, fd))) { 6852 int e = errno; 6853 close(fd); 6854 errno = e; 6855 return r; 6856 } 6857 lseek(fd, 0, SEEK_SET); 6858 6859 return fd; 6860 } 6861 6862 return safe_openat(dirfd, path(pathname), flags, mode); 6863 } 6864 6865 #define TIMER_MAGIC 0x0caf0000 6866 #define TIMER_MAGIC_MASK 0xffff0000 6867 6868 /* Convert QEMU provided timer ID back to internal 16bit index format */ 6869 static target_timer_t get_timer_id(abi_long arg) 6870 { 6871 target_timer_t timerid = arg; 6872 6873 if ((timerid & TIMER_MAGIC_MASK) != TIMER_MAGIC) { 6874 return -TARGET_EINVAL; 6875 } 6876 6877 timerid &= 0xffff; 6878 6879 if (timerid >= ARRAY_SIZE(g_posix_timers)) { 6880 return -TARGET_EINVAL; 6881 } 6882 6883 return timerid; 6884 } 6885 6886 static int target_to_host_cpu_mask(unsigned long *host_mask, 6887 size_t host_size, 6888 abi_ulong target_addr, 6889 size_t target_size) 6890 { 6891 unsigned target_bits = sizeof(abi_ulong) * 8; 6892 unsigned host_bits = sizeof(*host_mask) * 8; 6893 abi_ulong *target_mask; 6894 unsigned i, j; 6895 6896 assert(host_size >= target_size); 6897 6898 target_mask = lock_user(VERIFY_READ, target_addr, target_size, 1); 6899 if (!target_mask) { 6900 return -TARGET_EFAULT; 6901 } 6902 memset(host_mask, 0, host_size); 6903 6904 for (i = 0 ; i < target_size / sizeof(abi_ulong); i++) { 6905 unsigned bit = i * target_bits; 6906 abi_ulong val; 6907 6908 __get_user(val, &target_mask[i]); 6909 for (j = 0; j < target_bits; j++, bit++) { 6910 if (val & (1UL << j)) { 6911 host_mask[bit / host_bits] |= 1UL << (bit % host_bits); 6912 } 6913 } 6914 } 6915 6916 unlock_user(target_mask, target_addr, 0); 6917 return 0; 6918 } 6919 6920 static int host_to_target_cpu_mask(const unsigned long *host_mask, 6921 size_t host_size, 6922 abi_ulong target_addr, 6923 size_t target_size) 6924 { 6925 unsigned target_bits = sizeof(abi_ulong) * 8; 6926 unsigned host_bits = sizeof(*host_mask) * 8; 6927 abi_ulong *target_mask; 6928 unsigned i, j; 6929 6930 assert(host_size >= target_size); 6931 6932 target_mask = lock_user(VERIFY_WRITE, target_addr, target_size, 0); 6933 if (!target_mask) { 6934 return -TARGET_EFAULT; 6935 } 6936 6937 for (i = 0 ; i < target_size / sizeof(abi_ulong); i++) { 6938 unsigned bit = i * target_bits; 6939 abi_ulong val = 0; 6940 6941 for (j = 0; j < target_bits; j++, bit++) { 6942 if (host_mask[bit / host_bits] & (1UL << (bit % host_bits))) { 6943 val |= 1UL << j; 6944 } 6945 } 6946 __put_user(val, &target_mask[i]); 6947 } 6948 6949 unlock_user(target_mask, target_addr, target_size); 6950 return 0; 6951 } 6952 6953 /* This is an internal helper for do_syscall so that it is easier 6954 * to have a single return point, so that actions, such as logging 6955 * of syscall results, can be performed. 6956 * All errnos that do_syscall() returns must be -TARGET_<errcode>. 6957 */ 6958 static abi_long do_syscall1(void *cpu_env, int num, abi_long arg1, 6959 abi_long arg2, abi_long arg3, abi_long arg4, 6960 abi_long arg5, abi_long arg6, abi_long arg7, 6961 abi_long arg8) 6962 { 6963 CPUState *cpu = ENV_GET_CPU(cpu_env); 6964 abi_long ret; 6965 #if defined(TARGET_NR_stat) || defined(TARGET_NR_stat64) \ 6966 || defined(TARGET_NR_lstat) || defined(TARGET_NR_lstat64) \ 6967 || defined(TARGET_NR_fstat) || defined(TARGET_NR_fstat64) 6968 struct stat st; 6969 #endif 6970 #if defined(TARGET_NR_statfs) || defined(TARGET_NR_statfs64) \ 6971 || defined(TARGET_NR_fstatfs) 6972 struct statfs stfs; 6973 #endif 6974 void *p; 6975 6976 switch(num) { 6977 case TARGET_NR_exit: 6978 /* In old applications this may be used to implement _exit(2). 6979 However in threaded applictions it is used for thread termination, 6980 and _exit_group is used for application termination. 6981 Do thread termination if we have more then one thread. */ 6982 6983 if (block_signals()) { 6984 return -TARGET_ERESTARTSYS; 6985 } 6986 6987 cpu_list_lock(); 6988 6989 if (CPU_NEXT(first_cpu)) { 6990 TaskState *ts; 6991 6992 /* Remove the CPU from the list. */ 6993 QTAILQ_REMOVE_RCU(&cpus, cpu, node); 6994 6995 cpu_list_unlock(); 6996 6997 ts = cpu->opaque; 6998 if (ts->child_tidptr) { 6999 put_user_u32(0, ts->child_tidptr); 7000 sys_futex(g2h(ts->child_tidptr), FUTEX_WAKE, INT_MAX, 7001 NULL, NULL, 0); 7002 } 7003 thread_cpu = NULL; 7004 object_unref(OBJECT(cpu)); 7005 g_free(ts); 7006 rcu_unregister_thread(); 7007 pthread_exit(NULL); 7008 } 7009 7010 cpu_list_unlock(); 7011 preexit_cleanup(cpu_env, arg1); 7012 _exit(arg1); 7013 return 0; /* avoid warning */ 7014 case TARGET_NR_read: 7015 if (arg2 == 0 && arg3 == 0) { 7016 return get_errno(safe_read(arg1, 0, 0)); 7017 } else { 7018 if (!(p = lock_user(VERIFY_WRITE, arg2, arg3, 0))) 7019 return -TARGET_EFAULT; 7020 ret = get_errno(safe_read(arg1, p, arg3)); 7021 if (ret >= 0 && 7022 fd_trans_host_to_target_data(arg1)) { 7023 ret = fd_trans_host_to_target_data(arg1)(p, ret); 7024 } 7025 unlock_user(p, arg2, ret); 7026 } 7027 return ret; 7028 case TARGET_NR_write: 7029 if (arg2 == 0 && arg3 == 0) { 7030 return get_errno(safe_write(arg1, 0, 0)); 7031 } 7032 if (!(p = lock_user(VERIFY_READ, arg2, arg3, 1))) 7033 return -TARGET_EFAULT; 7034 if (fd_trans_target_to_host_data(arg1)) { 7035 void *copy = g_malloc(arg3); 7036 memcpy(copy, p, arg3); 7037 ret = fd_trans_target_to_host_data(arg1)(copy, arg3); 7038 if (ret >= 0) { 7039 ret = get_errno(safe_write(arg1, copy, ret)); 7040 } 7041 g_free(copy); 7042 } else { 7043 ret = get_errno(safe_write(arg1, p, arg3)); 7044 } 7045 unlock_user(p, arg2, 0); 7046 return ret; 7047 7048 #ifdef TARGET_NR_open 7049 case TARGET_NR_open: 7050 if (!(p = lock_user_string(arg1))) 7051 return -TARGET_EFAULT; 7052 ret = get_errno(do_openat(cpu_env, AT_FDCWD, p, 7053 target_to_host_bitmask(arg2, fcntl_flags_tbl), 7054 arg3)); 7055 fd_trans_unregister(ret); 7056 unlock_user(p, arg1, 0); 7057 return ret; 7058 #endif 7059 case TARGET_NR_openat: 7060 if (!(p = lock_user_string(arg2))) 7061 return -TARGET_EFAULT; 7062 ret = get_errno(do_openat(cpu_env, arg1, p, 7063 target_to_host_bitmask(arg3, fcntl_flags_tbl), 7064 arg4)); 7065 fd_trans_unregister(ret); 7066 unlock_user(p, arg2, 0); 7067 return ret; 7068 #if defined(TARGET_NR_name_to_handle_at) && defined(CONFIG_OPEN_BY_HANDLE) 7069 case TARGET_NR_name_to_handle_at: 7070 ret = do_name_to_handle_at(arg1, arg2, arg3, arg4, arg5); 7071 return ret; 7072 #endif 7073 #if defined(TARGET_NR_open_by_handle_at) && defined(CONFIG_OPEN_BY_HANDLE) 7074 case TARGET_NR_open_by_handle_at: 7075 ret = do_open_by_handle_at(arg1, arg2, arg3); 7076 fd_trans_unregister(ret); 7077 return ret; 7078 #endif 7079 case TARGET_NR_close: 7080 fd_trans_unregister(arg1); 7081 return get_errno(close(arg1)); 7082 7083 case TARGET_NR_brk: 7084 return do_brk(arg1); 7085 #ifdef TARGET_NR_fork 7086 case TARGET_NR_fork: 7087 return get_errno(do_fork(cpu_env, TARGET_SIGCHLD, 0, 0, 0, 0)); 7088 #endif 7089 #ifdef TARGET_NR_waitpid 7090 case TARGET_NR_waitpid: 7091 { 7092 int status; 7093 ret = get_errno(safe_wait4(arg1, &status, arg3, 0)); 7094 if (!is_error(ret) && arg2 && ret 7095 && put_user_s32(host_to_target_waitstatus(status), arg2)) 7096 return -TARGET_EFAULT; 7097 } 7098 return ret; 7099 #endif 7100 #ifdef TARGET_NR_waitid 7101 case TARGET_NR_waitid: 7102 { 7103 siginfo_t info; 7104 info.si_pid = 0; 7105 ret = get_errno(safe_waitid(arg1, arg2, &info, arg4, NULL)); 7106 if (!is_error(ret) && arg3 && info.si_pid != 0) { 7107 if (!(p = lock_user(VERIFY_WRITE, arg3, sizeof(target_siginfo_t), 0))) 7108 return -TARGET_EFAULT; 7109 host_to_target_siginfo(p, &info); 7110 unlock_user(p, arg3, sizeof(target_siginfo_t)); 7111 } 7112 } 7113 return ret; 7114 #endif 7115 #ifdef TARGET_NR_creat /* not on alpha */ 7116 case TARGET_NR_creat: 7117 if (!(p = lock_user_string(arg1))) 7118 return -TARGET_EFAULT; 7119 ret = get_errno(creat(p, arg2)); 7120 fd_trans_unregister(ret); 7121 unlock_user(p, arg1, 0); 7122 return ret; 7123 #endif 7124 #ifdef TARGET_NR_link 7125 case TARGET_NR_link: 7126 { 7127 void * p2; 7128 p = lock_user_string(arg1); 7129 p2 = lock_user_string(arg2); 7130 if (!p || !p2) 7131 ret = -TARGET_EFAULT; 7132 else 7133 ret = get_errno(link(p, p2)); 7134 unlock_user(p2, arg2, 0); 7135 unlock_user(p, arg1, 0); 7136 } 7137 return ret; 7138 #endif 7139 #if defined(TARGET_NR_linkat) 7140 case TARGET_NR_linkat: 7141 { 7142 void * p2 = NULL; 7143 if (!arg2 || !arg4) 7144 return -TARGET_EFAULT; 7145 p = lock_user_string(arg2); 7146 p2 = lock_user_string(arg4); 7147 if (!p || !p2) 7148 ret = -TARGET_EFAULT; 7149 else 7150 ret = get_errno(linkat(arg1, p, arg3, p2, arg5)); 7151 unlock_user(p, arg2, 0); 7152 unlock_user(p2, arg4, 0); 7153 } 7154 return ret; 7155 #endif 7156 #ifdef TARGET_NR_unlink 7157 case TARGET_NR_unlink: 7158 if (!(p = lock_user_string(arg1))) 7159 return -TARGET_EFAULT; 7160 ret = get_errno(unlink(p)); 7161 unlock_user(p, arg1, 0); 7162 return ret; 7163 #endif 7164 #if defined(TARGET_NR_unlinkat) 7165 case TARGET_NR_unlinkat: 7166 if (!(p = lock_user_string(arg2))) 7167 return -TARGET_EFAULT; 7168 ret = get_errno(unlinkat(arg1, p, arg3)); 7169 unlock_user(p, arg2, 0); 7170 return ret; 7171 #endif 7172 case TARGET_NR_execve: 7173 { 7174 char **argp, **envp; 7175 int argc, envc; 7176 abi_ulong gp; 7177 abi_ulong guest_argp; 7178 abi_ulong guest_envp; 7179 abi_ulong addr; 7180 char **q; 7181 int total_size = 0; 7182 7183 argc = 0; 7184 guest_argp = arg2; 7185 for (gp = guest_argp; gp; gp += sizeof(abi_ulong)) { 7186 if (get_user_ual(addr, gp)) 7187 return -TARGET_EFAULT; 7188 if (!addr) 7189 break; 7190 argc++; 7191 } 7192 envc = 0; 7193 guest_envp = arg3; 7194 for (gp = guest_envp; gp; gp += sizeof(abi_ulong)) { 7195 if (get_user_ual(addr, gp)) 7196 return -TARGET_EFAULT; 7197 if (!addr) 7198 break; 7199 envc++; 7200 } 7201 7202 argp = g_new0(char *, argc + 1); 7203 envp = g_new0(char *, envc + 1); 7204 7205 for (gp = guest_argp, q = argp; gp; 7206 gp += sizeof(abi_ulong), q++) { 7207 if (get_user_ual(addr, gp)) 7208 goto execve_efault; 7209 if (!addr) 7210 break; 7211 if (!(*q = lock_user_string(addr))) 7212 goto execve_efault; 7213 total_size += strlen(*q) + 1; 7214 } 7215 *q = NULL; 7216 7217 for (gp = guest_envp, q = envp; gp; 7218 gp += sizeof(abi_ulong), q++) { 7219 if (get_user_ual(addr, gp)) 7220 goto execve_efault; 7221 if (!addr) 7222 break; 7223 if (!(*q = lock_user_string(addr))) 7224 goto execve_efault; 7225 total_size += strlen(*q) + 1; 7226 } 7227 *q = NULL; 7228 7229 if (!(p = lock_user_string(arg1))) 7230 goto execve_efault; 7231 /* Although execve() is not an interruptible syscall it is 7232 * a special case where we must use the safe_syscall wrapper: 7233 * if we allow a signal to happen before we make the host 7234 * syscall then we will 'lose' it, because at the point of 7235 * execve the process leaves QEMU's control. So we use the 7236 * safe syscall wrapper to ensure that we either take the 7237 * signal as a guest signal, or else it does not happen 7238 * before the execve completes and makes it the other 7239 * program's problem. 7240 */ 7241 ret = get_errno(safe_execve(p, argp, envp)); 7242 unlock_user(p, arg1, 0); 7243 7244 goto execve_end; 7245 7246 execve_efault: 7247 ret = -TARGET_EFAULT; 7248 7249 execve_end: 7250 for (gp = guest_argp, q = argp; *q; 7251 gp += sizeof(abi_ulong), q++) { 7252 if (get_user_ual(addr, gp) 7253 || !addr) 7254 break; 7255 unlock_user(*q, addr, 0); 7256 } 7257 for (gp = guest_envp, q = envp; *q; 7258 gp += sizeof(abi_ulong), q++) { 7259 if (get_user_ual(addr, gp) 7260 || !addr) 7261 break; 7262 unlock_user(*q, addr, 0); 7263 } 7264 7265 g_free(argp); 7266 g_free(envp); 7267 } 7268 return ret; 7269 case TARGET_NR_chdir: 7270 if (!(p = lock_user_string(arg1))) 7271 return -TARGET_EFAULT; 7272 ret = get_errno(chdir(p)); 7273 unlock_user(p, arg1, 0); 7274 return ret; 7275 #ifdef TARGET_NR_time 7276 case TARGET_NR_time: 7277 { 7278 time_t host_time; 7279 ret = get_errno(time(&host_time)); 7280 if (!is_error(ret) 7281 && arg1 7282 && put_user_sal(host_time, arg1)) 7283 return -TARGET_EFAULT; 7284 } 7285 return ret; 7286 #endif 7287 #ifdef TARGET_NR_mknod 7288 case TARGET_NR_mknod: 7289 if (!(p = lock_user_string(arg1))) 7290 return -TARGET_EFAULT; 7291 ret = get_errno(mknod(p, arg2, arg3)); 7292 unlock_user(p, arg1, 0); 7293 return ret; 7294 #endif 7295 #if defined(TARGET_NR_mknodat) 7296 case TARGET_NR_mknodat: 7297 if (!(p = lock_user_string(arg2))) 7298 return -TARGET_EFAULT; 7299 ret = get_errno(mknodat(arg1, p, arg3, arg4)); 7300 unlock_user(p, arg2, 0); 7301 return ret; 7302 #endif 7303 #ifdef TARGET_NR_chmod 7304 case TARGET_NR_chmod: 7305 if (!(p = lock_user_string(arg1))) 7306 return -TARGET_EFAULT; 7307 ret = get_errno(chmod(p, arg2)); 7308 unlock_user(p, arg1, 0); 7309 return ret; 7310 #endif 7311 #ifdef TARGET_NR_lseek 7312 case TARGET_NR_lseek: 7313 return get_errno(lseek(arg1, arg2, arg3)); 7314 #endif 7315 #if defined(TARGET_NR_getxpid) && defined(TARGET_ALPHA) 7316 /* Alpha specific */ 7317 case TARGET_NR_getxpid: 7318 ((CPUAlphaState *)cpu_env)->ir[IR_A4] = getppid(); 7319 return get_errno(getpid()); 7320 #endif 7321 #ifdef TARGET_NR_getpid 7322 case TARGET_NR_getpid: 7323 return get_errno(getpid()); 7324 #endif 7325 case TARGET_NR_mount: 7326 { 7327 /* need to look at the data field */ 7328 void *p2, *p3; 7329 7330 if (arg1) { 7331 p = lock_user_string(arg1); 7332 if (!p) { 7333 return -TARGET_EFAULT; 7334 } 7335 } else { 7336 p = NULL; 7337 } 7338 7339 p2 = lock_user_string(arg2); 7340 if (!p2) { 7341 if (arg1) { 7342 unlock_user(p, arg1, 0); 7343 } 7344 return -TARGET_EFAULT; 7345 } 7346 7347 if (arg3) { 7348 p3 = lock_user_string(arg3); 7349 if (!p3) { 7350 if (arg1) { 7351 unlock_user(p, arg1, 0); 7352 } 7353 unlock_user(p2, arg2, 0); 7354 return -TARGET_EFAULT; 7355 } 7356 } else { 7357 p3 = NULL; 7358 } 7359 7360 /* FIXME - arg5 should be locked, but it isn't clear how to 7361 * do that since it's not guaranteed to be a NULL-terminated 7362 * string. 7363 */ 7364 if (!arg5) { 7365 ret = mount(p, p2, p3, (unsigned long)arg4, NULL); 7366 } else { 7367 ret = mount(p, p2, p3, (unsigned long)arg4, g2h(arg5)); 7368 } 7369 ret = get_errno(ret); 7370 7371 if (arg1) { 7372 unlock_user(p, arg1, 0); 7373 } 7374 unlock_user(p2, arg2, 0); 7375 if (arg3) { 7376 unlock_user(p3, arg3, 0); 7377 } 7378 } 7379 return ret; 7380 #ifdef TARGET_NR_umount 7381 case TARGET_NR_umount: 7382 if (!(p = lock_user_string(arg1))) 7383 return -TARGET_EFAULT; 7384 ret = get_errno(umount(p)); 7385 unlock_user(p, arg1, 0); 7386 return ret; 7387 #endif 7388 #ifdef TARGET_NR_stime /* not on alpha */ 7389 case TARGET_NR_stime: 7390 { 7391 time_t host_time; 7392 if (get_user_sal(host_time, arg1)) 7393 return -TARGET_EFAULT; 7394 return get_errno(stime(&host_time)); 7395 } 7396 #endif 7397 #ifdef TARGET_NR_alarm /* not on alpha */ 7398 case TARGET_NR_alarm: 7399 return alarm(arg1); 7400 #endif 7401 #ifdef TARGET_NR_pause /* not on alpha */ 7402 case TARGET_NR_pause: 7403 if (!block_signals()) { 7404 sigsuspend(&((TaskState *)cpu->opaque)->signal_mask); 7405 } 7406 return -TARGET_EINTR; 7407 #endif 7408 #ifdef TARGET_NR_utime 7409 case TARGET_NR_utime: 7410 { 7411 struct utimbuf tbuf, *host_tbuf; 7412 struct target_utimbuf *target_tbuf; 7413 if (arg2) { 7414 if (!lock_user_struct(VERIFY_READ, target_tbuf, arg2, 1)) 7415 return -TARGET_EFAULT; 7416 tbuf.actime = tswapal(target_tbuf->actime); 7417 tbuf.modtime = tswapal(target_tbuf->modtime); 7418 unlock_user_struct(target_tbuf, arg2, 0); 7419 host_tbuf = &tbuf; 7420 } else { 7421 host_tbuf = NULL; 7422 } 7423 if (!(p = lock_user_string(arg1))) 7424 return -TARGET_EFAULT; 7425 ret = get_errno(utime(p, host_tbuf)); 7426 unlock_user(p, arg1, 0); 7427 } 7428 return ret; 7429 #endif 7430 #ifdef TARGET_NR_utimes 7431 case TARGET_NR_utimes: 7432 { 7433 struct timeval *tvp, tv[2]; 7434 if (arg2) { 7435 if (copy_from_user_timeval(&tv[0], arg2) 7436 || copy_from_user_timeval(&tv[1], 7437 arg2 + sizeof(struct target_timeval))) 7438 return -TARGET_EFAULT; 7439 tvp = tv; 7440 } else { 7441 tvp = NULL; 7442 } 7443 if (!(p = lock_user_string(arg1))) 7444 return -TARGET_EFAULT; 7445 ret = get_errno(utimes(p, tvp)); 7446 unlock_user(p, arg1, 0); 7447 } 7448 return ret; 7449 #endif 7450 #if defined(TARGET_NR_futimesat) 7451 case TARGET_NR_futimesat: 7452 { 7453 struct timeval *tvp, tv[2]; 7454 if (arg3) { 7455 if (copy_from_user_timeval(&tv[0], arg3) 7456 || copy_from_user_timeval(&tv[1], 7457 arg3 + sizeof(struct target_timeval))) 7458 return -TARGET_EFAULT; 7459 tvp = tv; 7460 } else { 7461 tvp = NULL; 7462 } 7463 if (!(p = lock_user_string(arg2))) { 7464 return -TARGET_EFAULT; 7465 } 7466 ret = get_errno(futimesat(arg1, path(p), tvp)); 7467 unlock_user(p, arg2, 0); 7468 } 7469 return ret; 7470 #endif 7471 #ifdef TARGET_NR_access 7472 case TARGET_NR_access: 7473 if (!(p = lock_user_string(arg1))) { 7474 return -TARGET_EFAULT; 7475 } 7476 ret = get_errno(access(path(p), arg2)); 7477 unlock_user(p, arg1, 0); 7478 return ret; 7479 #endif 7480 #if defined(TARGET_NR_faccessat) && defined(__NR_faccessat) 7481 case TARGET_NR_faccessat: 7482 if (!(p = lock_user_string(arg2))) { 7483 return -TARGET_EFAULT; 7484 } 7485 ret = get_errno(faccessat(arg1, p, arg3, 0)); 7486 unlock_user(p, arg2, 0); 7487 return ret; 7488 #endif 7489 #ifdef TARGET_NR_nice /* not on alpha */ 7490 case TARGET_NR_nice: 7491 return get_errno(nice(arg1)); 7492 #endif 7493 case TARGET_NR_sync: 7494 sync(); 7495 return 0; 7496 #if defined(TARGET_NR_syncfs) && defined(CONFIG_SYNCFS) 7497 case TARGET_NR_syncfs: 7498 return get_errno(syncfs(arg1)); 7499 #endif 7500 case TARGET_NR_kill: 7501 return get_errno(safe_kill(arg1, target_to_host_signal(arg2))); 7502 #ifdef TARGET_NR_rename 7503 case TARGET_NR_rename: 7504 { 7505 void *p2; 7506 p = lock_user_string(arg1); 7507 p2 = lock_user_string(arg2); 7508 if (!p || !p2) 7509 ret = -TARGET_EFAULT; 7510 else 7511 ret = get_errno(rename(p, p2)); 7512 unlock_user(p2, arg2, 0); 7513 unlock_user(p, arg1, 0); 7514 } 7515 return ret; 7516 #endif 7517 #if defined(TARGET_NR_renameat) 7518 case TARGET_NR_renameat: 7519 { 7520 void *p2; 7521 p = lock_user_string(arg2); 7522 p2 = lock_user_string(arg4); 7523 if (!p || !p2) 7524 ret = -TARGET_EFAULT; 7525 else 7526 ret = get_errno(renameat(arg1, p, arg3, p2)); 7527 unlock_user(p2, arg4, 0); 7528 unlock_user(p, arg2, 0); 7529 } 7530 return ret; 7531 #endif 7532 #if defined(TARGET_NR_renameat2) 7533 case TARGET_NR_renameat2: 7534 { 7535 void *p2; 7536 p = lock_user_string(arg2); 7537 p2 = lock_user_string(arg4); 7538 if (!p || !p2) { 7539 ret = -TARGET_EFAULT; 7540 } else { 7541 ret = get_errno(sys_renameat2(arg1, p, arg3, p2, arg5)); 7542 } 7543 unlock_user(p2, arg4, 0); 7544 unlock_user(p, arg2, 0); 7545 } 7546 return ret; 7547 #endif 7548 #ifdef TARGET_NR_mkdir 7549 case TARGET_NR_mkdir: 7550 if (!(p = lock_user_string(arg1))) 7551 return -TARGET_EFAULT; 7552 ret = get_errno(mkdir(p, arg2)); 7553 unlock_user(p, arg1, 0); 7554 return ret; 7555 #endif 7556 #if defined(TARGET_NR_mkdirat) 7557 case TARGET_NR_mkdirat: 7558 if (!(p = lock_user_string(arg2))) 7559 return -TARGET_EFAULT; 7560 ret = get_errno(mkdirat(arg1, p, arg3)); 7561 unlock_user(p, arg2, 0); 7562 return ret; 7563 #endif 7564 #ifdef TARGET_NR_rmdir 7565 case TARGET_NR_rmdir: 7566 if (!(p = lock_user_string(arg1))) 7567 return -TARGET_EFAULT; 7568 ret = get_errno(rmdir(p)); 7569 unlock_user(p, arg1, 0); 7570 return ret; 7571 #endif 7572 case TARGET_NR_dup: 7573 ret = get_errno(dup(arg1)); 7574 if (ret >= 0) { 7575 fd_trans_dup(arg1, ret); 7576 } 7577 return ret; 7578 #ifdef TARGET_NR_pipe 7579 case TARGET_NR_pipe: 7580 return do_pipe(cpu_env, arg1, 0, 0); 7581 #endif 7582 #ifdef TARGET_NR_pipe2 7583 case TARGET_NR_pipe2: 7584 return do_pipe(cpu_env, arg1, 7585 target_to_host_bitmask(arg2, fcntl_flags_tbl), 1); 7586 #endif 7587 case TARGET_NR_times: 7588 { 7589 struct target_tms *tmsp; 7590 struct tms tms; 7591 ret = get_errno(times(&tms)); 7592 if (arg1) { 7593 tmsp = lock_user(VERIFY_WRITE, arg1, sizeof(struct target_tms), 0); 7594 if (!tmsp) 7595 return -TARGET_EFAULT; 7596 tmsp->tms_utime = tswapal(host_to_target_clock_t(tms.tms_utime)); 7597 tmsp->tms_stime = tswapal(host_to_target_clock_t(tms.tms_stime)); 7598 tmsp->tms_cutime = tswapal(host_to_target_clock_t(tms.tms_cutime)); 7599 tmsp->tms_cstime = tswapal(host_to_target_clock_t(tms.tms_cstime)); 7600 } 7601 if (!is_error(ret)) 7602 ret = host_to_target_clock_t(ret); 7603 } 7604 return ret; 7605 case TARGET_NR_acct: 7606 if (arg1 == 0) { 7607 ret = get_errno(acct(NULL)); 7608 } else { 7609 if (!(p = lock_user_string(arg1))) { 7610 return -TARGET_EFAULT; 7611 } 7612 ret = get_errno(acct(path(p))); 7613 unlock_user(p, arg1, 0); 7614 } 7615 return ret; 7616 #ifdef TARGET_NR_umount2 7617 case TARGET_NR_umount2: 7618 if (!(p = lock_user_string(arg1))) 7619 return -TARGET_EFAULT; 7620 ret = get_errno(umount2(p, arg2)); 7621 unlock_user(p, arg1, 0); 7622 return ret; 7623 #endif 7624 case TARGET_NR_ioctl: 7625 return do_ioctl(arg1, arg2, arg3); 7626 #ifdef TARGET_NR_fcntl 7627 case TARGET_NR_fcntl: 7628 return do_fcntl(arg1, arg2, arg3); 7629 #endif 7630 case TARGET_NR_setpgid: 7631 return get_errno(setpgid(arg1, arg2)); 7632 case TARGET_NR_umask: 7633 return get_errno(umask(arg1)); 7634 case TARGET_NR_chroot: 7635 if (!(p = lock_user_string(arg1))) 7636 return -TARGET_EFAULT; 7637 ret = get_errno(chroot(p)); 7638 unlock_user(p, arg1, 0); 7639 return ret; 7640 #ifdef TARGET_NR_dup2 7641 case TARGET_NR_dup2: 7642 ret = get_errno(dup2(arg1, arg2)); 7643 if (ret >= 0) { 7644 fd_trans_dup(arg1, arg2); 7645 } 7646 return ret; 7647 #endif 7648 #if defined(CONFIG_DUP3) && defined(TARGET_NR_dup3) 7649 case TARGET_NR_dup3: 7650 { 7651 int host_flags; 7652 7653 if ((arg3 & ~TARGET_O_CLOEXEC) != 0) { 7654 return -EINVAL; 7655 } 7656 host_flags = target_to_host_bitmask(arg3, fcntl_flags_tbl); 7657 ret = get_errno(dup3(arg1, arg2, host_flags)); 7658 if (ret >= 0) { 7659 fd_trans_dup(arg1, arg2); 7660 } 7661 return ret; 7662 } 7663 #endif 7664 #ifdef TARGET_NR_getppid /* not on alpha */ 7665 case TARGET_NR_getppid: 7666 return get_errno(getppid()); 7667 #endif 7668 #ifdef TARGET_NR_getpgrp 7669 case TARGET_NR_getpgrp: 7670 return get_errno(getpgrp()); 7671 #endif 7672 case TARGET_NR_setsid: 7673 return get_errno(setsid()); 7674 #ifdef TARGET_NR_sigaction 7675 case TARGET_NR_sigaction: 7676 { 7677 #if defined(TARGET_ALPHA) 7678 struct target_sigaction act, oact, *pact = 0; 7679 struct target_old_sigaction *old_act; 7680 if (arg2) { 7681 if (!lock_user_struct(VERIFY_READ, old_act, arg2, 1)) 7682 return -TARGET_EFAULT; 7683 act._sa_handler = old_act->_sa_handler; 7684 target_siginitset(&act.sa_mask, old_act->sa_mask); 7685 act.sa_flags = old_act->sa_flags; 7686 act.sa_restorer = 0; 7687 unlock_user_struct(old_act, arg2, 0); 7688 pact = &act; 7689 } 7690 ret = get_errno(do_sigaction(arg1, pact, &oact)); 7691 if (!is_error(ret) && arg3) { 7692 if (!lock_user_struct(VERIFY_WRITE, old_act, arg3, 0)) 7693 return -TARGET_EFAULT; 7694 old_act->_sa_handler = oact._sa_handler; 7695 old_act->sa_mask = oact.sa_mask.sig[0]; 7696 old_act->sa_flags = oact.sa_flags; 7697 unlock_user_struct(old_act, arg3, 1); 7698 } 7699 #elif defined(TARGET_MIPS) 7700 struct target_sigaction act, oact, *pact, *old_act; 7701 7702 if (arg2) { 7703 if (!lock_user_struct(VERIFY_READ, old_act, arg2, 1)) 7704 return -TARGET_EFAULT; 7705 act._sa_handler = old_act->_sa_handler; 7706 target_siginitset(&act.sa_mask, old_act->sa_mask.sig[0]); 7707 act.sa_flags = old_act->sa_flags; 7708 unlock_user_struct(old_act, arg2, 0); 7709 pact = &act; 7710 } else { 7711 pact = NULL; 7712 } 7713 7714 ret = get_errno(do_sigaction(arg1, pact, &oact)); 7715 7716 if (!is_error(ret) && arg3) { 7717 if (!lock_user_struct(VERIFY_WRITE, old_act, arg3, 0)) 7718 return -TARGET_EFAULT; 7719 old_act->_sa_handler = oact._sa_handler; 7720 old_act->sa_flags = oact.sa_flags; 7721 old_act->sa_mask.sig[0] = oact.sa_mask.sig[0]; 7722 old_act->sa_mask.sig[1] = 0; 7723 old_act->sa_mask.sig[2] = 0; 7724 old_act->sa_mask.sig[3] = 0; 7725 unlock_user_struct(old_act, arg3, 1); 7726 } 7727 #else 7728 struct target_old_sigaction *old_act; 7729 struct target_sigaction act, oact, *pact; 7730 if (arg2) { 7731 if (!lock_user_struct(VERIFY_READ, old_act, arg2, 1)) 7732 return -TARGET_EFAULT; 7733 act._sa_handler = old_act->_sa_handler; 7734 target_siginitset(&act.sa_mask, old_act->sa_mask); 7735 act.sa_flags = old_act->sa_flags; 7736 act.sa_restorer = old_act->sa_restorer; 7737 #ifdef TARGET_ARCH_HAS_KA_RESTORER 7738 act.ka_restorer = 0; 7739 #endif 7740 unlock_user_struct(old_act, arg2, 0); 7741 pact = &act; 7742 } else { 7743 pact = NULL; 7744 } 7745 ret = get_errno(do_sigaction(arg1, pact, &oact)); 7746 if (!is_error(ret) && arg3) { 7747 if (!lock_user_struct(VERIFY_WRITE, old_act, arg3, 0)) 7748 return -TARGET_EFAULT; 7749 old_act->_sa_handler = oact._sa_handler; 7750 old_act->sa_mask = oact.sa_mask.sig[0]; 7751 old_act->sa_flags = oact.sa_flags; 7752 old_act->sa_restorer = oact.sa_restorer; 7753 unlock_user_struct(old_act, arg3, 1); 7754 } 7755 #endif 7756 } 7757 return ret; 7758 #endif 7759 case TARGET_NR_rt_sigaction: 7760 { 7761 #if defined(TARGET_ALPHA) 7762 /* For Alpha and SPARC this is a 5 argument syscall, with 7763 * a 'restorer' parameter which must be copied into the 7764 * sa_restorer field of the sigaction struct. 7765 * For Alpha that 'restorer' is arg5; for SPARC it is arg4, 7766 * and arg5 is the sigsetsize. 7767 * Alpha also has a separate rt_sigaction struct that it uses 7768 * here; SPARC uses the usual sigaction struct. 7769 */ 7770 struct target_rt_sigaction *rt_act; 7771 struct target_sigaction act, oact, *pact = 0; 7772 7773 if (arg4 != sizeof(target_sigset_t)) { 7774 return -TARGET_EINVAL; 7775 } 7776 if (arg2) { 7777 if (!lock_user_struct(VERIFY_READ, rt_act, arg2, 1)) 7778 return -TARGET_EFAULT; 7779 act._sa_handler = rt_act->_sa_handler; 7780 act.sa_mask = rt_act->sa_mask; 7781 act.sa_flags = rt_act->sa_flags; 7782 act.sa_restorer = arg5; 7783 unlock_user_struct(rt_act, arg2, 0); 7784 pact = &act; 7785 } 7786 ret = get_errno(do_sigaction(arg1, pact, &oact)); 7787 if (!is_error(ret) && arg3) { 7788 if (!lock_user_struct(VERIFY_WRITE, rt_act, arg3, 0)) 7789 return -TARGET_EFAULT; 7790 rt_act->_sa_handler = oact._sa_handler; 7791 rt_act->sa_mask = oact.sa_mask; 7792 rt_act->sa_flags = oact.sa_flags; 7793 unlock_user_struct(rt_act, arg3, 1); 7794 } 7795 #else 7796 #ifdef TARGET_SPARC 7797 target_ulong restorer = arg4; 7798 target_ulong sigsetsize = arg5; 7799 #else 7800 target_ulong sigsetsize = arg4; 7801 #endif 7802 struct target_sigaction *act; 7803 struct target_sigaction *oact; 7804 7805 if (sigsetsize != sizeof(target_sigset_t)) { 7806 return -TARGET_EINVAL; 7807 } 7808 if (arg2) { 7809 if (!lock_user_struct(VERIFY_READ, act, arg2, 1)) { 7810 return -TARGET_EFAULT; 7811 } 7812 #ifdef TARGET_ARCH_HAS_KA_RESTORER 7813 act->ka_restorer = restorer; 7814 #endif 7815 } else { 7816 act = NULL; 7817 } 7818 if (arg3) { 7819 if (!lock_user_struct(VERIFY_WRITE, oact, arg3, 0)) { 7820 ret = -TARGET_EFAULT; 7821 goto rt_sigaction_fail; 7822 } 7823 } else 7824 oact = NULL; 7825 ret = get_errno(do_sigaction(arg1, act, oact)); 7826 rt_sigaction_fail: 7827 if (act) 7828 unlock_user_struct(act, arg2, 0); 7829 if (oact) 7830 unlock_user_struct(oact, arg3, 1); 7831 #endif 7832 } 7833 return ret; 7834 #ifdef TARGET_NR_sgetmask /* not on alpha */ 7835 case TARGET_NR_sgetmask: 7836 { 7837 sigset_t cur_set; 7838 abi_ulong target_set; 7839 ret = do_sigprocmask(0, NULL, &cur_set); 7840 if (!ret) { 7841 host_to_target_old_sigset(&target_set, &cur_set); 7842 ret = target_set; 7843 } 7844 } 7845 return ret; 7846 #endif 7847 #ifdef TARGET_NR_ssetmask /* not on alpha */ 7848 case TARGET_NR_ssetmask: 7849 { 7850 sigset_t set, oset; 7851 abi_ulong target_set = arg1; 7852 target_to_host_old_sigset(&set, &target_set); 7853 ret = do_sigprocmask(SIG_SETMASK, &set, &oset); 7854 if (!ret) { 7855 host_to_target_old_sigset(&target_set, &oset); 7856 ret = target_set; 7857 } 7858 } 7859 return ret; 7860 #endif 7861 #ifdef TARGET_NR_sigprocmask 7862 case TARGET_NR_sigprocmask: 7863 { 7864 #if defined(TARGET_ALPHA) 7865 sigset_t set, oldset; 7866 abi_ulong mask; 7867 int how; 7868 7869 switch (arg1) { 7870 case TARGET_SIG_BLOCK: 7871 how = SIG_BLOCK; 7872 break; 7873 case TARGET_SIG_UNBLOCK: 7874 how = SIG_UNBLOCK; 7875 break; 7876 case TARGET_SIG_SETMASK: 7877 how = SIG_SETMASK; 7878 break; 7879 default: 7880 return -TARGET_EINVAL; 7881 } 7882 mask = arg2; 7883 target_to_host_old_sigset(&set, &mask); 7884 7885 ret = do_sigprocmask(how, &set, &oldset); 7886 if (!is_error(ret)) { 7887 host_to_target_old_sigset(&mask, &oldset); 7888 ret = mask; 7889 ((CPUAlphaState *)cpu_env)->ir[IR_V0] = 0; /* force no error */ 7890 } 7891 #else 7892 sigset_t set, oldset, *set_ptr; 7893 int how; 7894 7895 if (arg2) { 7896 switch (arg1) { 7897 case TARGET_SIG_BLOCK: 7898 how = SIG_BLOCK; 7899 break; 7900 case TARGET_SIG_UNBLOCK: 7901 how = SIG_UNBLOCK; 7902 break; 7903 case TARGET_SIG_SETMASK: 7904 how = SIG_SETMASK; 7905 break; 7906 default: 7907 return -TARGET_EINVAL; 7908 } 7909 if (!(p = lock_user(VERIFY_READ, arg2, sizeof(target_sigset_t), 1))) 7910 return -TARGET_EFAULT; 7911 target_to_host_old_sigset(&set, p); 7912 unlock_user(p, arg2, 0); 7913 set_ptr = &set; 7914 } else { 7915 how = 0; 7916 set_ptr = NULL; 7917 } 7918 ret = do_sigprocmask(how, set_ptr, &oldset); 7919 if (!is_error(ret) && arg3) { 7920 if (!(p = lock_user(VERIFY_WRITE, arg3, sizeof(target_sigset_t), 0))) 7921 return -TARGET_EFAULT; 7922 host_to_target_old_sigset(p, &oldset); 7923 unlock_user(p, arg3, sizeof(target_sigset_t)); 7924 } 7925 #endif 7926 } 7927 return ret; 7928 #endif 7929 case TARGET_NR_rt_sigprocmask: 7930 { 7931 int how = arg1; 7932 sigset_t set, oldset, *set_ptr; 7933 7934 if (arg4 != sizeof(target_sigset_t)) { 7935 return -TARGET_EINVAL; 7936 } 7937 7938 if (arg2) { 7939 switch(how) { 7940 case TARGET_SIG_BLOCK: 7941 how = SIG_BLOCK; 7942 break; 7943 case TARGET_SIG_UNBLOCK: 7944 how = SIG_UNBLOCK; 7945 break; 7946 case TARGET_SIG_SETMASK: 7947 how = SIG_SETMASK; 7948 break; 7949 default: 7950 return -TARGET_EINVAL; 7951 } 7952 if (!(p = lock_user(VERIFY_READ, arg2, sizeof(target_sigset_t), 1))) 7953 return -TARGET_EFAULT; 7954 target_to_host_sigset(&set, p); 7955 unlock_user(p, arg2, 0); 7956 set_ptr = &set; 7957 } else { 7958 how = 0; 7959 set_ptr = NULL; 7960 } 7961 ret = do_sigprocmask(how, set_ptr, &oldset); 7962 if (!is_error(ret) && arg3) { 7963 if (!(p = lock_user(VERIFY_WRITE, arg3, sizeof(target_sigset_t), 0))) 7964 return -TARGET_EFAULT; 7965 host_to_target_sigset(p, &oldset); 7966 unlock_user(p, arg3, sizeof(target_sigset_t)); 7967 } 7968 } 7969 return ret; 7970 #ifdef TARGET_NR_sigpending 7971 case TARGET_NR_sigpending: 7972 { 7973 sigset_t set; 7974 ret = get_errno(sigpending(&set)); 7975 if (!is_error(ret)) { 7976 if (!(p = lock_user(VERIFY_WRITE, arg1, sizeof(target_sigset_t), 0))) 7977 return -TARGET_EFAULT; 7978 host_to_target_old_sigset(p, &set); 7979 unlock_user(p, arg1, sizeof(target_sigset_t)); 7980 } 7981 } 7982 return ret; 7983 #endif 7984 case TARGET_NR_rt_sigpending: 7985 { 7986 sigset_t set; 7987 7988 /* Yes, this check is >, not != like most. We follow the kernel's 7989 * logic and it does it like this because it implements 7990 * NR_sigpending through the same code path, and in that case 7991 * the old_sigset_t is smaller in size. 7992 */ 7993 if (arg2 > sizeof(target_sigset_t)) { 7994 return -TARGET_EINVAL; 7995 } 7996 7997 ret = get_errno(sigpending(&set)); 7998 if (!is_error(ret)) { 7999 if (!(p = lock_user(VERIFY_WRITE, arg1, sizeof(target_sigset_t), 0))) 8000 return -TARGET_EFAULT; 8001 host_to_target_sigset(p, &set); 8002 unlock_user(p, arg1, sizeof(target_sigset_t)); 8003 } 8004 } 8005 return ret; 8006 #ifdef TARGET_NR_sigsuspend 8007 case TARGET_NR_sigsuspend: 8008 { 8009 TaskState *ts = cpu->opaque; 8010 #if defined(TARGET_ALPHA) 8011 abi_ulong mask = arg1; 8012 target_to_host_old_sigset(&ts->sigsuspend_mask, &mask); 8013 #else 8014 if (!(p = lock_user(VERIFY_READ, arg1, sizeof(target_sigset_t), 1))) 8015 return -TARGET_EFAULT; 8016 target_to_host_old_sigset(&ts->sigsuspend_mask, p); 8017 unlock_user(p, arg1, 0); 8018 #endif 8019 ret = get_errno(safe_rt_sigsuspend(&ts->sigsuspend_mask, 8020 SIGSET_T_SIZE)); 8021 if (ret != -TARGET_ERESTARTSYS) { 8022 ts->in_sigsuspend = 1; 8023 } 8024 } 8025 return ret; 8026 #endif 8027 case TARGET_NR_rt_sigsuspend: 8028 { 8029 TaskState *ts = cpu->opaque; 8030 8031 if (arg2 != sizeof(target_sigset_t)) { 8032 return -TARGET_EINVAL; 8033 } 8034 if (!(p = lock_user(VERIFY_READ, arg1, sizeof(target_sigset_t), 1))) 8035 return -TARGET_EFAULT; 8036 target_to_host_sigset(&ts->sigsuspend_mask, p); 8037 unlock_user(p, arg1, 0); 8038 ret = get_errno(safe_rt_sigsuspend(&ts->sigsuspend_mask, 8039 SIGSET_T_SIZE)); 8040 if (ret != -TARGET_ERESTARTSYS) { 8041 ts->in_sigsuspend = 1; 8042 } 8043 } 8044 return ret; 8045 case TARGET_NR_rt_sigtimedwait: 8046 { 8047 sigset_t set; 8048 struct timespec uts, *puts; 8049 siginfo_t uinfo; 8050 8051 if (arg4 != sizeof(target_sigset_t)) { 8052 return -TARGET_EINVAL; 8053 } 8054 8055 if (!(p = lock_user(VERIFY_READ, arg1, sizeof(target_sigset_t), 1))) 8056 return -TARGET_EFAULT; 8057 target_to_host_sigset(&set, p); 8058 unlock_user(p, arg1, 0); 8059 if (arg3) { 8060 puts = &uts; 8061 target_to_host_timespec(puts, arg3); 8062 } else { 8063 puts = NULL; 8064 } 8065 ret = get_errno(safe_rt_sigtimedwait(&set, &uinfo, puts, 8066 SIGSET_T_SIZE)); 8067 if (!is_error(ret)) { 8068 if (arg2) { 8069 p = lock_user(VERIFY_WRITE, arg2, sizeof(target_siginfo_t), 8070 0); 8071 if (!p) { 8072 return -TARGET_EFAULT; 8073 } 8074 host_to_target_siginfo(p, &uinfo); 8075 unlock_user(p, arg2, sizeof(target_siginfo_t)); 8076 } 8077 ret = host_to_target_signal(ret); 8078 } 8079 } 8080 return ret; 8081 case TARGET_NR_rt_sigqueueinfo: 8082 { 8083 siginfo_t uinfo; 8084 8085 p = lock_user(VERIFY_READ, arg3, sizeof(target_siginfo_t), 1); 8086 if (!p) { 8087 return -TARGET_EFAULT; 8088 } 8089 target_to_host_siginfo(&uinfo, p); 8090 unlock_user(p, arg3, 0); 8091 ret = get_errno(sys_rt_sigqueueinfo(arg1, arg2, &uinfo)); 8092 } 8093 return ret; 8094 case TARGET_NR_rt_tgsigqueueinfo: 8095 { 8096 siginfo_t uinfo; 8097 8098 p = lock_user(VERIFY_READ, arg4, sizeof(target_siginfo_t), 1); 8099 if (!p) { 8100 return -TARGET_EFAULT; 8101 } 8102 target_to_host_siginfo(&uinfo, p); 8103 unlock_user(p, arg4, 0); 8104 ret = get_errno(sys_rt_tgsigqueueinfo(arg1, arg2, arg3, &uinfo)); 8105 } 8106 return ret; 8107 #ifdef TARGET_NR_sigreturn 8108 case TARGET_NR_sigreturn: 8109 if (block_signals()) { 8110 return -TARGET_ERESTARTSYS; 8111 } 8112 return do_sigreturn(cpu_env); 8113 #endif 8114 case TARGET_NR_rt_sigreturn: 8115 if (block_signals()) { 8116 return -TARGET_ERESTARTSYS; 8117 } 8118 return do_rt_sigreturn(cpu_env); 8119 case TARGET_NR_sethostname: 8120 if (!(p = lock_user_string(arg1))) 8121 return -TARGET_EFAULT; 8122 ret = get_errno(sethostname(p, arg2)); 8123 unlock_user(p, arg1, 0); 8124 return ret; 8125 #ifdef TARGET_NR_setrlimit 8126 case TARGET_NR_setrlimit: 8127 { 8128 int resource = target_to_host_resource(arg1); 8129 struct target_rlimit *target_rlim; 8130 struct rlimit rlim; 8131 if (!lock_user_struct(VERIFY_READ, target_rlim, arg2, 1)) 8132 return -TARGET_EFAULT; 8133 rlim.rlim_cur = target_to_host_rlim(target_rlim->rlim_cur); 8134 rlim.rlim_max = target_to_host_rlim(target_rlim->rlim_max); 8135 unlock_user_struct(target_rlim, arg2, 0); 8136 /* 8137 * If we just passed through resource limit settings for memory then 8138 * they would also apply to QEMU's own allocations, and QEMU will 8139 * crash or hang or die if its allocations fail. Ideally we would 8140 * track the guest allocations in QEMU and apply the limits ourselves. 8141 * For now, just tell the guest the call succeeded but don't actually 8142 * limit anything. 8143 */ 8144 if (resource != RLIMIT_AS && 8145 resource != RLIMIT_DATA && 8146 resource != RLIMIT_STACK) { 8147 return get_errno(setrlimit(resource, &rlim)); 8148 } else { 8149 return 0; 8150 } 8151 } 8152 #endif 8153 #ifdef TARGET_NR_getrlimit 8154 case TARGET_NR_getrlimit: 8155 { 8156 int resource = target_to_host_resource(arg1); 8157 struct target_rlimit *target_rlim; 8158 struct rlimit rlim; 8159 8160 ret = get_errno(getrlimit(resource, &rlim)); 8161 if (!is_error(ret)) { 8162 if (!lock_user_struct(VERIFY_WRITE, target_rlim, arg2, 0)) 8163 return -TARGET_EFAULT; 8164 target_rlim->rlim_cur = host_to_target_rlim(rlim.rlim_cur); 8165 target_rlim->rlim_max = host_to_target_rlim(rlim.rlim_max); 8166 unlock_user_struct(target_rlim, arg2, 1); 8167 } 8168 } 8169 return ret; 8170 #endif 8171 case TARGET_NR_getrusage: 8172 { 8173 struct rusage rusage; 8174 ret = get_errno(getrusage(arg1, &rusage)); 8175 if (!is_error(ret)) { 8176 ret = host_to_target_rusage(arg2, &rusage); 8177 } 8178 } 8179 return ret; 8180 case TARGET_NR_gettimeofday: 8181 { 8182 struct timeval tv; 8183 ret = get_errno(gettimeofday(&tv, NULL)); 8184 if (!is_error(ret)) { 8185 if (copy_to_user_timeval(arg1, &tv)) 8186 return -TARGET_EFAULT; 8187 } 8188 } 8189 return ret; 8190 case TARGET_NR_settimeofday: 8191 { 8192 struct timeval tv, *ptv = NULL; 8193 struct timezone tz, *ptz = NULL; 8194 8195 if (arg1) { 8196 if (copy_from_user_timeval(&tv, arg1)) { 8197 return -TARGET_EFAULT; 8198 } 8199 ptv = &tv; 8200 } 8201 8202 if (arg2) { 8203 if (copy_from_user_timezone(&tz, arg2)) { 8204 return -TARGET_EFAULT; 8205 } 8206 ptz = &tz; 8207 } 8208 8209 return get_errno(settimeofday(ptv, ptz)); 8210 } 8211 #if defined(TARGET_NR_select) 8212 case TARGET_NR_select: 8213 #if defined(TARGET_WANT_NI_OLD_SELECT) 8214 /* some architectures used to have old_select here 8215 * but now ENOSYS it. 8216 */ 8217 ret = -TARGET_ENOSYS; 8218 #elif defined(TARGET_WANT_OLD_SYS_SELECT) 8219 ret = do_old_select(arg1); 8220 #else 8221 ret = do_select(arg1, arg2, arg3, arg4, arg5); 8222 #endif 8223 return ret; 8224 #endif 8225 #ifdef TARGET_NR_pselect6 8226 case TARGET_NR_pselect6: 8227 { 8228 abi_long rfd_addr, wfd_addr, efd_addr, n, ts_addr; 8229 fd_set rfds, wfds, efds; 8230 fd_set *rfds_ptr, *wfds_ptr, *efds_ptr; 8231 struct timespec ts, *ts_ptr; 8232 8233 /* 8234 * The 6th arg is actually two args smashed together, 8235 * so we cannot use the C library. 8236 */ 8237 sigset_t set; 8238 struct { 8239 sigset_t *set; 8240 size_t size; 8241 } sig, *sig_ptr; 8242 8243 abi_ulong arg_sigset, arg_sigsize, *arg7; 8244 target_sigset_t *target_sigset; 8245 8246 n = arg1; 8247 rfd_addr = arg2; 8248 wfd_addr = arg3; 8249 efd_addr = arg4; 8250 ts_addr = arg5; 8251 8252 ret = copy_from_user_fdset_ptr(&rfds, &rfds_ptr, rfd_addr, n); 8253 if (ret) { 8254 return ret; 8255 } 8256 ret = copy_from_user_fdset_ptr(&wfds, &wfds_ptr, wfd_addr, n); 8257 if (ret) { 8258 return ret; 8259 } 8260 ret = copy_from_user_fdset_ptr(&efds, &efds_ptr, efd_addr, n); 8261 if (ret) { 8262 return ret; 8263 } 8264 8265 /* 8266 * This takes a timespec, and not a timeval, so we cannot 8267 * use the do_select() helper ... 8268 */ 8269 if (ts_addr) { 8270 if (target_to_host_timespec(&ts, ts_addr)) { 8271 return -TARGET_EFAULT; 8272 } 8273 ts_ptr = &ts; 8274 } else { 8275 ts_ptr = NULL; 8276 } 8277 8278 /* Extract the two packed args for the sigset */ 8279 if (arg6) { 8280 sig_ptr = &sig; 8281 sig.size = SIGSET_T_SIZE; 8282 8283 arg7 = lock_user(VERIFY_READ, arg6, sizeof(*arg7) * 2, 1); 8284 if (!arg7) { 8285 return -TARGET_EFAULT; 8286 } 8287 arg_sigset = tswapal(arg7[0]); 8288 arg_sigsize = tswapal(arg7[1]); 8289 unlock_user(arg7, arg6, 0); 8290 8291 if (arg_sigset) { 8292 sig.set = &set; 8293 if (arg_sigsize != sizeof(*target_sigset)) { 8294 /* Like the kernel, we enforce correct size sigsets */ 8295 return -TARGET_EINVAL; 8296 } 8297 target_sigset = lock_user(VERIFY_READ, arg_sigset, 8298 sizeof(*target_sigset), 1); 8299 if (!target_sigset) { 8300 return -TARGET_EFAULT; 8301 } 8302 target_to_host_sigset(&set, target_sigset); 8303 unlock_user(target_sigset, arg_sigset, 0); 8304 } else { 8305 sig.set = NULL; 8306 } 8307 } else { 8308 sig_ptr = NULL; 8309 } 8310 8311 ret = get_errno(safe_pselect6(n, rfds_ptr, wfds_ptr, efds_ptr, 8312 ts_ptr, sig_ptr)); 8313 8314 if (!is_error(ret)) { 8315 if (rfd_addr && copy_to_user_fdset(rfd_addr, &rfds, n)) 8316 return -TARGET_EFAULT; 8317 if (wfd_addr && copy_to_user_fdset(wfd_addr, &wfds, n)) 8318 return -TARGET_EFAULT; 8319 if (efd_addr && copy_to_user_fdset(efd_addr, &efds, n)) 8320 return -TARGET_EFAULT; 8321 8322 if (ts_addr && host_to_target_timespec(ts_addr, &ts)) 8323 return -TARGET_EFAULT; 8324 } 8325 } 8326 return ret; 8327 #endif 8328 #ifdef TARGET_NR_symlink 8329 case TARGET_NR_symlink: 8330 { 8331 void *p2; 8332 p = lock_user_string(arg1); 8333 p2 = lock_user_string(arg2); 8334 if (!p || !p2) 8335 ret = -TARGET_EFAULT; 8336 else 8337 ret = get_errno(symlink(p, p2)); 8338 unlock_user(p2, arg2, 0); 8339 unlock_user(p, arg1, 0); 8340 } 8341 return ret; 8342 #endif 8343 #if defined(TARGET_NR_symlinkat) 8344 case TARGET_NR_symlinkat: 8345 { 8346 void *p2; 8347 p = lock_user_string(arg1); 8348 p2 = lock_user_string(arg3); 8349 if (!p || !p2) 8350 ret = -TARGET_EFAULT; 8351 else 8352 ret = get_errno(symlinkat(p, arg2, p2)); 8353 unlock_user(p2, arg3, 0); 8354 unlock_user(p, arg1, 0); 8355 } 8356 return ret; 8357 #endif 8358 #ifdef TARGET_NR_readlink 8359 case TARGET_NR_readlink: 8360 { 8361 void *p2; 8362 p = lock_user_string(arg1); 8363 p2 = lock_user(VERIFY_WRITE, arg2, arg3, 0); 8364 if (!p || !p2) { 8365 ret = -TARGET_EFAULT; 8366 } else if (!arg3) { 8367 /* Short circuit this for the magic exe check. */ 8368 ret = -TARGET_EINVAL; 8369 } else if (is_proc_myself((const char *)p, "exe")) { 8370 char real[PATH_MAX], *temp; 8371 temp = realpath(exec_path, real); 8372 /* Return value is # of bytes that we wrote to the buffer. */ 8373 if (temp == NULL) { 8374 ret = get_errno(-1); 8375 } else { 8376 /* Don't worry about sign mismatch as earlier mapping 8377 * logic would have thrown a bad address error. */ 8378 ret = MIN(strlen(real), arg3); 8379 /* We cannot NUL terminate the string. */ 8380 memcpy(p2, real, ret); 8381 } 8382 } else { 8383 ret = get_errno(readlink(path(p), p2, arg3)); 8384 } 8385 unlock_user(p2, arg2, ret); 8386 unlock_user(p, arg1, 0); 8387 } 8388 return ret; 8389 #endif 8390 #if defined(TARGET_NR_readlinkat) 8391 case TARGET_NR_readlinkat: 8392 { 8393 void *p2; 8394 p = lock_user_string(arg2); 8395 p2 = lock_user(VERIFY_WRITE, arg3, arg4, 0); 8396 if (!p || !p2) { 8397 ret = -TARGET_EFAULT; 8398 } else if (is_proc_myself((const char *)p, "exe")) { 8399 char real[PATH_MAX], *temp; 8400 temp = realpath(exec_path, real); 8401 ret = temp == NULL ? get_errno(-1) : strlen(real) ; 8402 snprintf((char *)p2, arg4, "%s", real); 8403 } else { 8404 ret = get_errno(readlinkat(arg1, path(p), p2, arg4)); 8405 } 8406 unlock_user(p2, arg3, ret); 8407 unlock_user(p, arg2, 0); 8408 } 8409 return ret; 8410 #endif 8411 #ifdef TARGET_NR_swapon 8412 case TARGET_NR_swapon: 8413 if (!(p = lock_user_string(arg1))) 8414 return -TARGET_EFAULT; 8415 ret = get_errno(swapon(p, arg2)); 8416 unlock_user(p, arg1, 0); 8417 return ret; 8418 #endif 8419 case TARGET_NR_reboot: 8420 if (arg3 == LINUX_REBOOT_CMD_RESTART2) { 8421 /* arg4 must be ignored in all other cases */ 8422 p = lock_user_string(arg4); 8423 if (!p) { 8424 return -TARGET_EFAULT; 8425 } 8426 ret = get_errno(reboot(arg1, arg2, arg3, p)); 8427 unlock_user(p, arg4, 0); 8428 } else { 8429 ret = get_errno(reboot(arg1, arg2, arg3, NULL)); 8430 } 8431 return ret; 8432 #ifdef TARGET_NR_mmap 8433 case TARGET_NR_mmap: 8434 #if (defined(TARGET_I386) && defined(TARGET_ABI32)) || \ 8435 (defined(TARGET_ARM) && defined(TARGET_ABI32)) || \ 8436 defined(TARGET_M68K) || defined(TARGET_CRIS) || defined(TARGET_MICROBLAZE) \ 8437 || defined(TARGET_S390X) 8438 { 8439 abi_ulong *v; 8440 abi_ulong v1, v2, v3, v4, v5, v6; 8441 if (!(v = lock_user(VERIFY_READ, arg1, 6 * sizeof(abi_ulong), 1))) 8442 return -TARGET_EFAULT; 8443 v1 = tswapal(v[0]); 8444 v2 = tswapal(v[1]); 8445 v3 = tswapal(v[2]); 8446 v4 = tswapal(v[3]); 8447 v5 = tswapal(v[4]); 8448 v6 = tswapal(v[5]); 8449 unlock_user(v, arg1, 0); 8450 ret = get_errno(target_mmap(v1, v2, v3, 8451 target_to_host_bitmask(v4, mmap_flags_tbl), 8452 v5, v6)); 8453 } 8454 #else 8455 ret = get_errno(target_mmap(arg1, arg2, arg3, 8456 target_to_host_bitmask(arg4, mmap_flags_tbl), 8457 arg5, 8458 arg6)); 8459 #endif 8460 return ret; 8461 #endif 8462 #ifdef TARGET_NR_mmap2 8463 case TARGET_NR_mmap2: 8464 #ifndef MMAP_SHIFT 8465 #define MMAP_SHIFT 12 8466 #endif 8467 ret = target_mmap(arg1, arg2, arg3, 8468 target_to_host_bitmask(arg4, mmap_flags_tbl), 8469 arg5, arg6 << MMAP_SHIFT); 8470 return get_errno(ret); 8471 #endif 8472 case TARGET_NR_munmap: 8473 return get_errno(target_munmap(arg1, arg2)); 8474 case TARGET_NR_mprotect: 8475 { 8476 TaskState *ts = cpu->opaque; 8477 /* Special hack to detect libc making the stack executable. */ 8478 if ((arg3 & PROT_GROWSDOWN) 8479 && arg1 >= ts->info->stack_limit 8480 && arg1 <= ts->info->start_stack) { 8481 arg3 &= ~PROT_GROWSDOWN; 8482 arg2 = arg2 + arg1 - ts->info->stack_limit; 8483 arg1 = ts->info->stack_limit; 8484 } 8485 } 8486 return get_errno(target_mprotect(arg1, arg2, arg3)); 8487 #ifdef TARGET_NR_mremap 8488 case TARGET_NR_mremap: 8489 return get_errno(target_mremap(arg1, arg2, arg3, arg4, arg5)); 8490 #endif 8491 /* ??? msync/mlock/munlock are broken for softmmu. */ 8492 #ifdef TARGET_NR_msync 8493 case TARGET_NR_msync: 8494 return get_errno(msync(g2h(arg1), arg2, arg3)); 8495 #endif 8496 #ifdef TARGET_NR_mlock 8497 case TARGET_NR_mlock: 8498 return get_errno(mlock(g2h(arg1), arg2)); 8499 #endif 8500 #ifdef TARGET_NR_munlock 8501 case TARGET_NR_munlock: 8502 return get_errno(munlock(g2h(arg1), arg2)); 8503 #endif 8504 #ifdef TARGET_NR_mlockall 8505 case TARGET_NR_mlockall: 8506 return get_errno(mlockall(target_to_host_mlockall_arg(arg1))); 8507 #endif 8508 #ifdef TARGET_NR_munlockall 8509 case TARGET_NR_munlockall: 8510 return get_errno(munlockall()); 8511 #endif 8512 #ifdef TARGET_NR_truncate 8513 case TARGET_NR_truncate: 8514 if (!(p = lock_user_string(arg1))) 8515 return -TARGET_EFAULT; 8516 ret = get_errno(truncate(p, arg2)); 8517 unlock_user(p, arg1, 0); 8518 return ret; 8519 #endif 8520 #ifdef TARGET_NR_ftruncate 8521 case TARGET_NR_ftruncate: 8522 return get_errno(ftruncate(arg1, arg2)); 8523 #endif 8524 case TARGET_NR_fchmod: 8525 return get_errno(fchmod(arg1, arg2)); 8526 #if defined(TARGET_NR_fchmodat) 8527 case TARGET_NR_fchmodat: 8528 if (!(p = lock_user_string(arg2))) 8529 return -TARGET_EFAULT; 8530 ret = get_errno(fchmodat(arg1, p, arg3, 0)); 8531 unlock_user(p, arg2, 0); 8532 return ret; 8533 #endif 8534 case TARGET_NR_getpriority: 8535 /* Note that negative values are valid for getpriority, so we must 8536 differentiate based on errno settings. */ 8537 errno = 0; 8538 ret = getpriority(arg1, arg2); 8539 if (ret == -1 && errno != 0) { 8540 return -host_to_target_errno(errno); 8541 } 8542 #ifdef TARGET_ALPHA 8543 /* Return value is the unbiased priority. Signal no error. */ 8544 ((CPUAlphaState *)cpu_env)->ir[IR_V0] = 0; 8545 #else 8546 /* Return value is a biased priority to avoid negative numbers. */ 8547 ret = 20 - ret; 8548 #endif 8549 return ret; 8550 case TARGET_NR_setpriority: 8551 return get_errno(setpriority(arg1, arg2, arg3)); 8552 #ifdef TARGET_NR_statfs 8553 case TARGET_NR_statfs: 8554 if (!(p = lock_user_string(arg1))) { 8555 return -TARGET_EFAULT; 8556 } 8557 ret = get_errno(statfs(path(p), &stfs)); 8558 unlock_user(p, arg1, 0); 8559 convert_statfs: 8560 if (!is_error(ret)) { 8561 struct target_statfs *target_stfs; 8562 8563 if (!lock_user_struct(VERIFY_WRITE, target_stfs, arg2, 0)) 8564 return -TARGET_EFAULT; 8565 __put_user(stfs.f_type, &target_stfs->f_type); 8566 __put_user(stfs.f_bsize, &target_stfs->f_bsize); 8567 __put_user(stfs.f_blocks, &target_stfs->f_blocks); 8568 __put_user(stfs.f_bfree, &target_stfs->f_bfree); 8569 __put_user(stfs.f_bavail, &target_stfs->f_bavail); 8570 __put_user(stfs.f_files, &target_stfs->f_files); 8571 __put_user(stfs.f_ffree, &target_stfs->f_ffree); 8572 __put_user(stfs.f_fsid.__val[0], &target_stfs->f_fsid.val[0]); 8573 __put_user(stfs.f_fsid.__val[1], &target_stfs->f_fsid.val[1]); 8574 __put_user(stfs.f_namelen, &target_stfs->f_namelen); 8575 __put_user(stfs.f_frsize, &target_stfs->f_frsize); 8576 #ifdef _STATFS_F_FLAGS 8577 __put_user(stfs.f_flags, &target_stfs->f_flags); 8578 #else 8579 __put_user(0, &target_stfs->f_flags); 8580 #endif 8581 memset(target_stfs->f_spare, 0, sizeof(target_stfs->f_spare)); 8582 unlock_user_struct(target_stfs, arg2, 1); 8583 } 8584 return ret; 8585 #endif 8586 #ifdef TARGET_NR_fstatfs 8587 case TARGET_NR_fstatfs: 8588 ret = get_errno(fstatfs(arg1, &stfs)); 8589 goto convert_statfs; 8590 #endif 8591 #ifdef TARGET_NR_statfs64 8592 case TARGET_NR_statfs64: 8593 if (!(p = lock_user_string(arg1))) { 8594 return -TARGET_EFAULT; 8595 } 8596 ret = get_errno(statfs(path(p), &stfs)); 8597 unlock_user(p, arg1, 0); 8598 convert_statfs64: 8599 if (!is_error(ret)) { 8600 struct target_statfs64 *target_stfs; 8601 8602 if (!lock_user_struct(VERIFY_WRITE, target_stfs, arg3, 0)) 8603 return -TARGET_EFAULT; 8604 __put_user(stfs.f_type, &target_stfs->f_type); 8605 __put_user(stfs.f_bsize, &target_stfs->f_bsize); 8606 __put_user(stfs.f_blocks, &target_stfs->f_blocks); 8607 __put_user(stfs.f_bfree, &target_stfs->f_bfree); 8608 __put_user(stfs.f_bavail, &target_stfs->f_bavail); 8609 __put_user(stfs.f_files, &target_stfs->f_files); 8610 __put_user(stfs.f_ffree, &target_stfs->f_ffree); 8611 __put_user(stfs.f_fsid.__val[0], &target_stfs->f_fsid.val[0]); 8612 __put_user(stfs.f_fsid.__val[1], &target_stfs->f_fsid.val[1]); 8613 __put_user(stfs.f_namelen, &target_stfs->f_namelen); 8614 __put_user(stfs.f_frsize, &target_stfs->f_frsize); 8615 memset(target_stfs->f_spare, 0, sizeof(target_stfs->f_spare)); 8616 unlock_user_struct(target_stfs, arg3, 1); 8617 } 8618 return ret; 8619 case TARGET_NR_fstatfs64: 8620 ret = get_errno(fstatfs(arg1, &stfs)); 8621 goto convert_statfs64; 8622 #endif 8623 #ifdef TARGET_NR_socketcall 8624 case TARGET_NR_socketcall: 8625 return do_socketcall(arg1, arg2); 8626 #endif 8627 #ifdef TARGET_NR_accept 8628 case TARGET_NR_accept: 8629 return do_accept4(arg1, arg2, arg3, 0); 8630 #endif 8631 #ifdef TARGET_NR_accept4 8632 case TARGET_NR_accept4: 8633 return do_accept4(arg1, arg2, arg3, arg4); 8634 #endif 8635 #ifdef TARGET_NR_bind 8636 case TARGET_NR_bind: 8637 return do_bind(arg1, arg2, arg3); 8638 #endif 8639 #ifdef TARGET_NR_connect 8640 case TARGET_NR_connect: 8641 return do_connect(arg1, arg2, arg3); 8642 #endif 8643 #ifdef TARGET_NR_getpeername 8644 case TARGET_NR_getpeername: 8645 return do_getpeername(arg1, arg2, arg3); 8646 #endif 8647 #ifdef TARGET_NR_getsockname 8648 case TARGET_NR_getsockname: 8649 return do_getsockname(arg1, arg2, arg3); 8650 #endif 8651 #ifdef TARGET_NR_getsockopt 8652 case TARGET_NR_getsockopt: 8653 return do_getsockopt(arg1, arg2, arg3, arg4, arg5); 8654 #endif 8655 #ifdef TARGET_NR_listen 8656 case TARGET_NR_listen: 8657 return get_errno(listen(arg1, arg2)); 8658 #endif 8659 #ifdef TARGET_NR_recv 8660 case TARGET_NR_recv: 8661 return do_recvfrom(arg1, arg2, arg3, arg4, 0, 0); 8662 #endif 8663 #ifdef TARGET_NR_recvfrom 8664 case TARGET_NR_recvfrom: 8665 return do_recvfrom(arg1, arg2, arg3, arg4, arg5, arg6); 8666 #endif 8667 #ifdef TARGET_NR_recvmsg 8668 case TARGET_NR_recvmsg: 8669 return do_sendrecvmsg(arg1, arg2, arg3, 0); 8670 #endif 8671 #ifdef TARGET_NR_send 8672 case TARGET_NR_send: 8673 return do_sendto(arg1, arg2, arg3, arg4, 0, 0); 8674 #endif 8675 #ifdef TARGET_NR_sendmsg 8676 case TARGET_NR_sendmsg: 8677 return do_sendrecvmsg(arg1, arg2, arg3, 1); 8678 #endif 8679 #ifdef TARGET_NR_sendmmsg 8680 case TARGET_NR_sendmmsg: 8681 return do_sendrecvmmsg(arg1, arg2, arg3, arg4, 1); 8682 case TARGET_NR_recvmmsg: 8683 return do_sendrecvmmsg(arg1, arg2, arg3, arg4, 0); 8684 #endif 8685 #ifdef TARGET_NR_sendto 8686 case TARGET_NR_sendto: 8687 return do_sendto(arg1, arg2, arg3, arg4, arg5, arg6); 8688 #endif 8689 #ifdef TARGET_NR_shutdown 8690 case TARGET_NR_shutdown: 8691 return get_errno(shutdown(arg1, arg2)); 8692 #endif 8693 #if defined(TARGET_NR_getrandom) && defined(__NR_getrandom) 8694 case TARGET_NR_getrandom: 8695 p = lock_user(VERIFY_WRITE, arg1, arg2, 0); 8696 if (!p) { 8697 return -TARGET_EFAULT; 8698 } 8699 ret = get_errno(getrandom(p, arg2, arg3)); 8700 unlock_user(p, arg1, ret); 8701 return ret; 8702 #endif 8703 #ifdef TARGET_NR_socket 8704 case TARGET_NR_socket: 8705 return do_socket(arg1, arg2, arg3); 8706 #endif 8707 #ifdef TARGET_NR_socketpair 8708 case TARGET_NR_socketpair: 8709 return do_socketpair(arg1, arg2, arg3, arg4); 8710 #endif 8711 #ifdef TARGET_NR_setsockopt 8712 case TARGET_NR_setsockopt: 8713 return do_setsockopt(arg1, arg2, arg3, arg4, (socklen_t) arg5); 8714 #endif 8715 #if defined(TARGET_NR_syslog) 8716 case TARGET_NR_syslog: 8717 { 8718 int len = arg2; 8719 8720 switch (arg1) { 8721 case TARGET_SYSLOG_ACTION_CLOSE: /* Close log */ 8722 case TARGET_SYSLOG_ACTION_OPEN: /* Open log */ 8723 case TARGET_SYSLOG_ACTION_CLEAR: /* Clear ring buffer */ 8724 case TARGET_SYSLOG_ACTION_CONSOLE_OFF: /* Disable logging */ 8725 case TARGET_SYSLOG_ACTION_CONSOLE_ON: /* Enable logging */ 8726 case TARGET_SYSLOG_ACTION_CONSOLE_LEVEL: /* Set messages level */ 8727 case TARGET_SYSLOG_ACTION_SIZE_UNREAD: /* Number of chars */ 8728 case TARGET_SYSLOG_ACTION_SIZE_BUFFER: /* Size of the buffer */ 8729 return get_errno(sys_syslog((int)arg1, NULL, (int)arg3)); 8730 case TARGET_SYSLOG_ACTION_READ: /* Read from log */ 8731 case TARGET_SYSLOG_ACTION_READ_CLEAR: /* Read/clear msgs */ 8732 case TARGET_SYSLOG_ACTION_READ_ALL: /* Read last messages */ 8733 { 8734 if (len < 0) { 8735 return -TARGET_EINVAL; 8736 } 8737 if (len == 0) { 8738 return 0; 8739 } 8740 p = lock_user(VERIFY_WRITE, arg2, arg3, 0); 8741 if (!p) { 8742 return -TARGET_EFAULT; 8743 } 8744 ret = get_errno(sys_syslog((int)arg1, p, (int)arg3)); 8745 unlock_user(p, arg2, arg3); 8746 } 8747 return ret; 8748 default: 8749 return -TARGET_EINVAL; 8750 } 8751 } 8752 break; 8753 #endif 8754 case TARGET_NR_setitimer: 8755 { 8756 struct itimerval value, ovalue, *pvalue; 8757 8758 if (arg2) { 8759 pvalue = &value; 8760 if (copy_from_user_timeval(&pvalue->it_interval, arg2) 8761 || copy_from_user_timeval(&pvalue->it_value, 8762 arg2 + sizeof(struct target_timeval))) 8763 return -TARGET_EFAULT; 8764 } else { 8765 pvalue = NULL; 8766 } 8767 ret = get_errno(setitimer(arg1, pvalue, &ovalue)); 8768 if (!is_error(ret) && arg3) { 8769 if (copy_to_user_timeval(arg3, 8770 &ovalue.it_interval) 8771 || copy_to_user_timeval(arg3 + sizeof(struct target_timeval), 8772 &ovalue.it_value)) 8773 return -TARGET_EFAULT; 8774 } 8775 } 8776 return ret; 8777 case TARGET_NR_getitimer: 8778 { 8779 struct itimerval value; 8780 8781 ret = get_errno(getitimer(arg1, &value)); 8782 if (!is_error(ret) && arg2) { 8783 if (copy_to_user_timeval(arg2, 8784 &value.it_interval) 8785 || copy_to_user_timeval(arg2 + sizeof(struct target_timeval), 8786 &value.it_value)) 8787 return -TARGET_EFAULT; 8788 } 8789 } 8790 return ret; 8791 #ifdef TARGET_NR_stat 8792 case TARGET_NR_stat: 8793 if (!(p = lock_user_string(arg1))) { 8794 return -TARGET_EFAULT; 8795 } 8796 ret = get_errno(stat(path(p), &st)); 8797 unlock_user(p, arg1, 0); 8798 goto do_stat; 8799 #endif 8800 #ifdef TARGET_NR_lstat 8801 case TARGET_NR_lstat: 8802 if (!(p = lock_user_string(arg1))) { 8803 return -TARGET_EFAULT; 8804 } 8805 ret = get_errno(lstat(path(p), &st)); 8806 unlock_user(p, arg1, 0); 8807 goto do_stat; 8808 #endif 8809 #ifdef TARGET_NR_fstat 8810 case TARGET_NR_fstat: 8811 { 8812 ret = get_errno(fstat(arg1, &st)); 8813 #if defined(TARGET_NR_stat) || defined(TARGET_NR_lstat) 8814 do_stat: 8815 #endif 8816 if (!is_error(ret)) { 8817 struct target_stat *target_st; 8818 8819 if (!lock_user_struct(VERIFY_WRITE, target_st, arg2, 0)) 8820 return -TARGET_EFAULT; 8821 memset(target_st, 0, sizeof(*target_st)); 8822 __put_user(st.st_dev, &target_st->st_dev); 8823 __put_user(st.st_ino, &target_st->st_ino); 8824 __put_user(st.st_mode, &target_st->st_mode); 8825 __put_user(st.st_uid, &target_st->st_uid); 8826 __put_user(st.st_gid, &target_st->st_gid); 8827 __put_user(st.st_nlink, &target_st->st_nlink); 8828 __put_user(st.st_rdev, &target_st->st_rdev); 8829 __put_user(st.st_size, &target_st->st_size); 8830 __put_user(st.st_blksize, &target_st->st_blksize); 8831 __put_user(st.st_blocks, &target_st->st_blocks); 8832 __put_user(st.st_atime, &target_st->target_st_atime); 8833 __put_user(st.st_mtime, &target_st->target_st_mtime); 8834 __put_user(st.st_ctime, &target_st->target_st_ctime); 8835 unlock_user_struct(target_st, arg2, 1); 8836 } 8837 } 8838 return ret; 8839 #endif 8840 case TARGET_NR_vhangup: 8841 return get_errno(vhangup()); 8842 #ifdef TARGET_NR_syscall 8843 case TARGET_NR_syscall: 8844 return do_syscall(cpu_env, arg1 & 0xffff, arg2, arg3, arg4, arg5, 8845 arg6, arg7, arg8, 0); 8846 #endif 8847 case TARGET_NR_wait4: 8848 { 8849 int status; 8850 abi_long status_ptr = arg2; 8851 struct rusage rusage, *rusage_ptr; 8852 abi_ulong target_rusage = arg4; 8853 abi_long rusage_err; 8854 if (target_rusage) 8855 rusage_ptr = &rusage; 8856 else 8857 rusage_ptr = NULL; 8858 ret = get_errno(safe_wait4(arg1, &status, arg3, rusage_ptr)); 8859 if (!is_error(ret)) { 8860 if (status_ptr && ret) { 8861 status = host_to_target_waitstatus(status); 8862 if (put_user_s32(status, status_ptr)) 8863 return -TARGET_EFAULT; 8864 } 8865 if (target_rusage) { 8866 rusage_err = host_to_target_rusage(target_rusage, &rusage); 8867 if (rusage_err) { 8868 ret = rusage_err; 8869 } 8870 } 8871 } 8872 } 8873 return ret; 8874 #ifdef TARGET_NR_swapoff 8875 case TARGET_NR_swapoff: 8876 if (!(p = lock_user_string(arg1))) 8877 return -TARGET_EFAULT; 8878 ret = get_errno(swapoff(p)); 8879 unlock_user(p, arg1, 0); 8880 return ret; 8881 #endif 8882 case TARGET_NR_sysinfo: 8883 { 8884 struct target_sysinfo *target_value; 8885 struct sysinfo value; 8886 ret = get_errno(sysinfo(&value)); 8887 if (!is_error(ret) && arg1) 8888 { 8889 if (!lock_user_struct(VERIFY_WRITE, target_value, arg1, 0)) 8890 return -TARGET_EFAULT; 8891 __put_user(value.uptime, &target_value->uptime); 8892 __put_user(value.loads[0], &target_value->loads[0]); 8893 __put_user(value.loads[1], &target_value->loads[1]); 8894 __put_user(value.loads[2], &target_value->loads[2]); 8895 __put_user(value.totalram, &target_value->totalram); 8896 __put_user(value.freeram, &target_value->freeram); 8897 __put_user(value.sharedram, &target_value->sharedram); 8898 __put_user(value.bufferram, &target_value->bufferram); 8899 __put_user(value.totalswap, &target_value->totalswap); 8900 __put_user(value.freeswap, &target_value->freeswap); 8901 __put_user(value.procs, &target_value->procs); 8902 __put_user(value.totalhigh, &target_value->totalhigh); 8903 __put_user(value.freehigh, &target_value->freehigh); 8904 __put_user(value.mem_unit, &target_value->mem_unit); 8905 unlock_user_struct(target_value, arg1, 1); 8906 } 8907 } 8908 return ret; 8909 #ifdef TARGET_NR_ipc 8910 case TARGET_NR_ipc: 8911 return do_ipc(cpu_env, arg1, arg2, arg3, arg4, arg5, arg6); 8912 #endif 8913 #ifdef TARGET_NR_semget 8914 case TARGET_NR_semget: 8915 return get_errno(semget(arg1, arg2, arg3)); 8916 #endif 8917 #ifdef TARGET_NR_semop 8918 case TARGET_NR_semop: 8919 return do_semop(arg1, arg2, arg3); 8920 #endif 8921 #ifdef TARGET_NR_semctl 8922 case TARGET_NR_semctl: 8923 return do_semctl(arg1, arg2, arg3, arg4); 8924 #endif 8925 #ifdef TARGET_NR_msgctl 8926 case TARGET_NR_msgctl: 8927 return do_msgctl(arg1, arg2, arg3); 8928 #endif 8929 #ifdef TARGET_NR_msgget 8930 case TARGET_NR_msgget: 8931 return get_errno(msgget(arg1, arg2)); 8932 #endif 8933 #ifdef TARGET_NR_msgrcv 8934 case TARGET_NR_msgrcv: 8935 return do_msgrcv(arg1, arg2, arg3, arg4, arg5); 8936 #endif 8937 #ifdef TARGET_NR_msgsnd 8938 case TARGET_NR_msgsnd: 8939 return do_msgsnd(arg1, arg2, arg3, arg4); 8940 #endif 8941 #ifdef TARGET_NR_shmget 8942 case TARGET_NR_shmget: 8943 return get_errno(shmget(arg1, arg2, arg3)); 8944 #endif 8945 #ifdef TARGET_NR_shmctl 8946 case TARGET_NR_shmctl: 8947 return do_shmctl(arg1, arg2, arg3); 8948 #endif 8949 #ifdef TARGET_NR_shmat 8950 case TARGET_NR_shmat: 8951 return do_shmat(cpu_env, arg1, arg2, arg3); 8952 #endif 8953 #ifdef TARGET_NR_shmdt 8954 case TARGET_NR_shmdt: 8955 return do_shmdt(arg1); 8956 #endif 8957 case TARGET_NR_fsync: 8958 return get_errno(fsync(arg1)); 8959 case TARGET_NR_clone: 8960 /* Linux manages to have three different orderings for its 8961 * arguments to clone(); the BACKWARDS and BACKWARDS2 defines 8962 * match the kernel's CONFIG_CLONE_* settings. 8963 * Microblaze is further special in that it uses a sixth 8964 * implicit argument to clone for the TLS pointer. 8965 */ 8966 #if defined(TARGET_MICROBLAZE) 8967 ret = get_errno(do_fork(cpu_env, arg1, arg2, arg4, arg6, arg5)); 8968 #elif defined(TARGET_CLONE_BACKWARDS) 8969 ret = get_errno(do_fork(cpu_env, arg1, arg2, arg3, arg4, arg5)); 8970 #elif defined(TARGET_CLONE_BACKWARDS2) 8971 ret = get_errno(do_fork(cpu_env, arg2, arg1, arg3, arg5, arg4)); 8972 #else 8973 ret = get_errno(do_fork(cpu_env, arg1, arg2, arg3, arg5, arg4)); 8974 #endif 8975 return ret; 8976 #ifdef __NR_exit_group 8977 /* new thread calls */ 8978 case TARGET_NR_exit_group: 8979 preexit_cleanup(cpu_env, arg1); 8980 return get_errno(exit_group(arg1)); 8981 #endif 8982 case TARGET_NR_setdomainname: 8983 if (!(p = lock_user_string(arg1))) 8984 return -TARGET_EFAULT; 8985 ret = get_errno(setdomainname(p, arg2)); 8986 unlock_user(p, arg1, 0); 8987 return ret; 8988 case TARGET_NR_uname: 8989 /* no need to transcode because we use the linux syscall */ 8990 { 8991 struct new_utsname * buf; 8992 8993 if (!lock_user_struct(VERIFY_WRITE, buf, arg1, 0)) 8994 return -TARGET_EFAULT; 8995 ret = get_errno(sys_uname(buf)); 8996 if (!is_error(ret)) { 8997 /* Overwrite the native machine name with whatever is being 8998 emulated. */ 8999 g_strlcpy(buf->machine, cpu_to_uname_machine(cpu_env), 9000 sizeof(buf->machine)); 9001 /* Allow the user to override the reported release. */ 9002 if (qemu_uname_release && *qemu_uname_release) { 9003 g_strlcpy(buf->release, qemu_uname_release, 9004 sizeof(buf->release)); 9005 } 9006 } 9007 unlock_user_struct(buf, arg1, 1); 9008 } 9009 return ret; 9010 #ifdef TARGET_I386 9011 case TARGET_NR_modify_ldt: 9012 return do_modify_ldt(cpu_env, arg1, arg2, arg3); 9013 #if !defined(TARGET_X86_64) 9014 case TARGET_NR_vm86: 9015 return do_vm86(cpu_env, arg1, arg2); 9016 #endif 9017 #endif 9018 case TARGET_NR_adjtimex: 9019 { 9020 struct timex host_buf; 9021 9022 if (target_to_host_timex(&host_buf, arg1) != 0) { 9023 return -TARGET_EFAULT; 9024 } 9025 ret = get_errno(adjtimex(&host_buf)); 9026 if (!is_error(ret)) { 9027 if (host_to_target_timex(arg1, &host_buf) != 0) { 9028 return -TARGET_EFAULT; 9029 } 9030 } 9031 } 9032 return ret; 9033 #if defined(TARGET_NR_clock_adjtime) && defined(CONFIG_CLOCK_ADJTIME) 9034 case TARGET_NR_clock_adjtime: 9035 { 9036 struct timex htx, *phtx = &htx; 9037 9038 if (target_to_host_timex(phtx, arg2) != 0) { 9039 return -TARGET_EFAULT; 9040 } 9041 ret = get_errno(clock_adjtime(arg1, phtx)); 9042 if (!is_error(ret) && phtx) { 9043 if (host_to_target_timex(arg2, phtx) != 0) { 9044 return -TARGET_EFAULT; 9045 } 9046 } 9047 } 9048 return ret; 9049 #endif 9050 case TARGET_NR_getpgid: 9051 return get_errno(getpgid(arg1)); 9052 case TARGET_NR_fchdir: 9053 return get_errno(fchdir(arg1)); 9054 case TARGET_NR_personality: 9055 return get_errno(personality(arg1)); 9056 #ifdef TARGET_NR__llseek /* Not on alpha */ 9057 case TARGET_NR__llseek: 9058 { 9059 int64_t res; 9060 #if !defined(__NR_llseek) 9061 res = lseek(arg1, ((uint64_t)arg2 << 32) | (abi_ulong)arg3, arg5); 9062 if (res == -1) { 9063 ret = get_errno(res); 9064 } else { 9065 ret = 0; 9066 } 9067 #else 9068 ret = get_errno(_llseek(arg1, arg2, arg3, &res, arg5)); 9069 #endif 9070 if ((ret == 0) && put_user_s64(res, arg4)) { 9071 return -TARGET_EFAULT; 9072 } 9073 } 9074 return ret; 9075 #endif 9076 #ifdef TARGET_NR_getdents 9077 case TARGET_NR_getdents: 9078 #ifdef EMULATE_GETDENTS_WITH_GETDENTS 9079 #if TARGET_ABI_BITS == 32 && HOST_LONG_BITS == 64 9080 { 9081 struct target_dirent *target_dirp; 9082 struct linux_dirent *dirp; 9083 abi_long count = arg3; 9084 9085 dirp = g_try_malloc(count); 9086 if (!dirp) { 9087 return -TARGET_ENOMEM; 9088 } 9089 9090 ret = get_errno(sys_getdents(arg1, dirp, count)); 9091 if (!is_error(ret)) { 9092 struct linux_dirent *de; 9093 struct target_dirent *tde; 9094 int len = ret; 9095 int reclen, treclen; 9096 int count1, tnamelen; 9097 9098 count1 = 0; 9099 de = dirp; 9100 if (!(target_dirp = lock_user(VERIFY_WRITE, arg2, count, 0))) 9101 return -TARGET_EFAULT; 9102 tde = target_dirp; 9103 while (len > 0) { 9104 reclen = de->d_reclen; 9105 tnamelen = reclen - offsetof(struct linux_dirent, d_name); 9106 assert(tnamelen >= 0); 9107 treclen = tnamelen + offsetof(struct target_dirent, d_name); 9108 assert(count1 + treclen <= count); 9109 tde->d_reclen = tswap16(treclen); 9110 tde->d_ino = tswapal(de->d_ino); 9111 tde->d_off = tswapal(de->d_off); 9112 memcpy(tde->d_name, de->d_name, tnamelen); 9113 de = (struct linux_dirent *)((char *)de + reclen); 9114 len -= reclen; 9115 tde = (struct target_dirent *)((char *)tde + treclen); 9116 count1 += treclen; 9117 } 9118 ret = count1; 9119 unlock_user(target_dirp, arg2, ret); 9120 } 9121 g_free(dirp); 9122 } 9123 #else 9124 { 9125 struct linux_dirent *dirp; 9126 abi_long count = arg3; 9127 9128 if (!(dirp = lock_user(VERIFY_WRITE, arg2, count, 0))) 9129 return -TARGET_EFAULT; 9130 ret = get_errno(sys_getdents(arg1, dirp, count)); 9131 if (!is_error(ret)) { 9132 struct linux_dirent *de; 9133 int len = ret; 9134 int reclen; 9135 de = dirp; 9136 while (len > 0) { 9137 reclen = de->d_reclen; 9138 if (reclen > len) 9139 break; 9140 de->d_reclen = tswap16(reclen); 9141 tswapls(&de->d_ino); 9142 tswapls(&de->d_off); 9143 de = (struct linux_dirent *)((char *)de + reclen); 9144 len -= reclen; 9145 } 9146 } 9147 unlock_user(dirp, arg2, ret); 9148 } 9149 #endif 9150 #else 9151 /* Implement getdents in terms of getdents64 */ 9152 { 9153 struct linux_dirent64 *dirp; 9154 abi_long count = arg3; 9155 9156 dirp = lock_user(VERIFY_WRITE, arg2, count, 0); 9157 if (!dirp) { 9158 return -TARGET_EFAULT; 9159 } 9160 ret = get_errno(sys_getdents64(arg1, dirp, count)); 9161 if (!is_error(ret)) { 9162 /* Convert the dirent64 structs to target dirent. We do this 9163 * in-place, since we can guarantee that a target_dirent is no 9164 * larger than a dirent64; however this means we have to be 9165 * careful to read everything before writing in the new format. 9166 */ 9167 struct linux_dirent64 *de; 9168 struct target_dirent *tde; 9169 int len = ret; 9170 int tlen = 0; 9171 9172 de = dirp; 9173 tde = (struct target_dirent *)dirp; 9174 while (len > 0) { 9175 int namelen, treclen; 9176 int reclen = de->d_reclen; 9177 uint64_t ino = de->d_ino; 9178 int64_t off = de->d_off; 9179 uint8_t type = de->d_type; 9180 9181 namelen = strlen(de->d_name); 9182 treclen = offsetof(struct target_dirent, d_name) 9183 + namelen + 2; 9184 treclen = QEMU_ALIGN_UP(treclen, sizeof(abi_long)); 9185 9186 memmove(tde->d_name, de->d_name, namelen + 1); 9187 tde->d_ino = tswapal(ino); 9188 tde->d_off = tswapal(off); 9189 tde->d_reclen = tswap16(treclen); 9190 /* The target_dirent type is in what was formerly a padding 9191 * byte at the end of the structure: 9192 */ 9193 *(((char *)tde) + treclen - 1) = type; 9194 9195 de = (struct linux_dirent64 *)((char *)de + reclen); 9196 tde = (struct target_dirent *)((char *)tde + treclen); 9197 len -= reclen; 9198 tlen += treclen; 9199 } 9200 ret = tlen; 9201 } 9202 unlock_user(dirp, arg2, ret); 9203 } 9204 #endif 9205 return ret; 9206 #endif /* TARGET_NR_getdents */ 9207 #if defined(TARGET_NR_getdents64) && defined(__NR_getdents64) 9208 case TARGET_NR_getdents64: 9209 { 9210 struct linux_dirent64 *dirp; 9211 abi_long count = arg3; 9212 if (!(dirp = lock_user(VERIFY_WRITE, arg2, count, 0))) 9213 return -TARGET_EFAULT; 9214 ret = get_errno(sys_getdents64(arg1, dirp, count)); 9215 if (!is_error(ret)) { 9216 struct linux_dirent64 *de; 9217 int len = ret; 9218 int reclen; 9219 de = dirp; 9220 while (len > 0) { 9221 reclen = de->d_reclen; 9222 if (reclen > len) 9223 break; 9224 de->d_reclen = tswap16(reclen); 9225 tswap64s((uint64_t *)&de->d_ino); 9226 tswap64s((uint64_t *)&de->d_off); 9227 de = (struct linux_dirent64 *)((char *)de + reclen); 9228 len -= reclen; 9229 } 9230 } 9231 unlock_user(dirp, arg2, ret); 9232 } 9233 return ret; 9234 #endif /* TARGET_NR_getdents64 */ 9235 #if defined(TARGET_NR__newselect) 9236 case TARGET_NR__newselect: 9237 return do_select(arg1, arg2, arg3, arg4, arg5); 9238 #endif 9239 #if defined(TARGET_NR_poll) || defined(TARGET_NR_ppoll) 9240 # ifdef TARGET_NR_poll 9241 case TARGET_NR_poll: 9242 # endif 9243 # ifdef TARGET_NR_ppoll 9244 case TARGET_NR_ppoll: 9245 # endif 9246 { 9247 struct target_pollfd *target_pfd; 9248 unsigned int nfds = arg2; 9249 struct pollfd *pfd; 9250 unsigned int i; 9251 9252 pfd = NULL; 9253 target_pfd = NULL; 9254 if (nfds) { 9255 if (nfds > (INT_MAX / sizeof(struct target_pollfd))) { 9256 return -TARGET_EINVAL; 9257 } 9258 9259 target_pfd = lock_user(VERIFY_WRITE, arg1, 9260 sizeof(struct target_pollfd) * nfds, 1); 9261 if (!target_pfd) { 9262 return -TARGET_EFAULT; 9263 } 9264 9265 pfd = alloca(sizeof(struct pollfd) * nfds); 9266 for (i = 0; i < nfds; i++) { 9267 pfd[i].fd = tswap32(target_pfd[i].fd); 9268 pfd[i].events = tswap16(target_pfd[i].events); 9269 } 9270 } 9271 9272 switch (num) { 9273 # ifdef TARGET_NR_ppoll 9274 case TARGET_NR_ppoll: 9275 { 9276 struct timespec _timeout_ts, *timeout_ts = &_timeout_ts; 9277 target_sigset_t *target_set; 9278 sigset_t _set, *set = &_set; 9279 9280 if (arg3) { 9281 if (target_to_host_timespec(timeout_ts, arg3)) { 9282 unlock_user(target_pfd, arg1, 0); 9283 return -TARGET_EFAULT; 9284 } 9285 } else { 9286 timeout_ts = NULL; 9287 } 9288 9289 if (arg4) { 9290 if (arg5 != sizeof(target_sigset_t)) { 9291 unlock_user(target_pfd, arg1, 0); 9292 return -TARGET_EINVAL; 9293 } 9294 9295 target_set = lock_user(VERIFY_READ, arg4, sizeof(target_sigset_t), 1); 9296 if (!target_set) { 9297 unlock_user(target_pfd, arg1, 0); 9298 return -TARGET_EFAULT; 9299 } 9300 target_to_host_sigset(set, target_set); 9301 } else { 9302 set = NULL; 9303 } 9304 9305 ret = get_errno(safe_ppoll(pfd, nfds, timeout_ts, 9306 set, SIGSET_T_SIZE)); 9307 9308 if (!is_error(ret) && arg3) { 9309 host_to_target_timespec(arg3, timeout_ts); 9310 } 9311 if (arg4) { 9312 unlock_user(target_set, arg4, 0); 9313 } 9314 break; 9315 } 9316 # endif 9317 # ifdef TARGET_NR_poll 9318 case TARGET_NR_poll: 9319 { 9320 struct timespec ts, *pts; 9321 9322 if (arg3 >= 0) { 9323 /* Convert ms to secs, ns */ 9324 ts.tv_sec = arg3 / 1000; 9325 ts.tv_nsec = (arg3 % 1000) * 1000000LL; 9326 pts = &ts; 9327 } else { 9328 /* -ve poll() timeout means "infinite" */ 9329 pts = NULL; 9330 } 9331 ret = get_errno(safe_ppoll(pfd, nfds, pts, NULL, 0)); 9332 break; 9333 } 9334 # endif 9335 default: 9336 g_assert_not_reached(); 9337 } 9338 9339 if (!is_error(ret)) { 9340 for(i = 0; i < nfds; i++) { 9341 target_pfd[i].revents = tswap16(pfd[i].revents); 9342 } 9343 } 9344 unlock_user(target_pfd, arg1, sizeof(struct target_pollfd) * nfds); 9345 } 9346 return ret; 9347 #endif 9348 case TARGET_NR_flock: 9349 /* NOTE: the flock constant seems to be the same for every 9350 Linux platform */ 9351 return get_errno(safe_flock(arg1, arg2)); 9352 case TARGET_NR_readv: 9353 { 9354 struct iovec *vec = lock_iovec(VERIFY_WRITE, arg2, arg3, 0); 9355 if (vec != NULL) { 9356 ret = get_errno(safe_readv(arg1, vec, arg3)); 9357 unlock_iovec(vec, arg2, arg3, 1); 9358 } else { 9359 ret = -host_to_target_errno(errno); 9360 } 9361 } 9362 return ret; 9363 case TARGET_NR_writev: 9364 { 9365 struct iovec *vec = lock_iovec(VERIFY_READ, arg2, arg3, 1); 9366 if (vec != NULL) { 9367 ret = get_errno(safe_writev(arg1, vec, arg3)); 9368 unlock_iovec(vec, arg2, arg3, 0); 9369 } else { 9370 ret = -host_to_target_errno(errno); 9371 } 9372 } 9373 return ret; 9374 #if defined(TARGET_NR_preadv) 9375 case TARGET_NR_preadv: 9376 { 9377 struct iovec *vec = lock_iovec(VERIFY_WRITE, arg2, arg3, 0); 9378 if (vec != NULL) { 9379 unsigned long low, high; 9380 9381 target_to_host_low_high(arg4, arg5, &low, &high); 9382 ret = get_errno(safe_preadv(arg1, vec, arg3, low, high)); 9383 unlock_iovec(vec, arg2, arg3, 1); 9384 } else { 9385 ret = -host_to_target_errno(errno); 9386 } 9387 } 9388 return ret; 9389 #endif 9390 #if defined(TARGET_NR_pwritev) 9391 case TARGET_NR_pwritev: 9392 { 9393 struct iovec *vec = lock_iovec(VERIFY_READ, arg2, arg3, 1); 9394 if (vec != NULL) { 9395 unsigned long low, high; 9396 9397 target_to_host_low_high(arg4, arg5, &low, &high); 9398 ret = get_errno(safe_pwritev(arg1, vec, arg3, low, high)); 9399 unlock_iovec(vec, arg2, arg3, 0); 9400 } else { 9401 ret = -host_to_target_errno(errno); 9402 } 9403 } 9404 return ret; 9405 #endif 9406 case TARGET_NR_getsid: 9407 return get_errno(getsid(arg1)); 9408 #if defined(TARGET_NR_fdatasync) /* Not on alpha (osf_datasync ?) */ 9409 case TARGET_NR_fdatasync: 9410 return get_errno(fdatasync(arg1)); 9411 #endif 9412 #ifdef TARGET_NR__sysctl 9413 case TARGET_NR__sysctl: 9414 /* We don't implement this, but ENOTDIR is always a safe 9415 return value. */ 9416 return -TARGET_ENOTDIR; 9417 #endif 9418 case TARGET_NR_sched_getaffinity: 9419 { 9420 unsigned int mask_size; 9421 unsigned long *mask; 9422 9423 /* 9424 * sched_getaffinity needs multiples of ulong, so need to take 9425 * care of mismatches between target ulong and host ulong sizes. 9426 */ 9427 if (arg2 & (sizeof(abi_ulong) - 1)) { 9428 return -TARGET_EINVAL; 9429 } 9430 mask_size = (arg2 + (sizeof(*mask) - 1)) & ~(sizeof(*mask) - 1); 9431 9432 mask = alloca(mask_size); 9433 memset(mask, 0, mask_size); 9434 ret = get_errno(sys_sched_getaffinity(arg1, mask_size, mask)); 9435 9436 if (!is_error(ret)) { 9437 if (ret > arg2) { 9438 /* More data returned than the caller's buffer will fit. 9439 * This only happens if sizeof(abi_long) < sizeof(long) 9440 * and the caller passed us a buffer holding an odd number 9441 * of abi_longs. If the host kernel is actually using the 9442 * extra 4 bytes then fail EINVAL; otherwise we can just 9443 * ignore them and only copy the interesting part. 9444 */ 9445 int numcpus = sysconf(_SC_NPROCESSORS_CONF); 9446 if (numcpus > arg2 * 8) { 9447 return -TARGET_EINVAL; 9448 } 9449 ret = arg2; 9450 } 9451 9452 if (host_to_target_cpu_mask(mask, mask_size, arg3, ret)) { 9453 return -TARGET_EFAULT; 9454 } 9455 } 9456 } 9457 return ret; 9458 case TARGET_NR_sched_setaffinity: 9459 { 9460 unsigned int mask_size; 9461 unsigned long *mask; 9462 9463 /* 9464 * sched_setaffinity needs multiples of ulong, so need to take 9465 * care of mismatches between target ulong and host ulong sizes. 9466 */ 9467 if (arg2 & (sizeof(abi_ulong) - 1)) { 9468 return -TARGET_EINVAL; 9469 } 9470 mask_size = (arg2 + (sizeof(*mask) - 1)) & ~(sizeof(*mask) - 1); 9471 mask = alloca(mask_size); 9472 9473 ret = target_to_host_cpu_mask(mask, mask_size, arg3, arg2); 9474 if (ret) { 9475 return ret; 9476 } 9477 9478 return get_errno(sys_sched_setaffinity(arg1, mask_size, mask)); 9479 } 9480 case TARGET_NR_getcpu: 9481 { 9482 unsigned cpu, node; 9483 ret = get_errno(sys_getcpu(arg1 ? &cpu : NULL, 9484 arg2 ? &node : NULL, 9485 NULL)); 9486 if (is_error(ret)) { 9487 return ret; 9488 } 9489 if (arg1 && put_user_u32(cpu, arg1)) { 9490 return -TARGET_EFAULT; 9491 } 9492 if (arg2 && put_user_u32(node, arg2)) { 9493 return -TARGET_EFAULT; 9494 } 9495 } 9496 return ret; 9497 case TARGET_NR_sched_setparam: 9498 { 9499 struct sched_param *target_schp; 9500 struct sched_param schp; 9501 9502 if (arg2 == 0) { 9503 return -TARGET_EINVAL; 9504 } 9505 if (!lock_user_struct(VERIFY_READ, target_schp, arg2, 1)) 9506 return -TARGET_EFAULT; 9507 schp.sched_priority = tswap32(target_schp->sched_priority); 9508 unlock_user_struct(target_schp, arg2, 0); 9509 return get_errno(sched_setparam(arg1, &schp)); 9510 } 9511 case TARGET_NR_sched_getparam: 9512 { 9513 struct sched_param *target_schp; 9514 struct sched_param schp; 9515 9516 if (arg2 == 0) { 9517 return -TARGET_EINVAL; 9518 } 9519 ret = get_errno(sched_getparam(arg1, &schp)); 9520 if (!is_error(ret)) { 9521 if (!lock_user_struct(VERIFY_WRITE, target_schp, arg2, 0)) 9522 return -TARGET_EFAULT; 9523 target_schp->sched_priority = tswap32(schp.sched_priority); 9524 unlock_user_struct(target_schp, arg2, 1); 9525 } 9526 } 9527 return ret; 9528 case TARGET_NR_sched_setscheduler: 9529 { 9530 struct sched_param *target_schp; 9531 struct sched_param schp; 9532 if (arg3 == 0) { 9533 return -TARGET_EINVAL; 9534 } 9535 if (!lock_user_struct(VERIFY_READ, target_schp, arg3, 1)) 9536 return -TARGET_EFAULT; 9537 schp.sched_priority = tswap32(target_schp->sched_priority); 9538 unlock_user_struct(target_schp, arg3, 0); 9539 return get_errno(sched_setscheduler(arg1, arg2, &schp)); 9540 } 9541 case TARGET_NR_sched_getscheduler: 9542 return get_errno(sched_getscheduler(arg1)); 9543 case TARGET_NR_sched_yield: 9544 return get_errno(sched_yield()); 9545 case TARGET_NR_sched_get_priority_max: 9546 return get_errno(sched_get_priority_max(arg1)); 9547 case TARGET_NR_sched_get_priority_min: 9548 return get_errno(sched_get_priority_min(arg1)); 9549 case TARGET_NR_sched_rr_get_interval: 9550 { 9551 struct timespec ts; 9552 ret = get_errno(sched_rr_get_interval(arg1, &ts)); 9553 if (!is_error(ret)) { 9554 ret = host_to_target_timespec(arg2, &ts); 9555 } 9556 } 9557 return ret; 9558 case TARGET_NR_nanosleep: 9559 { 9560 struct timespec req, rem; 9561 target_to_host_timespec(&req, arg1); 9562 ret = get_errno(safe_nanosleep(&req, &rem)); 9563 if (is_error(ret) && arg2) { 9564 host_to_target_timespec(arg2, &rem); 9565 } 9566 } 9567 return ret; 9568 case TARGET_NR_prctl: 9569 switch (arg1) { 9570 case PR_GET_PDEATHSIG: 9571 { 9572 int deathsig; 9573 ret = get_errno(prctl(arg1, &deathsig, arg3, arg4, arg5)); 9574 if (!is_error(ret) && arg2 9575 && put_user_ual(deathsig, arg2)) { 9576 return -TARGET_EFAULT; 9577 } 9578 return ret; 9579 } 9580 #ifdef PR_GET_NAME 9581 case PR_GET_NAME: 9582 { 9583 void *name = lock_user(VERIFY_WRITE, arg2, 16, 1); 9584 if (!name) { 9585 return -TARGET_EFAULT; 9586 } 9587 ret = get_errno(prctl(arg1, (unsigned long)name, 9588 arg3, arg4, arg5)); 9589 unlock_user(name, arg2, 16); 9590 return ret; 9591 } 9592 case PR_SET_NAME: 9593 { 9594 void *name = lock_user(VERIFY_READ, arg2, 16, 1); 9595 if (!name) { 9596 return -TARGET_EFAULT; 9597 } 9598 ret = get_errno(prctl(arg1, (unsigned long)name, 9599 arg3, arg4, arg5)); 9600 unlock_user(name, arg2, 0); 9601 return ret; 9602 } 9603 #endif 9604 #ifdef TARGET_MIPS 9605 case TARGET_PR_GET_FP_MODE: 9606 { 9607 CPUMIPSState *env = ((CPUMIPSState *)cpu_env); 9608 ret = 0; 9609 if (env->CP0_Status & (1 << CP0St_FR)) { 9610 ret |= TARGET_PR_FP_MODE_FR; 9611 } 9612 if (env->CP0_Config5 & (1 << CP0C5_FRE)) { 9613 ret |= TARGET_PR_FP_MODE_FRE; 9614 } 9615 return ret; 9616 } 9617 case TARGET_PR_SET_FP_MODE: 9618 { 9619 CPUMIPSState *env = ((CPUMIPSState *)cpu_env); 9620 bool old_fr = env->CP0_Status & (1 << CP0St_FR); 9621 bool old_fre = env->CP0_Config5 & (1 << CP0C5_FRE); 9622 bool new_fr = arg2 & TARGET_PR_FP_MODE_FR; 9623 bool new_fre = arg2 & TARGET_PR_FP_MODE_FRE; 9624 9625 const unsigned int known_bits = TARGET_PR_FP_MODE_FR | 9626 TARGET_PR_FP_MODE_FRE; 9627 9628 /* If nothing to change, return right away, successfully. */ 9629 if (old_fr == new_fr && old_fre == new_fre) { 9630 return 0; 9631 } 9632 /* Check the value is valid */ 9633 if (arg2 & ~known_bits) { 9634 return -TARGET_EOPNOTSUPP; 9635 } 9636 /* Setting FRE without FR is not supported. */ 9637 if (new_fre && !new_fr) { 9638 return -TARGET_EOPNOTSUPP; 9639 } 9640 if (new_fr && !(env->active_fpu.fcr0 & (1 << FCR0_F64))) { 9641 /* FR1 is not supported */ 9642 return -TARGET_EOPNOTSUPP; 9643 } 9644 if (!new_fr && (env->active_fpu.fcr0 & (1 << FCR0_F64)) 9645 && !(env->CP0_Status_rw_bitmask & (1 << CP0St_FR))) { 9646 /* cannot set FR=0 */ 9647 return -TARGET_EOPNOTSUPP; 9648 } 9649 if (new_fre && !(env->active_fpu.fcr0 & (1 << FCR0_FREP))) { 9650 /* Cannot set FRE=1 */ 9651 return -TARGET_EOPNOTSUPP; 9652 } 9653 9654 int i; 9655 fpr_t *fpr = env->active_fpu.fpr; 9656 for (i = 0; i < 32 ; i += 2) { 9657 if (!old_fr && new_fr) { 9658 fpr[i].w[!FP_ENDIAN_IDX] = fpr[i + 1].w[FP_ENDIAN_IDX]; 9659 } else if (old_fr && !new_fr) { 9660 fpr[i + 1].w[FP_ENDIAN_IDX] = fpr[i].w[!FP_ENDIAN_IDX]; 9661 } 9662 } 9663 9664 if (new_fr) { 9665 env->CP0_Status |= (1 << CP0St_FR); 9666 env->hflags |= MIPS_HFLAG_F64; 9667 } else { 9668 env->CP0_Status &= ~(1 << CP0St_FR); 9669 env->hflags &= ~MIPS_HFLAG_F64; 9670 } 9671 if (new_fre) { 9672 env->CP0_Config5 |= (1 << CP0C5_FRE); 9673 if (env->active_fpu.fcr0 & (1 << FCR0_FREP)) { 9674 env->hflags |= MIPS_HFLAG_FRE; 9675 } 9676 } else { 9677 env->CP0_Config5 &= ~(1 << CP0C5_FRE); 9678 env->hflags &= ~MIPS_HFLAG_FRE; 9679 } 9680 9681 return 0; 9682 } 9683 #endif /* MIPS */ 9684 #ifdef TARGET_AARCH64 9685 case TARGET_PR_SVE_SET_VL: 9686 /* 9687 * We cannot support either PR_SVE_SET_VL_ONEXEC or 9688 * PR_SVE_VL_INHERIT. Note the kernel definition 9689 * of sve_vl_valid allows for VQ=512, i.e. VL=8192, 9690 * even though the current architectural maximum is VQ=16. 9691 */ 9692 ret = -TARGET_EINVAL; 9693 if (cpu_isar_feature(aa64_sve, arm_env_get_cpu(cpu_env)) 9694 && arg2 >= 0 && arg2 <= 512 * 16 && !(arg2 & 15)) { 9695 CPUARMState *env = cpu_env; 9696 ARMCPU *cpu = arm_env_get_cpu(env); 9697 uint32_t vq, old_vq; 9698 9699 old_vq = (env->vfp.zcr_el[1] & 0xf) + 1; 9700 vq = MAX(arg2 / 16, 1); 9701 vq = MIN(vq, cpu->sve_max_vq); 9702 9703 if (vq < old_vq) { 9704 aarch64_sve_narrow_vq(env, vq); 9705 } 9706 env->vfp.zcr_el[1] = vq - 1; 9707 ret = vq * 16; 9708 } 9709 return ret; 9710 case TARGET_PR_SVE_GET_VL: 9711 ret = -TARGET_EINVAL; 9712 { 9713 ARMCPU *cpu = arm_env_get_cpu(cpu_env); 9714 if (cpu_isar_feature(aa64_sve, cpu)) { 9715 ret = ((cpu->env.vfp.zcr_el[1] & 0xf) + 1) * 16; 9716 } 9717 } 9718 return ret; 9719 case TARGET_PR_PAC_RESET_KEYS: 9720 { 9721 CPUARMState *env = cpu_env; 9722 ARMCPU *cpu = arm_env_get_cpu(env); 9723 9724 if (arg3 || arg4 || arg5) { 9725 return -TARGET_EINVAL; 9726 } 9727 if (cpu_isar_feature(aa64_pauth, cpu)) { 9728 int all = (TARGET_PR_PAC_APIAKEY | TARGET_PR_PAC_APIBKEY | 9729 TARGET_PR_PAC_APDAKEY | TARGET_PR_PAC_APDBKEY | 9730 TARGET_PR_PAC_APGAKEY); 9731 if (arg2 == 0) { 9732 arg2 = all; 9733 } else if (arg2 & ~all) { 9734 return -TARGET_EINVAL; 9735 } 9736 if (arg2 & TARGET_PR_PAC_APIAKEY) { 9737 arm_init_pauth_key(&env->apia_key); 9738 } 9739 if (arg2 & TARGET_PR_PAC_APIBKEY) { 9740 arm_init_pauth_key(&env->apib_key); 9741 } 9742 if (arg2 & TARGET_PR_PAC_APDAKEY) { 9743 arm_init_pauth_key(&env->apda_key); 9744 } 9745 if (arg2 & TARGET_PR_PAC_APDBKEY) { 9746 arm_init_pauth_key(&env->apdb_key); 9747 } 9748 if (arg2 & TARGET_PR_PAC_APGAKEY) { 9749 arm_init_pauth_key(&env->apga_key); 9750 } 9751 return 0; 9752 } 9753 } 9754 return -TARGET_EINVAL; 9755 #endif /* AARCH64 */ 9756 case PR_GET_SECCOMP: 9757 case PR_SET_SECCOMP: 9758 /* Disable seccomp to prevent the target disabling syscalls we 9759 * need. */ 9760 return -TARGET_EINVAL; 9761 default: 9762 /* Most prctl options have no pointer arguments */ 9763 return get_errno(prctl(arg1, arg2, arg3, arg4, arg5)); 9764 } 9765 break; 9766 #ifdef TARGET_NR_arch_prctl 9767 case TARGET_NR_arch_prctl: 9768 #if defined(TARGET_I386) && !defined(TARGET_ABI32) 9769 return do_arch_prctl(cpu_env, arg1, arg2); 9770 #else 9771 #error unreachable 9772 #endif 9773 #endif 9774 #ifdef TARGET_NR_pread64 9775 case TARGET_NR_pread64: 9776 if (regpairs_aligned(cpu_env, num)) { 9777 arg4 = arg5; 9778 arg5 = arg6; 9779 } 9780 if (arg2 == 0 && arg3 == 0) { 9781 /* Special-case NULL buffer and zero length, which should succeed */ 9782 p = 0; 9783 } else { 9784 p = lock_user(VERIFY_WRITE, arg2, arg3, 0); 9785 if (!p) { 9786 return -TARGET_EFAULT; 9787 } 9788 } 9789 ret = get_errno(pread64(arg1, p, arg3, target_offset64(arg4, arg5))); 9790 unlock_user(p, arg2, ret); 9791 return ret; 9792 case TARGET_NR_pwrite64: 9793 if (regpairs_aligned(cpu_env, num)) { 9794 arg4 = arg5; 9795 arg5 = arg6; 9796 } 9797 if (arg2 == 0 && arg3 == 0) { 9798 /* Special-case NULL buffer and zero length, which should succeed */ 9799 p = 0; 9800 } else { 9801 p = lock_user(VERIFY_READ, arg2, arg3, 1); 9802 if (!p) { 9803 return -TARGET_EFAULT; 9804 } 9805 } 9806 ret = get_errno(pwrite64(arg1, p, arg3, target_offset64(arg4, arg5))); 9807 unlock_user(p, arg2, 0); 9808 return ret; 9809 #endif 9810 case TARGET_NR_getcwd: 9811 if (!(p = lock_user(VERIFY_WRITE, arg1, arg2, 0))) 9812 return -TARGET_EFAULT; 9813 ret = get_errno(sys_getcwd1(p, arg2)); 9814 unlock_user(p, arg1, ret); 9815 return ret; 9816 case TARGET_NR_capget: 9817 case TARGET_NR_capset: 9818 { 9819 struct target_user_cap_header *target_header; 9820 struct target_user_cap_data *target_data = NULL; 9821 struct __user_cap_header_struct header; 9822 struct __user_cap_data_struct data[2]; 9823 struct __user_cap_data_struct *dataptr = NULL; 9824 int i, target_datalen; 9825 int data_items = 1; 9826 9827 if (!lock_user_struct(VERIFY_WRITE, target_header, arg1, 1)) { 9828 return -TARGET_EFAULT; 9829 } 9830 header.version = tswap32(target_header->version); 9831 header.pid = tswap32(target_header->pid); 9832 9833 if (header.version != _LINUX_CAPABILITY_VERSION) { 9834 /* Version 2 and up takes pointer to two user_data structs */ 9835 data_items = 2; 9836 } 9837 9838 target_datalen = sizeof(*target_data) * data_items; 9839 9840 if (arg2) { 9841 if (num == TARGET_NR_capget) { 9842 target_data = lock_user(VERIFY_WRITE, arg2, target_datalen, 0); 9843 } else { 9844 target_data = lock_user(VERIFY_READ, arg2, target_datalen, 1); 9845 } 9846 if (!target_data) { 9847 unlock_user_struct(target_header, arg1, 0); 9848 return -TARGET_EFAULT; 9849 } 9850 9851 if (num == TARGET_NR_capset) { 9852 for (i = 0; i < data_items; i++) { 9853 data[i].effective = tswap32(target_data[i].effective); 9854 data[i].permitted = tswap32(target_data[i].permitted); 9855 data[i].inheritable = tswap32(target_data[i].inheritable); 9856 } 9857 } 9858 9859 dataptr = data; 9860 } 9861 9862 if (num == TARGET_NR_capget) { 9863 ret = get_errno(capget(&header, dataptr)); 9864 } else { 9865 ret = get_errno(capset(&header, dataptr)); 9866 } 9867 9868 /* The kernel always updates version for both capget and capset */ 9869 target_header->version = tswap32(header.version); 9870 unlock_user_struct(target_header, arg1, 1); 9871 9872 if (arg2) { 9873 if (num == TARGET_NR_capget) { 9874 for (i = 0; i < data_items; i++) { 9875 target_data[i].effective = tswap32(data[i].effective); 9876 target_data[i].permitted = tswap32(data[i].permitted); 9877 target_data[i].inheritable = tswap32(data[i].inheritable); 9878 } 9879 unlock_user(target_data, arg2, target_datalen); 9880 } else { 9881 unlock_user(target_data, arg2, 0); 9882 } 9883 } 9884 return ret; 9885 } 9886 case TARGET_NR_sigaltstack: 9887 return do_sigaltstack(arg1, arg2, 9888 get_sp_from_cpustate((CPUArchState *)cpu_env)); 9889 9890 #ifdef CONFIG_SENDFILE 9891 #ifdef TARGET_NR_sendfile 9892 case TARGET_NR_sendfile: 9893 { 9894 off_t *offp = NULL; 9895 off_t off; 9896 if (arg3) { 9897 ret = get_user_sal(off, arg3); 9898 if (is_error(ret)) { 9899 return ret; 9900 } 9901 offp = &off; 9902 } 9903 ret = get_errno(sendfile(arg1, arg2, offp, arg4)); 9904 if (!is_error(ret) && arg3) { 9905 abi_long ret2 = put_user_sal(off, arg3); 9906 if (is_error(ret2)) { 9907 ret = ret2; 9908 } 9909 } 9910 return ret; 9911 } 9912 #endif 9913 #ifdef TARGET_NR_sendfile64 9914 case TARGET_NR_sendfile64: 9915 { 9916 off_t *offp = NULL; 9917 off_t off; 9918 if (arg3) { 9919 ret = get_user_s64(off, arg3); 9920 if (is_error(ret)) { 9921 return ret; 9922 } 9923 offp = &off; 9924 } 9925 ret = get_errno(sendfile(arg1, arg2, offp, arg4)); 9926 if (!is_error(ret) && arg3) { 9927 abi_long ret2 = put_user_s64(off, arg3); 9928 if (is_error(ret2)) { 9929 ret = ret2; 9930 } 9931 } 9932 return ret; 9933 } 9934 #endif 9935 #endif 9936 #ifdef TARGET_NR_vfork 9937 case TARGET_NR_vfork: 9938 return get_errno(do_fork(cpu_env, 9939 CLONE_VFORK | CLONE_VM | TARGET_SIGCHLD, 9940 0, 0, 0, 0)); 9941 #endif 9942 #ifdef TARGET_NR_ugetrlimit 9943 case TARGET_NR_ugetrlimit: 9944 { 9945 struct rlimit rlim; 9946 int resource = target_to_host_resource(arg1); 9947 ret = get_errno(getrlimit(resource, &rlim)); 9948 if (!is_error(ret)) { 9949 struct target_rlimit *target_rlim; 9950 if (!lock_user_struct(VERIFY_WRITE, target_rlim, arg2, 0)) 9951 return -TARGET_EFAULT; 9952 target_rlim->rlim_cur = host_to_target_rlim(rlim.rlim_cur); 9953 target_rlim->rlim_max = host_to_target_rlim(rlim.rlim_max); 9954 unlock_user_struct(target_rlim, arg2, 1); 9955 } 9956 return ret; 9957 } 9958 #endif 9959 #ifdef TARGET_NR_truncate64 9960 case TARGET_NR_truncate64: 9961 if (!(p = lock_user_string(arg1))) 9962 return -TARGET_EFAULT; 9963 ret = target_truncate64(cpu_env, p, arg2, arg3, arg4); 9964 unlock_user(p, arg1, 0); 9965 return ret; 9966 #endif 9967 #ifdef TARGET_NR_ftruncate64 9968 case TARGET_NR_ftruncate64: 9969 return target_ftruncate64(cpu_env, arg1, arg2, arg3, arg4); 9970 #endif 9971 #ifdef TARGET_NR_stat64 9972 case TARGET_NR_stat64: 9973 if (!(p = lock_user_string(arg1))) { 9974 return -TARGET_EFAULT; 9975 } 9976 ret = get_errno(stat(path(p), &st)); 9977 unlock_user(p, arg1, 0); 9978 if (!is_error(ret)) 9979 ret = host_to_target_stat64(cpu_env, arg2, &st); 9980 return ret; 9981 #endif 9982 #ifdef TARGET_NR_lstat64 9983 case TARGET_NR_lstat64: 9984 if (!(p = lock_user_string(arg1))) { 9985 return -TARGET_EFAULT; 9986 } 9987 ret = get_errno(lstat(path(p), &st)); 9988 unlock_user(p, arg1, 0); 9989 if (!is_error(ret)) 9990 ret = host_to_target_stat64(cpu_env, arg2, &st); 9991 return ret; 9992 #endif 9993 #ifdef TARGET_NR_fstat64 9994 case TARGET_NR_fstat64: 9995 ret = get_errno(fstat(arg1, &st)); 9996 if (!is_error(ret)) 9997 ret = host_to_target_stat64(cpu_env, arg2, &st); 9998 return ret; 9999 #endif 10000 #if (defined(TARGET_NR_fstatat64) || defined(TARGET_NR_newfstatat)) 10001 #ifdef TARGET_NR_fstatat64 10002 case TARGET_NR_fstatat64: 10003 #endif 10004 #ifdef TARGET_NR_newfstatat 10005 case TARGET_NR_newfstatat: 10006 #endif 10007 if (!(p = lock_user_string(arg2))) { 10008 return -TARGET_EFAULT; 10009 } 10010 ret = get_errno(fstatat(arg1, path(p), &st, arg4)); 10011 unlock_user(p, arg2, 0); 10012 if (!is_error(ret)) 10013 ret = host_to_target_stat64(cpu_env, arg3, &st); 10014 return ret; 10015 #endif 10016 #ifdef TARGET_NR_lchown 10017 case TARGET_NR_lchown: 10018 if (!(p = lock_user_string(arg1))) 10019 return -TARGET_EFAULT; 10020 ret = get_errno(lchown(p, low2highuid(arg2), low2highgid(arg3))); 10021 unlock_user(p, arg1, 0); 10022 return ret; 10023 #endif 10024 #ifdef TARGET_NR_getuid 10025 case TARGET_NR_getuid: 10026 return get_errno(high2lowuid(getuid())); 10027 #endif 10028 #ifdef TARGET_NR_getgid 10029 case TARGET_NR_getgid: 10030 return get_errno(high2lowgid(getgid())); 10031 #endif 10032 #ifdef TARGET_NR_geteuid 10033 case TARGET_NR_geteuid: 10034 return get_errno(high2lowuid(geteuid())); 10035 #endif 10036 #ifdef TARGET_NR_getegid 10037 case TARGET_NR_getegid: 10038 return get_errno(high2lowgid(getegid())); 10039 #endif 10040 case TARGET_NR_setreuid: 10041 return get_errno(setreuid(low2highuid(arg1), low2highuid(arg2))); 10042 case TARGET_NR_setregid: 10043 return get_errno(setregid(low2highgid(arg1), low2highgid(arg2))); 10044 case TARGET_NR_getgroups: 10045 { 10046 int gidsetsize = arg1; 10047 target_id *target_grouplist; 10048 gid_t *grouplist; 10049 int i; 10050 10051 grouplist = alloca(gidsetsize * sizeof(gid_t)); 10052 ret = get_errno(getgroups(gidsetsize, grouplist)); 10053 if (gidsetsize == 0) 10054 return ret; 10055 if (!is_error(ret)) { 10056 target_grouplist = lock_user(VERIFY_WRITE, arg2, gidsetsize * sizeof(target_id), 0); 10057 if (!target_grouplist) 10058 return -TARGET_EFAULT; 10059 for(i = 0;i < ret; i++) 10060 target_grouplist[i] = tswapid(high2lowgid(grouplist[i])); 10061 unlock_user(target_grouplist, arg2, gidsetsize * sizeof(target_id)); 10062 } 10063 } 10064 return ret; 10065 case TARGET_NR_setgroups: 10066 { 10067 int gidsetsize = arg1; 10068 target_id *target_grouplist; 10069 gid_t *grouplist = NULL; 10070 int i; 10071 if (gidsetsize) { 10072 grouplist = alloca(gidsetsize * sizeof(gid_t)); 10073 target_grouplist = lock_user(VERIFY_READ, arg2, gidsetsize * sizeof(target_id), 1); 10074 if (!target_grouplist) { 10075 return -TARGET_EFAULT; 10076 } 10077 for (i = 0; i < gidsetsize; i++) { 10078 grouplist[i] = low2highgid(tswapid(target_grouplist[i])); 10079 } 10080 unlock_user(target_grouplist, arg2, 0); 10081 } 10082 return get_errno(setgroups(gidsetsize, grouplist)); 10083 } 10084 case TARGET_NR_fchown: 10085 return get_errno(fchown(arg1, low2highuid(arg2), low2highgid(arg3))); 10086 #if defined(TARGET_NR_fchownat) 10087 case TARGET_NR_fchownat: 10088 if (!(p = lock_user_string(arg2))) 10089 return -TARGET_EFAULT; 10090 ret = get_errno(fchownat(arg1, p, low2highuid(arg3), 10091 low2highgid(arg4), arg5)); 10092 unlock_user(p, arg2, 0); 10093 return ret; 10094 #endif 10095 #ifdef TARGET_NR_setresuid 10096 case TARGET_NR_setresuid: 10097 return get_errno(sys_setresuid(low2highuid(arg1), 10098 low2highuid(arg2), 10099 low2highuid(arg3))); 10100 #endif 10101 #ifdef TARGET_NR_getresuid 10102 case TARGET_NR_getresuid: 10103 { 10104 uid_t ruid, euid, suid; 10105 ret = get_errno(getresuid(&ruid, &euid, &suid)); 10106 if (!is_error(ret)) { 10107 if (put_user_id(high2lowuid(ruid), arg1) 10108 || put_user_id(high2lowuid(euid), arg2) 10109 || put_user_id(high2lowuid(suid), arg3)) 10110 return -TARGET_EFAULT; 10111 } 10112 } 10113 return ret; 10114 #endif 10115 #ifdef TARGET_NR_getresgid 10116 case TARGET_NR_setresgid: 10117 return get_errno(sys_setresgid(low2highgid(arg1), 10118 low2highgid(arg2), 10119 low2highgid(arg3))); 10120 #endif 10121 #ifdef TARGET_NR_getresgid 10122 case TARGET_NR_getresgid: 10123 { 10124 gid_t rgid, egid, sgid; 10125 ret = get_errno(getresgid(&rgid, &egid, &sgid)); 10126 if (!is_error(ret)) { 10127 if (put_user_id(high2lowgid(rgid), arg1) 10128 || put_user_id(high2lowgid(egid), arg2) 10129 || put_user_id(high2lowgid(sgid), arg3)) 10130 return -TARGET_EFAULT; 10131 } 10132 } 10133 return ret; 10134 #endif 10135 #ifdef TARGET_NR_chown 10136 case TARGET_NR_chown: 10137 if (!(p = lock_user_string(arg1))) 10138 return -TARGET_EFAULT; 10139 ret = get_errno(chown(p, low2highuid(arg2), low2highgid(arg3))); 10140 unlock_user(p, arg1, 0); 10141 return ret; 10142 #endif 10143 case TARGET_NR_setuid: 10144 return get_errno(sys_setuid(low2highuid(arg1))); 10145 case TARGET_NR_setgid: 10146 return get_errno(sys_setgid(low2highgid(arg1))); 10147 case TARGET_NR_setfsuid: 10148 return get_errno(setfsuid(arg1)); 10149 case TARGET_NR_setfsgid: 10150 return get_errno(setfsgid(arg1)); 10151 10152 #ifdef TARGET_NR_lchown32 10153 case TARGET_NR_lchown32: 10154 if (!(p = lock_user_string(arg1))) 10155 return -TARGET_EFAULT; 10156 ret = get_errno(lchown(p, arg2, arg3)); 10157 unlock_user(p, arg1, 0); 10158 return ret; 10159 #endif 10160 #ifdef TARGET_NR_getuid32 10161 case TARGET_NR_getuid32: 10162 return get_errno(getuid()); 10163 #endif 10164 10165 #if defined(TARGET_NR_getxuid) && defined(TARGET_ALPHA) 10166 /* Alpha specific */ 10167 case TARGET_NR_getxuid: 10168 { 10169 uid_t euid; 10170 euid=geteuid(); 10171 ((CPUAlphaState *)cpu_env)->ir[IR_A4]=euid; 10172 } 10173 return get_errno(getuid()); 10174 #endif 10175 #if defined(TARGET_NR_getxgid) && defined(TARGET_ALPHA) 10176 /* Alpha specific */ 10177 case TARGET_NR_getxgid: 10178 { 10179 uid_t egid; 10180 egid=getegid(); 10181 ((CPUAlphaState *)cpu_env)->ir[IR_A4]=egid; 10182 } 10183 return get_errno(getgid()); 10184 #endif 10185 #if defined(TARGET_NR_osf_getsysinfo) && defined(TARGET_ALPHA) 10186 /* Alpha specific */ 10187 case TARGET_NR_osf_getsysinfo: 10188 ret = -TARGET_EOPNOTSUPP; 10189 switch (arg1) { 10190 case TARGET_GSI_IEEE_FP_CONTROL: 10191 { 10192 uint64_t swcr, fpcr = cpu_alpha_load_fpcr (cpu_env); 10193 10194 /* Copied from linux ieee_fpcr_to_swcr. */ 10195 swcr = (fpcr >> 35) & SWCR_STATUS_MASK; 10196 swcr |= (fpcr >> 36) & SWCR_MAP_DMZ; 10197 swcr |= (~fpcr >> 48) & (SWCR_TRAP_ENABLE_INV 10198 | SWCR_TRAP_ENABLE_DZE 10199 | SWCR_TRAP_ENABLE_OVF); 10200 swcr |= (~fpcr >> 57) & (SWCR_TRAP_ENABLE_UNF 10201 | SWCR_TRAP_ENABLE_INE); 10202 swcr |= (fpcr >> 47) & SWCR_MAP_UMZ; 10203 swcr |= (~fpcr >> 41) & SWCR_TRAP_ENABLE_DNO; 10204 10205 if (put_user_u64 (swcr, arg2)) 10206 return -TARGET_EFAULT; 10207 ret = 0; 10208 } 10209 break; 10210 10211 /* case GSI_IEEE_STATE_AT_SIGNAL: 10212 -- Not implemented in linux kernel. 10213 case GSI_UACPROC: 10214 -- Retrieves current unaligned access state; not much used. 10215 case GSI_PROC_TYPE: 10216 -- Retrieves implver information; surely not used. 10217 case GSI_GET_HWRPB: 10218 -- Grabs a copy of the HWRPB; surely not used. 10219 */ 10220 } 10221 return ret; 10222 #endif 10223 #if defined(TARGET_NR_osf_setsysinfo) && defined(TARGET_ALPHA) 10224 /* Alpha specific */ 10225 case TARGET_NR_osf_setsysinfo: 10226 ret = -TARGET_EOPNOTSUPP; 10227 switch (arg1) { 10228 case TARGET_SSI_IEEE_FP_CONTROL: 10229 { 10230 uint64_t swcr, fpcr, orig_fpcr; 10231 10232 if (get_user_u64 (swcr, arg2)) { 10233 return -TARGET_EFAULT; 10234 } 10235 orig_fpcr = cpu_alpha_load_fpcr(cpu_env); 10236 fpcr = orig_fpcr & FPCR_DYN_MASK; 10237 10238 /* Copied from linux ieee_swcr_to_fpcr. */ 10239 fpcr |= (swcr & SWCR_STATUS_MASK) << 35; 10240 fpcr |= (swcr & SWCR_MAP_DMZ) << 36; 10241 fpcr |= (~swcr & (SWCR_TRAP_ENABLE_INV 10242 | SWCR_TRAP_ENABLE_DZE 10243 | SWCR_TRAP_ENABLE_OVF)) << 48; 10244 fpcr |= (~swcr & (SWCR_TRAP_ENABLE_UNF 10245 | SWCR_TRAP_ENABLE_INE)) << 57; 10246 fpcr |= (swcr & SWCR_MAP_UMZ ? FPCR_UNDZ | FPCR_UNFD : 0); 10247 fpcr |= (~swcr & SWCR_TRAP_ENABLE_DNO) << 41; 10248 10249 cpu_alpha_store_fpcr(cpu_env, fpcr); 10250 ret = 0; 10251 } 10252 break; 10253 10254 case TARGET_SSI_IEEE_RAISE_EXCEPTION: 10255 { 10256 uint64_t exc, fpcr, orig_fpcr; 10257 int si_code; 10258 10259 if (get_user_u64(exc, arg2)) { 10260 return -TARGET_EFAULT; 10261 } 10262 10263 orig_fpcr = cpu_alpha_load_fpcr(cpu_env); 10264 10265 /* We only add to the exception status here. */ 10266 fpcr = orig_fpcr | ((exc & SWCR_STATUS_MASK) << 35); 10267 10268 cpu_alpha_store_fpcr(cpu_env, fpcr); 10269 ret = 0; 10270 10271 /* Old exceptions are not signaled. */ 10272 fpcr &= ~(orig_fpcr & FPCR_STATUS_MASK); 10273 10274 /* If any exceptions set by this call, 10275 and are unmasked, send a signal. */ 10276 si_code = 0; 10277 if ((fpcr & (FPCR_INE | FPCR_INED)) == FPCR_INE) { 10278 si_code = TARGET_FPE_FLTRES; 10279 } 10280 if ((fpcr & (FPCR_UNF | FPCR_UNFD)) == FPCR_UNF) { 10281 si_code = TARGET_FPE_FLTUND; 10282 } 10283 if ((fpcr & (FPCR_OVF | FPCR_OVFD)) == FPCR_OVF) { 10284 si_code = TARGET_FPE_FLTOVF; 10285 } 10286 if ((fpcr & (FPCR_DZE | FPCR_DZED)) == FPCR_DZE) { 10287 si_code = TARGET_FPE_FLTDIV; 10288 } 10289 if ((fpcr & (FPCR_INV | FPCR_INVD)) == FPCR_INV) { 10290 si_code = TARGET_FPE_FLTINV; 10291 } 10292 if (si_code != 0) { 10293 target_siginfo_t info; 10294 info.si_signo = SIGFPE; 10295 info.si_errno = 0; 10296 info.si_code = si_code; 10297 info._sifields._sigfault._addr 10298 = ((CPUArchState *)cpu_env)->pc; 10299 queue_signal((CPUArchState *)cpu_env, info.si_signo, 10300 QEMU_SI_FAULT, &info); 10301 } 10302 } 10303 break; 10304 10305 /* case SSI_NVPAIRS: 10306 -- Used with SSIN_UACPROC to enable unaligned accesses. 10307 case SSI_IEEE_STATE_AT_SIGNAL: 10308 case SSI_IEEE_IGNORE_STATE_AT_SIGNAL: 10309 -- Not implemented in linux kernel 10310 */ 10311 } 10312 return ret; 10313 #endif 10314 #ifdef TARGET_NR_osf_sigprocmask 10315 /* Alpha specific. */ 10316 case TARGET_NR_osf_sigprocmask: 10317 { 10318 abi_ulong mask; 10319 int how; 10320 sigset_t set, oldset; 10321 10322 switch(arg1) { 10323 case TARGET_SIG_BLOCK: 10324 how = SIG_BLOCK; 10325 break; 10326 case TARGET_SIG_UNBLOCK: 10327 how = SIG_UNBLOCK; 10328 break; 10329 case TARGET_SIG_SETMASK: 10330 how = SIG_SETMASK; 10331 break; 10332 default: 10333 return -TARGET_EINVAL; 10334 } 10335 mask = arg2; 10336 target_to_host_old_sigset(&set, &mask); 10337 ret = do_sigprocmask(how, &set, &oldset); 10338 if (!ret) { 10339 host_to_target_old_sigset(&mask, &oldset); 10340 ret = mask; 10341 } 10342 } 10343 return ret; 10344 #endif 10345 10346 #ifdef TARGET_NR_getgid32 10347 case TARGET_NR_getgid32: 10348 return get_errno(getgid()); 10349 #endif 10350 #ifdef TARGET_NR_geteuid32 10351 case TARGET_NR_geteuid32: 10352 return get_errno(geteuid()); 10353 #endif 10354 #ifdef TARGET_NR_getegid32 10355 case TARGET_NR_getegid32: 10356 return get_errno(getegid()); 10357 #endif 10358 #ifdef TARGET_NR_setreuid32 10359 case TARGET_NR_setreuid32: 10360 return get_errno(setreuid(arg1, arg2)); 10361 #endif 10362 #ifdef TARGET_NR_setregid32 10363 case TARGET_NR_setregid32: 10364 return get_errno(setregid(arg1, arg2)); 10365 #endif 10366 #ifdef TARGET_NR_getgroups32 10367 case TARGET_NR_getgroups32: 10368 { 10369 int gidsetsize = arg1; 10370 uint32_t *target_grouplist; 10371 gid_t *grouplist; 10372 int i; 10373 10374 grouplist = alloca(gidsetsize * sizeof(gid_t)); 10375 ret = get_errno(getgroups(gidsetsize, grouplist)); 10376 if (gidsetsize == 0) 10377 return ret; 10378 if (!is_error(ret)) { 10379 target_grouplist = lock_user(VERIFY_WRITE, arg2, gidsetsize * 4, 0); 10380 if (!target_grouplist) { 10381 return -TARGET_EFAULT; 10382 } 10383 for(i = 0;i < ret; i++) 10384 target_grouplist[i] = tswap32(grouplist[i]); 10385 unlock_user(target_grouplist, arg2, gidsetsize * 4); 10386 } 10387 } 10388 return ret; 10389 #endif 10390 #ifdef TARGET_NR_setgroups32 10391 case TARGET_NR_setgroups32: 10392 { 10393 int gidsetsize = arg1; 10394 uint32_t *target_grouplist; 10395 gid_t *grouplist; 10396 int i; 10397 10398 grouplist = alloca(gidsetsize * sizeof(gid_t)); 10399 target_grouplist = lock_user(VERIFY_READ, arg2, gidsetsize * 4, 1); 10400 if (!target_grouplist) { 10401 return -TARGET_EFAULT; 10402 } 10403 for(i = 0;i < gidsetsize; i++) 10404 grouplist[i] = tswap32(target_grouplist[i]); 10405 unlock_user(target_grouplist, arg2, 0); 10406 return get_errno(setgroups(gidsetsize, grouplist)); 10407 } 10408 #endif 10409 #ifdef TARGET_NR_fchown32 10410 case TARGET_NR_fchown32: 10411 return get_errno(fchown(arg1, arg2, arg3)); 10412 #endif 10413 #ifdef TARGET_NR_setresuid32 10414 case TARGET_NR_setresuid32: 10415 return get_errno(sys_setresuid(arg1, arg2, arg3)); 10416 #endif 10417 #ifdef TARGET_NR_getresuid32 10418 case TARGET_NR_getresuid32: 10419 { 10420 uid_t ruid, euid, suid; 10421 ret = get_errno(getresuid(&ruid, &euid, &suid)); 10422 if (!is_error(ret)) { 10423 if (put_user_u32(ruid, arg1) 10424 || put_user_u32(euid, arg2) 10425 || put_user_u32(suid, arg3)) 10426 return -TARGET_EFAULT; 10427 } 10428 } 10429 return ret; 10430 #endif 10431 #ifdef TARGET_NR_setresgid32 10432 case TARGET_NR_setresgid32: 10433 return get_errno(sys_setresgid(arg1, arg2, arg3)); 10434 #endif 10435 #ifdef TARGET_NR_getresgid32 10436 case TARGET_NR_getresgid32: 10437 { 10438 gid_t rgid, egid, sgid; 10439 ret = get_errno(getresgid(&rgid, &egid, &sgid)); 10440 if (!is_error(ret)) { 10441 if (put_user_u32(rgid, arg1) 10442 || put_user_u32(egid, arg2) 10443 || put_user_u32(sgid, arg3)) 10444 return -TARGET_EFAULT; 10445 } 10446 } 10447 return ret; 10448 #endif 10449 #ifdef TARGET_NR_chown32 10450 case TARGET_NR_chown32: 10451 if (!(p = lock_user_string(arg1))) 10452 return -TARGET_EFAULT; 10453 ret = get_errno(chown(p, arg2, arg3)); 10454 unlock_user(p, arg1, 0); 10455 return ret; 10456 #endif 10457 #ifdef TARGET_NR_setuid32 10458 case TARGET_NR_setuid32: 10459 return get_errno(sys_setuid(arg1)); 10460 #endif 10461 #ifdef TARGET_NR_setgid32 10462 case TARGET_NR_setgid32: 10463 return get_errno(sys_setgid(arg1)); 10464 #endif 10465 #ifdef TARGET_NR_setfsuid32 10466 case TARGET_NR_setfsuid32: 10467 return get_errno(setfsuid(arg1)); 10468 #endif 10469 #ifdef TARGET_NR_setfsgid32 10470 case TARGET_NR_setfsgid32: 10471 return get_errno(setfsgid(arg1)); 10472 #endif 10473 #ifdef TARGET_NR_mincore 10474 case TARGET_NR_mincore: 10475 { 10476 void *a = lock_user(VERIFY_READ, arg1, arg2, 0); 10477 if (!a) { 10478 return -TARGET_ENOMEM; 10479 } 10480 p = lock_user_string(arg3); 10481 if (!p) { 10482 ret = -TARGET_EFAULT; 10483 } else { 10484 ret = get_errno(mincore(a, arg2, p)); 10485 unlock_user(p, arg3, ret); 10486 } 10487 unlock_user(a, arg1, 0); 10488 } 10489 return ret; 10490 #endif 10491 #ifdef TARGET_NR_arm_fadvise64_64 10492 case TARGET_NR_arm_fadvise64_64: 10493 /* arm_fadvise64_64 looks like fadvise64_64 but 10494 * with different argument order: fd, advice, offset, len 10495 * rather than the usual fd, offset, len, advice. 10496 * Note that offset and len are both 64-bit so appear as 10497 * pairs of 32-bit registers. 10498 */ 10499 ret = posix_fadvise(arg1, target_offset64(arg3, arg4), 10500 target_offset64(arg5, arg6), arg2); 10501 return -host_to_target_errno(ret); 10502 #endif 10503 10504 #if TARGET_ABI_BITS == 32 10505 10506 #ifdef TARGET_NR_fadvise64_64 10507 case TARGET_NR_fadvise64_64: 10508 #if defined(TARGET_PPC) || defined(TARGET_XTENSA) 10509 /* 6 args: fd, advice, offset (high, low), len (high, low) */ 10510 ret = arg2; 10511 arg2 = arg3; 10512 arg3 = arg4; 10513 arg4 = arg5; 10514 arg5 = arg6; 10515 arg6 = ret; 10516 #else 10517 /* 6 args: fd, offset (high, low), len (high, low), advice */ 10518 if (regpairs_aligned(cpu_env, num)) { 10519 /* offset is in (3,4), len in (5,6) and advice in 7 */ 10520 arg2 = arg3; 10521 arg3 = arg4; 10522 arg4 = arg5; 10523 arg5 = arg6; 10524 arg6 = arg7; 10525 } 10526 #endif 10527 ret = posix_fadvise(arg1, target_offset64(arg2, arg3), 10528 target_offset64(arg4, arg5), arg6); 10529 return -host_to_target_errno(ret); 10530 #endif 10531 10532 #ifdef TARGET_NR_fadvise64 10533 case TARGET_NR_fadvise64: 10534 /* 5 args: fd, offset (high, low), len, advice */ 10535 if (regpairs_aligned(cpu_env, num)) { 10536 /* offset is in (3,4), len in 5 and advice in 6 */ 10537 arg2 = arg3; 10538 arg3 = arg4; 10539 arg4 = arg5; 10540 arg5 = arg6; 10541 } 10542 ret = posix_fadvise(arg1, target_offset64(arg2, arg3), arg4, arg5); 10543 return -host_to_target_errno(ret); 10544 #endif 10545 10546 #else /* not a 32-bit ABI */ 10547 #if defined(TARGET_NR_fadvise64_64) || defined(TARGET_NR_fadvise64) 10548 #ifdef TARGET_NR_fadvise64_64 10549 case TARGET_NR_fadvise64_64: 10550 #endif 10551 #ifdef TARGET_NR_fadvise64 10552 case TARGET_NR_fadvise64: 10553 #endif 10554 #ifdef TARGET_S390X 10555 switch (arg4) { 10556 case 4: arg4 = POSIX_FADV_NOREUSE + 1; break; /* make sure it's an invalid value */ 10557 case 5: arg4 = POSIX_FADV_NOREUSE + 2; break; /* ditto */ 10558 case 6: arg4 = POSIX_FADV_DONTNEED; break; 10559 case 7: arg4 = POSIX_FADV_NOREUSE; break; 10560 default: break; 10561 } 10562 #endif 10563 return -host_to_target_errno(posix_fadvise(arg1, arg2, arg3, arg4)); 10564 #endif 10565 #endif /* end of 64-bit ABI fadvise handling */ 10566 10567 #ifdef TARGET_NR_madvise 10568 case TARGET_NR_madvise: 10569 /* A straight passthrough may not be safe because qemu sometimes 10570 turns private file-backed mappings into anonymous mappings. 10571 This will break MADV_DONTNEED. 10572 This is a hint, so ignoring and returning success is ok. */ 10573 return 0; 10574 #endif 10575 #if TARGET_ABI_BITS == 32 10576 case TARGET_NR_fcntl64: 10577 { 10578 int cmd; 10579 struct flock64 fl; 10580 from_flock64_fn *copyfrom = copy_from_user_flock64; 10581 to_flock64_fn *copyto = copy_to_user_flock64; 10582 10583 #ifdef TARGET_ARM 10584 if (!((CPUARMState *)cpu_env)->eabi) { 10585 copyfrom = copy_from_user_oabi_flock64; 10586 copyto = copy_to_user_oabi_flock64; 10587 } 10588 #endif 10589 10590 cmd = target_to_host_fcntl_cmd(arg2); 10591 if (cmd == -TARGET_EINVAL) { 10592 return cmd; 10593 } 10594 10595 switch(arg2) { 10596 case TARGET_F_GETLK64: 10597 ret = copyfrom(&fl, arg3); 10598 if (ret) { 10599 break; 10600 } 10601 ret = get_errno(safe_fcntl(arg1, cmd, &fl)); 10602 if (ret == 0) { 10603 ret = copyto(arg3, &fl); 10604 } 10605 break; 10606 10607 case TARGET_F_SETLK64: 10608 case TARGET_F_SETLKW64: 10609 ret = copyfrom(&fl, arg3); 10610 if (ret) { 10611 break; 10612 } 10613 ret = get_errno(safe_fcntl(arg1, cmd, &fl)); 10614 break; 10615 default: 10616 ret = do_fcntl(arg1, arg2, arg3); 10617 break; 10618 } 10619 return ret; 10620 } 10621 #endif 10622 #ifdef TARGET_NR_cacheflush 10623 case TARGET_NR_cacheflush: 10624 /* self-modifying code is handled automatically, so nothing needed */ 10625 return 0; 10626 #endif 10627 #ifdef TARGET_NR_getpagesize 10628 case TARGET_NR_getpagesize: 10629 return TARGET_PAGE_SIZE; 10630 #endif 10631 case TARGET_NR_gettid: 10632 return get_errno(gettid()); 10633 #ifdef TARGET_NR_readahead 10634 case TARGET_NR_readahead: 10635 #if TARGET_ABI_BITS == 32 10636 if (regpairs_aligned(cpu_env, num)) { 10637 arg2 = arg3; 10638 arg3 = arg4; 10639 arg4 = arg5; 10640 } 10641 ret = get_errno(readahead(arg1, target_offset64(arg2, arg3) , arg4)); 10642 #else 10643 ret = get_errno(readahead(arg1, arg2, arg3)); 10644 #endif 10645 return ret; 10646 #endif 10647 #ifdef CONFIG_ATTR 10648 #ifdef TARGET_NR_setxattr 10649 case TARGET_NR_listxattr: 10650 case TARGET_NR_llistxattr: 10651 { 10652 void *p, *b = 0; 10653 if (arg2) { 10654 b = lock_user(VERIFY_WRITE, arg2, arg3, 0); 10655 if (!b) { 10656 return -TARGET_EFAULT; 10657 } 10658 } 10659 p = lock_user_string(arg1); 10660 if (p) { 10661 if (num == TARGET_NR_listxattr) { 10662 ret = get_errno(listxattr(p, b, arg3)); 10663 } else { 10664 ret = get_errno(llistxattr(p, b, arg3)); 10665 } 10666 } else { 10667 ret = -TARGET_EFAULT; 10668 } 10669 unlock_user(p, arg1, 0); 10670 unlock_user(b, arg2, arg3); 10671 return ret; 10672 } 10673 case TARGET_NR_flistxattr: 10674 { 10675 void *b = 0; 10676 if (arg2) { 10677 b = lock_user(VERIFY_WRITE, arg2, arg3, 0); 10678 if (!b) { 10679 return -TARGET_EFAULT; 10680 } 10681 } 10682 ret = get_errno(flistxattr(arg1, b, arg3)); 10683 unlock_user(b, arg2, arg3); 10684 return ret; 10685 } 10686 case TARGET_NR_setxattr: 10687 case TARGET_NR_lsetxattr: 10688 { 10689 void *p, *n, *v = 0; 10690 if (arg3) { 10691 v = lock_user(VERIFY_READ, arg3, arg4, 1); 10692 if (!v) { 10693 return -TARGET_EFAULT; 10694 } 10695 } 10696 p = lock_user_string(arg1); 10697 n = lock_user_string(arg2); 10698 if (p && n) { 10699 if (num == TARGET_NR_setxattr) { 10700 ret = get_errno(setxattr(p, n, v, arg4, arg5)); 10701 } else { 10702 ret = get_errno(lsetxattr(p, n, v, arg4, arg5)); 10703 } 10704 } else { 10705 ret = -TARGET_EFAULT; 10706 } 10707 unlock_user(p, arg1, 0); 10708 unlock_user(n, arg2, 0); 10709 unlock_user(v, arg3, 0); 10710 } 10711 return ret; 10712 case TARGET_NR_fsetxattr: 10713 { 10714 void *n, *v = 0; 10715 if (arg3) { 10716 v = lock_user(VERIFY_READ, arg3, arg4, 1); 10717 if (!v) { 10718 return -TARGET_EFAULT; 10719 } 10720 } 10721 n = lock_user_string(arg2); 10722 if (n) { 10723 ret = get_errno(fsetxattr(arg1, n, v, arg4, arg5)); 10724 } else { 10725 ret = -TARGET_EFAULT; 10726 } 10727 unlock_user(n, arg2, 0); 10728 unlock_user(v, arg3, 0); 10729 } 10730 return ret; 10731 case TARGET_NR_getxattr: 10732 case TARGET_NR_lgetxattr: 10733 { 10734 void *p, *n, *v = 0; 10735 if (arg3) { 10736 v = lock_user(VERIFY_WRITE, arg3, arg4, 0); 10737 if (!v) { 10738 return -TARGET_EFAULT; 10739 } 10740 } 10741 p = lock_user_string(arg1); 10742 n = lock_user_string(arg2); 10743 if (p && n) { 10744 if (num == TARGET_NR_getxattr) { 10745 ret = get_errno(getxattr(p, n, v, arg4)); 10746 } else { 10747 ret = get_errno(lgetxattr(p, n, v, arg4)); 10748 } 10749 } else { 10750 ret = -TARGET_EFAULT; 10751 } 10752 unlock_user(p, arg1, 0); 10753 unlock_user(n, arg2, 0); 10754 unlock_user(v, arg3, arg4); 10755 } 10756 return ret; 10757 case TARGET_NR_fgetxattr: 10758 { 10759 void *n, *v = 0; 10760 if (arg3) { 10761 v = lock_user(VERIFY_WRITE, arg3, arg4, 0); 10762 if (!v) { 10763 return -TARGET_EFAULT; 10764 } 10765 } 10766 n = lock_user_string(arg2); 10767 if (n) { 10768 ret = get_errno(fgetxattr(arg1, n, v, arg4)); 10769 } else { 10770 ret = -TARGET_EFAULT; 10771 } 10772 unlock_user(n, arg2, 0); 10773 unlock_user(v, arg3, arg4); 10774 } 10775 return ret; 10776 case TARGET_NR_removexattr: 10777 case TARGET_NR_lremovexattr: 10778 { 10779 void *p, *n; 10780 p = lock_user_string(arg1); 10781 n = lock_user_string(arg2); 10782 if (p && n) { 10783 if (num == TARGET_NR_removexattr) { 10784 ret = get_errno(removexattr(p, n)); 10785 } else { 10786 ret = get_errno(lremovexattr(p, n)); 10787 } 10788 } else { 10789 ret = -TARGET_EFAULT; 10790 } 10791 unlock_user(p, arg1, 0); 10792 unlock_user(n, arg2, 0); 10793 } 10794 return ret; 10795 case TARGET_NR_fremovexattr: 10796 { 10797 void *n; 10798 n = lock_user_string(arg2); 10799 if (n) { 10800 ret = get_errno(fremovexattr(arg1, n)); 10801 } else { 10802 ret = -TARGET_EFAULT; 10803 } 10804 unlock_user(n, arg2, 0); 10805 } 10806 return ret; 10807 #endif 10808 #endif /* CONFIG_ATTR */ 10809 #ifdef TARGET_NR_set_thread_area 10810 case TARGET_NR_set_thread_area: 10811 #if defined(TARGET_MIPS) 10812 ((CPUMIPSState *) cpu_env)->active_tc.CP0_UserLocal = arg1; 10813 return 0; 10814 #elif defined(TARGET_CRIS) 10815 if (arg1 & 0xff) 10816 ret = -TARGET_EINVAL; 10817 else { 10818 ((CPUCRISState *) cpu_env)->pregs[PR_PID] = arg1; 10819 ret = 0; 10820 } 10821 return ret; 10822 #elif defined(TARGET_I386) && defined(TARGET_ABI32) 10823 return do_set_thread_area(cpu_env, arg1); 10824 #elif defined(TARGET_M68K) 10825 { 10826 TaskState *ts = cpu->opaque; 10827 ts->tp_value = arg1; 10828 return 0; 10829 } 10830 #else 10831 return -TARGET_ENOSYS; 10832 #endif 10833 #endif 10834 #ifdef TARGET_NR_get_thread_area 10835 case TARGET_NR_get_thread_area: 10836 #if defined(TARGET_I386) && defined(TARGET_ABI32) 10837 return do_get_thread_area(cpu_env, arg1); 10838 #elif defined(TARGET_M68K) 10839 { 10840 TaskState *ts = cpu->opaque; 10841 return ts->tp_value; 10842 } 10843 #else 10844 return -TARGET_ENOSYS; 10845 #endif 10846 #endif 10847 #ifdef TARGET_NR_getdomainname 10848 case TARGET_NR_getdomainname: 10849 return -TARGET_ENOSYS; 10850 #endif 10851 10852 #ifdef TARGET_NR_clock_settime 10853 case TARGET_NR_clock_settime: 10854 { 10855 struct timespec ts; 10856 10857 ret = target_to_host_timespec(&ts, arg2); 10858 if (!is_error(ret)) { 10859 ret = get_errno(clock_settime(arg1, &ts)); 10860 } 10861 return ret; 10862 } 10863 #endif 10864 #ifdef TARGET_NR_clock_gettime 10865 case TARGET_NR_clock_gettime: 10866 { 10867 struct timespec ts; 10868 ret = get_errno(clock_gettime(arg1, &ts)); 10869 if (!is_error(ret)) { 10870 ret = host_to_target_timespec(arg2, &ts); 10871 } 10872 return ret; 10873 } 10874 #endif 10875 #ifdef TARGET_NR_clock_getres 10876 case TARGET_NR_clock_getres: 10877 { 10878 struct timespec ts; 10879 ret = get_errno(clock_getres(arg1, &ts)); 10880 if (!is_error(ret)) { 10881 host_to_target_timespec(arg2, &ts); 10882 } 10883 return ret; 10884 } 10885 #endif 10886 #ifdef TARGET_NR_clock_nanosleep 10887 case TARGET_NR_clock_nanosleep: 10888 { 10889 struct timespec ts; 10890 target_to_host_timespec(&ts, arg3); 10891 ret = get_errno(safe_clock_nanosleep(arg1, arg2, 10892 &ts, arg4 ? &ts : NULL)); 10893 if (arg4) 10894 host_to_target_timespec(arg4, &ts); 10895 10896 #if defined(TARGET_PPC) 10897 /* clock_nanosleep is odd in that it returns positive errno values. 10898 * On PPC, CR0 bit 3 should be set in such a situation. */ 10899 if (ret && ret != -TARGET_ERESTARTSYS) { 10900 ((CPUPPCState *)cpu_env)->crf[0] |= 1; 10901 } 10902 #endif 10903 return ret; 10904 } 10905 #endif 10906 10907 #if defined(TARGET_NR_set_tid_address) && defined(__NR_set_tid_address) 10908 case TARGET_NR_set_tid_address: 10909 return get_errno(set_tid_address((int *)g2h(arg1))); 10910 #endif 10911 10912 case TARGET_NR_tkill: 10913 return get_errno(safe_tkill((int)arg1, target_to_host_signal(arg2))); 10914 10915 case TARGET_NR_tgkill: 10916 return get_errno(safe_tgkill((int)arg1, (int)arg2, 10917 target_to_host_signal(arg3))); 10918 10919 #ifdef TARGET_NR_set_robust_list 10920 case TARGET_NR_set_robust_list: 10921 case TARGET_NR_get_robust_list: 10922 /* The ABI for supporting robust futexes has userspace pass 10923 * the kernel a pointer to a linked list which is updated by 10924 * userspace after the syscall; the list is walked by the kernel 10925 * when the thread exits. Since the linked list in QEMU guest 10926 * memory isn't a valid linked list for the host and we have 10927 * no way to reliably intercept the thread-death event, we can't 10928 * support these. Silently return ENOSYS so that guest userspace 10929 * falls back to a non-robust futex implementation (which should 10930 * be OK except in the corner case of the guest crashing while 10931 * holding a mutex that is shared with another process via 10932 * shared memory). 10933 */ 10934 return -TARGET_ENOSYS; 10935 #endif 10936 10937 #if defined(TARGET_NR_utimensat) 10938 case TARGET_NR_utimensat: 10939 { 10940 struct timespec *tsp, ts[2]; 10941 if (!arg3) { 10942 tsp = NULL; 10943 } else { 10944 target_to_host_timespec(ts, arg3); 10945 target_to_host_timespec(ts+1, arg3+sizeof(struct target_timespec)); 10946 tsp = ts; 10947 } 10948 if (!arg2) 10949 ret = get_errno(sys_utimensat(arg1, NULL, tsp, arg4)); 10950 else { 10951 if (!(p = lock_user_string(arg2))) { 10952 return -TARGET_EFAULT; 10953 } 10954 ret = get_errno(sys_utimensat(arg1, path(p), tsp, arg4)); 10955 unlock_user(p, arg2, 0); 10956 } 10957 } 10958 return ret; 10959 #endif 10960 case TARGET_NR_futex: 10961 return do_futex(arg1, arg2, arg3, arg4, arg5, arg6); 10962 #if defined(TARGET_NR_inotify_init) && defined(__NR_inotify_init) 10963 case TARGET_NR_inotify_init: 10964 ret = get_errno(sys_inotify_init()); 10965 if (ret >= 0) { 10966 fd_trans_register(ret, &target_inotify_trans); 10967 } 10968 return ret; 10969 #endif 10970 #ifdef CONFIG_INOTIFY1 10971 #if defined(TARGET_NR_inotify_init1) && defined(__NR_inotify_init1) 10972 case TARGET_NR_inotify_init1: 10973 ret = get_errno(sys_inotify_init1(target_to_host_bitmask(arg1, 10974 fcntl_flags_tbl))); 10975 if (ret >= 0) { 10976 fd_trans_register(ret, &target_inotify_trans); 10977 } 10978 return ret; 10979 #endif 10980 #endif 10981 #if defined(TARGET_NR_inotify_add_watch) && defined(__NR_inotify_add_watch) 10982 case TARGET_NR_inotify_add_watch: 10983 p = lock_user_string(arg2); 10984 ret = get_errno(sys_inotify_add_watch(arg1, path(p), arg3)); 10985 unlock_user(p, arg2, 0); 10986 return ret; 10987 #endif 10988 #if defined(TARGET_NR_inotify_rm_watch) && defined(__NR_inotify_rm_watch) 10989 case TARGET_NR_inotify_rm_watch: 10990 return get_errno(sys_inotify_rm_watch(arg1, arg2)); 10991 #endif 10992 10993 #if defined(TARGET_NR_mq_open) && defined(__NR_mq_open) 10994 case TARGET_NR_mq_open: 10995 { 10996 struct mq_attr posix_mq_attr; 10997 struct mq_attr *pposix_mq_attr; 10998 int host_flags; 10999 11000 host_flags = target_to_host_bitmask(arg2, fcntl_flags_tbl); 11001 pposix_mq_attr = NULL; 11002 if (arg4) { 11003 if (copy_from_user_mq_attr(&posix_mq_attr, arg4) != 0) { 11004 return -TARGET_EFAULT; 11005 } 11006 pposix_mq_attr = &posix_mq_attr; 11007 } 11008 p = lock_user_string(arg1 - 1); 11009 if (!p) { 11010 return -TARGET_EFAULT; 11011 } 11012 ret = get_errno(mq_open(p, host_flags, arg3, pposix_mq_attr)); 11013 unlock_user (p, arg1, 0); 11014 } 11015 return ret; 11016 11017 case TARGET_NR_mq_unlink: 11018 p = lock_user_string(arg1 - 1); 11019 if (!p) { 11020 return -TARGET_EFAULT; 11021 } 11022 ret = get_errno(mq_unlink(p)); 11023 unlock_user (p, arg1, 0); 11024 return ret; 11025 11026 case TARGET_NR_mq_timedsend: 11027 { 11028 struct timespec ts; 11029 11030 p = lock_user (VERIFY_READ, arg2, arg3, 1); 11031 if (arg5 != 0) { 11032 target_to_host_timespec(&ts, arg5); 11033 ret = get_errno(safe_mq_timedsend(arg1, p, arg3, arg4, &ts)); 11034 host_to_target_timespec(arg5, &ts); 11035 } else { 11036 ret = get_errno(safe_mq_timedsend(arg1, p, arg3, arg4, NULL)); 11037 } 11038 unlock_user (p, arg2, arg3); 11039 } 11040 return ret; 11041 11042 case TARGET_NR_mq_timedreceive: 11043 { 11044 struct timespec ts; 11045 unsigned int prio; 11046 11047 p = lock_user (VERIFY_READ, arg2, arg3, 1); 11048 if (arg5 != 0) { 11049 target_to_host_timespec(&ts, arg5); 11050 ret = get_errno(safe_mq_timedreceive(arg1, p, arg3, 11051 &prio, &ts)); 11052 host_to_target_timespec(arg5, &ts); 11053 } else { 11054 ret = get_errno(safe_mq_timedreceive(arg1, p, arg3, 11055 &prio, NULL)); 11056 } 11057 unlock_user (p, arg2, arg3); 11058 if (arg4 != 0) 11059 put_user_u32(prio, arg4); 11060 } 11061 return ret; 11062 11063 /* Not implemented for now... */ 11064 /* case TARGET_NR_mq_notify: */ 11065 /* break; */ 11066 11067 case TARGET_NR_mq_getsetattr: 11068 { 11069 struct mq_attr posix_mq_attr_in, posix_mq_attr_out; 11070 ret = 0; 11071 if (arg2 != 0) { 11072 copy_from_user_mq_attr(&posix_mq_attr_in, arg2); 11073 ret = get_errno(mq_setattr(arg1, &posix_mq_attr_in, 11074 &posix_mq_attr_out)); 11075 } else if (arg3 != 0) { 11076 ret = get_errno(mq_getattr(arg1, &posix_mq_attr_out)); 11077 } 11078 if (ret == 0 && arg3 != 0) { 11079 copy_to_user_mq_attr(arg3, &posix_mq_attr_out); 11080 } 11081 } 11082 return ret; 11083 #endif 11084 11085 #ifdef CONFIG_SPLICE 11086 #ifdef TARGET_NR_tee 11087 case TARGET_NR_tee: 11088 { 11089 ret = get_errno(tee(arg1,arg2,arg3,arg4)); 11090 } 11091 return ret; 11092 #endif 11093 #ifdef TARGET_NR_splice 11094 case TARGET_NR_splice: 11095 { 11096 loff_t loff_in, loff_out; 11097 loff_t *ploff_in = NULL, *ploff_out = NULL; 11098 if (arg2) { 11099 if (get_user_u64(loff_in, arg2)) { 11100 return -TARGET_EFAULT; 11101 } 11102 ploff_in = &loff_in; 11103 } 11104 if (arg4) { 11105 if (get_user_u64(loff_out, arg4)) { 11106 return -TARGET_EFAULT; 11107 } 11108 ploff_out = &loff_out; 11109 } 11110 ret = get_errno(splice(arg1, ploff_in, arg3, ploff_out, arg5, arg6)); 11111 if (arg2) { 11112 if (put_user_u64(loff_in, arg2)) { 11113 return -TARGET_EFAULT; 11114 } 11115 } 11116 if (arg4) { 11117 if (put_user_u64(loff_out, arg4)) { 11118 return -TARGET_EFAULT; 11119 } 11120 } 11121 } 11122 return ret; 11123 #endif 11124 #ifdef TARGET_NR_vmsplice 11125 case TARGET_NR_vmsplice: 11126 { 11127 struct iovec *vec = lock_iovec(VERIFY_READ, arg2, arg3, 1); 11128 if (vec != NULL) { 11129 ret = get_errno(vmsplice(arg1, vec, arg3, arg4)); 11130 unlock_iovec(vec, arg2, arg3, 0); 11131 } else { 11132 ret = -host_to_target_errno(errno); 11133 } 11134 } 11135 return ret; 11136 #endif 11137 #endif /* CONFIG_SPLICE */ 11138 #ifdef CONFIG_EVENTFD 11139 #if defined(TARGET_NR_eventfd) 11140 case TARGET_NR_eventfd: 11141 ret = get_errno(eventfd(arg1, 0)); 11142 if (ret >= 0) { 11143 fd_trans_register(ret, &target_eventfd_trans); 11144 } 11145 return ret; 11146 #endif 11147 #if defined(TARGET_NR_eventfd2) 11148 case TARGET_NR_eventfd2: 11149 { 11150 int host_flags = arg2 & (~(TARGET_O_NONBLOCK | TARGET_O_CLOEXEC)); 11151 if (arg2 & TARGET_O_NONBLOCK) { 11152 host_flags |= O_NONBLOCK; 11153 } 11154 if (arg2 & TARGET_O_CLOEXEC) { 11155 host_flags |= O_CLOEXEC; 11156 } 11157 ret = get_errno(eventfd(arg1, host_flags)); 11158 if (ret >= 0) { 11159 fd_trans_register(ret, &target_eventfd_trans); 11160 } 11161 return ret; 11162 } 11163 #endif 11164 #endif /* CONFIG_EVENTFD */ 11165 #if defined(CONFIG_FALLOCATE) && defined(TARGET_NR_fallocate) 11166 case TARGET_NR_fallocate: 11167 #if TARGET_ABI_BITS == 32 11168 ret = get_errno(fallocate(arg1, arg2, target_offset64(arg3, arg4), 11169 target_offset64(arg5, arg6))); 11170 #else 11171 ret = get_errno(fallocate(arg1, arg2, arg3, arg4)); 11172 #endif 11173 return ret; 11174 #endif 11175 #if defined(CONFIG_SYNC_FILE_RANGE) 11176 #if defined(TARGET_NR_sync_file_range) 11177 case TARGET_NR_sync_file_range: 11178 #if TARGET_ABI_BITS == 32 11179 #if defined(TARGET_MIPS) 11180 ret = get_errno(sync_file_range(arg1, target_offset64(arg3, arg4), 11181 target_offset64(arg5, arg6), arg7)); 11182 #else 11183 ret = get_errno(sync_file_range(arg1, target_offset64(arg2, arg3), 11184 target_offset64(arg4, arg5), arg6)); 11185 #endif /* !TARGET_MIPS */ 11186 #else 11187 ret = get_errno(sync_file_range(arg1, arg2, arg3, arg4)); 11188 #endif 11189 return ret; 11190 #endif 11191 #if defined(TARGET_NR_sync_file_range2) 11192 case TARGET_NR_sync_file_range2: 11193 /* This is like sync_file_range but the arguments are reordered */ 11194 #if TARGET_ABI_BITS == 32 11195 ret = get_errno(sync_file_range(arg1, target_offset64(arg3, arg4), 11196 target_offset64(arg5, arg6), arg2)); 11197 #else 11198 ret = get_errno(sync_file_range(arg1, arg3, arg4, arg2)); 11199 #endif 11200 return ret; 11201 #endif 11202 #endif 11203 #if defined(TARGET_NR_signalfd4) 11204 case TARGET_NR_signalfd4: 11205 return do_signalfd4(arg1, arg2, arg4); 11206 #endif 11207 #if defined(TARGET_NR_signalfd) 11208 case TARGET_NR_signalfd: 11209 return do_signalfd4(arg1, arg2, 0); 11210 #endif 11211 #if defined(CONFIG_EPOLL) 11212 #if defined(TARGET_NR_epoll_create) 11213 case TARGET_NR_epoll_create: 11214 return get_errno(epoll_create(arg1)); 11215 #endif 11216 #if defined(TARGET_NR_epoll_create1) && defined(CONFIG_EPOLL_CREATE1) 11217 case TARGET_NR_epoll_create1: 11218 return get_errno(epoll_create1(arg1)); 11219 #endif 11220 #if defined(TARGET_NR_epoll_ctl) 11221 case TARGET_NR_epoll_ctl: 11222 { 11223 struct epoll_event ep; 11224 struct epoll_event *epp = 0; 11225 if (arg4) { 11226 struct target_epoll_event *target_ep; 11227 if (!lock_user_struct(VERIFY_READ, target_ep, arg4, 1)) { 11228 return -TARGET_EFAULT; 11229 } 11230 ep.events = tswap32(target_ep->events); 11231 /* The epoll_data_t union is just opaque data to the kernel, 11232 * so we transfer all 64 bits across and need not worry what 11233 * actual data type it is. 11234 */ 11235 ep.data.u64 = tswap64(target_ep->data.u64); 11236 unlock_user_struct(target_ep, arg4, 0); 11237 epp = &ep; 11238 } 11239 return get_errno(epoll_ctl(arg1, arg2, arg3, epp)); 11240 } 11241 #endif 11242 11243 #if defined(TARGET_NR_epoll_wait) || defined(TARGET_NR_epoll_pwait) 11244 #if defined(TARGET_NR_epoll_wait) 11245 case TARGET_NR_epoll_wait: 11246 #endif 11247 #if defined(TARGET_NR_epoll_pwait) 11248 case TARGET_NR_epoll_pwait: 11249 #endif 11250 { 11251 struct target_epoll_event *target_ep; 11252 struct epoll_event *ep; 11253 int epfd = arg1; 11254 int maxevents = arg3; 11255 int timeout = arg4; 11256 11257 if (maxevents <= 0 || maxevents > TARGET_EP_MAX_EVENTS) { 11258 return -TARGET_EINVAL; 11259 } 11260 11261 target_ep = lock_user(VERIFY_WRITE, arg2, 11262 maxevents * sizeof(struct target_epoll_event), 1); 11263 if (!target_ep) { 11264 return -TARGET_EFAULT; 11265 } 11266 11267 ep = g_try_new(struct epoll_event, maxevents); 11268 if (!ep) { 11269 unlock_user(target_ep, arg2, 0); 11270 return -TARGET_ENOMEM; 11271 } 11272 11273 switch (num) { 11274 #if defined(TARGET_NR_epoll_pwait) 11275 case TARGET_NR_epoll_pwait: 11276 { 11277 target_sigset_t *target_set; 11278 sigset_t _set, *set = &_set; 11279 11280 if (arg5) { 11281 if (arg6 != sizeof(target_sigset_t)) { 11282 ret = -TARGET_EINVAL; 11283 break; 11284 } 11285 11286 target_set = lock_user(VERIFY_READ, arg5, 11287 sizeof(target_sigset_t), 1); 11288 if (!target_set) { 11289 ret = -TARGET_EFAULT; 11290 break; 11291 } 11292 target_to_host_sigset(set, target_set); 11293 unlock_user(target_set, arg5, 0); 11294 } else { 11295 set = NULL; 11296 } 11297 11298 ret = get_errno(safe_epoll_pwait(epfd, ep, maxevents, timeout, 11299 set, SIGSET_T_SIZE)); 11300 break; 11301 } 11302 #endif 11303 #if defined(TARGET_NR_epoll_wait) 11304 case TARGET_NR_epoll_wait: 11305 ret = get_errno(safe_epoll_pwait(epfd, ep, maxevents, timeout, 11306 NULL, 0)); 11307 break; 11308 #endif 11309 default: 11310 ret = -TARGET_ENOSYS; 11311 } 11312 if (!is_error(ret)) { 11313 int i; 11314 for (i = 0; i < ret; i++) { 11315 target_ep[i].events = tswap32(ep[i].events); 11316 target_ep[i].data.u64 = tswap64(ep[i].data.u64); 11317 } 11318 unlock_user(target_ep, arg2, 11319 ret * sizeof(struct target_epoll_event)); 11320 } else { 11321 unlock_user(target_ep, arg2, 0); 11322 } 11323 g_free(ep); 11324 return ret; 11325 } 11326 #endif 11327 #endif 11328 #ifdef TARGET_NR_prlimit64 11329 case TARGET_NR_prlimit64: 11330 { 11331 /* args: pid, resource number, ptr to new rlimit, ptr to old rlimit */ 11332 struct target_rlimit64 *target_rnew, *target_rold; 11333 struct host_rlimit64 rnew, rold, *rnewp = 0; 11334 int resource = target_to_host_resource(arg2); 11335 if (arg3) { 11336 if (!lock_user_struct(VERIFY_READ, target_rnew, arg3, 1)) { 11337 return -TARGET_EFAULT; 11338 } 11339 rnew.rlim_cur = tswap64(target_rnew->rlim_cur); 11340 rnew.rlim_max = tswap64(target_rnew->rlim_max); 11341 unlock_user_struct(target_rnew, arg3, 0); 11342 rnewp = &rnew; 11343 } 11344 11345 ret = get_errno(sys_prlimit64(arg1, resource, rnewp, arg4 ? &rold : 0)); 11346 if (!is_error(ret) && arg4) { 11347 if (!lock_user_struct(VERIFY_WRITE, target_rold, arg4, 1)) { 11348 return -TARGET_EFAULT; 11349 } 11350 target_rold->rlim_cur = tswap64(rold.rlim_cur); 11351 target_rold->rlim_max = tswap64(rold.rlim_max); 11352 unlock_user_struct(target_rold, arg4, 1); 11353 } 11354 return ret; 11355 } 11356 #endif 11357 #ifdef TARGET_NR_gethostname 11358 case TARGET_NR_gethostname: 11359 { 11360 char *name = lock_user(VERIFY_WRITE, arg1, arg2, 0); 11361 if (name) { 11362 ret = get_errno(gethostname(name, arg2)); 11363 unlock_user(name, arg1, arg2); 11364 } else { 11365 ret = -TARGET_EFAULT; 11366 } 11367 return ret; 11368 } 11369 #endif 11370 #ifdef TARGET_NR_atomic_cmpxchg_32 11371 case TARGET_NR_atomic_cmpxchg_32: 11372 { 11373 /* should use start_exclusive from main.c */ 11374 abi_ulong mem_value; 11375 if (get_user_u32(mem_value, arg6)) { 11376 target_siginfo_t info; 11377 info.si_signo = SIGSEGV; 11378 info.si_errno = 0; 11379 info.si_code = TARGET_SEGV_MAPERR; 11380 info._sifields._sigfault._addr = arg6; 11381 queue_signal((CPUArchState *)cpu_env, info.si_signo, 11382 QEMU_SI_FAULT, &info); 11383 ret = 0xdeadbeef; 11384 11385 } 11386 if (mem_value == arg2) 11387 put_user_u32(arg1, arg6); 11388 return mem_value; 11389 } 11390 #endif 11391 #ifdef TARGET_NR_atomic_barrier 11392 case TARGET_NR_atomic_barrier: 11393 /* Like the kernel implementation and the 11394 qemu arm barrier, no-op this? */ 11395 return 0; 11396 #endif 11397 11398 #ifdef TARGET_NR_timer_create 11399 case TARGET_NR_timer_create: 11400 { 11401 /* args: clockid_t clockid, struct sigevent *sevp, timer_t *timerid */ 11402 11403 struct sigevent host_sevp = { {0}, }, *phost_sevp = NULL; 11404 11405 int clkid = arg1; 11406 int timer_index = next_free_host_timer(); 11407 11408 if (timer_index < 0) { 11409 ret = -TARGET_EAGAIN; 11410 } else { 11411 timer_t *phtimer = g_posix_timers + timer_index; 11412 11413 if (arg2) { 11414 phost_sevp = &host_sevp; 11415 ret = target_to_host_sigevent(phost_sevp, arg2); 11416 if (ret != 0) { 11417 return ret; 11418 } 11419 } 11420 11421 ret = get_errno(timer_create(clkid, phost_sevp, phtimer)); 11422 if (ret) { 11423 phtimer = NULL; 11424 } else { 11425 if (put_user(TIMER_MAGIC | timer_index, arg3, target_timer_t)) { 11426 return -TARGET_EFAULT; 11427 } 11428 } 11429 } 11430 return ret; 11431 } 11432 #endif 11433 11434 #ifdef TARGET_NR_timer_settime 11435 case TARGET_NR_timer_settime: 11436 { 11437 /* args: timer_t timerid, int flags, const struct itimerspec *new_value, 11438 * struct itimerspec * old_value */ 11439 target_timer_t timerid = get_timer_id(arg1); 11440 11441 if (timerid < 0) { 11442 ret = timerid; 11443 } else if (arg3 == 0) { 11444 ret = -TARGET_EINVAL; 11445 } else { 11446 timer_t htimer = g_posix_timers[timerid]; 11447 struct itimerspec hspec_new = {{0},}, hspec_old = {{0},}; 11448 11449 if (target_to_host_itimerspec(&hspec_new, arg3)) { 11450 return -TARGET_EFAULT; 11451 } 11452 ret = get_errno( 11453 timer_settime(htimer, arg2, &hspec_new, &hspec_old)); 11454 if (arg4 && host_to_target_itimerspec(arg4, &hspec_old)) { 11455 return -TARGET_EFAULT; 11456 } 11457 } 11458 return ret; 11459 } 11460 #endif 11461 11462 #ifdef TARGET_NR_timer_gettime 11463 case TARGET_NR_timer_gettime: 11464 { 11465 /* args: timer_t timerid, struct itimerspec *curr_value */ 11466 target_timer_t timerid = get_timer_id(arg1); 11467 11468 if (timerid < 0) { 11469 ret = timerid; 11470 } else if (!arg2) { 11471 ret = -TARGET_EFAULT; 11472 } else { 11473 timer_t htimer = g_posix_timers[timerid]; 11474 struct itimerspec hspec; 11475 ret = get_errno(timer_gettime(htimer, &hspec)); 11476 11477 if (host_to_target_itimerspec(arg2, &hspec)) { 11478 ret = -TARGET_EFAULT; 11479 } 11480 } 11481 return ret; 11482 } 11483 #endif 11484 11485 #ifdef TARGET_NR_timer_getoverrun 11486 case TARGET_NR_timer_getoverrun: 11487 { 11488 /* args: timer_t timerid */ 11489 target_timer_t timerid = get_timer_id(arg1); 11490 11491 if (timerid < 0) { 11492 ret = timerid; 11493 } else { 11494 timer_t htimer = g_posix_timers[timerid]; 11495 ret = get_errno(timer_getoverrun(htimer)); 11496 } 11497 fd_trans_unregister(ret); 11498 return ret; 11499 } 11500 #endif 11501 11502 #ifdef TARGET_NR_timer_delete 11503 case TARGET_NR_timer_delete: 11504 { 11505 /* args: timer_t timerid */ 11506 target_timer_t timerid = get_timer_id(arg1); 11507 11508 if (timerid < 0) { 11509 ret = timerid; 11510 } else { 11511 timer_t htimer = g_posix_timers[timerid]; 11512 ret = get_errno(timer_delete(htimer)); 11513 g_posix_timers[timerid] = 0; 11514 } 11515 return ret; 11516 } 11517 #endif 11518 11519 #if defined(TARGET_NR_timerfd_create) && defined(CONFIG_TIMERFD) 11520 case TARGET_NR_timerfd_create: 11521 return get_errno(timerfd_create(arg1, 11522 target_to_host_bitmask(arg2, fcntl_flags_tbl))); 11523 #endif 11524 11525 #if defined(TARGET_NR_timerfd_gettime) && defined(CONFIG_TIMERFD) 11526 case TARGET_NR_timerfd_gettime: 11527 { 11528 struct itimerspec its_curr; 11529 11530 ret = get_errno(timerfd_gettime(arg1, &its_curr)); 11531 11532 if (arg2 && host_to_target_itimerspec(arg2, &its_curr)) { 11533 return -TARGET_EFAULT; 11534 } 11535 } 11536 return ret; 11537 #endif 11538 11539 #if defined(TARGET_NR_timerfd_settime) && defined(CONFIG_TIMERFD) 11540 case TARGET_NR_timerfd_settime: 11541 { 11542 struct itimerspec its_new, its_old, *p_new; 11543 11544 if (arg3) { 11545 if (target_to_host_itimerspec(&its_new, arg3)) { 11546 return -TARGET_EFAULT; 11547 } 11548 p_new = &its_new; 11549 } else { 11550 p_new = NULL; 11551 } 11552 11553 ret = get_errno(timerfd_settime(arg1, arg2, p_new, &its_old)); 11554 11555 if (arg4 && host_to_target_itimerspec(arg4, &its_old)) { 11556 return -TARGET_EFAULT; 11557 } 11558 } 11559 return ret; 11560 #endif 11561 11562 #if defined(TARGET_NR_ioprio_get) && defined(__NR_ioprio_get) 11563 case TARGET_NR_ioprio_get: 11564 return get_errno(ioprio_get(arg1, arg2)); 11565 #endif 11566 11567 #if defined(TARGET_NR_ioprio_set) && defined(__NR_ioprio_set) 11568 case TARGET_NR_ioprio_set: 11569 return get_errno(ioprio_set(arg1, arg2, arg3)); 11570 #endif 11571 11572 #if defined(TARGET_NR_setns) && defined(CONFIG_SETNS) 11573 case TARGET_NR_setns: 11574 return get_errno(setns(arg1, arg2)); 11575 #endif 11576 #if defined(TARGET_NR_unshare) && defined(CONFIG_SETNS) 11577 case TARGET_NR_unshare: 11578 return get_errno(unshare(arg1)); 11579 #endif 11580 #if defined(TARGET_NR_kcmp) && defined(__NR_kcmp) 11581 case TARGET_NR_kcmp: 11582 return get_errno(kcmp(arg1, arg2, arg3, arg4, arg5)); 11583 #endif 11584 #ifdef TARGET_NR_swapcontext 11585 case TARGET_NR_swapcontext: 11586 /* PowerPC specific. */ 11587 return do_swapcontext(cpu_env, arg1, arg2, arg3); 11588 #endif 11589 11590 default: 11591 qemu_log_mask(LOG_UNIMP, "Unsupported syscall: %d\n", num); 11592 return -TARGET_ENOSYS; 11593 } 11594 return ret; 11595 } 11596 11597 abi_long do_syscall(void *cpu_env, int num, abi_long arg1, 11598 abi_long arg2, abi_long arg3, abi_long arg4, 11599 abi_long arg5, abi_long arg6, abi_long arg7, 11600 abi_long arg8) 11601 { 11602 CPUState *cpu = ENV_GET_CPU(cpu_env); 11603 abi_long ret; 11604 11605 #ifdef DEBUG_ERESTARTSYS 11606 /* Debug-only code for exercising the syscall-restart code paths 11607 * in the per-architecture cpu main loops: restart every syscall 11608 * the guest makes once before letting it through. 11609 */ 11610 { 11611 static bool flag; 11612 flag = !flag; 11613 if (flag) { 11614 return -TARGET_ERESTARTSYS; 11615 } 11616 } 11617 #endif 11618 11619 trace_guest_user_syscall(cpu, num, arg1, arg2, arg3, arg4, 11620 arg5, arg6, arg7, arg8); 11621 11622 if (unlikely(do_strace)) { 11623 print_syscall(num, arg1, arg2, arg3, arg4, arg5, arg6); 11624 ret = do_syscall1(cpu_env, num, arg1, arg2, arg3, arg4, 11625 arg5, arg6, arg7, arg8); 11626 print_syscall_ret(num, ret); 11627 } else { 11628 ret = do_syscall1(cpu_env, num, arg1, arg2, arg3, arg4, 11629 arg5, arg6, arg7, arg8); 11630 } 11631 11632 trace_guest_user_syscall_ret(cpu, num, ret); 11633 return ret; 11634 } 11635