1 /* 2 * Linux syscalls 3 * 4 * Copyright (c) 2003 Fabrice Bellard 5 * 6 * This program is free software; you can redistribute it and/or modify 7 * it under the terms of the GNU General Public License as published by 8 * the Free Software Foundation; either version 2 of the License, or 9 * (at your option) any later version. 10 * 11 * This program is distributed in the hope that it will be useful, 12 * but WITHOUT ANY WARRANTY; without even the implied warranty of 13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 14 * GNU General Public License for more details. 15 * 16 * You should have received a copy of the GNU General Public License 17 * along with this program; if not, see <http://www.gnu.org/licenses/>. 18 */ 19 #define _ATFILE_SOURCE 20 #include <stdlib.h> 21 #include <stdio.h> 22 #include <stdarg.h> 23 #include <string.h> 24 #include <elf.h> 25 #include <endian.h> 26 #include <errno.h> 27 #include <unistd.h> 28 #include <fcntl.h> 29 #include <time.h> 30 #include <limits.h> 31 #include <sys/types.h> 32 #include <sys/ipc.h> 33 #include <sys/msg.h> 34 #include <sys/wait.h> 35 #include <sys/time.h> 36 #include <sys/stat.h> 37 #include <sys/mount.h> 38 #include <sys/prctl.h> 39 #include <sys/resource.h> 40 #include <sys/mman.h> 41 #include <sys/swap.h> 42 #include <signal.h> 43 #include <sched.h> 44 #ifdef __ia64__ 45 int __clone2(int (*fn)(void *), void *child_stack_base, 46 size_t stack_size, int flags, void *arg, ...); 47 #endif 48 #include <sys/socket.h> 49 #include <sys/un.h> 50 #include <sys/uio.h> 51 #include <sys/poll.h> 52 #include <sys/times.h> 53 #include <sys/shm.h> 54 #include <sys/sem.h> 55 #include <sys/statfs.h> 56 #include <utime.h> 57 #include <sys/sysinfo.h> 58 #include <sys/utsname.h> 59 //#include <sys/user.h> 60 #include <netinet/ip.h> 61 #include <netinet/tcp.h> 62 #include <linux/wireless.h> 63 #include <qemu-common.h> 64 #ifdef TARGET_GPROF 65 #include <sys/gmon.h> 66 #endif 67 #ifdef CONFIG_EVENTFD 68 #include <sys/eventfd.h> 69 #endif 70 #ifdef CONFIG_EPOLL 71 #include <sys/epoll.h> 72 #endif 73 74 #define termios host_termios 75 #define winsize host_winsize 76 #define termio host_termio 77 #define sgttyb host_sgttyb /* same as target */ 78 #define tchars host_tchars /* same as target */ 79 #define ltchars host_ltchars /* same as target */ 80 81 #include <linux/termios.h> 82 #include <linux/unistd.h> 83 #include <linux/utsname.h> 84 #include <linux/cdrom.h> 85 #include <linux/hdreg.h> 86 #include <linux/soundcard.h> 87 #include <linux/kd.h> 88 #include <linux/mtio.h> 89 #include <linux/fs.h> 90 #if defined(CONFIG_FIEMAP) 91 #include <linux/fiemap.h> 92 #endif 93 #include <linux/fb.h> 94 #include <linux/vt.h> 95 #include "linux_loop.h" 96 #include "cpu-uname.h" 97 98 #include "qemu.h" 99 #include "qemu-common.h" 100 101 #if defined(CONFIG_USE_NPTL) 102 #define CLONE_NPTL_FLAGS2 (CLONE_SETTLS | \ 103 CLONE_PARENT_SETTID | CLONE_CHILD_SETTID | CLONE_CHILD_CLEARTID) 104 #else 105 /* XXX: Hardcode the above values. */ 106 #define CLONE_NPTL_FLAGS2 0 107 #endif 108 109 //#define DEBUG 110 111 //#include <linux/msdos_fs.h> 112 #define VFAT_IOCTL_READDIR_BOTH _IOR('r', 1, struct linux_dirent [2]) 113 #define VFAT_IOCTL_READDIR_SHORT _IOR('r', 2, struct linux_dirent [2]) 114 115 116 #undef _syscall0 117 #undef _syscall1 118 #undef _syscall2 119 #undef _syscall3 120 #undef _syscall4 121 #undef _syscall5 122 #undef _syscall6 123 124 #define _syscall0(type,name) \ 125 static type name (void) \ 126 { \ 127 return syscall(__NR_##name); \ 128 } 129 130 #define _syscall1(type,name,type1,arg1) \ 131 static type name (type1 arg1) \ 132 { \ 133 return syscall(__NR_##name, arg1); \ 134 } 135 136 #define _syscall2(type,name,type1,arg1,type2,arg2) \ 137 static type name (type1 arg1,type2 arg2) \ 138 { \ 139 return syscall(__NR_##name, arg1, arg2); \ 140 } 141 142 #define _syscall3(type,name,type1,arg1,type2,arg2,type3,arg3) \ 143 static type name (type1 arg1,type2 arg2,type3 arg3) \ 144 { \ 145 return syscall(__NR_##name, arg1, arg2, arg3); \ 146 } 147 148 #define _syscall4(type,name,type1,arg1,type2,arg2,type3,arg3,type4,arg4) \ 149 static type name (type1 arg1,type2 arg2,type3 arg3,type4 arg4) \ 150 { \ 151 return syscall(__NR_##name, arg1, arg2, arg3, arg4); \ 152 } 153 154 #define _syscall5(type,name,type1,arg1,type2,arg2,type3,arg3,type4,arg4, \ 155 type5,arg5) \ 156 static type name (type1 arg1,type2 arg2,type3 arg3,type4 arg4,type5 arg5) \ 157 { \ 158 return syscall(__NR_##name, arg1, arg2, arg3, arg4, arg5); \ 159 } 160 161 162 #define _syscall6(type,name,type1,arg1,type2,arg2,type3,arg3,type4,arg4, \ 163 type5,arg5,type6,arg6) \ 164 static type name (type1 arg1,type2 arg2,type3 arg3,type4 arg4,type5 arg5, \ 165 type6 arg6) \ 166 { \ 167 return syscall(__NR_##name, arg1, arg2, arg3, arg4, arg5, arg6); \ 168 } 169 170 171 #define __NR_sys_uname __NR_uname 172 #define __NR_sys_faccessat __NR_faccessat 173 #define __NR_sys_fchmodat __NR_fchmodat 174 #define __NR_sys_fchownat __NR_fchownat 175 #define __NR_sys_fstatat64 __NR_fstatat64 176 #define __NR_sys_futimesat __NR_futimesat 177 #define __NR_sys_getcwd1 __NR_getcwd 178 #define __NR_sys_getdents __NR_getdents 179 #define __NR_sys_getdents64 __NR_getdents64 180 #define __NR_sys_getpriority __NR_getpriority 181 #define __NR_sys_linkat __NR_linkat 182 #define __NR_sys_mkdirat __NR_mkdirat 183 #define __NR_sys_mknodat __NR_mknodat 184 #define __NR_sys_newfstatat __NR_newfstatat 185 #define __NR_sys_openat __NR_openat 186 #define __NR_sys_readlinkat __NR_readlinkat 187 #define __NR_sys_renameat __NR_renameat 188 #define __NR_sys_rt_sigqueueinfo __NR_rt_sigqueueinfo 189 #define __NR_sys_symlinkat __NR_symlinkat 190 #define __NR_sys_syslog __NR_syslog 191 #define __NR_sys_tgkill __NR_tgkill 192 #define __NR_sys_tkill __NR_tkill 193 #define __NR_sys_unlinkat __NR_unlinkat 194 #define __NR_sys_utimensat __NR_utimensat 195 #define __NR_sys_futex __NR_futex 196 #define __NR_sys_inotify_init __NR_inotify_init 197 #define __NR_sys_inotify_add_watch __NR_inotify_add_watch 198 #define __NR_sys_inotify_rm_watch __NR_inotify_rm_watch 199 200 #if defined(__alpha__) || defined (__ia64__) || defined(__x86_64__) || \ 201 defined(__s390x__) 202 #define __NR__llseek __NR_lseek 203 #endif 204 205 #ifdef __NR_gettid 206 _syscall0(int, gettid) 207 #else 208 /* This is a replacement for the host gettid() and must return a host 209 errno. */ 210 static int gettid(void) { 211 return -ENOSYS; 212 } 213 #endif 214 _syscall3(int, sys_getdents, uint, fd, struct linux_dirent *, dirp, uint, count); 215 #if defined(TARGET_NR_getdents64) && defined(__NR_getdents64) 216 _syscall3(int, sys_getdents64, uint, fd, struct linux_dirent64 *, dirp, uint, count); 217 #endif 218 _syscall2(int, sys_getpriority, int, which, int, who); 219 #if defined(TARGET_NR__llseek) && defined(__NR_llseek) 220 _syscall5(int, _llseek, uint, fd, ulong, hi, ulong, lo, 221 loff_t *, res, uint, wh); 222 #endif 223 _syscall3(int,sys_rt_sigqueueinfo,int,pid,int,sig,siginfo_t *,uinfo) 224 _syscall3(int,sys_syslog,int,type,char*,bufp,int,len) 225 #if defined(TARGET_NR_tgkill) && defined(__NR_tgkill) 226 _syscall3(int,sys_tgkill,int,tgid,int,pid,int,sig) 227 #endif 228 #if defined(TARGET_NR_tkill) && defined(__NR_tkill) 229 _syscall2(int,sys_tkill,int,tid,int,sig) 230 #endif 231 #ifdef __NR_exit_group 232 _syscall1(int,exit_group,int,error_code) 233 #endif 234 #if defined(TARGET_NR_set_tid_address) && defined(__NR_set_tid_address) 235 _syscall1(int,set_tid_address,int *,tidptr) 236 #endif 237 #if defined(CONFIG_USE_NPTL) 238 #if defined(TARGET_NR_futex) && defined(__NR_futex) 239 _syscall6(int,sys_futex,int *,uaddr,int,op,int,val, 240 const struct timespec *,timeout,int *,uaddr2,int,val3) 241 #endif 242 #endif 243 #define __NR_sys_sched_getaffinity __NR_sched_getaffinity 244 _syscall3(int, sys_sched_getaffinity, pid_t, pid, unsigned int, len, 245 unsigned long *, user_mask_ptr); 246 #define __NR_sys_sched_setaffinity __NR_sched_setaffinity 247 _syscall3(int, sys_sched_setaffinity, pid_t, pid, unsigned int, len, 248 unsigned long *, user_mask_ptr); 249 250 static bitmask_transtbl fcntl_flags_tbl[] = { 251 { TARGET_O_ACCMODE, TARGET_O_WRONLY, O_ACCMODE, O_WRONLY, }, 252 { TARGET_O_ACCMODE, TARGET_O_RDWR, O_ACCMODE, O_RDWR, }, 253 { TARGET_O_CREAT, TARGET_O_CREAT, O_CREAT, O_CREAT, }, 254 { TARGET_O_EXCL, TARGET_O_EXCL, O_EXCL, O_EXCL, }, 255 { TARGET_O_NOCTTY, TARGET_O_NOCTTY, O_NOCTTY, O_NOCTTY, }, 256 { TARGET_O_TRUNC, TARGET_O_TRUNC, O_TRUNC, O_TRUNC, }, 257 { TARGET_O_APPEND, TARGET_O_APPEND, O_APPEND, O_APPEND, }, 258 { TARGET_O_NONBLOCK, TARGET_O_NONBLOCK, O_NONBLOCK, O_NONBLOCK, }, 259 { TARGET_O_SYNC, TARGET_O_SYNC, O_SYNC, O_SYNC, }, 260 { TARGET_FASYNC, TARGET_FASYNC, FASYNC, FASYNC, }, 261 { TARGET_O_DIRECTORY, TARGET_O_DIRECTORY, O_DIRECTORY, O_DIRECTORY, }, 262 { TARGET_O_NOFOLLOW, TARGET_O_NOFOLLOW, O_NOFOLLOW, O_NOFOLLOW, }, 263 { TARGET_O_LARGEFILE, TARGET_O_LARGEFILE, O_LARGEFILE, O_LARGEFILE, }, 264 #if defined(O_DIRECT) 265 { TARGET_O_DIRECT, TARGET_O_DIRECT, O_DIRECT, O_DIRECT, }, 266 #endif 267 { 0, 0, 0, 0 } 268 }; 269 270 #define COPY_UTSNAME_FIELD(dest, src) \ 271 do { \ 272 /* __NEW_UTS_LEN doesn't include terminating null */ \ 273 (void) strncpy((dest), (src), __NEW_UTS_LEN); \ 274 (dest)[__NEW_UTS_LEN] = '\0'; \ 275 } while (0) 276 277 static int sys_uname(struct new_utsname *buf) 278 { 279 struct utsname uts_buf; 280 281 if (uname(&uts_buf) < 0) 282 return (-1); 283 284 /* 285 * Just in case these have some differences, we 286 * translate utsname to new_utsname (which is the 287 * struct linux kernel uses). 288 */ 289 290 bzero(buf, sizeof (*buf)); 291 COPY_UTSNAME_FIELD(buf->sysname, uts_buf.sysname); 292 COPY_UTSNAME_FIELD(buf->nodename, uts_buf.nodename); 293 COPY_UTSNAME_FIELD(buf->release, uts_buf.release); 294 COPY_UTSNAME_FIELD(buf->version, uts_buf.version); 295 COPY_UTSNAME_FIELD(buf->machine, uts_buf.machine); 296 #ifdef _GNU_SOURCE 297 COPY_UTSNAME_FIELD(buf->domainname, uts_buf.domainname); 298 #endif 299 return (0); 300 301 #undef COPY_UTSNAME_FIELD 302 } 303 304 static int sys_getcwd1(char *buf, size_t size) 305 { 306 if (getcwd(buf, size) == NULL) { 307 /* getcwd() sets errno */ 308 return (-1); 309 } 310 return strlen(buf)+1; 311 } 312 313 #ifdef CONFIG_ATFILE 314 /* 315 * Host system seems to have atfile syscall stubs available. We 316 * now enable them one by one as specified by target syscall_nr.h. 317 */ 318 319 #ifdef TARGET_NR_faccessat 320 static int sys_faccessat(int dirfd, const char *pathname, int mode) 321 { 322 return (faccessat(dirfd, pathname, mode, 0)); 323 } 324 #endif 325 #ifdef TARGET_NR_fchmodat 326 static int sys_fchmodat(int dirfd, const char *pathname, mode_t mode) 327 { 328 return (fchmodat(dirfd, pathname, mode, 0)); 329 } 330 #endif 331 #if defined(TARGET_NR_fchownat) && defined(USE_UID16) 332 static int sys_fchownat(int dirfd, const char *pathname, uid_t owner, 333 gid_t group, int flags) 334 { 335 return (fchownat(dirfd, pathname, owner, group, flags)); 336 } 337 #endif 338 #ifdef __NR_fstatat64 339 static int sys_fstatat64(int dirfd, const char *pathname, struct stat *buf, 340 int flags) 341 { 342 return (fstatat(dirfd, pathname, buf, flags)); 343 } 344 #endif 345 #ifdef __NR_newfstatat 346 static int sys_newfstatat(int dirfd, const char *pathname, struct stat *buf, 347 int flags) 348 { 349 return (fstatat(dirfd, pathname, buf, flags)); 350 } 351 #endif 352 #ifdef TARGET_NR_futimesat 353 static int sys_futimesat(int dirfd, const char *pathname, 354 const struct timeval times[2]) 355 { 356 return (futimesat(dirfd, pathname, times)); 357 } 358 #endif 359 #ifdef TARGET_NR_linkat 360 static int sys_linkat(int olddirfd, const char *oldpath, 361 int newdirfd, const char *newpath, int flags) 362 { 363 return (linkat(olddirfd, oldpath, newdirfd, newpath, flags)); 364 } 365 #endif 366 #ifdef TARGET_NR_mkdirat 367 static int sys_mkdirat(int dirfd, const char *pathname, mode_t mode) 368 { 369 return (mkdirat(dirfd, pathname, mode)); 370 } 371 #endif 372 #ifdef TARGET_NR_mknodat 373 static int sys_mknodat(int dirfd, const char *pathname, mode_t mode, 374 dev_t dev) 375 { 376 return (mknodat(dirfd, pathname, mode, dev)); 377 } 378 #endif 379 #ifdef TARGET_NR_openat 380 static int sys_openat(int dirfd, const char *pathname, int flags, ...) 381 { 382 /* 383 * open(2) has extra parameter 'mode' when called with 384 * flag O_CREAT. 385 */ 386 if ((flags & O_CREAT) != 0) { 387 va_list ap; 388 mode_t mode; 389 390 /* 391 * Get the 'mode' parameter and translate it to 392 * host bits. 393 */ 394 va_start(ap, flags); 395 mode = va_arg(ap, mode_t); 396 mode = target_to_host_bitmask(mode, fcntl_flags_tbl); 397 va_end(ap); 398 399 return (openat(dirfd, pathname, flags, mode)); 400 } 401 return (openat(dirfd, pathname, flags)); 402 } 403 #endif 404 #ifdef TARGET_NR_readlinkat 405 static int sys_readlinkat(int dirfd, const char *pathname, char *buf, size_t bufsiz) 406 { 407 return (readlinkat(dirfd, pathname, buf, bufsiz)); 408 } 409 #endif 410 #ifdef TARGET_NR_renameat 411 static int sys_renameat(int olddirfd, const char *oldpath, 412 int newdirfd, const char *newpath) 413 { 414 return (renameat(olddirfd, oldpath, newdirfd, newpath)); 415 } 416 #endif 417 #ifdef TARGET_NR_symlinkat 418 static int sys_symlinkat(const char *oldpath, int newdirfd, const char *newpath) 419 { 420 return (symlinkat(oldpath, newdirfd, newpath)); 421 } 422 #endif 423 #ifdef TARGET_NR_unlinkat 424 static int sys_unlinkat(int dirfd, const char *pathname, int flags) 425 { 426 return (unlinkat(dirfd, pathname, flags)); 427 } 428 #endif 429 #else /* !CONFIG_ATFILE */ 430 431 /* 432 * Try direct syscalls instead 433 */ 434 #if defined(TARGET_NR_faccessat) && defined(__NR_faccessat) 435 _syscall3(int,sys_faccessat,int,dirfd,const char *,pathname,int,mode) 436 #endif 437 #if defined(TARGET_NR_fchmodat) && defined(__NR_fchmodat) 438 _syscall3(int,sys_fchmodat,int,dirfd,const char *,pathname, mode_t,mode) 439 #endif 440 #if defined(TARGET_NR_fchownat) && defined(__NR_fchownat) && defined(USE_UID16) 441 _syscall5(int,sys_fchownat,int,dirfd,const char *,pathname, 442 uid_t,owner,gid_t,group,int,flags) 443 #endif 444 #if (defined(TARGET_NR_fstatat64) || defined(TARGET_NR_newfstatat)) && \ 445 defined(__NR_fstatat64) 446 _syscall4(int,sys_fstatat64,int,dirfd,const char *,pathname, 447 struct stat *,buf,int,flags) 448 #endif 449 #if defined(TARGET_NR_futimesat) && defined(__NR_futimesat) 450 _syscall3(int,sys_futimesat,int,dirfd,const char *,pathname, 451 const struct timeval *,times) 452 #endif 453 #if (defined(TARGET_NR_newfstatat) || defined(TARGET_NR_fstatat64) ) && \ 454 defined(__NR_newfstatat) 455 _syscall4(int,sys_newfstatat,int,dirfd,const char *,pathname, 456 struct stat *,buf,int,flags) 457 #endif 458 #if defined(TARGET_NR_linkat) && defined(__NR_linkat) 459 _syscall5(int,sys_linkat,int,olddirfd,const char *,oldpath, 460 int,newdirfd,const char *,newpath,int,flags) 461 #endif 462 #if defined(TARGET_NR_mkdirat) && defined(__NR_mkdirat) 463 _syscall3(int,sys_mkdirat,int,dirfd,const char *,pathname,mode_t,mode) 464 #endif 465 #if defined(TARGET_NR_mknodat) && defined(__NR_mknodat) 466 _syscall4(int,sys_mknodat,int,dirfd,const char *,pathname, 467 mode_t,mode,dev_t,dev) 468 #endif 469 #if defined(TARGET_NR_openat) && defined(__NR_openat) 470 _syscall4(int,sys_openat,int,dirfd,const char *,pathname,int,flags,mode_t,mode) 471 #endif 472 #if defined(TARGET_NR_readlinkat) && defined(__NR_readlinkat) 473 _syscall4(int,sys_readlinkat,int,dirfd,const char *,pathname, 474 char *,buf,size_t,bufsize) 475 #endif 476 #if defined(TARGET_NR_renameat) && defined(__NR_renameat) 477 _syscall4(int,sys_renameat,int,olddirfd,const char *,oldpath, 478 int,newdirfd,const char *,newpath) 479 #endif 480 #if defined(TARGET_NR_symlinkat) && defined(__NR_symlinkat) 481 _syscall3(int,sys_symlinkat,const char *,oldpath, 482 int,newdirfd,const char *,newpath) 483 #endif 484 #if defined(TARGET_NR_unlinkat) && defined(__NR_unlinkat) 485 _syscall3(int,sys_unlinkat,int,dirfd,const char *,pathname,int,flags) 486 #endif 487 488 #endif /* CONFIG_ATFILE */ 489 490 #ifdef CONFIG_UTIMENSAT 491 static int sys_utimensat(int dirfd, const char *pathname, 492 const struct timespec times[2], int flags) 493 { 494 if (pathname == NULL) 495 return futimens(dirfd, times); 496 else 497 return utimensat(dirfd, pathname, times, flags); 498 } 499 #else 500 #if defined(TARGET_NR_utimensat) && defined(__NR_utimensat) 501 _syscall4(int,sys_utimensat,int,dirfd,const char *,pathname, 502 const struct timespec *,tsp,int,flags) 503 #endif 504 #endif /* CONFIG_UTIMENSAT */ 505 506 #ifdef CONFIG_INOTIFY 507 #include <sys/inotify.h> 508 509 #if defined(TARGET_NR_inotify_init) && defined(__NR_inotify_init) 510 static int sys_inotify_init(void) 511 { 512 return (inotify_init()); 513 } 514 #endif 515 #if defined(TARGET_NR_inotify_add_watch) && defined(__NR_inotify_add_watch) 516 static int sys_inotify_add_watch(int fd,const char *pathname, int32_t mask) 517 { 518 return (inotify_add_watch(fd, pathname, mask)); 519 } 520 #endif 521 #if defined(TARGET_NR_inotify_rm_watch) && defined(__NR_inotify_rm_watch) 522 static int sys_inotify_rm_watch(int fd, int32_t wd) 523 { 524 return (inotify_rm_watch(fd, wd)); 525 } 526 #endif 527 #ifdef CONFIG_INOTIFY1 528 #if defined(TARGET_NR_inotify_init1) && defined(__NR_inotify_init1) 529 static int sys_inotify_init1(int flags) 530 { 531 return (inotify_init1(flags)); 532 } 533 #endif 534 #endif 535 #else 536 /* Userspace can usually survive runtime without inotify */ 537 #undef TARGET_NR_inotify_init 538 #undef TARGET_NR_inotify_init1 539 #undef TARGET_NR_inotify_add_watch 540 #undef TARGET_NR_inotify_rm_watch 541 #endif /* CONFIG_INOTIFY */ 542 543 #if defined(TARGET_NR_ppoll) 544 #ifndef __NR_ppoll 545 # define __NR_ppoll -1 546 #endif 547 #define __NR_sys_ppoll __NR_ppoll 548 _syscall5(int, sys_ppoll, struct pollfd *, fds, nfds_t, nfds, 549 struct timespec *, timeout, const __sigset_t *, sigmask, 550 size_t, sigsetsize) 551 #endif 552 553 extern int personality(int); 554 extern int flock(int, int); 555 extern int setfsuid(int); 556 extern int setfsgid(int); 557 extern int setgroups(int, gid_t *); 558 559 #define ERRNO_TABLE_SIZE 1200 560 561 /* target_to_host_errno_table[] is initialized from 562 * host_to_target_errno_table[] in syscall_init(). */ 563 static uint16_t target_to_host_errno_table[ERRNO_TABLE_SIZE] = { 564 }; 565 566 /* 567 * This list is the union of errno values overridden in asm-<arch>/errno.h 568 * minus the errnos that are not actually generic to all archs. 569 */ 570 static uint16_t host_to_target_errno_table[ERRNO_TABLE_SIZE] = { 571 [EIDRM] = TARGET_EIDRM, 572 [ECHRNG] = TARGET_ECHRNG, 573 [EL2NSYNC] = TARGET_EL2NSYNC, 574 [EL3HLT] = TARGET_EL3HLT, 575 [EL3RST] = TARGET_EL3RST, 576 [ELNRNG] = TARGET_ELNRNG, 577 [EUNATCH] = TARGET_EUNATCH, 578 [ENOCSI] = TARGET_ENOCSI, 579 [EL2HLT] = TARGET_EL2HLT, 580 [EDEADLK] = TARGET_EDEADLK, 581 [ENOLCK] = TARGET_ENOLCK, 582 [EBADE] = TARGET_EBADE, 583 [EBADR] = TARGET_EBADR, 584 [EXFULL] = TARGET_EXFULL, 585 [ENOANO] = TARGET_ENOANO, 586 [EBADRQC] = TARGET_EBADRQC, 587 [EBADSLT] = TARGET_EBADSLT, 588 [EBFONT] = TARGET_EBFONT, 589 [ENOSTR] = TARGET_ENOSTR, 590 [ENODATA] = TARGET_ENODATA, 591 [ETIME] = TARGET_ETIME, 592 [ENOSR] = TARGET_ENOSR, 593 [ENONET] = TARGET_ENONET, 594 [ENOPKG] = TARGET_ENOPKG, 595 [EREMOTE] = TARGET_EREMOTE, 596 [ENOLINK] = TARGET_ENOLINK, 597 [EADV] = TARGET_EADV, 598 [ESRMNT] = TARGET_ESRMNT, 599 [ECOMM] = TARGET_ECOMM, 600 [EPROTO] = TARGET_EPROTO, 601 [EDOTDOT] = TARGET_EDOTDOT, 602 [EMULTIHOP] = TARGET_EMULTIHOP, 603 [EBADMSG] = TARGET_EBADMSG, 604 [ENAMETOOLONG] = TARGET_ENAMETOOLONG, 605 [EOVERFLOW] = TARGET_EOVERFLOW, 606 [ENOTUNIQ] = TARGET_ENOTUNIQ, 607 [EBADFD] = TARGET_EBADFD, 608 [EREMCHG] = TARGET_EREMCHG, 609 [ELIBACC] = TARGET_ELIBACC, 610 [ELIBBAD] = TARGET_ELIBBAD, 611 [ELIBSCN] = TARGET_ELIBSCN, 612 [ELIBMAX] = TARGET_ELIBMAX, 613 [ELIBEXEC] = TARGET_ELIBEXEC, 614 [EILSEQ] = TARGET_EILSEQ, 615 [ENOSYS] = TARGET_ENOSYS, 616 [ELOOP] = TARGET_ELOOP, 617 [ERESTART] = TARGET_ERESTART, 618 [ESTRPIPE] = TARGET_ESTRPIPE, 619 [ENOTEMPTY] = TARGET_ENOTEMPTY, 620 [EUSERS] = TARGET_EUSERS, 621 [ENOTSOCK] = TARGET_ENOTSOCK, 622 [EDESTADDRREQ] = TARGET_EDESTADDRREQ, 623 [EMSGSIZE] = TARGET_EMSGSIZE, 624 [EPROTOTYPE] = TARGET_EPROTOTYPE, 625 [ENOPROTOOPT] = TARGET_ENOPROTOOPT, 626 [EPROTONOSUPPORT] = TARGET_EPROTONOSUPPORT, 627 [ESOCKTNOSUPPORT] = TARGET_ESOCKTNOSUPPORT, 628 [EOPNOTSUPP] = TARGET_EOPNOTSUPP, 629 [EPFNOSUPPORT] = TARGET_EPFNOSUPPORT, 630 [EAFNOSUPPORT] = TARGET_EAFNOSUPPORT, 631 [EADDRINUSE] = TARGET_EADDRINUSE, 632 [EADDRNOTAVAIL] = TARGET_EADDRNOTAVAIL, 633 [ENETDOWN] = TARGET_ENETDOWN, 634 [ENETUNREACH] = TARGET_ENETUNREACH, 635 [ENETRESET] = TARGET_ENETRESET, 636 [ECONNABORTED] = TARGET_ECONNABORTED, 637 [ECONNRESET] = TARGET_ECONNRESET, 638 [ENOBUFS] = TARGET_ENOBUFS, 639 [EISCONN] = TARGET_EISCONN, 640 [ENOTCONN] = TARGET_ENOTCONN, 641 [EUCLEAN] = TARGET_EUCLEAN, 642 [ENOTNAM] = TARGET_ENOTNAM, 643 [ENAVAIL] = TARGET_ENAVAIL, 644 [EISNAM] = TARGET_EISNAM, 645 [EREMOTEIO] = TARGET_EREMOTEIO, 646 [ESHUTDOWN] = TARGET_ESHUTDOWN, 647 [ETOOMANYREFS] = TARGET_ETOOMANYREFS, 648 [ETIMEDOUT] = TARGET_ETIMEDOUT, 649 [ECONNREFUSED] = TARGET_ECONNREFUSED, 650 [EHOSTDOWN] = TARGET_EHOSTDOWN, 651 [EHOSTUNREACH] = TARGET_EHOSTUNREACH, 652 [EALREADY] = TARGET_EALREADY, 653 [EINPROGRESS] = TARGET_EINPROGRESS, 654 [ESTALE] = TARGET_ESTALE, 655 [ECANCELED] = TARGET_ECANCELED, 656 [ENOMEDIUM] = TARGET_ENOMEDIUM, 657 [EMEDIUMTYPE] = TARGET_EMEDIUMTYPE, 658 #ifdef ENOKEY 659 [ENOKEY] = TARGET_ENOKEY, 660 #endif 661 #ifdef EKEYEXPIRED 662 [EKEYEXPIRED] = TARGET_EKEYEXPIRED, 663 #endif 664 #ifdef EKEYREVOKED 665 [EKEYREVOKED] = TARGET_EKEYREVOKED, 666 #endif 667 #ifdef EKEYREJECTED 668 [EKEYREJECTED] = TARGET_EKEYREJECTED, 669 #endif 670 #ifdef EOWNERDEAD 671 [EOWNERDEAD] = TARGET_EOWNERDEAD, 672 #endif 673 #ifdef ENOTRECOVERABLE 674 [ENOTRECOVERABLE] = TARGET_ENOTRECOVERABLE, 675 #endif 676 }; 677 678 static inline int host_to_target_errno(int err) 679 { 680 if(host_to_target_errno_table[err]) 681 return host_to_target_errno_table[err]; 682 return err; 683 } 684 685 static inline int target_to_host_errno(int err) 686 { 687 if (target_to_host_errno_table[err]) 688 return target_to_host_errno_table[err]; 689 return err; 690 } 691 692 static inline abi_long get_errno(abi_long ret) 693 { 694 if (ret == -1) 695 return -host_to_target_errno(errno); 696 else 697 return ret; 698 } 699 700 static inline int is_error(abi_long ret) 701 { 702 return (abi_ulong)ret >= (abi_ulong)(-4096); 703 } 704 705 char *target_strerror(int err) 706 { 707 return strerror(target_to_host_errno(err)); 708 } 709 710 static abi_ulong target_brk; 711 static abi_ulong target_original_brk; 712 713 void target_set_brk(abi_ulong new_brk) 714 { 715 target_original_brk = target_brk = HOST_PAGE_ALIGN(new_brk); 716 } 717 718 /* do_brk() must return target values and target errnos. */ 719 abi_long do_brk(abi_ulong new_brk) 720 { 721 abi_ulong brk_page; 722 abi_long mapped_addr; 723 int new_alloc_size; 724 725 if (!new_brk) 726 return target_brk; 727 if (new_brk < target_original_brk) 728 return target_brk; 729 730 brk_page = HOST_PAGE_ALIGN(target_brk); 731 732 /* If the new brk is less than this, set it and we're done... */ 733 if (new_brk < brk_page) { 734 target_brk = new_brk; 735 return target_brk; 736 } 737 738 /* We need to allocate more memory after the brk... */ 739 new_alloc_size = HOST_PAGE_ALIGN(new_brk - brk_page + 1); 740 mapped_addr = get_errno(target_mmap(brk_page, new_alloc_size, 741 PROT_READ|PROT_WRITE, 742 MAP_ANON|MAP_FIXED|MAP_PRIVATE, 0, 0)); 743 744 #if defined(TARGET_ALPHA) 745 /* We (partially) emulate OSF/1 on Alpha, which requires we 746 return a proper errno, not an unchanged brk value. */ 747 if (is_error(mapped_addr)) { 748 return -TARGET_ENOMEM; 749 } 750 #endif 751 752 if (!is_error(mapped_addr)) { 753 target_brk = new_brk; 754 } 755 return target_brk; 756 } 757 758 static inline abi_long copy_from_user_fdset(fd_set *fds, 759 abi_ulong target_fds_addr, 760 int n) 761 { 762 int i, nw, j, k; 763 abi_ulong b, *target_fds; 764 765 nw = (n + TARGET_ABI_BITS - 1) / TARGET_ABI_BITS; 766 if (!(target_fds = lock_user(VERIFY_READ, 767 target_fds_addr, 768 sizeof(abi_ulong) * nw, 769 1))) 770 return -TARGET_EFAULT; 771 772 FD_ZERO(fds); 773 k = 0; 774 for (i = 0; i < nw; i++) { 775 /* grab the abi_ulong */ 776 __get_user(b, &target_fds[i]); 777 for (j = 0; j < TARGET_ABI_BITS; j++) { 778 /* check the bit inside the abi_ulong */ 779 if ((b >> j) & 1) 780 FD_SET(k, fds); 781 k++; 782 } 783 } 784 785 unlock_user(target_fds, target_fds_addr, 0); 786 787 return 0; 788 } 789 790 static inline abi_long copy_to_user_fdset(abi_ulong target_fds_addr, 791 const fd_set *fds, 792 int n) 793 { 794 int i, nw, j, k; 795 abi_long v; 796 abi_ulong *target_fds; 797 798 nw = (n + TARGET_ABI_BITS - 1) / TARGET_ABI_BITS; 799 if (!(target_fds = lock_user(VERIFY_WRITE, 800 target_fds_addr, 801 sizeof(abi_ulong) * nw, 802 0))) 803 return -TARGET_EFAULT; 804 805 k = 0; 806 for (i = 0; i < nw; i++) { 807 v = 0; 808 for (j = 0; j < TARGET_ABI_BITS; j++) { 809 v |= ((FD_ISSET(k, fds) != 0) << j); 810 k++; 811 } 812 __put_user(v, &target_fds[i]); 813 } 814 815 unlock_user(target_fds, target_fds_addr, sizeof(abi_ulong) * nw); 816 817 return 0; 818 } 819 820 #if defined(__alpha__) 821 #define HOST_HZ 1024 822 #else 823 #define HOST_HZ 100 824 #endif 825 826 static inline abi_long host_to_target_clock_t(long ticks) 827 { 828 #if HOST_HZ == TARGET_HZ 829 return ticks; 830 #else 831 return ((int64_t)ticks * TARGET_HZ) / HOST_HZ; 832 #endif 833 } 834 835 static inline abi_long host_to_target_rusage(abi_ulong target_addr, 836 const struct rusage *rusage) 837 { 838 struct target_rusage *target_rusage; 839 840 if (!lock_user_struct(VERIFY_WRITE, target_rusage, target_addr, 0)) 841 return -TARGET_EFAULT; 842 target_rusage->ru_utime.tv_sec = tswapl(rusage->ru_utime.tv_sec); 843 target_rusage->ru_utime.tv_usec = tswapl(rusage->ru_utime.tv_usec); 844 target_rusage->ru_stime.tv_sec = tswapl(rusage->ru_stime.tv_sec); 845 target_rusage->ru_stime.tv_usec = tswapl(rusage->ru_stime.tv_usec); 846 target_rusage->ru_maxrss = tswapl(rusage->ru_maxrss); 847 target_rusage->ru_ixrss = tswapl(rusage->ru_ixrss); 848 target_rusage->ru_idrss = tswapl(rusage->ru_idrss); 849 target_rusage->ru_isrss = tswapl(rusage->ru_isrss); 850 target_rusage->ru_minflt = tswapl(rusage->ru_minflt); 851 target_rusage->ru_majflt = tswapl(rusage->ru_majflt); 852 target_rusage->ru_nswap = tswapl(rusage->ru_nswap); 853 target_rusage->ru_inblock = tswapl(rusage->ru_inblock); 854 target_rusage->ru_oublock = tswapl(rusage->ru_oublock); 855 target_rusage->ru_msgsnd = tswapl(rusage->ru_msgsnd); 856 target_rusage->ru_msgrcv = tswapl(rusage->ru_msgrcv); 857 target_rusage->ru_nsignals = tswapl(rusage->ru_nsignals); 858 target_rusage->ru_nvcsw = tswapl(rusage->ru_nvcsw); 859 target_rusage->ru_nivcsw = tswapl(rusage->ru_nivcsw); 860 unlock_user_struct(target_rusage, target_addr, 1); 861 862 return 0; 863 } 864 865 static inline rlim_t target_to_host_rlim(target_ulong target_rlim) 866 { 867 if (target_rlim == TARGET_RLIM_INFINITY) 868 return RLIM_INFINITY; 869 else 870 return tswapl(target_rlim); 871 } 872 873 static inline target_ulong host_to_target_rlim(rlim_t rlim) 874 { 875 if (rlim == RLIM_INFINITY || rlim != (target_long)rlim) 876 return TARGET_RLIM_INFINITY; 877 else 878 return tswapl(rlim); 879 } 880 881 static inline abi_long copy_from_user_timeval(struct timeval *tv, 882 abi_ulong target_tv_addr) 883 { 884 struct target_timeval *target_tv; 885 886 if (!lock_user_struct(VERIFY_READ, target_tv, target_tv_addr, 1)) 887 return -TARGET_EFAULT; 888 889 __get_user(tv->tv_sec, &target_tv->tv_sec); 890 __get_user(tv->tv_usec, &target_tv->tv_usec); 891 892 unlock_user_struct(target_tv, target_tv_addr, 0); 893 894 return 0; 895 } 896 897 static inline abi_long copy_to_user_timeval(abi_ulong target_tv_addr, 898 const struct timeval *tv) 899 { 900 struct target_timeval *target_tv; 901 902 if (!lock_user_struct(VERIFY_WRITE, target_tv, target_tv_addr, 0)) 903 return -TARGET_EFAULT; 904 905 __put_user(tv->tv_sec, &target_tv->tv_sec); 906 __put_user(tv->tv_usec, &target_tv->tv_usec); 907 908 unlock_user_struct(target_tv, target_tv_addr, 1); 909 910 return 0; 911 } 912 913 #if defined(TARGET_NR_mq_open) && defined(__NR_mq_open) 914 #include <mqueue.h> 915 916 static inline abi_long copy_from_user_mq_attr(struct mq_attr *attr, 917 abi_ulong target_mq_attr_addr) 918 { 919 struct target_mq_attr *target_mq_attr; 920 921 if (!lock_user_struct(VERIFY_READ, target_mq_attr, 922 target_mq_attr_addr, 1)) 923 return -TARGET_EFAULT; 924 925 __get_user(attr->mq_flags, &target_mq_attr->mq_flags); 926 __get_user(attr->mq_maxmsg, &target_mq_attr->mq_maxmsg); 927 __get_user(attr->mq_msgsize, &target_mq_attr->mq_msgsize); 928 __get_user(attr->mq_curmsgs, &target_mq_attr->mq_curmsgs); 929 930 unlock_user_struct(target_mq_attr, target_mq_attr_addr, 0); 931 932 return 0; 933 } 934 935 static inline abi_long copy_to_user_mq_attr(abi_ulong target_mq_attr_addr, 936 const struct mq_attr *attr) 937 { 938 struct target_mq_attr *target_mq_attr; 939 940 if (!lock_user_struct(VERIFY_WRITE, target_mq_attr, 941 target_mq_attr_addr, 0)) 942 return -TARGET_EFAULT; 943 944 __put_user(attr->mq_flags, &target_mq_attr->mq_flags); 945 __put_user(attr->mq_maxmsg, &target_mq_attr->mq_maxmsg); 946 __put_user(attr->mq_msgsize, &target_mq_attr->mq_msgsize); 947 __put_user(attr->mq_curmsgs, &target_mq_attr->mq_curmsgs); 948 949 unlock_user_struct(target_mq_attr, target_mq_attr_addr, 1); 950 951 return 0; 952 } 953 #endif 954 955 /* do_select() must return target values and target errnos. */ 956 static abi_long do_select(int n, 957 abi_ulong rfd_addr, abi_ulong wfd_addr, 958 abi_ulong efd_addr, abi_ulong target_tv_addr) 959 { 960 fd_set rfds, wfds, efds; 961 fd_set *rfds_ptr, *wfds_ptr, *efds_ptr; 962 struct timeval tv, *tv_ptr; 963 abi_long ret; 964 965 if (rfd_addr) { 966 if (copy_from_user_fdset(&rfds, rfd_addr, n)) 967 return -TARGET_EFAULT; 968 rfds_ptr = &rfds; 969 } else { 970 rfds_ptr = NULL; 971 } 972 if (wfd_addr) { 973 if (copy_from_user_fdset(&wfds, wfd_addr, n)) 974 return -TARGET_EFAULT; 975 wfds_ptr = &wfds; 976 } else { 977 wfds_ptr = NULL; 978 } 979 if (efd_addr) { 980 if (copy_from_user_fdset(&efds, efd_addr, n)) 981 return -TARGET_EFAULT; 982 efds_ptr = &efds; 983 } else { 984 efds_ptr = NULL; 985 } 986 987 if (target_tv_addr) { 988 if (copy_from_user_timeval(&tv, target_tv_addr)) 989 return -TARGET_EFAULT; 990 tv_ptr = &tv; 991 } else { 992 tv_ptr = NULL; 993 } 994 995 ret = get_errno(select(n, rfds_ptr, wfds_ptr, efds_ptr, tv_ptr)); 996 997 if (!is_error(ret)) { 998 if (rfd_addr && copy_to_user_fdset(rfd_addr, &rfds, n)) 999 return -TARGET_EFAULT; 1000 if (wfd_addr && copy_to_user_fdset(wfd_addr, &wfds, n)) 1001 return -TARGET_EFAULT; 1002 if (efd_addr && copy_to_user_fdset(efd_addr, &efds, n)) 1003 return -TARGET_EFAULT; 1004 1005 if (target_tv_addr && copy_to_user_timeval(target_tv_addr, &tv)) 1006 return -TARGET_EFAULT; 1007 } 1008 1009 return ret; 1010 } 1011 1012 static abi_long do_pipe2(int host_pipe[], int flags) 1013 { 1014 #ifdef CONFIG_PIPE2 1015 return pipe2(host_pipe, flags); 1016 #else 1017 return -ENOSYS; 1018 #endif 1019 } 1020 1021 static abi_long do_pipe(void *cpu_env, abi_ulong pipedes, 1022 int flags, int is_pipe2) 1023 { 1024 int host_pipe[2]; 1025 abi_long ret; 1026 ret = flags ? do_pipe2(host_pipe, flags) : pipe(host_pipe); 1027 1028 if (is_error(ret)) 1029 return get_errno(ret); 1030 1031 /* Several targets have special calling conventions for the original 1032 pipe syscall, but didn't replicate this into the pipe2 syscall. */ 1033 if (!is_pipe2) { 1034 #if defined(TARGET_ALPHA) 1035 ((CPUAlphaState *)cpu_env)->ir[IR_A4] = host_pipe[1]; 1036 return host_pipe[0]; 1037 #elif defined(TARGET_MIPS) 1038 ((CPUMIPSState*)cpu_env)->active_tc.gpr[3] = host_pipe[1]; 1039 return host_pipe[0]; 1040 #elif defined(TARGET_SH4) 1041 ((CPUSH4State*)cpu_env)->gregs[1] = host_pipe[1]; 1042 return host_pipe[0]; 1043 #endif 1044 } 1045 1046 if (put_user_s32(host_pipe[0], pipedes) 1047 || put_user_s32(host_pipe[1], pipedes + sizeof(host_pipe[0]))) 1048 return -TARGET_EFAULT; 1049 return get_errno(ret); 1050 } 1051 1052 static inline abi_long target_to_host_ip_mreq(struct ip_mreqn *mreqn, 1053 abi_ulong target_addr, 1054 socklen_t len) 1055 { 1056 struct target_ip_mreqn *target_smreqn; 1057 1058 target_smreqn = lock_user(VERIFY_READ, target_addr, len, 1); 1059 if (!target_smreqn) 1060 return -TARGET_EFAULT; 1061 mreqn->imr_multiaddr.s_addr = target_smreqn->imr_multiaddr.s_addr; 1062 mreqn->imr_address.s_addr = target_smreqn->imr_address.s_addr; 1063 if (len == sizeof(struct target_ip_mreqn)) 1064 mreqn->imr_ifindex = tswapl(target_smreqn->imr_ifindex); 1065 unlock_user(target_smreqn, target_addr, 0); 1066 1067 return 0; 1068 } 1069 1070 static inline abi_long target_to_host_sockaddr(struct sockaddr *addr, 1071 abi_ulong target_addr, 1072 socklen_t len) 1073 { 1074 const socklen_t unix_maxlen = sizeof (struct sockaddr_un); 1075 sa_family_t sa_family; 1076 struct target_sockaddr *target_saddr; 1077 1078 target_saddr = lock_user(VERIFY_READ, target_addr, len, 1); 1079 if (!target_saddr) 1080 return -TARGET_EFAULT; 1081 1082 sa_family = tswap16(target_saddr->sa_family); 1083 1084 /* Oops. The caller might send a incomplete sun_path; sun_path 1085 * must be terminated by \0 (see the manual page), but 1086 * unfortunately it is quite common to specify sockaddr_un 1087 * length as "strlen(x->sun_path)" while it should be 1088 * "strlen(...) + 1". We'll fix that here if needed. 1089 * Linux kernel has a similar feature. 1090 */ 1091 1092 if (sa_family == AF_UNIX) { 1093 if (len < unix_maxlen && len > 0) { 1094 char *cp = (char*)target_saddr; 1095 1096 if ( cp[len-1] && !cp[len] ) 1097 len++; 1098 } 1099 if (len > unix_maxlen) 1100 len = unix_maxlen; 1101 } 1102 1103 memcpy(addr, target_saddr, len); 1104 addr->sa_family = sa_family; 1105 unlock_user(target_saddr, target_addr, 0); 1106 1107 return 0; 1108 } 1109 1110 static inline abi_long host_to_target_sockaddr(abi_ulong target_addr, 1111 struct sockaddr *addr, 1112 socklen_t len) 1113 { 1114 struct target_sockaddr *target_saddr; 1115 1116 target_saddr = lock_user(VERIFY_WRITE, target_addr, len, 0); 1117 if (!target_saddr) 1118 return -TARGET_EFAULT; 1119 memcpy(target_saddr, addr, len); 1120 target_saddr->sa_family = tswap16(addr->sa_family); 1121 unlock_user(target_saddr, target_addr, len); 1122 1123 return 0; 1124 } 1125 1126 /* ??? Should this also swap msgh->name? */ 1127 static inline abi_long target_to_host_cmsg(struct msghdr *msgh, 1128 struct target_msghdr *target_msgh) 1129 { 1130 struct cmsghdr *cmsg = CMSG_FIRSTHDR(msgh); 1131 abi_long msg_controllen; 1132 abi_ulong target_cmsg_addr; 1133 struct target_cmsghdr *target_cmsg; 1134 socklen_t space = 0; 1135 1136 msg_controllen = tswapl(target_msgh->msg_controllen); 1137 if (msg_controllen < sizeof (struct target_cmsghdr)) 1138 goto the_end; 1139 target_cmsg_addr = tswapl(target_msgh->msg_control); 1140 target_cmsg = lock_user(VERIFY_READ, target_cmsg_addr, msg_controllen, 1); 1141 if (!target_cmsg) 1142 return -TARGET_EFAULT; 1143 1144 while (cmsg && target_cmsg) { 1145 void *data = CMSG_DATA(cmsg); 1146 void *target_data = TARGET_CMSG_DATA(target_cmsg); 1147 1148 int len = tswapl(target_cmsg->cmsg_len) 1149 - TARGET_CMSG_ALIGN(sizeof (struct target_cmsghdr)); 1150 1151 space += CMSG_SPACE(len); 1152 if (space > msgh->msg_controllen) { 1153 space -= CMSG_SPACE(len); 1154 gemu_log("Host cmsg overflow\n"); 1155 break; 1156 } 1157 1158 cmsg->cmsg_level = tswap32(target_cmsg->cmsg_level); 1159 cmsg->cmsg_type = tswap32(target_cmsg->cmsg_type); 1160 cmsg->cmsg_len = CMSG_LEN(len); 1161 1162 if (cmsg->cmsg_level != TARGET_SOL_SOCKET || cmsg->cmsg_type != SCM_RIGHTS) { 1163 gemu_log("Unsupported ancillary data: %d/%d\n", cmsg->cmsg_level, cmsg->cmsg_type); 1164 memcpy(data, target_data, len); 1165 } else { 1166 int *fd = (int *)data; 1167 int *target_fd = (int *)target_data; 1168 int i, numfds = len / sizeof(int); 1169 1170 for (i = 0; i < numfds; i++) 1171 fd[i] = tswap32(target_fd[i]); 1172 } 1173 1174 cmsg = CMSG_NXTHDR(msgh, cmsg); 1175 target_cmsg = TARGET_CMSG_NXTHDR(target_msgh, target_cmsg); 1176 } 1177 unlock_user(target_cmsg, target_cmsg_addr, 0); 1178 the_end: 1179 msgh->msg_controllen = space; 1180 return 0; 1181 } 1182 1183 /* ??? Should this also swap msgh->name? */ 1184 static inline abi_long host_to_target_cmsg(struct target_msghdr *target_msgh, 1185 struct msghdr *msgh) 1186 { 1187 struct cmsghdr *cmsg = CMSG_FIRSTHDR(msgh); 1188 abi_long msg_controllen; 1189 abi_ulong target_cmsg_addr; 1190 struct target_cmsghdr *target_cmsg; 1191 socklen_t space = 0; 1192 1193 msg_controllen = tswapl(target_msgh->msg_controllen); 1194 if (msg_controllen < sizeof (struct target_cmsghdr)) 1195 goto the_end; 1196 target_cmsg_addr = tswapl(target_msgh->msg_control); 1197 target_cmsg = lock_user(VERIFY_WRITE, target_cmsg_addr, msg_controllen, 0); 1198 if (!target_cmsg) 1199 return -TARGET_EFAULT; 1200 1201 while (cmsg && target_cmsg) { 1202 void *data = CMSG_DATA(cmsg); 1203 void *target_data = TARGET_CMSG_DATA(target_cmsg); 1204 1205 int len = cmsg->cmsg_len - CMSG_ALIGN(sizeof (struct cmsghdr)); 1206 1207 space += TARGET_CMSG_SPACE(len); 1208 if (space > msg_controllen) { 1209 space -= TARGET_CMSG_SPACE(len); 1210 gemu_log("Target cmsg overflow\n"); 1211 break; 1212 } 1213 1214 target_cmsg->cmsg_level = tswap32(cmsg->cmsg_level); 1215 target_cmsg->cmsg_type = tswap32(cmsg->cmsg_type); 1216 target_cmsg->cmsg_len = tswapl(TARGET_CMSG_LEN(len)); 1217 1218 if (cmsg->cmsg_level != TARGET_SOL_SOCKET || cmsg->cmsg_type != SCM_RIGHTS) { 1219 gemu_log("Unsupported ancillary data: %d/%d\n", cmsg->cmsg_level, cmsg->cmsg_type); 1220 memcpy(target_data, data, len); 1221 } else { 1222 int *fd = (int *)data; 1223 int *target_fd = (int *)target_data; 1224 int i, numfds = len / sizeof(int); 1225 1226 for (i = 0; i < numfds; i++) 1227 target_fd[i] = tswap32(fd[i]); 1228 } 1229 1230 cmsg = CMSG_NXTHDR(msgh, cmsg); 1231 target_cmsg = TARGET_CMSG_NXTHDR(target_msgh, target_cmsg); 1232 } 1233 unlock_user(target_cmsg, target_cmsg_addr, space); 1234 the_end: 1235 target_msgh->msg_controllen = tswapl(space); 1236 return 0; 1237 } 1238 1239 /* do_setsockopt() Must return target values and target errnos. */ 1240 static abi_long do_setsockopt(int sockfd, int level, int optname, 1241 abi_ulong optval_addr, socklen_t optlen) 1242 { 1243 abi_long ret; 1244 int val; 1245 struct ip_mreqn *ip_mreq; 1246 struct ip_mreq_source *ip_mreq_source; 1247 1248 switch(level) { 1249 case SOL_TCP: 1250 /* TCP options all take an 'int' value. */ 1251 if (optlen < sizeof(uint32_t)) 1252 return -TARGET_EINVAL; 1253 1254 if (get_user_u32(val, optval_addr)) 1255 return -TARGET_EFAULT; 1256 ret = get_errno(setsockopt(sockfd, level, optname, &val, sizeof(val))); 1257 break; 1258 case SOL_IP: 1259 switch(optname) { 1260 case IP_TOS: 1261 case IP_TTL: 1262 case IP_HDRINCL: 1263 case IP_ROUTER_ALERT: 1264 case IP_RECVOPTS: 1265 case IP_RETOPTS: 1266 case IP_PKTINFO: 1267 case IP_MTU_DISCOVER: 1268 case IP_RECVERR: 1269 case IP_RECVTOS: 1270 #ifdef IP_FREEBIND 1271 case IP_FREEBIND: 1272 #endif 1273 case IP_MULTICAST_TTL: 1274 case IP_MULTICAST_LOOP: 1275 val = 0; 1276 if (optlen >= sizeof(uint32_t)) { 1277 if (get_user_u32(val, optval_addr)) 1278 return -TARGET_EFAULT; 1279 } else if (optlen >= 1) { 1280 if (get_user_u8(val, optval_addr)) 1281 return -TARGET_EFAULT; 1282 } 1283 ret = get_errno(setsockopt(sockfd, level, optname, &val, sizeof(val))); 1284 break; 1285 case IP_ADD_MEMBERSHIP: 1286 case IP_DROP_MEMBERSHIP: 1287 if (optlen < sizeof (struct target_ip_mreq) || 1288 optlen > sizeof (struct target_ip_mreqn)) 1289 return -TARGET_EINVAL; 1290 1291 ip_mreq = (struct ip_mreqn *) alloca(optlen); 1292 target_to_host_ip_mreq(ip_mreq, optval_addr, optlen); 1293 ret = get_errno(setsockopt(sockfd, level, optname, ip_mreq, optlen)); 1294 break; 1295 1296 case IP_BLOCK_SOURCE: 1297 case IP_UNBLOCK_SOURCE: 1298 case IP_ADD_SOURCE_MEMBERSHIP: 1299 case IP_DROP_SOURCE_MEMBERSHIP: 1300 if (optlen != sizeof (struct target_ip_mreq_source)) 1301 return -TARGET_EINVAL; 1302 1303 ip_mreq_source = lock_user(VERIFY_READ, optval_addr, optlen, 1); 1304 ret = get_errno(setsockopt(sockfd, level, optname, ip_mreq_source, optlen)); 1305 unlock_user (ip_mreq_source, optval_addr, 0); 1306 break; 1307 1308 default: 1309 goto unimplemented; 1310 } 1311 break; 1312 case TARGET_SOL_SOCKET: 1313 switch (optname) { 1314 /* Options with 'int' argument. */ 1315 case TARGET_SO_DEBUG: 1316 optname = SO_DEBUG; 1317 break; 1318 case TARGET_SO_REUSEADDR: 1319 optname = SO_REUSEADDR; 1320 break; 1321 case TARGET_SO_TYPE: 1322 optname = SO_TYPE; 1323 break; 1324 case TARGET_SO_ERROR: 1325 optname = SO_ERROR; 1326 break; 1327 case TARGET_SO_DONTROUTE: 1328 optname = SO_DONTROUTE; 1329 break; 1330 case TARGET_SO_BROADCAST: 1331 optname = SO_BROADCAST; 1332 break; 1333 case TARGET_SO_SNDBUF: 1334 optname = SO_SNDBUF; 1335 break; 1336 case TARGET_SO_RCVBUF: 1337 optname = SO_RCVBUF; 1338 break; 1339 case TARGET_SO_KEEPALIVE: 1340 optname = SO_KEEPALIVE; 1341 break; 1342 case TARGET_SO_OOBINLINE: 1343 optname = SO_OOBINLINE; 1344 break; 1345 case TARGET_SO_NO_CHECK: 1346 optname = SO_NO_CHECK; 1347 break; 1348 case TARGET_SO_PRIORITY: 1349 optname = SO_PRIORITY; 1350 break; 1351 #ifdef SO_BSDCOMPAT 1352 case TARGET_SO_BSDCOMPAT: 1353 optname = SO_BSDCOMPAT; 1354 break; 1355 #endif 1356 case TARGET_SO_PASSCRED: 1357 optname = SO_PASSCRED; 1358 break; 1359 case TARGET_SO_TIMESTAMP: 1360 optname = SO_TIMESTAMP; 1361 break; 1362 case TARGET_SO_RCVLOWAT: 1363 optname = SO_RCVLOWAT; 1364 break; 1365 case TARGET_SO_RCVTIMEO: 1366 optname = SO_RCVTIMEO; 1367 break; 1368 case TARGET_SO_SNDTIMEO: 1369 optname = SO_SNDTIMEO; 1370 break; 1371 break; 1372 default: 1373 goto unimplemented; 1374 } 1375 if (optlen < sizeof(uint32_t)) 1376 return -TARGET_EINVAL; 1377 1378 if (get_user_u32(val, optval_addr)) 1379 return -TARGET_EFAULT; 1380 ret = get_errno(setsockopt(sockfd, SOL_SOCKET, optname, &val, sizeof(val))); 1381 break; 1382 default: 1383 unimplemented: 1384 gemu_log("Unsupported setsockopt level=%d optname=%d \n", level, optname); 1385 ret = -TARGET_ENOPROTOOPT; 1386 } 1387 return ret; 1388 } 1389 1390 /* do_getsockopt() Must return target values and target errnos. */ 1391 static abi_long do_getsockopt(int sockfd, int level, int optname, 1392 abi_ulong optval_addr, abi_ulong optlen) 1393 { 1394 abi_long ret; 1395 int len, val; 1396 socklen_t lv; 1397 1398 switch(level) { 1399 case TARGET_SOL_SOCKET: 1400 level = SOL_SOCKET; 1401 switch (optname) { 1402 /* These don't just return a single integer */ 1403 case TARGET_SO_LINGER: 1404 case TARGET_SO_RCVTIMEO: 1405 case TARGET_SO_SNDTIMEO: 1406 case TARGET_SO_PEERCRED: 1407 case TARGET_SO_PEERNAME: 1408 goto unimplemented; 1409 /* Options with 'int' argument. */ 1410 case TARGET_SO_DEBUG: 1411 optname = SO_DEBUG; 1412 goto int_case; 1413 case TARGET_SO_REUSEADDR: 1414 optname = SO_REUSEADDR; 1415 goto int_case; 1416 case TARGET_SO_TYPE: 1417 optname = SO_TYPE; 1418 goto int_case; 1419 case TARGET_SO_ERROR: 1420 optname = SO_ERROR; 1421 goto int_case; 1422 case TARGET_SO_DONTROUTE: 1423 optname = SO_DONTROUTE; 1424 goto int_case; 1425 case TARGET_SO_BROADCAST: 1426 optname = SO_BROADCAST; 1427 goto int_case; 1428 case TARGET_SO_SNDBUF: 1429 optname = SO_SNDBUF; 1430 goto int_case; 1431 case TARGET_SO_RCVBUF: 1432 optname = SO_RCVBUF; 1433 goto int_case; 1434 case TARGET_SO_KEEPALIVE: 1435 optname = SO_KEEPALIVE; 1436 goto int_case; 1437 case TARGET_SO_OOBINLINE: 1438 optname = SO_OOBINLINE; 1439 goto int_case; 1440 case TARGET_SO_NO_CHECK: 1441 optname = SO_NO_CHECK; 1442 goto int_case; 1443 case TARGET_SO_PRIORITY: 1444 optname = SO_PRIORITY; 1445 goto int_case; 1446 #ifdef SO_BSDCOMPAT 1447 case TARGET_SO_BSDCOMPAT: 1448 optname = SO_BSDCOMPAT; 1449 goto int_case; 1450 #endif 1451 case TARGET_SO_PASSCRED: 1452 optname = SO_PASSCRED; 1453 goto int_case; 1454 case TARGET_SO_TIMESTAMP: 1455 optname = SO_TIMESTAMP; 1456 goto int_case; 1457 case TARGET_SO_RCVLOWAT: 1458 optname = SO_RCVLOWAT; 1459 goto int_case; 1460 default: 1461 goto int_case; 1462 } 1463 break; 1464 case SOL_TCP: 1465 /* TCP options all take an 'int' value. */ 1466 int_case: 1467 if (get_user_u32(len, optlen)) 1468 return -TARGET_EFAULT; 1469 if (len < 0) 1470 return -TARGET_EINVAL; 1471 lv = sizeof(lv); 1472 ret = get_errno(getsockopt(sockfd, level, optname, &val, &lv)); 1473 if (ret < 0) 1474 return ret; 1475 if (len > lv) 1476 len = lv; 1477 if (len == 4) { 1478 if (put_user_u32(val, optval_addr)) 1479 return -TARGET_EFAULT; 1480 } else { 1481 if (put_user_u8(val, optval_addr)) 1482 return -TARGET_EFAULT; 1483 } 1484 if (put_user_u32(len, optlen)) 1485 return -TARGET_EFAULT; 1486 break; 1487 case SOL_IP: 1488 switch(optname) { 1489 case IP_TOS: 1490 case IP_TTL: 1491 case IP_HDRINCL: 1492 case IP_ROUTER_ALERT: 1493 case IP_RECVOPTS: 1494 case IP_RETOPTS: 1495 case IP_PKTINFO: 1496 case IP_MTU_DISCOVER: 1497 case IP_RECVERR: 1498 case IP_RECVTOS: 1499 #ifdef IP_FREEBIND 1500 case IP_FREEBIND: 1501 #endif 1502 case IP_MULTICAST_TTL: 1503 case IP_MULTICAST_LOOP: 1504 if (get_user_u32(len, optlen)) 1505 return -TARGET_EFAULT; 1506 if (len < 0) 1507 return -TARGET_EINVAL; 1508 lv = sizeof(lv); 1509 ret = get_errno(getsockopt(sockfd, level, optname, &val, &lv)); 1510 if (ret < 0) 1511 return ret; 1512 if (len < sizeof(int) && len > 0 && val >= 0 && val < 255) { 1513 len = 1; 1514 if (put_user_u32(len, optlen) 1515 || put_user_u8(val, optval_addr)) 1516 return -TARGET_EFAULT; 1517 } else { 1518 if (len > sizeof(int)) 1519 len = sizeof(int); 1520 if (put_user_u32(len, optlen) 1521 || put_user_u32(val, optval_addr)) 1522 return -TARGET_EFAULT; 1523 } 1524 break; 1525 default: 1526 ret = -TARGET_ENOPROTOOPT; 1527 break; 1528 } 1529 break; 1530 default: 1531 unimplemented: 1532 gemu_log("getsockopt level=%d optname=%d not yet supported\n", 1533 level, optname); 1534 ret = -TARGET_EOPNOTSUPP; 1535 break; 1536 } 1537 return ret; 1538 } 1539 1540 /* FIXME 1541 * lock_iovec()/unlock_iovec() have a return code of 0 for success where 1542 * other lock functions have a return code of 0 for failure. 1543 */ 1544 static abi_long lock_iovec(int type, struct iovec *vec, abi_ulong target_addr, 1545 int count, int copy) 1546 { 1547 struct target_iovec *target_vec; 1548 abi_ulong base; 1549 int i; 1550 1551 target_vec = lock_user(VERIFY_READ, target_addr, count * sizeof(struct target_iovec), 1); 1552 if (!target_vec) 1553 return -TARGET_EFAULT; 1554 for(i = 0;i < count; i++) { 1555 base = tswapl(target_vec[i].iov_base); 1556 vec[i].iov_len = tswapl(target_vec[i].iov_len); 1557 if (vec[i].iov_len != 0) { 1558 vec[i].iov_base = lock_user(type, base, vec[i].iov_len, copy); 1559 /* Don't check lock_user return value. We must call writev even 1560 if a element has invalid base address. */ 1561 } else { 1562 /* zero length pointer is ignored */ 1563 vec[i].iov_base = NULL; 1564 } 1565 } 1566 unlock_user (target_vec, target_addr, 0); 1567 return 0; 1568 } 1569 1570 static abi_long unlock_iovec(struct iovec *vec, abi_ulong target_addr, 1571 int count, int copy) 1572 { 1573 struct target_iovec *target_vec; 1574 abi_ulong base; 1575 int i; 1576 1577 target_vec = lock_user(VERIFY_READ, target_addr, count * sizeof(struct target_iovec), 1); 1578 if (!target_vec) 1579 return -TARGET_EFAULT; 1580 for(i = 0;i < count; i++) { 1581 if (target_vec[i].iov_base) { 1582 base = tswapl(target_vec[i].iov_base); 1583 unlock_user(vec[i].iov_base, base, copy ? vec[i].iov_len : 0); 1584 } 1585 } 1586 unlock_user (target_vec, target_addr, 0); 1587 1588 return 0; 1589 } 1590 1591 /* do_socket() Must return target values and target errnos. */ 1592 static abi_long do_socket(int domain, int type, int protocol) 1593 { 1594 #if defined(TARGET_MIPS) 1595 switch(type) { 1596 case TARGET_SOCK_DGRAM: 1597 type = SOCK_DGRAM; 1598 break; 1599 case TARGET_SOCK_STREAM: 1600 type = SOCK_STREAM; 1601 break; 1602 case TARGET_SOCK_RAW: 1603 type = SOCK_RAW; 1604 break; 1605 case TARGET_SOCK_RDM: 1606 type = SOCK_RDM; 1607 break; 1608 case TARGET_SOCK_SEQPACKET: 1609 type = SOCK_SEQPACKET; 1610 break; 1611 case TARGET_SOCK_PACKET: 1612 type = SOCK_PACKET; 1613 break; 1614 } 1615 #endif 1616 if (domain == PF_NETLINK) 1617 return -EAFNOSUPPORT; /* do not NETLINK socket connections possible */ 1618 return get_errno(socket(domain, type, protocol)); 1619 } 1620 1621 /* do_bind() Must return target values and target errnos. */ 1622 static abi_long do_bind(int sockfd, abi_ulong target_addr, 1623 socklen_t addrlen) 1624 { 1625 void *addr; 1626 abi_long ret; 1627 1628 if ((int)addrlen < 0) { 1629 return -TARGET_EINVAL; 1630 } 1631 1632 addr = alloca(addrlen+1); 1633 1634 ret = target_to_host_sockaddr(addr, target_addr, addrlen); 1635 if (ret) 1636 return ret; 1637 1638 return get_errno(bind(sockfd, addr, addrlen)); 1639 } 1640 1641 /* do_connect() Must return target values and target errnos. */ 1642 static abi_long do_connect(int sockfd, abi_ulong target_addr, 1643 socklen_t addrlen) 1644 { 1645 void *addr; 1646 abi_long ret; 1647 1648 if ((int)addrlen < 0) { 1649 return -TARGET_EINVAL; 1650 } 1651 1652 addr = alloca(addrlen); 1653 1654 ret = target_to_host_sockaddr(addr, target_addr, addrlen); 1655 if (ret) 1656 return ret; 1657 1658 return get_errno(connect(sockfd, addr, addrlen)); 1659 } 1660 1661 /* do_sendrecvmsg() Must return target values and target errnos. */ 1662 static abi_long do_sendrecvmsg(int fd, abi_ulong target_msg, 1663 int flags, int send) 1664 { 1665 abi_long ret, len; 1666 struct target_msghdr *msgp; 1667 struct msghdr msg; 1668 int count; 1669 struct iovec *vec; 1670 abi_ulong target_vec; 1671 1672 /* FIXME */ 1673 if (!lock_user_struct(send ? VERIFY_READ : VERIFY_WRITE, 1674 msgp, 1675 target_msg, 1676 send ? 1 : 0)) 1677 return -TARGET_EFAULT; 1678 if (msgp->msg_name) { 1679 msg.msg_namelen = tswap32(msgp->msg_namelen); 1680 msg.msg_name = alloca(msg.msg_namelen); 1681 ret = target_to_host_sockaddr(msg.msg_name, tswapl(msgp->msg_name), 1682 msg.msg_namelen); 1683 if (ret) { 1684 unlock_user_struct(msgp, target_msg, send ? 0 : 1); 1685 return ret; 1686 } 1687 } else { 1688 msg.msg_name = NULL; 1689 msg.msg_namelen = 0; 1690 } 1691 msg.msg_controllen = 2 * tswapl(msgp->msg_controllen); 1692 msg.msg_control = alloca(msg.msg_controllen); 1693 msg.msg_flags = tswap32(msgp->msg_flags); 1694 1695 count = tswapl(msgp->msg_iovlen); 1696 vec = alloca(count * sizeof(struct iovec)); 1697 target_vec = tswapl(msgp->msg_iov); 1698 lock_iovec(send ? VERIFY_READ : VERIFY_WRITE, vec, target_vec, count, send); 1699 msg.msg_iovlen = count; 1700 msg.msg_iov = vec; 1701 1702 if (send) { 1703 ret = target_to_host_cmsg(&msg, msgp); 1704 if (ret == 0) 1705 ret = get_errno(sendmsg(fd, &msg, flags)); 1706 } else { 1707 ret = get_errno(recvmsg(fd, &msg, flags)); 1708 if (!is_error(ret)) { 1709 len = ret; 1710 ret = host_to_target_cmsg(msgp, &msg); 1711 if (!is_error(ret)) 1712 ret = len; 1713 } 1714 } 1715 unlock_iovec(vec, target_vec, count, !send); 1716 unlock_user_struct(msgp, target_msg, send ? 0 : 1); 1717 return ret; 1718 } 1719 1720 /* do_accept() Must return target values and target errnos. */ 1721 static abi_long do_accept(int fd, abi_ulong target_addr, 1722 abi_ulong target_addrlen_addr) 1723 { 1724 socklen_t addrlen; 1725 void *addr; 1726 abi_long ret; 1727 1728 if (target_addr == 0) 1729 return get_errno(accept(fd, NULL, NULL)); 1730 1731 /* linux returns EINVAL if addrlen pointer is invalid */ 1732 if (get_user_u32(addrlen, target_addrlen_addr)) 1733 return -TARGET_EINVAL; 1734 1735 if ((int)addrlen < 0) { 1736 return -TARGET_EINVAL; 1737 } 1738 1739 if (!access_ok(VERIFY_WRITE, target_addr, addrlen)) 1740 return -TARGET_EINVAL; 1741 1742 addr = alloca(addrlen); 1743 1744 ret = get_errno(accept(fd, addr, &addrlen)); 1745 if (!is_error(ret)) { 1746 host_to_target_sockaddr(target_addr, addr, addrlen); 1747 if (put_user_u32(addrlen, target_addrlen_addr)) 1748 ret = -TARGET_EFAULT; 1749 } 1750 return ret; 1751 } 1752 1753 /* do_getpeername() Must return target values and target errnos. */ 1754 static abi_long do_getpeername(int fd, abi_ulong target_addr, 1755 abi_ulong target_addrlen_addr) 1756 { 1757 socklen_t addrlen; 1758 void *addr; 1759 abi_long ret; 1760 1761 if (get_user_u32(addrlen, target_addrlen_addr)) 1762 return -TARGET_EFAULT; 1763 1764 if ((int)addrlen < 0) { 1765 return -TARGET_EINVAL; 1766 } 1767 1768 if (!access_ok(VERIFY_WRITE, target_addr, addrlen)) 1769 return -TARGET_EFAULT; 1770 1771 addr = alloca(addrlen); 1772 1773 ret = get_errno(getpeername(fd, addr, &addrlen)); 1774 if (!is_error(ret)) { 1775 host_to_target_sockaddr(target_addr, addr, addrlen); 1776 if (put_user_u32(addrlen, target_addrlen_addr)) 1777 ret = -TARGET_EFAULT; 1778 } 1779 return ret; 1780 } 1781 1782 /* do_getsockname() Must return target values and target errnos. */ 1783 static abi_long do_getsockname(int fd, abi_ulong target_addr, 1784 abi_ulong target_addrlen_addr) 1785 { 1786 socklen_t addrlen; 1787 void *addr; 1788 abi_long ret; 1789 1790 if (get_user_u32(addrlen, target_addrlen_addr)) 1791 return -TARGET_EFAULT; 1792 1793 if ((int)addrlen < 0) { 1794 return -TARGET_EINVAL; 1795 } 1796 1797 if (!access_ok(VERIFY_WRITE, target_addr, addrlen)) 1798 return -TARGET_EFAULT; 1799 1800 addr = alloca(addrlen); 1801 1802 ret = get_errno(getsockname(fd, addr, &addrlen)); 1803 if (!is_error(ret)) { 1804 host_to_target_sockaddr(target_addr, addr, addrlen); 1805 if (put_user_u32(addrlen, target_addrlen_addr)) 1806 ret = -TARGET_EFAULT; 1807 } 1808 return ret; 1809 } 1810 1811 /* do_socketpair() Must return target values and target errnos. */ 1812 static abi_long do_socketpair(int domain, int type, int protocol, 1813 abi_ulong target_tab_addr) 1814 { 1815 int tab[2]; 1816 abi_long ret; 1817 1818 ret = get_errno(socketpair(domain, type, protocol, tab)); 1819 if (!is_error(ret)) { 1820 if (put_user_s32(tab[0], target_tab_addr) 1821 || put_user_s32(tab[1], target_tab_addr + sizeof(tab[0]))) 1822 ret = -TARGET_EFAULT; 1823 } 1824 return ret; 1825 } 1826 1827 /* do_sendto() Must return target values and target errnos. */ 1828 static abi_long do_sendto(int fd, abi_ulong msg, size_t len, int flags, 1829 abi_ulong target_addr, socklen_t addrlen) 1830 { 1831 void *addr; 1832 void *host_msg; 1833 abi_long ret; 1834 1835 if ((int)addrlen < 0) { 1836 return -TARGET_EINVAL; 1837 } 1838 1839 host_msg = lock_user(VERIFY_READ, msg, len, 1); 1840 if (!host_msg) 1841 return -TARGET_EFAULT; 1842 if (target_addr) { 1843 addr = alloca(addrlen); 1844 ret = target_to_host_sockaddr(addr, target_addr, addrlen); 1845 if (ret) { 1846 unlock_user(host_msg, msg, 0); 1847 return ret; 1848 } 1849 ret = get_errno(sendto(fd, host_msg, len, flags, addr, addrlen)); 1850 } else { 1851 ret = get_errno(send(fd, host_msg, len, flags)); 1852 } 1853 unlock_user(host_msg, msg, 0); 1854 return ret; 1855 } 1856 1857 /* do_recvfrom() Must return target values and target errnos. */ 1858 static abi_long do_recvfrom(int fd, abi_ulong msg, size_t len, int flags, 1859 abi_ulong target_addr, 1860 abi_ulong target_addrlen) 1861 { 1862 socklen_t addrlen; 1863 void *addr; 1864 void *host_msg; 1865 abi_long ret; 1866 1867 host_msg = lock_user(VERIFY_WRITE, msg, len, 0); 1868 if (!host_msg) 1869 return -TARGET_EFAULT; 1870 if (target_addr) { 1871 if (get_user_u32(addrlen, target_addrlen)) { 1872 ret = -TARGET_EFAULT; 1873 goto fail; 1874 } 1875 if ((int)addrlen < 0) { 1876 ret = -TARGET_EINVAL; 1877 goto fail; 1878 } 1879 addr = alloca(addrlen); 1880 ret = get_errno(recvfrom(fd, host_msg, len, flags, addr, &addrlen)); 1881 } else { 1882 addr = NULL; /* To keep compiler quiet. */ 1883 ret = get_errno(recv(fd, host_msg, len, flags)); 1884 } 1885 if (!is_error(ret)) { 1886 if (target_addr) { 1887 host_to_target_sockaddr(target_addr, addr, addrlen); 1888 if (put_user_u32(addrlen, target_addrlen)) { 1889 ret = -TARGET_EFAULT; 1890 goto fail; 1891 } 1892 } 1893 unlock_user(host_msg, msg, len); 1894 } else { 1895 fail: 1896 unlock_user(host_msg, msg, 0); 1897 } 1898 return ret; 1899 } 1900 1901 #ifdef TARGET_NR_socketcall 1902 /* do_socketcall() Must return target values and target errnos. */ 1903 static abi_long do_socketcall(int num, abi_ulong vptr) 1904 { 1905 abi_long ret; 1906 const int n = sizeof(abi_ulong); 1907 1908 switch(num) { 1909 case SOCKOP_socket: 1910 { 1911 abi_ulong domain, type, protocol; 1912 1913 if (get_user_ual(domain, vptr) 1914 || get_user_ual(type, vptr + n) 1915 || get_user_ual(protocol, vptr + 2 * n)) 1916 return -TARGET_EFAULT; 1917 1918 ret = do_socket(domain, type, protocol); 1919 } 1920 break; 1921 case SOCKOP_bind: 1922 { 1923 abi_ulong sockfd; 1924 abi_ulong target_addr; 1925 socklen_t addrlen; 1926 1927 if (get_user_ual(sockfd, vptr) 1928 || get_user_ual(target_addr, vptr + n) 1929 || get_user_ual(addrlen, vptr + 2 * n)) 1930 return -TARGET_EFAULT; 1931 1932 ret = do_bind(sockfd, target_addr, addrlen); 1933 } 1934 break; 1935 case SOCKOP_connect: 1936 { 1937 abi_ulong sockfd; 1938 abi_ulong target_addr; 1939 socklen_t addrlen; 1940 1941 if (get_user_ual(sockfd, vptr) 1942 || get_user_ual(target_addr, vptr + n) 1943 || get_user_ual(addrlen, vptr + 2 * n)) 1944 return -TARGET_EFAULT; 1945 1946 ret = do_connect(sockfd, target_addr, addrlen); 1947 } 1948 break; 1949 case SOCKOP_listen: 1950 { 1951 abi_ulong sockfd, backlog; 1952 1953 if (get_user_ual(sockfd, vptr) 1954 || get_user_ual(backlog, vptr + n)) 1955 return -TARGET_EFAULT; 1956 1957 ret = get_errno(listen(sockfd, backlog)); 1958 } 1959 break; 1960 case SOCKOP_accept: 1961 { 1962 abi_ulong sockfd; 1963 abi_ulong target_addr, target_addrlen; 1964 1965 if (get_user_ual(sockfd, vptr) 1966 || get_user_ual(target_addr, vptr + n) 1967 || get_user_ual(target_addrlen, vptr + 2 * n)) 1968 return -TARGET_EFAULT; 1969 1970 ret = do_accept(sockfd, target_addr, target_addrlen); 1971 } 1972 break; 1973 case SOCKOP_getsockname: 1974 { 1975 abi_ulong sockfd; 1976 abi_ulong target_addr, target_addrlen; 1977 1978 if (get_user_ual(sockfd, vptr) 1979 || get_user_ual(target_addr, vptr + n) 1980 || get_user_ual(target_addrlen, vptr + 2 * n)) 1981 return -TARGET_EFAULT; 1982 1983 ret = do_getsockname(sockfd, target_addr, target_addrlen); 1984 } 1985 break; 1986 case SOCKOP_getpeername: 1987 { 1988 abi_ulong sockfd; 1989 abi_ulong target_addr, target_addrlen; 1990 1991 if (get_user_ual(sockfd, vptr) 1992 || get_user_ual(target_addr, vptr + n) 1993 || get_user_ual(target_addrlen, vptr + 2 * n)) 1994 return -TARGET_EFAULT; 1995 1996 ret = do_getpeername(sockfd, target_addr, target_addrlen); 1997 } 1998 break; 1999 case SOCKOP_socketpair: 2000 { 2001 abi_ulong domain, type, protocol; 2002 abi_ulong tab; 2003 2004 if (get_user_ual(domain, vptr) 2005 || get_user_ual(type, vptr + n) 2006 || get_user_ual(protocol, vptr + 2 * n) 2007 || get_user_ual(tab, vptr + 3 * n)) 2008 return -TARGET_EFAULT; 2009 2010 ret = do_socketpair(domain, type, protocol, tab); 2011 } 2012 break; 2013 case SOCKOP_send: 2014 { 2015 abi_ulong sockfd; 2016 abi_ulong msg; 2017 size_t len; 2018 abi_ulong flags; 2019 2020 if (get_user_ual(sockfd, vptr) 2021 || get_user_ual(msg, vptr + n) 2022 || get_user_ual(len, vptr + 2 * n) 2023 || get_user_ual(flags, vptr + 3 * n)) 2024 return -TARGET_EFAULT; 2025 2026 ret = do_sendto(sockfd, msg, len, flags, 0, 0); 2027 } 2028 break; 2029 case SOCKOP_recv: 2030 { 2031 abi_ulong sockfd; 2032 abi_ulong msg; 2033 size_t len; 2034 abi_ulong flags; 2035 2036 if (get_user_ual(sockfd, vptr) 2037 || get_user_ual(msg, vptr + n) 2038 || get_user_ual(len, vptr + 2 * n) 2039 || get_user_ual(flags, vptr + 3 * n)) 2040 return -TARGET_EFAULT; 2041 2042 ret = do_recvfrom(sockfd, msg, len, flags, 0, 0); 2043 } 2044 break; 2045 case SOCKOP_sendto: 2046 { 2047 abi_ulong sockfd; 2048 abi_ulong msg; 2049 size_t len; 2050 abi_ulong flags; 2051 abi_ulong addr; 2052 socklen_t addrlen; 2053 2054 if (get_user_ual(sockfd, vptr) 2055 || get_user_ual(msg, vptr + n) 2056 || get_user_ual(len, vptr + 2 * n) 2057 || get_user_ual(flags, vptr + 3 * n) 2058 || get_user_ual(addr, vptr + 4 * n) 2059 || get_user_ual(addrlen, vptr + 5 * n)) 2060 return -TARGET_EFAULT; 2061 2062 ret = do_sendto(sockfd, msg, len, flags, addr, addrlen); 2063 } 2064 break; 2065 case SOCKOP_recvfrom: 2066 { 2067 abi_ulong sockfd; 2068 abi_ulong msg; 2069 size_t len; 2070 abi_ulong flags; 2071 abi_ulong addr; 2072 socklen_t addrlen; 2073 2074 if (get_user_ual(sockfd, vptr) 2075 || get_user_ual(msg, vptr + n) 2076 || get_user_ual(len, vptr + 2 * n) 2077 || get_user_ual(flags, vptr + 3 * n) 2078 || get_user_ual(addr, vptr + 4 * n) 2079 || get_user_ual(addrlen, vptr + 5 * n)) 2080 return -TARGET_EFAULT; 2081 2082 ret = do_recvfrom(sockfd, msg, len, flags, addr, addrlen); 2083 } 2084 break; 2085 case SOCKOP_shutdown: 2086 { 2087 abi_ulong sockfd, how; 2088 2089 if (get_user_ual(sockfd, vptr) 2090 || get_user_ual(how, vptr + n)) 2091 return -TARGET_EFAULT; 2092 2093 ret = get_errno(shutdown(sockfd, how)); 2094 } 2095 break; 2096 case SOCKOP_sendmsg: 2097 case SOCKOP_recvmsg: 2098 { 2099 abi_ulong fd; 2100 abi_ulong target_msg; 2101 abi_ulong flags; 2102 2103 if (get_user_ual(fd, vptr) 2104 || get_user_ual(target_msg, vptr + n) 2105 || get_user_ual(flags, vptr + 2 * n)) 2106 return -TARGET_EFAULT; 2107 2108 ret = do_sendrecvmsg(fd, target_msg, flags, 2109 (num == SOCKOP_sendmsg)); 2110 } 2111 break; 2112 case SOCKOP_setsockopt: 2113 { 2114 abi_ulong sockfd; 2115 abi_ulong level; 2116 abi_ulong optname; 2117 abi_ulong optval; 2118 socklen_t optlen; 2119 2120 if (get_user_ual(sockfd, vptr) 2121 || get_user_ual(level, vptr + n) 2122 || get_user_ual(optname, vptr + 2 * n) 2123 || get_user_ual(optval, vptr + 3 * n) 2124 || get_user_ual(optlen, vptr + 4 * n)) 2125 return -TARGET_EFAULT; 2126 2127 ret = do_setsockopt(sockfd, level, optname, optval, optlen); 2128 } 2129 break; 2130 case SOCKOP_getsockopt: 2131 { 2132 abi_ulong sockfd; 2133 abi_ulong level; 2134 abi_ulong optname; 2135 abi_ulong optval; 2136 socklen_t optlen; 2137 2138 if (get_user_ual(sockfd, vptr) 2139 || get_user_ual(level, vptr + n) 2140 || get_user_ual(optname, vptr + 2 * n) 2141 || get_user_ual(optval, vptr + 3 * n) 2142 || get_user_ual(optlen, vptr + 4 * n)) 2143 return -TARGET_EFAULT; 2144 2145 ret = do_getsockopt(sockfd, level, optname, optval, optlen); 2146 } 2147 break; 2148 default: 2149 gemu_log("Unsupported socketcall: %d\n", num); 2150 ret = -TARGET_ENOSYS; 2151 break; 2152 } 2153 return ret; 2154 } 2155 #endif 2156 2157 #define N_SHM_REGIONS 32 2158 2159 static struct shm_region { 2160 abi_ulong start; 2161 abi_ulong size; 2162 } shm_regions[N_SHM_REGIONS]; 2163 2164 struct target_ipc_perm 2165 { 2166 abi_long __key; 2167 abi_ulong uid; 2168 abi_ulong gid; 2169 abi_ulong cuid; 2170 abi_ulong cgid; 2171 unsigned short int mode; 2172 unsigned short int __pad1; 2173 unsigned short int __seq; 2174 unsigned short int __pad2; 2175 abi_ulong __unused1; 2176 abi_ulong __unused2; 2177 }; 2178 2179 struct target_semid_ds 2180 { 2181 struct target_ipc_perm sem_perm; 2182 abi_ulong sem_otime; 2183 abi_ulong __unused1; 2184 abi_ulong sem_ctime; 2185 abi_ulong __unused2; 2186 abi_ulong sem_nsems; 2187 abi_ulong __unused3; 2188 abi_ulong __unused4; 2189 }; 2190 2191 static inline abi_long target_to_host_ipc_perm(struct ipc_perm *host_ip, 2192 abi_ulong target_addr) 2193 { 2194 struct target_ipc_perm *target_ip; 2195 struct target_semid_ds *target_sd; 2196 2197 if (!lock_user_struct(VERIFY_READ, target_sd, target_addr, 1)) 2198 return -TARGET_EFAULT; 2199 target_ip = &(target_sd->sem_perm); 2200 host_ip->__key = tswapl(target_ip->__key); 2201 host_ip->uid = tswapl(target_ip->uid); 2202 host_ip->gid = tswapl(target_ip->gid); 2203 host_ip->cuid = tswapl(target_ip->cuid); 2204 host_ip->cgid = tswapl(target_ip->cgid); 2205 host_ip->mode = tswapl(target_ip->mode); 2206 unlock_user_struct(target_sd, target_addr, 0); 2207 return 0; 2208 } 2209 2210 static inline abi_long host_to_target_ipc_perm(abi_ulong target_addr, 2211 struct ipc_perm *host_ip) 2212 { 2213 struct target_ipc_perm *target_ip; 2214 struct target_semid_ds *target_sd; 2215 2216 if (!lock_user_struct(VERIFY_WRITE, target_sd, target_addr, 0)) 2217 return -TARGET_EFAULT; 2218 target_ip = &(target_sd->sem_perm); 2219 target_ip->__key = tswapl(host_ip->__key); 2220 target_ip->uid = tswapl(host_ip->uid); 2221 target_ip->gid = tswapl(host_ip->gid); 2222 target_ip->cuid = tswapl(host_ip->cuid); 2223 target_ip->cgid = tswapl(host_ip->cgid); 2224 target_ip->mode = tswapl(host_ip->mode); 2225 unlock_user_struct(target_sd, target_addr, 1); 2226 return 0; 2227 } 2228 2229 static inline abi_long target_to_host_semid_ds(struct semid_ds *host_sd, 2230 abi_ulong target_addr) 2231 { 2232 struct target_semid_ds *target_sd; 2233 2234 if (!lock_user_struct(VERIFY_READ, target_sd, target_addr, 1)) 2235 return -TARGET_EFAULT; 2236 if (target_to_host_ipc_perm(&(host_sd->sem_perm),target_addr)) 2237 return -TARGET_EFAULT; 2238 host_sd->sem_nsems = tswapl(target_sd->sem_nsems); 2239 host_sd->sem_otime = tswapl(target_sd->sem_otime); 2240 host_sd->sem_ctime = tswapl(target_sd->sem_ctime); 2241 unlock_user_struct(target_sd, target_addr, 0); 2242 return 0; 2243 } 2244 2245 static inline abi_long host_to_target_semid_ds(abi_ulong target_addr, 2246 struct semid_ds *host_sd) 2247 { 2248 struct target_semid_ds *target_sd; 2249 2250 if (!lock_user_struct(VERIFY_WRITE, target_sd, target_addr, 0)) 2251 return -TARGET_EFAULT; 2252 if (host_to_target_ipc_perm(target_addr,&(host_sd->sem_perm))) 2253 return -TARGET_EFAULT;; 2254 target_sd->sem_nsems = tswapl(host_sd->sem_nsems); 2255 target_sd->sem_otime = tswapl(host_sd->sem_otime); 2256 target_sd->sem_ctime = tswapl(host_sd->sem_ctime); 2257 unlock_user_struct(target_sd, target_addr, 1); 2258 return 0; 2259 } 2260 2261 struct target_seminfo { 2262 int semmap; 2263 int semmni; 2264 int semmns; 2265 int semmnu; 2266 int semmsl; 2267 int semopm; 2268 int semume; 2269 int semusz; 2270 int semvmx; 2271 int semaem; 2272 }; 2273 2274 static inline abi_long host_to_target_seminfo(abi_ulong target_addr, 2275 struct seminfo *host_seminfo) 2276 { 2277 struct target_seminfo *target_seminfo; 2278 if (!lock_user_struct(VERIFY_WRITE, target_seminfo, target_addr, 0)) 2279 return -TARGET_EFAULT; 2280 __put_user(host_seminfo->semmap, &target_seminfo->semmap); 2281 __put_user(host_seminfo->semmni, &target_seminfo->semmni); 2282 __put_user(host_seminfo->semmns, &target_seminfo->semmns); 2283 __put_user(host_seminfo->semmnu, &target_seminfo->semmnu); 2284 __put_user(host_seminfo->semmsl, &target_seminfo->semmsl); 2285 __put_user(host_seminfo->semopm, &target_seminfo->semopm); 2286 __put_user(host_seminfo->semume, &target_seminfo->semume); 2287 __put_user(host_seminfo->semusz, &target_seminfo->semusz); 2288 __put_user(host_seminfo->semvmx, &target_seminfo->semvmx); 2289 __put_user(host_seminfo->semaem, &target_seminfo->semaem); 2290 unlock_user_struct(target_seminfo, target_addr, 1); 2291 return 0; 2292 } 2293 2294 union semun { 2295 int val; 2296 struct semid_ds *buf; 2297 unsigned short *array; 2298 struct seminfo *__buf; 2299 }; 2300 2301 union target_semun { 2302 int val; 2303 abi_ulong buf; 2304 abi_ulong array; 2305 abi_ulong __buf; 2306 }; 2307 2308 static inline abi_long target_to_host_semarray(int semid, unsigned short **host_array, 2309 abi_ulong target_addr) 2310 { 2311 int nsems; 2312 unsigned short *array; 2313 union semun semun; 2314 struct semid_ds semid_ds; 2315 int i, ret; 2316 2317 semun.buf = &semid_ds; 2318 2319 ret = semctl(semid, 0, IPC_STAT, semun); 2320 if (ret == -1) 2321 return get_errno(ret); 2322 2323 nsems = semid_ds.sem_nsems; 2324 2325 *host_array = malloc(nsems*sizeof(unsigned short)); 2326 array = lock_user(VERIFY_READ, target_addr, 2327 nsems*sizeof(unsigned short), 1); 2328 if (!array) 2329 return -TARGET_EFAULT; 2330 2331 for(i=0; i<nsems; i++) { 2332 __get_user((*host_array)[i], &array[i]); 2333 } 2334 unlock_user(array, target_addr, 0); 2335 2336 return 0; 2337 } 2338 2339 static inline abi_long host_to_target_semarray(int semid, abi_ulong target_addr, 2340 unsigned short **host_array) 2341 { 2342 int nsems; 2343 unsigned short *array; 2344 union semun semun; 2345 struct semid_ds semid_ds; 2346 int i, ret; 2347 2348 semun.buf = &semid_ds; 2349 2350 ret = semctl(semid, 0, IPC_STAT, semun); 2351 if (ret == -1) 2352 return get_errno(ret); 2353 2354 nsems = semid_ds.sem_nsems; 2355 2356 array = lock_user(VERIFY_WRITE, target_addr, 2357 nsems*sizeof(unsigned short), 0); 2358 if (!array) 2359 return -TARGET_EFAULT; 2360 2361 for(i=0; i<nsems; i++) { 2362 __put_user((*host_array)[i], &array[i]); 2363 } 2364 free(*host_array); 2365 unlock_user(array, target_addr, 1); 2366 2367 return 0; 2368 } 2369 2370 static inline abi_long do_semctl(int semid, int semnum, int cmd, 2371 union target_semun target_su) 2372 { 2373 union semun arg; 2374 struct semid_ds dsarg; 2375 unsigned short *array = NULL; 2376 struct seminfo seminfo; 2377 abi_long ret = -TARGET_EINVAL; 2378 abi_long err; 2379 cmd &= 0xff; 2380 2381 switch( cmd ) { 2382 case GETVAL: 2383 case SETVAL: 2384 arg.val = tswapl(target_su.val); 2385 ret = get_errno(semctl(semid, semnum, cmd, arg)); 2386 target_su.val = tswapl(arg.val); 2387 break; 2388 case GETALL: 2389 case SETALL: 2390 err = target_to_host_semarray(semid, &array, target_su.array); 2391 if (err) 2392 return err; 2393 arg.array = array; 2394 ret = get_errno(semctl(semid, semnum, cmd, arg)); 2395 err = host_to_target_semarray(semid, target_su.array, &array); 2396 if (err) 2397 return err; 2398 break; 2399 case IPC_STAT: 2400 case IPC_SET: 2401 case SEM_STAT: 2402 err = target_to_host_semid_ds(&dsarg, target_su.buf); 2403 if (err) 2404 return err; 2405 arg.buf = &dsarg; 2406 ret = get_errno(semctl(semid, semnum, cmd, arg)); 2407 err = host_to_target_semid_ds(target_su.buf, &dsarg); 2408 if (err) 2409 return err; 2410 break; 2411 case IPC_INFO: 2412 case SEM_INFO: 2413 arg.__buf = &seminfo; 2414 ret = get_errno(semctl(semid, semnum, cmd, arg)); 2415 err = host_to_target_seminfo(target_su.__buf, &seminfo); 2416 if (err) 2417 return err; 2418 break; 2419 case IPC_RMID: 2420 case GETPID: 2421 case GETNCNT: 2422 case GETZCNT: 2423 ret = get_errno(semctl(semid, semnum, cmd, NULL)); 2424 break; 2425 } 2426 2427 return ret; 2428 } 2429 2430 struct target_sembuf { 2431 unsigned short sem_num; 2432 short sem_op; 2433 short sem_flg; 2434 }; 2435 2436 static inline abi_long target_to_host_sembuf(struct sembuf *host_sembuf, 2437 abi_ulong target_addr, 2438 unsigned nsops) 2439 { 2440 struct target_sembuf *target_sembuf; 2441 int i; 2442 2443 target_sembuf = lock_user(VERIFY_READ, target_addr, 2444 nsops*sizeof(struct target_sembuf), 1); 2445 if (!target_sembuf) 2446 return -TARGET_EFAULT; 2447 2448 for(i=0; i<nsops; i++) { 2449 __get_user(host_sembuf[i].sem_num, &target_sembuf[i].sem_num); 2450 __get_user(host_sembuf[i].sem_op, &target_sembuf[i].sem_op); 2451 __get_user(host_sembuf[i].sem_flg, &target_sembuf[i].sem_flg); 2452 } 2453 2454 unlock_user(target_sembuf, target_addr, 0); 2455 2456 return 0; 2457 } 2458 2459 static inline abi_long do_semop(int semid, abi_long ptr, unsigned nsops) 2460 { 2461 struct sembuf sops[nsops]; 2462 2463 if (target_to_host_sembuf(sops, ptr, nsops)) 2464 return -TARGET_EFAULT; 2465 2466 return semop(semid, sops, nsops); 2467 } 2468 2469 struct target_msqid_ds 2470 { 2471 struct target_ipc_perm msg_perm; 2472 abi_ulong msg_stime; 2473 #if TARGET_ABI_BITS == 32 2474 abi_ulong __unused1; 2475 #endif 2476 abi_ulong msg_rtime; 2477 #if TARGET_ABI_BITS == 32 2478 abi_ulong __unused2; 2479 #endif 2480 abi_ulong msg_ctime; 2481 #if TARGET_ABI_BITS == 32 2482 abi_ulong __unused3; 2483 #endif 2484 abi_ulong __msg_cbytes; 2485 abi_ulong msg_qnum; 2486 abi_ulong msg_qbytes; 2487 abi_ulong msg_lspid; 2488 abi_ulong msg_lrpid; 2489 abi_ulong __unused4; 2490 abi_ulong __unused5; 2491 }; 2492 2493 static inline abi_long target_to_host_msqid_ds(struct msqid_ds *host_md, 2494 abi_ulong target_addr) 2495 { 2496 struct target_msqid_ds *target_md; 2497 2498 if (!lock_user_struct(VERIFY_READ, target_md, target_addr, 1)) 2499 return -TARGET_EFAULT; 2500 if (target_to_host_ipc_perm(&(host_md->msg_perm),target_addr)) 2501 return -TARGET_EFAULT; 2502 host_md->msg_stime = tswapl(target_md->msg_stime); 2503 host_md->msg_rtime = tswapl(target_md->msg_rtime); 2504 host_md->msg_ctime = tswapl(target_md->msg_ctime); 2505 host_md->__msg_cbytes = tswapl(target_md->__msg_cbytes); 2506 host_md->msg_qnum = tswapl(target_md->msg_qnum); 2507 host_md->msg_qbytes = tswapl(target_md->msg_qbytes); 2508 host_md->msg_lspid = tswapl(target_md->msg_lspid); 2509 host_md->msg_lrpid = tswapl(target_md->msg_lrpid); 2510 unlock_user_struct(target_md, target_addr, 0); 2511 return 0; 2512 } 2513 2514 static inline abi_long host_to_target_msqid_ds(abi_ulong target_addr, 2515 struct msqid_ds *host_md) 2516 { 2517 struct target_msqid_ds *target_md; 2518 2519 if (!lock_user_struct(VERIFY_WRITE, target_md, target_addr, 0)) 2520 return -TARGET_EFAULT; 2521 if (host_to_target_ipc_perm(target_addr,&(host_md->msg_perm))) 2522 return -TARGET_EFAULT; 2523 target_md->msg_stime = tswapl(host_md->msg_stime); 2524 target_md->msg_rtime = tswapl(host_md->msg_rtime); 2525 target_md->msg_ctime = tswapl(host_md->msg_ctime); 2526 target_md->__msg_cbytes = tswapl(host_md->__msg_cbytes); 2527 target_md->msg_qnum = tswapl(host_md->msg_qnum); 2528 target_md->msg_qbytes = tswapl(host_md->msg_qbytes); 2529 target_md->msg_lspid = tswapl(host_md->msg_lspid); 2530 target_md->msg_lrpid = tswapl(host_md->msg_lrpid); 2531 unlock_user_struct(target_md, target_addr, 1); 2532 return 0; 2533 } 2534 2535 struct target_msginfo { 2536 int msgpool; 2537 int msgmap; 2538 int msgmax; 2539 int msgmnb; 2540 int msgmni; 2541 int msgssz; 2542 int msgtql; 2543 unsigned short int msgseg; 2544 }; 2545 2546 static inline abi_long host_to_target_msginfo(abi_ulong target_addr, 2547 struct msginfo *host_msginfo) 2548 { 2549 struct target_msginfo *target_msginfo; 2550 if (!lock_user_struct(VERIFY_WRITE, target_msginfo, target_addr, 0)) 2551 return -TARGET_EFAULT; 2552 __put_user(host_msginfo->msgpool, &target_msginfo->msgpool); 2553 __put_user(host_msginfo->msgmap, &target_msginfo->msgmap); 2554 __put_user(host_msginfo->msgmax, &target_msginfo->msgmax); 2555 __put_user(host_msginfo->msgmnb, &target_msginfo->msgmnb); 2556 __put_user(host_msginfo->msgmni, &target_msginfo->msgmni); 2557 __put_user(host_msginfo->msgssz, &target_msginfo->msgssz); 2558 __put_user(host_msginfo->msgtql, &target_msginfo->msgtql); 2559 __put_user(host_msginfo->msgseg, &target_msginfo->msgseg); 2560 unlock_user_struct(target_msginfo, target_addr, 1); 2561 return 0; 2562 } 2563 2564 static inline abi_long do_msgctl(int msgid, int cmd, abi_long ptr) 2565 { 2566 struct msqid_ds dsarg; 2567 struct msginfo msginfo; 2568 abi_long ret = -TARGET_EINVAL; 2569 2570 cmd &= 0xff; 2571 2572 switch (cmd) { 2573 case IPC_STAT: 2574 case IPC_SET: 2575 case MSG_STAT: 2576 if (target_to_host_msqid_ds(&dsarg,ptr)) 2577 return -TARGET_EFAULT; 2578 ret = get_errno(msgctl(msgid, cmd, &dsarg)); 2579 if (host_to_target_msqid_ds(ptr,&dsarg)) 2580 return -TARGET_EFAULT; 2581 break; 2582 case IPC_RMID: 2583 ret = get_errno(msgctl(msgid, cmd, NULL)); 2584 break; 2585 case IPC_INFO: 2586 case MSG_INFO: 2587 ret = get_errno(msgctl(msgid, cmd, (struct msqid_ds *)&msginfo)); 2588 if (host_to_target_msginfo(ptr, &msginfo)) 2589 return -TARGET_EFAULT; 2590 break; 2591 } 2592 2593 return ret; 2594 } 2595 2596 struct target_msgbuf { 2597 abi_long mtype; 2598 char mtext[1]; 2599 }; 2600 2601 static inline abi_long do_msgsnd(int msqid, abi_long msgp, 2602 unsigned int msgsz, int msgflg) 2603 { 2604 struct target_msgbuf *target_mb; 2605 struct msgbuf *host_mb; 2606 abi_long ret = 0; 2607 2608 if (!lock_user_struct(VERIFY_READ, target_mb, msgp, 0)) 2609 return -TARGET_EFAULT; 2610 host_mb = malloc(msgsz+sizeof(long)); 2611 host_mb->mtype = (abi_long) tswapl(target_mb->mtype); 2612 memcpy(host_mb->mtext, target_mb->mtext, msgsz); 2613 ret = get_errno(msgsnd(msqid, host_mb, msgsz, msgflg)); 2614 free(host_mb); 2615 unlock_user_struct(target_mb, msgp, 0); 2616 2617 return ret; 2618 } 2619 2620 static inline abi_long do_msgrcv(int msqid, abi_long msgp, 2621 unsigned int msgsz, abi_long msgtyp, 2622 int msgflg) 2623 { 2624 struct target_msgbuf *target_mb; 2625 char *target_mtext; 2626 struct msgbuf *host_mb; 2627 abi_long ret = 0; 2628 2629 if (!lock_user_struct(VERIFY_WRITE, target_mb, msgp, 0)) 2630 return -TARGET_EFAULT; 2631 2632 host_mb = malloc(msgsz+sizeof(long)); 2633 ret = get_errno(msgrcv(msqid, host_mb, msgsz, tswapl(msgtyp), msgflg)); 2634 2635 if (ret > 0) { 2636 abi_ulong target_mtext_addr = msgp + sizeof(abi_ulong); 2637 target_mtext = lock_user(VERIFY_WRITE, target_mtext_addr, ret, 0); 2638 if (!target_mtext) { 2639 ret = -TARGET_EFAULT; 2640 goto end; 2641 } 2642 memcpy(target_mb->mtext, host_mb->mtext, ret); 2643 unlock_user(target_mtext, target_mtext_addr, ret); 2644 } 2645 2646 target_mb->mtype = tswapl(host_mb->mtype); 2647 free(host_mb); 2648 2649 end: 2650 if (target_mb) 2651 unlock_user_struct(target_mb, msgp, 1); 2652 return ret; 2653 } 2654 2655 struct target_shmid_ds 2656 { 2657 struct target_ipc_perm shm_perm; 2658 abi_ulong shm_segsz; 2659 abi_ulong shm_atime; 2660 #if TARGET_ABI_BITS == 32 2661 abi_ulong __unused1; 2662 #endif 2663 abi_ulong shm_dtime; 2664 #if TARGET_ABI_BITS == 32 2665 abi_ulong __unused2; 2666 #endif 2667 abi_ulong shm_ctime; 2668 #if TARGET_ABI_BITS == 32 2669 abi_ulong __unused3; 2670 #endif 2671 int shm_cpid; 2672 int shm_lpid; 2673 abi_ulong shm_nattch; 2674 unsigned long int __unused4; 2675 unsigned long int __unused5; 2676 }; 2677 2678 static inline abi_long target_to_host_shmid_ds(struct shmid_ds *host_sd, 2679 abi_ulong target_addr) 2680 { 2681 struct target_shmid_ds *target_sd; 2682 2683 if (!lock_user_struct(VERIFY_READ, target_sd, target_addr, 1)) 2684 return -TARGET_EFAULT; 2685 if (target_to_host_ipc_perm(&(host_sd->shm_perm), target_addr)) 2686 return -TARGET_EFAULT; 2687 __get_user(host_sd->shm_segsz, &target_sd->shm_segsz); 2688 __get_user(host_sd->shm_atime, &target_sd->shm_atime); 2689 __get_user(host_sd->shm_dtime, &target_sd->shm_dtime); 2690 __get_user(host_sd->shm_ctime, &target_sd->shm_ctime); 2691 __get_user(host_sd->shm_cpid, &target_sd->shm_cpid); 2692 __get_user(host_sd->shm_lpid, &target_sd->shm_lpid); 2693 __get_user(host_sd->shm_nattch, &target_sd->shm_nattch); 2694 unlock_user_struct(target_sd, target_addr, 0); 2695 return 0; 2696 } 2697 2698 static inline abi_long host_to_target_shmid_ds(abi_ulong target_addr, 2699 struct shmid_ds *host_sd) 2700 { 2701 struct target_shmid_ds *target_sd; 2702 2703 if (!lock_user_struct(VERIFY_WRITE, target_sd, target_addr, 0)) 2704 return -TARGET_EFAULT; 2705 if (host_to_target_ipc_perm(target_addr, &(host_sd->shm_perm))) 2706 return -TARGET_EFAULT; 2707 __put_user(host_sd->shm_segsz, &target_sd->shm_segsz); 2708 __put_user(host_sd->shm_atime, &target_sd->shm_atime); 2709 __put_user(host_sd->shm_dtime, &target_sd->shm_dtime); 2710 __put_user(host_sd->shm_ctime, &target_sd->shm_ctime); 2711 __put_user(host_sd->shm_cpid, &target_sd->shm_cpid); 2712 __put_user(host_sd->shm_lpid, &target_sd->shm_lpid); 2713 __put_user(host_sd->shm_nattch, &target_sd->shm_nattch); 2714 unlock_user_struct(target_sd, target_addr, 1); 2715 return 0; 2716 } 2717 2718 struct target_shminfo { 2719 abi_ulong shmmax; 2720 abi_ulong shmmin; 2721 abi_ulong shmmni; 2722 abi_ulong shmseg; 2723 abi_ulong shmall; 2724 }; 2725 2726 static inline abi_long host_to_target_shminfo(abi_ulong target_addr, 2727 struct shminfo *host_shminfo) 2728 { 2729 struct target_shminfo *target_shminfo; 2730 if (!lock_user_struct(VERIFY_WRITE, target_shminfo, target_addr, 0)) 2731 return -TARGET_EFAULT; 2732 __put_user(host_shminfo->shmmax, &target_shminfo->shmmax); 2733 __put_user(host_shminfo->shmmin, &target_shminfo->shmmin); 2734 __put_user(host_shminfo->shmmni, &target_shminfo->shmmni); 2735 __put_user(host_shminfo->shmseg, &target_shminfo->shmseg); 2736 __put_user(host_shminfo->shmall, &target_shminfo->shmall); 2737 unlock_user_struct(target_shminfo, target_addr, 1); 2738 return 0; 2739 } 2740 2741 struct target_shm_info { 2742 int used_ids; 2743 abi_ulong shm_tot; 2744 abi_ulong shm_rss; 2745 abi_ulong shm_swp; 2746 abi_ulong swap_attempts; 2747 abi_ulong swap_successes; 2748 }; 2749 2750 static inline abi_long host_to_target_shm_info(abi_ulong target_addr, 2751 struct shm_info *host_shm_info) 2752 { 2753 struct target_shm_info *target_shm_info; 2754 if (!lock_user_struct(VERIFY_WRITE, target_shm_info, target_addr, 0)) 2755 return -TARGET_EFAULT; 2756 __put_user(host_shm_info->used_ids, &target_shm_info->used_ids); 2757 __put_user(host_shm_info->shm_tot, &target_shm_info->shm_tot); 2758 __put_user(host_shm_info->shm_rss, &target_shm_info->shm_rss); 2759 __put_user(host_shm_info->shm_swp, &target_shm_info->shm_swp); 2760 __put_user(host_shm_info->swap_attempts, &target_shm_info->swap_attempts); 2761 __put_user(host_shm_info->swap_successes, &target_shm_info->swap_successes); 2762 unlock_user_struct(target_shm_info, target_addr, 1); 2763 return 0; 2764 } 2765 2766 static inline abi_long do_shmctl(int shmid, int cmd, abi_long buf) 2767 { 2768 struct shmid_ds dsarg; 2769 struct shminfo shminfo; 2770 struct shm_info shm_info; 2771 abi_long ret = -TARGET_EINVAL; 2772 2773 cmd &= 0xff; 2774 2775 switch(cmd) { 2776 case IPC_STAT: 2777 case IPC_SET: 2778 case SHM_STAT: 2779 if (target_to_host_shmid_ds(&dsarg, buf)) 2780 return -TARGET_EFAULT; 2781 ret = get_errno(shmctl(shmid, cmd, &dsarg)); 2782 if (host_to_target_shmid_ds(buf, &dsarg)) 2783 return -TARGET_EFAULT; 2784 break; 2785 case IPC_INFO: 2786 ret = get_errno(shmctl(shmid, cmd, (struct shmid_ds *)&shminfo)); 2787 if (host_to_target_shminfo(buf, &shminfo)) 2788 return -TARGET_EFAULT; 2789 break; 2790 case SHM_INFO: 2791 ret = get_errno(shmctl(shmid, cmd, (struct shmid_ds *)&shm_info)); 2792 if (host_to_target_shm_info(buf, &shm_info)) 2793 return -TARGET_EFAULT; 2794 break; 2795 case IPC_RMID: 2796 case SHM_LOCK: 2797 case SHM_UNLOCK: 2798 ret = get_errno(shmctl(shmid, cmd, NULL)); 2799 break; 2800 } 2801 2802 return ret; 2803 } 2804 2805 static inline abi_ulong do_shmat(int shmid, abi_ulong shmaddr, int shmflg) 2806 { 2807 abi_long raddr; 2808 void *host_raddr; 2809 struct shmid_ds shm_info; 2810 int i,ret; 2811 2812 /* find out the length of the shared memory segment */ 2813 ret = get_errno(shmctl(shmid, IPC_STAT, &shm_info)); 2814 if (is_error(ret)) { 2815 /* can't get length, bail out */ 2816 return ret; 2817 } 2818 2819 mmap_lock(); 2820 2821 if (shmaddr) 2822 host_raddr = shmat(shmid, (void *)g2h(shmaddr), shmflg); 2823 else { 2824 abi_ulong mmap_start; 2825 2826 mmap_start = mmap_find_vma(0, shm_info.shm_segsz); 2827 2828 if (mmap_start == -1) { 2829 errno = ENOMEM; 2830 host_raddr = (void *)-1; 2831 } else 2832 host_raddr = shmat(shmid, g2h(mmap_start), shmflg | SHM_REMAP); 2833 } 2834 2835 if (host_raddr == (void *)-1) { 2836 mmap_unlock(); 2837 return get_errno((long)host_raddr); 2838 } 2839 raddr=h2g((unsigned long)host_raddr); 2840 2841 page_set_flags(raddr, raddr + shm_info.shm_segsz, 2842 PAGE_VALID | PAGE_READ | 2843 ((shmflg & SHM_RDONLY)? 0 : PAGE_WRITE)); 2844 2845 for (i = 0; i < N_SHM_REGIONS; i++) { 2846 if (shm_regions[i].start == 0) { 2847 shm_regions[i].start = raddr; 2848 shm_regions[i].size = shm_info.shm_segsz; 2849 break; 2850 } 2851 } 2852 2853 mmap_unlock(); 2854 return raddr; 2855 2856 } 2857 2858 static inline abi_long do_shmdt(abi_ulong shmaddr) 2859 { 2860 int i; 2861 2862 for (i = 0; i < N_SHM_REGIONS; ++i) { 2863 if (shm_regions[i].start == shmaddr) { 2864 shm_regions[i].start = 0; 2865 page_set_flags(shmaddr, shmaddr + shm_regions[i].size, 0); 2866 break; 2867 } 2868 } 2869 2870 return get_errno(shmdt(g2h(shmaddr))); 2871 } 2872 2873 #ifdef TARGET_NR_ipc 2874 /* ??? This only works with linear mappings. */ 2875 /* do_ipc() must return target values and target errnos. */ 2876 static abi_long do_ipc(unsigned int call, int first, 2877 int second, int third, 2878 abi_long ptr, abi_long fifth) 2879 { 2880 int version; 2881 abi_long ret = 0; 2882 2883 version = call >> 16; 2884 call &= 0xffff; 2885 2886 switch (call) { 2887 case IPCOP_semop: 2888 ret = do_semop(first, ptr, second); 2889 break; 2890 2891 case IPCOP_semget: 2892 ret = get_errno(semget(first, second, third)); 2893 break; 2894 2895 case IPCOP_semctl: 2896 ret = do_semctl(first, second, third, (union target_semun)(abi_ulong) ptr); 2897 break; 2898 2899 case IPCOP_msgget: 2900 ret = get_errno(msgget(first, second)); 2901 break; 2902 2903 case IPCOP_msgsnd: 2904 ret = do_msgsnd(first, ptr, second, third); 2905 break; 2906 2907 case IPCOP_msgctl: 2908 ret = do_msgctl(first, second, ptr); 2909 break; 2910 2911 case IPCOP_msgrcv: 2912 switch (version) { 2913 case 0: 2914 { 2915 struct target_ipc_kludge { 2916 abi_long msgp; 2917 abi_long msgtyp; 2918 } *tmp; 2919 2920 if (!lock_user_struct(VERIFY_READ, tmp, ptr, 1)) { 2921 ret = -TARGET_EFAULT; 2922 break; 2923 } 2924 2925 ret = do_msgrcv(first, tmp->msgp, second, tmp->msgtyp, third); 2926 2927 unlock_user_struct(tmp, ptr, 0); 2928 break; 2929 } 2930 default: 2931 ret = do_msgrcv(first, ptr, second, fifth, third); 2932 } 2933 break; 2934 2935 case IPCOP_shmat: 2936 switch (version) { 2937 default: 2938 { 2939 abi_ulong raddr; 2940 raddr = do_shmat(first, ptr, second); 2941 if (is_error(raddr)) 2942 return get_errno(raddr); 2943 if (put_user_ual(raddr, third)) 2944 return -TARGET_EFAULT; 2945 break; 2946 } 2947 case 1: 2948 ret = -TARGET_EINVAL; 2949 break; 2950 } 2951 break; 2952 case IPCOP_shmdt: 2953 ret = do_shmdt(ptr); 2954 break; 2955 2956 case IPCOP_shmget: 2957 /* IPC_* flag values are the same on all linux platforms */ 2958 ret = get_errno(shmget(first, second, third)); 2959 break; 2960 2961 /* IPC_* and SHM_* command values are the same on all linux platforms */ 2962 case IPCOP_shmctl: 2963 ret = do_shmctl(first, second, third); 2964 break; 2965 default: 2966 gemu_log("Unsupported ipc call: %d (version %d)\n", call, version); 2967 ret = -TARGET_ENOSYS; 2968 break; 2969 } 2970 return ret; 2971 } 2972 #endif 2973 2974 /* kernel structure types definitions */ 2975 2976 #define STRUCT(name, ...) STRUCT_ ## name, 2977 #define STRUCT_SPECIAL(name) STRUCT_ ## name, 2978 enum { 2979 #include "syscall_types.h" 2980 }; 2981 #undef STRUCT 2982 #undef STRUCT_SPECIAL 2983 2984 #define STRUCT(name, ...) static const argtype struct_ ## name ## _def[] = { __VA_ARGS__, TYPE_NULL }; 2985 #define STRUCT_SPECIAL(name) 2986 #include "syscall_types.h" 2987 #undef STRUCT 2988 #undef STRUCT_SPECIAL 2989 2990 typedef struct IOCTLEntry IOCTLEntry; 2991 2992 typedef abi_long do_ioctl_fn(const IOCTLEntry *ie, uint8_t *buf_temp, 2993 int fd, abi_long cmd, abi_long arg); 2994 2995 struct IOCTLEntry { 2996 unsigned int target_cmd; 2997 unsigned int host_cmd; 2998 const char *name; 2999 int access; 3000 do_ioctl_fn *do_ioctl; 3001 const argtype arg_type[5]; 3002 }; 3003 3004 #define IOC_R 0x0001 3005 #define IOC_W 0x0002 3006 #define IOC_RW (IOC_R | IOC_W) 3007 3008 #define MAX_STRUCT_SIZE 4096 3009 3010 #ifdef CONFIG_FIEMAP 3011 /* So fiemap access checks don't overflow on 32 bit systems. 3012 * This is very slightly smaller than the limit imposed by 3013 * the underlying kernel. 3014 */ 3015 #define FIEMAP_MAX_EXTENTS ((UINT_MAX - sizeof(struct fiemap)) \ 3016 / sizeof(struct fiemap_extent)) 3017 3018 static abi_long do_ioctl_fs_ioc_fiemap(const IOCTLEntry *ie, uint8_t *buf_temp, 3019 int fd, abi_long cmd, abi_long arg) 3020 { 3021 /* The parameter for this ioctl is a struct fiemap followed 3022 * by an array of struct fiemap_extent whose size is set 3023 * in fiemap->fm_extent_count. The array is filled in by the 3024 * ioctl. 3025 */ 3026 int target_size_in, target_size_out; 3027 struct fiemap *fm; 3028 const argtype *arg_type = ie->arg_type; 3029 const argtype extent_arg_type[] = { MK_STRUCT(STRUCT_fiemap_extent) }; 3030 void *argptr, *p; 3031 abi_long ret; 3032 int i, extent_size = thunk_type_size(extent_arg_type, 0); 3033 uint32_t outbufsz; 3034 int free_fm = 0; 3035 3036 assert(arg_type[0] == TYPE_PTR); 3037 assert(ie->access == IOC_RW); 3038 arg_type++; 3039 target_size_in = thunk_type_size(arg_type, 0); 3040 argptr = lock_user(VERIFY_READ, arg, target_size_in, 1); 3041 if (!argptr) { 3042 return -TARGET_EFAULT; 3043 } 3044 thunk_convert(buf_temp, argptr, arg_type, THUNK_HOST); 3045 unlock_user(argptr, arg, 0); 3046 fm = (struct fiemap *)buf_temp; 3047 if (fm->fm_extent_count > FIEMAP_MAX_EXTENTS) { 3048 return -TARGET_EINVAL; 3049 } 3050 3051 outbufsz = sizeof (*fm) + 3052 (sizeof(struct fiemap_extent) * fm->fm_extent_count); 3053 3054 if (outbufsz > MAX_STRUCT_SIZE) { 3055 /* We can't fit all the extents into the fixed size buffer. 3056 * Allocate one that is large enough and use it instead. 3057 */ 3058 fm = malloc(outbufsz); 3059 if (!fm) { 3060 return -TARGET_ENOMEM; 3061 } 3062 memcpy(fm, buf_temp, sizeof(struct fiemap)); 3063 free_fm = 1; 3064 } 3065 ret = get_errno(ioctl(fd, ie->host_cmd, fm)); 3066 if (!is_error(ret)) { 3067 target_size_out = target_size_in; 3068 /* An extent_count of 0 means we were only counting the extents 3069 * so there are no structs to copy 3070 */ 3071 if (fm->fm_extent_count != 0) { 3072 target_size_out += fm->fm_mapped_extents * extent_size; 3073 } 3074 argptr = lock_user(VERIFY_WRITE, arg, target_size_out, 0); 3075 if (!argptr) { 3076 ret = -TARGET_EFAULT; 3077 } else { 3078 /* Convert the struct fiemap */ 3079 thunk_convert(argptr, fm, arg_type, THUNK_TARGET); 3080 if (fm->fm_extent_count != 0) { 3081 p = argptr + target_size_in; 3082 /* ...and then all the struct fiemap_extents */ 3083 for (i = 0; i < fm->fm_mapped_extents; i++) { 3084 thunk_convert(p, &fm->fm_extents[i], extent_arg_type, 3085 THUNK_TARGET); 3086 p += extent_size; 3087 } 3088 } 3089 unlock_user(argptr, arg, target_size_out); 3090 } 3091 } 3092 if (free_fm) { 3093 free(fm); 3094 } 3095 return ret; 3096 } 3097 #endif 3098 3099 static abi_long do_ioctl_ifconf(const IOCTLEntry *ie, uint8_t *buf_temp, 3100 int fd, abi_long cmd, abi_long arg) 3101 { 3102 const argtype *arg_type = ie->arg_type; 3103 int target_size; 3104 void *argptr; 3105 int ret; 3106 struct ifconf *host_ifconf; 3107 uint32_t outbufsz; 3108 const argtype ifreq_arg_type[] = { MK_STRUCT(STRUCT_sockaddr_ifreq) }; 3109 int target_ifreq_size; 3110 int nb_ifreq; 3111 int free_buf = 0; 3112 int i; 3113 int target_ifc_len; 3114 abi_long target_ifc_buf; 3115 int host_ifc_len; 3116 char *host_ifc_buf; 3117 3118 assert(arg_type[0] == TYPE_PTR); 3119 assert(ie->access == IOC_RW); 3120 3121 arg_type++; 3122 target_size = thunk_type_size(arg_type, 0); 3123 3124 argptr = lock_user(VERIFY_READ, arg, target_size, 1); 3125 if (!argptr) 3126 return -TARGET_EFAULT; 3127 thunk_convert(buf_temp, argptr, arg_type, THUNK_HOST); 3128 unlock_user(argptr, arg, 0); 3129 3130 host_ifconf = (struct ifconf *)(unsigned long)buf_temp; 3131 target_ifc_len = host_ifconf->ifc_len; 3132 target_ifc_buf = (abi_long)(unsigned long)host_ifconf->ifc_buf; 3133 3134 target_ifreq_size = thunk_type_size(ifreq_arg_type, 0); 3135 nb_ifreq = target_ifc_len / target_ifreq_size; 3136 host_ifc_len = nb_ifreq * sizeof(struct ifreq); 3137 3138 outbufsz = sizeof(*host_ifconf) + host_ifc_len; 3139 if (outbufsz > MAX_STRUCT_SIZE) { 3140 /* We can't fit all the extents into the fixed size buffer. 3141 * Allocate one that is large enough and use it instead. 3142 */ 3143 host_ifconf = malloc(outbufsz); 3144 if (!host_ifconf) { 3145 return -TARGET_ENOMEM; 3146 } 3147 memcpy(host_ifconf, buf_temp, sizeof(*host_ifconf)); 3148 free_buf = 1; 3149 } 3150 host_ifc_buf = (char*)host_ifconf + sizeof(*host_ifconf); 3151 3152 host_ifconf->ifc_len = host_ifc_len; 3153 host_ifconf->ifc_buf = host_ifc_buf; 3154 3155 ret = get_errno(ioctl(fd, ie->host_cmd, host_ifconf)); 3156 if (!is_error(ret)) { 3157 /* convert host ifc_len to target ifc_len */ 3158 3159 nb_ifreq = host_ifconf->ifc_len / sizeof(struct ifreq); 3160 target_ifc_len = nb_ifreq * target_ifreq_size; 3161 host_ifconf->ifc_len = target_ifc_len; 3162 3163 /* restore target ifc_buf */ 3164 3165 host_ifconf->ifc_buf = (char *)(unsigned long)target_ifc_buf; 3166 3167 /* copy struct ifconf to target user */ 3168 3169 argptr = lock_user(VERIFY_WRITE, arg, target_size, 0); 3170 if (!argptr) 3171 return -TARGET_EFAULT; 3172 thunk_convert(argptr, host_ifconf, arg_type, THUNK_TARGET); 3173 unlock_user(argptr, arg, target_size); 3174 3175 /* copy ifreq[] to target user */ 3176 3177 argptr = lock_user(VERIFY_WRITE, target_ifc_buf, target_ifc_len, 0); 3178 for (i = 0; i < nb_ifreq ; i++) { 3179 thunk_convert(argptr + i * target_ifreq_size, 3180 host_ifc_buf + i * sizeof(struct ifreq), 3181 ifreq_arg_type, THUNK_TARGET); 3182 } 3183 unlock_user(argptr, target_ifc_buf, target_ifc_len); 3184 } 3185 3186 if (free_buf) { 3187 free(host_ifconf); 3188 } 3189 3190 return ret; 3191 } 3192 3193 static IOCTLEntry ioctl_entries[] = { 3194 #define IOCTL(cmd, access, ...) \ 3195 { TARGET_ ## cmd, cmd, #cmd, access, 0, { __VA_ARGS__ } }, 3196 #define IOCTL_SPECIAL(cmd, access, dofn, ...) \ 3197 { TARGET_ ## cmd, cmd, #cmd, access, dofn, { __VA_ARGS__ } }, 3198 #include "ioctls.h" 3199 { 0, 0, }, 3200 }; 3201 3202 /* ??? Implement proper locking for ioctls. */ 3203 /* do_ioctl() Must return target values and target errnos. */ 3204 static abi_long do_ioctl(int fd, abi_long cmd, abi_long arg) 3205 { 3206 const IOCTLEntry *ie; 3207 const argtype *arg_type; 3208 abi_long ret; 3209 uint8_t buf_temp[MAX_STRUCT_SIZE]; 3210 int target_size; 3211 void *argptr; 3212 3213 ie = ioctl_entries; 3214 for(;;) { 3215 if (ie->target_cmd == 0) { 3216 gemu_log("Unsupported ioctl: cmd=0x%04lx\n", (long)cmd); 3217 return -TARGET_ENOSYS; 3218 } 3219 if (ie->target_cmd == cmd) 3220 break; 3221 ie++; 3222 } 3223 arg_type = ie->arg_type; 3224 #if defined(DEBUG) 3225 gemu_log("ioctl: cmd=0x%04lx (%s)\n", (long)cmd, ie->name); 3226 #endif 3227 if (ie->do_ioctl) { 3228 return ie->do_ioctl(ie, buf_temp, fd, cmd, arg); 3229 } 3230 3231 switch(arg_type[0]) { 3232 case TYPE_NULL: 3233 /* no argument */ 3234 ret = get_errno(ioctl(fd, ie->host_cmd)); 3235 break; 3236 case TYPE_PTRVOID: 3237 case TYPE_INT: 3238 /* int argment */ 3239 ret = get_errno(ioctl(fd, ie->host_cmd, arg)); 3240 break; 3241 case TYPE_PTR: 3242 arg_type++; 3243 target_size = thunk_type_size(arg_type, 0); 3244 switch(ie->access) { 3245 case IOC_R: 3246 ret = get_errno(ioctl(fd, ie->host_cmd, buf_temp)); 3247 if (!is_error(ret)) { 3248 argptr = lock_user(VERIFY_WRITE, arg, target_size, 0); 3249 if (!argptr) 3250 return -TARGET_EFAULT; 3251 thunk_convert(argptr, buf_temp, arg_type, THUNK_TARGET); 3252 unlock_user(argptr, arg, target_size); 3253 } 3254 break; 3255 case IOC_W: 3256 argptr = lock_user(VERIFY_READ, arg, target_size, 1); 3257 if (!argptr) 3258 return -TARGET_EFAULT; 3259 thunk_convert(buf_temp, argptr, arg_type, THUNK_HOST); 3260 unlock_user(argptr, arg, 0); 3261 ret = get_errno(ioctl(fd, ie->host_cmd, buf_temp)); 3262 break; 3263 default: 3264 case IOC_RW: 3265 argptr = lock_user(VERIFY_READ, arg, target_size, 1); 3266 if (!argptr) 3267 return -TARGET_EFAULT; 3268 thunk_convert(buf_temp, argptr, arg_type, THUNK_HOST); 3269 unlock_user(argptr, arg, 0); 3270 ret = get_errno(ioctl(fd, ie->host_cmd, buf_temp)); 3271 if (!is_error(ret)) { 3272 argptr = lock_user(VERIFY_WRITE, arg, target_size, 0); 3273 if (!argptr) 3274 return -TARGET_EFAULT; 3275 thunk_convert(argptr, buf_temp, arg_type, THUNK_TARGET); 3276 unlock_user(argptr, arg, target_size); 3277 } 3278 break; 3279 } 3280 break; 3281 default: 3282 gemu_log("Unsupported ioctl type: cmd=0x%04lx type=%d\n", 3283 (long)cmd, arg_type[0]); 3284 ret = -TARGET_ENOSYS; 3285 break; 3286 } 3287 return ret; 3288 } 3289 3290 static const bitmask_transtbl iflag_tbl[] = { 3291 { TARGET_IGNBRK, TARGET_IGNBRK, IGNBRK, IGNBRK }, 3292 { TARGET_BRKINT, TARGET_BRKINT, BRKINT, BRKINT }, 3293 { TARGET_IGNPAR, TARGET_IGNPAR, IGNPAR, IGNPAR }, 3294 { TARGET_PARMRK, TARGET_PARMRK, PARMRK, PARMRK }, 3295 { TARGET_INPCK, TARGET_INPCK, INPCK, INPCK }, 3296 { TARGET_ISTRIP, TARGET_ISTRIP, ISTRIP, ISTRIP }, 3297 { TARGET_INLCR, TARGET_INLCR, INLCR, INLCR }, 3298 { TARGET_IGNCR, TARGET_IGNCR, IGNCR, IGNCR }, 3299 { TARGET_ICRNL, TARGET_ICRNL, ICRNL, ICRNL }, 3300 { TARGET_IUCLC, TARGET_IUCLC, IUCLC, IUCLC }, 3301 { TARGET_IXON, TARGET_IXON, IXON, IXON }, 3302 { TARGET_IXANY, TARGET_IXANY, IXANY, IXANY }, 3303 { TARGET_IXOFF, TARGET_IXOFF, IXOFF, IXOFF }, 3304 { TARGET_IMAXBEL, TARGET_IMAXBEL, IMAXBEL, IMAXBEL }, 3305 { 0, 0, 0, 0 } 3306 }; 3307 3308 static const bitmask_transtbl oflag_tbl[] = { 3309 { TARGET_OPOST, TARGET_OPOST, OPOST, OPOST }, 3310 { TARGET_OLCUC, TARGET_OLCUC, OLCUC, OLCUC }, 3311 { TARGET_ONLCR, TARGET_ONLCR, ONLCR, ONLCR }, 3312 { TARGET_OCRNL, TARGET_OCRNL, OCRNL, OCRNL }, 3313 { TARGET_ONOCR, TARGET_ONOCR, ONOCR, ONOCR }, 3314 { TARGET_ONLRET, TARGET_ONLRET, ONLRET, ONLRET }, 3315 { TARGET_OFILL, TARGET_OFILL, OFILL, OFILL }, 3316 { TARGET_OFDEL, TARGET_OFDEL, OFDEL, OFDEL }, 3317 { TARGET_NLDLY, TARGET_NL0, NLDLY, NL0 }, 3318 { TARGET_NLDLY, TARGET_NL1, NLDLY, NL1 }, 3319 { TARGET_CRDLY, TARGET_CR0, CRDLY, CR0 }, 3320 { TARGET_CRDLY, TARGET_CR1, CRDLY, CR1 }, 3321 { TARGET_CRDLY, TARGET_CR2, CRDLY, CR2 }, 3322 { TARGET_CRDLY, TARGET_CR3, CRDLY, CR3 }, 3323 { TARGET_TABDLY, TARGET_TAB0, TABDLY, TAB0 }, 3324 { TARGET_TABDLY, TARGET_TAB1, TABDLY, TAB1 }, 3325 { TARGET_TABDLY, TARGET_TAB2, TABDLY, TAB2 }, 3326 { TARGET_TABDLY, TARGET_TAB3, TABDLY, TAB3 }, 3327 { TARGET_BSDLY, TARGET_BS0, BSDLY, BS0 }, 3328 { TARGET_BSDLY, TARGET_BS1, BSDLY, BS1 }, 3329 { TARGET_VTDLY, TARGET_VT0, VTDLY, VT0 }, 3330 { TARGET_VTDLY, TARGET_VT1, VTDLY, VT1 }, 3331 { TARGET_FFDLY, TARGET_FF0, FFDLY, FF0 }, 3332 { TARGET_FFDLY, TARGET_FF1, FFDLY, FF1 }, 3333 { 0, 0, 0, 0 } 3334 }; 3335 3336 static const bitmask_transtbl cflag_tbl[] = { 3337 { TARGET_CBAUD, TARGET_B0, CBAUD, B0 }, 3338 { TARGET_CBAUD, TARGET_B50, CBAUD, B50 }, 3339 { TARGET_CBAUD, TARGET_B75, CBAUD, B75 }, 3340 { TARGET_CBAUD, TARGET_B110, CBAUD, B110 }, 3341 { TARGET_CBAUD, TARGET_B134, CBAUD, B134 }, 3342 { TARGET_CBAUD, TARGET_B150, CBAUD, B150 }, 3343 { TARGET_CBAUD, TARGET_B200, CBAUD, B200 }, 3344 { TARGET_CBAUD, TARGET_B300, CBAUD, B300 }, 3345 { TARGET_CBAUD, TARGET_B600, CBAUD, B600 }, 3346 { TARGET_CBAUD, TARGET_B1200, CBAUD, B1200 }, 3347 { TARGET_CBAUD, TARGET_B1800, CBAUD, B1800 }, 3348 { TARGET_CBAUD, TARGET_B2400, CBAUD, B2400 }, 3349 { TARGET_CBAUD, TARGET_B4800, CBAUD, B4800 }, 3350 { TARGET_CBAUD, TARGET_B9600, CBAUD, B9600 }, 3351 { TARGET_CBAUD, TARGET_B19200, CBAUD, B19200 }, 3352 { TARGET_CBAUD, TARGET_B38400, CBAUD, B38400 }, 3353 { TARGET_CBAUD, TARGET_B57600, CBAUD, B57600 }, 3354 { TARGET_CBAUD, TARGET_B115200, CBAUD, B115200 }, 3355 { TARGET_CBAUD, TARGET_B230400, CBAUD, B230400 }, 3356 { TARGET_CBAUD, TARGET_B460800, CBAUD, B460800 }, 3357 { TARGET_CSIZE, TARGET_CS5, CSIZE, CS5 }, 3358 { TARGET_CSIZE, TARGET_CS6, CSIZE, CS6 }, 3359 { TARGET_CSIZE, TARGET_CS7, CSIZE, CS7 }, 3360 { TARGET_CSIZE, TARGET_CS8, CSIZE, CS8 }, 3361 { TARGET_CSTOPB, TARGET_CSTOPB, CSTOPB, CSTOPB }, 3362 { TARGET_CREAD, TARGET_CREAD, CREAD, CREAD }, 3363 { TARGET_PARENB, TARGET_PARENB, PARENB, PARENB }, 3364 { TARGET_PARODD, TARGET_PARODD, PARODD, PARODD }, 3365 { TARGET_HUPCL, TARGET_HUPCL, HUPCL, HUPCL }, 3366 { TARGET_CLOCAL, TARGET_CLOCAL, CLOCAL, CLOCAL }, 3367 { TARGET_CRTSCTS, TARGET_CRTSCTS, CRTSCTS, CRTSCTS }, 3368 { 0, 0, 0, 0 } 3369 }; 3370 3371 static const bitmask_transtbl lflag_tbl[] = { 3372 { TARGET_ISIG, TARGET_ISIG, ISIG, ISIG }, 3373 { TARGET_ICANON, TARGET_ICANON, ICANON, ICANON }, 3374 { TARGET_XCASE, TARGET_XCASE, XCASE, XCASE }, 3375 { TARGET_ECHO, TARGET_ECHO, ECHO, ECHO }, 3376 { TARGET_ECHOE, TARGET_ECHOE, ECHOE, ECHOE }, 3377 { TARGET_ECHOK, TARGET_ECHOK, ECHOK, ECHOK }, 3378 { TARGET_ECHONL, TARGET_ECHONL, ECHONL, ECHONL }, 3379 { TARGET_NOFLSH, TARGET_NOFLSH, NOFLSH, NOFLSH }, 3380 { TARGET_TOSTOP, TARGET_TOSTOP, TOSTOP, TOSTOP }, 3381 { TARGET_ECHOCTL, TARGET_ECHOCTL, ECHOCTL, ECHOCTL }, 3382 { TARGET_ECHOPRT, TARGET_ECHOPRT, ECHOPRT, ECHOPRT }, 3383 { TARGET_ECHOKE, TARGET_ECHOKE, ECHOKE, ECHOKE }, 3384 { TARGET_FLUSHO, TARGET_FLUSHO, FLUSHO, FLUSHO }, 3385 { TARGET_PENDIN, TARGET_PENDIN, PENDIN, PENDIN }, 3386 { TARGET_IEXTEN, TARGET_IEXTEN, IEXTEN, IEXTEN }, 3387 { 0, 0, 0, 0 } 3388 }; 3389 3390 static void target_to_host_termios (void *dst, const void *src) 3391 { 3392 struct host_termios *host = dst; 3393 const struct target_termios *target = src; 3394 3395 host->c_iflag = 3396 target_to_host_bitmask(tswap32(target->c_iflag), iflag_tbl); 3397 host->c_oflag = 3398 target_to_host_bitmask(tswap32(target->c_oflag), oflag_tbl); 3399 host->c_cflag = 3400 target_to_host_bitmask(tswap32(target->c_cflag), cflag_tbl); 3401 host->c_lflag = 3402 target_to_host_bitmask(tswap32(target->c_lflag), lflag_tbl); 3403 host->c_line = target->c_line; 3404 3405 memset(host->c_cc, 0, sizeof(host->c_cc)); 3406 host->c_cc[VINTR] = target->c_cc[TARGET_VINTR]; 3407 host->c_cc[VQUIT] = target->c_cc[TARGET_VQUIT]; 3408 host->c_cc[VERASE] = target->c_cc[TARGET_VERASE]; 3409 host->c_cc[VKILL] = target->c_cc[TARGET_VKILL]; 3410 host->c_cc[VEOF] = target->c_cc[TARGET_VEOF]; 3411 host->c_cc[VTIME] = target->c_cc[TARGET_VTIME]; 3412 host->c_cc[VMIN] = target->c_cc[TARGET_VMIN]; 3413 host->c_cc[VSWTC] = target->c_cc[TARGET_VSWTC]; 3414 host->c_cc[VSTART] = target->c_cc[TARGET_VSTART]; 3415 host->c_cc[VSTOP] = target->c_cc[TARGET_VSTOP]; 3416 host->c_cc[VSUSP] = target->c_cc[TARGET_VSUSP]; 3417 host->c_cc[VEOL] = target->c_cc[TARGET_VEOL]; 3418 host->c_cc[VREPRINT] = target->c_cc[TARGET_VREPRINT]; 3419 host->c_cc[VDISCARD] = target->c_cc[TARGET_VDISCARD]; 3420 host->c_cc[VWERASE] = target->c_cc[TARGET_VWERASE]; 3421 host->c_cc[VLNEXT] = target->c_cc[TARGET_VLNEXT]; 3422 host->c_cc[VEOL2] = target->c_cc[TARGET_VEOL2]; 3423 } 3424 3425 static void host_to_target_termios (void *dst, const void *src) 3426 { 3427 struct target_termios *target = dst; 3428 const struct host_termios *host = src; 3429 3430 target->c_iflag = 3431 tswap32(host_to_target_bitmask(host->c_iflag, iflag_tbl)); 3432 target->c_oflag = 3433 tswap32(host_to_target_bitmask(host->c_oflag, oflag_tbl)); 3434 target->c_cflag = 3435 tswap32(host_to_target_bitmask(host->c_cflag, cflag_tbl)); 3436 target->c_lflag = 3437 tswap32(host_to_target_bitmask(host->c_lflag, lflag_tbl)); 3438 target->c_line = host->c_line; 3439 3440 memset(target->c_cc, 0, sizeof(target->c_cc)); 3441 target->c_cc[TARGET_VINTR] = host->c_cc[VINTR]; 3442 target->c_cc[TARGET_VQUIT] = host->c_cc[VQUIT]; 3443 target->c_cc[TARGET_VERASE] = host->c_cc[VERASE]; 3444 target->c_cc[TARGET_VKILL] = host->c_cc[VKILL]; 3445 target->c_cc[TARGET_VEOF] = host->c_cc[VEOF]; 3446 target->c_cc[TARGET_VTIME] = host->c_cc[VTIME]; 3447 target->c_cc[TARGET_VMIN] = host->c_cc[VMIN]; 3448 target->c_cc[TARGET_VSWTC] = host->c_cc[VSWTC]; 3449 target->c_cc[TARGET_VSTART] = host->c_cc[VSTART]; 3450 target->c_cc[TARGET_VSTOP] = host->c_cc[VSTOP]; 3451 target->c_cc[TARGET_VSUSP] = host->c_cc[VSUSP]; 3452 target->c_cc[TARGET_VEOL] = host->c_cc[VEOL]; 3453 target->c_cc[TARGET_VREPRINT] = host->c_cc[VREPRINT]; 3454 target->c_cc[TARGET_VDISCARD] = host->c_cc[VDISCARD]; 3455 target->c_cc[TARGET_VWERASE] = host->c_cc[VWERASE]; 3456 target->c_cc[TARGET_VLNEXT] = host->c_cc[VLNEXT]; 3457 target->c_cc[TARGET_VEOL2] = host->c_cc[VEOL2]; 3458 } 3459 3460 static const StructEntry struct_termios_def = { 3461 .convert = { host_to_target_termios, target_to_host_termios }, 3462 .size = { sizeof(struct target_termios), sizeof(struct host_termios) }, 3463 .align = { __alignof__(struct target_termios), __alignof__(struct host_termios) }, 3464 }; 3465 3466 static bitmask_transtbl mmap_flags_tbl[] = { 3467 { TARGET_MAP_SHARED, TARGET_MAP_SHARED, MAP_SHARED, MAP_SHARED }, 3468 { TARGET_MAP_PRIVATE, TARGET_MAP_PRIVATE, MAP_PRIVATE, MAP_PRIVATE }, 3469 { TARGET_MAP_FIXED, TARGET_MAP_FIXED, MAP_FIXED, MAP_FIXED }, 3470 { TARGET_MAP_ANONYMOUS, TARGET_MAP_ANONYMOUS, MAP_ANONYMOUS, MAP_ANONYMOUS }, 3471 { TARGET_MAP_GROWSDOWN, TARGET_MAP_GROWSDOWN, MAP_GROWSDOWN, MAP_GROWSDOWN }, 3472 { TARGET_MAP_DENYWRITE, TARGET_MAP_DENYWRITE, MAP_DENYWRITE, MAP_DENYWRITE }, 3473 { TARGET_MAP_EXECUTABLE, TARGET_MAP_EXECUTABLE, MAP_EXECUTABLE, MAP_EXECUTABLE }, 3474 { TARGET_MAP_LOCKED, TARGET_MAP_LOCKED, MAP_LOCKED, MAP_LOCKED }, 3475 { 0, 0, 0, 0 } 3476 }; 3477 3478 #if defined(TARGET_I386) 3479 3480 /* NOTE: there is really one LDT for all the threads */ 3481 static uint8_t *ldt_table; 3482 3483 static abi_long read_ldt(abi_ulong ptr, unsigned long bytecount) 3484 { 3485 int size; 3486 void *p; 3487 3488 if (!ldt_table) 3489 return 0; 3490 size = TARGET_LDT_ENTRIES * TARGET_LDT_ENTRY_SIZE; 3491 if (size > bytecount) 3492 size = bytecount; 3493 p = lock_user(VERIFY_WRITE, ptr, size, 0); 3494 if (!p) 3495 return -TARGET_EFAULT; 3496 /* ??? Should this by byteswapped? */ 3497 memcpy(p, ldt_table, size); 3498 unlock_user(p, ptr, size); 3499 return size; 3500 } 3501 3502 /* XXX: add locking support */ 3503 static abi_long write_ldt(CPUX86State *env, 3504 abi_ulong ptr, unsigned long bytecount, int oldmode) 3505 { 3506 struct target_modify_ldt_ldt_s ldt_info; 3507 struct target_modify_ldt_ldt_s *target_ldt_info; 3508 int seg_32bit, contents, read_exec_only, limit_in_pages; 3509 int seg_not_present, useable, lm; 3510 uint32_t *lp, entry_1, entry_2; 3511 3512 if (bytecount != sizeof(ldt_info)) 3513 return -TARGET_EINVAL; 3514 if (!lock_user_struct(VERIFY_READ, target_ldt_info, ptr, 1)) 3515 return -TARGET_EFAULT; 3516 ldt_info.entry_number = tswap32(target_ldt_info->entry_number); 3517 ldt_info.base_addr = tswapl(target_ldt_info->base_addr); 3518 ldt_info.limit = tswap32(target_ldt_info->limit); 3519 ldt_info.flags = tswap32(target_ldt_info->flags); 3520 unlock_user_struct(target_ldt_info, ptr, 0); 3521 3522 if (ldt_info.entry_number >= TARGET_LDT_ENTRIES) 3523 return -TARGET_EINVAL; 3524 seg_32bit = ldt_info.flags & 1; 3525 contents = (ldt_info.flags >> 1) & 3; 3526 read_exec_only = (ldt_info.flags >> 3) & 1; 3527 limit_in_pages = (ldt_info.flags >> 4) & 1; 3528 seg_not_present = (ldt_info.flags >> 5) & 1; 3529 useable = (ldt_info.flags >> 6) & 1; 3530 #ifdef TARGET_ABI32 3531 lm = 0; 3532 #else 3533 lm = (ldt_info.flags >> 7) & 1; 3534 #endif 3535 if (contents == 3) { 3536 if (oldmode) 3537 return -TARGET_EINVAL; 3538 if (seg_not_present == 0) 3539 return -TARGET_EINVAL; 3540 } 3541 /* allocate the LDT */ 3542 if (!ldt_table) { 3543 env->ldt.base = target_mmap(0, 3544 TARGET_LDT_ENTRIES * TARGET_LDT_ENTRY_SIZE, 3545 PROT_READ|PROT_WRITE, 3546 MAP_ANONYMOUS|MAP_PRIVATE, -1, 0); 3547 if (env->ldt.base == -1) 3548 return -TARGET_ENOMEM; 3549 memset(g2h(env->ldt.base), 0, 3550 TARGET_LDT_ENTRIES * TARGET_LDT_ENTRY_SIZE); 3551 env->ldt.limit = 0xffff; 3552 ldt_table = g2h(env->ldt.base); 3553 } 3554 3555 /* NOTE: same code as Linux kernel */ 3556 /* Allow LDTs to be cleared by the user. */ 3557 if (ldt_info.base_addr == 0 && ldt_info.limit == 0) { 3558 if (oldmode || 3559 (contents == 0 && 3560 read_exec_only == 1 && 3561 seg_32bit == 0 && 3562 limit_in_pages == 0 && 3563 seg_not_present == 1 && 3564 useable == 0 )) { 3565 entry_1 = 0; 3566 entry_2 = 0; 3567 goto install; 3568 } 3569 } 3570 3571 entry_1 = ((ldt_info.base_addr & 0x0000ffff) << 16) | 3572 (ldt_info.limit & 0x0ffff); 3573 entry_2 = (ldt_info.base_addr & 0xff000000) | 3574 ((ldt_info.base_addr & 0x00ff0000) >> 16) | 3575 (ldt_info.limit & 0xf0000) | 3576 ((read_exec_only ^ 1) << 9) | 3577 (contents << 10) | 3578 ((seg_not_present ^ 1) << 15) | 3579 (seg_32bit << 22) | 3580 (limit_in_pages << 23) | 3581 (lm << 21) | 3582 0x7000; 3583 if (!oldmode) 3584 entry_2 |= (useable << 20); 3585 3586 /* Install the new entry ... */ 3587 install: 3588 lp = (uint32_t *)(ldt_table + (ldt_info.entry_number << 3)); 3589 lp[0] = tswap32(entry_1); 3590 lp[1] = tswap32(entry_2); 3591 return 0; 3592 } 3593 3594 /* specific and weird i386 syscalls */ 3595 static abi_long do_modify_ldt(CPUX86State *env, int func, abi_ulong ptr, 3596 unsigned long bytecount) 3597 { 3598 abi_long ret; 3599 3600 switch (func) { 3601 case 0: 3602 ret = read_ldt(ptr, bytecount); 3603 break; 3604 case 1: 3605 ret = write_ldt(env, ptr, bytecount, 1); 3606 break; 3607 case 0x11: 3608 ret = write_ldt(env, ptr, bytecount, 0); 3609 break; 3610 default: 3611 ret = -TARGET_ENOSYS; 3612 break; 3613 } 3614 return ret; 3615 } 3616 3617 #if defined(TARGET_I386) && defined(TARGET_ABI32) 3618 static abi_long do_set_thread_area(CPUX86State *env, abi_ulong ptr) 3619 { 3620 uint64_t *gdt_table = g2h(env->gdt.base); 3621 struct target_modify_ldt_ldt_s ldt_info; 3622 struct target_modify_ldt_ldt_s *target_ldt_info; 3623 int seg_32bit, contents, read_exec_only, limit_in_pages; 3624 int seg_not_present, useable, lm; 3625 uint32_t *lp, entry_1, entry_2; 3626 int i; 3627 3628 lock_user_struct(VERIFY_WRITE, target_ldt_info, ptr, 1); 3629 if (!target_ldt_info) 3630 return -TARGET_EFAULT; 3631 ldt_info.entry_number = tswap32(target_ldt_info->entry_number); 3632 ldt_info.base_addr = tswapl(target_ldt_info->base_addr); 3633 ldt_info.limit = tswap32(target_ldt_info->limit); 3634 ldt_info.flags = tswap32(target_ldt_info->flags); 3635 if (ldt_info.entry_number == -1) { 3636 for (i=TARGET_GDT_ENTRY_TLS_MIN; i<=TARGET_GDT_ENTRY_TLS_MAX; i++) { 3637 if (gdt_table[i] == 0) { 3638 ldt_info.entry_number = i; 3639 target_ldt_info->entry_number = tswap32(i); 3640 break; 3641 } 3642 } 3643 } 3644 unlock_user_struct(target_ldt_info, ptr, 1); 3645 3646 if (ldt_info.entry_number < TARGET_GDT_ENTRY_TLS_MIN || 3647 ldt_info.entry_number > TARGET_GDT_ENTRY_TLS_MAX) 3648 return -TARGET_EINVAL; 3649 seg_32bit = ldt_info.flags & 1; 3650 contents = (ldt_info.flags >> 1) & 3; 3651 read_exec_only = (ldt_info.flags >> 3) & 1; 3652 limit_in_pages = (ldt_info.flags >> 4) & 1; 3653 seg_not_present = (ldt_info.flags >> 5) & 1; 3654 useable = (ldt_info.flags >> 6) & 1; 3655 #ifdef TARGET_ABI32 3656 lm = 0; 3657 #else 3658 lm = (ldt_info.flags >> 7) & 1; 3659 #endif 3660 3661 if (contents == 3) { 3662 if (seg_not_present == 0) 3663 return -TARGET_EINVAL; 3664 } 3665 3666 /* NOTE: same code as Linux kernel */ 3667 /* Allow LDTs to be cleared by the user. */ 3668 if (ldt_info.base_addr == 0 && ldt_info.limit == 0) { 3669 if ((contents == 0 && 3670 read_exec_only == 1 && 3671 seg_32bit == 0 && 3672 limit_in_pages == 0 && 3673 seg_not_present == 1 && 3674 useable == 0 )) { 3675 entry_1 = 0; 3676 entry_2 = 0; 3677 goto install; 3678 } 3679 } 3680 3681 entry_1 = ((ldt_info.base_addr & 0x0000ffff) << 16) | 3682 (ldt_info.limit & 0x0ffff); 3683 entry_2 = (ldt_info.base_addr & 0xff000000) | 3684 ((ldt_info.base_addr & 0x00ff0000) >> 16) | 3685 (ldt_info.limit & 0xf0000) | 3686 ((read_exec_only ^ 1) << 9) | 3687 (contents << 10) | 3688 ((seg_not_present ^ 1) << 15) | 3689 (seg_32bit << 22) | 3690 (limit_in_pages << 23) | 3691 (useable << 20) | 3692 (lm << 21) | 3693 0x7000; 3694 3695 /* Install the new entry ... */ 3696 install: 3697 lp = (uint32_t *)(gdt_table + ldt_info.entry_number); 3698 lp[0] = tswap32(entry_1); 3699 lp[1] = tswap32(entry_2); 3700 return 0; 3701 } 3702 3703 static abi_long do_get_thread_area(CPUX86State *env, abi_ulong ptr) 3704 { 3705 struct target_modify_ldt_ldt_s *target_ldt_info; 3706 uint64_t *gdt_table = g2h(env->gdt.base); 3707 uint32_t base_addr, limit, flags; 3708 int seg_32bit, contents, read_exec_only, limit_in_pages, idx; 3709 int seg_not_present, useable, lm; 3710 uint32_t *lp, entry_1, entry_2; 3711 3712 lock_user_struct(VERIFY_WRITE, target_ldt_info, ptr, 1); 3713 if (!target_ldt_info) 3714 return -TARGET_EFAULT; 3715 idx = tswap32(target_ldt_info->entry_number); 3716 if (idx < TARGET_GDT_ENTRY_TLS_MIN || 3717 idx > TARGET_GDT_ENTRY_TLS_MAX) { 3718 unlock_user_struct(target_ldt_info, ptr, 1); 3719 return -TARGET_EINVAL; 3720 } 3721 lp = (uint32_t *)(gdt_table + idx); 3722 entry_1 = tswap32(lp[0]); 3723 entry_2 = tswap32(lp[1]); 3724 3725 read_exec_only = ((entry_2 >> 9) & 1) ^ 1; 3726 contents = (entry_2 >> 10) & 3; 3727 seg_not_present = ((entry_2 >> 15) & 1) ^ 1; 3728 seg_32bit = (entry_2 >> 22) & 1; 3729 limit_in_pages = (entry_2 >> 23) & 1; 3730 useable = (entry_2 >> 20) & 1; 3731 #ifdef TARGET_ABI32 3732 lm = 0; 3733 #else 3734 lm = (entry_2 >> 21) & 1; 3735 #endif 3736 flags = (seg_32bit << 0) | (contents << 1) | 3737 (read_exec_only << 3) | (limit_in_pages << 4) | 3738 (seg_not_present << 5) | (useable << 6) | (lm << 7); 3739 limit = (entry_1 & 0xffff) | (entry_2 & 0xf0000); 3740 base_addr = (entry_1 >> 16) | 3741 (entry_2 & 0xff000000) | 3742 ((entry_2 & 0xff) << 16); 3743 target_ldt_info->base_addr = tswapl(base_addr); 3744 target_ldt_info->limit = tswap32(limit); 3745 target_ldt_info->flags = tswap32(flags); 3746 unlock_user_struct(target_ldt_info, ptr, 1); 3747 return 0; 3748 } 3749 #endif /* TARGET_I386 && TARGET_ABI32 */ 3750 3751 #ifndef TARGET_ABI32 3752 static abi_long do_arch_prctl(CPUX86State *env, int code, abi_ulong addr) 3753 { 3754 abi_long ret; 3755 abi_ulong val; 3756 int idx; 3757 3758 switch(code) { 3759 case TARGET_ARCH_SET_GS: 3760 case TARGET_ARCH_SET_FS: 3761 if (code == TARGET_ARCH_SET_GS) 3762 idx = R_GS; 3763 else 3764 idx = R_FS; 3765 cpu_x86_load_seg(env, idx, 0); 3766 env->segs[idx].base = addr; 3767 break; 3768 case TARGET_ARCH_GET_GS: 3769 case TARGET_ARCH_GET_FS: 3770 if (code == TARGET_ARCH_GET_GS) 3771 idx = R_GS; 3772 else 3773 idx = R_FS; 3774 val = env->segs[idx].base; 3775 if (put_user(val, addr, abi_ulong)) 3776 return -TARGET_EFAULT; 3777 break; 3778 default: 3779 ret = -TARGET_EINVAL; 3780 break; 3781 } 3782 return 0; 3783 } 3784 #endif 3785 3786 #endif /* defined(TARGET_I386) */ 3787 3788 #define NEW_STACK_SIZE 0x40000 3789 3790 #if defined(CONFIG_USE_NPTL) 3791 3792 static pthread_mutex_t clone_lock = PTHREAD_MUTEX_INITIALIZER; 3793 typedef struct { 3794 CPUState *env; 3795 pthread_mutex_t mutex; 3796 pthread_cond_t cond; 3797 pthread_t thread; 3798 uint32_t tid; 3799 abi_ulong child_tidptr; 3800 abi_ulong parent_tidptr; 3801 sigset_t sigmask; 3802 } new_thread_info; 3803 3804 static void *clone_func(void *arg) 3805 { 3806 new_thread_info *info = arg; 3807 CPUState *env; 3808 TaskState *ts; 3809 3810 env = info->env; 3811 thread_env = env; 3812 ts = (TaskState *)thread_env->opaque; 3813 info->tid = gettid(); 3814 env->host_tid = info->tid; 3815 task_settid(ts); 3816 if (info->child_tidptr) 3817 put_user_u32(info->tid, info->child_tidptr); 3818 if (info->parent_tidptr) 3819 put_user_u32(info->tid, info->parent_tidptr); 3820 /* Enable signals. */ 3821 sigprocmask(SIG_SETMASK, &info->sigmask, NULL); 3822 /* Signal to the parent that we're ready. */ 3823 pthread_mutex_lock(&info->mutex); 3824 pthread_cond_broadcast(&info->cond); 3825 pthread_mutex_unlock(&info->mutex); 3826 /* Wait until the parent has finshed initializing the tls state. */ 3827 pthread_mutex_lock(&clone_lock); 3828 pthread_mutex_unlock(&clone_lock); 3829 cpu_loop(env); 3830 /* never exits */ 3831 return NULL; 3832 } 3833 #else 3834 3835 static int clone_func(void *arg) 3836 { 3837 CPUState *env = arg; 3838 cpu_loop(env); 3839 /* never exits */ 3840 return 0; 3841 } 3842 #endif 3843 3844 /* do_fork() Must return host values and target errnos (unlike most 3845 do_*() functions). */ 3846 static int do_fork(CPUState *env, unsigned int flags, abi_ulong newsp, 3847 abi_ulong parent_tidptr, target_ulong newtls, 3848 abi_ulong child_tidptr) 3849 { 3850 int ret; 3851 TaskState *ts; 3852 CPUState *new_env; 3853 #if defined(CONFIG_USE_NPTL) 3854 unsigned int nptl_flags; 3855 sigset_t sigmask; 3856 #else 3857 uint8_t *new_stack; 3858 #endif 3859 3860 /* Emulate vfork() with fork() */ 3861 if (flags & CLONE_VFORK) 3862 flags &= ~(CLONE_VFORK | CLONE_VM); 3863 3864 if (flags & CLONE_VM) { 3865 TaskState *parent_ts = (TaskState *)env->opaque; 3866 #if defined(CONFIG_USE_NPTL) 3867 new_thread_info info; 3868 pthread_attr_t attr; 3869 #endif 3870 ts = qemu_mallocz(sizeof(TaskState)); 3871 init_task_state(ts); 3872 /* we create a new CPU instance. */ 3873 new_env = cpu_copy(env); 3874 #if defined(TARGET_I386) || defined(TARGET_SPARC) || defined(TARGET_PPC) 3875 cpu_reset(new_env); 3876 #endif 3877 /* Init regs that differ from the parent. */ 3878 cpu_clone_regs(new_env, newsp); 3879 new_env->opaque = ts; 3880 ts->bprm = parent_ts->bprm; 3881 ts->info = parent_ts->info; 3882 #if defined(CONFIG_USE_NPTL) 3883 nptl_flags = flags; 3884 flags &= ~CLONE_NPTL_FLAGS2; 3885 3886 if (nptl_flags & CLONE_CHILD_CLEARTID) { 3887 ts->child_tidptr = child_tidptr; 3888 } 3889 3890 if (nptl_flags & CLONE_SETTLS) 3891 cpu_set_tls (new_env, newtls); 3892 3893 /* Grab a mutex so that thread setup appears atomic. */ 3894 pthread_mutex_lock(&clone_lock); 3895 3896 memset(&info, 0, sizeof(info)); 3897 pthread_mutex_init(&info.mutex, NULL); 3898 pthread_mutex_lock(&info.mutex); 3899 pthread_cond_init(&info.cond, NULL); 3900 info.env = new_env; 3901 if (nptl_flags & CLONE_CHILD_SETTID) 3902 info.child_tidptr = child_tidptr; 3903 if (nptl_flags & CLONE_PARENT_SETTID) 3904 info.parent_tidptr = parent_tidptr; 3905 3906 ret = pthread_attr_init(&attr); 3907 ret = pthread_attr_setstacksize(&attr, NEW_STACK_SIZE); 3908 ret = pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_DETACHED); 3909 /* It is not safe to deliver signals until the child has finished 3910 initializing, so temporarily block all signals. */ 3911 sigfillset(&sigmask); 3912 sigprocmask(SIG_BLOCK, &sigmask, &info.sigmask); 3913 3914 ret = pthread_create(&info.thread, &attr, clone_func, &info); 3915 /* TODO: Free new CPU state if thread creation failed. */ 3916 3917 sigprocmask(SIG_SETMASK, &info.sigmask, NULL); 3918 pthread_attr_destroy(&attr); 3919 if (ret == 0) { 3920 /* Wait for the child to initialize. */ 3921 pthread_cond_wait(&info.cond, &info.mutex); 3922 ret = info.tid; 3923 if (flags & CLONE_PARENT_SETTID) 3924 put_user_u32(ret, parent_tidptr); 3925 } else { 3926 ret = -1; 3927 } 3928 pthread_mutex_unlock(&info.mutex); 3929 pthread_cond_destroy(&info.cond); 3930 pthread_mutex_destroy(&info.mutex); 3931 pthread_mutex_unlock(&clone_lock); 3932 #else 3933 if (flags & CLONE_NPTL_FLAGS2) 3934 return -EINVAL; 3935 /* This is probably going to die very quickly, but do it anyway. */ 3936 new_stack = qemu_mallocz (NEW_STACK_SIZE); 3937 #ifdef __ia64__ 3938 ret = __clone2(clone_func, new_stack, NEW_STACK_SIZE, flags, new_env); 3939 #else 3940 ret = clone(clone_func, new_stack + NEW_STACK_SIZE, flags, new_env); 3941 #endif 3942 #endif 3943 } else { 3944 /* if no CLONE_VM, we consider it is a fork */ 3945 if ((flags & ~(CSIGNAL | CLONE_NPTL_FLAGS2)) != 0) 3946 return -EINVAL; 3947 fork_start(); 3948 ret = fork(); 3949 if (ret == 0) { 3950 /* Child Process. */ 3951 cpu_clone_regs(env, newsp); 3952 fork_end(1); 3953 #if defined(CONFIG_USE_NPTL) 3954 /* There is a race condition here. The parent process could 3955 theoretically read the TID in the child process before the child 3956 tid is set. This would require using either ptrace 3957 (not implemented) or having *_tidptr to point at a shared memory 3958 mapping. We can't repeat the spinlock hack used above because 3959 the child process gets its own copy of the lock. */ 3960 if (flags & CLONE_CHILD_SETTID) 3961 put_user_u32(gettid(), child_tidptr); 3962 if (flags & CLONE_PARENT_SETTID) 3963 put_user_u32(gettid(), parent_tidptr); 3964 ts = (TaskState *)env->opaque; 3965 if (flags & CLONE_SETTLS) 3966 cpu_set_tls (env, newtls); 3967 if (flags & CLONE_CHILD_CLEARTID) 3968 ts->child_tidptr = child_tidptr; 3969 #endif 3970 } else { 3971 fork_end(0); 3972 } 3973 } 3974 return ret; 3975 } 3976 3977 /* warning : doesn't handle linux specific flags... */ 3978 static int target_to_host_fcntl_cmd(int cmd) 3979 { 3980 switch(cmd) { 3981 case TARGET_F_DUPFD: 3982 case TARGET_F_GETFD: 3983 case TARGET_F_SETFD: 3984 case TARGET_F_GETFL: 3985 case TARGET_F_SETFL: 3986 return cmd; 3987 case TARGET_F_GETLK: 3988 return F_GETLK; 3989 case TARGET_F_SETLK: 3990 return F_SETLK; 3991 case TARGET_F_SETLKW: 3992 return F_SETLKW; 3993 case TARGET_F_GETOWN: 3994 return F_GETOWN; 3995 case TARGET_F_SETOWN: 3996 return F_SETOWN; 3997 case TARGET_F_GETSIG: 3998 return F_GETSIG; 3999 case TARGET_F_SETSIG: 4000 return F_SETSIG; 4001 #if TARGET_ABI_BITS == 32 4002 case TARGET_F_GETLK64: 4003 return F_GETLK64; 4004 case TARGET_F_SETLK64: 4005 return F_SETLK64; 4006 case TARGET_F_SETLKW64: 4007 return F_SETLKW64; 4008 #endif 4009 case TARGET_F_SETLEASE: 4010 return F_SETLEASE; 4011 case TARGET_F_GETLEASE: 4012 return F_GETLEASE; 4013 #ifdef F_DUPFD_CLOEXEC 4014 case TARGET_F_DUPFD_CLOEXEC: 4015 return F_DUPFD_CLOEXEC; 4016 #endif 4017 case TARGET_F_NOTIFY: 4018 return F_NOTIFY; 4019 default: 4020 return -TARGET_EINVAL; 4021 } 4022 return -TARGET_EINVAL; 4023 } 4024 4025 static abi_long do_fcntl(int fd, int cmd, abi_ulong arg) 4026 { 4027 struct flock fl; 4028 struct target_flock *target_fl; 4029 struct flock64 fl64; 4030 struct target_flock64 *target_fl64; 4031 abi_long ret; 4032 int host_cmd = target_to_host_fcntl_cmd(cmd); 4033 4034 if (host_cmd == -TARGET_EINVAL) 4035 return host_cmd; 4036 4037 switch(cmd) { 4038 case TARGET_F_GETLK: 4039 if (!lock_user_struct(VERIFY_READ, target_fl, arg, 1)) 4040 return -TARGET_EFAULT; 4041 fl.l_type = tswap16(target_fl->l_type); 4042 fl.l_whence = tswap16(target_fl->l_whence); 4043 fl.l_start = tswapl(target_fl->l_start); 4044 fl.l_len = tswapl(target_fl->l_len); 4045 fl.l_pid = tswap32(target_fl->l_pid); 4046 unlock_user_struct(target_fl, arg, 0); 4047 ret = get_errno(fcntl(fd, host_cmd, &fl)); 4048 if (ret == 0) { 4049 if (!lock_user_struct(VERIFY_WRITE, target_fl, arg, 0)) 4050 return -TARGET_EFAULT; 4051 target_fl->l_type = tswap16(fl.l_type); 4052 target_fl->l_whence = tswap16(fl.l_whence); 4053 target_fl->l_start = tswapl(fl.l_start); 4054 target_fl->l_len = tswapl(fl.l_len); 4055 target_fl->l_pid = tswap32(fl.l_pid); 4056 unlock_user_struct(target_fl, arg, 1); 4057 } 4058 break; 4059 4060 case TARGET_F_SETLK: 4061 case TARGET_F_SETLKW: 4062 if (!lock_user_struct(VERIFY_READ, target_fl, arg, 1)) 4063 return -TARGET_EFAULT; 4064 fl.l_type = tswap16(target_fl->l_type); 4065 fl.l_whence = tswap16(target_fl->l_whence); 4066 fl.l_start = tswapl(target_fl->l_start); 4067 fl.l_len = tswapl(target_fl->l_len); 4068 fl.l_pid = tswap32(target_fl->l_pid); 4069 unlock_user_struct(target_fl, arg, 0); 4070 ret = get_errno(fcntl(fd, host_cmd, &fl)); 4071 break; 4072 4073 case TARGET_F_GETLK64: 4074 if (!lock_user_struct(VERIFY_READ, target_fl64, arg, 1)) 4075 return -TARGET_EFAULT; 4076 fl64.l_type = tswap16(target_fl64->l_type) >> 1; 4077 fl64.l_whence = tswap16(target_fl64->l_whence); 4078 fl64.l_start = tswapl(target_fl64->l_start); 4079 fl64.l_len = tswapl(target_fl64->l_len); 4080 fl64.l_pid = tswap32(target_fl64->l_pid); 4081 unlock_user_struct(target_fl64, arg, 0); 4082 ret = get_errno(fcntl(fd, host_cmd, &fl64)); 4083 if (ret == 0) { 4084 if (!lock_user_struct(VERIFY_WRITE, target_fl64, arg, 0)) 4085 return -TARGET_EFAULT; 4086 target_fl64->l_type = tswap16(fl64.l_type) >> 1; 4087 target_fl64->l_whence = tswap16(fl64.l_whence); 4088 target_fl64->l_start = tswapl(fl64.l_start); 4089 target_fl64->l_len = tswapl(fl64.l_len); 4090 target_fl64->l_pid = tswap32(fl64.l_pid); 4091 unlock_user_struct(target_fl64, arg, 1); 4092 } 4093 break; 4094 case TARGET_F_SETLK64: 4095 case TARGET_F_SETLKW64: 4096 if (!lock_user_struct(VERIFY_READ, target_fl64, arg, 1)) 4097 return -TARGET_EFAULT; 4098 fl64.l_type = tswap16(target_fl64->l_type) >> 1; 4099 fl64.l_whence = tswap16(target_fl64->l_whence); 4100 fl64.l_start = tswapl(target_fl64->l_start); 4101 fl64.l_len = tswapl(target_fl64->l_len); 4102 fl64.l_pid = tswap32(target_fl64->l_pid); 4103 unlock_user_struct(target_fl64, arg, 0); 4104 ret = get_errno(fcntl(fd, host_cmd, &fl64)); 4105 break; 4106 4107 case TARGET_F_GETFL: 4108 ret = get_errno(fcntl(fd, host_cmd, arg)); 4109 if (ret >= 0) { 4110 ret = host_to_target_bitmask(ret, fcntl_flags_tbl); 4111 } 4112 break; 4113 4114 case TARGET_F_SETFL: 4115 ret = get_errno(fcntl(fd, host_cmd, target_to_host_bitmask(arg, fcntl_flags_tbl))); 4116 break; 4117 4118 case TARGET_F_SETOWN: 4119 case TARGET_F_GETOWN: 4120 case TARGET_F_SETSIG: 4121 case TARGET_F_GETSIG: 4122 case TARGET_F_SETLEASE: 4123 case TARGET_F_GETLEASE: 4124 ret = get_errno(fcntl(fd, host_cmd, arg)); 4125 break; 4126 4127 default: 4128 ret = get_errno(fcntl(fd, cmd, arg)); 4129 break; 4130 } 4131 return ret; 4132 } 4133 4134 #ifdef USE_UID16 4135 4136 static inline int high2lowuid(int uid) 4137 { 4138 if (uid > 65535) 4139 return 65534; 4140 else 4141 return uid; 4142 } 4143 4144 static inline int high2lowgid(int gid) 4145 { 4146 if (gid > 65535) 4147 return 65534; 4148 else 4149 return gid; 4150 } 4151 4152 static inline int low2highuid(int uid) 4153 { 4154 if ((int16_t)uid == -1) 4155 return -1; 4156 else 4157 return uid; 4158 } 4159 4160 static inline int low2highgid(int gid) 4161 { 4162 if ((int16_t)gid == -1) 4163 return -1; 4164 else 4165 return gid; 4166 } 4167 4168 #endif /* USE_UID16 */ 4169 4170 void syscall_init(void) 4171 { 4172 IOCTLEntry *ie; 4173 const argtype *arg_type; 4174 int size; 4175 int i; 4176 4177 #define STRUCT(name, ...) thunk_register_struct(STRUCT_ ## name, #name, struct_ ## name ## _def); 4178 #define STRUCT_SPECIAL(name) thunk_register_struct_direct(STRUCT_ ## name, #name, &struct_ ## name ## _def); 4179 #include "syscall_types.h" 4180 #undef STRUCT 4181 #undef STRUCT_SPECIAL 4182 4183 /* we patch the ioctl size if necessary. We rely on the fact that 4184 no ioctl has all the bits at '1' in the size field */ 4185 ie = ioctl_entries; 4186 while (ie->target_cmd != 0) { 4187 if (((ie->target_cmd >> TARGET_IOC_SIZESHIFT) & TARGET_IOC_SIZEMASK) == 4188 TARGET_IOC_SIZEMASK) { 4189 arg_type = ie->arg_type; 4190 if (arg_type[0] != TYPE_PTR) { 4191 fprintf(stderr, "cannot patch size for ioctl 0x%x\n", 4192 ie->target_cmd); 4193 exit(1); 4194 } 4195 arg_type++; 4196 size = thunk_type_size(arg_type, 0); 4197 ie->target_cmd = (ie->target_cmd & 4198 ~(TARGET_IOC_SIZEMASK << TARGET_IOC_SIZESHIFT)) | 4199 (size << TARGET_IOC_SIZESHIFT); 4200 } 4201 4202 /* Build target_to_host_errno_table[] table from 4203 * host_to_target_errno_table[]. */ 4204 for (i=0; i < ERRNO_TABLE_SIZE; i++) 4205 target_to_host_errno_table[host_to_target_errno_table[i]] = i; 4206 4207 /* automatic consistency check if same arch */ 4208 #if (defined(__i386__) && defined(TARGET_I386) && defined(TARGET_ABI32)) || \ 4209 (defined(__x86_64__) && defined(TARGET_X86_64)) 4210 if (unlikely(ie->target_cmd != ie->host_cmd)) { 4211 fprintf(stderr, "ERROR: ioctl(%s): target=0x%x host=0x%x\n", 4212 ie->name, ie->target_cmd, ie->host_cmd); 4213 } 4214 #endif 4215 ie++; 4216 } 4217 } 4218 4219 #if TARGET_ABI_BITS == 32 4220 static inline uint64_t target_offset64(uint32_t word0, uint32_t word1) 4221 { 4222 #ifdef TARGET_WORDS_BIGENDIAN 4223 return ((uint64_t)word0 << 32) | word1; 4224 #else 4225 return ((uint64_t)word1 << 32) | word0; 4226 #endif 4227 } 4228 #else /* TARGET_ABI_BITS == 32 */ 4229 static inline uint64_t target_offset64(uint64_t word0, uint64_t word1) 4230 { 4231 return word0; 4232 } 4233 #endif /* TARGET_ABI_BITS != 32 */ 4234 4235 #ifdef TARGET_NR_truncate64 4236 static inline abi_long target_truncate64(void *cpu_env, const char *arg1, 4237 abi_long arg2, 4238 abi_long arg3, 4239 abi_long arg4) 4240 { 4241 #ifdef TARGET_ARM 4242 if (((CPUARMState *)cpu_env)->eabi) 4243 { 4244 arg2 = arg3; 4245 arg3 = arg4; 4246 } 4247 #endif 4248 return get_errno(truncate64(arg1, target_offset64(arg2, arg3))); 4249 } 4250 #endif 4251 4252 #ifdef TARGET_NR_ftruncate64 4253 static inline abi_long target_ftruncate64(void *cpu_env, abi_long arg1, 4254 abi_long arg2, 4255 abi_long arg3, 4256 abi_long arg4) 4257 { 4258 #ifdef TARGET_ARM 4259 if (((CPUARMState *)cpu_env)->eabi) 4260 { 4261 arg2 = arg3; 4262 arg3 = arg4; 4263 } 4264 #endif 4265 return get_errno(ftruncate64(arg1, target_offset64(arg2, arg3))); 4266 } 4267 #endif 4268 4269 static inline abi_long target_to_host_timespec(struct timespec *host_ts, 4270 abi_ulong target_addr) 4271 { 4272 struct target_timespec *target_ts; 4273 4274 if (!lock_user_struct(VERIFY_READ, target_ts, target_addr, 1)) 4275 return -TARGET_EFAULT; 4276 host_ts->tv_sec = tswapl(target_ts->tv_sec); 4277 host_ts->tv_nsec = tswapl(target_ts->tv_nsec); 4278 unlock_user_struct(target_ts, target_addr, 0); 4279 return 0; 4280 } 4281 4282 static inline abi_long host_to_target_timespec(abi_ulong target_addr, 4283 struct timespec *host_ts) 4284 { 4285 struct target_timespec *target_ts; 4286 4287 if (!lock_user_struct(VERIFY_WRITE, target_ts, target_addr, 0)) 4288 return -TARGET_EFAULT; 4289 target_ts->tv_sec = tswapl(host_ts->tv_sec); 4290 target_ts->tv_nsec = tswapl(host_ts->tv_nsec); 4291 unlock_user_struct(target_ts, target_addr, 1); 4292 return 0; 4293 } 4294 4295 #if defined(TARGET_NR_stat64) || defined(TARGET_NR_newfstatat) 4296 static inline abi_long host_to_target_stat64(void *cpu_env, 4297 abi_ulong target_addr, 4298 struct stat *host_st) 4299 { 4300 #ifdef TARGET_ARM 4301 if (((CPUARMState *)cpu_env)->eabi) { 4302 struct target_eabi_stat64 *target_st; 4303 4304 if (!lock_user_struct(VERIFY_WRITE, target_st, target_addr, 0)) 4305 return -TARGET_EFAULT; 4306 memset(target_st, 0, sizeof(struct target_eabi_stat64)); 4307 __put_user(host_st->st_dev, &target_st->st_dev); 4308 __put_user(host_st->st_ino, &target_st->st_ino); 4309 #ifdef TARGET_STAT64_HAS_BROKEN_ST_INO 4310 __put_user(host_st->st_ino, &target_st->__st_ino); 4311 #endif 4312 __put_user(host_st->st_mode, &target_st->st_mode); 4313 __put_user(host_st->st_nlink, &target_st->st_nlink); 4314 __put_user(host_st->st_uid, &target_st->st_uid); 4315 __put_user(host_st->st_gid, &target_st->st_gid); 4316 __put_user(host_st->st_rdev, &target_st->st_rdev); 4317 __put_user(host_st->st_size, &target_st->st_size); 4318 __put_user(host_st->st_blksize, &target_st->st_blksize); 4319 __put_user(host_st->st_blocks, &target_st->st_blocks); 4320 __put_user(host_st->st_atime, &target_st->target_st_atime); 4321 __put_user(host_st->st_mtime, &target_st->target_st_mtime); 4322 __put_user(host_st->st_ctime, &target_st->target_st_ctime); 4323 unlock_user_struct(target_st, target_addr, 1); 4324 } else 4325 #endif 4326 { 4327 #if TARGET_ABI_BITS == 64 && !defined(TARGET_ALPHA) 4328 struct target_stat *target_st; 4329 #else 4330 struct target_stat64 *target_st; 4331 #endif 4332 4333 if (!lock_user_struct(VERIFY_WRITE, target_st, target_addr, 0)) 4334 return -TARGET_EFAULT; 4335 memset(target_st, 0, sizeof(*target_st)); 4336 __put_user(host_st->st_dev, &target_st->st_dev); 4337 __put_user(host_st->st_ino, &target_st->st_ino); 4338 #ifdef TARGET_STAT64_HAS_BROKEN_ST_INO 4339 __put_user(host_st->st_ino, &target_st->__st_ino); 4340 #endif 4341 __put_user(host_st->st_mode, &target_st->st_mode); 4342 __put_user(host_st->st_nlink, &target_st->st_nlink); 4343 __put_user(host_st->st_uid, &target_st->st_uid); 4344 __put_user(host_st->st_gid, &target_st->st_gid); 4345 __put_user(host_st->st_rdev, &target_st->st_rdev); 4346 /* XXX: better use of kernel struct */ 4347 __put_user(host_st->st_size, &target_st->st_size); 4348 __put_user(host_st->st_blksize, &target_st->st_blksize); 4349 __put_user(host_st->st_blocks, &target_st->st_blocks); 4350 __put_user(host_st->st_atime, &target_st->target_st_atime); 4351 __put_user(host_st->st_mtime, &target_st->target_st_mtime); 4352 __put_user(host_st->st_ctime, &target_st->target_st_ctime); 4353 unlock_user_struct(target_st, target_addr, 1); 4354 } 4355 4356 return 0; 4357 } 4358 #endif 4359 4360 #if defined(CONFIG_USE_NPTL) 4361 /* ??? Using host futex calls even when target atomic operations 4362 are not really atomic probably breaks things. However implementing 4363 futexes locally would make futexes shared between multiple processes 4364 tricky. However they're probably useless because guest atomic 4365 operations won't work either. */ 4366 static int do_futex(target_ulong uaddr, int op, int val, target_ulong timeout, 4367 target_ulong uaddr2, int val3) 4368 { 4369 struct timespec ts, *pts; 4370 int base_op; 4371 4372 /* ??? We assume FUTEX_* constants are the same on both host 4373 and target. */ 4374 #ifdef FUTEX_CMD_MASK 4375 base_op = op & FUTEX_CMD_MASK; 4376 #else 4377 base_op = op; 4378 #endif 4379 switch (base_op) { 4380 case FUTEX_WAIT: 4381 if (timeout) { 4382 pts = &ts; 4383 target_to_host_timespec(pts, timeout); 4384 } else { 4385 pts = NULL; 4386 } 4387 return get_errno(sys_futex(g2h(uaddr), op, tswap32(val), 4388 pts, NULL, 0)); 4389 case FUTEX_WAKE: 4390 return get_errno(sys_futex(g2h(uaddr), op, val, NULL, NULL, 0)); 4391 case FUTEX_FD: 4392 return get_errno(sys_futex(g2h(uaddr), op, val, NULL, NULL, 0)); 4393 case FUTEX_REQUEUE: 4394 case FUTEX_CMP_REQUEUE: 4395 case FUTEX_WAKE_OP: 4396 /* For FUTEX_REQUEUE, FUTEX_CMP_REQUEUE, and FUTEX_WAKE_OP, the 4397 TIMEOUT parameter is interpreted as a uint32_t by the kernel. 4398 But the prototype takes a `struct timespec *'; insert casts 4399 to satisfy the compiler. We do not need to tswap TIMEOUT 4400 since it's not compared to guest memory. */ 4401 pts = (struct timespec *)(uintptr_t) timeout; 4402 return get_errno(sys_futex(g2h(uaddr), op, val, pts, 4403 g2h(uaddr2), 4404 (base_op == FUTEX_CMP_REQUEUE 4405 ? tswap32(val3) 4406 : val3))); 4407 default: 4408 return -TARGET_ENOSYS; 4409 } 4410 } 4411 #endif 4412 4413 /* Map host to target signal numbers for the wait family of syscalls. 4414 Assume all other status bits are the same. */ 4415 static int host_to_target_waitstatus(int status) 4416 { 4417 if (WIFSIGNALED(status)) { 4418 return host_to_target_signal(WTERMSIG(status)) | (status & ~0x7f); 4419 } 4420 if (WIFSTOPPED(status)) { 4421 return (host_to_target_signal(WSTOPSIG(status)) << 8) 4422 | (status & 0xff); 4423 } 4424 return status; 4425 } 4426 4427 int get_osversion(void) 4428 { 4429 static int osversion; 4430 struct new_utsname buf; 4431 const char *s; 4432 int i, n, tmp; 4433 if (osversion) 4434 return osversion; 4435 if (qemu_uname_release && *qemu_uname_release) { 4436 s = qemu_uname_release; 4437 } else { 4438 if (sys_uname(&buf)) 4439 return 0; 4440 s = buf.release; 4441 } 4442 tmp = 0; 4443 for (i = 0; i < 3; i++) { 4444 n = 0; 4445 while (*s >= '0' && *s <= '9') { 4446 n *= 10; 4447 n += *s - '0'; 4448 s++; 4449 } 4450 tmp = (tmp << 8) + n; 4451 if (*s == '.') 4452 s++; 4453 } 4454 osversion = tmp; 4455 return osversion; 4456 } 4457 4458 /* do_syscall() should always have a single exit point at the end so 4459 that actions, such as logging of syscall results, can be performed. 4460 All errnos that do_syscall() returns must be -TARGET_<errcode>. */ 4461 abi_long do_syscall(void *cpu_env, int num, abi_long arg1, 4462 abi_long arg2, abi_long arg3, abi_long arg4, 4463 abi_long arg5, abi_long arg6) 4464 { 4465 abi_long ret; 4466 struct stat st; 4467 struct statfs stfs; 4468 void *p; 4469 4470 #ifdef DEBUG 4471 gemu_log("syscall %d", num); 4472 #endif 4473 if(do_strace) 4474 print_syscall(num, arg1, arg2, arg3, arg4, arg5, arg6); 4475 4476 switch(num) { 4477 case TARGET_NR_exit: 4478 #ifdef CONFIG_USE_NPTL 4479 /* In old applications this may be used to implement _exit(2). 4480 However in threaded applictions it is used for thread termination, 4481 and _exit_group is used for application termination. 4482 Do thread termination if we have more then one thread. */ 4483 /* FIXME: This probably breaks if a signal arrives. We should probably 4484 be disabling signals. */ 4485 if (first_cpu->next_cpu) { 4486 TaskState *ts; 4487 CPUState **lastp; 4488 CPUState *p; 4489 4490 cpu_list_lock(); 4491 lastp = &first_cpu; 4492 p = first_cpu; 4493 while (p && p != (CPUState *)cpu_env) { 4494 lastp = &p->next_cpu; 4495 p = p->next_cpu; 4496 } 4497 /* If we didn't find the CPU for this thread then something is 4498 horribly wrong. */ 4499 if (!p) 4500 abort(); 4501 /* Remove the CPU from the list. */ 4502 *lastp = p->next_cpu; 4503 cpu_list_unlock(); 4504 ts = ((CPUState *)cpu_env)->opaque; 4505 if (ts->child_tidptr) { 4506 put_user_u32(0, ts->child_tidptr); 4507 sys_futex(g2h(ts->child_tidptr), FUTEX_WAKE, INT_MAX, 4508 NULL, NULL, 0); 4509 } 4510 thread_env = NULL; 4511 qemu_free(cpu_env); 4512 qemu_free(ts); 4513 pthread_exit(NULL); 4514 } 4515 #endif 4516 #ifdef TARGET_GPROF 4517 _mcleanup(); 4518 #endif 4519 gdb_exit(cpu_env, arg1); 4520 _exit(arg1); 4521 ret = 0; /* avoid warning */ 4522 break; 4523 case TARGET_NR_read: 4524 if (arg3 == 0) 4525 ret = 0; 4526 else { 4527 if (!(p = lock_user(VERIFY_WRITE, arg2, arg3, 0))) 4528 goto efault; 4529 ret = get_errno(read(arg1, p, arg3)); 4530 unlock_user(p, arg2, ret); 4531 } 4532 break; 4533 case TARGET_NR_write: 4534 if (!(p = lock_user(VERIFY_READ, arg2, arg3, 1))) 4535 goto efault; 4536 ret = get_errno(write(arg1, p, arg3)); 4537 unlock_user(p, arg2, 0); 4538 break; 4539 case TARGET_NR_open: 4540 if (!(p = lock_user_string(arg1))) 4541 goto efault; 4542 ret = get_errno(open(path(p), 4543 target_to_host_bitmask(arg2, fcntl_flags_tbl), 4544 arg3)); 4545 unlock_user(p, arg1, 0); 4546 break; 4547 #if defined(TARGET_NR_openat) && defined(__NR_openat) 4548 case TARGET_NR_openat: 4549 if (!(p = lock_user_string(arg2))) 4550 goto efault; 4551 ret = get_errno(sys_openat(arg1, 4552 path(p), 4553 target_to_host_bitmask(arg3, fcntl_flags_tbl), 4554 arg4)); 4555 unlock_user(p, arg2, 0); 4556 break; 4557 #endif 4558 case TARGET_NR_close: 4559 ret = get_errno(close(arg1)); 4560 break; 4561 case TARGET_NR_brk: 4562 ret = do_brk(arg1); 4563 break; 4564 case TARGET_NR_fork: 4565 ret = get_errno(do_fork(cpu_env, SIGCHLD, 0, 0, 0, 0)); 4566 break; 4567 #ifdef TARGET_NR_waitpid 4568 case TARGET_NR_waitpid: 4569 { 4570 int status; 4571 ret = get_errno(waitpid(arg1, &status, arg3)); 4572 if (!is_error(ret) && arg2 4573 && put_user_s32(host_to_target_waitstatus(status), arg2)) 4574 goto efault; 4575 } 4576 break; 4577 #endif 4578 #ifdef TARGET_NR_waitid 4579 case TARGET_NR_waitid: 4580 { 4581 siginfo_t info; 4582 info.si_pid = 0; 4583 ret = get_errno(waitid(arg1, arg2, &info, arg4)); 4584 if (!is_error(ret) && arg3 && info.si_pid != 0) { 4585 if (!(p = lock_user(VERIFY_WRITE, arg3, sizeof(target_siginfo_t), 0))) 4586 goto efault; 4587 host_to_target_siginfo(p, &info); 4588 unlock_user(p, arg3, sizeof(target_siginfo_t)); 4589 } 4590 } 4591 break; 4592 #endif 4593 #ifdef TARGET_NR_creat /* not on alpha */ 4594 case TARGET_NR_creat: 4595 if (!(p = lock_user_string(arg1))) 4596 goto efault; 4597 ret = get_errno(creat(p, arg2)); 4598 unlock_user(p, arg1, 0); 4599 break; 4600 #endif 4601 case TARGET_NR_link: 4602 { 4603 void * p2; 4604 p = lock_user_string(arg1); 4605 p2 = lock_user_string(arg2); 4606 if (!p || !p2) 4607 ret = -TARGET_EFAULT; 4608 else 4609 ret = get_errno(link(p, p2)); 4610 unlock_user(p2, arg2, 0); 4611 unlock_user(p, arg1, 0); 4612 } 4613 break; 4614 #if defined(TARGET_NR_linkat) && defined(__NR_linkat) 4615 case TARGET_NR_linkat: 4616 { 4617 void * p2 = NULL; 4618 if (!arg2 || !arg4) 4619 goto efault; 4620 p = lock_user_string(arg2); 4621 p2 = lock_user_string(arg4); 4622 if (!p || !p2) 4623 ret = -TARGET_EFAULT; 4624 else 4625 ret = get_errno(sys_linkat(arg1, p, arg3, p2, arg5)); 4626 unlock_user(p, arg2, 0); 4627 unlock_user(p2, arg4, 0); 4628 } 4629 break; 4630 #endif 4631 case TARGET_NR_unlink: 4632 if (!(p = lock_user_string(arg1))) 4633 goto efault; 4634 ret = get_errno(unlink(p)); 4635 unlock_user(p, arg1, 0); 4636 break; 4637 #if defined(TARGET_NR_unlinkat) && defined(__NR_unlinkat) 4638 case TARGET_NR_unlinkat: 4639 if (!(p = lock_user_string(arg2))) 4640 goto efault; 4641 ret = get_errno(sys_unlinkat(arg1, p, arg3)); 4642 unlock_user(p, arg2, 0); 4643 break; 4644 #endif 4645 case TARGET_NR_execve: 4646 { 4647 char **argp, **envp; 4648 int argc, envc; 4649 abi_ulong gp; 4650 abi_ulong guest_argp; 4651 abi_ulong guest_envp; 4652 abi_ulong addr; 4653 char **q; 4654 4655 argc = 0; 4656 guest_argp = arg2; 4657 for (gp = guest_argp; gp; gp += sizeof(abi_ulong)) { 4658 if (get_user_ual(addr, gp)) 4659 goto efault; 4660 if (!addr) 4661 break; 4662 argc++; 4663 } 4664 envc = 0; 4665 guest_envp = arg3; 4666 for (gp = guest_envp; gp; gp += sizeof(abi_ulong)) { 4667 if (get_user_ual(addr, gp)) 4668 goto efault; 4669 if (!addr) 4670 break; 4671 envc++; 4672 } 4673 4674 argp = alloca((argc + 1) * sizeof(void *)); 4675 envp = alloca((envc + 1) * sizeof(void *)); 4676 4677 for (gp = guest_argp, q = argp; gp; 4678 gp += sizeof(abi_ulong), q++) { 4679 if (get_user_ual(addr, gp)) 4680 goto execve_efault; 4681 if (!addr) 4682 break; 4683 if (!(*q = lock_user_string(addr))) 4684 goto execve_efault; 4685 } 4686 *q = NULL; 4687 4688 for (gp = guest_envp, q = envp; gp; 4689 gp += sizeof(abi_ulong), q++) { 4690 if (get_user_ual(addr, gp)) 4691 goto execve_efault; 4692 if (!addr) 4693 break; 4694 if (!(*q = lock_user_string(addr))) 4695 goto execve_efault; 4696 } 4697 *q = NULL; 4698 4699 if (!(p = lock_user_string(arg1))) 4700 goto execve_efault; 4701 ret = get_errno(execve(p, argp, envp)); 4702 unlock_user(p, arg1, 0); 4703 4704 goto execve_end; 4705 4706 execve_efault: 4707 ret = -TARGET_EFAULT; 4708 4709 execve_end: 4710 for (gp = guest_argp, q = argp; *q; 4711 gp += sizeof(abi_ulong), q++) { 4712 if (get_user_ual(addr, gp) 4713 || !addr) 4714 break; 4715 unlock_user(*q, addr, 0); 4716 } 4717 for (gp = guest_envp, q = envp; *q; 4718 gp += sizeof(abi_ulong), q++) { 4719 if (get_user_ual(addr, gp) 4720 || !addr) 4721 break; 4722 unlock_user(*q, addr, 0); 4723 } 4724 } 4725 break; 4726 case TARGET_NR_chdir: 4727 if (!(p = lock_user_string(arg1))) 4728 goto efault; 4729 ret = get_errno(chdir(p)); 4730 unlock_user(p, arg1, 0); 4731 break; 4732 #ifdef TARGET_NR_time 4733 case TARGET_NR_time: 4734 { 4735 time_t host_time; 4736 ret = get_errno(time(&host_time)); 4737 if (!is_error(ret) 4738 && arg1 4739 && put_user_sal(host_time, arg1)) 4740 goto efault; 4741 } 4742 break; 4743 #endif 4744 case TARGET_NR_mknod: 4745 if (!(p = lock_user_string(arg1))) 4746 goto efault; 4747 ret = get_errno(mknod(p, arg2, arg3)); 4748 unlock_user(p, arg1, 0); 4749 break; 4750 #if defined(TARGET_NR_mknodat) && defined(__NR_mknodat) 4751 case TARGET_NR_mknodat: 4752 if (!(p = lock_user_string(arg2))) 4753 goto efault; 4754 ret = get_errno(sys_mknodat(arg1, p, arg3, arg4)); 4755 unlock_user(p, arg2, 0); 4756 break; 4757 #endif 4758 case TARGET_NR_chmod: 4759 if (!(p = lock_user_string(arg1))) 4760 goto efault; 4761 ret = get_errno(chmod(p, arg2)); 4762 unlock_user(p, arg1, 0); 4763 break; 4764 #ifdef TARGET_NR_break 4765 case TARGET_NR_break: 4766 goto unimplemented; 4767 #endif 4768 #ifdef TARGET_NR_oldstat 4769 case TARGET_NR_oldstat: 4770 goto unimplemented; 4771 #endif 4772 case TARGET_NR_lseek: 4773 ret = get_errno(lseek(arg1, arg2, arg3)); 4774 break; 4775 #if defined(TARGET_NR_getxpid) && defined(TARGET_ALPHA) 4776 /* Alpha specific */ 4777 case TARGET_NR_getxpid: 4778 ((CPUAlphaState *)cpu_env)->ir[IR_A4] = getppid(); 4779 ret = get_errno(getpid()); 4780 break; 4781 #endif 4782 #ifdef TARGET_NR_getpid 4783 case TARGET_NR_getpid: 4784 ret = get_errno(getpid()); 4785 break; 4786 #endif 4787 case TARGET_NR_mount: 4788 { 4789 /* need to look at the data field */ 4790 void *p2, *p3; 4791 p = lock_user_string(arg1); 4792 p2 = lock_user_string(arg2); 4793 p3 = lock_user_string(arg3); 4794 if (!p || !p2 || !p3) 4795 ret = -TARGET_EFAULT; 4796 else { 4797 /* FIXME - arg5 should be locked, but it isn't clear how to 4798 * do that since it's not guaranteed to be a NULL-terminated 4799 * string. 4800 */ 4801 if ( ! arg5 ) 4802 ret = get_errno(mount(p, p2, p3, (unsigned long)arg4, NULL)); 4803 else 4804 ret = get_errno(mount(p, p2, p3, (unsigned long)arg4, g2h(arg5))); 4805 } 4806 unlock_user(p, arg1, 0); 4807 unlock_user(p2, arg2, 0); 4808 unlock_user(p3, arg3, 0); 4809 break; 4810 } 4811 #ifdef TARGET_NR_umount 4812 case TARGET_NR_umount: 4813 if (!(p = lock_user_string(arg1))) 4814 goto efault; 4815 ret = get_errno(umount(p)); 4816 unlock_user(p, arg1, 0); 4817 break; 4818 #endif 4819 #ifdef TARGET_NR_stime /* not on alpha */ 4820 case TARGET_NR_stime: 4821 { 4822 time_t host_time; 4823 if (get_user_sal(host_time, arg1)) 4824 goto efault; 4825 ret = get_errno(stime(&host_time)); 4826 } 4827 break; 4828 #endif 4829 case TARGET_NR_ptrace: 4830 goto unimplemented; 4831 #ifdef TARGET_NR_alarm /* not on alpha */ 4832 case TARGET_NR_alarm: 4833 ret = alarm(arg1); 4834 break; 4835 #endif 4836 #ifdef TARGET_NR_oldfstat 4837 case TARGET_NR_oldfstat: 4838 goto unimplemented; 4839 #endif 4840 #ifdef TARGET_NR_pause /* not on alpha */ 4841 case TARGET_NR_pause: 4842 ret = get_errno(pause()); 4843 break; 4844 #endif 4845 #ifdef TARGET_NR_utime 4846 case TARGET_NR_utime: 4847 { 4848 struct utimbuf tbuf, *host_tbuf; 4849 struct target_utimbuf *target_tbuf; 4850 if (arg2) { 4851 if (!lock_user_struct(VERIFY_READ, target_tbuf, arg2, 1)) 4852 goto efault; 4853 tbuf.actime = tswapl(target_tbuf->actime); 4854 tbuf.modtime = tswapl(target_tbuf->modtime); 4855 unlock_user_struct(target_tbuf, arg2, 0); 4856 host_tbuf = &tbuf; 4857 } else { 4858 host_tbuf = NULL; 4859 } 4860 if (!(p = lock_user_string(arg1))) 4861 goto efault; 4862 ret = get_errno(utime(p, host_tbuf)); 4863 unlock_user(p, arg1, 0); 4864 } 4865 break; 4866 #endif 4867 case TARGET_NR_utimes: 4868 { 4869 struct timeval *tvp, tv[2]; 4870 if (arg2) { 4871 if (copy_from_user_timeval(&tv[0], arg2) 4872 || copy_from_user_timeval(&tv[1], 4873 arg2 + sizeof(struct target_timeval))) 4874 goto efault; 4875 tvp = tv; 4876 } else { 4877 tvp = NULL; 4878 } 4879 if (!(p = lock_user_string(arg1))) 4880 goto efault; 4881 ret = get_errno(utimes(p, tvp)); 4882 unlock_user(p, arg1, 0); 4883 } 4884 break; 4885 #if defined(TARGET_NR_futimesat) && defined(__NR_futimesat) 4886 case TARGET_NR_futimesat: 4887 { 4888 struct timeval *tvp, tv[2]; 4889 if (arg3) { 4890 if (copy_from_user_timeval(&tv[0], arg3) 4891 || copy_from_user_timeval(&tv[1], 4892 arg3 + sizeof(struct target_timeval))) 4893 goto efault; 4894 tvp = tv; 4895 } else { 4896 tvp = NULL; 4897 } 4898 if (!(p = lock_user_string(arg2))) 4899 goto efault; 4900 ret = get_errno(sys_futimesat(arg1, path(p), tvp)); 4901 unlock_user(p, arg2, 0); 4902 } 4903 break; 4904 #endif 4905 #ifdef TARGET_NR_stty 4906 case TARGET_NR_stty: 4907 goto unimplemented; 4908 #endif 4909 #ifdef TARGET_NR_gtty 4910 case TARGET_NR_gtty: 4911 goto unimplemented; 4912 #endif 4913 case TARGET_NR_access: 4914 if (!(p = lock_user_string(arg1))) 4915 goto efault; 4916 ret = get_errno(access(path(p), arg2)); 4917 unlock_user(p, arg1, 0); 4918 break; 4919 #if defined(TARGET_NR_faccessat) && defined(__NR_faccessat) 4920 case TARGET_NR_faccessat: 4921 if (!(p = lock_user_string(arg2))) 4922 goto efault; 4923 ret = get_errno(sys_faccessat(arg1, p, arg3)); 4924 unlock_user(p, arg2, 0); 4925 break; 4926 #endif 4927 #ifdef TARGET_NR_nice /* not on alpha */ 4928 case TARGET_NR_nice: 4929 ret = get_errno(nice(arg1)); 4930 break; 4931 #endif 4932 #ifdef TARGET_NR_ftime 4933 case TARGET_NR_ftime: 4934 goto unimplemented; 4935 #endif 4936 case TARGET_NR_sync: 4937 sync(); 4938 ret = 0; 4939 break; 4940 case TARGET_NR_kill: 4941 ret = get_errno(kill(arg1, target_to_host_signal(arg2))); 4942 break; 4943 case TARGET_NR_rename: 4944 { 4945 void *p2; 4946 p = lock_user_string(arg1); 4947 p2 = lock_user_string(arg2); 4948 if (!p || !p2) 4949 ret = -TARGET_EFAULT; 4950 else 4951 ret = get_errno(rename(p, p2)); 4952 unlock_user(p2, arg2, 0); 4953 unlock_user(p, arg1, 0); 4954 } 4955 break; 4956 #if defined(TARGET_NR_renameat) && defined(__NR_renameat) 4957 case TARGET_NR_renameat: 4958 { 4959 void *p2; 4960 p = lock_user_string(arg2); 4961 p2 = lock_user_string(arg4); 4962 if (!p || !p2) 4963 ret = -TARGET_EFAULT; 4964 else 4965 ret = get_errno(sys_renameat(arg1, p, arg3, p2)); 4966 unlock_user(p2, arg4, 0); 4967 unlock_user(p, arg2, 0); 4968 } 4969 break; 4970 #endif 4971 case TARGET_NR_mkdir: 4972 if (!(p = lock_user_string(arg1))) 4973 goto efault; 4974 ret = get_errno(mkdir(p, arg2)); 4975 unlock_user(p, arg1, 0); 4976 break; 4977 #if defined(TARGET_NR_mkdirat) && defined(__NR_mkdirat) 4978 case TARGET_NR_mkdirat: 4979 if (!(p = lock_user_string(arg2))) 4980 goto efault; 4981 ret = get_errno(sys_mkdirat(arg1, p, arg3)); 4982 unlock_user(p, arg2, 0); 4983 break; 4984 #endif 4985 case TARGET_NR_rmdir: 4986 if (!(p = lock_user_string(arg1))) 4987 goto efault; 4988 ret = get_errno(rmdir(p)); 4989 unlock_user(p, arg1, 0); 4990 break; 4991 case TARGET_NR_dup: 4992 ret = get_errno(dup(arg1)); 4993 break; 4994 case TARGET_NR_pipe: 4995 ret = do_pipe(cpu_env, arg1, 0, 0); 4996 break; 4997 #ifdef TARGET_NR_pipe2 4998 case TARGET_NR_pipe2: 4999 ret = do_pipe(cpu_env, arg1, arg2, 1); 5000 break; 5001 #endif 5002 case TARGET_NR_times: 5003 { 5004 struct target_tms *tmsp; 5005 struct tms tms; 5006 ret = get_errno(times(&tms)); 5007 if (arg1) { 5008 tmsp = lock_user(VERIFY_WRITE, arg1, sizeof(struct target_tms), 0); 5009 if (!tmsp) 5010 goto efault; 5011 tmsp->tms_utime = tswapl(host_to_target_clock_t(tms.tms_utime)); 5012 tmsp->tms_stime = tswapl(host_to_target_clock_t(tms.tms_stime)); 5013 tmsp->tms_cutime = tswapl(host_to_target_clock_t(tms.tms_cutime)); 5014 tmsp->tms_cstime = tswapl(host_to_target_clock_t(tms.tms_cstime)); 5015 } 5016 if (!is_error(ret)) 5017 ret = host_to_target_clock_t(ret); 5018 } 5019 break; 5020 #ifdef TARGET_NR_prof 5021 case TARGET_NR_prof: 5022 goto unimplemented; 5023 #endif 5024 #ifdef TARGET_NR_signal 5025 case TARGET_NR_signal: 5026 goto unimplemented; 5027 #endif 5028 case TARGET_NR_acct: 5029 if (arg1 == 0) { 5030 ret = get_errno(acct(NULL)); 5031 } else { 5032 if (!(p = lock_user_string(arg1))) 5033 goto efault; 5034 ret = get_errno(acct(path(p))); 5035 unlock_user(p, arg1, 0); 5036 } 5037 break; 5038 #ifdef TARGET_NR_umount2 /* not on alpha */ 5039 case TARGET_NR_umount2: 5040 if (!(p = lock_user_string(arg1))) 5041 goto efault; 5042 ret = get_errno(umount2(p, arg2)); 5043 unlock_user(p, arg1, 0); 5044 break; 5045 #endif 5046 #ifdef TARGET_NR_lock 5047 case TARGET_NR_lock: 5048 goto unimplemented; 5049 #endif 5050 case TARGET_NR_ioctl: 5051 ret = do_ioctl(arg1, arg2, arg3); 5052 break; 5053 case TARGET_NR_fcntl: 5054 ret = do_fcntl(arg1, arg2, arg3); 5055 break; 5056 #ifdef TARGET_NR_mpx 5057 case TARGET_NR_mpx: 5058 goto unimplemented; 5059 #endif 5060 case TARGET_NR_setpgid: 5061 ret = get_errno(setpgid(arg1, arg2)); 5062 break; 5063 #ifdef TARGET_NR_ulimit 5064 case TARGET_NR_ulimit: 5065 goto unimplemented; 5066 #endif 5067 #ifdef TARGET_NR_oldolduname 5068 case TARGET_NR_oldolduname: 5069 goto unimplemented; 5070 #endif 5071 case TARGET_NR_umask: 5072 ret = get_errno(umask(arg1)); 5073 break; 5074 case TARGET_NR_chroot: 5075 if (!(p = lock_user_string(arg1))) 5076 goto efault; 5077 ret = get_errno(chroot(p)); 5078 unlock_user(p, arg1, 0); 5079 break; 5080 case TARGET_NR_ustat: 5081 goto unimplemented; 5082 case TARGET_NR_dup2: 5083 ret = get_errno(dup2(arg1, arg2)); 5084 break; 5085 #if defined(CONFIG_DUP3) && defined(TARGET_NR_dup3) 5086 case TARGET_NR_dup3: 5087 ret = get_errno(dup3(arg1, arg2, arg3)); 5088 break; 5089 #endif 5090 #ifdef TARGET_NR_getppid /* not on alpha */ 5091 case TARGET_NR_getppid: 5092 ret = get_errno(getppid()); 5093 break; 5094 #endif 5095 case TARGET_NR_getpgrp: 5096 ret = get_errno(getpgrp()); 5097 break; 5098 case TARGET_NR_setsid: 5099 ret = get_errno(setsid()); 5100 break; 5101 #ifdef TARGET_NR_sigaction 5102 case TARGET_NR_sigaction: 5103 { 5104 #if defined(TARGET_ALPHA) 5105 struct target_sigaction act, oact, *pact = 0; 5106 struct target_old_sigaction *old_act; 5107 if (arg2) { 5108 if (!lock_user_struct(VERIFY_READ, old_act, arg2, 1)) 5109 goto efault; 5110 act._sa_handler = old_act->_sa_handler; 5111 target_siginitset(&act.sa_mask, old_act->sa_mask); 5112 act.sa_flags = old_act->sa_flags; 5113 act.sa_restorer = 0; 5114 unlock_user_struct(old_act, arg2, 0); 5115 pact = &act; 5116 } 5117 ret = get_errno(do_sigaction(arg1, pact, &oact)); 5118 if (!is_error(ret) && arg3) { 5119 if (!lock_user_struct(VERIFY_WRITE, old_act, arg3, 0)) 5120 goto efault; 5121 old_act->_sa_handler = oact._sa_handler; 5122 old_act->sa_mask = oact.sa_mask.sig[0]; 5123 old_act->sa_flags = oact.sa_flags; 5124 unlock_user_struct(old_act, arg3, 1); 5125 } 5126 #elif defined(TARGET_MIPS) 5127 struct target_sigaction act, oact, *pact, *old_act; 5128 5129 if (arg2) { 5130 if (!lock_user_struct(VERIFY_READ, old_act, arg2, 1)) 5131 goto efault; 5132 act._sa_handler = old_act->_sa_handler; 5133 target_siginitset(&act.sa_mask, old_act->sa_mask.sig[0]); 5134 act.sa_flags = old_act->sa_flags; 5135 unlock_user_struct(old_act, arg2, 0); 5136 pact = &act; 5137 } else { 5138 pact = NULL; 5139 } 5140 5141 ret = get_errno(do_sigaction(arg1, pact, &oact)); 5142 5143 if (!is_error(ret) && arg3) { 5144 if (!lock_user_struct(VERIFY_WRITE, old_act, arg3, 0)) 5145 goto efault; 5146 old_act->_sa_handler = oact._sa_handler; 5147 old_act->sa_flags = oact.sa_flags; 5148 old_act->sa_mask.sig[0] = oact.sa_mask.sig[0]; 5149 old_act->sa_mask.sig[1] = 0; 5150 old_act->sa_mask.sig[2] = 0; 5151 old_act->sa_mask.sig[3] = 0; 5152 unlock_user_struct(old_act, arg3, 1); 5153 } 5154 #else 5155 struct target_old_sigaction *old_act; 5156 struct target_sigaction act, oact, *pact; 5157 if (arg2) { 5158 if (!lock_user_struct(VERIFY_READ, old_act, arg2, 1)) 5159 goto efault; 5160 act._sa_handler = old_act->_sa_handler; 5161 target_siginitset(&act.sa_mask, old_act->sa_mask); 5162 act.sa_flags = old_act->sa_flags; 5163 act.sa_restorer = old_act->sa_restorer; 5164 unlock_user_struct(old_act, arg2, 0); 5165 pact = &act; 5166 } else { 5167 pact = NULL; 5168 } 5169 ret = get_errno(do_sigaction(arg1, pact, &oact)); 5170 if (!is_error(ret) && arg3) { 5171 if (!lock_user_struct(VERIFY_WRITE, old_act, arg3, 0)) 5172 goto efault; 5173 old_act->_sa_handler = oact._sa_handler; 5174 old_act->sa_mask = oact.sa_mask.sig[0]; 5175 old_act->sa_flags = oact.sa_flags; 5176 old_act->sa_restorer = oact.sa_restorer; 5177 unlock_user_struct(old_act, arg3, 1); 5178 } 5179 #endif 5180 } 5181 break; 5182 #endif 5183 case TARGET_NR_rt_sigaction: 5184 { 5185 #if defined(TARGET_ALPHA) 5186 struct target_sigaction act, oact, *pact = 0; 5187 struct target_rt_sigaction *rt_act; 5188 /* ??? arg4 == sizeof(sigset_t). */ 5189 if (arg2) { 5190 if (!lock_user_struct(VERIFY_READ, rt_act, arg2, 1)) 5191 goto efault; 5192 act._sa_handler = rt_act->_sa_handler; 5193 act.sa_mask = rt_act->sa_mask; 5194 act.sa_flags = rt_act->sa_flags; 5195 act.sa_restorer = arg5; 5196 unlock_user_struct(rt_act, arg2, 0); 5197 pact = &act; 5198 } 5199 ret = get_errno(do_sigaction(arg1, pact, &oact)); 5200 if (!is_error(ret) && arg3) { 5201 if (!lock_user_struct(VERIFY_WRITE, rt_act, arg3, 0)) 5202 goto efault; 5203 rt_act->_sa_handler = oact._sa_handler; 5204 rt_act->sa_mask = oact.sa_mask; 5205 rt_act->sa_flags = oact.sa_flags; 5206 unlock_user_struct(rt_act, arg3, 1); 5207 } 5208 #else 5209 struct target_sigaction *act; 5210 struct target_sigaction *oact; 5211 5212 if (arg2) { 5213 if (!lock_user_struct(VERIFY_READ, act, arg2, 1)) 5214 goto efault; 5215 } else 5216 act = NULL; 5217 if (arg3) { 5218 if (!lock_user_struct(VERIFY_WRITE, oact, arg3, 0)) { 5219 ret = -TARGET_EFAULT; 5220 goto rt_sigaction_fail; 5221 } 5222 } else 5223 oact = NULL; 5224 ret = get_errno(do_sigaction(arg1, act, oact)); 5225 rt_sigaction_fail: 5226 if (act) 5227 unlock_user_struct(act, arg2, 0); 5228 if (oact) 5229 unlock_user_struct(oact, arg3, 1); 5230 #endif 5231 } 5232 break; 5233 #ifdef TARGET_NR_sgetmask /* not on alpha */ 5234 case TARGET_NR_sgetmask: 5235 { 5236 sigset_t cur_set; 5237 abi_ulong target_set; 5238 sigprocmask(0, NULL, &cur_set); 5239 host_to_target_old_sigset(&target_set, &cur_set); 5240 ret = target_set; 5241 } 5242 break; 5243 #endif 5244 #ifdef TARGET_NR_ssetmask /* not on alpha */ 5245 case TARGET_NR_ssetmask: 5246 { 5247 sigset_t set, oset, cur_set; 5248 abi_ulong target_set = arg1; 5249 sigprocmask(0, NULL, &cur_set); 5250 target_to_host_old_sigset(&set, &target_set); 5251 sigorset(&set, &set, &cur_set); 5252 sigprocmask(SIG_SETMASK, &set, &oset); 5253 host_to_target_old_sigset(&target_set, &oset); 5254 ret = target_set; 5255 } 5256 break; 5257 #endif 5258 #ifdef TARGET_NR_sigprocmask 5259 case TARGET_NR_sigprocmask: 5260 { 5261 #if defined(TARGET_ALPHA) 5262 sigset_t set, oldset; 5263 abi_ulong mask; 5264 int how; 5265 5266 switch (arg1) { 5267 case TARGET_SIG_BLOCK: 5268 how = SIG_BLOCK; 5269 break; 5270 case TARGET_SIG_UNBLOCK: 5271 how = SIG_UNBLOCK; 5272 break; 5273 case TARGET_SIG_SETMASK: 5274 how = SIG_SETMASK; 5275 break; 5276 default: 5277 ret = -TARGET_EINVAL; 5278 goto fail; 5279 } 5280 mask = arg2; 5281 target_to_host_old_sigset(&set, &mask); 5282 5283 ret = get_errno(sigprocmask(how, &set, &oldset)); 5284 5285 if (!is_error(ret)) { 5286 host_to_target_old_sigset(&mask, &oldset); 5287 ret = mask; 5288 ((CPUAlphaState *)cpu_env)->[IR_V0] = 0; /* force no error */ 5289 } 5290 #else 5291 sigset_t set, oldset, *set_ptr; 5292 int how; 5293 5294 if (arg2) { 5295 switch (arg1) { 5296 case TARGET_SIG_BLOCK: 5297 how = SIG_BLOCK; 5298 break; 5299 case TARGET_SIG_UNBLOCK: 5300 how = SIG_UNBLOCK; 5301 break; 5302 case TARGET_SIG_SETMASK: 5303 how = SIG_SETMASK; 5304 break; 5305 default: 5306 ret = -TARGET_EINVAL; 5307 goto fail; 5308 } 5309 if (!(p = lock_user(VERIFY_READ, arg2, sizeof(target_sigset_t), 1))) 5310 goto efault; 5311 target_to_host_old_sigset(&set, p); 5312 unlock_user(p, arg2, 0); 5313 set_ptr = &set; 5314 } else { 5315 how = 0; 5316 set_ptr = NULL; 5317 } 5318 ret = get_errno(sigprocmask(how, set_ptr, &oldset)); 5319 if (!is_error(ret) && arg3) { 5320 if (!(p = lock_user(VERIFY_WRITE, arg3, sizeof(target_sigset_t), 0))) 5321 goto efault; 5322 host_to_target_old_sigset(p, &oldset); 5323 unlock_user(p, arg3, sizeof(target_sigset_t)); 5324 } 5325 #endif 5326 } 5327 break; 5328 #endif 5329 case TARGET_NR_rt_sigprocmask: 5330 { 5331 int how = arg1; 5332 sigset_t set, oldset, *set_ptr; 5333 5334 if (arg2) { 5335 switch(how) { 5336 case TARGET_SIG_BLOCK: 5337 how = SIG_BLOCK; 5338 break; 5339 case TARGET_SIG_UNBLOCK: 5340 how = SIG_UNBLOCK; 5341 break; 5342 case TARGET_SIG_SETMASK: 5343 how = SIG_SETMASK; 5344 break; 5345 default: 5346 ret = -TARGET_EINVAL; 5347 goto fail; 5348 } 5349 if (!(p = lock_user(VERIFY_READ, arg2, sizeof(target_sigset_t), 1))) 5350 goto efault; 5351 target_to_host_sigset(&set, p); 5352 unlock_user(p, arg2, 0); 5353 set_ptr = &set; 5354 } else { 5355 how = 0; 5356 set_ptr = NULL; 5357 } 5358 ret = get_errno(sigprocmask(how, set_ptr, &oldset)); 5359 if (!is_error(ret) && arg3) { 5360 if (!(p = lock_user(VERIFY_WRITE, arg3, sizeof(target_sigset_t), 0))) 5361 goto efault; 5362 host_to_target_sigset(p, &oldset); 5363 unlock_user(p, arg3, sizeof(target_sigset_t)); 5364 } 5365 } 5366 break; 5367 #ifdef TARGET_NR_sigpending 5368 case TARGET_NR_sigpending: 5369 { 5370 sigset_t set; 5371 ret = get_errno(sigpending(&set)); 5372 if (!is_error(ret)) { 5373 if (!(p = lock_user(VERIFY_WRITE, arg1, sizeof(target_sigset_t), 0))) 5374 goto efault; 5375 host_to_target_old_sigset(p, &set); 5376 unlock_user(p, arg1, sizeof(target_sigset_t)); 5377 } 5378 } 5379 break; 5380 #endif 5381 case TARGET_NR_rt_sigpending: 5382 { 5383 sigset_t set; 5384 ret = get_errno(sigpending(&set)); 5385 if (!is_error(ret)) { 5386 if (!(p = lock_user(VERIFY_WRITE, arg1, sizeof(target_sigset_t), 0))) 5387 goto efault; 5388 host_to_target_sigset(p, &set); 5389 unlock_user(p, arg1, sizeof(target_sigset_t)); 5390 } 5391 } 5392 break; 5393 #ifdef TARGET_NR_sigsuspend 5394 case TARGET_NR_sigsuspend: 5395 { 5396 sigset_t set; 5397 #if defined(TARGET_ALPHA) 5398 abi_ulong mask = arg1; 5399 target_to_host_old_sigset(&set, &mask); 5400 #else 5401 if (!(p = lock_user(VERIFY_READ, arg1, sizeof(target_sigset_t), 1))) 5402 goto efault; 5403 target_to_host_old_sigset(&set, p); 5404 unlock_user(p, arg1, 0); 5405 #endif 5406 ret = get_errno(sigsuspend(&set)); 5407 } 5408 break; 5409 #endif 5410 case TARGET_NR_rt_sigsuspend: 5411 { 5412 sigset_t set; 5413 if (!(p = lock_user(VERIFY_READ, arg1, sizeof(target_sigset_t), 1))) 5414 goto efault; 5415 target_to_host_sigset(&set, p); 5416 unlock_user(p, arg1, 0); 5417 ret = get_errno(sigsuspend(&set)); 5418 } 5419 break; 5420 case TARGET_NR_rt_sigtimedwait: 5421 { 5422 sigset_t set; 5423 struct timespec uts, *puts; 5424 siginfo_t uinfo; 5425 5426 if (!(p = lock_user(VERIFY_READ, arg1, sizeof(target_sigset_t), 1))) 5427 goto efault; 5428 target_to_host_sigset(&set, p); 5429 unlock_user(p, arg1, 0); 5430 if (arg3) { 5431 puts = &uts; 5432 target_to_host_timespec(puts, arg3); 5433 } else { 5434 puts = NULL; 5435 } 5436 ret = get_errno(sigtimedwait(&set, &uinfo, puts)); 5437 if (!is_error(ret) && arg2) { 5438 if (!(p = lock_user(VERIFY_WRITE, arg2, sizeof(target_siginfo_t), 0))) 5439 goto efault; 5440 host_to_target_siginfo(p, &uinfo); 5441 unlock_user(p, arg2, sizeof(target_siginfo_t)); 5442 } 5443 } 5444 break; 5445 case TARGET_NR_rt_sigqueueinfo: 5446 { 5447 siginfo_t uinfo; 5448 if (!(p = lock_user(VERIFY_READ, arg3, sizeof(target_sigset_t), 1))) 5449 goto efault; 5450 target_to_host_siginfo(&uinfo, p); 5451 unlock_user(p, arg1, 0); 5452 ret = get_errno(sys_rt_sigqueueinfo(arg1, arg2, &uinfo)); 5453 } 5454 break; 5455 #ifdef TARGET_NR_sigreturn 5456 case TARGET_NR_sigreturn: 5457 /* NOTE: ret is eax, so not transcoding must be done */ 5458 ret = do_sigreturn(cpu_env); 5459 break; 5460 #endif 5461 case TARGET_NR_rt_sigreturn: 5462 /* NOTE: ret is eax, so not transcoding must be done */ 5463 ret = do_rt_sigreturn(cpu_env); 5464 break; 5465 case TARGET_NR_sethostname: 5466 if (!(p = lock_user_string(arg1))) 5467 goto efault; 5468 ret = get_errno(sethostname(p, arg2)); 5469 unlock_user(p, arg1, 0); 5470 break; 5471 case TARGET_NR_setrlimit: 5472 { 5473 int resource = arg1; 5474 struct target_rlimit *target_rlim; 5475 struct rlimit rlim; 5476 if (!lock_user_struct(VERIFY_READ, target_rlim, arg2, 1)) 5477 goto efault; 5478 rlim.rlim_cur = target_to_host_rlim(target_rlim->rlim_cur); 5479 rlim.rlim_max = target_to_host_rlim(target_rlim->rlim_max); 5480 unlock_user_struct(target_rlim, arg2, 0); 5481 ret = get_errno(setrlimit(resource, &rlim)); 5482 } 5483 break; 5484 case TARGET_NR_getrlimit: 5485 { 5486 int resource = arg1; 5487 struct target_rlimit *target_rlim; 5488 struct rlimit rlim; 5489 5490 ret = get_errno(getrlimit(resource, &rlim)); 5491 if (!is_error(ret)) { 5492 if (!lock_user_struct(VERIFY_WRITE, target_rlim, arg2, 0)) 5493 goto efault; 5494 target_rlim->rlim_cur = host_to_target_rlim(rlim.rlim_cur); 5495 target_rlim->rlim_max = host_to_target_rlim(rlim.rlim_max); 5496 unlock_user_struct(target_rlim, arg2, 1); 5497 } 5498 } 5499 break; 5500 case TARGET_NR_getrusage: 5501 { 5502 struct rusage rusage; 5503 ret = get_errno(getrusage(arg1, &rusage)); 5504 if (!is_error(ret)) { 5505 host_to_target_rusage(arg2, &rusage); 5506 } 5507 } 5508 break; 5509 case TARGET_NR_gettimeofday: 5510 { 5511 struct timeval tv; 5512 ret = get_errno(gettimeofday(&tv, NULL)); 5513 if (!is_error(ret)) { 5514 if (copy_to_user_timeval(arg1, &tv)) 5515 goto efault; 5516 } 5517 } 5518 break; 5519 case TARGET_NR_settimeofday: 5520 { 5521 struct timeval tv; 5522 if (copy_from_user_timeval(&tv, arg1)) 5523 goto efault; 5524 ret = get_errno(settimeofday(&tv, NULL)); 5525 } 5526 break; 5527 #ifdef TARGET_NR_select 5528 case TARGET_NR_select: 5529 { 5530 struct target_sel_arg_struct *sel; 5531 abi_ulong inp, outp, exp, tvp; 5532 long nsel; 5533 5534 if (!lock_user_struct(VERIFY_READ, sel, arg1, 1)) 5535 goto efault; 5536 nsel = tswapl(sel->n); 5537 inp = tswapl(sel->inp); 5538 outp = tswapl(sel->outp); 5539 exp = tswapl(sel->exp); 5540 tvp = tswapl(sel->tvp); 5541 unlock_user_struct(sel, arg1, 0); 5542 ret = do_select(nsel, inp, outp, exp, tvp); 5543 } 5544 break; 5545 #endif 5546 #ifdef TARGET_NR_pselect6 5547 case TARGET_NR_pselect6: 5548 goto unimplemented_nowarn; 5549 #endif 5550 case TARGET_NR_symlink: 5551 { 5552 void *p2; 5553 p = lock_user_string(arg1); 5554 p2 = lock_user_string(arg2); 5555 if (!p || !p2) 5556 ret = -TARGET_EFAULT; 5557 else 5558 ret = get_errno(symlink(p, p2)); 5559 unlock_user(p2, arg2, 0); 5560 unlock_user(p, arg1, 0); 5561 } 5562 break; 5563 #if defined(TARGET_NR_symlinkat) && defined(__NR_symlinkat) 5564 case TARGET_NR_symlinkat: 5565 { 5566 void *p2; 5567 p = lock_user_string(arg1); 5568 p2 = lock_user_string(arg3); 5569 if (!p || !p2) 5570 ret = -TARGET_EFAULT; 5571 else 5572 ret = get_errno(sys_symlinkat(p, arg2, p2)); 5573 unlock_user(p2, arg3, 0); 5574 unlock_user(p, arg1, 0); 5575 } 5576 break; 5577 #endif 5578 #ifdef TARGET_NR_oldlstat 5579 case TARGET_NR_oldlstat: 5580 goto unimplemented; 5581 #endif 5582 case TARGET_NR_readlink: 5583 { 5584 void *p2, *temp; 5585 p = lock_user_string(arg1); 5586 p2 = lock_user(VERIFY_WRITE, arg2, arg3, 0); 5587 if (!p || !p2) 5588 ret = -TARGET_EFAULT; 5589 else { 5590 if (strncmp((const char *)p, "/proc/self/exe", 14) == 0) { 5591 char real[PATH_MAX]; 5592 temp = realpath(exec_path,real); 5593 ret = (temp==NULL) ? get_errno(-1) : strlen(real) ; 5594 snprintf((char *)p2, arg3, "%s", real); 5595 } 5596 else 5597 ret = get_errno(readlink(path(p), p2, arg3)); 5598 } 5599 unlock_user(p2, arg2, ret); 5600 unlock_user(p, arg1, 0); 5601 } 5602 break; 5603 #if defined(TARGET_NR_readlinkat) && defined(__NR_readlinkat) 5604 case TARGET_NR_readlinkat: 5605 { 5606 void *p2; 5607 p = lock_user_string(arg2); 5608 p2 = lock_user(VERIFY_WRITE, arg3, arg4, 0); 5609 if (!p || !p2) 5610 ret = -TARGET_EFAULT; 5611 else 5612 ret = get_errno(sys_readlinkat(arg1, path(p), p2, arg4)); 5613 unlock_user(p2, arg3, ret); 5614 unlock_user(p, arg2, 0); 5615 } 5616 break; 5617 #endif 5618 #ifdef TARGET_NR_uselib 5619 case TARGET_NR_uselib: 5620 goto unimplemented; 5621 #endif 5622 #ifdef TARGET_NR_swapon 5623 case TARGET_NR_swapon: 5624 if (!(p = lock_user_string(arg1))) 5625 goto efault; 5626 ret = get_errno(swapon(p, arg2)); 5627 unlock_user(p, arg1, 0); 5628 break; 5629 #endif 5630 case TARGET_NR_reboot: 5631 goto unimplemented; 5632 #ifdef TARGET_NR_readdir 5633 case TARGET_NR_readdir: 5634 goto unimplemented; 5635 #endif 5636 #ifdef TARGET_NR_mmap 5637 case TARGET_NR_mmap: 5638 #if (defined(TARGET_I386) && defined(TARGET_ABI32)) || defined(TARGET_ARM) || defined(TARGET_M68K) || defined(TARGET_CRIS) || defined(TARGET_MICROBLAZE) 5639 { 5640 abi_ulong *v; 5641 abi_ulong v1, v2, v3, v4, v5, v6; 5642 if (!(v = lock_user(VERIFY_READ, arg1, 6 * sizeof(abi_ulong), 1))) 5643 goto efault; 5644 v1 = tswapl(v[0]); 5645 v2 = tswapl(v[1]); 5646 v3 = tswapl(v[2]); 5647 v4 = tswapl(v[3]); 5648 v5 = tswapl(v[4]); 5649 v6 = tswapl(v[5]); 5650 unlock_user(v, arg1, 0); 5651 ret = get_errno(target_mmap(v1, v2, v3, 5652 target_to_host_bitmask(v4, mmap_flags_tbl), 5653 v5, v6)); 5654 } 5655 #else 5656 ret = get_errno(target_mmap(arg1, arg2, arg3, 5657 target_to_host_bitmask(arg4, mmap_flags_tbl), 5658 arg5, 5659 arg6)); 5660 #endif 5661 break; 5662 #endif 5663 #ifdef TARGET_NR_mmap2 5664 case TARGET_NR_mmap2: 5665 #ifndef MMAP_SHIFT 5666 #define MMAP_SHIFT 12 5667 #endif 5668 ret = get_errno(target_mmap(arg1, arg2, arg3, 5669 target_to_host_bitmask(arg4, mmap_flags_tbl), 5670 arg5, 5671 arg6 << MMAP_SHIFT)); 5672 break; 5673 #endif 5674 case TARGET_NR_munmap: 5675 ret = get_errno(target_munmap(arg1, arg2)); 5676 break; 5677 case TARGET_NR_mprotect: 5678 { 5679 TaskState *ts = ((CPUState *)cpu_env)->opaque; 5680 /* Special hack to detect libc making the stack executable. */ 5681 if ((arg3 & PROT_GROWSDOWN) 5682 && arg1 >= ts->info->stack_limit 5683 && arg1 <= ts->info->start_stack) { 5684 arg3 &= ~PROT_GROWSDOWN; 5685 arg2 = arg2 + arg1 - ts->info->stack_limit; 5686 arg1 = ts->info->stack_limit; 5687 } 5688 } 5689 ret = get_errno(target_mprotect(arg1, arg2, arg3)); 5690 break; 5691 #ifdef TARGET_NR_mremap 5692 case TARGET_NR_mremap: 5693 ret = get_errno(target_mremap(arg1, arg2, arg3, arg4, arg5)); 5694 break; 5695 #endif 5696 /* ??? msync/mlock/munlock are broken for softmmu. */ 5697 #ifdef TARGET_NR_msync 5698 case TARGET_NR_msync: 5699 ret = get_errno(msync(g2h(arg1), arg2, arg3)); 5700 break; 5701 #endif 5702 #ifdef TARGET_NR_mlock 5703 case TARGET_NR_mlock: 5704 ret = get_errno(mlock(g2h(arg1), arg2)); 5705 break; 5706 #endif 5707 #ifdef TARGET_NR_munlock 5708 case TARGET_NR_munlock: 5709 ret = get_errno(munlock(g2h(arg1), arg2)); 5710 break; 5711 #endif 5712 #ifdef TARGET_NR_mlockall 5713 case TARGET_NR_mlockall: 5714 ret = get_errno(mlockall(arg1)); 5715 break; 5716 #endif 5717 #ifdef TARGET_NR_munlockall 5718 case TARGET_NR_munlockall: 5719 ret = get_errno(munlockall()); 5720 break; 5721 #endif 5722 case TARGET_NR_truncate: 5723 if (!(p = lock_user_string(arg1))) 5724 goto efault; 5725 ret = get_errno(truncate(p, arg2)); 5726 unlock_user(p, arg1, 0); 5727 break; 5728 case TARGET_NR_ftruncate: 5729 ret = get_errno(ftruncate(arg1, arg2)); 5730 break; 5731 case TARGET_NR_fchmod: 5732 ret = get_errno(fchmod(arg1, arg2)); 5733 break; 5734 #if defined(TARGET_NR_fchmodat) && defined(__NR_fchmodat) 5735 case TARGET_NR_fchmodat: 5736 if (!(p = lock_user_string(arg2))) 5737 goto efault; 5738 ret = get_errno(sys_fchmodat(arg1, p, arg3)); 5739 unlock_user(p, arg2, 0); 5740 break; 5741 #endif 5742 case TARGET_NR_getpriority: 5743 /* libc does special remapping of the return value of 5744 * sys_getpriority() so it's just easiest to call 5745 * sys_getpriority() directly rather than through libc. */ 5746 ret = get_errno(sys_getpriority(arg1, arg2)); 5747 break; 5748 case TARGET_NR_setpriority: 5749 ret = get_errno(setpriority(arg1, arg2, arg3)); 5750 break; 5751 #ifdef TARGET_NR_profil 5752 case TARGET_NR_profil: 5753 goto unimplemented; 5754 #endif 5755 case TARGET_NR_statfs: 5756 if (!(p = lock_user_string(arg1))) 5757 goto efault; 5758 ret = get_errno(statfs(path(p), &stfs)); 5759 unlock_user(p, arg1, 0); 5760 convert_statfs: 5761 if (!is_error(ret)) { 5762 struct target_statfs *target_stfs; 5763 5764 if (!lock_user_struct(VERIFY_WRITE, target_stfs, arg2, 0)) 5765 goto efault; 5766 __put_user(stfs.f_type, &target_stfs->f_type); 5767 __put_user(stfs.f_bsize, &target_stfs->f_bsize); 5768 __put_user(stfs.f_blocks, &target_stfs->f_blocks); 5769 __put_user(stfs.f_bfree, &target_stfs->f_bfree); 5770 __put_user(stfs.f_bavail, &target_stfs->f_bavail); 5771 __put_user(stfs.f_files, &target_stfs->f_files); 5772 __put_user(stfs.f_ffree, &target_stfs->f_ffree); 5773 __put_user(stfs.f_fsid.__val[0], &target_stfs->f_fsid.val[0]); 5774 __put_user(stfs.f_fsid.__val[1], &target_stfs->f_fsid.val[1]); 5775 __put_user(stfs.f_namelen, &target_stfs->f_namelen); 5776 unlock_user_struct(target_stfs, arg2, 1); 5777 } 5778 break; 5779 case TARGET_NR_fstatfs: 5780 ret = get_errno(fstatfs(arg1, &stfs)); 5781 goto convert_statfs; 5782 #ifdef TARGET_NR_statfs64 5783 case TARGET_NR_statfs64: 5784 if (!(p = lock_user_string(arg1))) 5785 goto efault; 5786 ret = get_errno(statfs(path(p), &stfs)); 5787 unlock_user(p, arg1, 0); 5788 convert_statfs64: 5789 if (!is_error(ret)) { 5790 struct target_statfs64 *target_stfs; 5791 5792 if (!lock_user_struct(VERIFY_WRITE, target_stfs, arg3, 0)) 5793 goto efault; 5794 __put_user(stfs.f_type, &target_stfs->f_type); 5795 __put_user(stfs.f_bsize, &target_stfs->f_bsize); 5796 __put_user(stfs.f_blocks, &target_stfs->f_blocks); 5797 __put_user(stfs.f_bfree, &target_stfs->f_bfree); 5798 __put_user(stfs.f_bavail, &target_stfs->f_bavail); 5799 __put_user(stfs.f_files, &target_stfs->f_files); 5800 __put_user(stfs.f_ffree, &target_stfs->f_ffree); 5801 __put_user(stfs.f_fsid.__val[0], &target_stfs->f_fsid.val[0]); 5802 __put_user(stfs.f_fsid.__val[1], &target_stfs->f_fsid.val[1]); 5803 __put_user(stfs.f_namelen, &target_stfs->f_namelen); 5804 unlock_user_struct(target_stfs, arg3, 1); 5805 } 5806 break; 5807 case TARGET_NR_fstatfs64: 5808 ret = get_errno(fstatfs(arg1, &stfs)); 5809 goto convert_statfs64; 5810 #endif 5811 #ifdef TARGET_NR_ioperm 5812 case TARGET_NR_ioperm: 5813 goto unimplemented; 5814 #endif 5815 #ifdef TARGET_NR_socketcall 5816 case TARGET_NR_socketcall: 5817 ret = do_socketcall(arg1, arg2); 5818 break; 5819 #endif 5820 #ifdef TARGET_NR_accept 5821 case TARGET_NR_accept: 5822 ret = do_accept(arg1, arg2, arg3); 5823 break; 5824 #endif 5825 #ifdef TARGET_NR_bind 5826 case TARGET_NR_bind: 5827 ret = do_bind(arg1, arg2, arg3); 5828 break; 5829 #endif 5830 #ifdef TARGET_NR_connect 5831 case TARGET_NR_connect: 5832 ret = do_connect(arg1, arg2, arg3); 5833 break; 5834 #endif 5835 #ifdef TARGET_NR_getpeername 5836 case TARGET_NR_getpeername: 5837 ret = do_getpeername(arg1, arg2, arg3); 5838 break; 5839 #endif 5840 #ifdef TARGET_NR_getsockname 5841 case TARGET_NR_getsockname: 5842 ret = do_getsockname(arg1, arg2, arg3); 5843 break; 5844 #endif 5845 #ifdef TARGET_NR_getsockopt 5846 case TARGET_NR_getsockopt: 5847 ret = do_getsockopt(arg1, arg2, arg3, arg4, arg5); 5848 break; 5849 #endif 5850 #ifdef TARGET_NR_listen 5851 case TARGET_NR_listen: 5852 ret = get_errno(listen(arg1, arg2)); 5853 break; 5854 #endif 5855 #ifdef TARGET_NR_recv 5856 case TARGET_NR_recv: 5857 ret = do_recvfrom(arg1, arg2, arg3, arg4, 0, 0); 5858 break; 5859 #endif 5860 #ifdef TARGET_NR_recvfrom 5861 case TARGET_NR_recvfrom: 5862 ret = do_recvfrom(arg1, arg2, arg3, arg4, arg5, arg6); 5863 break; 5864 #endif 5865 #ifdef TARGET_NR_recvmsg 5866 case TARGET_NR_recvmsg: 5867 ret = do_sendrecvmsg(arg1, arg2, arg3, 0); 5868 break; 5869 #endif 5870 #ifdef TARGET_NR_send 5871 case TARGET_NR_send: 5872 ret = do_sendto(arg1, arg2, arg3, arg4, 0, 0); 5873 break; 5874 #endif 5875 #ifdef TARGET_NR_sendmsg 5876 case TARGET_NR_sendmsg: 5877 ret = do_sendrecvmsg(arg1, arg2, arg3, 1); 5878 break; 5879 #endif 5880 #ifdef TARGET_NR_sendto 5881 case TARGET_NR_sendto: 5882 ret = do_sendto(arg1, arg2, arg3, arg4, arg5, arg6); 5883 break; 5884 #endif 5885 #ifdef TARGET_NR_shutdown 5886 case TARGET_NR_shutdown: 5887 ret = get_errno(shutdown(arg1, arg2)); 5888 break; 5889 #endif 5890 #ifdef TARGET_NR_socket 5891 case TARGET_NR_socket: 5892 ret = do_socket(arg1, arg2, arg3); 5893 break; 5894 #endif 5895 #ifdef TARGET_NR_socketpair 5896 case TARGET_NR_socketpair: 5897 ret = do_socketpair(arg1, arg2, arg3, arg4); 5898 break; 5899 #endif 5900 #ifdef TARGET_NR_setsockopt 5901 case TARGET_NR_setsockopt: 5902 ret = do_setsockopt(arg1, arg2, arg3, arg4, (socklen_t) arg5); 5903 break; 5904 #endif 5905 5906 case TARGET_NR_syslog: 5907 if (!(p = lock_user_string(arg2))) 5908 goto efault; 5909 ret = get_errno(sys_syslog((int)arg1, p, (int)arg3)); 5910 unlock_user(p, arg2, 0); 5911 break; 5912 5913 case TARGET_NR_setitimer: 5914 { 5915 struct itimerval value, ovalue, *pvalue; 5916 5917 if (arg2) { 5918 pvalue = &value; 5919 if (copy_from_user_timeval(&pvalue->it_interval, arg2) 5920 || copy_from_user_timeval(&pvalue->it_value, 5921 arg2 + sizeof(struct target_timeval))) 5922 goto efault; 5923 } else { 5924 pvalue = NULL; 5925 } 5926 ret = get_errno(setitimer(arg1, pvalue, &ovalue)); 5927 if (!is_error(ret) && arg3) { 5928 if (copy_to_user_timeval(arg3, 5929 &ovalue.it_interval) 5930 || copy_to_user_timeval(arg3 + sizeof(struct target_timeval), 5931 &ovalue.it_value)) 5932 goto efault; 5933 } 5934 } 5935 break; 5936 case TARGET_NR_getitimer: 5937 { 5938 struct itimerval value; 5939 5940 ret = get_errno(getitimer(arg1, &value)); 5941 if (!is_error(ret) && arg2) { 5942 if (copy_to_user_timeval(arg2, 5943 &value.it_interval) 5944 || copy_to_user_timeval(arg2 + sizeof(struct target_timeval), 5945 &value.it_value)) 5946 goto efault; 5947 } 5948 } 5949 break; 5950 case TARGET_NR_stat: 5951 if (!(p = lock_user_string(arg1))) 5952 goto efault; 5953 ret = get_errno(stat(path(p), &st)); 5954 unlock_user(p, arg1, 0); 5955 goto do_stat; 5956 case TARGET_NR_lstat: 5957 if (!(p = lock_user_string(arg1))) 5958 goto efault; 5959 ret = get_errno(lstat(path(p), &st)); 5960 unlock_user(p, arg1, 0); 5961 goto do_stat; 5962 case TARGET_NR_fstat: 5963 { 5964 ret = get_errno(fstat(arg1, &st)); 5965 do_stat: 5966 if (!is_error(ret)) { 5967 struct target_stat *target_st; 5968 5969 if (!lock_user_struct(VERIFY_WRITE, target_st, arg2, 0)) 5970 goto efault; 5971 memset(target_st, 0, sizeof(*target_st)); 5972 __put_user(st.st_dev, &target_st->st_dev); 5973 __put_user(st.st_ino, &target_st->st_ino); 5974 __put_user(st.st_mode, &target_st->st_mode); 5975 __put_user(st.st_uid, &target_st->st_uid); 5976 __put_user(st.st_gid, &target_st->st_gid); 5977 __put_user(st.st_nlink, &target_st->st_nlink); 5978 __put_user(st.st_rdev, &target_st->st_rdev); 5979 __put_user(st.st_size, &target_st->st_size); 5980 __put_user(st.st_blksize, &target_st->st_blksize); 5981 __put_user(st.st_blocks, &target_st->st_blocks); 5982 __put_user(st.st_atime, &target_st->target_st_atime); 5983 __put_user(st.st_mtime, &target_st->target_st_mtime); 5984 __put_user(st.st_ctime, &target_st->target_st_ctime); 5985 unlock_user_struct(target_st, arg2, 1); 5986 } 5987 } 5988 break; 5989 #ifdef TARGET_NR_olduname 5990 case TARGET_NR_olduname: 5991 goto unimplemented; 5992 #endif 5993 #ifdef TARGET_NR_iopl 5994 case TARGET_NR_iopl: 5995 goto unimplemented; 5996 #endif 5997 case TARGET_NR_vhangup: 5998 ret = get_errno(vhangup()); 5999 break; 6000 #ifdef TARGET_NR_idle 6001 case TARGET_NR_idle: 6002 goto unimplemented; 6003 #endif 6004 #ifdef TARGET_NR_syscall 6005 case TARGET_NR_syscall: 6006 ret = do_syscall(cpu_env,arg1 & 0xffff,arg2,arg3,arg4,arg5,arg6,0); 6007 break; 6008 #endif 6009 case TARGET_NR_wait4: 6010 { 6011 int status; 6012 abi_long status_ptr = arg2; 6013 struct rusage rusage, *rusage_ptr; 6014 abi_ulong target_rusage = arg4; 6015 if (target_rusage) 6016 rusage_ptr = &rusage; 6017 else 6018 rusage_ptr = NULL; 6019 ret = get_errno(wait4(arg1, &status, arg3, rusage_ptr)); 6020 if (!is_error(ret)) { 6021 if (status_ptr) { 6022 status = host_to_target_waitstatus(status); 6023 if (put_user_s32(status, status_ptr)) 6024 goto efault; 6025 } 6026 if (target_rusage) 6027 host_to_target_rusage(target_rusage, &rusage); 6028 } 6029 } 6030 break; 6031 #ifdef TARGET_NR_swapoff 6032 case TARGET_NR_swapoff: 6033 if (!(p = lock_user_string(arg1))) 6034 goto efault; 6035 ret = get_errno(swapoff(p)); 6036 unlock_user(p, arg1, 0); 6037 break; 6038 #endif 6039 case TARGET_NR_sysinfo: 6040 { 6041 struct target_sysinfo *target_value; 6042 struct sysinfo value; 6043 ret = get_errno(sysinfo(&value)); 6044 if (!is_error(ret) && arg1) 6045 { 6046 if (!lock_user_struct(VERIFY_WRITE, target_value, arg1, 0)) 6047 goto efault; 6048 __put_user(value.uptime, &target_value->uptime); 6049 __put_user(value.loads[0], &target_value->loads[0]); 6050 __put_user(value.loads[1], &target_value->loads[1]); 6051 __put_user(value.loads[2], &target_value->loads[2]); 6052 __put_user(value.totalram, &target_value->totalram); 6053 __put_user(value.freeram, &target_value->freeram); 6054 __put_user(value.sharedram, &target_value->sharedram); 6055 __put_user(value.bufferram, &target_value->bufferram); 6056 __put_user(value.totalswap, &target_value->totalswap); 6057 __put_user(value.freeswap, &target_value->freeswap); 6058 __put_user(value.procs, &target_value->procs); 6059 __put_user(value.totalhigh, &target_value->totalhigh); 6060 __put_user(value.freehigh, &target_value->freehigh); 6061 __put_user(value.mem_unit, &target_value->mem_unit); 6062 unlock_user_struct(target_value, arg1, 1); 6063 } 6064 } 6065 break; 6066 #ifdef TARGET_NR_ipc 6067 case TARGET_NR_ipc: 6068 ret = do_ipc(arg1, arg2, arg3, arg4, arg5, arg6); 6069 break; 6070 #endif 6071 #ifdef TARGET_NR_semget 6072 case TARGET_NR_semget: 6073 ret = get_errno(semget(arg1, arg2, arg3)); 6074 break; 6075 #endif 6076 #ifdef TARGET_NR_semop 6077 case TARGET_NR_semop: 6078 ret = get_errno(do_semop(arg1, arg2, arg3)); 6079 break; 6080 #endif 6081 #ifdef TARGET_NR_semctl 6082 case TARGET_NR_semctl: 6083 ret = do_semctl(arg1, arg2, arg3, (union target_semun)(abi_ulong)arg4); 6084 break; 6085 #endif 6086 #ifdef TARGET_NR_msgctl 6087 case TARGET_NR_msgctl: 6088 ret = do_msgctl(arg1, arg2, arg3); 6089 break; 6090 #endif 6091 #ifdef TARGET_NR_msgget 6092 case TARGET_NR_msgget: 6093 ret = get_errno(msgget(arg1, arg2)); 6094 break; 6095 #endif 6096 #ifdef TARGET_NR_msgrcv 6097 case TARGET_NR_msgrcv: 6098 ret = do_msgrcv(arg1, arg2, arg3, arg4, arg5); 6099 break; 6100 #endif 6101 #ifdef TARGET_NR_msgsnd 6102 case TARGET_NR_msgsnd: 6103 ret = do_msgsnd(arg1, arg2, arg3, arg4); 6104 break; 6105 #endif 6106 #ifdef TARGET_NR_shmget 6107 case TARGET_NR_shmget: 6108 ret = get_errno(shmget(arg1, arg2, arg3)); 6109 break; 6110 #endif 6111 #ifdef TARGET_NR_shmctl 6112 case TARGET_NR_shmctl: 6113 ret = do_shmctl(arg1, arg2, arg3); 6114 break; 6115 #endif 6116 #ifdef TARGET_NR_shmat 6117 case TARGET_NR_shmat: 6118 ret = do_shmat(arg1, arg2, arg3); 6119 break; 6120 #endif 6121 #ifdef TARGET_NR_shmdt 6122 case TARGET_NR_shmdt: 6123 ret = do_shmdt(arg1); 6124 break; 6125 #endif 6126 case TARGET_NR_fsync: 6127 ret = get_errno(fsync(arg1)); 6128 break; 6129 case TARGET_NR_clone: 6130 #if defined(TARGET_SH4) || defined(TARGET_ALPHA) 6131 ret = get_errno(do_fork(cpu_env, arg1, arg2, arg3, arg5, arg4)); 6132 #elif defined(TARGET_CRIS) 6133 ret = get_errno(do_fork(cpu_env, arg2, arg1, arg3, arg4, arg5)); 6134 #else 6135 ret = get_errno(do_fork(cpu_env, arg1, arg2, arg3, arg4, arg5)); 6136 #endif 6137 break; 6138 #ifdef __NR_exit_group 6139 /* new thread calls */ 6140 case TARGET_NR_exit_group: 6141 #ifdef TARGET_GPROF 6142 _mcleanup(); 6143 #endif 6144 gdb_exit(cpu_env, arg1); 6145 ret = get_errno(exit_group(arg1)); 6146 break; 6147 #endif 6148 case TARGET_NR_setdomainname: 6149 if (!(p = lock_user_string(arg1))) 6150 goto efault; 6151 ret = get_errno(setdomainname(p, arg2)); 6152 unlock_user(p, arg1, 0); 6153 break; 6154 case TARGET_NR_uname: 6155 /* no need to transcode because we use the linux syscall */ 6156 { 6157 struct new_utsname * buf; 6158 6159 if (!lock_user_struct(VERIFY_WRITE, buf, arg1, 0)) 6160 goto efault; 6161 ret = get_errno(sys_uname(buf)); 6162 if (!is_error(ret)) { 6163 /* Overrite the native machine name with whatever is being 6164 emulated. */ 6165 strcpy (buf->machine, cpu_to_uname_machine(cpu_env)); 6166 /* Allow the user to override the reported release. */ 6167 if (qemu_uname_release && *qemu_uname_release) 6168 strcpy (buf->release, qemu_uname_release); 6169 } 6170 unlock_user_struct(buf, arg1, 1); 6171 } 6172 break; 6173 #ifdef TARGET_I386 6174 case TARGET_NR_modify_ldt: 6175 ret = do_modify_ldt(cpu_env, arg1, arg2, arg3); 6176 break; 6177 #if !defined(TARGET_X86_64) 6178 case TARGET_NR_vm86old: 6179 goto unimplemented; 6180 case TARGET_NR_vm86: 6181 ret = do_vm86(cpu_env, arg1, arg2); 6182 break; 6183 #endif 6184 #endif 6185 case TARGET_NR_adjtimex: 6186 goto unimplemented; 6187 #ifdef TARGET_NR_create_module 6188 case TARGET_NR_create_module: 6189 #endif 6190 case TARGET_NR_init_module: 6191 case TARGET_NR_delete_module: 6192 #ifdef TARGET_NR_get_kernel_syms 6193 case TARGET_NR_get_kernel_syms: 6194 #endif 6195 goto unimplemented; 6196 case TARGET_NR_quotactl: 6197 goto unimplemented; 6198 case TARGET_NR_getpgid: 6199 ret = get_errno(getpgid(arg1)); 6200 break; 6201 case TARGET_NR_fchdir: 6202 ret = get_errno(fchdir(arg1)); 6203 break; 6204 #ifdef TARGET_NR_bdflush /* not on x86_64 */ 6205 case TARGET_NR_bdflush: 6206 goto unimplemented; 6207 #endif 6208 #ifdef TARGET_NR_sysfs 6209 case TARGET_NR_sysfs: 6210 goto unimplemented; 6211 #endif 6212 case TARGET_NR_personality: 6213 ret = get_errno(personality(arg1)); 6214 break; 6215 #ifdef TARGET_NR_afs_syscall 6216 case TARGET_NR_afs_syscall: 6217 goto unimplemented; 6218 #endif 6219 #ifdef TARGET_NR__llseek /* Not on alpha */ 6220 case TARGET_NR__llseek: 6221 { 6222 int64_t res; 6223 #if !defined(__NR_llseek) 6224 res = lseek(arg1, ((uint64_t)arg2 << 32) | arg3, arg5); 6225 if (res == -1) { 6226 ret = get_errno(res); 6227 } else { 6228 ret = 0; 6229 } 6230 #else 6231 ret = get_errno(_llseek(arg1, arg2, arg3, &res, arg5)); 6232 #endif 6233 if ((ret == 0) && put_user_s64(res, arg4)) { 6234 goto efault; 6235 } 6236 } 6237 break; 6238 #endif 6239 case TARGET_NR_getdents: 6240 #if TARGET_ABI_BITS == 32 && HOST_LONG_BITS == 64 6241 { 6242 struct target_dirent *target_dirp; 6243 struct linux_dirent *dirp; 6244 abi_long count = arg3; 6245 6246 dirp = malloc(count); 6247 if (!dirp) { 6248 ret = -TARGET_ENOMEM; 6249 goto fail; 6250 } 6251 6252 ret = get_errno(sys_getdents(arg1, dirp, count)); 6253 if (!is_error(ret)) { 6254 struct linux_dirent *de; 6255 struct target_dirent *tde; 6256 int len = ret; 6257 int reclen, treclen; 6258 int count1, tnamelen; 6259 6260 count1 = 0; 6261 de = dirp; 6262 if (!(target_dirp = lock_user(VERIFY_WRITE, arg2, count, 0))) 6263 goto efault; 6264 tde = target_dirp; 6265 while (len > 0) { 6266 reclen = de->d_reclen; 6267 treclen = reclen - (2 * (sizeof(long) - sizeof(abi_long))); 6268 tde->d_reclen = tswap16(treclen); 6269 tde->d_ino = tswapl(de->d_ino); 6270 tde->d_off = tswapl(de->d_off); 6271 tnamelen = treclen - (2 * sizeof(abi_long) + 2); 6272 if (tnamelen > 256) 6273 tnamelen = 256; 6274 /* XXX: may not be correct */ 6275 pstrcpy(tde->d_name, tnamelen, de->d_name); 6276 de = (struct linux_dirent *)((char *)de + reclen); 6277 len -= reclen; 6278 tde = (struct target_dirent *)((char *)tde + treclen); 6279 count1 += treclen; 6280 } 6281 ret = count1; 6282 unlock_user(target_dirp, arg2, ret); 6283 } 6284 free(dirp); 6285 } 6286 #else 6287 { 6288 struct linux_dirent *dirp; 6289 abi_long count = arg3; 6290 6291 if (!(dirp = lock_user(VERIFY_WRITE, arg2, count, 0))) 6292 goto efault; 6293 ret = get_errno(sys_getdents(arg1, dirp, count)); 6294 if (!is_error(ret)) { 6295 struct linux_dirent *de; 6296 int len = ret; 6297 int reclen; 6298 de = dirp; 6299 while (len > 0) { 6300 reclen = de->d_reclen; 6301 if (reclen > len) 6302 break; 6303 de->d_reclen = tswap16(reclen); 6304 tswapls(&de->d_ino); 6305 tswapls(&de->d_off); 6306 de = (struct linux_dirent *)((char *)de + reclen); 6307 len -= reclen; 6308 } 6309 } 6310 unlock_user(dirp, arg2, ret); 6311 } 6312 #endif 6313 break; 6314 #if defined(TARGET_NR_getdents64) && defined(__NR_getdents64) 6315 case TARGET_NR_getdents64: 6316 { 6317 struct linux_dirent64 *dirp; 6318 abi_long count = arg3; 6319 if (!(dirp = lock_user(VERIFY_WRITE, arg2, count, 0))) 6320 goto efault; 6321 ret = get_errno(sys_getdents64(arg1, dirp, count)); 6322 if (!is_error(ret)) { 6323 struct linux_dirent64 *de; 6324 int len = ret; 6325 int reclen; 6326 de = dirp; 6327 while (len > 0) { 6328 reclen = de->d_reclen; 6329 if (reclen > len) 6330 break; 6331 de->d_reclen = tswap16(reclen); 6332 tswap64s((uint64_t *)&de->d_ino); 6333 tswap64s((uint64_t *)&de->d_off); 6334 de = (struct linux_dirent64 *)((char *)de + reclen); 6335 len -= reclen; 6336 } 6337 } 6338 unlock_user(dirp, arg2, ret); 6339 } 6340 break; 6341 #endif /* TARGET_NR_getdents64 */ 6342 #ifdef TARGET_NR__newselect 6343 case TARGET_NR__newselect: 6344 ret = do_select(arg1, arg2, arg3, arg4, arg5); 6345 break; 6346 #endif 6347 #if defined(TARGET_NR_poll) || defined(TARGET_NR_ppoll) 6348 # ifdef TARGET_NR_poll 6349 case TARGET_NR_poll: 6350 # endif 6351 # ifdef TARGET_NR_ppoll 6352 case TARGET_NR_ppoll: 6353 # endif 6354 { 6355 struct target_pollfd *target_pfd; 6356 unsigned int nfds = arg2; 6357 int timeout = arg3; 6358 struct pollfd *pfd; 6359 unsigned int i; 6360 6361 target_pfd = lock_user(VERIFY_WRITE, arg1, sizeof(struct target_pollfd) * nfds, 1); 6362 if (!target_pfd) 6363 goto efault; 6364 6365 pfd = alloca(sizeof(struct pollfd) * nfds); 6366 for(i = 0; i < nfds; i++) { 6367 pfd[i].fd = tswap32(target_pfd[i].fd); 6368 pfd[i].events = tswap16(target_pfd[i].events); 6369 } 6370 6371 # ifdef TARGET_NR_ppoll 6372 if (num == TARGET_NR_ppoll) { 6373 struct timespec _timeout_ts, *timeout_ts = &_timeout_ts; 6374 target_sigset_t *target_set; 6375 sigset_t _set, *set = &_set; 6376 6377 if (arg3) { 6378 if (target_to_host_timespec(timeout_ts, arg3)) { 6379 unlock_user(target_pfd, arg1, 0); 6380 goto efault; 6381 } 6382 } else { 6383 timeout_ts = NULL; 6384 } 6385 6386 if (arg4) { 6387 target_set = lock_user(VERIFY_READ, arg4, sizeof(target_sigset_t), 1); 6388 if (!target_set) { 6389 unlock_user(target_pfd, arg1, 0); 6390 goto efault; 6391 } 6392 target_to_host_sigset(set, target_set); 6393 } else { 6394 set = NULL; 6395 } 6396 6397 ret = get_errno(sys_ppoll(pfd, nfds, timeout_ts, set, _NSIG/8)); 6398 6399 if (!is_error(ret) && arg3) { 6400 host_to_target_timespec(arg3, timeout_ts); 6401 } 6402 if (arg4) { 6403 unlock_user(target_set, arg4, 0); 6404 } 6405 } else 6406 # endif 6407 ret = get_errno(poll(pfd, nfds, timeout)); 6408 6409 if (!is_error(ret)) { 6410 for(i = 0; i < nfds; i++) { 6411 target_pfd[i].revents = tswap16(pfd[i].revents); 6412 } 6413 } 6414 unlock_user(target_pfd, arg1, sizeof(struct target_pollfd) * nfds); 6415 } 6416 break; 6417 #endif 6418 case TARGET_NR_flock: 6419 /* NOTE: the flock constant seems to be the same for every 6420 Linux platform */ 6421 ret = get_errno(flock(arg1, arg2)); 6422 break; 6423 case TARGET_NR_readv: 6424 { 6425 int count = arg3; 6426 struct iovec *vec; 6427 6428 vec = alloca(count * sizeof(struct iovec)); 6429 if (lock_iovec(VERIFY_WRITE, vec, arg2, count, 0) < 0) 6430 goto efault; 6431 ret = get_errno(readv(arg1, vec, count)); 6432 unlock_iovec(vec, arg2, count, 1); 6433 } 6434 break; 6435 case TARGET_NR_writev: 6436 { 6437 int count = arg3; 6438 struct iovec *vec; 6439 6440 vec = alloca(count * sizeof(struct iovec)); 6441 if (lock_iovec(VERIFY_READ, vec, arg2, count, 1) < 0) 6442 goto efault; 6443 ret = get_errno(writev(arg1, vec, count)); 6444 unlock_iovec(vec, arg2, count, 0); 6445 } 6446 break; 6447 case TARGET_NR_getsid: 6448 ret = get_errno(getsid(arg1)); 6449 break; 6450 #if defined(TARGET_NR_fdatasync) /* Not on alpha (osf_datasync ?) */ 6451 case TARGET_NR_fdatasync: 6452 ret = get_errno(fdatasync(arg1)); 6453 break; 6454 #endif 6455 case TARGET_NR__sysctl: 6456 /* We don't implement this, but ENOTDIR is always a safe 6457 return value. */ 6458 ret = -TARGET_ENOTDIR; 6459 break; 6460 case TARGET_NR_sched_getaffinity: 6461 { 6462 unsigned int mask_size; 6463 unsigned long *mask; 6464 6465 /* 6466 * sched_getaffinity needs multiples of ulong, so need to take 6467 * care of mismatches between target ulong and host ulong sizes. 6468 */ 6469 if (arg2 & (sizeof(abi_ulong) - 1)) { 6470 ret = -TARGET_EINVAL; 6471 break; 6472 } 6473 mask_size = (arg2 + (sizeof(*mask) - 1)) & ~(sizeof(*mask) - 1); 6474 6475 mask = alloca(mask_size); 6476 ret = get_errno(sys_sched_getaffinity(arg1, mask_size, mask)); 6477 6478 if (!is_error(ret)) { 6479 if (arg2 > ret) { 6480 /* Zero out any extra space kernel didn't fill */ 6481 unsigned long zero = arg2 - ret; 6482 p = alloca(zero); 6483 memset(p, 0, zero); 6484 if (copy_to_user(arg3 + zero, p, zero)) { 6485 goto efault; 6486 } 6487 arg2 = ret; 6488 } 6489 if (copy_to_user(arg3, mask, arg2)) { 6490 goto efault; 6491 } 6492 ret = arg2; 6493 } 6494 } 6495 break; 6496 case TARGET_NR_sched_setaffinity: 6497 { 6498 unsigned int mask_size; 6499 unsigned long *mask; 6500 6501 /* 6502 * sched_setaffinity needs multiples of ulong, so need to take 6503 * care of mismatches between target ulong and host ulong sizes. 6504 */ 6505 if (arg2 & (sizeof(abi_ulong) - 1)) { 6506 ret = -TARGET_EINVAL; 6507 break; 6508 } 6509 mask_size = (arg2 + (sizeof(*mask) - 1)) & ~(sizeof(*mask) - 1); 6510 6511 mask = alloca(mask_size); 6512 if (!lock_user_struct(VERIFY_READ, p, arg3, 1)) { 6513 goto efault; 6514 } 6515 memcpy(mask, p, arg2); 6516 unlock_user_struct(p, arg2, 0); 6517 6518 ret = get_errno(sys_sched_setaffinity(arg1, mask_size, mask)); 6519 } 6520 break; 6521 case TARGET_NR_sched_setparam: 6522 { 6523 struct sched_param *target_schp; 6524 struct sched_param schp; 6525 6526 if (!lock_user_struct(VERIFY_READ, target_schp, arg2, 1)) 6527 goto efault; 6528 schp.sched_priority = tswap32(target_schp->sched_priority); 6529 unlock_user_struct(target_schp, arg2, 0); 6530 ret = get_errno(sched_setparam(arg1, &schp)); 6531 } 6532 break; 6533 case TARGET_NR_sched_getparam: 6534 { 6535 struct sched_param *target_schp; 6536 struct sched_param schp; 6537 ret = get_errno(sched_getparam(arg1, &schp)); 6538 if (!is_error(ret)) { 6539 if (!lock_user_struct(VERIFY_WRITE, target_schp, arg2, 0)) 6540 goto efault; 6541 target_schp->sched_priority = tswap32(schp.sched_priority); 6542 unlock_user_struct(target_schp, arg2, 1); 6543 } 6544 } 6545 break; 6546 case TARGET_NR_sched_setscheduler: 6547 { 6548 struct sched_param *target_schp; 6549 struct sched_param schp; 6550 if (!lock_user_struct(VERIFY_READ, target_schp, arg3, 1)) 6551 goto efault; 6552 schp.sched_priority = tswap32(target_schp->sched_priority); 6553 unlock_user_struct(target_schp, arg3, 0); 6554 ret = get_errno(sched_setscheduler(arg1, arg2, &schp)); 6555 } 6556 break; 6557 case TARGET_NR_sched_getscheduler: 6558 ret = get_errno(sched_getscheduler(arg1)); 6559 break; 6560 case TARGET_NR_sched_yield: 6561 ret = get_errno(sched_yield()); 6562 break; 6563 case TARGET_NR_sched_get_priority_max: 6564 ret = get_errno(sched_get_priority_max(arg1)); 6565 break; 6566 case TARGET_NR_sched_get_priority_min: 6567 ret = get_errno(sched_get_priority_min(arg1)); 6568 break; 6569 case TARGET_NR_sched_rr_get_interval: 6570 { 6571 struct timespec ts; 6572 ret = get_errno(sched_rr_get_interval(arg1, &ts)); 6573 if (!is_error(ret)) { 6574 host_to_target_timespec(arg2, &ts); 6575 } 6576 } 6577 break; 6578 case TARGET_NR_nanosleep: 6579 { 6580 struct timespec req, rem; 6581 target_to_host_timespec(&req, arg1); 6582 ret = get_errno(nanosleep(&req, &rem)); 6583 if (is_error(ret) && arg2) { 6584 host_to_target_timespec(arg2, &rem); 6585 } 6586 } 6587 break; 6588 #ifdef TARGET_NR_query_module 6589 case TARGET_NR_query_module: 6590 goto unimplemented; 6591 #endif 6592 #ifdef TARGET_NR_nfsservctl 6593 case TARGET_NR_nfsservctl: 6594 goto unimplemented; 6595 #endif 6596 case TARGET_NR_prctl: 6597 switch (arg1) 6598 { 6599 case PR_GET_PDEATHSIG: 6600 { 6601 int deathsig; 6602 ret = get_errno(prctl(arg1, &deathsig, arg3, arg4, arg5)); 6603 if (!is_error(ret) && arg2 6604 && put_user_ual(deathsig, arg2)) 6605 goto efault; 6606 } 6607 break; 6608 default: 6609 ret = get_errno(prctl(arg1, arg2, arg3, arg4, arg5)); 6610 break; 6611 } 6612 break; 6613 #ifdef TARGET_NR_arch_prctl 6614 case TARGET_NR_arch_prctl: 6615 #if defined(TARGET_I386) && !defined(TARGET_ABI32) 6616 ret = do_arch_prctl(cpu_env, arg1, arg2); 6617 break; 6618 #else 6619 goto unimplemented; 6620 #endif 6621 #endif 6622 #ifdef TARGET_NR_pread 6623 case TARGET_NR_pread: 6624 #ifdef TARGET_ARM 6625 if (((CPUARMState *)cpu_env)->eabi) 6626 arg4 = arg5; 6627 #endif 6628 if (!(p = lock_user(VERIFY_WRITE, arg2, arg3, 0))) 6629 goto efault; 6630 ret = get_errno(pread(arg1, p, arg3, arg4)); 6631 unlock_user(p, arg2, ret); 6632 break; 6633 case TARGET_NR_pwrite: 6634 #ifdef TARGET_ARM 6635 if (((CPUARMState *)cpu_env)->eabi) 6636 arg4 = arg5; 6637 #endif 6638 if (!(p = lock_user(VERIFY_READ, arg2, arg3, 1))) 6639 goto efault; 6640 ret = get_errno(pwrite(arg1, p, arg3, arg4)); 6641 unlock_user(p, arg2, 0); 6642 break; 6643 #endif 6644 #ifdef TARGET_NR_pread64 6645 case TARGET_NR_pread64: 6646 if (!(p = lock_user(VERIFY_WRITE, arg2, arg3, 0))) 6647 goto efault; 6648 ret = get_errno(pread64(arg1, p, arg3, target_offset64(arg4, arg5))); 6649 unlock_user(p, arg2, ret); 6650 break; 6651 case TARGET_NR_pwrite64: 6652 if (!(p = lock_user(VERIFY_READ, arg2, arg3, 1))) 6653 goto efault; 6654 ret = get_errno(pwrite64(arg1, p, arg3, target_offset64(arg4, arg5))); 6655 unlock_user(p, arg2, 0); 6656 break; 6657 #endif 6658 case TARGET_NR_getcwd: 6659 if (!(p = lock_user(VERIFY_WRITE, arg1, arg2, 0))) 6660 goto efault; 6661 ret = get_errno(sys_getcwd1(p, arg2)); 6662 unlock_user(p, arg1, ret); 6663 break; 6664 case TARGET_NR_capget: 6665 goto unimplemented; 6666 case TARGET_NR_capset: 6667 goto unimplemented; 6668 case TARGET_NR_sigaltstack: 6669 #if defined(TARGET_I386) || defined(TARGET_ARM) || defined(TARGET_MIPS) || \ 6670 defined(TARGET_SPARC) || defined(TARGET_PPC) || defined(TARGET_ALPHA) || \ 6671 defined(TARGET_M68K) 6672 ret = do_sigaltstack(arg1, arg2, get_sp_from_cpustate((CPUState *)cpu_env)); 6673 break; 6674 #else 6675 goto unimplemented; 6676 #endif 6677 case TARGET_NR_sendfile: 6678 goto unimplemented; 6679 #ifdef TARGET_NR_getpmsg 6680 case TARGET_NR_getpmsg: 6681 goto unimplemented; 6682 #endif 6683 #ifdef TARGET_NR_putpmsg 6684 case TARGET_NR_putpmsg: 6685 goto unimplemented; 6686 #endif 6687 #ifdef TARGET_NR_vfork 6688 case TARGET_NR_vfork: 6689 ret = get_errno(do_fork(cpu_env, CLONE_VFORK | CLONE_VM | SIGCHLD, 6690 0, 0, 0, 0)); 6691 break; 6692 #endif 6693 #ifdef TARGET_NR_ugetrlimit 6694 case TARGET_NR_ugetrlimit: 6695 { 6696 struct rlimit rlim; 6697 ret = get_errno(getrlimit(arg1, &rlim)); 6698 if (!is_error(ret)) { 6699 struct target_rlimit *target_rlim; 6700 if (!lock_user_struct(VERIFY_WRITE, target_rlim, arg2, 0)) 6701 goto efault; 6702 target_rlim->rlim_cur = host_to_target_rlim(rlim.rlim_cur); 6703 target_rlim->rlim_max = host_to_target_rlim(rlim.rlim_max); 6704 unlock_user_struct(target_rlim, arg2, 1); 6705 } 6706 break; 6707 } 6708 #endif 6709 #ifdef TARGET_NR_truncate64 6710 case TARGET_NR_truncate64: 6711 if (!(p = lock_user_string(arg1))) 6712 goto efault; 6713 ret = target_truncate64(cpu_env, p, arg2, arg3, arg4); 6714 unlock_user(p, arg1, 0); 6715 break; 6716 #endif 6717 #ifdef TARGET_NR_ftruncate64 6718 case TARGET_NR_ftruncate64: 6719 ret = target_ftruncate64(cpu_env, arg1, arg2, arg3, arg4); 6720 break; 6721 #endif 6722 #ifdef TARGET_NR_stat64 6723 case TARGET_NR_stat64: 6724 if (!(p = lock_user_string(arg1))) 6725 goto efault; 6726 ret = get_errno(stat(path(p), &st)); 6727 unlock_user(p, arg1, 0); 6728 if (!is_error(ret)) 6729 ret = host_to_target_stat64(cpu_env, arg2, &st); 6730 break; 6731 #endif 6732 #ifdef TARGET_NR_lstat64 6733 case TARGET_NR_lstat64: 6734 if (!(p = lock_user_string(arg1))) 6735 goto efault; 6736 ret = get_errno(lstat(path(p), &st)); 6737 unlock_user(p, arg1, 0); 6738 if (!is_error(ret)) 6739 ret = host_to_target_stat64(cpu_env, arg2, &st); 6740 break; 6741 #endif 6742 #ifdef TARGET_NR_fstat64 6743 case TARGET_NR_fstat64: 6744 ret = get_errno(fstat(arg1, &st)); 6745 if (!is_error(ret)) 6746 ret = host_to_target_stat64(cpu_env, arg2, &st); 6747 break; 6748 #endif 6749 #if (defined(TARGET_NR_fstatat64) || defined(TARGET_NR_newfstatat)) && \ 6750 (defined(__NR_fstatat64) || defined(__NR_newfstatat)) 6751 #ifdef TARGET_NR_fstatat64 6752 case TARGET_NR_fstatat64: 6753 #endif 6754 #ifdef TARGET_NR_newfstatat 6755 case TARGET_NR_newfstatat: 6756 #endif 6757 if (!(p = lock_user_string(arg2))) 6758 goto efault; 6759 #ifdef __NR_fstatat64 6760 ret = get_errno(sys_fstatat64(arg1, path(p), &st, arg4)); 6761 #else 6762 ret = get_errno(sys_newfstatat(arg1, path(p), &st, arg4)); 6763 #endif 6764 if (!is_error(ret)) 6765 ret = host_to_target_stat64(cpu_env, arg3, &st); 6766 break; 6767 #endif 6768 #ifdef USE_UID16 6769 case TARGET_NR_lchown: 6770 if (!(p = lock_user_string(arg1))) 6771 goto efault; 6772 ret = get_errno(lchown(p, low2highuid(arg2), low2highgid(arg3))); 6773 unlock_user(p, arg1, 0); 6774 break; 6775 case TARGET_NR_getuid: 6776 ret = get_errno(high2lowuid(getuid())); 6777 break; 6778 case TARGET_NR_getgid: 6779 ret = get_errno(high2lowgid(getgid())); 6780 break; 6781 case TARGET_NR_geteuid: 6782 ret = get_errno(high2lowuid(geteuid())); 6783 break; 6784 case TARGET_NR_getegid: 6785 ret = get_errno(high2lowgid(getegid())); 6786 break; 6787 case TARGET_NR_setreuid: 6788 ret = get_errno(setreuid(low2highuid(arg1), low2highuid(arg2))); 6789 break; 6790 case TARGET_NR_setregid: 6791 ret = get_errno(setregid(low2highgid(arg1), low2highgid(arg2))); 6792 break; 6793 case TARGET_NR_getgroups: 6794 { 6795 int gidsetsize = arg1; 6796 uint16_t *target_grouplist; 6797 gid_t *grouplist; 6798 int i; 6799 6800 grouplist = alloca(gidsetsize * sizeof(gid_t)); 6801 ret = get_errno(getgroups(gidsetsize, grouplist)); 6802 if (gidsetsize == 0) 6803 break; 6804 if (!is_error(ret)) { 6805 target_grouplist = lock_user(VERIFY_WRITE, arg2, gidsetsize * 2, 0); 6806 if (!target_grouplist) 6807 goto efault; 6808 for(i = 0;i < ret; i++) 6809 target_grouplist[i] = tswap16(grouplist[i]); 6810 unlock_user(target_grouplist, arg2, gidsetsize * 2); 6811 } 6812 } 6813 break; 6814 case TARGET_NR_setgroups: 6815 { 6816 int gidsetsize = arg1; 6817 uint16_t *target_grouplist; 6818 gid_t *grouplist; 6819 int i; 6820 6821 grouplist = alloca(gidsetsize * sizeof(gid_t)); 6822 target_grouplist = lock_user(VERIFY_READ, arg2, gidsetsize * 2, 1); 6823 if (!target_grouplist) { 6824 ret = -TARGET_EFAULT; 6825 goto fail; 6826 } 6827 for(i = 0;i < gidsetsize; i++) 6828 grouplist[i] = tswap16(target_grouplist[i]); 6829 unlock_user(target_grouplist, arg2, 0); 6830 ret = get_errno(setgroups(gidsetsize, grouplist)); 6831 } 6832 break; 6833 case TARGET_NR_fchown: 6834 ret = get_errno(fchown(arg1, low2highuid(arg2), low2highgid(arg3))); 6835 break; 6836 #if defined(TARGET_NR_fchownat) && defined(__NR_fchownat) 6837 case TARGET_NR_fchownat: 6838 if (!(p = lock_user_string(arg2))) 6839 goto efault; 6840 ret = get_errno(sys_fchownat(arg1, p, low2highuid(arg3), low2highgid(arg4), arg5)); 6841 unlock_user(p, arg2, 0); 6842 break; 6843 #endif 6844 #ifdef TARGET_NR_setresuid 6845 case TARGET_NR_setresuid: 6846 ret = get_errno(setresuid(low2highuid(arg1), 6847 low2highuid(arg2), 6848 low2highuid(arg3))); 6849 break; 6850 #endif 6851 #ifdef TARGET_NR_getresuid 6852 case TARGET_NR_getresuid: 6853 { 6854 uid_t ruid, euid, suid; 6855 ret = get_errno(getresuid(&ruid, &euid, &suid)); 6856 if (!is_error(ret)) { 6857 if (put_user_u16(high2lowuid(ruid), arg1) 6858 || put_user_u16(high2lowuid(euid), arg2) 6859 || put_user_u16(high2lowuid(suid), arg3)) 6860 goto efault; 6861 } 6862 } 6863 break; 6864 #endif 6865 #ifdef TARGET_NR_getresgid 6866 case TARGET_NR_setresgid: 6867 ret = get_errno(setresgid(low2highgid(arg1), 6868 low2highgid(arg2), 6869 low2highgid(arg3))); 6870 break; 6871 #endif 6872 #ifdef TARGET_NR_getresgid 6873 case TARGET_NR_getresgid: 6874 { 6875 gid_t rgid, egid, sgid; 6876 ret = get_errno(getresgid(&rgid, &egid, &sgid)); 6877 if (!is_error(ret)) { 6878 if (put_user_u16(high2lowgid(rgid), arg1) 6879 || put_user_u16(high2lowgid(egid), arg2) 6880 || put_user_u16(high2lowgid(sgid), arg3)) 6881 goto efault; 6882 } 6883 } 6884 break; 6885 #endif 6886 case TARGET_NR_chown: 6887 if (!(p = lock_user_string(arg1))) 6888 goto efault; 6889 ret = get_errno(chown(p, low2highuid(arg2), low2highgid(arg3))); 6890 unlock_user(p, arg1, 0); 6891 break; 6892 case TARGET_NR_setuid: 6893 ret = get_errno(setuid(low2highuid(arg1))); 6894 break; 6895 case TARGET_NR_setgid: 6896 ret = get_errno(setgid(low2highgid(arg1))); 6897 break; 6898 case TARGET_NR_setfsuid: 6899 ret = get_errno(setfsuid(arg1)); 6900 break; 6901 case TARGET_NR_setfsgid: 6902 ret = get_errno(setfsgid(arg1)); 6903 break; 6904 #endif /* USE_UID16 */ 6905 6906 #ifdef TARGET_NR_lchown32 6907 case TARGET_NR_lchown32: 6908 if (!(p = lock_user_string(arg1))) 6909 goto efault; 6910 ret = get_errno(lchown(p, arg2, arg3)); 6911 unlock_user(p, arg1, 0); 6912 break; 6913 #endif 6914 #ifdef TARGET_NR_getuid32 6915 case TARGET_NR_getuid32: 6916 ret = get_errno(getuid()); 6917 break; 6918 #endif 6919 6920 #if defined(TARGET_NR_getxuid) && defined(TARGET_ALPHA) 6921 /* Alpha specific */ 6922 case TARGET_NR_getxuid: 6923 { 6924 uid_t euid; 6925 euid=geteuid(); 6926 ((CPUAlphaState *)cpu_env)->ir[IR_A4]=euid; 6927 } 6928 ret = get_errno(getuid()); 6929 break; 6930 #endif 6931 #if defined(TARGET_NR_getxgid) && defined(TARGET_ALPHA) 6932 /* Alpha specific */ 6933 case TARGET_NR_getxgid: 6934 { 6935 uid_t egid; 6936 egid=getegid(); 6937 ((CPUAlphaState *)cpu_env)->ir[IR_A4]=egid; 6938 } 6939 ret = get_errno(getgid()); 6940 break; 6941 #endif 6942 #if defined(TARGET_NR_osf_getsysinfo) && defined(TARGET_ALPHA) 6943 /* Alpha specific */ 6944 case TARGET_NR_osf_getsysinfo: 6945 ret = -TARGET_EOPNOTSUPP; 6946 switch (arg1) { 6947 case TARGET_GSI_IEEE_FP_CONTROL: 6948 { 6949 uint64_t swcr, fpcr = cpu_alpha_load_fpcr (cpu_env); 6950 6951 /* Copied from linux ieee_fpcr_to_swcr. */ 6952 swcr = (fpcr >> 35) & SWCR_STATUS_MASK; 6953 swcr |= (fpcr >> 36) & SWCR_MAP_DMZ; 6954 swcr |= (~fpcr >> 48) & (SWCR_TRAP_ENABLE_INV 6955 | SWCR_TRAP_ENABLE_DZE 6956 | SWCR_TRAP_ENABLE_OVF); 6957 swcr |= (~fpcr >> 57) & (SWCR_TRAP_ENABLE_UNF 6958 | SWCR_TRAP_ENABLE_INE); 6959 swcr |= (fpcr >> 47) & SWCR_MAP_UMZ; 6960 swcr |= (~fpcr >> 41) & SWCR_TRAP_ENABLE_DNO; 6961 6962 if (put_user_u64 (swcr, arg2)) 6963 goto efault; 6964 ret = 0; 6965 } 6966 break; 6967 6968 /* case GSI_IEEE_STATE_AT_SIGNAL: 6969 -- Not implemented in linux kernel. 6970 case GSI_UACPROC: 6971 -- Retrieves current unaligned access state; not much used. 6972 case GSI_PROC_TYPE: 6973 -- Retrieves implver information; surely not used. 6974 case GSI_GET_HWRPB: 6975 -- Grabs a copy of the HWRPB; surely not used. 6976 */ 6977 } 6978 break; 6979 #endif 6980 #if defined(TARGET_NR_osf_setsysinfo) && defined(TARGET_ALPHA) 6981 /* Alpha specific */ 6982 case TARGET_NR_osf_setsysinfo: 6983 ret = -TARGET_EOPNOTSUPP; 6984 switch (arg1) { 6985 case TARGET_SSI_IEEE_FP_CONTROL: 6986 case TARGET_SSI_IEEE_RAISE_EXCEPTION: 6987 { 6988 uint64_t swcr, fpcr, orig_fpcr; 6989 6990 if (get_user_u64 (swcr, arg2)) 6991 goto efault; 6992 orig_fpcr = cpu_alpha_load_fpcr (cpu_env); 6993 fpcr = orig_fpcr & FPCR_DYN_MASK; 6994 6995 /* Copied from linux ieee_swcr_to_fpcr. */ 6996 fpcr |= (swcr & SWCR_STATUS_MASK) << 35; 6997 fpcr |= (swcr & SWCR_MAP_DMZ) << 36; 6998 fpcr |= (~swcr & (SWCR_TRAP_ENABLE_INV 6999 | SWCR_TRAP_ENABLE_DZE 7000 | SWCR_TRAP_ENABLE_OVF)) << 48; 7001 fpcr |= (~swcr & (SWCR_TRAP_ENABLE_UNF 7002 | SWCR_TRAP_ENABLE_INE)) << 57; 7003 fpcr |= (swcr & SWCR_MAP_UMZ ? FPCR_UNDZ | FPCR_UNFD : 0); 7004 fpcr |= (~swcr & SWCR_TRAP_ENABLE_DNO) << 41; 7005 7006 cpu_alpha_store_fpcr (cpu_env, fpcr); 7007 ret = 0; 7008 7009 if (arg1 == TARGET_SSI_IEEE_RAISE_EXCEPTION) { 7010 /* Old exceptions are not signaled. */ 7011 fpcr &= ~(orig_fpcr & FPCR_STATUS_MASK); 7012 7013 /* If any exceptions set by this call, and are unmasked, 7014 send a signal. */ 7015 /* ??? FIXME */ 7016 } 7017 } 7018 break; 7019 7020 /* case SSI_NVPAIRS: 7021 -- Used with SSIN_UACPROC to enable unaligned accesses. 7022 case SSI_IEEE_STATE_AT_SIGNAL: 7023 case SSI_IEEE_IGNORE_STATE_AT_SIGNAL: 7024 -- Not implemented in linux kernel 7025 */ 7026 } 7027 break; 7028 #endif 7029 #ifdef TARGET_NR_osf_sigprocmask 7030 /* Alpha specific. */ 7031 case TARGET_NR_osf_sigprocmask: 7032 { 7033 abi_ulong mask; 7034 int how = arg1; 7035 sigset_t set, oldset; 7036 7037 switch(arg1) { 7038 case TARGET_SIG_BLOCK: 7039 how = SIG_BLOCK; 7040 break; 7041 case TARGET_SIG_UNBLOCK: 7042 how = SIG_UNBLOCK; 7043 break; 7044 case TARGET_SIG_SETMASK: 7045 how = SIG_SETMASK; 7046 break; 7047 default: 7048 ret = -TARGET_EINVAL; 7049 goto fail; 7050 } 7051 mask = arg2; 7052 target_to_host_old_sigset(&set, &mask); 7053 sigprocmask(arg1, &set, &oldset); 7054 host_to_target_old_sigset(&mask, &oldset); 7055 ret = mask; 7056 } 7057 break; 7058 #endif 7059 7060 #ifdef TARGET_NR_getgid32 7061 case TARGET_NR_getgid32: 7062 ret = get_errno(getgid()); 7063 break; 7064 #endif 7065 #ifdef TARGET_NR_geteuid32 7066 case TARGET_NR_geteuid32: 7067 ret = get_errno(geteuid()); 7068 break; 7069 #endif 7070 #ifdef TARGET_NR_getegid32 7071 case TARGET_NR_getegid32: 7072 ret = get_errno(getegid()); 7073 break; 7074 #endif 7075 #ifdef TARGET_NR_setreuid32 7076 case TARGET_NR_setreuid32: 7077 ret = get_errno(setreuid(arg1, arg2)); 7078 break; 7079 #endif 7080 #ifdef TARGET_NR_setregid32 7081 case TARGET_NR_setregid32: 7082 ret = get_errno(setregid(arg1, arg2)); 7083 break; 7084 #endif 7085 #ifdef TARGET_NR_getgroups32 7086 case TARGET_NR_getgroups32: 7087 { 7088 int gidsetsize = arg1; 7089 uint32_t *target_grouplist; 7090 gid_t *grouplist; 7091 int i; 7092 7093 grouplist = alloca(gidsetsize * sizeof(gid_t)); 7094 ret = get_errno(getgroups(gidsetsize, grouplist)); 7095 if (gidsetsize == 0) 7096 break; 7097 if (!is_error(ret)) { 7098 target_grouplist = lock_user(VERIFY_WRITE, arg2, gidsetsize * 4, 0); 7099 if (!target_grouplist) { 7100 ret = -TARGET_EFAULT; 7101 goto fail; 7102 } 7103 for(i = 0;i < ret; i++) 7104 target_grouplist[i] = tswap32(grouplist[i]); 7105 unlock_user(target_grouplist, arg2, gidsetsize * 4); 7106 } 7107 } 7108 break; 7109 #endif 7110 #ifdef TARGET_NR_setgroups32 7111 case TARGET_NR_setgroups32: 7112 { 7113 int gidsetsize = arg1; 7114 uint32_t *target_grouplist; 7115 gid_t *grouplist; 7116 int i; 7117 7118 grouplist = alloca(gidsetsize * sizeof(gid_t)); 7119 target_grouplist = lock_user(VERIFY_READ, arg2, gidsetsize * 4, 1); 7120 if (!target_grouplist) { 7121 ret = -TARGET_EFAULT; 7122 goto fail; 7123 } 7124 for(i = 0;i < gidsetsize; i++) 7125 grouplist[i] = tswap32(target_grouplist[i]); 7126 unlock_user(target_grouplist, arg2, 0); 7127 ret = get_errno(setgroups(gidsetsize, grouplist)); 7128 } 7129 break; 7130 #endif 7131 #ifdef TARGET_NR_fchown32 7132 case TARGET_NR_fchown32: 7133 ret = get_errno(fchown(arg1, arg2, arg3)); 7134 break; 7135 #endif 7136 #ifdef TARGET_NR_setresuid32 7137 case TARGET_NR_setresuid32: 7138 ret = get_errno(setresuid(arg1, arg2, arg3)); 7139 break; 7140 #endif 7141 #ifdef TARGET_NR_getresuid32 7142 case TARGET_NR_getresuid32: 7143 { 7144 uid_t ruid, euid, suid; 7145 ret = get_errno(getresuid(&ruid, &euid, &suid)); 7146 if (!is_error(ret)) { 7147 if (put_user_u32(ruid, arg1) 7148 || put_user_u32(euid, arg2) 7149 || put_user_u32(suid, arg3)) 7150 goto efault; 7151 } 7152 } 7153 break; 7154 #endif 7155 #ifdef TARGET_NR_setresgid32 7156 case TARGET_NR_setresgid32: 7157 ret = get_errno(setresgid(arg1, arg2, arg3)); 7158 break; 7159 #endif 7160 #ifdef TARGET_NR_getresgid32 7161 case TARGET_NR_getresgid32: 7162 { 7163 gid_t rgid, egid, sgid; 7164 ret = get_errno(getresgid(&rgid, &egid, &sgid)); 7165 if (!is_error(ret)) { 7166 if (put_user_u32(rgid, arg1) 7167 || put_user_u32(egid, arg2) 7168 || put_user_u32(sgid, arg3)) 7169 goto efault; 7170 } 7171 } 7172 break; 7173 #endif 7174 #ifdef TARGET_NR_chown32 7175 case TARGET_NR_chown32: 7176 if (!(p = lock_user_string(arg1))) 7177 goto efault; 7178 ret = get_errno(chown(p, arg2, arg3)); 7179 unlock_user(p, arg1, 0); 7180 break; 7181 #endif 7182 #ifdef TARGET_NR_setuid32 7183 case TARGET_NR_setuid32: 7184 ret = get_errno(setuid(arg1)); 7185 break; 7186 #endif 7187 #ifdef TARGET_NR_setgid32 7188 case TARGET_NR_setgid32: 7189 ret = get_errno(setgid(arg1)); 7190 break; 7191 #endif 7192 #ifdef TARGET_NR_setfsuid32 7193 case TARGET_NR_setfsuid32: 7194 ret = get_errno(setfsuid(arg1)); 7195 break; 7196 #endif 7197 #ifdef TARGET_NR_setfsgid32 7198 case TARGET_NR_setfsgid32: 7199 ret = get_errno(setfsgid(arg1)); 7200 break; 7201 #endif 7202 7203 case TARGET_NR_pivot_root: 7204 goto unimplemented; 7205 #ifdef TARGET_NR_mincore 7206 case TARGET_NR_mincore: 7207 { 7208 void *a; 7209 ret = -TARGET_EFAULT; 7210 if (!(a = lock_user(VERIFY_READ, arg1,arg2, 0))) 7211 goto efault; 7212 if (!(p = lock_user_string(arg3))) 7213 goto mincore_fail; 7214 ret = get_errno(mincore(a, arg2, p)); 7215 unlock_user(p, arg3, ret); 7216 mincore_fail: 7217 unlock_user(a, arg1, 0); 7218 } 7219 break; 7220 #endif 7221 #ifdef TARGET_NR_arm_fadvise64_64 7222 case TARGET_NR_arm_fadvise64_64: 7223 { 7224 /* 7225 * arm_fadvise64_64 looks like fadvise64_64 but 7226 * with different argument order 7227 */ 7228 abi_long temp; 7229 temp = arg3; 7230 arg3 = arg4; 7231 arg4 = temp; 7232 } 7233 #endif 7234 #if defined(TARGET_NR_fadvise64_64) || defined(TARGET_NR_arm_fadvise64_64) || defined(TARGET_NR_fadvise64) 7235 #ifdef TARGET_NR_fadvise64_64 7236 case TARGET_NR_fadvise64_64: 7237 #endif 7238 #ifdef TARGET_NR_fadvise64 7239 case TARGET_NR_fadvise64: 7240 #endif 7241 #ifdef TARGET_S390X 7242 switch (arg4) { 7243 case 4: arg4 = POSIX_FADV_NOREUSE + 1; break; /* make sure it's an invalid value */ 7244 case 5: arg4 = POSIX_FADV_NOREUSE + 2; break; /* ditto */ 7245 case 6: arg4 = POSIX_FADV_DONTNEED; break; 7246 case 7: arg4 = POSIX_FADV_NOREUSE; break; 7247 default: break; 7248 } 7249 #endif 7250 ret = -posix_fadvise(arg1, arg2, arg3, arg4); 7251 break; 7252 #endif 7253 #ifdef TARGET_NR_madvise 7254 case TARGET_NR_madvise: 7255 /* A straight passthrough may not be safe because qemu sometimes 7256 turns private flie-backed mappings into anonymous mappings. 7257 This will break MADV_DONTNEED. 7258 This is a hint, so ignoring and returning success is ok. */ 7259 ret = get_errno(0); 7260 break; 7261 #endif 7262 #if TARGET_ABI_BITS == 32 7263 case TARGET_NR_fcntl64: 7264 { 7265 int cmd; 7266 struct flock64 fl; 7267 struct target_flock64 *target_fl; 7268 #ifdef TARGET_ARM 7269 struct target_eabi_flock64 *target_efl; 7270 #endif 7271 7272 cmd = target_to_host_fcntl_cmd(arg2); 7273 if (cmd == -TARGET_EINVAL) 7274 return cmd; 7275 7276 switch(arg2) { 7277 case TARGET_F_GETLK64: 7278 #ifdef TARGET_ARM 7279 if (((CPUARMState *)cpu_env)->eabi) { 7280 if (!lock_user_struct(VERIFY_READ, target_efl, arg3, 1)) 7281 goto efault; 7282 fl.l_type = tswap16(target_efl->l_type); 7283 fl.l_whence = tswap16(target_efl->l_whence); 7284 fl.l_start = tswap64(target_efl->l_start); 7285 fl.l_len = tswap64(target_efl->l_len); 7286 fl.l_pid = tswap32(target_efl->l_pid); 7287 unlock_user_struct(target_efl, arg3, 0); 7288 } else 7289 #endif 7290 { 7291 if (!lock_user_struct(VERIFY_READ, target_fl, arg3, 1)) 7292 goto efault; 7293 fl.l_type = tswap16(target_fl->l_type); 7294 fl.l_whence = tswap16(target_fl->l_whence); 7295 fl.l_start = tswap64(target_fl->l_start); 7296 fl.l_len = tswap64(target_fl->l_len); 7297 fl.l_pid = tswap32(target_fl->l_pid); 7298 unlock_user_struct(target_fl, arg3, 0); 7299 } 7300 ret = get_errno(fcntl(arg1, cmd, &fl)); 7301 if (ret == 0) { 7302 #ifdef TARGET_ARM 7303 if (((CPUARMState *)cpu_env)->eabi) { 7304 if (!lock_user_struct(VERIFY_WRITE, target_efl, arg3, 0)) 7305 goto efault; 7306 target_efl->l_type = tswap16(fl.l_type); 7307 target_efl->l_whence = tswap16(fl.l_whence); 7308 target_efl->l_start = tswap64(fl.l_start); 7309 target_efl->l_len = tswap64(fl.l_len); 7310 target_efl->l_pid = tswap32(fl.l_pid); 7311 unlock_user_struct(target_efl, arg3, 1); 7312 } else 7313 #endif 7314 { 7315 if (!lock_user_struct(VERIFY_WRITE, target_fl, arg3, 0)) 7316 goto efault; 7317 target_fl->l_type = tswap16(fl.l_type); 7318 target_fl->l_whence = tswap16(fl.l_whence); 7319 target_fl->l_start = tswap64(fl.l_start); 7320 target_fl->l_len = tswap64(fl.l_len); 7321 target_fl->l_pid = tswap32(fl.l_pid); 7322 unlock_user_struct(target_fl, arg3, 1); 7323 } 7324 } 7325 break; 7326 7327 case TARGET_F_SETLK64: 7328 case TARGET_F_SETLKW64: 7329 #ifdef TARGET_ARM 7330 if (((CPUARMState *)cpu_env)->eabi) { 7331 if (!lock_user_struct(VERIFY_READ, target_efl, arg3, 1)) 7332 goto efault; 7333 fl.l_type = tswap16(target_efl->l_type); 7334 fl.l_whence = tswap16(target_efl->l_whence); 7335 fl.l_start = tswap64(target_efl->l_start); 7336 fl.l_len = tswap64(target_efl->l_len); 7337 fl.l_pid = tswap32(target_efl->l_pid); 7338 unlock_user_struct(target_efl, arg3, 0); 7339 } else 7340 #endif 7341 { 7342 if (!lock_user_struct(VERIFY_READ, target_fl, arg3, 1)) 7343 goto efault; 7344 fl.l_type = tswap16(target_fl->l_type); 7345 fl.l_whence = tswap16(target_fl->l_whence); 7346 fl.l_start = tswap64(target_fl->l_start); 7347 fl.l_len = tswap64(target_fl->l_len); 7348 fl.l_pid = tswap32(target_fl->l_pid); 7349 unlock_user_struct(target_fl, arg3, 0); 7350 } 7351 ret = get_errno(fcntl(arg1, cmd, &fl)); 7352 break; 7353 default: 7354 ret = do_fcntl(arg1, arg2, arg3); 7355 break; 7356 } 7357 break; 7358 } 7359 #endif 7360 #ifdef TARGET_NR_cacheflush 7361 case TARGET_NR_cacheflush: 7362 /* self-modifying code is handled automatically, so nothing needed */ 7363 ret = 0; 7364 break; 7365 #endif 7366 #ifdef TARGET_NR_security 7367 case TARGET_NR_security: 7368 goto unimplemented; 7369 #endif 7370 #ifdef TARGET_NR_getpagesize 7371 case TARGET_NR_getpagesize: 7372 ret = TARGET_PAGE_SIZE; 7373 break; 7374 #endif 7375 case TARGET_NR_gettid: 7376 ret = get_errno(gettid()); 7377 break; 7378 #ifdef TARGET_NR_readahead 7379 case TARGET_NR_readahead: 7380 #if TARGET_ABI_BITS == 32 7381 #ifdef TARGET_ARM 7382 if (((CPUARMState *)cpu_env)->eabi) 7383 { 7384 arg2 = arg3; 7385 arg3 = arg4; 7386 arg4 = arg5; 7387 } 7388 #endif 7389 ret = get_errno(readahead(arg1, ((off64_t)arg3 << 32) | arg2, arg4)); 7390 #else 7391 ret = get_errno(readahead(arg1, arg2, arg3)); 7392 #endif 7393 break; 7394 #endif 7395 #ifdef TARGET_NR_setxattr 7396 case TARGET_NR_setxattr: 7397 case TARGET_NR_lsetxattr: 7398 case TARGET_NR_fsetxattr: 7399 case TARGET_NR_getxattr: 7400 case TARGET_NR_lgetxattr: 7401 case TARGET_NR_fgetxattr: 7402 case TARGET_NR_listxattr: 7403 case TARGET_NR_llistxattr: 7404 case TARGET_NR_flistxattr: 7405 case TARGET_NR_removexattr: 7406 case TARGET_NR_lremovexattr: 7407 case TARGET_NR_fremovexattr: 7408 ret = -TARGET_EOPNOTSUPP; 7409 break; 7410 #endif 7411 #ifdef TARGET_NR_set_thread_area 7412 case TARGET_NR_set_thread_area: 7413 #if defined(TARGET_MIPS) 7414 ((CPUMIPSState *) cpu_env)->tls_value = arg1; 7415 ret = 0; 7416 break; 7417 #elif defined(TARGET_CRIS) 7418 if (arg1 & 0xff) 7419 ret = -TARGET_EINVAL; 7420 else { 7421 ((CPUCRISState *) cpu_env)->pregs[PR_PID] = arg1; 7422 ret = 0; 7423 } 7424 break; 7425 #elif defined(TARGET_I386) && defined(TARGET_ABI32) 7426 ret = do_set_thread_area(cpu_env, arg1); 7427 break; 7428 #else 7429 goto unimplemented_nowarn; 7430 #endif 7431 #endif 7432 #ifdef TARGET_NR_get_thread_area 7433 case TARGET_NR_get_thread_area: 7434 #if defined(TARGET_I386) && defined(TARGET_ABI32) 7435 ret = do_get_thread_area(cpu_env, arg1); 7436 #else 7437 goto unimplemented_nowarn; 7438 #endif 7439 #endif 7440 #ifdef TARGET_NR_getdomainname 7441 case TARGET_NR_getdomainname: 7442 goto unimplemented_nowarn; 7443 #endif 7444 7445 #ifdef TARGET_NR_clock_gettime 7446 case TARGET_NR_clock_gettime: 7447 { 7448 struct timespec ts; 7449 ret = get_errno(clock_gettime(arg1, &ts)); 7450 if (!is_error(ret)) { 7451 host_to_target_timespec(arg2, &ts); 7452 } 7453 break; 7454 } 7455 #endif 7456 #ifdef TARGET_NR_clock_getres 7457 case TARGET_NR_clock_getres: 7458 { 7459 struct timespec ts; 7460 ret = get_errno(clock_getres(arg1, &ts)); 7461 if (!is_error(ret)) { 7462 host_to_target_timespec(arg2, &ts); 7463 } 7464 break; 7465 } 7466 #endif 7467 #ifdef TARGET_NR_clock_nanosleep 7468 case TARGET_NR_clock_nanosleep: 7469 { 7470 struct timespec ts; 7471 target_to_host_timespec(&ts, arg3); 7472 ret = get_errno(clock_nanosleep(arg1, arg2, &ts, arg4 ? &ts : NULL)); 7473 if (arg4) 7474 host_to_target_timespec(arg4, &ts); 7475 break; 7476 } 7477 #endif 7478 7479 #if defined(TARGET_NR_set_tid_address) && defined(__NR_set_tid_address) 7480 case TARGET_NR_set_tid_address: 7481 ret = get_errno(set_tid_address((int *)g2h(arg1))); 7482 break; 7483 #endif 7484 7485 #if defined(TARGET_NR_tkill) && defined(__NR_tkill) 7486 case TARGET_NR_tkill: 7487 ret = get_errno(sys_tkill((int)arg1, target_to_host_signal(arg2))); 7488 break; 7489 #endif 7490 7491 #if defined(TARGET_NR_tgkill) && defined(__NR_tgkill) 7492 case TARGET_NR_tgkill: 7493 ret = get_errno(sys_tgkill((int)arg1, (int)arg2, 7494 target_to_host_signal(arg3))); 7495 break; 7496 #endif 7497 7498 #ifdef TARGET_NR_set_robust_list 7499 case TARGET_NR_set_robust_list: 7500 goto unimplemented_nowarn; 7501 #endif 7502 7503 #if defined(TARGET_NR_utimensat) && defined(__NR_utimensat) 7504 case TARGET_NR_utimensat: 7505 { 7506 struct timespec *tsp, ts[2]; 7507 if (!arg3) { 7508 tsp = NULL; 7509 } else { 7510 target_to_host_timespec(ts, arg3); 7511 target_to_host_timespec(ts+1, arg3+sizeof(struct target_timespec)); 7512 tsp = ts; 7513 } 7514 if (!arg2) 7515 ret = get_errno(sys_utimensat(arg1, NULL, tsp, arg4)); 7516 else { 7517 if (!(p = lock_user_string(arg2))) { 7518 ret = -TARGET_EFAULT; 7519 goto fail; 7520 } 7521 ret = get_errno(sys_utimensat(arg1, path(p), tsp, arg4)); 7522 unlock_user(p, arg2, 0); 7523 } 7524 } 7525 break; 7526 #endif 7527 #if defined(CONFIG_USE_NPTL) 7528 case TARGET_NR_futex: 7529 ret = do_futex(arg1, arg2, arg3, arg4, arg5, arg6); 7530 break; 7531 #endif 7532 #if defined(TARGET_NR_inotify_init) && defined(__NR_inotify_init) 7533 case TARGET_NR_inotify_init: 7534 ret = get_errno(sys_inotify_init()); 7535 break; 7536 #endif 7537 #ifdef CONFIG_INOTIFY1 7538 #if defined(TARGET_NR_inotify_init1) && defined(__NR_inotify_init1) 7539 case TARGET_NR_inotify_init1: 7540 ret = get_errno(sys_inotify_init1(arg1)); 7541 break; 7542 #endif 7543 #endif 7544 #if defined(TARGET_NR_inotify_add_watch) && defined(__NR_inotify_add_watch) 7545 case TARGET_NR_inotify_add_watch: 7546 p = lock_user_string(arg2); 7547 ret = get_errno(sys_inotify_add_watch(arg1, path(p), arg3)); 7548 unlock_user(p, arg2, 0); 7549 break; 7550 #endif 7551 #if defined(TARGET_NR_inotify_rm_watch) && defined(__NR_inotify_rm_watch) 7552 case TARGET_NR_inotify_rm_watch: 7553 ret = get_errno(sys_inotify_rm_watch(arg1, arg2)); 7554 break; 7555 #endif 7556 7557 #if defined(TARGET_NR_mq_open) && defined(__NR_mq_open) 7558 case TARGET_NR_mq_open: 7559 { 7560 struct mq_attr posix_mq_attr; 7561 7562 p = lock_user_string(arg1 - 1); 7563 if (arg4 != 0) 7564 copy_from_user_mq_attr (&posix_mq_attr, arg4); 7565 ret = get_errno(mq_open(p, arg2, arg3, &posix_mq_attr)); 7566 unlock_user (p, arg1, 0); 7567 } 7568 break; 7569 7570 case TARGET_NR_mq_unlink: 7571 p = lock_user_string(arg1 - 1); 7572 ret = get_errno(mq_unlink(p)); 7573 unlock_user (p, arg1, 0); 7574 break; 7575 7576 case TARGET_NR_mq_timedsend: 7577 { 7578 struct timespec ts; 7579 7580 p = lock_user (VERIFY_READ, arg2, arg3, 1); 7581 if (arg5 != 0) { 7582 target_to_host_timespec(&ts, arg5); 7583 ret = get_errno(mq_timedsend(arg1, p, arg3, arg4, &ts)); 7584 host_to_target_timespec(arg5, &ts); 7585 } 7586 else 7587 ret = get_errno(mq_send(arg1, p, arg3, arg4)); 7588 unlock_user (p, arg2, arg3); 7589 } 7590 break; 7591 7592 case TARGET_NR_mq_timedreceive: 7593 { 7594 struct timespec ts; 7595 unsigned int prio; 7596 7597 p = lock_user (VERIFY_READ, arg2, arg3, 1); 7598 if (arg5 != 0) { 7599 target_to_host_timespec(&ts, arg5); 7600 ret = get_errno(mq_timedreceive(arg1, p, arg3, &prio, &ts)); 7601 host_to_target_timespec(arg5, &ts); 7602 } 7603 else 7604 ret = get_errno(mq_receive(arg1, p, arg3, &prio)); 7605 unlock_user (p, arg2, arg3); 7606 if (arg4 != 0) 7607 put_user_u32(prio, arg4); 7608 } 7609 break; 7610 7611 /* Not implemented for now... */ 7612 /* case TARGET_NR_mq_notify: */ 7613 /* break; */ 7614 7615 case TARGET_NR_mq_getsetattr: 7616 { 7617 struct mq_attr posix_mq_attr_in, posix_mq_attr_out; 7618 ret = 0; 7619 if (arg3 != 0) { 7620 ret = mq_getattr(arg1, &posix_mq_attr_out); 7621 copy_to_user_mq_attr(arg3, &posix_mq_attr_out); 7622 } 7623 if (arg2 != 0) { 7624 copy_from_user_mq_attr(&posix_mq_attr_in, arg2); 7625 ret |= mq_setattr(arg1, &posix_mq_attr_in, &posix_mq_attr_out); 7626 } 7627 7628 } 7629 break; 7630 #endif 7631 7632 #ifdef CONFIG_SPLICE 7633 #ifdef TARGET_NR_tee 7634 case TARGET_NR_tee: 7635 { 7636 ret = get_errno(tee(arg1,arg2,arg3,arg4)); 7637 } 7638 break; 7639 #endif 7640 #ifdef TARGET_NR_splice 7641 case TARGET_NR_splice: 7642 { 7643 loff_t loff_in, loff_out; 7644 loff_t *ploff_in = NULL, *ploff_out = NULL; 7645 if(arg2) { 7646 get_user_u64(loff_in, arg2); 7647 ploff_in = &loff_in; 7648 } 7649 if(arg4) { 7650 get_user_u64(loff_out, arg2); 7651 ploff_out = &loff_out; 7652 } 7653 ret = get_errno(splice(arg1, ploff_in, arg3, ploff_out, arg5, arg6)); 7654 } 7655 break; 7656 #endif 7657 #ifdef TARGET_NR_vmsplice 7658 case TARGET_NR_vmsplice: 7659 { 7660 int count = arg3; 7661 struct iovec *vec; 7662 7663 vec = alloca(count * sizeof(struct iovec)); 7664 if (lock_iovec(VERIFY_READ, vec, arg2, count, 1) < 0) 7665 goto efault; 7666 ret = get_errno(vmsplice(arg1, vec, count, arg4)); 7667 unlock_iovec(vec, arg2, count, 0); 7668 } 7669 break; 7670 #endif 7671 #endif /* CONFIG_SPLICE */ 7672 #ifdef CONFIG_EVENTFD 7673 #if defined(TARGET_NR_eventfd) 7674 case TARGET_NR_eventfd: 7675 ret = get_errno(eventfd(arg1, 0)); 7676 break; 7677 #endif 7678 #if defined(TARGET_NR_eventfd2) 7679 case TARGET_NR_eventfd2: 7680 ret = get_errno(eventfd(arg1, arg2)); 7681 break; 7682 #endif 7683 #endif /* CONFIG_EVENTFD */ 7684 #if defined(CONFIG_FALLOCATE) && defined(TARGET_NR_fallocate) 7685 case TARGET_NR_fallocate: 7686 ret = get_errno(fallocate(arg1, arg2, arg3, arg4)); 7687 break; 7688 #endif 7689 #if defined(CONFIG_SYNC_FILE_RANGE) 7690 #if defined(TARGET_NR_sync_file_range) 7691 case TARGET_NR_sync_file_range: 7692 #if TARGET_ABI_BITS == 32 7693 ret = get_errno(sync_file_range(arg1, target_offset64(arg2, arg3), 7694 target_offset64(arg4, arg5), arg6)); 7695 #else 7696 ret = get_errno(sync_file_range(arg1, arg2, arg3, arg4)); 7697 #endif 7698 break; 7699 #endif 7700 #if defined(TARGET_NR_sync_file_range2) 7701 case TARGET_NR_sync_file_range2: 7702 /* This is like sync_file_range but the arguments are reordered */ 7703 #if TARGET_ABI_BITS == 32 7704 ret = get_errno(sync_file_range(arg1, target_offset64(arg3, arg4), 7705 target_offset64(arg5, arg6), arg2)); 7706 #else 7707 ret = get_errno(sync_file_range(arg1, arg3, arg4, arg2)); 7708 #endif 7709 break; 7710 #endif 7711 #endif 7712 #if defined(CONFIG_EPOLL) 7713 #if defined(TARGET_NR_epoll_create) 7714 case TARGET_NR_epoll_create: 7715 ret = get_errno(epoll_create(arg1)); 7716 break; 7717 #endif 7718 #if defined(TARGET_NR_epoll_create1) && defined(CONFIG_EPOLL_CREATE1) 7719 case TARGET_NR_epoll_create1: 7720 ret = get_errno(epoll_create1(arg1)); 7721 break; 7722 #endif 7723 #if defined(TARGET_NR_epoll_ctl) 7724 case TARGET_NR_epoll_ctl: 7725 { 7726 struct epoll_event ep; 7727 struct epoll_event *epp = 0; 7728 if (arg4) { 7729 struct target_epoll_event *target_ep; 7730 if (!lock_user_struct(VERIFY_READ, target_ep, arg4, 1)) { 7731 goto efault; 7732 } 7733 ep.events = tswap32(target_ep->events); 7734 /* The epoll_data_t union is just opaque data to the kernel, 7735 * so we transfer all 64 bits across and need not worry what 7736 * actual data type it is. 7737 */ 7738 ep.data.u64 = tswap64(target_ep->data.u64); 7739 unlock_user_struct(target_ep, arg4, 0); 7740 epp = &ep; 7741 } 7742 ret = get_errno(epoll_ctl(arg1, arg2, arg3, epp)); 7743 break; 7744 } 7745 #endif 7746 7747 #if defined(TARGET_NR_epoll_pwait) && defined(CONFIG_EPOLL_PWAIT) 7748 #define IMPLEMENT_EPOLL_PWAIT 7749 #endif 7750 #if defined(TARGET_NR_epoll_wait) || defined(IMPLEMENT_EPOLL_PWAIT) 7751 #if defined(TARGET_NR_epoll_wait) 7752 case TARGET_NR_epoll_wait: 7753 #endif 7754 #if defined(IMPLEMENT_EPOLL_PWAIT) 7755 case TARGET_NR_epoll_pwait: 7756 #endif 7757 { 7758 struct target_epoll_event *target_ep; 7759 struct epoll_event *ep; 7760 int epfd = arg1; 7761 int maxevents = arg3; 7762 int timeout = arg4; 7763 7764 target_ep = lock_user(VERIFY_WRITE, arg2, 7765 maxevents * sizeof(struct target_epoll_event), 1); 7766 if (!target_ep) { 7767 goto efault; 7768 } 7769 7770 ep = alloca(maxevents * sizeof(struct epoll_event)); 7771 7772 switch (num) { 7773 #if defined(IMPLEMENT_EPOLL_PWAIT) 7774 case TARGET_NR_epoll_pwait: 7775 { 7776 target_sigset_t *target_set; 7777 sigset_t _set, *set = &_set; 7778 7779 if (arg5) { 7780 target_set = lock_user(VERIFY_READ, arg5, 7781 sizeof(target_sigset_t), 1); 7782 if (!target_set) { 7783 unlock_user(target_ep, arg2, 0); 7784 goto efault; 7785 } 7786 target_to_host_sigset(set, target_set); 7787 unlock_user(target_set, arg5, 0); 7788 } else { 7789 set = NULL; 7790 } 7791 7792 ret = get_errno(epoll_pwait(epfd, ep, maxevents, timeout, set)); 7793 break; 7794 } 7795 #endif 7796 #if defined(TARGET_NR_epoll_wait) 7797 case TARGET_NR_epoll_wait: 7798 ret = get_errno(epoll_wait(epfd, ep, maxevents, timeout)); 7799 break; 7800 #endif 7801 default: 7802 ret = -TARGET_ENOSYS; 7803 } 7804 if (!is_error(ret)) { 7805 int i; 7806 for (i = 0; i < ret; i++) { 7807 target_ep[i].events = tswap32(ep[i].events); 7808 target_ep[i].data.u64 = tswap64(ep[i].data.u64); 7809 } 7810 } 7811 unlock_user(target_ep, arg2, ret * sizeof(struct target_epoll_event)); 7812 break; 7813 } 7814 #endif 7815 #endif 7816 default: 7817 unimplemented: 7818 gemu_log("qemu: Unsupported syscall: %d\n", num); 7819 #if defined(TARGET_NR_setxattr) || defined(TARGET_NR_get_thread_area) || defined(TARGET_NR_getdomainname) || defined(TARGET_NR_set_robust_list) 7820 unimplemented_nowarn: 7821 #endif 7822 ret = -TARGET_ENOSYS; 7823 break; 7824 } 7825 fail: 7826 #ifdef DEBUG 7827 gemu_log(" = " TARGET_ABI_FMT_ld "\n", ret); 7828 #endif 7829 if(do_strace) 7830 print_syscall_ret(num, ret); 7831 return ret; 7832 efault: 7833 ret = -TARGET_EFAULT; 7834 goto fail; 7835 } 7836