1 /* 2 * Emulation of Linux signals 3 * 4 * Copyright (c) 2003 Fabrice Bellard 5 * 6 * This program is free software; you can redistribute it and/or modify 7 * it under the terms of the GNU General Public License as published by 8 * the Free Software Foundation; either version 2 of the License, or 9 * (at your option) any later version. 10 * 11 * This program is distributed in the hope that it will be useful, 12 * but WITHOUT ANY WARRANTY; without even the implied warranty of 13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 14 * GNU General Public License for more details. 15 * 16 * You should have received a copy of the GNU General Public License 17 * along with this program; if not, see <http://www.gnu.org/licenses/>. 18 */ 19 #include "qemu/osdep.h" 20 #include "qemu/bitops.h" 21 #include <sys/ucontext.h> 22 #include <sys/resource.h> 23 24 #include "qemu.h" 25 #include "qemu-common.h" 26 #include "target_signal.h" 27 #include "trace.h" 28 29 static struct target_sigaltstack target_sigaltstack_used = { 30 .ss_sp = 0, 31 .ss_size = 0, 32 .ss_flags = TARGET_SS_DISABLE, 33 }; 34 35 static struct target_sigaction sigact_table[TARGET_NSIG]; 36 37 static void host_signal_handler(int host_signum, siginfo_t *info, 38 void *puc); 39 40 static uint8_t host_to_target_signal_table[_NSIG] = { 41 [SIGHUP] = TARGET_SIGHUP, 42 [SIGINT] = TARGET_SIGINT, 43 [SIGQUIT] = TARGET_SIGQUIT, 44 [SIGILL] = TARGET_SIGILL, 45 [SIGTRAP] = TARGET_SIGTRAP, 46 [SIGABRT] = TARGET_SIGABRT, 47 /* [SIGIOT] = TARGET_SIGIOT,*/ 48 [SIGBUS] = TARGET_SIGBUS, 49 [SIGFPE] = TARGET_SIGFPE, 50 [SIGKILL] = TARGET_SIGKILL, 51 [SIGUSR1] = TARGET_SIGUSR1, 52 [SIGSEGV] = TARGET_SIGSEGV, 53 [SIGUSR2] = TARGET_SIGUSR2, 54 [SIGPIPE] = TARGET_SIGPIPE, 55 [SIGALRM] = TARGET_SIGALRM, 56 [SIGTERM] = TARGET_SIGTERM, 57 #ifdef SIGSTKFLT 58 [SIGSTKFLT] = TARGET_SIGSTKFLT, 59 #endif 60 [SIGCHLD] = TARGET_SIGCHLD, 61 [SIGCONT] = TARGET_SIGCONT, 62 [SIGSTOP] = TARGET_SIGSTOP, 63 [SIGTSTP] = TARGET_SIGTSTP, 64 [SIGTTIN] = TARGET_SIGTTIN, 65 [SIGTTOU] = TARGET_SIGTTOU, 66 [SIGURG] = TARGET_SIGURG, 67 [SIGXCPU] = TARGET_SIGXCPU, 68 [SIGXFSZ] = TARGET_SIGXFSZ, 69 [SIGVTALRM] = TARGET_SIGVTALRM, 70 [SIGPROF] = TARGET_SIGPROF, 71 [SIGWINCH] = TARGET_SIGWINCH, 72 [SIGIO] = TARGET_SIGIO, 73 [SIGPWR] = TARGET_SIGPWR, 74 [SIGSYS] = TARGET_SIGSYS, 75 /* next signals stay the same */ 76 /* Nasty hack: Reverse SIGRTMIN and SIGRTMAX to avoid overlap with 77 host libpthread signals. This assumes no one actually uses SIGRTMAX :-/ 78 To fix this properly we need to do manual signal delivery multiplexed 79 over a single host signal. */ 80 [__SIGRTMIN] = __SIGRTMAX, 81 [__SIGRTMAX] = __SIGRTMIN, 82 }; 83 static uint8_t target_to_host_signal_table[_NSIG]; 84 85 static inline int on_sig_stack(unsigned long sp) 86 { 87 return (sp - target_sigaltstack_used.ss_sp 88 < target_sigaltstack_used.ss_size); 89 } 90 91 static inline int sas_ss_flags(unsigned long sp) 92 { 93 return (target_sigaltstack_used.ss_size == 0 ? SS_DISABLE 94 : on_sig_stack(sp) ? SS_ONSTACK : 0); 95 } 96 97 int host_to_target_signal(int sig) 98 { 99 if (sig < 0 || sig >= _NSIG) 100 return sig; 101 return host_to_target_signal_table[sig]; 102 } 103 104 int target_to_host_signal(int sig) 105 { 106 if (sig < 0 || sig >= _NSIG) 107 return sig; 108 return target_to_host_signal_table[sig]; 109 } 110 111 static inline void target_sigemptyset(target_sigset_t *set) 112 { 113 memset(set, 0, sizeof(*set)); 114 } 115 116 static inline void target_sigaddset(target_sigset_t *set, int signum) 117 { 118 signum--; 119 abi_ulong mask = (abi_ulong)1 << (signum % TARGET_NSIG_BPW); 120 set->sig[signum / TARGET_NSIG_BPW] |= mask; 121 } 122 123 static inline int target_sigismember(const target_sigset_t *set, int signum) 124 { 125 signum--; 126 abi_ulong mask = (abi_ulong)1 << (signum % TARGET_NSIG_BPW); 127 return ((set->sig[signum / TARGET_NSIG_BPW] & mask) != 0); 128 } 129 130 static void host_to_target_sigset_internal(target_sigset_t *d, 131 const sigset_t *s) 132 { 133 int i; 134 target_sigemptyset(d); 135 for (i = 1; i <= TARGET_NSIG; i++) { 136 if (sigismember(s, i)) { 137 target_sigaddset(d, host_to_target_signal(i)); 138 } 139 } 140 } 141 142 void host_to_target_sigset(target_sigset_t *d, const sigset_t *s) 143 { 144 target_sigset_t d1; 145 int i; 146 147 host_to_target_sigset_internal(&d1, s); 148 for(i = 0;i < TARGET_NSIG_WORDS; i++) 149 d->sig[i] = tswapal(d1.sig[i]); 150 } 151 152 static void target_to_host_sigset_internal(sigset_t *d, 153 const target_sigset_t *s) 154 { 155 int i; 156 sigemptyset(d); 157 for (i = 1; i <= TARGET_NSIG; i++) { 158 if (target_sigismember(s, i)) { 159 sigaddset(d, target_to_host_signal(i)); 160 } 161 } 162 } 163 164 void target_to_host_sigset(sigset_t *d, const target_sigset_t *s) 165 { 166 target_sigset_t s1; 167 int i; 168 169 for(i = 0;i < TARGET_NSIG_WORDS; i++) 170 s1.sig[i] = tswapal(s->sig[i]); 171 target_to_host_sigset_internal(d, &s1); 172 } 173 174 void host_to_target_old_sigset(abi_ulong *old_sigset, 175 const sigset_t *sigset) 176 { 177 target_sigset_t d; 178 host_to_target_sigset(&d, sigset); 179 *old_sigset = d.sig[0]; 180 } 181 182 void target_to_host_old_sigset(sigset_t *sigset, 183 const abi_ulong *old_sigset) 184 { 185 target_sigset_t d; 186 int i; 187 188 d.sig[0] = *old_sigset; 189 for(i = 1;i < TARGET_NSIG_WORDS; i++) 190 d.sig[i] = 0; 191 target_to_host_sigset(sigset, &d); 192 } 193 194 int block_signals(void) 195 { 196 TaskState *ts = (TaskState *)thread_cpu->opaque; 197 sigset_t set; 198 199 /* It's OK to block everything including SIGSEGV, because we won't 200 * run any further guest code before unblocking signals in 201 * process_pending_signals(). 202 */ 203 sigfillset(&set); 204 sigprocmask(SIG_SETMASK, &set, 0); 205 206 return atomic_xchg(&ts->signal_pending, 1); 207 } 208 209 /* Wrapper for sigprocmask function 210 * Emulates a sigprocmask in a safe way for the guest. Note that set and oldset 211 * are host signal set, not guest ones. Returns -TARGET_ERESTARTSYS if 212 * a signal was already pending and the syscall must be restarted, or 213 * 0 on success. 214 * If set is NULL, this is guaranteed not to fail. 215 */ 216 int do_sigprocmask(int how, const sigset_t *set, sigset_t *oldset) 217 { 218 TaskState *ts = (TaskState *)thread_cpu->opaque; 219 220 if (oldset) { 221 *oldset = ts->signal_mask; 222 } 223 224 if (set) { 225 int i; 226 227 if (block_signals()) { 228 return -TARGET_ERESTARTSYS; 229 } 230 231 switch (how) { 232 case SIG_BLOCK: 233 sigorset(&ts->signal_mask, &ts->signal_mask, set); 234 break; 235 case SIG_UNBLOCK: 236 for (i = 1; i <= NSIG; ++i) { 237 if (sigismember(set, i)) { 238 sigdelset(&ts->signal_mask, i); 239 } 240 } 241 break; 242 case SIG_SETMASK: 243 ts->signal_mask = *set; 244 break; 245 default: 246 g_assert_not_reached(); 247 } 248 249 /* Silently ignore attempts to change blocking status of KILL or STOP */ 250 sigdelset(&ts->signal_mask, SIGKILL); 251 sigdelset(&ts->signal_mask, SIGSTOP); 252 } 253 return 0; 254 } 255 256 #if !defined(TARGET_OPENRISC) && !defined(TARGET_UNICORE32) && \ 257 !defined(TARGET_X86_64) 258 /* Just set the guest's signal mask to the specified value; the 259 * caller is assumed to have called block_signals() already. 260 */ 261 static void set_sigmask(const sigset_t *set) 262 { 263 TaskState *ts = (TaskState *)thread_cpu->opaque; 264 265 ts->signal_mask = *set; 266 } 267 #endif 268 269 /* siginfo conversion */ 270 271 static inline void host_to_target_siginfo_noswap(target_siginfo_t *tinfo, 272 const siginfo_t *info) 273 { 274 int sig = host_to_target_signal(info->si_signo); 275 int si_code = info->si_code; 276 int si_type; 277 tinfo->si_signo = sig; 278 tinfo->si_errno = 0; 279 tinfo->si_code = info->si_code; 280 281 /* This memset serves two purposes: 282 * (1) ensure we don't leak random junk to the guest later 283 * (2) placate false positives from gcc about fields 284 * being used uninitialized if it chooses to inline both this 285 * function and tswap_siginfo() into host_to_target_siginfo(). 286 */ 287 memset(tinfo->_sifields._pad, 0, sizeof(tinfo->_sifields._pad)); 288 289 /* This is awkward, because we have to use a combination of 290 * the si_code and si_signo to figure out which of the union's 291 * members are valid. (Within the host kernel it is always possible 292 * to tell, but the kernel carefully avoids giving userspace the 293 * high 16 bits of si_code, so we don't have the information to 294 * do this the easy way...) We therefore make our best guess, 295 * bearing in mind that a guest can spoof most of the si_codes 296 * via rt_sigqueueinfo() if it likes. 297 * 298 * Once we have made our guess, we record it in the top 16 bits of 299 * the si_code, so that tswap_siginfo() later can use it. 300 * tswap_siginfo() will strip these top bits out before writing 301 * si_code to the guest (sign-extending the lower bits). 302 */ 303 304 switch (si_code) { 305 case SI_USER: 306 case SI_TKILL: 307 case SI_KERNEL: 308 /* Sent via kill(), tkill() or tgkill(), or direct from the kernel. 309 * These are the only unspoofable si_code values. 310 */ 311 tinfo->_sifields._kill._pid = info->si_pid; 312 tinfo->_sifields._kill._uid = info->si_uid; 313 si_type = QEMU_SI_KILL; 314 break; 315 default: 316 /* Everything else is spoofable. Make best guess based on signal */ 317 switch (sig) { 318 case TARGET_SIGCHLD: 319 tinfo->_sifields._sigchld._pid = info->si_pid; 320 tinfo->_sifields._sigchld._uid = info->si_uid; 321 tinfo->_sifields._sigchld._status 322 = host_to_target_waitstatus(info->si_status); 323 tinfo->_sifields._sigchld._utime = info->si_utime; 324 tinfo->_sifields._sigchld._stime = info->si_stime; 325 si_type = QEMU_SI_CHLD; 326 break; 327 case TARGET_SIGIO: 328 tinfo->_sifields._sigpoll._band = info->si_band; 329 tinfo->_sifields._sigpoll._fd = info->si_fd; 330 si_type = QEMU_SI_POLL; 331 break; 332 default: 333 /* Assume a sigqueue()/mq_notify()/rt_sigqueueinfo() source. */ 334 tinfo->_sifields._rt._pid = info->si_pid; 335 tinfo->_sifields._rt._uid = info->si_uid; 336 /* XXX: potential problem if 64 bit */ 337 tinfo->_sifields._rt._sigval.sival_ptr 338 = (abi_ulong)(unsigned long)info->si_value.sival_ptr; 339 si_type = QEMU_SI_RT; 340 break; 341 } 342 break; 343 } 344 345 tinfo->si_code = deposit32(si_code, 16, 16, si_type); 346 } 347 348 static void tswap_siginfo(target_siginfo_t *tinfo, 349 const target_siginfo_t *info) 350 { 351 int si_type = extract32(info->si_code, 16, 16); 352 int si_code = sextract32(info->si_code, 0, 16); 353 354 __put_user(info->si_signo, &tinfo->si_signo); 355 __put_user(info->si_errno, &tinfo->si_errno); 356 __put_user(si_code, &tinfo->si_code); 357 358 /* We can use our internal marker of which fields in the structure 359 * are valid, rather than duplicating the guesswork of 360 * host_to_target_siginfo_noswap() here. 361 */ 362 switch (si_type) { 363 case QEMU_SI_KILL: 364 __put_user(info->_sifields._kill._pid, &tinfo->_sifields._kill._pid); 365 __put_user(info->_sifields._kill._uid, &tinfo->_sifields._kill._uid); 366 break; 367 case QEMU_SI_TIMER: 368 __put_user(info->_sifields._timer._timer1, 369 &tinfo->_sifields._timer._timer1); 370 __put_user(info->_sifields._timer._timer2, 371 &tinfo->_sifields._timer._timer2); 372 break; 373 case QEMU_SI_POLL: 374 __put_user(info->_sifields._sigpoll._band, 375 &tinfo->_sifields._sigpoll._band); 376 __put_user(info->_sifields._sigpoll._fd, 377 &tinfo->_sifields._sigpoll._fd); 378 break; 379 case QEMU_SI_FAULT: 380 __put_user(info->_sifields._sigfault._addr, 381 &tinfo->_sifields._sigfault._addr); 382 break; 383 case QEMU_SI_CHLD: 384 __put_user(info->_sifields._sigchld._pid, 385 &tinfo->_sifields._sigchld._pid); 386 __put_user(info->_sifields._sigchld._uid, 387 &tinfo->_sifields._sigchld._uid); 388 __put_user(info->_sifields._sigchld._status, 389 &tinfo->_sifields._sigchld._status); 390 __put_user(info->_sifields._sigchld._utime, 391 &tinfo->_sifields._sigchld._utime); 392 __put_user(info->_sifields._sigchld._stime, 393 &tinfo->_sifields._sigchld._stime); 394 break; 395 case QEMU_SI_RT: 396 __put_user(info->_sifields._rt._pid, &tinfo->_sifields._rt._pid); 397 __put_user(info->_sifields._rt._uid, &tinfo->_sifields._rt._uid); 398 __put_user(info->_sifields._rt._sigval.sival_ptr, 399 &tinfo->_sifields._rt._sigval.sival_ptr); 400 break; 401 default: 402 g_assert_not_reached(); 403 } 404 } 405 406 void host_to_target_siginfo(target_siginfo_t *tinfo, const siginfo_t *info) 407 { 408 target_siginfo_t tgt_tmp; 409 host_to_target_siginfo_noswap(&tgt_tmp, info); 410 tswap_siginfo(tinfo, &tgt_tmp); 411 } 412 413 /* XXX: we support only POSIX RT signals are used. */ 414 /* XXX: find a solution for 64 bit (additional malloced data is needed) */ 415 void target_to_host_siginfo(siginfo_t *info, const target_siginfo_t *tinfo) 416 { 417 /* This conversion is used only for the rt_sigqueueinfo syscall, 418 * and so we know that the _rt fields are the valid ones. 419 */ 420 abi_ulong sival_ptr; 421 422 __get_user(info->si_signo, &tinfo->si_signo); 423 __get_user(info->si_errno, &tinfo->si_errno); 424 __get_user(info->si_code, &tinfo->si_code); 425 __get_user(info->si_pid, &tinfo->_sifields._rt._pid); 426 __get_user(info->si_uid, &tinfo->_sifields._rt._uid); 427 __get_user(sival_ptr, &tinfo->_sifields._rt._sigval.sival_ptr); 428 info->si_value.sival_ptr = (void *)(long)sival_ptr; 429 } 430 431 static int fatal_signal (int sig) 432 { 433 switch (sig) { 434 case TARGET_SIGCHLD: 435 case TARGET_SIGURG: 436 case TARGET_SIGWINCH: 437 /* Ignored by default. */ 438 return 0; 439 case TARGET_SIGCONT: 440 case TARGET_SIGSTOP: 441 case TARGET_SIGTSTP: 442 case TARGET_SIGTTIN: 443 case TARGET_SIGTTOU: 444 /* Job control signals. */ 445 return 0; 446 default: 447 return 1; 448 } 449 } 450 451 /* returns 1 if given signal should dump core if not handled */ 452 static int core_dump_signal(int sig) 453 { 454 switch (sig) { 455 case TARGET_SIGABRT: 456 case TARGET_SIGFPE: 457 case TARGET_SIGILL: 458 case TARGET_SIGQUIT: 459 case TARGET_SIGSEGV: 460 case TARGET_SIGTRAP: 461 case TARGET_SIGBUS: 462 return (1); 463 default: 464 return (0); 465 } 466 } 467 468 void signal_init(void) 469 { 470 TaskState *ts = (TaskState *)thread_cpu->opaque; 471 struct sigaction act; 472 struct sigaction oact; 473 int i, j; 474 int host_sig; 475 476 /* generate signal conversion tables */ 477 for(i = 1; i < _NSIG; i++) { 478 if (host_to_target_signal_table[i] == 0) 479 host_to_target_signal_table[i] = i; 480 } 481 for(i = 1; i < _NSIG; i++) { 482 j = host_to_target_signal_table[i]; 483 target_to_host_signal_table[j] = i; 484 } 485 486 /* Set the signal mask from the host mask. */ 487 sigprocmask(0, 0, &ts->signal_mask); 488 489 /* set all host signal handlers. ALL signals are blocked during 490 the handlers to serialize them. */ 491 memset(sigact_table, 0, sizeof(sigact_table)); 492 493 sigfillset(&act.sa_mask); 494 act.sa_flags = SA_SIGINFO; 495 act.sa_sigaction = host_signal_handler; 496 for(i = 1; i <= TARGET_NSIG; i++) { 497 host_sig = target_to_host_signal(i); 498 sigaction(host_sig, NULL, &oact); 499 if (oact.sa_sigaction == (void *)SIG_IGN) { 500 sigact_table[i - 1]._sa_handler = TARGET_SIG_IGN; 501 } else if (oact.sa_sigaction == (void *)SIG_DFL) { 502 sigact_table[i - 1]._sa_handler = TARGET_SIG_DFL; 503 } 504 /* If there's already a handler installed then something has 505 gone horribly wrong, so don't even try to handle that case. */ 506 /* Install some handlers for our own use. We need at least 507 SIGSEGV and SIGBUS, to detect exceptions. We can not just 508 trap all signals because it affects syscall interrupt 509 behavior. But do trap all default-fatal signals. */ 510 if (fatal_signal (i)) 511 sigaction(host_sig, &act, NULL); 512 } 513 } 514 515 516 /* abort execution with signal */ 517 static void QEMU_NORETURN force_sig(int target_sig) 518 { 519 CPUState *cpu = thread_cpu; 520 CPUArchState *env = cpu->env_ptr; 521 TaskState *ts = (TaskState *)cpu->opaque; 522 int host_sig, core_dumped = 0; 523 struct sigaction act; 524 525 host_sig = target_to_host_signal(target_sig); 526 trace_user_force_sig(env, target_sig, host_sig); 527 gdb_signalled(env, target_sig); 528 529 /* dump core if supported by target binary format */ 530 if (core_dump_signal(target_sig) && (ts->bprm->core_dump != NULL)) { 531 stop_all_tasks(); 532 core_dumped = 533 ((*ts->bprm->core_dump)(target_sig, env) == 0); 534 } 535 if (core_dumped) { 536 /* we already dumped the core of target process, we don't want 537 * a coredump of qemu itself */ 538 struct rlimit nodump; 539 getrlimit(RLIMIT_CORE, &nodump); 540 nodump.rlim_cur=0; 541 setrlimit(RLIMIT_CORE, &nodump); 542 (void) fprintf(stderr, "qemu: uncaught target signal %d (%s) - %s\n", 543 target_sig, strsignal(host_sig), "core dumped" ); 544 } 545 546 /* The proper exit code for dying from an uncaught signal is 547 * -<signal>. The kernel doesn't allow exit() or _exit() to pass 548 * a negative value. To get the proper exit code we need to 549 * actually die from an uncaught signal. Here the default signal 550 * handler is installed, we send ourself a signal and we wait for 551 * it to arrive. */ 552 sigfillset(&act.sa_mask); 553 act.sa_handler = SIG_DFL; 554 act.sa_flags = 0; 555 sigaction(host_sig, &act, NULL); 556 557 /* For some reason raise(host_sig) doesn't send the signal when 558 * statically linked on x86-64. */ 559 kill(getpid(), host_sig); 560 561 /* Make sure the signal isn't masked (just reuse the mask inside 562 of act) */ 563 sigdelset(&act.sa_mask, host_sig); 564 sigsuspend(&act.sa_mask); 565 566 /* unreachable */ 567 abort(); 568 } 569 570 /* queue a signal so that it will be send to the virtual CPU as soon 571 as possible */ 572 int queue_signal(CPUArchState *env, int sig, target_siginfo_t *info) 573 { 574 CPUState *cpu = ENV_GET_CPU(env); 575 TaskState *ts = cpu->opaque; 576 577 trace_user_queue_signal(env, sig); 578 579 /* Currently all callers define siginfo structures which 580 * use the _sifields._sigfault union member, so we can 581 * set the type here. If that changes we should push this 582 * out so the si_type is passed in by callers. 583 */ 584 info->si_code = deposit32(info->si_code, 16, 16, QEMU_SI_FAULT); 585 586 ts->sync_signal.info = *info; 587 ts->sync_signal.pending = sig; 588 /* signal that a new signal is pending */ 589 atomic_set(&ts->signal_pending, 1); 590 return 1; /* indicates that the signal was queued */ 591 } 592 593 #ifndef HAVE_SAFE_SYSCALL 594 static inline void rewind_if_in_safe_syscall(void *puc) 595 { 596 /* Default version: never rewind */ 597 } 598 #endif 599 600 static void host_signal_handler(int host_signum, siginfo_t *info, 601 void *puc) 602 { 603 CPUArchState *env = thread_cpu->env_ptr; 604 CPUState *cpu = ENV_GET_CPU(env); 605 TaskState *ts = cpu->opaque; 606 607 int sig; 608 target_siginfo_t tinfo; 609 ucontext_t *uc = puc; 610 struct emulated_sigtable *k; 611 612 /* the CPU emulator uses some host signals to detect exceptions, 613 we forward to it some signals */ 614 if ((host_signum == SIGSEGV || host_signum == SIGBUS) 615 && info->si_code > 0) { 616 if (cpu_signal_handler(host_signum, info, puc)) 617 return; 618 } 619 620 /* get target signal number */ 621 sig = host_to_target_signal(host_signum); 622 if (sig < 1 || sig > TARGET_NSIG) 623 return; 624 trace_user_host_signal(env, host_signum, sig); 625 626 rewind_if_in_safe_syscall(puc); 627 628 host_to_target_siginfo_noswap(&tinfo, info); 629 k = &ts->sigtab[sig - 1]; 630 k->info = tinfo; 631 k->pending = sig; 632 ts->signal_pending = 1; 633 634 /* Block host signals until target signal handler entered. We 635 * can't block SIGSEGV or SIGBUS while we're executing guest 636 * code in case the guest code provokes one in the window between 637 * now and it getting out to the main loop. Signals will be 638 * unblocked again in process_pending_signals(). 639 * 640 * WARNING: we cannot use sigfillset() here because the uc_sigmask 641 * field is a kernel sigset_t, which is much smaller than the 642 * libc sigset_t which sigfillset() operates on. Using sigfillset() 643 * would write 0xff bytes off the end of the structure and trash 644 * data on the struct. 645 * We can't use sizeof(uc->uc_sigmask) either, because the libc 646 * headers define the struct field with the wrong (too large) type. 647 */ 648 memset(&uc->uc_sigmask, 0xff, SIGSET_T_SIZE); 649 sigdelset(&uc->uc_sigmask, SIGSEGV); 650 sigdelset(&uc->uc_sigmask, SIGBUS); 651 652 /* interrupt the virtual CPU as soon as possible */ 653 cpu_exit(thread_cpu); 654 } 655 656 /* do_sigaltstack() returns target values and errnos. */ 657 /* compare linux/kernel/signal.c:do_sigaltstack() */ 658 abi_long do_sigaltstack(abi_ulong uss_addr, abi_ulong uoss_addr, abi_ulong sp) 659 { 660 int ret; 661 struct target_sigaltstack oss; 662 663 /* XXX: test errors */ 664 if(uoss_addr) 665 { 666 __put_user(target_sigaltstack_used.ss_sp, &oss.ss_sp); 667 __put_user(target_sigaltstack_used.ss_size, &oss.ss_size); 668 __put_user(sas_ss_flags(sp), &oss.ss_flags); 669 } 670 671 if(uss_addr) 672 { 673 struct target_sigaltstack *uss; 674 struct target_sigaltstack ss; 675 size_t minstacksize = TARGET_MINSIGSTKSZ; 676 677 #if defined(TARGET_PPC64) 678 /* ELF V2 for PPC64 has a 4K minimum stack size for signal handlers */ 679 struct image_info *image = ((TaskState *)thread_cpu->opaque)->info; 680 if (get_ppc64_abi(image) > 1) { 681 minstacksize = 4096; 682 } 683 #endif 684 685 ret = -TARGET_EFAULT; 686 if (!lock_user_struct(VERIFY_READ, uss, uss_addr, 1)) { 687 goto out; 688 } 689 __get_user(ss.ss_sp, &uss->ss_sp); 690 __get_user(ss.ss_size, &uss->ss_size); 691 __get_user(ss.ss_flags, &uss->ss_flags); 692 unlock_user_struct(uss, uss_addr, 0); 693 694 ret = -TARGET_EPERM; 695 if (on_sig_stack(sp)) 696 goto out; 697 698 ret = -TARGET_EINVAL; 699 if (ss.ss_flags != TARGET_SS_DISABLE 700 && ss.ss_flags != TARGET_SS_ONSTACK 701 && ss.ss_flags != 0) 702 goto out; 703 704 if (ss.ss_flags == TARGET_SS_DISABLE) { 705 ss.ss_size = 0; 706 ss.ss_sp = 0; 707 } else { 708 ret = -TARGET_ENOMEM; 709 if (ss.ss_size < minstacksize) { 710 goto out; 711 } 712 } 713 714 target_sigaltstack_used.ss_sp = ss.ss_sp; 715 target_sigaltstack_used.ss_size = ss.ss_size; 716 } 717 718 if (uoss_addr) { 719 ret = -TARGET_EFAULT; 720 if (copy_to_user(uoss_addr, &oss, sizeof(oss))) 721 goto out; 722 } 723 724 ret = 0; 725 out: 726 return ret; 727 } 728 729 /* do_sigaction() return target values and host errnos */ 730 int do_sigaction(int sig, const struct target_sigaction *act, 731 struct target_sigaction *oact) 732 { 733 struct target_sigaction *k; 734 struct sigaction act1; 735 int host_sig; 736 int ret = 0; 737 738 if (sig < 1 || sig > TARGET_NSIG || sig == TARGET_SIGKILL || sig == TARGET_SIGSTOP) { 739 return -TARGET_EINVAL; 740 } 741 742 if (block_signals()) { 743 return -TARGET_ERESTARTSYS; 744 } 745 746 k = &sigact_table[sig - 1]; 747 if (oact) { 748 __put_user(k->_sa_handler, &oact->_sa_handler); 749 __put_user(k->sa_flags, &oact->sa_flags); 750 #if !defined(TARGET_MIPS) 751 __put_user(k->sa_restorer, &oact->sa_restorer); 752 #endif 753 /* Not swapped. */ 754 oact->sa_mask = k->sa_mask; 755 } 756 if (act) { 757 /* FIXME: This is not threadsafe. */ 758 __get_user(k->_sa_handler, &act->_sa_handler); 759 __get_user(k->sa_flags, &act->sa_flags); 760 #if !defined(TARGET_MIPS) 761 __get_user(k->sa_restorer, &act->sa_restorer); 762 #endif 763 /* To be swapped in target_to_host_sigset. */ 764 k->sa_mask = act->sa_mask; 765 766 /* we update the host linux signal state */ 767 host_sig = target_to_host_signal(sig); 768 if (host_sig != SIGSEGV && host_sig != SIGBUS) { 769 sigfillset(&act1.sa_mask); 770 act1.sa_flags = SA_SIGINFO; 771 if (k->sa_flags & TARGET_SA_RESTART) 772 act1.sa_flags |= SA_RESTART; 773 /* NOTE: it is important to update the host kernel signal 774 ignore state to avoid getting unexpected interrupted 775 syscalls */ 776 if (k->_sa_handler == TARGET_SIG_IGN) { 777 act1.sa_sigaction = (void *)SIG_IGN; 778 } else if (k->_sa_handler == TARGET_SIG_DFL) { 779 if (fatal_signal (sig)) 780 act1.sa_sigaction = host_signal_handler; 781 else 782 act1.sa_sigaction = (void *)SIG_DFL; 783 } else { 784 act1.sa_sigaction = host_signal_handler; 785 } 786 ret = sigaction(host_sig, &act1, NULL); 787 } 788 } 789 return ret; 790 } 791 792 #if defined(TARGET_I386) && TARGET_ABI_BITS == 32 793 794 /* from the Linux kernel */ 795 796 struct target_fpreg { 797 uint16_t significand[4]; 798 uint16_t exponent; 799 }; 800 801 struct target_fpxreg { 802 uint16_t significand[4]; 803 uint16_t exponent; 804 uint16_t padding[3]; 805 }; 806 807 struct target_xmmreg { 808 abi_ulong element[4]; 809 }; 810 811 struct target_fpstate { 812 /* Regular FPU environment */ 813 abi_ulong cw; 814 abi_ulong sw; 815 abi_ulong tag; 816 abi_ulong ipoff; 817 abi_ulong cssel; 818 abi_ulong dataoff; 819 abi_ulong datasel; 820 struct target_fpreg _st[8]; 821 uint16_t status; 822 uint16_t magic; /* 0xffff = regular FPU data only */ 823 824 /* FXSR FPU environment */ 825 abi_ulong _fxsr_env[6]; /* FXSR FPU env is ignored */ 826 abi_ulong mxcsr; 827 abi_ulong reserved; 828 struct target_fpxreg _fxsr_st[8]; /* FXSR FPU reg data is ignored */ 829 struct target_xmmreg _xmm[8]; 830 abi_ulong padding[56]; 831 }; 832 833 #define X86_FXSR_MAGIC 0x0000 834 835 struct target_sigcontext { 836 uint16_t gs, __gsh; 837 uint16_t fs, __fsh; 838 uint16_t es, __esh; 839 uint16_t ds, __dsh; 840 abi_ulong edi; 841 abi_ulong esi; 842 abi_ulong ebp; 843 abi_ulong esp; 844 abi_ulong ebx; 845 abi_ulong edx; 846 abi_ulong ecx; 847 abi_ulong eax; 848 abi_ulong trapno; 849 abi_ulong err; 850 abi_ulong eip; 851 uint16_t cs, __csh; 852 abi_ulong eflags; 853 abi_ulong esp_at_signal; 854 uint16_t ss, __ssh; 855 abi_ulong fpstate; /* pointer */ 856 abi_ulong oldmask; 857 abi_ulong cr2; 858 }; 859 860 struct target_ucontext { 861 abi_ulong tuc_flags; 862 abi_ulong tuc_link; 863 target_stack_t tuc_stack; 864 struct target_sigcontext tuc_mcontext; 865 target_sigset_t tuc_sigmask; /* mask last for extensibility */ 866 }; 867 868 struct sigframe 869 { 870 abi_ulong pretcode; 871 int sig; 872 struct target_sigcontext sc; 873 struct target_fpstate fpstate; 874 abi_ulong extramask[TARGET_NSIG_WORDS-1]; 875 char retcode[8]; 876 }; 877 878 struct rt_sigframe 879 { 880 abi_ulong pretcode; 881 int sig; 882 abi_ulong pinfo; 883 abi_ulong puc; 884 struct target_siginfo info; 885 struct target_ucontext uc; 886 struct target_fpstate fpstate; 887 char retcode[8]; 888 }; 889 890 /* 891 * Set up a signal frame. 892 */ 893 894 /* XXX: save x87 state */ 895 static void setup_sigcontext(struct target_sigcontext *sc, 896 struct target_fpstate *fpstate, CPUX86State *env, abi_ulong mask, 897 abi_ulong fpstate_addr) 898 { 899 CPUState *cs = CPU(x86_env_get_cpu(env)); 900 uint16_t magic; 901 902 /* already locked in setup_frame() */ 903 __put_user(env->segs[R_GS].selector, (unsigned int *)&sc->gs); 904 __put_user(env->segs[R_FS].selector, (unsigned int *)&sc->fs); 905 __put_user(env->segs[R_ES].selector, (unsigned int *)&sc->es); 906 __put_user(env->segs[R_DS].selector, (unsigned int *)&sc->ds); 907 __put_user(env->regs[R_EDI], &sc->edi); 908 __put_user(env->regs[R_ESI], &sc->esi); 909 __put_user(env->regs[R_EBP], &sc->ebp); 910 __put_user(env->regs[R_ESP], &sc->esp); 911 __put_user(env->regs[R_EBX], &sc->ebx); 912 __put_user(env->regs[R_EDX], &sc->edx); 913 __put_user(env->regs[R_ECX], &sc->ecx); 914 __put_user(env->regs[R_EAX], &sc->eax); 915 __put_user(cs->exception_index, &sc->trapno); 916 __put_user(env->error_code, &sc->err); 917 __put_user(env->eip, &sc->eip); 918 __put_user(env->segs[R_CS].selector, (unsigned int *)&sc->cs); 919 __put_user(env->eflags, &sc->eflags); 920 __put_user(env->regs[R_ESP], &sc->esp_at_signal); 921 __put_user(env->segs[R_SS].selector, (unsigned int *)&sc->ss); 922 923 cpu_x86_fsave(env, fpstate_addr, 1); 924 fpstate->status = fpstate->sw; 925 magic = 0xffff; 926 __put_user(magic, &fpstate->magic); 927 __put_user(fpstate_addr, &sc->fpstate); 928 929 /* non-iBCS2 extensions.. */ 930 __put_user(mask, &sc->oldmask); 931 __put_user(env->cr[2], &sc->cr2); 932 } 933 934 /* 935 * Determine which stack to use.. 936 */ 937 938 static inline abi_ulong 939 get_sigframe(struct target_sigaction *ka, CPUX86State *env, size_t frame_size) 940 { 941 unsigned long esp; 942 943 /* Default to using normal stack */ 944 esp = env->regs[R_ESP]; 945 /* This is the X/Open sanctioned signal stack switching. */ 946 if (ka->sa_flags & TARGET_SA_ONSTACK) { 947 if (sas_ss_flags(esp) == 0) { 948 esp = target_sigaltstack_used.ss_sp + target_sigaltstack_used.ss_size; 949 } 950 } else { 951 952 /* This is the legacy signal stack switching. */ 953 if ((env->segs[R_SS].selector & 0xffff) != __USER_DS && 954 !(ka->sa_flags & TARGET_SA_RESTORER) && 955 ka->sa_restorer) { 956 esp = (unsigned long) ka->sa_restorer; 957 } 958 } 959 return (esp - frame_size) & -8ul; 960 } 961 962 /* compare linux/arch/i386/kernel/signal.c:setup_frame() */ 963 static void setup_frame(int sig, struct target_sigaction *ka, 964 target_sigset_t *set, CPUX86State *env) 965 { 966 abi_ulong frame_addr; 967 struct sigframe *frame; 968 int i; 969 970 frame_addr = get_sigframe(ka, env, sizeof(*frame)); 971 trace_user_setup_frame(env, frame_addr); 972 973 if (!lock_user_struct(VERIFY_WRITE, frame, frame_addr, 0)) 974 goto give_sigsegv; 975 976 __put_user(sig, &frame->sig); 977 978 setup_sigcontext(&frame->sc, &frame->fpstate, env, set->sig[0], 979 frame_addr + offsetof(struct sigframe, fpstate)); 980 981 for(i = 1; i < TARGET_NSIG_WORDS; i++) { 982 __put_user(set->sig[i], &frame->extramask[i - 1]); 983 } 984 985 /* Set up to return from userspace. If provided, use a stub 986 already in userspace. */ 987 if (ka->sa_flags & TARGET_SA_RESTORER) { 988 __put_user(ka->sa_restorer, &frame->pretcode); 989 } else { 990 uint16_t val16; 991 abi_ulong retcode_addr; 992 retcode_addr = frame_addr + offsetof(struct sigframe, retcode); 993 __put_user(retcode_addr, &frame->pretcode); 994 /* This is popl %eax ; movl $,%eax ; int $0x80 */ 995 val16 = 0xb858; 996 __put_user(val16, (uint16_t *)(frame->retcode+0)); 997 __put_user(TARGET_NR_sigreturn, (int *)(frame->retcode+2)); 998 val16 = 0x80cd; 999 __put_user(val16, (uint16_t *)(frame->retcode+6)); 1000 } 1001 1002 1003 /* Set up registers for signal handler */ 1004 env->regs[R_ESP] = frame_addr; 1005 env->eip = ka->_sa_handler; 1006 1007 cpu_x86_load_seg(env, R_DS, __USER_DS); 1008 cpu_x86_load_seg(env, R_ES, __USER_DS); 1009 cpu_x86_load_seg(env, R_SS, __USER_DS); 1010 cpu_x86_load_seg(env, R_CS, __USER_CS); 1011 env->eflags &= ~TF_MASK; 1012 1013 unlock_user_struct(frame, frame_addr, 1); 1014 1015 return; 1016 1017 give_sigsegv: 1018 if (sig == TARGET_SIGSEGV) { 1019 ka->_sa_handler = TARGET_SIG_DFL; 1020 } 1021 force_sig(TARGET_SIGSEGV /* , current */); 1022 } 1023 1024 /* compare linux/arch/i386/kernel/signal.c:setup_rt_frame() */ 1025 static void setup_rt_frame(int sig, struct target_sigaction *ka, 1026 target_siginfo_t *info, 1027 target_sigset_t *set, CPUX86State *env) 1028 { 1029 abi_ulong frame_addr, addr; 1030 struct rt_sigframe *frame; 1031 int i; 1032 1033 frame_addr = get_sigframe(ka, env, sizeof(*frame)); 1034 trace_user_setup_rt_frame(env, frame_addr); 1035 1036 if (!lock_user_struct(VERIFY_WRITE, frame, frame_addr, 0)) 1037 goto give_sigsegv; 1038 1039 __put_user(sig, &frame->sig); 1040 addr = frame_addr + offsetof(struct rt_sigframe, info); 1041 __put_user(addr, &frame->pinfo); 1042 addr = frame_addr + offsetof(struct rt_sigframe, uc); 1043 __put_user(addr, &frame->puc); 1044 tswap_siginfo(&frame->info, info); 1045 1046 /* Create the ucontext. */ 1047 __put_user(0, &frame->uc.tuc_flags); 1048 __put_user(0, &frame->uc.tuc_link); 1049 __put_user(target_sigaltstack_used.ss_sp, &frame->uc.tuc_stack.ss_sp); 1050 __put_user(sas_ss_flags(get_sp_from_cpustate(env)), 1051 &frame->uc.tuc_stack.ss_flags); 1052 __put_user(target_sigaltstack_used.ss_size, 1053 &frame->uc.tuc_stack.ss_size); 1054 setup_sigcontext(&frame->uc.tuc_mcontext, &frame->fpstate, env, 1055 set->sig[0], frame_addr + offsetof(struct rt_sigframe, fpstate)); 1056 1057 for(i = 0; i < TARGET_NSIG_WORDS; i++) { 1058 __put_user(set->sig[i], &frame->uc.tuc_sigmask.sig[i]); 1059 } 1060 1061 /* Set up to return from userspace. If provided, use a stub 1062 already in userspace. */ 1063 if (ka->sa_flags & TARGET_SA_RESTORER) { 1064 __put_user(ka->sa_restorer, &frame->pretcode); 1065 } else { 1066 uint16_t val16; 1067 addr = frame_addr + offsetof(struct rt_sigframe, retcode); 1068 __put_user(addr, &frame->pretcode); 1069 /* This is movl $,%eax ; int $0x80 */ 1070 __put_user(0xb8, (char *)(frame->retcode+0)); 1071 __put_user(TARGET_NR_rt_sigreturn, (int *)(frame->retcode+1)); 1072 val16 = 0x80cd; 1073 __put_user(val16, (uint16_t *)(frame->retcode+5)); 1074 } 1075 1076 /* Set up registers for signal handler */ 1077 env->regs[R_ESP] = frame_addr; 1078 env->eip = ka->_sa_handler; 1079 1080 cpu_x86_load_seg(env, R_DS, __USER_DS); 1081 cpu_x86_load_seg(env, R_ES, __USER_DS); 1082 cpu_x86_load_seg(env, R_SS, __USER_DS); 1083 cpu_x86_load_seg(env, R_CS, __USER_CS); 1084 env->eflags &= ~TF_MASK; 1085 1086 unlock_user_struct(frame, frame_addr, 1); 1087 1088 return; 1089 1090 give_sigsegv: 1091 if (sig == TARGET_SIGSEGV) { 1092 ka->_sa_handler = TARGET_SIG_DFL; 1093 } 1094 force_sig(TARGET_SIGSEGV /* , current */); 1095 } 1096 1097 static int 1098 restore_sigcontext(CPUX86State *env, struct target_sigcontext *sc) 1099 { 1100 unsigned int err = 0; 1101 abi_ulong fpstate_addr; 1102 unsigned int tmpflags; 1103 1104 cpu_x86_load_seg(env, R_GS, tswap16(sc->gs)); 1105 cpu_x86_load_seg(env, R_FS, tswap16(sc->fs)); 1106 cpu_x86_load_seg(env, R_ES, tswap16(sc->es)); 1107 cpu_x86_load_seg(env, R_DS, tswap16(sc->ds)); 1108 1109 env->regs[R_EDI] = tswapl(sc->edi); 1110 env->regs[R_ESI] = tswapl(sc->esi); 1111 env->regs[R_EBP] = tswapl(sc->ebp); 1112 env->regs[R_ESP] = tswapl(sc->esp); 1113 env->regs[R_EBX] = tswapl(sc->ebx); 1114 env->regs[R_EDX] = tswapl(sc->edx); 1115 env->regs[R_ECX] = tswapl(sc->ecx); 1116 env->regs[R_EAX] = tswapl(sc->eax); 1117 env->eip = tswapl(sc->eip); 1118 1119 cpu_x86_load_seg(env, R_CS, lduw_p(&sc->cs) | 3); 1120 cpu_x86_load_seg(env, R_SS, lduw_p(&sc->ss) | 3); 1121 1122 tmpflags = tswapl(sc->eflags); 1123 env->eflags = (env->eflags & ~0x40DD5) | (tmpflags & 0x40DD5); 1124 // regs->orig_eax = -1; /* disable syscall checks */ 1125 1126 fpstate_addr = tswapl(sc->fpstate); 1127 if (fpstate_addr != 0) { 1128 if (!access_ok(VERIFY_READ, fpstate_addr, 1129 sizeof(struct target_fpstate))) 1130 goto badframe; 1131 cpu_x86_frstor(env, fpstate_addr, 1); 1132 } 1133 1134 return err; 1135 badframe: 1136 return 1; 1137 } 1138 1139 long do_sigreturn(CPUX86State *env) 1140 { 1141 struct sigframe *frame; 1142 abi_ulong frame_addr = env->regs[R_ESP] - 8; 1143 target_sigset_t target_set; 1144 sigset_t set; 1145 int i; 1146 1147 trace_user_do_sigreturn(env, frame_addr); 1148 if (!lock_user_struct(VERIFY_READ, frame, frame_addr, 1)) 1149 goto badframe; 1150 /* set blocked signals */ 1151 __get_user(target_set.sig[0], &frame->sc.oldmask); 1152 for(i = 1; i < TARGET_NSIG_WORDS; i++) { 1153 __get_user(target_set.sig[i], &frame->extramask[i - 1]); 1154 } 1155 1156 target_to_host_sigset_internal(&set, &target_set); 1157 set_sigmask(&set); 1158 1159 /* restore registers */ 1160 if (restore_sigcontext(env, &frame->sc)) 1161 goto badframe; 1162 unlock_user_struct(frame, frame_addr, 0); 1163 return -TARGET_QEMU_ESIGRETURN; 1164 1165 badframe: 1166 unlock_user_struct(frame, frame_addr, 0); 1167 force_sig(TARGET_SIGSEGV); 1168 return 0; 1169 } 1170 1171 long do_rt_sigreturn(CPUX86State *env) 1172 { 1173 abi_ulong frame_addr; 1174 struct rt_sigframe *frame; 1175 sigset_t set; 1176 1177 frame_addr = env->regs[R_ESP] - 4; 1178 trace_user_do_rt_sigreturn(env, frame_addr); 1179 if (!lock_user_struct(VERIFY_READ, frame, frame_addr, 1)) 1180 goto badframe; 1181 target_to_host_sigset(&set, &frame->uc.tuc_sigmask); 1182 set_sigmask(&set); 1183 1184 if (restore_sigcontext(env, &frame->uc.tuc_mcontext)) { 1185 goto badframe; 1186 } 1187 1188 if (do_sigaltstack(frame_addr + offsetof(struct rt_sigframe, uc.tuc_stack), 0, 1189 get_sp_from_cpustate(env)) == -EFAULT) { 1190 goto badframe; 1191 } 1192 1193 unlock_user_struct(frame, frame_addr, 0); 1194 return -TARGET_QEMU_ESIGRETURN; 1195 1196 badframe: 1197 unlock_user_struct(frame, frame_addr, 0); 1198 force_sig(TARGET_SIGSEGV); 1199 return 0; 1200 } 1201 1202 #elif defined(TARGET_AARCH64) 1203 1204 struct target_sigcontext { 1205 uint64_t fault_address; 1206 /* AArch64 registers */ 1207 uint64_t regs[31]; 1208 uint64_t sp; 1209 uint64_t pc; 1210 uint64_t pstate; 1211 /* 4K reserved for FP/SIMD state and future expansion */ 1212 char __reserved[4096] __attribute__((__aligned__(16))); 1213 }; 1214 1215 struct target_ucontext { 1216 abi_ulong tuc_flags; 1217 abi_ulong tuc_link; 1218 target_stack_t tuc_stack; 1219 target_sigset_t tuc_sigmask; 1220 /* glibc uses a 1024-bit sigset_t */ 1221 char __unused[1024 / 8 - sizeof(target_sigset_t)]; 1222 /* last for future expansion */ 1223 struct target_sigcontext tuc_mcontext; 1224 }; 1225 1226 /* 1227 * Header to be used at the beginning of structures extending the user 1228 * context. Such structures must be placed after the rt_sigframe on the stack 1229 * and be 16-byte aligned. The last structure must be a dummy one with the 1230 * magic and size set to 0. 1231 */ 1232 struct target_aarch64_ctx { 1233 uint32_t magic; 1234 uint32_t size; 1235 }; 1236 1237 #define TARGET_FPSIMD_MAGIC 0x46508001 1238 1239 struct target_fpsimd_context { 1240 struct target_aarch64_ctx head; 1241 uint32_t fpsr; 1242 uint32_t fpcr; 1243 uint64_t vregs[32 * 2]; /* really uint128_t vregs[32] */ 1244 }; 1245 1246 /* 1247 * Auxiliary context saved in the sigcontext.__reserved array. Not exported to 1248 * user space as it will change with the addition of new context. User space 1249 * should check the magic/size information. 1250 */ 1251 struct target_aux_context { 1252 struct target_fpsimd_context fpsimd; 1253 /* additional context to be added before "end" */ 1254 struct target_aarch64_ctx end; 1255 }; 1256 1257 struct target_rt_sigframe { 1258 struct target_siginfo info; 1259 struct target_ucontext uc; 1260 uint64_t fp; 1261 uint64_t lr; 1262 uint32_t tramp[2]; 1263 }; 1264 1265 static int target_setup_sigframe(struct target_rt_sigframe *sf, 1266 CPUARMState *env, target_sigset_t *set) 1267 { 1268 int i; 1269 struct target_aux_context *aux = 1270 (struct target_aux_context *)sf->uc.tuc_mcontext.__reserved; 1271 1272 /* set up the stack frame for unwinding */ 1273 __put_user(env->xregs[29], &sf->fp); 1274 __put_user(env->xregs[30], &sf->lr); 1275 1276 for (i = 0; i < 31; i++) { 1277 __put_user(env->xregs[i], &sf->uc.tuc_mcontext.regs[i]); 1278 } 1279 __put_user(env->xregs[31], &sf->uc.tuc_mcontext.sp); 1280 __put_user(env->pc, &sf->uc.tuc_mcontext.pc); 1281 __put_user(pstate_read(env), &sf->uc.tuc_mcontext.pstate); 1282 1283 __put_user(env->exception.vaddress, &sf->uc.tuc_mcontext.fault_address); 1284 1285 for (i = 0; i < TARGET_NSIG_WORDS; i++) { 1286 __put_user(set->sig[i], &sf->uc.tuc_sigmask.sig[i]); 1287 } 1288 1289 for (i = 0; i < 32; i++) { 1290 #ifdef TARGET_WORDS_BIGENDIAN 1291 __put_user(env->vfp.regs[i * 2], &aux->fpsimd.vregs[i * 2 + 1]); 1292 __put_user(env->vfp.regs[i * 2 + 1], &aux->fpsimd.vregs[i * 2]); 1293 #else 1294 __put_user(env->vfp.regs[i * 2], &aux->fpsimd.vregs[i * 2]); 1295 __put_user(env->vfp.regs[i * 2 + 1], &aux->fpsimd.vregs[i * 2 + 1]); 1296 #endif 1297 } 1298 __put_user(vfp_get_fpsr(env), &aux->fpsimd.fpsr); 1299 __put_user(vfp_get_fpcr(env), &aux->fpsimd.fpcr); 1300 __put_user(TARGET_FPSIMD_MAGIC, &aux->fpsimd.head.magic); 1301 __put_user(sizeof(struct target_fpsimd_context), 1302 &aux->fpsimd.head.size); 1303 1304 /* set the "end" magic */ 1305 __put_user(0, &aux->end.magic); 1306 __put_user(0, &aux->end.size); 1307 1308 return 0; 1309 } 1310 1311 static int target_restore_sigframe(CPUARMState *env, 1312 struct target_rt_sigframe *sf) 1313 { 1314 sigset_t set; 1315 int i; 1316 struct target_aux_context *aux = 1317 (struct target_aux_context *)sf->uc.tuc_mcontext.__reserved; 1318 uint32_t magic, size, fpsr, fpcr; 1319 uint64_t pstate; 1320 1321 target_to_host_sigset(&set, &sf->uc.tuc_sigmask); 1322 set_sigmask(&set); 1323 1324 for (i = 0; i < 31; i++) { 1325 __get_user(env->xregs[i], &sf->uc.tuc_mcontext.regs[i]); 1326 } 1327 1328 __get_user(env->xregs[31], &sf->uc.tuc_mcontext.sp); 1329 __get_user(env->pc, &sf->uc.tuc_mcontext.pc); 1330 __get_user(pstate, &sf->uc.tuc_mcontext.pstate); 1331 pstate_write(env, pstate); 1332 1333 __get_user(magic, &aux->fpsimd.head.magic); 1334 __get_user(size, &aux->fpsimd.head.size); 1335 1336 if (magic != TARGET_FPSIMD_MAGIC 1337 || size != sizeof(struct target_fpsimd_context)) { 1338 return 1; 1339 } 1340 1341 for (i = 0; i < 32; i++) { 1342 #ifdef TARGET_WORDS_BIGENDIAN 1343 __get_user(env->vfp.regs[i * 2], &aux->fpsimd.vregs[i * 2 + 1]); 1344 __get_user(env->vfp.regs[i * 2 + 1], &aux->fpsimd.vregs[i * 2]); 1345 #else 1346 __get_user(env->vfp.regs[i * 2], &aux->fpsimd.vregs[i * 2]); 1347 __get_user(env->vfp.regs[i * 2 + 1], &aux->fpsimd.vregs[i * 2 + 1]); 1348 #endif 1349 } 1350 __get_user(fpsr, &aux->fpsimd.fpsr); 1351 vfp_set_fpsr(env, fpsr); 1352 __get_user(fpcr, &aux->fpsimd.fpcr); 1353 vfp_set_fpcr(env, fpcr); 1354 1355 return 0; 1356 } 1357 1358 static abi_ulong get_sigframe(struct target_sigaction *ka, CPUARMState *env) 1359 { 1360 abi_ulong sp; 1361 1362 sp = env->xregs[31]; 1363 1364 /* 1365 * This is the X/Open sanctioned signal stack switching. 1366 */ 1367 if ((ka->sa_flags & TARGET_SA_ONSTACK) && !sas_ss_flags(sp)) { 1368 sp = target_sigaltstack_used.ss_sp + target_sigaltstack_used.ss_size; 1369 } 1370 1371 sp = (sp - sizeof(struct target_rt_sigframe)) & ~15; 1372 1373 return sp; 1374 } 1375 1376 static void target_setup_frame(int usig, struct target_sigaction *ka, 1377 target_siginfo_t *info, target_sigset_t *set, 1378 CPUARMState *env) 1379 { 1380 struct target_rt_sigframe *frame; 1381 abi_ulong frame_addr, return_addr; 1382 1383 frame_addr = get_sigframe(ka, env); 1384 trace_user_setup_frame(env, frame_addr); 1385 if (!lock_user_struct(VERIFY_WRITE, frame, frame_addr, 0)) { 1386 goto give_sigsegv; 1387 } 1388 1389 __put_user(0, &frame->uc.tuc_flags); 1390 __put_user(0, &frame->uc.tuc_link); 1391 1392 __put_user(target_sigaltstack_used.ss_sp, 1393 &frame->uc.tuc_stack.ss_sp); 1394 __put_user(sas_ss_flags(env->xregs[31]), 1395 &frame->uc.tuc_stack.ss_flags); 1396 __put_user(target_sigaltstack_used.ss_size, 1397 &frame->uc.tuc_stack.ss_size); 1398 target_setup_sigframe(frame, env, set); 1399 if (ka->sa_flags & TARGET_SA_RESTORER) { 1400 return_addr = ka->sa_restorer; 1401 } else { 1402 /* mov x8,#__NR_rt_sigreturn; svc #0 */ 1403 __put_user(0xd2801168, &frame->tramp[0]); 1404 __put_user(0xd4000001, &frame->tramp[1]); 1405 return_addr = frame_addr + offsetof(struct target_rt_sigframe, tramp); 1406 } 1407 env->xregs[0] = usig; 1408 env->xregs[31] = frame_addr; 1409 env->xregs[29] = env->xregs[31] + offsetof(struct target_rt_sigframe, fp); 1410 env->pc = ka->_sa_handler; 1411 env->xregs[30] = return_addr; 1412 if (info) { 1413 tswap_siginfo(&frame->info, info); 1414 env->xregs[1] = frame_addr + offsetof(struct target_rt_sigframe, info); 1415 env->xregs[2] = frame_addr + offsetof(struct target_rt_sigframe, uc); 1416 } 1417 1418 unlock_user_struct(frame, frame_addr, 1); 1419 return; 1420 1421 give_sigsegv: 1422 unlock_user_struct(frame, frame_addr, 1); 1423 force_sig(TARGET_SIGSEGV); 1424 } 1425 1426 static void setup_rt_frame(int sig, struct target_sigaction *ka, 1427 target_siginfo_t *info, target_sigset_t *set, 1428 CPUARMState *env) 1429 { 1430 target_setup_frame(sig, ka, info, set, env); 1431 } 1432 1433 static void setup_frame(int sig, struct target_sigaction *ka, 1434 target_sigset_t *set, CPUARMState *env) 1435 { 1436 target_setup_frame(sig, ka, 0, set, env); 1437 } 1438 1439 long do_rt_sigreturn(CPUARMState *env) 1440 { 1441 struct target_rt_sigframe *frame = NULL; 1442 abi_ulong frame_addr = env->xregs[31]; 1443 1444 trace_user_do_rt_sigreturn(env, frame_addr); 1445 if (frame_addr & 15) { 1446 goto badframe; 1447 } 1448 1449 if (!lock_user_struct(VERIFY_READ, frame, frame_addr, 1)) { 1450 goto badframe; 1451 } 1452 1453 if (target_restore_sigframe(env, frame)) { 1454 goto badframe; 1455 } 1456 1457 if (do_sigaltstack(frame_addr + 1458 offsetof(struct target_rt_sigframe, uc.tuc_stack), 1459 0, get_sp_from_cpustate(env)) == -EFAULT) { 1460 goto badframe; 1461 } 1462 1463 unlock_user_struct(frame, frame_addr, 0); 1464 return -TARGET_QEMU_ESIGRETURN; 1465 1466 badframe: 1467 unlock_user_struct(frame, frame_addr, 0); 1468 force_sig(TARGET_SIGSEGV); 1469 return 0; 1470 } 1471 1472 long do_sigreturn(CPUARMState *env) 1473 { 1474 return do_rt_sigreturn(env); 1475 } 1476 1477 #elif defined(TARGET_ARM) 1478 1479 struct target_sigcontext { 1480 abi_ulong trap_no; 1481 abi_ulong error_code; 1482 abi_ulong oldmask; 1483 abi_ulong arm_r0; 1484 abi_ulong arm_r1; 1485 abi_ulong arm_r2; 1486 abi_ulong arm_r3; 1487 abi_ulong arm_r4; 1488 abi_ulong arm_r5; 1489 abi_ulong arm_r6; 1490 abi_ulong arm_r7; 1491 abi_ulong arm_r8; 1492 abi_ulong arm_r9; 1493 abi_ulong arm_r10; 1494 abi_ulong arm_fp; 1495 abi_ulong arm_ip; 1496 abi_ulong arm_sp; 1497 abi_ulong arm_lr; 1498 abi_ulong arm_pc; 1499 abi_ulong arm_cpsr; 1500 abi_ulong fault_address; 1501 }; 1502 1503 struct target_ucontext_v1 { 1504 abi_ulong tuc_flags; 1505 abi_ulong tuc_link; 1506 target_stack_t tuc_stack; 1507 struct target_sigcontext tuc_mcontext; 1508 target_sigset_t tuc_sigmask; /* mask last for extensibility */ 1509 }; 1510 1511 struct target_ucontext_v2 { 1512 abi_ulong tuc_flags; 1513 abi_ulong tuc_link; 1514 target_stack_t tuc_stack; 1515 struct target_sigcontext tuc_mcontext; 1516 target_sigset_t tuc_sigmask; /* mask last for extensibility */ 1517 char __unused[128 - sizeof(target_sigset_t)]; 1518 abi_ulong tuc_regspace[128] __attribute__((__aligned__(8))); 1519 }; 1520 1521 struct target_user_vfp { 1522 uint64_t fpregs[32]; 1523 abi_ulong fpscr; 1524 }; 1525 1526 struct target_user_vfp_exc { 1527 abi_ulong fpexc; 1528 abi_ulong fpinst; 1529 abi_ulong fpinst2; 1530 }; 1531 1532 struct target_vfp_sigframe { 1533 abi_ulong magic; 1534 abi_ulong size; 1535 struct target_user_vfp ufp; 1536 struct target_user_vfp_exc ufp_exc; 1537 } __attribute__((__aligned__(8))); 1538 1539 struct target_iwmmxt_sigframe { 1540 abi_ulong magic; 1541 abi_ulong size; 1542 uint64_t regs[16]; 1543 /* Note that not all the coprocessor control registers are stored here */ 1544 uint32_t wcssf; 1545 uint32_t wcasf; 1546 uint32_t wcgr0; 1547 uint32_t wcgr1; 1548 uint32_t wcgr2; 1549 uint32_t wcgr3; 1550 } __attribute__((__aligned__(8))); 1551 1552 #define TARGET_VFP_MAGIC 0x56465001 1553 #define TARGET_IWMMXT_MAGIC 0x12ef842a 1554 1555 struct sigframe_v1 1556 { 1557 struct target_sigcontext sc; 1558 abi_ulong extramask[TARGET_NSIG_WORDS-1]; 1559 abi_ulong retcode; 1560 }; 1561 1562 struct sigframe_v2 1563 { 1564 struct target_ucontext_v2 uc; 1565 abi_ulong retcode; 1566 }; 1567 1568 struct rt_sigframe_v1 1569 { 1570 abi_ulong pinfo; 1571 abi_ulong puc; 1572 struct target_siginfo info; 1573 struct target_ucontext_v1 uc; 1574 abi_ulong retcode; 1575 }; 1576 1577 struct rt_sigframe_v2 1578 { 1579 struct target_siginfo info; 1580 struct target_ucontext_v2 uc; 1581 abi_ulong retcode; 1582 }; 1583 1584 #define TARGET_CONFIG_CPU_32 1 1585 1586 /* 1587 * For ARM syscalls, we encode the syscall number into the instruction. 1588 */ 1589 #define SWI_SYS_SIGRETURN (0xef000000|(TARGET_NR_sigreturn + ARM_SYSCALL_BASE)) 1590 #define SWI_SYS_RT_SIGRETURN (0xef000000|(TARGET_NR_rt_sigreturn + ARM_SYSCALL_BASE)) 1591 1592 /* 1593 * For Thumb syscalls, we pass the syscall number via r7. We therefore 1594 * need two 16-bit instructions. 1595 */ 1596 #define SWI_THUMB_SIGRETURN (0xdf00 << 16 | 0x2700 | (TARGET_NR_sigreturn)) 1597 #define SWI_THUMB_RT_SIGRETURN (0xdf00 << 16 | 0x2700 | (TARGET_NR_rt_sigreturn)) 1598 1599 static const abi_ulong retcodes[4] = { 1600 SWI_SYS_SIGRETURN, SWI_THUMB_SIGRETURN, 1601 SWI_SYS_RT_SIGRETURN, SWI_THUMB_RT_SIGRETURN 1602 }; 1603 1604 1605 static inline int valid_user_regs(CPUARMState *regs) 1606 { 1607 return 1; 1608 } 1609 1610 static void 1611 setup_sigcontext(struct target_sigcontext *sc, /*struct _fpstate *fpstate,*/ 1612 CPUARMState *env, abi_ulong mask) 1613 { 1614 __put_user(env->regs[0], &sc->arm_r0); 1615 __put_user(env->regs[1], &sc->arm_r1); 1616 __put_user(env->regs[2], &sc->arm_r2); 1617 __put_user(env->regs[3], &sc->arm_r3); 1618 __put_user(env->regs[4], &sc->arm_r4); 1619 __put_user(env->regs[5], &sc->arm_r5); 1620 __put_user(env->regs[6], &sc->arm_r6); 1621 __put_user(env->regs[7], &sc->arm_r7); 1622 __put_user(env->regs[8], &sc->arm_r8); 1623 __put_user(env->regs[9], &sc->arm_r9); 1624 __put_user(env->regs[10], &sc->arm_r10); 1625 __put_user(env->regs[11], &sc->arm_fp); 1626 __put_user(env->regs[12], &sc->arm_ip); 1627 __put_user(env->regs[13], &sc->arm_sp); 1628 __put_user(env->regs[14], &sc->arm_lr); 1629 __put_user(env->regs[15], &sc->arm_pc); 1630 #ifdef TARGET_CONFIG_CPU_32 1631 __put_user(cpsr_read(env), &sc->arm_cpsr); 1632 #endif 1633 1634 __put_user(/* current->thread.trap_no */ 0, &sc->trap_no); 1635 __put_user(/* current->thread.error_code */ 0, &sc->error_code); 1636 __put_user(/* current->thread.address */ 0, &sc->fault_address); 1637 __put_user(mask, &sc->oldmask); 1638 } 1639 1640 static inline abi_ulong 1641 get_sigframe(struct target_sigaction *ka, CPUARMState *regs, int framesize) 1642 { 1643 unsigned long sp = regs->regs[13]; 1644 1645 /* 1646 * This is the X/Open sanctioned signal stack switching. 1647 */ 1648 if ((ka->sa_flags & TARGET_SA_ONSTACK) && !sas_ss_flags(sp)) { 1649 sp = target_sigaltstack_used.ss_sp + target_sigaltstack_used.ss_size; 1650 } 1651 /* 1652 * ATPCS B01 mandates 8-byte alignment 1653 */ 1654 return (sp - framesize) & ~7; 1655 } 1656 1657 static void 1658 setup_return(CPUARMState *env, struct target_sigaction *ka, 1659 abi_ulong *rc, abi_ulong frame_addr, int usig, abi_ulong rc_addr) 1660 { 1661 abi_ulong handler = ka->_sa_handler; 1662 abi_ulong retcode; 1663 int thumb = handler & 1; 1664 uint32_t cpsr = cpsr_read(env); 1665 1666 cpsr &= ~CPSR_IT; 1667 if (thumb) { 1668 cpsr |= CPSR_T; 1669 } else { 1670 cpsr &= ~CPSR_T; 1671 } 1672 1673 if (ka->sa_flags & TARGET_SA_RESTORER) { 1674 retcode = ka->sa_restorer; 1675 } else { 1676 unsigned int idx = thumb; 1677 1678 if (ka->sa_flags & TARGET_SA_SIGINFO) { 1679 idx += 2; 1680 } 1681 1682 __put_user(retcodes[idx], rc); 1683 1684 retcode = rc_addr + thumb; 1685 } 1686 1687 env->regs[0] = usig; 1688 env->regs[13] = frame_addr; 1689 env->regs[14] = retcode; 1690 env->regs[15] = handler & (thumb ? ~1 : ~3); 1691 cpsr_write(env, cpsr, CPSR_IT | CPSR_T, CPSRWriteByInstr); 1692 } 1693 1694 static abi_ulong *setup_sigframe_v2_vfp(abi_ulong *regspace, CPUARMState *env) 1695 { 1696 int i; 1697 struct target_vfp_sigframe *vfpframe; 1698 vfpframe = (struct target_vfp_sigframe *)regspace; 1699 __put_user(TARGET_VFP_MAGIC, &vfpframe->magic); 1700 __put_user(sizeof(*vfpframe), &vfpframe->size); 1701 for (i = 0; i < 32; i++) { 1702 __put_user(float64_val(env->vfp.regs[i]), &vfpframe->ufp.fpregs[i]); 1703 } 1704 __put_user(vfp_get_fpscr(env), &vfpframe->ufp.fpscr); 1705 __put_user(env->vfp.xregs[ARM_VFP_FPEXC], &vfpframe->ufp_exc.fpexc); 1706 __put_user(env->vfp.xregs[ARM_VFP_FPINST], &vfpframe->ufp_exc.fpinst); 1707 __put_user(env->vfp.xregs[ARM_VFP_FPINST2], &vfpframe->ufp_exc.fpinst2); 1708 return (abi_ulong*)(vfpframe+1); 1709 } 1710 1711 static abi_ulong *setup_sigframe_v2_iwmmxt(abi_ulong *regspace, 1712 CPUARMState *env) 1713 { 1714 int i; 1715 struct target_iwmmxt_sigframe *iwmmxtframe; 1716 iwmmxtframe = (struct target_iwmmxt_sigframe *)regspace; 1717 __put_user(TARGET_IWMMXT_MAGIC, &iwmmxtframe->magic); 1718 __put_user(sizeof(*iwmmxtframe), &iwmmxtframe->size); 1719 for (i = 0; i < 16; i++) { 1720 __put_user(env->iwmmxt.regs[i], &iwmmxtframe->regs[i]); 1721 } 1722 __put_user(env->vfp.xregs[ARM_IWMMXT_wCSSF], &iwmmxtframe->wcssf); 1723 __put_user(env->vfp.xregs[ARM_IWMMXT_wCASF], &iwmmxtframe->wcssf); 1724 __put_user(env->vfp.xregs[ARM_IWMMXT_wCGR0], &iwmmxtframe->wcgr0); 1725 __put_user(env->vfp.xregs[ARM_IWMMXT_wCGR1], &iwmmxtframe->wcgr1); 1726 __put_user(env->vfp.xregs[ARM_IWMMXT_wCGR2], &iwmmxtframe->wcgr2); 1727 __put_user(env->vfp.xregs[ARM_IWMMXT_wCGR3], &iwmmxtframe->wcgr3); 1728 return (abi_ulong*)(iwmmxtframe+1); 1729 } 1730 1731 static void setup_sigframe_v2(struct target_ucontext_v2 *uc, 1732 target_sigset_t *set, CPUARMState *env) 1733 { 1734 struct target_sigaltstack stack; 1735 int i; 1736 abi_ulong *regspace; 1737 1738 /* Clear all the bits of the ucontext we don't use. */ 1739 memset(uc, 0, offsetof(struct target_ucontext_v2, tuc_mcontext)); 1740 1741 memset(&stack, 0, sizeof(stack)); 1742 __put_user(target_sigaltstack_used.ss_sp, &stack.ss_sp); 1743 __put_user(target_sigaltstack_used.ss_size, &stack.ss_size); 1744 __put_user(sas_ss_flags(get_sp_from_cpustate(env)), &stack.ss_flags); 1745 memcpy(&uc->tuc_stack, &stack, sizeof(stack)); 1746 1747 setup_sigcontext(&uc->tuc_mcontext, env, set->sig[0]); 1748 /* Save coprocessor signal frame. */ 1749 regspace = uc->tuc_regspace; 1750 if (arm_feature(env, ARM_FEATURE_VFP)) { 1751 regspace = setup_sigframe_v2_vfp(regspace, env); 1752 } 1753 if (arm_feature(env, ARM_FEATURE_IWMMXT)) { 1754 regspace = setup_sigframe_v2_iwmmxt(regspace, env); 1755 } 1756 1757 /* Write terminating magic word */ 1758 __put_user(0, regspace); 1759 1760 for(i = 0; i < TARGET_NSIG_WORDS; i++) { 1761 __put_user(set->sig[i], &uc->tuc_sigmask.sig[i]); 1762 } 1763 } 1764 1765 /* compare linux/arch/arm/kernel/signal.c:setup_frame() */ 1766 static void setup_frame_v1(int usig, struct target_sigaction *ka, 1767 target_sigset_t *set, CPUARMState *regs) 1768 { 1769 struct sigframe_v1 *frame; 1770 abi_ulong frame_addr = get_sigframe(ka, regs, sizeof(*frame)); 1771 int i; 1772 1773 trace_user_setup_frame(regs, frame_addr); 1774 if (!lock_user_struct(VERIFY_WRITE, frame, frame_addr, 0)) { 1775 return; 1776 } 1777 1778 setup_sigcontext(&frame->sc, regs, set->sig[0]); 1779 1780 for(i = 1; i < TARGET_NSIG_WORDS; i++) { 1781 __put_user(set->sig[i], &frame->extramask[i - 1]); 1782 } 1783 1784 setup_return(regs, ka, &frame->retcode, frame_addr, usig, 1785 frame_addr + offsetof(struct sigframe_v1, retcode)); 1786 1787 unlock_user_struct(frame, frame_addr, 1); 1788 } 1789 1790 static void setup_frame_v2(int usig, struct target_sigaction *ka, 1791 target_sigset_t *set, CPUARMState *regs) 1792 { 1793 struct sigframe_v2 *frame; 1794 abi_ulong frame_addr = get_sigframe(ka, regs, sizeof(*frame)); 1795 1796 trace_user_setup_frame(regs, frame_addr); 1797 if (!lock_user_struct(VERIFY_WRITE, frame, frame_addr, 0)) { 1798 return; 1799 } 1800 1801 setup_sigframe_v2(&frame->uc, set, regs); 1802 1803 setup_return(regs, ka, &frame->retcode, frame_addr, usig, 1804 frame_addr + offsetof(struct sigframe_v2, retcode)); 1805 1806 unlock_user_struct(frame, frame_addr, 1); 1807 } 1808 1809 static void setup_frame(int usig, struct target_sigaction *ka, 1810 target_sigset_t *set, CPUARMState *regs) 1811 { 1812 if (get_osversion() >= 0x020612) { 1813 setup_frame_v2(usig, ka, set, regs); 1814 } else { 1815 setup_frame_v1(usig, ka, set, regs); 1816 } 1817 } 1818 1819 /* compare linux/arch/arm/kernel/signal.c:setup_rt_frame() */ 1820 static void setup_rt_frame_v1(int usig, struct target_sigaction *ka, 1821 target_siginfo_t *info, 1822 target_sigset_t *set, CPUARMState *env) 1823 { 1824 struct rt_sigframe_v1 *frame; 1825 abi_ulong frame_addr = get_sigframe(ka, env, sizeof(*frame)); 1826 struct target_sigaltstack stack; 1827 int i; 1828 abi_ulong info_addr, uc_addr; 1829 1830 trace_user_setup_rt_frame(env, frame_addr); 1831 if (!lock_user_struct(VERIFY_WRITE, frame, frame_addr, 0)) { 1832 return /* 1 */; 1833 } 1834 1835 info_addr = frame_addr + offsetof(struct rt_sigframe_v1, info); 1836 __put_user(info_addr, &frame->pinfo); 1837 uc_addr = frame_addr + offsetof(struct rt_sigframe_v1, uc); 1838 __put_user(uc_addr, &frame->puc); 1839 tswap_siginfo(&frame->info, info); 1840 1841 /* Clear all the bits of the ucontext we don't use. */ 1842 memset(&frame->uc, 0, offsetof(struct target_ucontext_v1, tuc_mcontext)); 1843 1844 memset(&stack, 0, sizeof(stack)); 1845 __put_user(target_sigaltstack_used.ss_sp, &stack.ss_sp); 1846 __put_user(target_sigaltstack_used.ss_size, &stack.ss_size); 1847 __put_user(sas_ss_flags(get_sp_from_cpustate(env)), &stack.ss_flags); 1848 memcpy(&frame->uc.tuc_stack, &stack, sizeof(stack)); 1849 1850 setup_sigcontext(&frame->uc.tuc_mcontext, env, set->sig[0]); 1851 for(i = 0; i < TARGET_NSIG_WORDS; i++) { 1852 __put_user(set->sig[i], &frame->uc.tuc_sigmask.sig[i]); 1853 } 1854 1855 setup_return(env, ka, &frame->retcode, frame_addr, usig, 1856 frame_addr + offsetof(struct rt_sigframe_v1, retcode)); 1857 1858 env->regs[1] = info_addr; 1859 env->regs[2] = uc_addr; 1860 1861 unlock_user_struct(frame, frame_addr, 1); 1862 } 1863 1864 static void setup_rt_frame_v2(int usig, struct target_sigaction *ka, 1865 target_siginfo_t *info, 1866 target_sigset_t *set, CPUARMState *env) 1867 { 1868 struct rt_sigframe_v2 *frame; 1869 abi_ulong frame_addr = get_sigframe(ka, env, sizeof(*frame)); 1870 abi_ulong info_addr, uc_addr; 1871 1872 trace_user_setup_rt_frame(env, frame_addr); 1873 if (!lock_user_struct(VERIFY_WRITE, frame, frame_addr, 0)) { 1874 return /* 1 */; 1875 } 1876 1877 info_addr = frame_addr + offsetof(struct rt_sigframe_v2, info); 1878 uc_addr = frame_addr + offsetof(struct rt_sigframe_v2, uc); 1879 tswap_siginfo(&frame->info, info); 1880 1881 setup_sigframe_v2(&frame->uc, set, env); 1882 1883 setup_return(env, ka, &frame->retcode, frame_addr, usig, 1884 frame_addr + offsetof(struct rt_sigframe_v2, retcode)); 1885 1886 env->regs[1] = info_addr; 1887 env->regs[2] = uc_addr; 1888 1889 unlock_user_struct(frame, frame_addr, 1); 1890 } 1891 1892 static void setup_rt_frame(int usig, struct target_sigaction *ka, 1893 target_siginfo_t *info, 1894 target_sigset_t *set, CPUARMState *env) 1895 { 1896 if (get_osversion() >= 0x020612) { 1897 setup_rt_frame_v2(usig, ka, info, set, env); 1898 } else { 1899 setup_rt_frame_v1(usig, ka, info, set, env); 1900 } 1901 } 1902 1903 static int 1904 restore_sigcontext(CPUARMState *env, struct target_sigcontext *sc) 1905 { 1906 int err = 0; 1907 uint32_t cpsr; 1908 1909 __get_user(env->regs[0], &sc->arm_r0); 1910 __get_user(env->regs[1], &sc->arm_r1); 1911 __get_user(env->regs[2], &sc->arm_r2); 1912 __get_user(env->regs[3], &sc->arm_r3); 1913 __get_user(env->regs[4], &sc->arm_r4); 1914 __get_user(env->regs[5], &sc->arm_r5); 1915 __get_user(env->regs[6], &sc->arm_r6); 1916 __get_user(env->regs[7], &sc->arm_r7); 1917 __get_user(env->regs[8], &sc->arm_r8); 1918 __get_user(env->regs[9], &sc->arm_r9); 1919 __get_user(env->regs[10], &sc->arm_r10); 1920 __get_user(env->regs[11], &sc->arm_fp); 1921 __get_user(env->regs[12], &sc->arm_ip); 1922 __get_user(env->regs[13], &sc->arm_sp); 1923 __get_user(env->regs[14], &sc->arm_lr); 1924 __get_user(env->regs[15], &sc->arm_pc); 1925 #ifdef TARGET_CONFIG_CPU_32 1926 __get_user(cpsr, &sc->arm_cpsr); 1927 cpsr_write(env, cpsr, CPSR_USER | CPSR_EXEC, CPSRWriteByInstr); 1928 #endif 1929 1930 err |= !valid_user_regs(env); 1931 1932 return err; 1933 } 1934 1935 static long do_sigreturn_v1(CPUARMState *env) 1936 { 1937 abi_ulong frame_addr; 1938 struct sigframe_v1 *frame = NULL; 1939 target_sigset_t set; 1940 sigset_t host_set; 1941 int i; 1942 1943 /* 1944 * Since we stacked the signal on a 64-bit boundary, 1945 * then 'sp' should be word aligned here. If it's 1946 * not, then the user is trying to mess with us. 1947 */ 1948 frame_addr = env->regs[13]; 1949 trace_user_do_sigreturn(env, frame_addr); 1950 if (frame_addr & 7) { 1951 goto badframe; 1952 } 1953 1954 if (!lock_user_struct(VERIFY_READ, frame, frame_addr, 1)) { 1955 goto badframe; 1956 } 1957 1958 __get_user(set.sig[0], &frame->sc.oldmask); 1959 for(i = 1; i < TARGET_NSIG_WORDS; i++) { 1960 __get_user(set.sig[i], &frame->extramask[i - 1]); 1961 } 1962 1963 target_to_host_sigset_internal(&host_set, &set); 1964 set_sigmask(&host_set); 1965 1966 if (restore_sigcontext(env, &frame->sc)) { 1967 goto badframe; 1968 } 1969 1970 #if 0 1971 /* Send SIGTRAP if we're single-stepping */ 1972 if (ptrace_cancel_bpt(current)) 1973 send_sig(SIGTRAP, current, 1); 1974 #endif 1975 unlock_user_struct(frame, frame_addr, 0); 1976 return -TARGET_QEMU_ESIGRETURN; 1977 1978 badframe: 1979 force_sig(TARGET_SIGSEGV /* , current */); 1980 return 0; 1981 } 1982 1983 static abi_ulong *restore_sigframe_v2_vfp(CPUARMState *env, abi_ulong *regspace) 1984 { 1985 int i; 1986 abi_ulong magic, sz; 1987 uint32_t fpscr, fpexc; 1988 struct target_vfp_sigframe *vfpframe; 1989 vfpframe = (struct target_vfp_sigframe *)regspace; 1990 1991 __get_user(magic, &vfpframe->magic); 1992 __get_user(sz, &vfpframe->size); 1993 if (magic != TARGET_VFP_MAGIC || sz != sizeof(*vfpframe)) { 1994 return 0; 1995 } 1996 for (i = 0; i < 32; i++) { 1997 __get_user(float64_val(env->vfp.regs[i]), &vfpframe->ufp.fpregs[i]); 1998 } 1999 __get_user(fpscr, &vfpframe->ufp.fpscr); 2000 vfp_set_fpscr(env, fpscr); 2001 __get_user(fpexc, &vfpframe->ufp_exc.fpexc); 2002 /* Sanitise FPEXC: ensure VFP is enabled, FPINST2 is invalid 2003 * and the exception flag is cleared 2004 */ 2005 fpexc |= (1 << 30); 2006 fpexc &= ~((1 << 31) | (1 << 28)); 2007 env->vfp.xregs[ARM_VFP_FPEXC] = fpexc; 2008 __get_user(env->vfp.xregs[ARM_VFP_FPINST], &vfpframe->ufp_exc.fpinst); 2009 __get_user(env->vfp.xregs[ARM_VFP_FPINST2], &vfpframe->ufp_exc.fpinst2); 2010 return (abi_ulong*)(vfpframe + 1); 2011 } 2012 2013 static abi_ulong *restore_sigframe_v2_iwmmxt(CPUARMState *env, 2014 abi_ulong *regspace) 2015 { 2016 int i; 2017 abi_ulong magic, sz; 2018 struct target_iwmmxt_sigframe *iwmmxtframe; 2019 iwmmxtframe = (struct target_iwmmxt_sigframe *)regspace; 2020 2021 __get_user(magic, &iwmmxtframe->magic); 2022 __get_user(sz, &iwmmxtframe->size); 2023 if (magic != TARGET_IWMMXT_MAGIC || sz != sizeof(*iwmmxtframe)) { 2024 return 0; 2025 } 2026 for (i = 0; i < 16; i++) { 2027 __get_user(env->iwmmxt.regs[i], &iwmmxtframe->regs[i]); 2028 } 2029 __get_user(env->vfp.xregs[ARM_IWMMXT_wCSSF], &iwmmxtframe->wcssf); 2030 __get_user(env->vfp.xregs[ARM_IWMMXT_wCASF], &iwmmxtframe->wcssf); 2031 __get_user(env->vfp.xregs[ARM_IWMMXT_wCGR0], &iwmmxtframe->wcgr0); 2032 __get_user(env->vfp.xregs[ARM_IWMMXT_wCGR1], &iwmmxtframe->wcgr1); 2033 __get_user(env->vfp.xregs[ARM_IWMMXT_wCGR2], &iwmmxtframe->wcgr2); 2034 __get_user(env->vfp.xregs[ARM_IWMMXT_wCGR3], &iwmmxtframe->wcgr3); 2035 return (abi_ulong*)(iwmmxtframe + 1); 2036 } 2037 2038 static int do_sigframe_return_v2(CPUARMState *env, target_ulong frame_addr, 2039 struct target_ucontext_v2 *uc) 2040 { 2041 sigset_t host_set; 2042 abi_ulong *regspace; 2043 2044 target_to_host_sigset(&host_set, &uc->tuc_sigmask); 2045 set_sigmask(&host_set); 2046 2047 if (restore_sigcontext(env, &uc->tuc_mcontext)) 2048 return 1; 2049 2050 /* Restore coprocessor signal frame */ 2051 regspace = uc->tuc_regspace; 2052 if (arm_feature(env, ARM_FEATURE_VFP)) { 2053 regspace = restore_sigframe_v2_vfp(env, regspace); 2054 if (!regspace) { 2055 return 1; 2056 } 2057 } 2058 if (arm_feature(env, ARM_FEATURE_IWMMXT)) { 2059 regspace = restore_sigframe_v2_iwmmxt(env, regspace); 2060 if (!regspace) { 2061 return 1; 2062 } 2063 } 2064 2065 if (do_sigaltstack(frame_addr + offsetof(struct target_ucontext_v2, tuc_stack), 0, get_sp_from_cpustate(env)) == -EFAULT) 2066 return 1; 2067 2068 #if 0 2069 /* Send SIGTRAP if we're single-stepping */ 2070 if (ptrace_cancel_bpt(current)) 2071 send_sig(SIGTRAP, current, 1); 2072 #endif 2073 2074 return 0; 2075 } 2076 2077 static long do_sigreturn_v2(CPUARMState *env) 2078 { 2079 abi_ulong frame_addr; 2080 struct sigframe_v2 *frame = NULL; 2081 2082 /* 2083 * Since we stacked the signal on a 64-bit boundary, 2084 * then 'sp' should be word aligned here. If it's 2085 * not, then the user is trying to mess with us. 2086 */ 2087 frame_addr = env->regs[13]; 2088 trace_user_do_sigreturn(env, frame_addr); 2089 if (frame_addr & 7) { 2090 goto badframe; 2091 } 2092 2093 if (!lock_user_struct(VERIFY_READ, frame, frame_addr, 1)) { 2094 goto badframe; 2095 } 2096 2097 if (do_sigframe_return_v2(env, frame_addr, &frame->uc)) { 2098 goto badframe; 2099 } 2100 2101 unlock_user_struct(frame, frame_addr, 0); 2102 return -TARGET_QEMU_ESIGRETURN; 2103 2104 badframe: 2105 unlock_user_struct(frame, frame_addr, 0); 2106 force_sig(TARGET_SIGSEGV /* , current */); 2107 return 0; 2108 } 2109 2110 long do_sigreturn(CPUARMState *env) 2111 { 2112 if (get_osversion() >= 0x020612) { 2113 return do_sigreturn_v2(env); 2114 } else { 2115 return do_sigreturn_v1(env); 2116 } 2117 } 2118 2119 static long do_rt_sigreturn_v1(CPUARMState *env) 2120 { 2121 abi_ulong frame_addr; 2122 struct rt_sigframe_v1 *frame = NULL; 2123 sigset_t host_set; 2124 2125 /* 2126 * Since we stacked the signal on a 64-bit boundary, 2127 * then 'sp' should be word aligned here. If it's 2128 * not, then the user is trying to mess with us. 2129 */ 2130 frame_addr = env->regs[13]; 2131 trace_user_do_rt_sigreturn(env, frame_addr); 2132 if (frame_addr & 7) { 2133 goto badframe; 2134 } 2135 2136 if (!lock_user_struct(VERIFY_READ, frame, frame_addr, 1)) { 2137 goto badframe; 2138 } 2139 2140 target_to_host_sigset(&host_set, &frame->uc.tuc_sigmask); 2141 set_sigmask(&host_set); 2142 2143 if (restore_sigcontext(env, &frame->uc.tuc_mcontext)) { 2144 goto badframe; 2145 } 2146 2147 if (do_sigaltstack(frame_addr + offsetof(struct rt_sigframe_v1, uc.tuc_stack), 0, get_sp_from_cpustate(env)) == -EFAULT) 2148 goto badframe; 2149 2150 #if 0 2151 /* Send SIGTRAP if we're single-stepping */ 2152 if (ptrace_cancel_bpt(current)) 2153 send_sig(SIGTRAP, current, 1); 2154 #endif 2155 unlock_user_struct(frame, frame_addr, 0); 2156 return -TARGET_QEMU_ESIGRETURN; 2157 2158 badframe: 2159 unlock_user_struct(frame, frame_addr, 0); 2160 force_sig(TARGET_SIGSEGV /* , current */); 2161 return 0; 2162 } 2163 2164 static long do_rt_sigreturn_v2(CPUARMState *env) 2165 { 2166 abi_ulong frame_addr; 2167 struct rt_sigframe_v2 *frame = NULL; 2168 2169 /* 2170 * Since we stacked the signal on a 64-bit boundary, 2171 * then 'sp' should be word aligned here. If it's 2172 * not, then the user is trying to mess with us. 2173 */ 2174 frame_addr = env->regs[13]; 2175 trace_user_do_rt_sigreturn(env, frame_addr); 2176 if (frame_addr & 7) { 2177 goto badframe; 2178 } 2179 2180 if (!lock_user_struct(VERIFY_READ, frame, frame_addr, 1)) { 2181 goto badframe; 2182 } 2183 2184 if (do_sigframe_return_v2(env, frame_addr, &frame->uc)) { 2185 goto badframe; 2186 } 2187 2188 unlock_user_struct(frame, frame_addr, 0); 2189 return -TARGET_QEMU_ESIGRETURN; 2190 2191 badframe: 2192 unlock_user_struct(frame, frame_addr, 0); 2193 force_sig(TARGET_SIGSEGV /* , current */); 2194 return 0; 2195 } 2196 2197 long do_rt_sigreturn(CPUARMState *env) 2198 { 2199 if (get_osversion() >= 0x020612) { 2200 return do_rt_sigreturn_v2(env); 2201 } else { 2202 return do_rt_sigreturn_v1(env); 2203 } 2204 } 2205 2206 #elif defined(TARGET_SPARC) 2207 2208 #define __SUNOS_MAXWIN 31 2209 2210 /* This is what SunOS does, so shall I. */ 2211 struct target_sigcontext { 2212 abi_ulong sigc_onstack; /* state to restore */ 2213 2214 abi_ulong sigc_mask; /* sigmask to restore */ 2215 abi_ulong sigc_sp; /* stack pointer */ 2216 abi_ulong sigc_pc; /* program counter */ 2217 abi_ulong sigc_npc; /* next program counter */ 2218 abi_ulong sigc_psr; /* for condition codes etc */ 2219 abi_ulong sigc_g1; /* User uses these two registers */ 2220 abi_ulong sigc_o0; /* within the trampoline code. */ 2221 2222 /* Now comes information regarding the users window set 2223 * at the time of the signal. 2224 */ 2225 abi_ulong sigc_oswins; /* outstanding windows */ 2226 2227 /* stack ptrs for each regwin buf */ 2228 char *sigc_spbuf[__SUNOS_MAXWIN]; 2229 2230 /* Windows to restore after signal */ 2231 struct { 2232 abi_ulong locals[8]; 2233 abi_ulong ins[8]; 2234 } sigc_wbuf[__SUNOS_MAXWIN]; 2235 }; 2236 /* A Sparc stack frame */ 2237 struct sparc_stackf { 2238 abi_ulong locals[8]; 2239 abi_ulong ins[8]; 2240 /* It's simpler to treat fp and callers_pc as elements of ins[] 2241 * since we never need to access them ourselves. 2242 */ 2243 char *structptr; 2244 abi_ulong xargs[6]; 2245 abi_ulong xxargs[1]; 2246 }; 2247 2248 typedef struct { 2249 struct { 2250 abi_ulong psr; 2251 abi_ulong pc; 2252 abi_ulong npc; 2253 abi_ulong y; 2254 abi_ulong u_regs[16]; /* globals and ins */ 2255 } si_regs; 2256 int si_mask; 2257 } __siginfo_t; 2258 2259 typedef struct { 2260 abi_ulong si_float_regs[32]; 2261 unsigned long si_fsr; 2262 unsigned long si_fpqdepth; 2263 struct { 2264 unsigned long *insn_addr; 2265 unsigned long insn; 2266 } si_fpqueue [16]; 2267 } qemu_siginfo_fpu_t; 2268 2269 2270 struct target_signal_frame { 2271 struct sparc_stackf ss; 2272 __siginfo_t info; 2273 abi_ulong fpu_save; 2274 abi_ulong insns[2] __attribute__ ((aligned (8))); 2275 abi_ulong extramask[TARGET_NSIG_WORDS - 1]; 2276 abi_ulong extra_size; /* Should be 0 */ 2277 qemu_siginfo_fpu_t fpu_state; 2278 }; 2279 struct target_rt_signal_frame { 2280 struct sparc_stackf ss; 2281 siginfo_t info; 2282 abi_ulong regs[20]; 2283 sigset_t mask; 2284 abi_ulong fpu_save; 2285 unsigned int insns[2]; 2286 stack_t stack; 2287 unsigned int extra_size; /* Should be 0 */ 2288 qemu_siginfo_fpu_t fpu_state; 2289 }; 2290 2291 #define UREG_O0 16 2292 #define UREG_O6 22 2293 #define UREG_I0 0 2294 #define UREG_I1 1 2295 #define UREG_I2 2 2296 #define UREG_I3 3 2297 #define UREG_I4 4 2298 #define UREG_I5 5 2299 #define UREG_I6 6 2300 #define UREG_I7 7 2301 #define UREG_L0 8 2302 #define UREG_FP UREG_I6 2303 #define UREG_SP UREG_O6 2304 2305 static inline abi_ulong get_sigframe(struct target_sigaction *sa, 2306 CPUSPARCState *env, 2307 unsigned long framesize) 2308 { 2309 abi_ulong sp; 2310 2311 sp = env->regwptr[UREG_FP]; 2312 2313 /* This is the X/Open sanctioned signal stack switching. */ 2314 if (sa->sa_flags & TARGET_SA_ONSTACK) { 2315 if (!on_sig_stack(sp) 2316 && !((target_sigaltstack_used.ss_sp + target_sigaltstack_used.ss_size) & 7)) { 2317 sp = target_sigaltstack_used.ss_sp + target_sigaltstack_used.ss_size; 2318 } 2319 } 2320 return sp - framesize; 2321 } 2322 2323 static int 2324 setup___siginfo(__siginfo_t *si, CPUSPARCState *env, abi_ulong mask) 2325 { 2326 int err = 0, i; 2327 2328 __put_user(env->psr, &si->si_regs.psr); 2329 __put_user(env->pc, &si->si_regs.pc); 2330 __put_user(env->npc, &si->si_regs.npc); 2331 __put_user(env->y, &si->si_regs.y); 2332 for (i=0; i < 8; i++) { 2333 __put_user(env->gregs[i], &si->si_regs.u_regs[i]); 2334 } 2335 for (i=0; i < 8; i++) { 2336 __put_user(env->regwptr[UREG_I0 + i], &si->si_regs.u_regs[i+8]); 2337 } 2338 __put_user(mask, &si->si_mask); 2339 return err; 2340 } 2341 2342 #if 0 2343 static int 2344 setup_sigcontext(struct target_sigcontext *sc, /*struct _fpstate *fpstate,*/ 2345 CPUSPARCState *env, unsigned long mask) 2346 { 2347 int err = 0; 2348 2349 __put_user(mask, &sc->sigc_mask); 2350 __put_user(env->regwptr[UREG_SP], &sc->sigc_sp); 2351 __put_user(env->pc, &sc->sigc_pc); 2352 __put_user(env->npc, &sc->sigc_npc); 2353 __put_user(env->psr, &sc->sigc_psr); 2354 __put_user(env->gregs[1], &sc->sigc_g1); 2355 __put_user(env->regwptr[UREG_O0], &sc->sigc_o0); 2356 2357 return err; 2358 } 2359 #endif 2360 #define NF_ALIGNEDSZ (((sizeof(struct target_signal_frame) + 7) & (~7))) 2361 2362 static void setup_frame(int sig, struct target_sigaction *ka, 2363 target_sigset_t *set, CPUSPARCState *env) 2364 { 2365 abi_ulong sf_addr; 2366 struct target_signal_frame *sf; 2367 int sigframe_size, err, i; 2368 2369 /* 1. Make sure everything is clean */ 2370 //synchronize_user_stack(); 2371 2372 sigframe_size = NF_ALIGNEDSZ; 2373 sf_addr = get_sigframe(ka, env, sigframe_size); 2374 trace_user_setup_frame(env, sf_addr); 2375 2376 sf = lock_user(VERIFY_WRITE, sf_addr, 2377 sizeof(struct target_signal_frame), 0); 2378 if (!sf) { 2379 goto sigsegv; 2380 } 2381 #if 0 2382 if (invalid_frame_pointer(sf, sigframe_size)) 2383 goto sigill_and_return; 2384 #endif 2385 /* 2. Save the current process state */ 2386 err = setup___siginfo(&sf->info, env, set->sig[0]); 2387 __put_user(0, &sf->extra_size); 2388 2389 //save_fpu_state(regs, &sf->fpu_state); 2390 //__put_user(&sf->fpu_state, &sf->fpu_save); 2391 2392 __put_user(set->sig[0], &sf->info.si_mask); 2393 for (i = 0; i < TARGET_NSIG_WORDS - 1; i++) { 2394 __put_user(set->sig[i + 1], &sf->extramask[i]); 2395 } 2396 2397 for (i = 0; i < 8; i++) { 2398 __put_user(env->regwptr[i + UREG_L0], &sf->ss.locals[i]); 2399 } 2400 for (i = 0; i < 8; i++) { 2401 __put_user(env->regwptr[i + UREG_I0], &sf->ss.ins[i]); 2402 } 2403 if (err) 2404 goto sigsegv; 2405 2406 /* 3. signal handler back-trampoline and parameters */ 2407 env->regwptr[UREG_FP] = sf_addr; 2408 env->regwptr[UREG_I0] = sig; 2409 env->regwptr[UREG_I1] = sf_addr + 2410 offsetof(struct target_signal_frame, info); 2411 env->regwptr[UREG_I2] = sf_addr + 2412 offsetof(struct target_signal_frame, info); 2413 2414 /* 4. signal handler */ 2415 env->pc = ka->_sa_handler; 2416 env->npc = (env->pc + 4); 2417 /* 5. return to kernel instructions */ 2418 if (ka->sa_restorer) { 2419 env->regwptr[UREG_I7] = ka->sa_restorer; 2420 } else { 2421 uint32_t val32; 2422 2423 env->regwptr[UREG_I7] = sf_addr + 2424 offsetof(struct target_signal_frame, insns) - 2 * 4; 2425 2426 /* mov __NR_sigreturn, %g1 */ 2427 val32 = 0x821020d8; 2428 __put_user(val32, &sf->insns[0]); 2429 2430 /* t 0x10 */ 2431 val32 = 0x91d02010; 2432 __put_user(val32, &sf->insns[1]); 2433 if (err) 2434 goto sigsegv; 2435 2436 /* Flush instruction space. */ 2437 // flush_sig_insns(current->mm, (unsigned long) &(sf->insns[0])); 2438 // tb_flush(env); 2439 } 2440 unlock_user(sf, sf_addr, sizeof(struct target_signal_frame)); 2441 return; 2442 #if 0 2443 sigill_and_return: 2444 force_sig(TARGET_SIGILL); 2445 #endif 2446 sigsegv: 2447 unlock_user(sf, sf_addr, sizeof(struct target_signal_frame)); 2448 force_sig(TARGET_SIGSEGV); 2449 } 2450 2451 static void setup_rt_frame(int sig, struct target_sigaction *ka, 2452 target_siginfo_t *info, 2453 target_sigset_t *set, CPUSPARCState *env) 2454 { 2455 fprintf(stderr, "setup_rt_frame: not implemented\n"); 2456 } 2457 2458 long do_sigreturn(CPUSPARCState *env) 2459 { 2460 abi_ulong sf_addr; 2461 struct target_signal_frame *sf; 2462 uint32_t up_psr, pc, npc; 2463 target_sigset_t set; 2464 sigset_t host_set; 2465 int err=0, i; 2466 2467 sf_addr = env->regwptr[UREG_FP]; 2468 trace_user_do_sigreturn(env, sf_addr); 2469 if (!lock_user_struct(VERIFY_READ, sf, sf_addr, 1)) { 2470 goto segv_and_exit; 2471 } 2472 2473 /* 1. Make sure we are not getting garbage from the user */ 2474 2475 if (sf_addr & 3) 2476 goto segv_and_exit; 2477 2478 __get_user(pc, &sf->info.si_regs.pc); 2479 __get_user(npc, &sf->info.si_regs.npc); 2480 2481 if ((pc | npc) & 3) { 2482 goto segv_and_exit; 2483 } 2484 2485 /* 2. Restore the state */ 2486 __get_user(up_psr, &sf->info.si_regs.psr); 2487 2488 /* User can only change condition codes and FPU enabling in %psr. */ 2489 env->psr = (up_psr & (PSR_ICC /* | PSR_EF */)) 2490 | (env->psr & ~(PSR_ICC /* | PSR_EF */)); 2491 2492 env->pc = pc; 2493 env->npc = npc; 2494 __get_user(env->y, &sf->info.si_regs.y); 2495 for (i=0; i < 8; i++) { 2496 __get_user(env->gregs[i], &sf->info.si_regs.u_regs[i]); 2497 } 2498 for (i=0; i < 8; i++) { 2499 __get_user(env->regwptr[i + UREG_I0], &sf->info.si_regs.u_regs[i+8]); 2500 } 2501 2502 /* FIXME: implement FPU save/restore: 2503 * __get_user(fpu_save, &sf->fpu_save); 2504 * if (fpu_save) 2505 * err |= restore_fpu_state(env, fpu_save); 2506 */ 2507 2508 /* This is pretty much atomic, no amount locking would prevent 2509 * the races which exist anyways. 2510 */ 2511 __get_user(set.sig[0], &sf->info.si_mask); 2512 for(i = 1; i < TARGET_NSIG_WORDS; i++) { 2513 __get_user(set.sig[i], &sf->extramask[i - 1]); 2514 } 2515 2516 target_to_host_sigset_internal(&host_set, &set); 2517 set_sigmask(&host_set); 2518 2519 if (err) { 2520 goto segv_and_exit; 2521 } 2522 unlock_user_struct(sf, sf_addr, 0); 2523 return -TARGET_QEMU_ESIGRETURN; 2524 2525 segv_and_exit: 2526 unlock_user_struct(sf, sf_addr, 0); 2527 force_sig(TARGET_SIGSEGV); 2528 } 2529 2530 long do_rt_sigreturn(CPUSPARCState *env) 2531 { 2532 trace_user_do_rt_sigreturn(env, 0); 2533 fprintf(stderr, "do_rt_sigreturn: not implemented\n"); 2534 return -TARGET_ENOSYS; 2535 } 2536 2537 #if defined(TARGET_SPARC64) && !defined(TARGET_ABI32) 2538 #define MC_TSTATE 0 2539 #define MC_PC 1 2540 #define MC_NPC 2 2541 #define MC_Y 3 2542 #define MC_G1 4 2543 #define MC_G2 5 2544 #define MC_G3 6 2545 #define MC_G4 7 2546 #define MC_G5 8 2547 #define MC_G6 9 2548 #define MC_G7 10 2549 #define MC_O0 11 2550 #define MC_O1 12 2551 #define MC_O2 13 2552 #define MC_O3 14 2553 #define MC_O4 15 2554 #define MC_O5 16 2555 #define MC_O6 17 2556 #define MC_O7 18 2557 #define MC_NGREG 19 2558 2559 typedef abi_ulong target_mc_greg_t; 2560 typedef target_mc_greg_t target_mc_gregset_t[MC_NGREG]; 2561 2562 struct target_mc_fq { 2563 abi_ulong *mcfq_addr; 2564 uint32_t mcfq_insn; 2565 }; 2566 2567 struct target_mc_fpu { 2568 union { 2569 uint32_t sregs[32]; 2570 uint64_t dregs[32]; 2571 //uint128_t qregs[16]; 2572 } mcfpu_fregs; 2573 abi_ulong mcfpu_fsr; 2574 abi_ulong mcfpu_fprs; 2575 abi_ulong mcfpu_gsr; 2576 struct target_mc_fq *mcfpu_fq; 2577 unsigned char mcfpu_qcnt; 2578 unsigned char mcfpu_qentsz; 2579 unsigned char mcfpu_enab; 2580 }; 2581 typedef struct target_mc_fpu target_mc_fpu_t; 2582 2583 typedef struct { 2584 target_mc_gregset_t mc_gregs; 2585 target_mc_greg_t mc_fp; 2586 target_mc_greg_t mc_i7; 2587 target_mc_fpu_t mc_fpregs; 2588 } target_mcontext_t; 2589 2590 struct target_ucontext { 2591 struct target_ucontext *tuc_link; 2592 abi_ulong tuc_flags; 2593 target_sigset_t tuc_sigmask; 2594 target_mcontext_t tuc_mcontext; 2595 }; 2596 2597 /* A V9 register window */ 2598 struct target_reg_window { 2599 abi_ulong locals[8]; 2600 abi_ulong ins[8]; 2601 }; 2602 2603 #define TARGET_STACK_BIAS 2047 2604 2605 /* {set, get}context() needed for 64-bit SparcLinux userland. */ 2606 void sparc64_set_context(CPUSPARCState *env) 2607 { 2608 abi_ulong ucp_addr; 2609 struct target_ucontext *ucp; 2610 target_mc_gregset_t *grp; 2611 abi_ulong pc, npc, tstate; 2612 abi_ulong fp, i7, w_addr; 2613 unsigned int i; 2614 2615 ucp_addr = env->regwptr[UREG_I0]; 2616 if (!lock_user_struct(VERIFY_READ, ucp, ucp_addr, 1)) { 2617 goto do_sigsegv; 2618 } 2619 grp = &ucp->tuc_mcontext.mc_gregs; 2620 __get_user(pc, &((*grp)[MC_PC])); 2621 __get_user(npc, &((*grp)[MC_NPC])); 2622 if ((pc | npc) & 3) { 2623 goto do_sigsegv; 2624 } 2625 if (env->regwptr[UREG_I1]) { 2626 target_sigset_t target_set; 2627 sigset_t set; 2628 2629 if (TARGET_NSIG_WORDS == 1) { 2630 __get_user(target_set.sig[0], &ucp->tuc_sigmask.sig[0]); 2631 } else { 2632 abi_ulong *src, *dst; 2633 src = ucp->tuc_sigmask.sig; 2634 dst = target_set.sig; 2635 for (i = 0; i < TARGET_NSIG_WORDS; i++, dst++, src++) { 2636 __get_user(*dst, src); 2637 } 2638 } 2639 target_to_host_sigset_internal(&set, &target_set); 2640 set_sigmask(&set); 2641 } 2642 env->pc = pc; 2643 env->npc = npc; 2644 __get_user(env->y, &((*grp)[MC_Y])); 2645 __get_user(tstate, &((*grp)[MC_TSTATE])); 2646 env->asi = (tstate >> 24) & 0xff; 2647 cpu_put_ccr(env, tstate >> 32); 2648 cpu_put_cwp64(env, tstate & 0x1f); 2649 __get_user(env->gregs[1], (&(*grp)[MC_G1])); 2650 __get_user(env->gregs[2], (&(*grp)[MC_G2])); 2651 __get_user(env->gregs[3], (&(*grp)[MC_G3])); 2652 __get_user(env->gregs[4], (&(*grp)[MC_G4])); 2653 __get_user(env->gregs[5], (&(*grp)[MC_G5])); 2654 __get_user(env->gregs[6], (&(*grp)[MC_G6])); 2655 __get_user(env->gregs[7], (&(*grp)[MC_G7])); 2656 __get_user(env->regwptr[UREG_I0], (&(*grp)[MC_O0])); 2657 __get_user(env->regwptr[UREG_I1], (&(*grp)[MC_O1])); 2658 __get_user(env->regwptr[UREG_I2], (&(*grp)[MC_O2])); 2659 __get_user(env->regwptr[UREG_I3], (&(*grp)[MC_O3])); 2660 __get_user(env->regwptr[UREG_I4], (&(*grp)[MC_O4])); 2661 __get_user(env->regwptr[UREG_I5], (&(*grp)[MC_O5])); 2662 __get_user(env->regwptr[UREG_I6], (&(*grp)[MC_O6])); 2663 __get_user(env->regwptr[UREG_I7], (&(*grp)[MC_O7])); 2664 2665 __get_user(fp, &(ucp->tuc_mcontext.mc_fp)); 2666 __get_user(i7, &(ucp->tuc_mcontext.mc_i7)); 2667 2668 w_addr = TARGET_STACK_BIAS+env->regwptr[UREG_I6]; 2669 if (put_user(fp, w_addr + offsetof(struct target_reg_window, ins[6]), 2670 abi_ulong) != 0) { 2671 goto do_sigsegv; 2672 } 2673 if (put_user(i7, w_addr + offsetof(struct target_reg_window, ins[7]), 2674 abi_ulong) != 0) { 2675 goto do_sigsegv; 2676 } 2677 /* FIXME this does not match how the kernel handles the FPU in 2678 * its sparc64_set_context implementation. In particular the FPU 2679 * is only restored if fenab is non-zero in: 2680 * __get_user(fenab, &(ucp->tuc_mcontext.mc_fpregs.mcfpu_enab)); 2681 */ 2682 __get_user(env->fprs, &(ucp->tuc_mcontext.mc_fpregs.mcfpu_fprs)); 2683 { 2684 uint32_t *src = ucp->tuc_mcontext.mc_fpregs.mcfpu_fregs.sregs; 2685 for (i = 0; i < 64; i++, src++) { 2686 if (i & 1) { 2687 __get_user(env->fpr[i/2].l.lower, src); 2688 } else { 2689 __get_user(env->fpr[i/2].l.upper, src); 2690 } 2691 } 2692 } 2693 __get_user(env->fsr, 2694 &(ucp->tuc_mcontext.mc_fpregs.mcfpu_fsr)); 2695 __get_user(env->gsr, 2696 &(ucp->tuc_mcontext.mc_fpregs.mcfpu_gsr)); 2697 unlock_user_struct(ucp, ucp_addr, 0); 2698 return; 2699 do_sigsegv: 2700 unlock_user_struct(ucp, ucp_addr, 0); 2701 force_sig(TARGET_SIGSEGV); 2702 } 2703 2704 void sparc64_get_context(CPUSPARCState *env) 2705 { 2706 abi_ulong ucp_addr; 2707 struct target_ucontext *ucp; 2708 target_mc_gregset_t *grp; 2709 target_mcontext_t *mcp; 2710 abi_ulong fp, i7, w_addr; 2711 int err; 2712 unsigned int i; 2713 target_sigset_t target_set; 2714 sigset_t set; 2715 2716 ucp_addr = env->regwptr[UREG_I0]; 2717 if (!lock_user_struct(VERIFY_WRITE, ucp, ucp_addr, 0)) { 2718 goto do_sigsegv; 2719 } 2720 2721 mcp = &ucp->tuc_mcontext; 2722 grp = &mcp->mc_gregs; 2723 2724 /* Skip over the trap instruction, first. */ 2725 env->pc = env->npc; 2726 env->npc += 4; 2727 2728 /* If we're only reading the signal mask then do_sigprocmask() 2729 * is guaranteed not to fail, which is important because we don't 2730 * have any way to signal a failure or restart this operation since 2731 * this is not a normal syscall. 2732 */ 2733 err = do_sigprocmask(0, NULL, &set); 2734 assert(err == 0); 2735 host_to_target_sigset_internal(&target_set, &set); 2736 if (TARGET_NSIG_WORDS == 1) { 2737 __put_user(target_set.sig[0], 2738 (abi_ulong *)&ucp->tuc_sigmask); 2739 } else { 2740 abi_ulong *src, *dst; 2741 src = target_set.sig; 2742 dst = ucp->tuc_sigmask.sig; 2743 for (i = 0; i < TARGET_NSIG_WORDS; i++, dst++, src++) { 2744 __put_user(*src, dst); 2745 } 2746 if (err) 2747 goto do_sigsegv; 2748 } 2749 2750 /* XXX: tstate must be saved properly */ 2751 // __put_user(env->tstate, &((*grp)[MC_TSTATE])); 2752 __put_user(env->pc, &((*grp)[MC_PC])); 2753 __put_user(env->npc, &((*grp)[MC_NPC])); 2754 __put_user(env->y, &((*grp)[MC_Y])); 2755 __put_user(env->gregs[1], &((*grp)[MC_G1])); 2756 __put_user(env->gregs[2], &((*grp)[MC_G2])); 2757 __put_user(env->gregs[3], &((*grp)[MC_G3])); 2758 __put_user(env->gregs[4], &((*grp)[MC_G4])); 2759 __put_user(env->gregs[5], &((*grp)[MC_G5])); 2760 __put_user(env->gregs[6], &((*grp)[MC_G6])); 2761 __put_user(env->gregs[7], &((*grp)[MC_G7])); 2762 __put_user(env->regwptr[UREG_I0], &((*grp)[MC_O0])); 2763 __put_user(env->regwptr[UREG_I1], &((*grp)[MC_O1])); 2764 __put_user(env->regwptr[UREG_I2], &((*grp)[MC_O2])); 2765 __put_user(env->regwptr[UREG_I3], &((*grp)[MC_O3])); 2766 __put_user(env->regwptr[UREG_I4], &((*grp)[MC_O4])); 2767 __put_user(env->regwptr[UREG_I5], &((*grp)[MC_O5])); 2768 __put_user(env->regwptr[UREG_I6], &((*grp)[MC_O6])); 2769 __put_user(env->regwptr[UREG_I7], &((*grp)[MC_O7])); 2770 2771 w_addr = TARGET_STACK_BIAS+env->regwptr[UREG_I6]; 2772 fp = i7 = 0; 2773 if (get_user(fp, w_addr + offsetof(struct target_reg_window, ins[6]), 2774 abi_ulong) != 0) { 2775 goto do_sigsegv; 2776 } 2777 if (get_user(i7, w_addr + offsetof(struct target_reg_window, ins[7]), 2778 abi_ulong) != 0) { 2779 goto do_sigsegv; 2780 } 2781 __put_user(fp, &(mcp->mc_fp)); 2782 __put_user(i7, &(mcp->mc_i7)); 2783 2784 { 2785 uint32_t *dst = ucp->tuc_mcontext.mc_fpregs.mcfpu_fregs.sregs; 2786 for (i = 0; i < 64; i++, dst++) { 2787 if (i & 1) { 2788 __put_user(env->fpr[i/2].l.lower, dst); 2789 } else { 2790 __put_user(env->fpr[i/2].l.upper, dst); 2791 } 2792 } 2793 } 2794 __put_user(env->fsr, &(mcp->mc_fpregs.mcfpu_fsr)); 2795 __put_user(env->gsr, &(mcp->mc_fpregs.mcfpu_gsr)); 2796 __put_user(env->fprs, &(mcp->mc_fpregs.mcfpu_fprs)); 2797 2798 if (err) 2799 goto do_sigsegv; 2800 unlock_user_struct(ucp, ucp_addr, 1); 2801 return; 2802 do_sigsegv: 2803 unlock_user_struct(ucp, ucp_addr, 1); 2804 force_sig(TARGET_SIGSEGV); 2805 } 2806 #endif 2807 #elif defined(TARGET_MIPS) || defined(TARGET_MIPS64) 2808 2809 # if defined(TARGET_ABI_MIPSO32) 2810 struct target_sigcontext { 2811 uint32_t sc_regmask; /* Unused */ 2812 uint32_t sc_status; 2813 uint64_t sc_pc; 2814 uint64_t sc_regs[32]; 2815 uint64_t sc_fpregs[32]; 2816 uint32_t sc_ownedfp; /* Unused */ 2817 uint32_t sc_fpc_csr; 2818 uint32_t sc_fpc_eir; /* Unused */ 2819 uint32_t sc_used_math; 2820 uint32_t sc_dsp; /* dsp status, was sc_ssflags */ 2821 uint32_t pad0; 2822 uint64_t sc_mdhi; 2823 uint64_t sc_mdlo; 2824 target_ulong sc_hi1; /* Was sc_cause */ 2825 target_ulong sc_lo1; /* Was sc_badvaddr */ 2826 target_ulong sc_hi2; /* Was sc_sigset[4] */ 2827 target_ulong sc_lo2; 2828 target_ulong sc_hi3; 2829 target_ulong sc_lo3; 2830 }; 2831 # else /* N32 || N64 */ 2832 struct target_sigcontext { 2833 uint64_t sc_regs[32]; 2834 uint64_t sc_fpregs[32]; 2835 uint64_t sc_mdhi; 2836 uint64_t sc_hi1; 2837 uint64_t sc_hi2; 2838 uint64_t sc_hi3; 2839 uint64_t sc_mdlo; 2840 uint64_t sc_lo1; 2841 uint64_t sc_lo2; 2842 uint64_t sc_lo3; 2843 uint64_t sc_pc; 2844 uint32_t sc_fpc_csr; 2845 uint32_t sc_used_math; 2846 uint32_t sc_dsp; 2847 uint32_t sc_reserved; 2848 }; 2849 # endif /* O32 */ 2850 2851 struct sigframe { 2852 uint32_t sf_ass[4]; /* argument save space for o32 */ 2853 uint32_t sf_code[2]; /* signal trampoline */ 2854 struct target_sigcontext sf_sc; 2855 target_sigset_t sf_mask; 2856 }; 2857 2858 struct target_ucontext { 2859 target_ulong tuc_flags; 2860 target_ulong tuc_link; 2861 target_stack_t tuc_stack; 2862 target_ulong pad0; 2863 struct target_sigcontext tuc_mcontext; 2864 target_sigset_t tuc_sigmask; 2865 }; 2866 2867 struct target_rt_sigframe { 2868 uint32_t rs_ass[4]; /* argument save space for o32 */ 2869 uint32_t rs_code[2]; /* signal trampoline */ 2870 struct target_siginfo rs_info; 2871 struct target_ucontext rs_uc; 2872 }; 2873 2874 /* Install trampoline to jump back from signal handler */ 2875 static inline int install_sigtramp(unsigned int *tramp, unsigned int syscall) 2876 { 2877 int err = 0; 2878 2879 /* 2880 * Set up the return code ... 2881 * 2882 * li v0, __NR__foo_sigreturn 2883 * syscall 2884 */ 2885 2886 __put_user(0x24020000 + syscall, tramp + 0); 2887 __put_user(0x0000000c , tramp + 1); 2888 return err; 2889 } 2890 2891 static inline void setup_sigcontext(CPUMIPSState *regs, 2892 struct target_sigcontext *sc) 2893 { 2894 int i; 2895 2896 __put_user(exception_resume_pc(regs), &sc->sc_pc); 2897 regs->hflags &= ~MIPS_HFLAG_BMASK; 2898 2899 __put_user(0, &sc->sc_regs[0]); 2900 for (i = 1; i < 32; ++i) { 2901 __put_user(regs->active_tc.gpr[i], &sc->sc_regs[i]); 2902 } 2903 2904 __put_user(regs->active_tc.HI[0], &sc->sc_mdhi); 2905 __put_user(regs->active_tc.LO[0], &sc->sc_mdlo); 2906 2907 /* Rather than checking for dsp existence, always copy. The storage 2908 would just be garbage otherwise. */ 2909 __put_user(regs->active_tc.HI[1], &sc->sc_hi1); 2910 __put_user(regs->active_tc.HI[2], &sc->sc_hi2); 2911 __put_user(regs->active_tc.HI[3], &sc->sc_hi3); 2912 __put_user(regs->active_tc.LO[1], &sc->sc_lo1); 2913 __put_user(regs->active_tc.LO[2], &sc->sc_lo2); 2914 __put_user(regs->active_tc.LO[3], &sc->sc_lo3); 2915 { 2916 uint32_t dsp = cpu_rddsp(0x3ff, regs); 2917 __put_user(dsp, &sc->sc_dsp); 2918 } 2919 2920 __put_user(1, &sc->sc_used_math); 2921 2922 for (i = 0; i < 32; ++i) { 2923 __put_user(regs->active_fpu.fpr[i].d, &sc->sc_fpregs[i]); 2924 } 2925 } 2926 2927 static inline void 2928 restore_sigcontext(CPUMIPSState *regs, struct target_sigcontext *sc) 2929 { 2930 int i; 2931 2932 __get_user(regs->CP0_EPC, &sc->sc_pc); 2933 2934 __get_user(regs->active_tc.HI[0], &sc->sc_mdhi); 2935 __get_user(regs->active_tc.LO[0], &sc->sc_mdlo); 2936 2937 for (i = 1; i < 32; ++i) { 2938 __get_user(regs->active_tc.gpr[i], &sc->sc_regs[i]); 2939 } 2940 2941 __get_user(regs->active_tc.HI[1], &sc->sc_hi1); 2942 __get_user(regs->active_tc.HI[2], &sc->sc_hi2); 2943 __get_user(regs->active_tc.HI[3], &sc->sc_hi3); 2944 __get_user(regs->active_tc.LO[1], &sc->sc_lo1); 2945 __get_user(regs->active_tc.LO[2], &sc->sc_lo2); 2946 __get_user(regs->active_tc.LO[3], &sc->sc_lo3); 2947 { 2948 uint32_t dsp; 2949 __get_user(dsp, &sc->sc_dsp); 2950 cpu_wrdsp(dsp, 0x3ff, regs); 2951 } 2952 2953 for (i = 0; i < 32; ++i) { 2954 __get_user(regs->active_fpu.fpr[i].d, &sc->sc_fpregs[i]); 2955 } 2956 } 2957 2958 /* 2959 * Determine which stack to use.. 2960 */ 2961 static inline abi_ulong 2962 get_sigframe(struct target_sigaction *ka, CPUMIPSState *regs, size_t frame_size) 2963 { 2964 unsigned long sp; 2965 2966 /* Default to using normal stack */ 2967 sp = regs->active_tc.gpr[29]; 2968 2969 /* 2970 * FPU emulator may have its own trampoline active just 2971 * above the user stack, 16-bytes before the next lowest 2972 * 16 byte boundary. Try to avoid trashing it. 2973 */ 2974 sp -= 32; 2975 2976 /* This is the X/Open sanctioned signal stack switching. */ 2977 if ((ka->sa_flags & TARGET_SA_ONSTACK) && (sas_ss_flags (sp) == 0)) { 2978 sp = target_sigaltstack_used.ss_sp + target_sigaltstack_used.ss_size; 2979 } 2980 2981 return (sp - frame_size) & ~7; 2982 } 2983 2984 static void mips_set_hflags_isa_mode_from_pc(CPUMIPSState *env) 2985 { 2986 if (env->insn_flags & (ASE_MIPS16 | ASE_MICROMIPS)) { 2987 env->hflags &= ~MIPS_HFLAG_M16; 2988 env->hflags |= (env->active_tc.PC & 1) << MIPS_HFLAG_M16_SHIFT; 2989 env->active_tc.PC &= ~(target_ulong) 1; 2990 } 2991 } 2992 2993 # if defined(TARGET_ABI_MIPSO32) 2994 /* compare linux/arch/mips/kernel/signal.c:setup_frame() */ 2995 static void setup_frame(int sig, struct target_sigaction * ka, 2996 target_sigset_t *set, CPUMIPSState *regs) 2997 { 2998 struct sigframe *frame; 2999 abi_ulong frame_addr; 3000 int i; 3001 3002 frame_addr = get_sigframe(ka, regs, sizeof(*frame)); 3003 trace_user_setup_frame(regs, frame_addr); 3004 if (!lock_user_struct(VERIFY_WRITE, frame, frame_addr, 0)) { 3005 goto give_sigsegv; 3006 } 3007 3008 install_sigtramp(frame->sf_code, TARGET_NR_sigreturn); 3009 3010 setup_sigcontext(regs, &frame->sf_sc); 3011 3012 for(i = 0; i < TARGET_NSIG_WORDS; i++) { 3013 __put_user(set->sig[i], &frame->sf_mask.sig[i]); 3014 } 3015 3016 /* 3017 * Arguments to signal handler: 3018 * 3019 * a0 = signal number 3020 * a1 = 0 (should be cause) 3021 * a2 = pointer to struct sigcontext 3022 * 3023 * $25 and PC point to the signal handler, $29 points to the 3024 * struct sigframe. 3025 */ 3026 regs->active_tc.gpr[ 4] = sig; 3027 regs->active_tc.gpr[ 5] = 0; 3028 regs->active_tc.gpr[ 6] = frame_addr + offsetof(struct sigframe, sf_sc); 3029 regs->active_tc.gpr[29] = frame_addr; 3030 regs->active_tc.gpr[31] = frame_addr + offsetof(struct sigframe, sf_code); 3031 /* The original kernel code sets CP0_EPC to the handler 3032 * since it returns to userland using eret 3033 * we cannot do this here, and we must set PC directly */ 3034 regs->active_tc.PC = regs->active_tc.gpr[25] = ka->_sa_handler; 3035 mips_set_hflags_isa_mode_from_pc(regs); 3036 unlock_user_struct(frame, frame_addr, 1); 3037 return; 3038 3039 give_sigsegv: 3040 force_sig(TARGET_SIGSEGV/*, current*/); 3041 } 3042 3043 long do_sigreturn(CPUMIPSState *regs) 3044 { 3045 struct sigframe *frame; 3046 abi_ulong frame_addr; 3047 sigset_t blocked; 3048 target_sigset_t target_set; 3049 int i; 3050 3051 frame_addr = regs->active_tc.gpr[29]; 3052 trace_user_do_sigreturn(regs, frame_addr); 3053 if (!lock_user_struct(VERIFY_READ, frame, frame_addr, 1)) 3054 goto badframe; 3055 3056 for(i = 0; i < TARGET_NSIG_WORDS; i++) { 3057 __get_user(target_set.sig[i], &frame->sf_mask.sig[i]); 3058 } 3059 3060 target_to_host_sigset_internal(&blocked, &target_set); 3061 set_sigmask(&blocked); 3062 3063 restore_sigcontext(regs, &frame->sf_sc); 3064 3065 #if 0 3066 /* 3067 * Don't let your children do this ... 3068 */ 3069 __asm__ __volatile__( 3070 "move\t$29, %0\n\t" 3071 "j\tsyscall_exit" 3072 :/* no outputs */ 3073 :"r" (®s)); 3074 /* Unreached */ 3075 #endif 3076 3077 regs->active_tc.PC = regs->CP0_EPC; 3078 mips_set_hflags_isa_mode_from_pc(regs); 3079 /* I am not sure this is right, but it seems to work 3080 * maybe a problem with nested signals ? */ 3081 regs->CP0_EPC = 0; 3082 return -TARGET_QEMU_ESIGRETURN; 3083 3084 badframe: 3085 force_sig(TARGET_SIGSEGV/*, current*/); 3086 return 0; 3087 } 3088 # endif /* O32 */ 3089 3090 static void setup_rt_frame(int sig, struct target_sigaction *ka, 3091 target_siginfo_t *info, 3092 target_sigset_t *set, CPUMIPSState *env) 3093 { 3094 struct target_rt_sigframe *frame; 3095 abi_ulong frame_addr; 3096 int i; 3097 3098 frame_addr = get_sigframe(ka, env, sizeof(*frame)); 3099 trace_user_setup_rt_frame(env, frame_addr); 3100 if (!lock_user_struct(VERIFY_WRITE, frame, frame_addr, 0)) { 3101 goto give_sigsegv; 3102 } 3103 3104 install_sigtramp(frame->rs_code, TARGET_NR_rt_sigreturn); 3105 3106 tswap_siginfo(&frame->rs_info, info); 3107 3108 __put_user(0, &frame->rs_uc.tuc_flags); 3109 __put_user(0, &frame->rs_uc.tuc_link); 3110 __put_user(target_sigaltstack_used.ss_sp, &frame->rs_uc.tuc_stack.ss_sp); 3111 __put_user(target_sigaltstack_used.ss_size, &frame->rs_uc.tuc_stack.ss_size); 3112 __put_user(sas_ss_flags(get_sp_from_cpustate(env)), 3113 &frame->rs_uc.tuc_stack.ss_flags); 3114 3115 setup_sigcontext(env, &frame->rs_uc.tuc_mcontext); 3116 3117 for(i = 0; i < TARGET_NSIG_WORDS; i++) { 3118 __put_user(set->sig[i], &frame->rs_uc.tuc_sigmask.sig[i]); 3119 } 3120 3121 /* 3122 * Arguments to signal handler: 3123 * 3124 * a0 = signal number 3125 * a1 = pointer to siginfo_t 3126 * a2 = pointer to struct ucontext 3127 * 3128 * $25 and PC point to the signal handler, $29 points to the 3129 * struct sigframe. 3130 */ 3131 env->active_tc.gpr[ 4] = sig; 3132 env->active_tc.gpr[ 5] = frame_addr 3133 + offsetof(struct target_rt_sigframe, rs_info); 3134 env->active_tc.gpr[ 6] = frame_addr 3135 + offsetof(struct target_rt_sigframe, rs_uc); 3136 env->active_tc.gpr[29] = frame_addr; 3137 env->active_tc.gpr[31] = frame_addr 3138 + offsetof(struct target_rt_sigframe, rs_code); 3139 /* The original kernel code sets CP0_EPC to the handler 3140 * since it returns to userland using eret 3141 * we cannot do this here, and we must set PC directly */ 3142 env->active_tc.PC = env->active_tc.gpr[25] = ka->_sa_handler; 3143 mips_set_hflags_isa_mode_from_pc(env); 3144 unlock_user_struct(frame, frame_addr, 1); 3145 return; 3146 3147 give_sigsegv: 3148 unlock_user_struct(frame, frame_addr, 1); 3149 force_sig(TARGET_SIGSEGV/*, current*/); 3150 } 3151 3152 long do_rt_sigreturn(CPUMIPSState *env) 3153 { 3154 struct target_rt_sigframe *frame; 3155 abi_ulong frame_addr; 3156 sigset_t blocked; 3157 3158 frame_addr = env->active_tc.gpr[29]; 3159 trace_user_do_rt_sigreturn(env, frame_addr); 3160 if (!lock_user_struct(VERIFY_READ, frame, frame_addr, 1)) { 3161 goto badframe; 3162 } 3163 3164 target_to_host_sigset(&blocked, &frame->rs_uc.tuc_sigmask); 3165 set_sigmask(&blocked); 3166 3167 restore_sigcontext(env, &frame->rs_uc.tuc_mcontext); 3168 3169 if (do_sigaltstack(frame_addr + 3170 offsetof(struct target_rt_sigframe, rs_uc.tuc_stack), 3171 0, get_sp_from_cpustate(env)) == -EFAULT) 3172 goto badframe; 3173 3174 env->active_tc.PC = env->CP0_EPC; 3175 mips_set_hflags_isa_mode_from_pc(env); 3176 /* I am not sure this is right, but it seems to work 3177 * maybe a problem with nested signals ? */ 3178 env->CP0_EPC = 0; 3179 return -TARGET_QEMU_ESIGRETURN; 3180 3181 badframe: 3182 force_sig(TARGET_SIGSEGV/*, current*/); 3183 return 0; 3184 } 3185 3186 #elif defined(TARGET_SH4) 3187 3188 /* 3189 * code and data structures from linux kernel: 3190 * include/asm-sh/sigcontext.h 3191 * arch/sh/kernel/signal.c 3192 */ 3193 3194 struct target_sigcontext { 3195 target_ulong oldmask; 3196 3197 /* CPU registers */ 3198 target_ulong sc_gregs[16]; 3199 target_ulong sc_pc; 3200 target_ulong sc_pr; 3201 target_ulong sc_sr; 3202 target_ulong sc_gbr; 3203 target_ulong sc_mach; 3204 target_ulong sc_macl; 3205 3206 /* FPU registers */ 3207 target_ulong sc_fpregs[16]; 3208 target_ulong sc_xfpregs[16]; 3209 unsigned int sc_fpscr; 3210 unsigned int sc_fpul; 3211 unsigned int sc_ownedfp; 3212 }; 3213 3214 struct target_sigframe 3215 { 3216 struct target_sigcontext sc; 3217 target_ulong extramask[TARGET_NSIG_WORDS-1]; 3218 uint16_t retcode[3]; 3219 }; 3220 3221 3222 struct target_ucontext { 3223 target_ulong tuc_flags; 3224 struct target_ucontext *tuc_link; 3225 target_stack_t tuc_stack; 3226 struct target_sigcontext tuc_mcontext; 3227 target_sigset_t tuc_sigmask; /* mask last for extensibility */ 3228 }; 3229 3230 struct target_rt_sigframe 3231 { 3232 struct target_siginfo info; 3233 struct target_ucontext uc; 3234 uint16_t retcode[3]; 3235 }; 3236 3237 3238 #define MOVW(n) (0x9300|((n)-2)) /* Move mem word at PC+n to R3 */ 3239 #define TRAP_NOARG 0xc310 /* Syscall w/no args (NR in R3) SH3/4 */ 3240 3241 static abi_ulong get_sigframe(struct target_sigaction *ka, 3242 unsigned long sp, size_t frame_size) 3243 { 3244 if ((ka->sa_flags & TARGET_SA_ONSTACK) && (sas_ss_flags(sp) == 0)) { 3245 sp = target_sigaltstack_used.ss_sp + target_sigaltstack_used.ss_size; 3246 } 3247 3248 return (sp - frame_size) & -8ul; 3249 } 3250 3251 static void setup_sigcontext(struct target_sigcontext *sc, 3252 CPUSH4State *regs, unsigned long mask) 3253 { 3254 int i; 3255 3256 #define COPY(x) __put_user(regs->x, &sc->sc_##x) 3257 COPY(gregs[0]); COPY(gregs[1]); 3258 COPY(gregs[2]); COPY(gregs[3]); 3259 COPY(gregs[4]); COPY(gregs[5]); 3260 COPY(gregs[6]); COPY(gregs[7]); 3261 COPY(gregs[8]); COPY(gregs[9]); 3262 COPY(gregs[10]); COPY(gregs[11]); 3263 COPY(gregs[12]); COPY(gregs[13]); 3264 COPY(gregs[14]); COPY(gregs[15]); 3265 COPY(gbr); COPY(mach); 3266 COPY(macl); COPY(pr); 3267 COPY(sr); COPY(pc); 3268 #undef COPY 3269 3270 for (i=0; i<16; i++) { 3271 __put_user(regs->fregs[i], &sc->sc_fpregs[i]); 3272 } 3273 __put_user(regs->fpscr, &sc->sc_fpscr); 3274 __put_user(regs->fpul, &sc->sc_fpul); 3275 3276 /* non-iBCS2 extensions.. */ 3277 __put_user(mask, &sc->oldmask); 3278 } 3279 3280 static void restore_sigcontext(CPUSH4State *regs, struct target_sigcontext *sc) 3281 { 3282 int i; 3283 3284 #define COPY(x) __get_user(regs->x, &sc->sc_##x) 3285 COPY(gregs[0]); COPY(gregs[1]); 3286 COPY(gregs[2]); COPY(gregs[3]); 3287 COPY(gregs[4]); COPY(gregs[5]); 3288 COPY(gregs[6]); COPY(gregs[7]); 3289 COPY(gregs[8]); COPY(gregs[9]); 3290 COPY(gregs[10]); COPY(gregs[11]); 3291 COPY(gregs[12]); COPY(gregs[13]); 3292 COPY(gregs[14]); COPY(gregs[15]); 3293 COPY(gbr); COPY(mach); 3294 COPY(macl); COPY(pr); 3295 COPY(sr); COPY(pc); 3296 #undef COPY 3297 3298 for (i=0; i<16; i++) { 3299 __get_user(regs->fregs[i], &sc->sc_fpregs[i]); 3300 } 3301 __get_user(regs->fpscr, &sc->sc_fpscr); 3302 __get_user(regs->fpul, &sc->sc_fpul); 3303 3304 regs->tra = -1; /* disable syscall checks */ 3305 } 3306 3307 static void setup_frame(int sig, struct target_sigaction *ka, 3308 target_sigset_t *set, CPUSH4State *regs) 3309 { 3310 struct target_sigframe *frame; 3311 abi_ulong frame_addr; 3312 int i; 3313 3314 frame_addr = get_sigframe(ka, regs->gregs[15], sizeof(*frame)); 3315 trace_user_setup_frame(regs, frame_addr); 3316 if (!lock_user_struct(VERIFY_WRITE, frame, frame_addr, 0)) { 3317 goto give_sigsegv; 3318 } 3319 3320 setup_sigcontext(&frame->sc, regs, set->sig[0]); 3321 3322 for (i = 0; i < TARGET_NSIG_WORDS - 1; i++) { 3323 __put_user(set->sig[i + 1], &frame->extramask[i]); 3324 } 3325 3326 /* Set up to return from userspace. If provided, use a stub 3327 already in userspace. */ 3328 if (ka->sa_flags & TARGET_SA_RESTORER) { 3329 regs->pr = (unsigned long) ka->sa_restorer; 3330 } else { 3331 /* Generate return code (system call to sigreturn) */ 3332 abi_ulong retcode_addr = frame_addr + 3333 offsetof(struct target_sigframe, retcode); 3334 __put_user(MOVW(2), &frame->retcode[0]); 3335 __put_user(TRAP_NOARG, &frame->retcode[1]); 3336 __put_user((TARGET_NR_sigreturn), &frame->retcode[2]); 3337 regs->pr = (unsigned long) retcode_addr; 3338 } 3339 3340 /* Set up registers for signal handler */ 3341 regs->gregs[15] = frame_addr; 3342 regs->gregs[4] = sig; /* Arg for signal handler */ 3343 regs->gregs[5] = 0; 3344 regs->gregs[6] = frame_addr += offsetof(typeof(*frame), sc); 3345 regs->pc = (unsigned long) ka->_sa_handler; 3346 3347 unlock_user_struct(frame, frame_addr, 1); 3348 return; 3349 3350 give_sigsegv: 3351 unlock_user_struct(frame, frame_addr, 1); 3352 force_sig(TARGET_SIGSEGV); 3353 } 3354 3355 static void setup_rt_frame(int sig, struct target_sigaction *ka, 3356 target_siginfo_t *info, 3357 target_sigset_t *set, CPUSH4State *regs) 3358 { 3359 struct target_rt_sigframe *frame; 3360 abi_ulong frame_addr; 3361 int i; 3362 3363 frame_addr = get_sigframe(ka, regs->gregs[15], sizeof(*frame)); 3364 trace_user_setup_rt_frame(regs, frame_addr); 3365 if (!lock_user_struct(VERIFY_WRITE, frame, frame_addr, 0)) { 3366 goto give_sigsegv; 3367 } 3368 3369 tswap_siginfo(&frame->info, info); 3370 3371 /* Create the ucontext. */ 3372 __put_user(0, &frame->uc.tuc_flags); 3373 __put_user(0, (unsigned long *)&frame->uc.tuc_link); 3374 __put_user((unsigned long)target_sigaltstack_used.ss_sp, 3375 &frame->uc.tuc_stack.ss_sp); 3376 __put_user(sas_ss_flags(regs->gregs[15]), 3377 &frame->uc.tuc_stack.ss_flags); 3378 __put_user(target_sigaltstack_used.ss_size, 3379 &frame->uc.tuc_stack.ss_size); 3380 setup_sigcontext(&frame->uc.tuc_mcontext, 3381 regs, set->sig[0]); 3382 for(i = 0; i < TARGET_NSIG_WORDS; i++) { 3383 __put_user(set->sig[i], &frame->uc.tuc_sigmask.sig[i]); 3384 } 3385 3386 /* Set up to return from userspace. If provided, use a stub 3387 already in userspace. */ 3388 if (ka->sa_flags & TARGET_SA_RESTORER) { 3389 regs->pr = (unsigned long) ka->sa_restorer; 3390 } else { 3391 /* Generate return code (system call to sigreturn) */ 3392 abi_ulong retcode_addr = frame_addr + 3393 offsetof(struct target_rt_sigframe, retcode); 3394 __put_user(MOVW(2), &frame->retcode[0]); 3395 __put_user(TRAP_NOARG, &frame->retcode[1]); 3396 __put_user((TARGET_NR_rt_sigreturn), &frame->retcode[2]); 3397 regs->pr = (unsigned long) retcode_addr; 3398 } 3399 3400 /* Set up registers for signal handler */ 3401 regs->gregs[15] = frame_addr; 3402 regs->gregs[4] = sig; /* Arg for signal handler */ 3403 regs->gregs[5] = frame_addr + offsetof(typeof(*frame), info); 3404 regs->gregs[6] = frame_addr + offsetof(typeof(*frame), uc); 3405 regs->pc = (unsigned long) ka->_sa_handler; 3406 3407 unlock_user_struct(frame, frame_addr, 1); 3408 return; 3409 3410 give_sigsegv: 3411 unlock_user_struct(frame, frame_addr, 1); 3412 force_sig(TARGET_SIGSEGV); 3413 } 3414 3415 long do_sigreturn(CPUSH4State *regs) 3416 { 3417 struct target_sigframe *frame; 3418 abi_ulong frame_addr; 3419 sigset_t blocked; 3420 target_sigset_t target_set; 3421 int i; 3422 int err = 0; 3423 3424 frame_addr = regs->gregs[15]; 3425 trace_user_do_sigreturn(regs, frame_addr); 3426 if (!lock_user_struct(VERIFY_READ, frame, frame_addr, 1)) { 3427 goto badframe; 3428 } 3429 3430 __get_user(target_set.sig[0], &frame->sc.oldmask); 3431 for(i = 1; i < TARGET_NSIG_WORDS; i++) { 3432 __get_user(target_set.sig[i], &frame->extramask[i - 1]); 3433 } 3434 3435 if (err) 3436 goto badframe; 3437 3438 target_to_host_sigset_internal(&blocked, &target_set); 3439 set_sigmask(&blocked); 3440 3441 restore_sigcontext(regs, &frame->sc); 3442 3443 unlock_user_struct(frame, frame_addr, 0); 3444 return -TARGET_QEMU_ESIGRETURN; 3445 3446 badframe: 3447 unlock_user_struct(frame, frame_addr, 0); 3448 force_sig(TARGET_SIGSEGV); 3449 return 0; 3450 } 3451 3452 long do_rt_sigreturn(CPUSH4State *regs) 3453 { 3454 struct target_rt_sigframe *frame; 3455 abi_ulong frame_addr; 3456 sigset_t blocked; 3457 3458 frame_addr = regs->gregs[15]; 3459 trace_user_do_rt_sigreturn(regs, frame_addr); 3460 if (!lock_user_struct(VERIFY_READ, frame, frame_addr, 1)) { 3461 goto badframe; 3462 } 3463 3464 target_to_host_sigset(&blocked, &frame->uc.tuc_sigmask); 3465 set_sigmask(&blocked); 3466 3467 restore_sigcontext(regs, &frame->uc.tuc_mcontext); 3468 3469 if (do_sigaltstack(frame_addr + 3470 offsetof(struct target_rt_sigframe, uc.tuc_stack), 3471 0, get_sp_from_cpustate(regs)) == -EFAULT) { 3472 goto badframe; 3473 } 3474 3475 unlock_user_struct(frame, frame_addr, 0); 3476 return -TARGET_QEMU_ESIGRETURN; 3477 3478 badframe: 3479 unlock_user_struct(frame, frame_addr, 0); 3480 force_sig(TARGET_SIGSEGV); 3481 return 0; 3482 } 3483 #elif defined(TARGET_MICROBLAZE) 3484 3485 struct target_sigcontext { 3486 struct target_pt_regs regs; /* needs to be first */ 3487 uint32_t oldmask; 3488 }; 3489 3490 struct target_stack_t { 3491 abi_ulong ss_sp; 3492 int ss_flags; 3493 unsigned int ss_size; 3494 }; 3495 3496 struct target_ucontext { 3497 abi_ulong tuc_flags; 3498 abi_ulong tuc_link; 3499 struct target_stack_t tuc_stack; 3500 struct target_sigcontext tuc_mcontext; 3501 uint32_t tuc_extramask[TARGET_NSIG_WORDS - 1]; 3502 }; 3503 3504 /* Signal frames. */ 3505 struct target_signal_frame { 3506 struct target_ucontext uc; 3507 uint32_t extramask[TARGET_NSIG_WORDS - 1]; 3508 uint32_t tramp[2]; 3509 }; 3510 3511 struct rt_signal_frame { 3512 siginfo_t info; 3513 struct ucontext uc; 3514 uint32_t tramp[2]; 3515 }; 3516 3517 static void setup_sigcontext(struct target_sigcontext *sc, CPUMBState *env) 3518 { 3519 __put_user(env->regs[0], &sc->regs.r0); 3520 __put_user(env->regs[1], &sc->regs.r1); 3521 __put_user(env->regs[2], &sc->regs.r2); 3522 __put_user(env->regs[3], &sc->regs.r3); 3523 __put_user(env->regs[4], &sc->regs.r4); 3524 __put_user(env->regs[5], &sc->regs.r5); 3525 __put_user(env->regs[6], &sc->regs.r6); 3526 __put_user(env->regs[7], &sc->regs.r7); 3527 __put_user(env->regs[8], &sc->regs.r8); 3528 __put_user(env->regs[9], &sc->regs.r9); 3529 __put_user(env->regs[10], &sc->regs.r10); 3530 __put_user(env->regs[11], &sc->regs.r11); 3531 __put_user(env->regs[12], &sc->regs.r12); 3532 __put_user(env->regs[13], &sc->regs.r13); 3533 __put_user(env->regs[14], &sc->regs.r14); 3534 __put_user(env->regs[15], &sc->regs.r15); 3535 __put_user(env->regs[16], &sc->regs.r16); 3536 __put_user(env->regs[17], &sc->regs.r17); 3537 __put_user(env->regs[18], &sc->regs.r18); 3538 __put_user(env->regs[19], &sc->regs.r19); 3539 __put_user(env->regs[20], &sc->regs.r20); 3540 __put_user(env->regs[21], &sc->regs.r21); 3541 __put_user(env->regs[22], &sc->regs.r22); 3542 __put_user(env->regs[23], &sc->regs.r23); 3543 __put_user(env->regs[24], &sc->regs.r24); 3544 __put_user(env->regs[25], &sc->regs.r25); 3545 __put_user(env->regs[26], &sc->regs.r26); 3546 __put_user(env->regs[27], &sc->regs.r27); 3547 __put_user(env->regs[28], &sc->regs.r28); 3548 __put_user(env->regs[29], &sc->regs.r29); 3549 __put_user(env->regs[30], &sc->regs.r30); 3550 __put_user(env->regs[31], &sc->regs.r31); 3551 __put_user(env->sregs[SR_PC], &sc->regs.pc); 3552 } 3553 3554 static void restore_sigcontext(struct target_sigcontext *sc, CPUMBState *env) 3555 { 3556 __get_user(env->regs[0], &sc->regs.r0); 3557 __get_user(env->regs[1], &sc->regs.r1); 3558 __get_user(env->regs[2], &sc->regs.r2); 3559 __get_user(env->regs[3], &sc->regs.r3); 3560 __get_user(env->regs[4], &sc->regs.r4); 3561 __get_user(env->regs[5], &sc->regs.r5); 3562 __get_user(env->regs[6], &sc->regs.r6); 3563 __get_user(env->regs[7], &sc->regs.r7); 3564 __get_user(env->regs[8], &sc->regs.r8); 3565 __get_user(env->regs[9], &sc->regs.r9); 3566 __get_user(env->regs[10], &sc->regs.r10); 3567 __get_user(env->regs[11], &sc->regs.r11); 3568 __get_user(env->regs[12], &sc->regs.r12); 3569 __get_user(env->regs[13], &sc->regs.r13); 3570 __get_user(env->regs[14], &sc->regs.r14); 3571 __get_user(env->regs[15], &sc->regs.r15); 3572 __get_user(env->regs[16], &sc->regs.r16); 3573 __get_user(env->regs[17], &sc->regs.r17); 3574 __get_user(env->regs[18], &sc->regs.r18); 3575 __get_user(env->regs[19], &sc->regs.r19); 3576 __get_user(env->regs[20], &sc->regs.r20); 3577 __get_user(env->regs[21], &sc->regs.r21); 3578 __get_user(env->regs[22], &sc->regs.r22); 3579 __get_user(env->regs[23], &sc->regs.r23); 3580 __get_user(env->regs[24], &sc->regs.r24); 3581 __get_user(env->regs[25], &sc->regs.r25); 3582 __get_user(env->regs[26], &sc->regs.r26); 3583 __get_user(env->regs[27], &sc->regs.r27); 3584 __get_user(env->regs[28], &sc->regs.r28); 3585 __get_user(env->regs[29], &sc->regs.r29); 3586 __get_user(env->regs[30], &sc->regs.r30); 3587 __get_user(env->regs[31], &sc->regs.r31); 3588 __get_user(env->sregs[SR_PC], &sc->regs.pc); 3589 } 3590 3591 static abi_ulong get_sigframe(struct target_sigaction *ka, 3592 CPUMBState *env, int frame_size) 3593 { 3594 abi_ulong sp = env->regs[1]; 3595 3596 if ((ka->sa_flags & TARGET_SA_ONSTACK) != 0 && !on_sig_stack(sp)) { 3597 sp = target_sigaltstack_used.ss_sp + target_sigaltstack_used.ss_size; 3598 } 3599 3600 return ((sp - frame_size) & -8UL); 3601 } 3602 3603 static void setup_frame(int sig, struct target_sigaction *ka, 3604 target_sigset_t *set, CPUMBState *env) 3605 { 3606 struct target_signal_frame *frame; 3607 abi_ulong frame_addr; 3608 int i; 3609 3610 frame_addr = get_sigframe(ka, env, sizeof *frame); 3611 trace_user_setup_frame(env, frame_addr); 3612 if (!lock_user_struct(VERIFY_WRITE, frame, frame_addr, 0)) 3613 goto badframe; 3614 3615 /* Save the mask. */ 3616 __put_user(set->sig[0], &frame->uc.tuc_mcontext.oldmask); 3617 3618 for(i = 1; i < TARGET_NSIG_WORDS; i++) { 3619 __put_user(set->sig[i], &frame->extramask[i - 1]); 3620 } 3621 3622 setup_sigcontext(&frame->uc.tuc_mcontext, env); 3623 3624 /* Set up to return from userspace. If provided, use a stub 3625 already in userspace. */ 3626 /* minus 8 is offset to cater for "rtsd r15,8" offset */ 3627 if (ka->sa_flags & TARGET_SA_RESTORER) { 3628 env->regs[15] = ((unsigned long)ka->sa_restorer)-8; 3629 } else { 3630 uint32_t t; 3631 /* Note, these encodings are _big endian_! */ 3632 /* addi r12, r0, __NR_sigreturn */ 3633 t = 0x31800000UL | TARGET_NR_sigreturn; 3634 __put_user(t, frame->tramp + 0); 3635 /* brki r14, 0x8 */ 3636 t = 0xb9cc0008UL; 3637 __put_user(t, frame->tramp + 1); 3638 3639 /* Return from sighandler will jump to the tramp. 3640 Negative 8 offset because return is rtsd r15, 8 */ 3641 env->regs[15] = frame_addr + offsetof(struct target_signal_frame, tramp) 3642 - 8; 3643 } 3644 3645 /* Set up registers for signal handler */ 3646 env->regs[1] = frame_addr; 3647 /* Signal handler args: */ 3648 env->regs[5] = sig; /* Arg 0: signum */ 3649 env->regs[6] = 0; 3650 /* arg 1: sigcontext */ 3651 env->regs[7] = frame_addr += offsetof(typeof(*frame), uc); 3652 3653 /* Offset of 4 to handle microblaze rtid r14, 0 */ 3654 env->sregs[SR_PC] = (unsigned long)ka->_sa_handler; 3655 3656 unlock_user_struct(frame, frame_addr, 1); 3657 return; 3658 badframe: 3659 force_sig(TARGET_SIGSEGV); 3660 } 3661 3662 static void setup_rt_frame(int sig, struct target_sigaction *ka, 3663 target_siginfo_t *info, 3664 target_sigset_t *set, CPUMBState *env) 3665 { 3666 fprintf(stderr, "Microblaze setup_rt_frame: not implemented\n"); 3667 } 3668 3669 long do_sigreturn(CPUMBState *env) 3670 { 3671 struct target_signal_frame *frame; 3672 abi_ulong frame_addr; 3673 target_sigset_t target_set; 3674 sigset_t set; 3675 int i; 3676 3677 frame_addr = env->regs[R_SP]; 3678 trace_user_do_sigreturn(env, frame_addr); 3679 /* Make sure the guest isn't playing games. */ 3680 if (!lock_user_struct(VERIFY_WRITE, frame, frame_addr, 1)) 3681 goto badframe; 3682 3683 /* Restore blocked signals */ 3684 __get_user(target_set.sig[0], &frame->uc.tuc_mcontext.oldmask); 3685 for(i = 1; i < TARGET_NSIG_WORDS; i++) { 3686 __get_user(target_set.sig[i], &frame->extramask[i - 1]); 3687 } 3688 target_to_host_sigset_internal(&set, &target_set); 3689 set_sigmask(&set); 3690 3691 restore_sigcontext(&frame->uc.tuc_mcontext, env); 3692 /* We got here through a sigreturn syscall, our path back is via an 3693 rtb insn so setup r14 for that. */ 3694 env->regs[14] = env->sregs[SR_PC]; 3695 3696 unlock_user_struct(frame, frame_addr, 0); 3697 return -TARGET_QEMU_ESIGRETURN; 3698 badframe: 3699 force_sig(TARGET_SIGSEGV); 3700 } 3701 3702 long do_rt_sigreturn(CPUMBState *env) 3703 { 3704 trace_user_do_rt_sigreturn(env, 0); 3705 fprintf(stderr, "Microblaze do_rt_sigreturn: not implemented\n"); 3706 return -TARGET_ENOSYS; 3707 } 3708 3709 #elif defined(TARGET_CRIS) 3710 3711 struct target_sigcontext { 3712 struct target_pt_regs regs; /* needs to be first */ 3713 uint32_t oldmask; 3714 uint32_t usp; /* usp before stacking this gunk on it */ 3715 }; 3716 3717 /* Signal frames. */ 3718 struct target_signal_frame { 3719 struct target_sigcontext sc; 3720 uint32_t extramask[TARGET_NSIG_WORDS - 1]; 3721 uint16_t retcode[4]; /* Trampoline code. */ 3722 }; 3723 3724 struct rt_signal_frame { 3725 siginfo_t *pinfo; 3726 void *puc; 3727 siginfo_t info; 3728 struct ucontext uc; 3729 uint16_t retcode[4]; /* Trampoline code. */ 3730 }; 3731 3732 static void setup_sigcontext(struct target_sigcontext *sc, CPUCRISState *env) 3733 { 3734 __put_user(env->regs[0], &sc->regs.r0); 3735 __put_user(env->regs[1], &sc->regs.r1); 3736 __put_user(env->regs[2], &sc->regs.r2); 3737 __put_user(env->regs[3], &sc->regs.r3); 3738 __put_user(env->regs[4], &sc->regs.r4); 3739 __put_user(env->regs[5], &sc->regs.r5); 3740 __put_user(env->regs[6], &sc->regs.r6); 3741 __put_user(env->regs[7], &sc->regs.r7); 3742 __put_user(env->regs[8], &sc->regs.r8); 3743 __put_user(env->regs[9], &sc->regs.r9); 3744 __put_user(env->regs[10], &sc->regs.r10); 3745 __put_user(env->regs[11], &sc->regs.r11); 3746 __put_user(env->regs[12], &sc->regs.r12); 3747 __put_user(env->regs[13], &sc->regs.r13); 3748 __put_user(env->regs[14], &sc->usp); 3749 __put_user(env->regs[15], &sc->regs.acr); 3750 __put_user(env->pregs[PR_MOF], &sc->regs.mof); 3751 __put_user(env->pregs[PR_SRP], &sc->regs.srp); 3752 __put_user(env->pc, &sc->regs.erp); 3753 } 3754 3755 static void restore_sigcontext(struct target_sigcontext *sc, CPUCRISState *env) 3756 { 3757 __get_user(env->regs[0], &sc->regs.r0); 3758 __get_user(env->regs[1], &sc->regs.r1); 3759 __get_user(env->regs[2], &sc->regs.r2); 3760 __get_user(env->regs[3], &sc->regs.r3); 3761 __get_user(env->regs[4], &sc->regs.r4); 3762 __get_user(env->regs[5], &sc->regs.r5); 3763 __get_user(env->regs[6], &sc->regs.r6); 3764 __get_user(env->regs[7], &sc->regs.r7); 3765 __get_user(env->regs[8], &sc->regs.r8); 3766 __get_user(env->regs[9], &sc->regs.r9); 3767 __get_user(env->regs[10], &sc->regs.r10); 3768 __get_user(env->regs[11], &sc->regs.r11); 3769 __get_user(env->regs[12], &sc->regs.r12); 3770 __get_user(env->regs[13], &sc->regs.r13); 3771 __get_user(env->regs[14], &sc->usp); 3772 __get_user(env->regs[15], &sc->regs.acr); 3773 __get_user(env->pregs[PR_MOF], &sc->regs.mof); 3774 __get_user(env->pregs[PR_SRP], &sc->regs.srp); 3775 __get_user(env->pc, &sc->regs.erp); 3776 } 3777 3778 static abi_ulong get_sigframe(CPUCRISState *env, int framesize) 3779 { 3780 abi_ulong sp; 3781 /* Align the stack downwards to 4. */ 3782 sp = (env->regs[R_SP] & ~3); 3783 return sp - framesize; 3784 } 3785 3786 static void setup_frame(int sig, struct target_sigaction *ka, 3787 target_sigset_t *set, CPUCRISState *env) 3788 { 3789 struct target_signal_frame *frame; 3790 abi_ulong frame_addr; 3791 int i; 3792 3793 frame_addr = get_sigframe(env, sizeof *frame); 3794 trace_user_setup_frame(env, frame_addr); 3795 if (!lock_user_struct(VERIFY_WRITE, frame, frame_addr, 0)) 3796 goto badframe; 3797 3798 /* 3799 * The CRIS signal return trampoline. A real linux/CRIS kernel doesn't 3800 * use this trampoline anymore but it sets it up for GDB. 3801 * In QEMU, using the trampoline simplifies things a bit so we use it. 3802 * 3803 * This is movu.w __NR_sigreturn, r9; break 13; 3804 */ 3805 __put_user(0x9c5f, frame->retcode+0); 3806 __put_user(TARGET_NR_sigreturn, 3807 frame->retcode + 1); 3808 __put_user(0xe93d, frame->retcode + 2); 3809 3810 /* Save the mask. */ 3811 __put_user(set->sig[0], &frame->sc.oldmask); 3812 3813 for(i = 1; i < TARGET_NSIG_WORDS; i++) { 3814 __put_user(set->sig[i], &frame->extramask[i - 1]); 3815 } 3816 3817 setup_sigcontext(&frame->sc, env); 3818 3819 /* Move the stack and setup the arguments for the handler. */ 3820 env->regs[R_SP] = frame_addr; 3821 env->regs[10] = sig; 3822 env->pc = (unsigned long) ka->_sa_handler; 3823 /* Link SRP so the guest returns through the trampoline. */ 3824 env->pregs[PR_SRP] = frame_addr + offsetof(typeof(*frame), retcode); 3825 3826 unlock_user_struct(frame, frame_addr, 1); 3827 return; 3828 badframe: 3829 force_sig(TARGET_SIGSEGV); 3830 } 3831 3832 static void setup_rt_frame(int sig, struct target_sigaction *ka, 3833 target_siginfo_t *info, 3834 target_sigset_t *set, CPUCRISState *env) 3835 { 3836 fprintf(stderr, "CRIS setup_rt_frame: not implemented\n"); 3837 } 3838 3839 long do_sigreturn(CPUCRISState *env) 3840 { 3841 struct target_signal_frame *frame; 3842 abi_ulong frame_addr; 3843 target_sigset_t target_set; 3844 sigset_t set; 3845 int i; 3846 3847 frame_addr = env->regs[R_SP]; 3848 trace_user_do_sigreturn(env, frame_addr); 3849 /* Make sure the guest isn't playing games. */ 3850 if (!lock_user_struct(VERIFY_WRITE, frame, frame_addr, 1)) { 3851 goto badframe; 3852 } 3853 3854 /* Restore blocked signals */ 3855 __get_user(target_set.sig[0], &frame->sc.oldmask); 3856 for(i = 1; i < TARGET_NSIG_WORDS; i++) { 3857 __get_user(target_set.sig[i], &frame->extramask[i - 1]); 3858 } 3859 target_to_host_sigset_internal(&set, &target_set); 3860 set_sigmask(&set); 3861 3862 restore_sigcontext(&frame->sc, env); 3863 unlock_user_struct(frame, frame_addr, 0); 3864 return -TARGET_QEMU_ESIGRETURN; 3865 badframe: 3866 force_sig(TARGET_SIGSEGV); 3867 } 3868 3869 long do_rt_sigreturn(CPUCRISState *env) 3870 { 3871 trace_user_do_rt_sigreturn(env, 0); 3872 fprintf(stderr, "CRIS do_rt_sigreturn: not implemented\n"); 3873 return -TARGET_ENOSYS; 3874 } 3875 3876 #elif defined(TARGET_OPENRISC) 3877 3878 struct target_sigcontext { 3879 struct target_pt_regs regs; 3880 abi_ulong oldmask; 3881 abi_ulong usp; 3882 }; 3883 3884 struct target_ucontext { 3885 abi_ulong tuc_flags; 3886 abi_ulong tuc_link; 3887 target_stack_t tuc_stack; 3888 struct target_sigcontext tuc_mcontext; 3889 target_sigset_t tuc_sigmask; /* mask last for extensibility */ 3890 }; 3891 3892 struct target_rt_sigframe { 3893 abi_ulong pinfo; 3894 uint64_t puc; 3895 struct target_siginfo info; 3896 struct target_sigcontext sc; 3897 struct target_ucontext uc; 3898 unsigned char retcode[16]; /* trampoline code */ 3899 }; 3900 3901 /* This is the asm-generic/ucontext.h version */ 3902 #if 0 3903 static int restore_sigcontext(CPUOpenRISCState *regs, 3904 struct target_sigcontext *sc) 3905 { 3906 unsigned int err = 0; 3907 unsigned long old_usp; 3908 3909 /* Alwys make any pending restarted system call return -EINTR */ 3910 current_thread_info()->restart_block.fn = do_no_restart_syscall; 3911 3912 /* restore the regs from &sc->regs (same as sc, since regs is first) 3913 * (sc is already checked for VERIFY_READ since the sigframe was 3914 * checked in sys_sigreturn previously) 3915 */ 3916 3917 if (copy_from_user(regs, &sc, sizeof(struct target_pt_regs))) { 3918 goto badframe; 3919 } 3920 3921 /* make sure the U-flag is set so user-mode cannot fool us */ 3922 3923 regs->sr &= ~SR_SM; 3924 3925 /* restore the old USP as it was before we stacked the sc etc. 3926 * (we cannot just pop the sigcontext since we aligned the sp and 3927 * stuff after pushing it) 3928 */ 3929 3930 __get_user(old_usp, &sc->usp); 3931 phx_signal("old_usp 0x%lx", old_usp); 3932 3933 __PHX__ REALLY /* ??? */ 3934 wrusp(old_usp); 3935 regs->gpr[1] = old_usp; 3936 3937 /* TODO: the other ports use regs->orig_XX to disable syscall checks 3938 * after this completes, but we don't use that mechanism. maybe we can 3939 * use it now ? 3940 */ 3941 3942 return err; 3943 3944 badframe: 3945 return 1; 3946 } 3947 #endif 3948 3949 /* Set up a signal frame. */ 3950 3951 static void setup_sigcontext(struct target_sigcontext *sc, 3952 CPUOpenRISCState *regs, 3953 unsigned long mask) 3954 { 3955 unsigned long usp = regs->gpr[1]; 3956 3957 /* copy the regs. they are first in sc so we can use sc directly */ 3958 3959 /*copy_to_user(&sc, regs, sizeof(struct target_pt_regs));*/ 3960 3961 /* Set the frametype to CRIS_FRAME_NORMAL for the execution of 3962 the signal handler. The frametype will be restored to its previous 3963 value in restore_sigcontext. */ 3964 /*regs->frametype = CRIS_FRAME_NORMAL;*/ 3965 3966 /* then some other stuff */ 3967 __put_user(mask, &sc->oldmask); 3968 __put_user(usp, &sc->usp); 3969 } 3970 3971 static inline unsigned long align_sigframe(unsigned long sp) 3972 { 3973 return sp & ~3UL; 3974 } 3975 3976 static inline abi_ulong get_sigframe(struct target_sigaction *ka, 3977 CPUOpenRISCState *regs, 3978 size_t frame_size) 3979 { 3980 unsigned long sp = regs->gpr[1]; 3981 int onsigstack = on_sig_stack(sp); 3982 3983 /* redzone */ 3984 /* This is the X/Open sanctioned signal stack switching. */ 3985 if ((ka->sa_flags & TARGET_SA_ONSTACK) != 0 && !onsigstack) { 3986 sp = target_sigaltstack_used.ss_sp + target_sigaltstack_used.ss_size; 3987 } 3988 3989 sp = align_sigframe(sp - frame_size); 3990 3991 /* 3992 * If we are on the alternate signal stack and would overflow it, don't. 3993 * Return an always-bogus address instead so we will die with SIGSEGV. 3994 */ 3995 3996 if (onsigstack && !likely(on_sig_stack(sp))) { 3997 return -1L; 3998 } 3999 4000 return sp; 4001 } 4002 4003 static void setup_rt_frame(int sig, struct target_sigaction *ka, 4004 target_siginfo_t *info, 4005 target_sigset_t *set, CPUOpenRISCState *env) 4006 { 4007 int err = 0; 4008 abi_ulong frame_addr; 4009 unsigned long return_ip; 4010 struct target_rt_sigframe *frame; 4011 abi_ulong info_addr, uc_addr; 4012 4013 frame_addr = get_sigframe(ka, env, sizeof(*frame)); 4014 trace_user_setup_rt_frame(env, frame_addr); 4015 if (!lock_user_struct(VERIFY_WRITE, frame, frame_addr, 0)) { 4016 goto give_sigsegv; 4017 } 4018 4019 info_addr = frame_addr + offsetof(struct target_rt_sigframe, info); 4020 __put_user(info_addr, &frame->pinfo); 4021 uc_addr = frame_addr + offsetof(struct target_rt_sigframe, uc); 4022 __put_user(uc_addr, &frame->puc); 4023 4024 if (ka->sa_flags & SA_SIGINFO) { 4025 tswap_siginfo(&frame->info, info); 4026 } 4027 4028 /*err |= __clear_user(&frame->uc, offsetof(struct ucontext, uc_mcontext));*/ 4029 __put_user(0, &frame->uc.tuc_flags); 4030 __put_user(0, &frame->uc.tuc_link); 4031 __put_user(target_sigaltstack_used.ss_sp, 4032 &frame->uc.tuc_stack.ss_sp); 4033 __put_user(sas_ss_flags(env->gpr[1]), &frame->uc.tuc_stack.ss_flags); 4034 __put_user(target_sigaltstack_used.ss_size, 4035 &frame->uc.tuc_stack.ss_size); 4036 setup_sigcontext(&frame->sc, env, set->sig[0]); 4037 4038 /*err |= copy_to_user(frame->uc.tuc_sigmask, set, sizeof(*set));*/ 4039 4040 /* trampoline - the desired return ip is the retcode itself */ 4041 return_ip = (unsigned long)&frame->retcode; 4042 /* This is l.ori r11,r0,__NR_sigreturn, l.sys 1 */ 4043 __put_user(0xa960, (short *)(frame->retcode + 0)); 4044 __put_user(TARGET_NR_rt_sigreturn, (short *)(frame->retcode + 2)); 4045 __put_user(0x20000001, (unsigned long *)(frame->retcode + 4)); 4046 __put_user(0x15000000, (unsigned long *)(frame->retcode + 8)); 4047 4048 if (err) { 4049 goto give_sigsegv; 4050 } 4051 4052 /* TODO what is the current->exec_domain stuff and invmap ? */ 4053 4054 /* Set up registers for signal handler */ 4055 env->pc = (unsigned long)ka->_sa_handler; /* what we enter NOW */ 4056 env->gpr[9] = (unsigned long)return_ip; /* what we enter LATER */ 4057 env->gpr[3] = (unsigned long)sig; /* arg 1: signo */ 4058 env->gpr[4] = (unsigned long)&frame->info; /* arg 2: (siginfo_t*) */ 4059 env->gpr[5] = (unsigned long)&frame->uc; /* arg 3: ucontext */ 4060 4061 /* actually move the usp to reflect the stacked frame */ 4062 env->gpr[1] = (unsigned long)frame; 4063 4064 return; 4065 4066 give_sigsegv: 4067 unlock_user_struct(frame, frame_addr, 1); 4068 if (sig == TARGET_SIGSEGV) { 4069 ka->_sa_handler = TARGET_SIG_DFL; 4070 } 4071 force_sig(TARGET_SIGSEGV); 4072 } 4073 4074 long do_sigreturn(CPUOpenRISCState *env) 4075 { 4076 trace_user_do_sigreturn(env, 0); 4077 fprintf(stderr, "do_sigreturn: not implemented\n"); 4078 return -TARGET_ENOSYS; 4079 } 4080 4081 long do_rt_sigreturn(CPUOpenRISCState *env) 4082 { 4083 trace_user_do_rt_sigreturn(env, 0); 4084 fprintf(stderr, "do_rt_sigreturn: not implemented\n"); 4085 return -TARGET_ENOSYS; 4086 } 4087 /* TARGET_OPENRISC */ 4088 4089 #elif defined(TARGET_S390X) 4090 4091 #define __NUM_GPRS 16 4092 #define __NUM_FPRS 16 4093 #define __NUM_ACRS 16 4094 4095 #define S390_SYSCALL_SIZE 2 4096 #define __SIGNAL_FRAMESIZE 160 /* FIXME: 31-bit mode -> 96 */ 4097 4098 #define _SIGCONTEXT_NSIG 64 4099 #define _SIGCONTEXT_NSIG_BPW 64 /* FIXME: 31-bit mode -> 32 */ 4100 #define _SIGCONTEXT_NSIG_WORDS (_SIGCONTEXT_NSIG / _SIGCONTEXT_NSIG_BPW) 4101 #define _SIGMASK_COPY_SIZE (sizeof(unsigned long)*_SIGCONTEXT_NSIG_WORDS) 4102 #define PSW_ADDR_AMODE 0x0000000000000000UL /* 0x80000000UL for 31-bit */ 4103 #define S390_SYSCALL_OPCODE ((uint16_t)0x0a00) 4104 4105 typedef struct { 4106 target_psw_t psw; 4107 target_ulong gprs[__NUM_GPRS]; 4108 unsigned int acrs[__NUM_ACRS]; 4109 } target_s390_regs_common; 4110 4111 typedef struct { 4112 unsigned int fpc; 4113 double fprs[__NUM_FPRS]; 4114 } target_s390_fp_regs; 4115 4116 typedef struct { 4117 target_s390_regs_common regs; 4118 target_s390_fp_regs fpregs; 4119 } target_sigregs; 4120 4121 struct target_sigcontext { 4122 target_ulong oldmask[_SIGCONTEXT_NSIG_WORDS]; 4123 target_sigregs *sregs; 4124 }; 4125 4126 typedef struct { 4127 uint8_t callee_used_stack[__SIGNAL_FRAMESIZE]; 4128 struct target_sigcontext sc; 4129 target_sigregs sregs; 4130 int signo; 4131 uint8_t retcode[S390_SYSCALL_SIZE]; 4132 } sigframe; 4133 4134 struct target_ucontext { 4135 target_ulong tuc_flags; 4136 struct target_ucontext *tuc_link; 4137 target_stack_t tuc_stack; 4138 target_sigregs tuc_mcontext; 4139 target_sigset_t tuc_sigmask; /* mask last for extensibility */ 4140 }; 4141 4142 typedef struct { 4143 uint8_t callee_used_stack[__SIGNAL_FRAMESIZE]; 4144 uint8_t retcode[S390_SYSCALL_SIZE]; 4145 struct target_siginfo info; 4146 struct target_ucontext uc; 4147 } rt_sigframe; 4148 4149 static inline abi_ulong 4150 get_sigframe(struct target_sigaction *ka, CPUS390XState *env, size_t frame_size) 4151 { 4152 abi_ulong sp; 4153 4154 /* Default to using normal stack */ 4155 sp = env->regs[15]; 4156 4157 /* This is the X/Open sanctioned signal stack switching. */ 4158 if (ka->sa_flags & TARGET_SA_ONSTACK) { 4159 if (!sas_ss_flags(sp)) { 4160 sp = target_sigaltstack_used.ss_sp + 4161 target_sigaltstack_used.ss_size; 4162 } 4163 } 4164 4165 /* This is the legacy signal stack switching. */ 4166 else if (/* FIXME !user_mode(regs) */ 0 && 4167 !(ka->sa_flags & TARGET_SA_RESTORER) && 4168 ka->sa_restorer) { 4169 sp = (abi_ulong) ka->sa_restorer; 4170 } 4171 4172 return (sp - frame_size) & -8ul; 4173 } 4174 4175 static void save_sigregs(CPUS390XState *env, target_sigregs *sregs) 4176 { 4177 int i; 4178 //save_access_regs(current->thread.acrs); FIXME 4179 4180 /* Copy a 'clean' PSW mask to the user to avoid leaking 4181 information about whether PER is currently on. */ 4182 __put_user(env->psw.mask, &sregs->regs.psw.mask); 4183 __put_user(env->psw.addr, &sregs->regs.psw.addr); 4184 for (i = 0; i < 16; i++) { 4185 __put_user(env->regs[i], &sregs->regs.gprs[i]); 4186 } 4187 for (i = 0; i < 16; i++) { 4188 __put_user(env->aregs[i], &sregs->regs.acrs[i]); 4189 } 4190 /* 4191 * We have to store the fp registers to current->thread.fp_regs 4192 * to merge them with the emulated registers. 4193 */ 4194 //save_fp_regs(¤t->thread.fp_regs); FIXME 4195 for (i = 0; i < 16; i++) { 4196 __put_user(get_freg(env, i)->ll, &sregs->fpregs.fprs[i]); 4197 } 4198 } 4199 4200 static void setup_frame(int sig, struct target_sigaction *ka, 4201 target_sigset_t *set, CPUS390XState *env) 4202 { 4203 sigframe *frame; 4204 abi_ulong frame_addr; 4205 4206 frame_addr = get_sigframe(ka, env, sizeof(*frame)); 4207 trace_user_setup_frame(env, frame_addr); 4208 if (!lock_user_struct(VERIFY_WRITE, frame, frame_addr, 0)) { 4209 goto give_sigsegv; 4210 } 4211 4212 __put_user(set->sig[0], &frame->sc.oldmask[0]); 4213 4214 save_sigregs(env, &frame->sregs); 4215 4216 __put_user((abi_ulong)(unsigned long)&frame->sregs, 4217 (abi_ulong *)&frame->sc.sregs); 4218 4219 /* Set up to return from userspace. If provided, use a stub 4220 already in userspace. */ 4221 if (ka->sa_flags & TARGET_SA_RESTORER) { 4222 env->regs[14] = (unsigned long) 4223 ka->sa_restorer | PSW_ADDR_AMODE; 4224 } else { 4225 env->regs[14] = (frame_addr + offsetof(sigframe, retcode)) 4226 | PSW_ADDR_AMODE; 4227 __put_user(S390_SYSCALL_OPCODE | TARGET_NR_sigreturn, 4228 (uint16_t *)(frame->retcode)); 4229 } 4230 4231 /* Set up backchain. */ 4232 __put_user(env->regs[15], (abi_ulong *) frame); 4233 4234 /* Set up registers for signal handler */ 4235 env->regs[15] = frame_addr; 4236 env->psw.addr = (target_ulong) ka->_sa_handler | PSW_ADDR_AMODE; 4237 4238 env->regs[2] = sig; //map_signal(sig); 4239 env->regs[3] = frame_addr += offsetof(typeof(*frame), sc); 4240 4241 /* We forgot to include these in the sigcontext. 4242 To avoid breaking binary compatibility, they are passed as args. */ 4243 env->regs[4] = 0; // FIXME: no clue... current->thread.trap_no; 4244 env->regs[5] = 0; // FIXME: no clue... current->thread.prot_addr; 4245 4246 /* Place signal number on stack to allow backtrace from handler. */ 4247 __put_user(env->regs[2], (int *) &frame->signo); 4248 unlock_user_struct(frame, frame_addr, 1); 4249 return; 4250 4251 give_sigsegv: 4252 force_sig(TARGET_SIGSEGV); 4253 } 4254 4255 static void setup_rt_frame(int sig, struct target_sigaction *ka, 4256 target_siginfo_t *info, 4257 target_sigset_t *set, CPUS390XState *env) 4258 { 4259 int i; 4260 rt_sigframe *frame; 4261 abi_ulong frame_addr; 4262 4263 frame_addr = get_sigframe(ka, env, sizeof *frame); 4264 trace_user_setup_rt_frame(env, frame_addr); 4265 if (!lock_user_struct(VERIFY_WRITE, frame, frame_addr, 0)) { 4266 goto give_sigsegv; 4267 } 4268 4269 tswap_siginfo(&frame->info, info); 4270 4271 /* Create the ucontext. */ 4272 __put_user(0, &frame->uc.tuc_flags); 4273 __put_user((abi_ulong)0, (abi_ulong *)&frame->uc.tuc_link); 4274 __put_user(target_sigaltstack_used.ss_sp, &frame->uc.tuc_stack.ss_sp); 4275 __put_user(sas_ss_flags(get_sp_from_cpustate(env)), 4276 &frame->uc.tuc_stack.ss_flags); 4277 __put_user(target_sigaltstack_used.ss_size, &frame->uc.tuc_stack.ss_size); 4278 save_sigregs(env, &frame->uc.tuc_mcontext); 4279 for (i = 0; i < TARGET_NSIG_WORDS; i++) { 4280 __put_user((abi_ulong)set->sig[i], 4281 (abi_ulong *)&frame->uc.tuc_sigmask.sig[i]); 4282 } 4283 4284 /* Set up to return from userspace. If provided, use a stub 4285 already in userspace. */ 4286 if (ka->sa_flags & TARGET_SA_RESTORER) { 4287 env->regs[14] = (unsigned long) ka->sa_restorer | PSW_ADDR_AMODE; 4288 } else { 4289 env->regs[14] = (unsigned long) frame->retcode | PSW_ADDR_AMODE; 4290 __put_user(S390_SYSCALL_OPCODE | TARGET_NR_rt_sigreturn, 4291 (uint16_t *)(frame->retcode)); 4292 } 4293 4294 /* Set up backchain. */ 4295 __put_user(env->regs[15], (abi_ulong *) frame); 4296 4297 /* Set up registers for signal handler */ 4298 env->regs[15] = frame_addr; 4299 env->psw.addr = (target_ulong) ka->_sa_handler | PSW_ADDR_AMODE; 4300 4301 env->regs[2] = sig; //map_signal(sig); 4302 env->regs[3] = frame_addr + offsetof(typeof(*frame), info); 4303 env->regs[4] = frame_addr + offsetof(typeof(*frame), uc); 4304 return; 4305 4306 give_sigsegv: 4307 force_sig(TARGET_SIGSEGV); 4308 } 4309 4310 static int 4311 restore_sigregs(CPUS390XState *env, target_sigregs *sc) 4312 { 4313 int err = 0; 4314 int i; 4315 4316 for (i = 0; i < 16; i++) { 4317 __get_user(env->regs[i], &sc->regs.gprs[i]); 4318 } 4319 4320 __get_user(env->psw.mask, &sc->regs.psw.mask); 4321 trace_user_s390x_restore_sigregs(env, (unsigned long long)sc->regs.psw.addr, 4322 (unsigned long long)env->psw.addr); 4323 __get_user(env->psw.addr, &sc->regs.psw.addr); 4324 /* FIXME: 31-bit -> | PSW_ADDR_AMODE */ 4325 4326 for (i = 0; i < 16; i++) { 4327 __get_user(env->aregs[i], &sc->regs.acrs[i]); 4328 } 4329 for (i = 0; i < 16; i++) { 4330 __get_user(get_freg(env, i)->ll, &sc->fpregs.fprs[i]); 4331 } 4332 4333 return err; 4334 } 4335 4336 long do_sigreturn(CPUS390XState *env) 4337 { 4338 sigframe *frame; 4339 abi_ulong frame_addr = env->regs[15]; 4340 target_sigset_t target_set; 4341 sigset_t set; 4342 4343 trace_user_do_sigreturn(env, frame_addr); 4344 if (!lock_user_struct(VERIFY_READ, frame, frame_addr, 1)) { 4345 goto badframe; 4346 } 4347 __get_user(target_set.sig[0], &frame->sc.oldmask[0]); 4348 4349 target_to_host_sigset_internal(&set, &target_set); 4350 set_sigmask(&set); /* ~_BLOCKABLE? */ 4351 4352 if (restore_sigregs(env, &frame->sregs)) { 4353 goto badframe; 4354 } 4355 4356 unlock_user_struct(frame, frame_addr, 0); 4357 return -TARGET_QEMU_ESIGRETURN; 4358 4359 badframe: 4360 force_sig(TARGET_SIGSEGV); 4361 return 0; 4362 } 4363 4364 long do_rt_sigreturn(CPUS390XState *env) 4365 { 4366 rt_sigframe *frame; 4367 abi_ulong frame_addr = env->regs[15]; 4368 sigset_t set; 4369 4370 trace_user_do_rt_sigreturn(env, frame_addr); 4371 if (!lock_user_struct(VERIFY_READ, frame, frame_addr, 1)) { 4372 goto badframe; 4373 } 4374 target_to_host_sigset(&set, &frame->uc.tuc_sigmask); 4375 4376 set_sigmask(&set); /* ~_BLOCKABLE? */ 4377 4378 if (restore_sigregs(env, &frame->uc.tuc_mcontext)) { 4379 goto badframe; 4380 } 4381 4382 if (do_sigaltstack(frame_addr + offsetof(rt_sigframe, uc.tuc_stack), 0, 4383 get_sp_from_cpustate(env)) == -EFAULT) { 4384 goto badframe; 4385 } 4386 unlock_user_struct(frame, frame_addr, 0); 4387 return -TARGET_QEMU_ESIGRETURN; 4388 4389 badframe: 4390 unlock_user_struct(frame, frame_addr, 0); 4391 force_sig(TARGET_SIGSEGV); 4392 return 0; 4393 } 4394 4395 #elif defined(TARGET_PPC) 4396 4397 /* Size of dummy stack frame allocated when calling signal handler. 4398 See arch/powerpc/include/asm/ptrace.h. */ 4399 #if defined(TARGET_PPC64) 4400 #define SIGNAL_FRAMESIZE 128 4401 #else 4402 #define SIGNAL_FRAMESIZE 64 4403 #endif 4404 4405 /* See arch/powerpc/include/asm/ucontext.h. Only used for 32-bit PPC; 4406 on 64-bit PPC, sigcontext and mcontext are one and the same. */ 4407 struct target_mcontext { 4408 target_ulong mc_gregs[48]; 4409 /* Includes fpscr. */ 4410 uint64_t mc_fregs[33]; 4411 target_ulong mc_pad[2]; 4412 /* We need to handle Altivec and SPE at the same time, which no 4413 kernel needs to do. Fortunately, the kernel defines this bit to 4414 be Altivec-register-large all the time, rather than trying to 4415 twiddle it based on the specific platform. */ 4416 union { 4417 /* SPE vector registers. One extra for SPEFSCR. */ 4418 uint32_t spe[33]; 4419 /* Altivec vector registers. The packing of VSCR and VRSAVE 4420 varies depending on whether we're PPC64 or not: PPC64 splits 4421 them apart; PPC32 stuffs them together. */ 4422 #if defined(TARGET_PPC64) 4423 #define QEMU_NVRREG 34 4424 #else 4425 #define QEMU_NVRREG 33 4426 #endif 4427 ppc_avr_t altivec[QEMU_NVRREG]; 4428 #undef QEMU_NVRREG 4429 } mc_vregs __attribute__((__aligned__(16))); 4430 }; 4431 4432 /* See arch/powerpc/include/asm/sigcontext.h. */ 4433 struct target_sigcontext { 4434 target_ulong _unused[4]; 4435 int32_t signal; 4436 #if defined(TARGET_PPC64) 4437 int32_t pad0; 4438 #endif 4439 target_ulong handler; 4440 target_ulong oldmask; 4441 target_ulong regs; /* struct pt_regs __user * */ 4442 #if defined(TARGET_PPC64) 4443 struct target_mcontext mcontext; 4444 #endif 4445 }; 4446 4447 /* Indices for target_mcontext.mc_gregs, below. 4448 See arch/powerpc/include/asm/ptrace.h for details. */ 4449 enum { 4450 TARGET_PT_R0 = 0, 4451 TARGET_PT_R1 = 1, 4452 TARGET_PT_R2 = 2, 4453 TARGET_PT_R3 = 3, 4454 TARGET_PT_R4 = 4, 4455 TARGET_PT_R5 = 5, 4456 TARGET_PT_R6 = 6, 4457 TARGET_PT_R7 = 7, 4458 TARGET_PT_R8 = 8, 4459 TARGET_PT_R9 = 9, 4460 TARGET_PT_R10 = 10, 4461 TARGET_PT_R11 = 11, 4462 TARGET_PT_R12 = 12, 4463 TARGET_PT_R13 = 13, 4464 TARGET_PT_R14 = 14, 4465 TARGET_PT_R15 = 15, 4466 TARGET_PT_R16 = 16, 4467 TARGET_PT_R17 = 17, 4468 TARGET_PT_R18 = 18, 4469 TARGET_PT_R19 = 19, 4470 TARGET_PT_R20 = 20, 4471 TARGET_PT_R21 = 21, 4472 TARGET_PT_R22 = 22, 4473 TARGET_PT_R23 = 23, 4474 TARGET_PT_R24 = 24, 4475 TARGET_PT_R25 = 25, 4476 TARGET_PT_R26 = 26, 4477 TARGET_PT_R27 = 27, 4478 TARGET_PT_R28 = 28, 4479 TARGET_PT_R29 = 29, 4480 TARGET_PT_R30 = 30, 4481 TARGET_PT_R31 = 31, 4482 TARGET_PT_NIP = 32, 4483 TARGET_PT_MSR = 33, 4484 TARGET_PT_ORIG_R3 = 34, 4485 TARGET_PT_CTR = 35, 4486 TARGET_PT_LNK = 36, 4487 TARGET_PT_XER = 37, 4488 TARGET_PT_CCR = 38, 4489 /* Yes, there are two registers with #39. One is 64-bit only. */ 4490 TARGET_PT_MQ = 39, 4491 TARGET_PT_SOFTE = 39, 4492 TARGET_PT_TRAP = 40, 4493 TARGET_PT_DAR = 41, 4494 TARGET_PT_DSISR = 42, 4495 TARGET_PT_RESULT = 43, 4496 TARGET_PT_REGS_COUNT = 44 4497 }; 4498 4499 4500 struct target_ucontext { 4501 target_ulong tuc_flags; 4502 target_ulong tuc_link; /* struct ucontext __user * */ 4503 struct target_sigaltstack tuc_stack; 4504 #if !defined(TARGET_PPC64) 4505 int32_t tuc_pad[7]; 4506 target_ulong tuc_regs; /* struct mcontext __user * 4507 points to uc_mcontext field */ 4508 #endif 4509 target_sigset_t tuc_sigmask; 4510 #if defined(TARGET_PPC64) 4511 target_sigset_t unused[15]; /* Allow for uc_sigmask growth */ 4512 struct target_sigcontext tuc_sigcontext; 4513 #else 4514 int32_t tuc_maskext[30]; 4515 int32_t tuc_pad2[3]; 4516 struct target_mcontext tuc_mcontext; 4517 #endif 4518 }; 4519 4520 /* See arch/powerpc/kernel/signal_32.c. */ 4521 struct target_sigframe { 4522 struct target_sigcontext sctx; 4523 struct target_mcontext mctx; 4524 int32_t abigap[56]; 4525 }; 4526 4527 #if defined(TARGET_PPC64) 4528 4529 #define TARGET_TRAMP_SIZE 6 4530 4531 struct target_rt_sigframe { 4532 /* sys_rt_sigreturn requires the ucontext be the first field */ 4533 struct target_ucontext uc; 4534 target_ulong _unused[2]; 4535 uint32_t trampoline[TARGET_TRAMP_SIZE]; 4536 target_ulong pinfo; /* struct siginfo __user * */ 4537 target_ulong puc; /* void __user * */ 4538 struct target_siginfo info; 4539 /* 64 bit ABI allows for 288 bytes below sp before decrementing it. */ 4540 char abigap[288]; 4541 } __attribute__((aligned(16))); 4542 4543 #else 4544 4545 struct target_rt_sigframe { 4546 struct target_siginfo info; 4547 struct target_ucontext uc; 4548 int32_t abigap[56]; 4549 }; 4550 4551 #endif 4552 4553 #if defined(TARGET_PPC64) 4554 4555 struct target_func_ptr { 4556 target_ulong entry; 4557 target_ulong toc; 4558 }; 4559 4560 #endif 4561 4562 /* We use the mc_pad field for the signal return trampoline. */ 4563 #define tramp mc_pad 4564 4565 /* See arch/powerpc/kernel/signal.c. */ 4566 static target_ulong get_sigframe(struct target_sigaction *ka, 4567 CPUPPCState *env, 4568 int frame_size) 4569 { 4570 target_ulong oldsp; 4571 4572 oldsp = env->gpr[1]; 4573 4574 if ((ka->sa_flags & TARGET_SA_ONSTACK) && 4575 (sas_ss_flags(oldsp) == 0)) { 4576 oldsp = (target_sigaltstack_used.ss_sp 4577 + target_sigaltstack_used.ss_size); 4578 } 4579 4580 return (oldsp - frame_size) & ~0xFUL; 4581 } 4582 4583 static void save_user_regs(CPUPPCState *env, struct target_mcontext *frame) 4584 { 4585 target_ulong msr = env->msr; 4586 int i; 4587 target_ulong ccr = 0; 4588 4589 /* In general, the kernel attempts to be intelligent about what it 4590 needs to save for Altivec/FP/SPE registers. We don't care that 4591 much, so we just go ahead and save everything. */ 4592 4593 /* Save general registers. */ 4594 for (i = 0; i < ARRAY_SIZE(env->gpr); i++) { 4595 __put_user(env->gpr[i], &frame->mc_gregs[i]); 4596 } 4597 __put_user(env->nip, &frame->mc_gregs[TARGET_PT_NIP]); 4598 __put_user(env->ctr, &frame->mc_gregs[TARGET_PT_CTR]); 4599 __put_user(env->lr, &frame->mc_gregs[TARGET_PT_LNK]); 4600 __put_user(env->xer, &frame->mc_gregs[TARGET_PT_XER]); 4601 4602 for (i = 0; i < ARRAY_SIZE(env->crf); i++) { 4603 ccr |= env->crf[i] << (32 - ((i + 1) * 4)); 4604 } 4605 __put_user(ccr, &frame->mc_gregs[TARGET_PT_CCR]); 4606 4607 /* Save Altivec registers if necessary. */ 4608 if (env->insns_flags & PPC_ALTIVEC) { 4609 for (i = 0; i < ARRAY_SIZE(env->avr); i++) { 4610 ppc_avr_t *avr = &env->avr[i]; 4611 ppc_avr_t *vreg = &frame->mc_vregs.altivec[i]; 4612 4613 __put_user(avr->u64[0], &vreg->u64[0]); 4614 __put_user(avr->u64[1], &vreg->u64[1]); 4615 } 4616 /* Set MSR_VR in the saved MSR value to indicate that 4617 frame->mc_vregs contains valid data. */ 4618 msr |= MSR_VR; 4619 __put_user((uint32_t)env->spr[SPR_VRSAVE], 4620 &frame->mc_vregs.altivec[32].u32[3]); 4621 } 4622 4623 /* Save floating point registers. */ 4624 if (env->insns_flags & PPC_FLOAT) { 4625 for (i = 0; i < ARRAY_SIZE(env->fpr); i++) { 4626 __put_user(env->fpr[i], &frame->mc_fregs[i]); 4627 } 4628 __put_user((uint64_t) env->fpscr, &frame->mc_fregs[32]); 4629 } 4630 4631 /* Save SPE registers. The kernel only saves the high half. */ 4632 if (env->insns_flags & PPC_SPE) { 4633 #if defined(TARGET_PPC64) 4634 for (i = 0; i < ARRAY_SIZE(env->gpr); i++) { 4635 __put_user(env->gpr[i] >> 32, &frame->mc_vregs.spe[i]); 4636 } 4637 #else 4638 for (i = 0; i < ARRAY_SIZE(env->gprh); i++) { 4639 __put_user(env->gprh[i], &frame->mc_vregs.spe[i]); 4640 } 4641 #endif 4642 /* Set MSR_SPE in the saved MSR value to indicate that 4643 frame->mc_vregs contains valid data. */ 4644 msr |= MSR_SPE; 4645 __put_user(env->spe_fscr, &frame->mc_vregs.spe[32]); 4646 } 4647 4648 /* Store MSR. */ 4649 __put_user(msr, &frame->mc_gregs[TARGET_PT_MSR]); 4650 } 4651 4652 static void encode_trampoline(int sigret, uint32_t *tramp) 4653 { 4654 /* Set up the sigreturn trampoline: li r0,sigret; sc. */ 4655 if (sigret) { 4656 __put_user(0x38000000 | sigret, &tramp[0]); 4657 __put_user(0x44000002, &tramp[1]); 4658 } 4659 } 4660 4661 static void restore_user_regs(CPUPPCState *env, 4662 struct target_mcontext *frame, int sig) 4663 { 4664 target_ulong save_r2 = 0; 4665 target_ulong msr; 4666 target_ulong ccr; 4667 4668 int i; 4669 4670 if (!sig) { 4671 save_r2 = env->gpr[2]; 4672 } 4673 4674 /* Restore general registers. */ 4675 for (i = 0; i < ARRAY_SIZE(env->gpr); i++) { 4676 __get_user(env->gpr[i], &frame->mc_gregs[i]); 4677 } 4678 __get_user(env->nip, &frame->mc_gregs[TARGET_PT_NIP]); 4679 __get_user(env->ctr, &frame->mc_gregs[TARGET_PT_CTR]); 4680 __get_user(env->lr, &frame->mc_gregs[TARGET_PT_LNK]); 4681 __get_user(env->xer, &frame->mc_gregs[TARGET_PT_XER]); 4682 __get_user(ccr, &frame->mc_gregs[TARGET_PT_CCR]); 4683 4684 for (i = 0; i < ARRAY_SIZE(env->crf); i++) { 4685 env->crf[i] = (ccr >> (32 - ((i + 1) * 4))) & 0xf; 4686 } 4687 4688 if (!sig) { 4689 env->gpr[2] = save_r2; 4690 } 4691 /* Restore MSR. */ 4692 __get_user(msr, &frame->mc_gregs[TARGET_PT_MSR]); 4693 4694 /* If doing signal return, restore the previous little-endian mode. */ 4695 if (sig) 4696 env->msr = (env->msr & ~(1ull << MSR_LE)) | (msr & (1ull << MSR_LE)); 4697 4698 /* Restore Altivec registers if necessary. */ 4699 if (env->insns_flags & PPC_ALTIVEC) { 4700 for (i = 0; i < ARRAY_SIZE(env->avr); i++) { 4701 ppc_avr_t *avr = &env->avr[i]; 4702 ppc_avr_t *vreg = &frame->mc_vregs.altivec[i]; 4703 4704 __get_user(avr->u64[0], &vreg->u64[0]); 4705 __get_user(avr->u64[1], &vreg->u64[1]); 4706 } 4707 /* Set MSR_VEC in the saved MSR value to indicate that 4708 frame->mc_vregs contains valid data. */ 4709 __get_user(env->spr[SPR_VRSAVE], 4710 (target_ulong *)(&frame->mc_vregs.altivec[32].u32[3])); 4711 } 4712 4713 /* Restore floating point registers. */ 4714 if (env->insns_flags & PPC_FLOAT) { 4715 uint64_t fpscr; 4716 for (i = 0; i < ARRAY_SIZE(env->fpr); i++) { 4717 __get_user(env->fpr[i], &frame->mc_fregs[i]); 4718 } 4719 __get_user(fpscr, &frame->mc_fregs[32]); 4720 env->fpscr = (uint32_t) fpscr; 4721 } 4722 4723 /* Save SPE registers. The kernel only saves the high half. */ 4724 if (env->insns_flags & PPC_SPE) { 4725 #if defined(TARGET_PPC64) 4726 for (i = 0; i < ARRAY_SIZE(env->gpr); i++) { 4727 uint32_t hi; 4728 4729 __get_user(hi, &frame->mc_vregs.spe[i]); 4730 env->gpr[i] = ((uint64_t)hi << 32) | ((uint32_t) env->gpr[i]); 4731 } 4732 #else 4733 for (i = 0; i < ARRAY_SIZE(env->gprh); i++) { 4734 __get_user(env->gprh[i], &frame->mc_vregs.spe[i]); 4735 } 4736 #endif 4737 __get_user(env->spe_fscr, &frame->mc_vregs.spe[32]); 4738 } 4739 } 4740 4741 static void setup_frame(int sig, struct target_sigaction *ka, 4742 target_sigset_t *set, CPUPPCState *env) 4743 { 4744 struct target_sigframe *frame; 4745 struct target_sigcontext *sc; 4746 target_ulong frame_addr, newsp; 4747 int err = 0; 4748 #if defined(TARGET_PPC64) 4749 struct image_info *image = ((TaskState *)thread_cpu->opaque)->info; 4750 #endif 4751 4752 frame_addr = get_sigframe(ka, env, sizeof(*frame)); 4753 trace_user_setup_frame(env, frame_addr); 4754 if (!lock_user_struct(VERIFY_WRITE, frame, frame_addr, 1)) 4755 goto sigsegv; 4756 sc = &frame->sctx; 4757 4758 __put_user(ka->_sa_handler, &sc->handler); 4759 __put_user(set->sig[0], &sc->oldmask); 4760 #if TARGET_ABI_BITS == 64 4761 __put_user(set->sig[0] >> 32, &sc->_unused[3]); 4762 #else 4763 __put_user(set->sig[1], &sc->_unused[3]); 4764 #endif 4765 __put_user(h2g(&frame->mctx), &sc->regs); 4766 __put_user(sig, &sc->signal); 4767 4768 /* Save user regs. */ 4769 save_user_regs(env, &frame->mctx); 4770 4771 /* Construct the trampoline code on the stack. */ 4772 encode_trampoline(TARGET_NR_sigreturn, (uint32_t *)&frame->mctx.tramp); 4773 4774 /* The kernel checks for the presence of a VDSO here. We don't 4775 emulate a vdso, so use a sigreturn system call. */ 4776 env->lr = (target_ulong) h2g(frame->mctx.tramp); 4777 4778 /* Turn off all fp exceptions. */ 4779 env->fpscr = 0; 4780 4781 /* Create a stack frame for the caller of the handler. */ 4782 newsp = frame_addr - SIGNAL_FRAMESIZE; 4783 err |= put_user(env->gpr[1], newsp, target_ulong); 4784 4785 if (err) 4786 goto sigsegv; 4787 4788 /* Set up registers for signal handler. */ 4789 env->gpr[1] = newsp; 4790 env->gpr[3] = sig; 4791 env->gpr[4] = frame_addr + offsetof(struct target_sigframe, sctx); 4792 4793 #if defined(TARGET_PPC64) 4794 if (get_ppc64_abi(image) < 2) { 4795 /* ELFv1 PPC64 function pointers are pointers to OPD entries. */ 4796 struct target_func_ptr *handler = 4797 (struct target_func_ptr *)g2h(ka->_sa_handler); 4798 env->nip = tswapl(handler->entry); 4799 env->gpr[2] = tswapl(handler->toc); 4800 } else { 4801 /* ELFv2 PPC64 function pointers are entry points, but R12 4802 * must also be set */ 4803 env->nip = tswapl((target_ulong) ka->_sa_handler); 4804 env->gpr[12] = env->nip; 4805 } 4806 #else 4807 env->nip = (target_ulong) ka->_sa_handler; 4808 #endif 4809 4810 /* Signal handlers are entered in big-endian mode. */ 4811 env->msr &= ~(1ull << MSR_LE); 4812 4813 unlock_user_struct(frame, frame_addr, 1); 4814 return; 4815 4816 sigsegv: 4817 unlock_user_struct(frame, frame_addr, 1); 4818 force_sig(TARGET_SIGSEGV); 4819 } 4820 4821 static void setup_rt_frame(int sig, struct target_sigaction *ka, 4822 target_siginfo_t *info, 4823 target_sigset_t *set, CPUPPCState *env) 4824 { 4825 struct target_rt_sigframe *rt_sf; 4826 uint32_t *trampptr = 0; 4827 struct target_mcontext *mctx = 0; 4828 target_ulong rt_sf_addr, newsp = 0; 4829 int i, err = 0; 4830 #if defined(TARGET_PPC64) 4831 struct image_info *image = ((TaskState *)thread_cpu->opaque)->info; 4832 #endif 4833 4834 rt_sf_addr = get_sigframe(ka, env, sizeof(*rt_sf)); 4835 if (!lock_user_struct(VERIFY_WRITE, rt_sf, rt_sf_addr, 1)) 4836 goto sigsegv; 4837 4838 tswap_siginfo(&rt_sf->info, info); 4839 4840 __put_user(0, &rt_sf->uc.tuc_flags); 4841 __put_user(0, &rt_sf->uc.tuc_link); 4842 __put_user((target_ulong)target_sigaltstack_used.ss_sp, 4843 &rt_sf->uc.tuc_stack.ss_sp); 4844 __put_user(sas_ss_flags(env->gpr[1]), 4845 &rt_sf->uc.tuc_stack.ss_flags); 4846 __put_user(target_sigaltstack_used.ss_size, 4847 &rt_sf->uc.tuc_stack.ss_size); 4848 #if !defined(TARGET_PPC64) 4849 __put_user(h2g (&rt_sf->uc.tuc_mcontext), 4850 &rt_sf->uc.tuc_regs); 4851 #endif 4852 for(i = 0; i < TARGET_NSIG_WORDS; i++) { 4853 __put_user(set->sig[i], &rt_sf->uc.tuc_sigmask.sig[i]); 4854 } 4855 4856 #if defined(TARGET_PPC64) 4857 mctx = &rt_sf->uc.tuc_sigcontext.mcontext; 4858 trampptr = &rt_sf->trampoline[0]; 4859 #else 4860 mctx = &rt_sf->uc.tuc_mcontext; 4861 trampptr = (uint32_t *)&rt_sf->uc.tuc_mcontext.tramp; 4862 #endif 4863 4864 save_user_regs(env, mctx); 4865 encode_trampoline(TARGET_NR_rt_sigreturn, trampptr); 4866 4867 /* The kernel checks for the presence of a VDSO here. We don't 4868 emulate a vdso, so use a sigreturn system call. */ 4869 env->lr = (target_ulong) h2g(trampptr); 4870 4871 /* Turn off all fp exceptions. */ 4872 env->fpscr = 0; 4873 4874 /* Create a stack frame for the caller of the handler. */ 4875 newsp = rt_sf_addr - (SIGNAL_FRAMESIZE + 16); 4876 err |= put_user(env->gpr[1], newsp, target_ulong); 4877 4878 if (err) 4879 goto sigsegv; 4880 4881 /* Set up registers for signal handler. */ 4882 env->gpr[1] = newsp; 4883 env->gpr[3] = (target_ulong) sig; 4884 env->gpr[4] = (target_ulong) h2g(&rt_sf->info); 4885 env->gpr[5] = (target_ulong) h2g(&rt_sf->uc); 4886 env->gpr[6] = (target_ulong) h2g(rt_sf); 4887 4888 #if defined(TARGET_PPC64) 4889 if (get_ppc64_abi(image) < 2) { 4890 /* ELFv1 PPC64 function pointers are pointers to OPD entries. */ 4891 struct target_func_ptr *handler = 4892 (struct target_func_ptr *)g2h(ka->_sa_handler); 4893 env->nip = tswapl(handler->entry); 4894 env->gpr[2] = tswapl(handler->toc); 4895 } else { 4896 /* ELFv2 PPC64 function pointers are entry points, but R12 4897 * must also be set */ 4898 env->nip = tswapl((target_ulong) ka->_sa_handler); 4899 env->gpr[12] = env->nip; 4900 } 4901 #else 4902 env->nip = (target_ulong) ka->_sa_handler; 4903 #endif 4904 4905 /* Signal handlers are entered in big-endian mode. */ 4906 env->msr &= ~(1ull << MSR_LE); 4907 4908 unlock_user_struct(rt_sf, rt_sf_addr, 1); 4909 return; 4910 4911 sigsegv: 4912 unlock_user_struct(rt_sf, rt_sf_addr, 1); 4913 force_sig(TARGET_SIGSEGV); 4914 4915 } 4916 4917 long do_sigreturn(CPUPPCState *env) 4918 { 4919 struct target_sigcontext *sc = NULL; 4920 struct target_mcontext *sr = NULL; 4921 target_ulong sr_addr = 0, sc_addr; 4922 sigset_t blocked; 4923 target_sigset_t set; 4924 4925 sc_addr = env->gpr[1] + SIGNAL_FRAMESIZE; 4926 if (!lock_user_struct(VERIFY_READ, sc, sc_addr, 1)) 4927 goto sigsegv; 4928 4929 #if defined(TARGET_PPC64) 4930 set.sig[0] = sc->oldmask + ((uint64_t)(sc->_unused[3]) << 32); 4931 #else 4932 __get_user(set.sig[0], &sc->oldmask); 4933 __get_user(set.sig[1], &sc->_unused[3]); 4934 #endif 4935 target_to_host_sigset_internal(&blocked, &set); 4936 set_sigmask(&blocked); 4937 4938 __get_user(sr_addr, &sc->regs); 4939 if (!lock_user_struct(VERIFY_READ, sr, sr_addr, 1)) 4940 goto sigsegv; 4941 restore_user_regs(env, sr, 1); 4942 4943 unlock_user_struct(sr, sr_addr, 1); 4944 unlock_user_struct(sc, sc_addr, 1); 4945 return -TARGET_QEMU_ESIGRETURN; 4946 4947 sigsegv: 4948 unlock_user_struct(sr, sr_addr, 1); 4949 unlock_user_struct(sc, sc_addr, 1); 4950 force_sig(TARGET_SIGSEGV); 4951 return 0; 4952 } 4953 4954 /* See arch/powerpc/kernel/signal_32.c. */ 4955 static int do_setcontext(struct target_ucontext *ucp, CPUPPCState *env, int sig) 4956 { 4957 struct target_mcontext *mcp; 4958 target_ulong mcp_addr; 4959 sigset_t blocked; 4960 target_sigset_t set; 4961 4962 if (copy_from_user(&set, h2g(ucp) + offsetof(struct target_ucontext, tuc_sigmask), 4963 sizeof (set))) 4964 return 1; 4965 4966 #if defined(TARGET_PPC64) 4967 mcp_addr = h2g(ucp) + 4968 offsetof(struct target_ucontext, tuc_sigcontext.mcontext); 4969 #else 4970 __get_user(mcp_addr, &ucp->tuc_regs); 4971 #endif 4972 4973 if (!lock_user_struct(VERIFY_READ, mcp, mcp_addr, 1)) 4974 return 1; 4975 4976 target_to_host_sigset_internal(&blocked, &set); 4977 set_sigmask(&blocked); 4978 restore_user_regs(env, mcp, sig); 4979 4980 unlock_user_struct(mcp, mcp_addr, 1); 4981 return 0; 4982 } 4983 4984 long do_rt_sigreturn(CPUPPCState *env) 4985 { 4986 struct target_rt_sigframe *rt_sf = NULL; 4987 target_ulong rt_sf_addr; 4988 4989 rt_sf_addr = env->gpr[1] + SIGNAL_FRAMESIZE + 16; 4990 if (!lock_user_struct(VERIFY_READ, rt_sf, rt_sf_addr, 1)) 4991 goto sigsegv; 4992 4993 if (do_setcontext(&rt_sf->uc, env, 1)) 4994 goto sigsegv; 4995 4996 do_sigaltstack(rt_sf_addr 4997 + offsetof(struct target_rt_sigframe, uc.tuc_stack), 4998 0, env->gpr[1]); 4999 5000 unlock_user_struct(rt_sf, rt_sf_addr, 1); 5001 return -TARGET_QEMU_ESIGRETURN; 5002 5003 sigsegv: 5004 unlock_user_struct(rt_sf, rt_sf_addr, 1); 5005 force_sig(TARGET_SIGSEGV); 5006 return 0; 5007 } 5008 5009 #elif defined(TARGET_M68K) 5010 5011 struct target_sigcontext { 5012 abi_ulong sc_mask; 5013 abi_ulong sc_usp; 5014 abi_ulong sc_d0; 5015 abi_ulong sc_d1; 5016 abi_ulong sc_a0; 5017 abi_ulong sc_a1; 5018 unsigned short sc_sr; 5019 abi_ulong sc_pc; 5020 }; 5021 5022 struct target_sigframe 5023 { 5024 abi_ulong pretcode; 5025 int sig; 5026 int code; 5027 abi_ulong psc; 5028 char retcode[8]; 5029 abi_ulong extramask[TARGET_NSIG_WORDS-1]; 5030 struct target_sigcontext sc; 5031 }; 5032 5033 typedef int target_greg_t; 5034 #define TARGET_NGREG 18 5035 typedef target_greg_t target_gregset_t[TARGET_NGREG]; 5036 5037 typedef struct target_fpregset { 5038 int f_fpcntl[3]; 5039 int f_fpregs[8*3]; 5040 } target_fpregset_t; 5041 5042 struct target_mcontext { 5043 int version; 5044 target_gregset_t gregs; 5045 target_fpregset_t fpregs; 5046 }; 5047 5048 #define TARGET_MCONTEXT_VERSION 2 5049 5050 struct target_ucontext { 5051 abi_ulong tuc_flags; 5052 abi_ulong tuc_link; 5053 target_stack_t tuc_stack; 5054 struct target_mcontext tuc_mcontext; 5055 abi_long tuc_filler[80]; 5056 target_sigset_t tuc_sigmask; 5057 }; 5058 5059 struct target_rt_sigframe 5060 { 5061 abi_ulong pretcode; 5062 int sig; 5063 abi_ulong pinfo; 5064 abi_ulong puc; 5065 char retcode[8]; 5066 struct target_siginfo info; 5067 struct target_ucontext uc; 5068 }; 5069 5070 static void setup_sigcontext(struct target_sigcontext *sc, CPUM68KState *env, 5071 abi_ulong mask) 5072 { 5073 __put_user(mask, &sc->sc_mask); 5074 __put_user(env->aregs[7], &sc->sc_usp); 5075 __put_user(env->dregs[0], &sc->sc_d0); 5076 __put_user(env->dregs[1], &sc->sc_d1); 5077 __put_user(env->aregs[0], &sc->sc_a0); 5078 __put_user(env->aregs[1], &sc->sc_a1); 5079 __put_user(env->sr, &sc->sc_sr); 5080 __put_user(env->pc, &sc->sc_pc); 5081 } 5082 5083 static void 5084 restore_sigcontext(CPUM68KState *env, struct target_sigcontext *sc) 5085 { 5086 int temp; 5087 5088 __get_user(env->aregs[7], &sc->sc_usp); 5089 __get_user(env->dregs[0], &sc->sc_d0); 5090 __get_user(env->dregs[1], &sc->sc_d1); 5091 __get_user(env->aregs[0], &sc->sc_a0); 5092 __get_user(env->aregs[1], &sc->sc_a1); 5093 __get_user(env->pc, &sc->sc_pc); 5094 __get_user(temp, &sc->sc_sr); 5095 env->sr = (env->sr & 0xff00) | (temp & 0xff); 5096 } 5097 5098 /* 5099 * Determine which stack to use.. 5100 */ 5101 static inline abi_ulong 5102 get_sigframe(struct target_sigaction *ka, CPUM68KState *regs, 5103 size_t frame_size) 5104 { 5105 unsigned long sp; 5106 5107 sp = regs->aregs[7]; 5108 5109 /* This is the X/Open sanctioned signal stack switching. */ 5110 if ((ka->sa_flags & TARGET_SA_ONSTACK) && (sas_ss_flags (sp) == 0)) { 5111 sp = target_sigaltstack_used.ss_sp + target_sigaltstack_used.ss_size; 5112 } 5113 5114 return ((sp - frame_size) & -8UL); 5115 } 5116 5117 static void setup_frame(int sig, struct target_sigaction *ka, 5118 target_sigset_t *set, CPUM68KState *env) 5119 { 5120 struct target_sigframe *frame; 5121 abi_ulong frame_addr; 5122 abi_ulong retcode_addr; 5123 abi_ulong sc_addr; 5124 int i; 5125 5126 frame_addr = get_sigframe(ka, env, sizeof *frame); 5127 trace_user_setup_frame(env, frame_addr); 5128 if (!lock_user_struct(VERIFY_WRITE, frame, frame_addr, 0)) { 5129 goto give_sigsegv; 5130 } 5131 5132 __put_user(sig, &frame->sig); 5133 5134 sc_addr = frame_addr + offsetof(struct target_sigframe, sc); 5135 __put_user(sc_addr, &frame->psc); 5136 5137 setup_sigcontext(&frame->sc, env, set->sig[0]); 5138 5139 for(i = 1; i < TARGET_NSIG_WORDS; i++) { 5140 __put_user(set->sig[i], &frame->extramask[i - 1]); 5141 } 5142 5143 /* Set up to return from userspace. */ 5144 5145 retcode_addr = frame_addr + offsetof(struct target_sigframe, retcode); 5146 __put_user(retcode_addr, &frame->pretcode); 5147 5148 /* moveq #,d0; trap #0 */ 5149 5150 __put_user(0x70004e40 + (TARGET_NR_sigreturn << 16), 5151 (uint32_t *)(frame->retcode)); 5152 5153 /* Set up to return from userspace */ 5154 5155 env->aregs[7] = frame_addr; 5156 env->pc = ka->_sa_handler; 5157 5158 unlock_user_struct(frame, frame_addr, 1); 5159 return; 5160 5161 give_sigsegv: 5162 force_sig(TARGET_SIGSEGV); 5163 } 5164 5165 static inline int target_rt_setup_ucontext(struct target_ucontext *uc, 5166 CPUM68KState *env) 5167 { 5168 target_greg_t *gregs = uc->tuc_mcontext.gregs; 5169 5170 __put_user(TARGET_MCONTEXT_VERSION, &uc->tuc_mcontext.version); 5171 __put_user(env->dregs[0], &gregs[0]); 5172 __put_user(env->dregs[1], &gregs[1]); 5173 __put_user(env->dregs[2], &gregs[2]); 5174 __put_user(env->dregs[3], &gregs[3]); 5175 __put_user(env->dregs[4], &gregs[4]); 5176 __put_user(env->dregs[5], &gregs[5]); 5177 __put_user(env->dregs[6], &gregs[6]); 5178 __put_user(env->dregs[7], &gregs[7]); 5179 __put_user(env->aregs[0], &gregs[8]); 5180 __put_user(env->aregs[1], &gregs[9]); 5181 __put_user(env->aregs[2], &gregs[10]); 5182 __put_user(env->aregs[3], &gregs[11]); 5183 __put_user(env->aregs[4], &gregs[12]); 5184 __put_user(env->aregs[5], &gregs[13]); 5185 __put_user(env->aregs[6], &gregs[14]); 5186 __put_user(env->aregs[7], &gregs[15]); 5187 __put_user(env->pc, &gregs[16]); 5188 __put_user(env->sr, &gregs[17]); 5189 5190 return 0; 5191 } 5192 5193 static inline int target_rt_restore_ucontext(CPUM68KState *env, 5194 struct target_ucontext *uc) 5195 { 5196 int temp; 5197 target_greg_t *gregs = uc->tuc_mcontext.gregs; 5198 5199 __get_user(temp, &uc->tuc_mcontext.version); 5200 if (temp != TARGET_MCONTEXT_VERSION) 5201 goto badframe; 5202 5203 /* restore passed registers */ 5204 __get_user(env->dregs[0], &gregs[0]); 5205 __get_user(env->dregs[1], &gregs[1]); 5206 __get_user(env->dregs[2], &gregs[2]); 5207 __get_user(env->dregs[3], &gregs[3]); 5208 __get_user(env->dregs[4], &gregs[4]); 5209 __get_user(env->dregs[5], &gregs[5]); 5210 __get_user(env->dregs[6], &gregs[6]); 5211 __get_user(env->dregs[7], &gregs[7]); 5212 __get_user(env->aregs[0], &gregs[8]); 5213 __get_user(env->aregs[1], &gregs[9]); 5214 __get_user(env->aregs[2], &gregs[10]); 5215 __get_user(env->aregs[3], &gregs[11]); 5216 __get_user(env->aregs[4], &gregs[12]); 5217 __get_user(env->aregs[5], &gregs[13]); 5218 __get_user(env->aregs[6], &gregs[14]); 5219 __get_user(env->aregs[7], &gregs[15]); 5220 __get_user(env->pc, &gregs[16]); 5221 __get_user(temp, &gregs[17]); 5222 env->sr = (env->sr & 0xff00) | (temp & 0xff); 5223 5224 return 0; 5225 5226 badframe: 5227 return 1; 5228 } 5229 5230 static void setup_rt_frame(int sig, struct target_sigaction *ka, 5231 target_siginfo_t *info, 5232 target_sigset_t *set, CPUM68KState *env) 5233 { 5234 struct target_rt_sigframe *frame; 5235 abi_ulong frame_addr; 5236 abi_ulong retcode_addr; 5237 abi_ulong info_addr; 5238 abi_ulong uc_addr; 5239 int err = 0; 5240 int i; 5241 5242 frame_addr = get_sigframe(ka, env, sizeof *frame); 5243 trace_user_setup_rt_frame(env, frame_addr); 5244 if (!lock_user_struct(VERIFY_WRITE, frame, frame_addr, 0)) { 5245 goto give_sigsegv; 5246 } 5247 5248 __put_user(sig, &frame->sig); 5249 5250 info_addr = frame_addr + offsetof(struct target_rt_sigframe, info); 5251 __put_user(info_addr, &frame->pinfo); 5252 5253 uc_addr = frame_addr + offsetof(struct target_rt_sigframe, uc); 5254 __put_user(uc_addr, &frame->puc); 5255 5256 tswap_siginfo(&frame->info, info); 5257 5258 /* Create the ucontext */ 5259 5260 __put_user(0, &frame->uc.tuc_flags); 5261 __put_user(0, &frame->uc.tuc_link); 5262 __put_user(target_sigaltstack_used.ss_sp, 5263 &frame->uc.tuc_stack.ss_sp); 5264 __put_user(sas_ss_flags(env->aregs[7]), 5265 &frame->uc.tuc_stack.ss_flags); 5266 __put_user(target_sigaltstack_used.ss_size, 5267 &frame->uc.tuc_stack.ss_size); 5268 err |= target_rt_setup_ucontext(&frame->uc, env); 5269 5270 if (err) 5271 goto give_sigsegv; 5272 5273 for(i = 0; i < TARGET_NSIG_WORDS; i++) { 5274 __put_user(set->sig[i], &frame->uc.tuc_sigmask.sig[i]); 5275 } 5276 5277 /* Set up to return from userspace. */ 5278 5279 retcode_addr = frame_addr + offsetof(struct target_sigframe, retcode); 5280 __put_user(retcode_addr, &frame->pretcode); 5281 5282 /* moveq #,d0; notb d0; trap #0 */ 5283 5284 __put_user(0x70004600 + ((TARGET_NR_rt_sigreturn ^ 0xff) << 16), 5285 (uint32_t *)(frame->retcode + 0)); 5286 __put_user(0x4e40, (uint16_t *)(frame->retcode + 4)); 5287 5288 if (err) 5289 goto give_sigsegv; 5290 5291 /* Set up to return from userspace */ 5292 5293 env->aregs[7] = frame_addr; 5294 env->pc = ka->_sa_handler; 5295 5296 unlock_user_struct(frame, frame_addr, 1); 5297 return; 5298 5299 give_sigsegv: 5300 unlock_user_struct(frame, frame_addr, 1); 5301 force_sig(TARGET_SIGSEGV); 5302 } 5303 5304 long do_sigreturn(CPUM68KState *env) 5305 { 5306 struct target_sigframe *frame; 5307 abi_ulong frame_addr = env->aregs[7] - 4; 5308 target_sigset_t target_set; 5309 sigset_t set; 5310 int i; 5311 5312 trace_user_do_sigreturn(env, frame_addr); 5313 if (!lock_user_struct(VERIFY_READ, frame, frame_addr, 1)) 5314 goto badframe; 5315 5316 /* set blocked signals */ 5317 5318 __get_user(target_set.sig[0], &frame->sc.sc_mask); 5319 5320 for(i = 1; i < TARGET_NSIG_WORDS; i++) { 5321 __get_user(target_set.sig[i], &frame->extramask[i - 1]); 5322 } 5323 5324 target_to_host_sigset_internal(&set, &target_set); 5325 set_sigmask(&set); 5326 5327 /* restore registers */ 5328 5329 restore_sigcontext(env, &frame->sc); 5330 5331 unlock_user_struct(frame, frame_addr, 0); 5332 return -TARGET_QEMU_ESIGRETURN; 5333 5334 badframe: 5335 force_sig(TARGET_SIGSEGV); 5336 return 0; 5337 } 5338 5339 long do_rt_sigreturn(CPUM68KState *env) 5340 { 5341 struct target_rt_sigframe *frame; 5342 abi_ulong frame_addr = env->aregs[7] - 4; 5343 target_sigset_t target_set; 5344 sigset_t set; 5345 5346 trace_user_do_rt_sigreturn(env, frame_addr); 5347 if (!lock_user_struct(VERIFY_READ, frame, frame_addr, 1)) 5348 goto badframe; 5349 5350 target_to_host_sigset_internal(&set, &target_set); 5351 set_sigmask(&set); 5352 5353 /* restore registers */ 5354 5355 if (target_rt_restore_ucontext(env, &frame->uc)) 5356 goto badframe; 5357 5358 if (do_sigaltstack(frame_addr + 5359 offsetof(struct target_rt_sigframe, uc.tuc_stack), 5360 0, get_sp_from_cpustate(env)) == -EFAULT) 5361 goto badframe; 5362 5363 unlock_user_struct(frame, frame_addr, 0); 5364 return -TARGET_QEMU_ESIGRETURN; 5365 5366 badframe: 5367 unlock_user_struct(frame, frame_addr, 0); 5368 force_sig(TARGET_SIGSEGV); 5369 return 0; 5370 } 5371 5372 #elif defined(TARGET_ALPHA) 5373 5374 struct target_sigcontext { 5375 abi_long sc_onstack; 5376 abi_long sc_mask; 5377 abi_long sc_pc; 5378 abi_long sc_ps; 5379 abi_long sc_regs[32]; 5380 abi_long sc_ownedfp; 5381 abi_long sc_fpregs[32]; 5382 abi_ulong sc_fpcr; 5383 abi_ulong sc_fp_control; 5384 abi_ulong sc_reserved1; 5385 abi_ulong sc_reserved2; 5386 abi_ulong sc_ssize; 5387 abi_ulong sc_sbase; 5388 abi_ulong sc_traparg_a0; 5389 abi_ulong sc_traparg_a1; 5390 abi_ulong sc_traparg_a2; 5391 abi_ulong sc_fp_trap_pc; 5392 abi_ulong sc_fp_trigger_sum; 5393 abi_ulong sc_fp_trigger_inst; 5394 }; 5395 5396 struct target_ucontext { 5397 abi_ulong tuc_flags; 5398 abi_ulong tuc_link; 5399 abi_ulong tuc_osf_sigmask; 5400 target_stack_t tuc_stack; 5401 struct target_sigcontext tuc_mcontext; 5402 target_sigset_t tuc_sigmask; 5403 }; 5404 5405 struct target_sigframe { 5406 struct target_sigcontext sc; 5407 unsigned int retcode[3]; 5408 }; 5409 5410 struct target_rt_sigframe { 5411 target_siginfo_t info; 5412 struct target_ucontext uc; 5413 unsigned int retcode[3]; 5414 }; 5415 5416 #define INSN_MOV_R30_R16 0x47fe0410 5417 #define INSN_LDI_R0 0x201f0000 5418 #define INSN_CALLSYS 0x00000083 5419 5420 static void setup_sigcontext(struct target_sigcontext *sc, CPUAlphaState *env, 5421 abi_ulong frame_addr, target_sigset_t *set) 5422 { 5423 int i; 5424 5425 __put_user(on_sig_stack(frame_addr), &sc->sc_onstack); 5426 __put_user(set->sig[0], &sc->sc_mask); 5427 __put_user(env->pc, &sc->sc_pc); 5428 __put_user(8, &sc->sc_ps); 5429 5430 for (i = 0; i < 31; ++i) { 5431 __put_user(env->ir[i], &sc->sc_regs[i]); 5432 } 5433 __put_user(0, &sc->sc_regs[31]); 5434 5435 for (i = 0; i < 31; ++i) { 5436 __put_user(env->fir[i], &sc->sc_fpregs[i]); 5437 } 5438 __put_user(0, &sc->sc_fpregs[31]); 5439 __put_user(cpu_alpha_load_fpcr(env), &sc->sc_fpcr); 5440 5441 __put_user(0, &sc->sc_traparg_a0); /* FIXME */ 5442 __put_user(0, &sc->sc_traparg_a1); /* FIXME */ 5443 __put_user(0, &sc->sc_traparg_a2); /* FIXME */ 5444 } 5445 5446 static void restore_sigcontext(CPUAlphaState *env, 5447 struct target_sigcontext *sc) 5448 { 5449 uint64_t fpcr; 5450 int i; 5451 5452 __get_user(env->pc, &sc->sc_pc); 5453 5454 for (i = 0; i < 31; ++i) { 5455 __get_user(env->ir[i], &sc->sc_regs[i]); 5456 } 5457 for (i = 0; i < 31; ++i) { 5458 __get_user(env->fir[i], &sc->sc_fpregs[i]); 5459 } 5460 5461 __get_user(fpcr, &sc->sc_fpcr); 5462 cpu_alpha_store_fpcr(env, fpcr); 5463 } 5464 5465 static inline abi_ulong get_sigframe(struct target_sigaction *sa, 5466 CPUAlphaState *env, 5467 unsigned long framesize) 5468 { 5469 abi_ulong sp = env->ir[IR_SP]; 5470 5471 /* This is the X/Open sanctioned signal stack switching. */ 5472 if ((sa->sa_flags & TARGET_SA_ONSTACK) != 0 && !sas_ss_flags(sp)) { 5473 sp = target_sigaltstack_used.ss_sp + target_sigaltstack_used.ss_size; 5474 } 5475 return (sp - framesize) & -32; 5476 } 5477 5478 static void setup_frame(int sig, struct target_sigaction *ka, 5479 target_sigset_t *set, CPUAlphaState *env) 5480 { 5481 abi_ulong frame_addr, r26; 5482 struct target_sigframe *frame; 5483 int err = 0; 5484 5485 frame_addr = get_sigframe(ka, env, sizeof(*frame)); 5486 trace_user_setup_frame(env, frame_addr); 5487 if (!lock_user_struct(VERIFY_WRITE, frame, frame_addr, 0)) { 5488 goto give_sigsegv; 5489 } 5490 5491 setup_sigcontext(&frame->sc, env, frame_addr, set); 5492 5493 if (ka->sa_restorer) { 5494 r26 = ka->sa_restorer; 5495 } else { 5496 __put_user(INSN_MOV_R30_R16, &frame->retcode[0]); 5497 __put_user(INSN_LDI_R0 + TARGET_NR_sigreturn, 5498 &frame->retcode[1]); 5499 __put_user(INSN_CALLSYS, &frame->retcode[2]); 5500 /* imb() */ 5501 r26 = frame_addr; 5502 } 5503 5504 unlock_user_struct(frame, frame_addr, 1); 5505 5506 if (err) { 5507 give_sigsegv: 5508 if (sig == TARGET_SIGSEGV) { 5509 ka->_sa_handler = TARGET_SIG_DFL; 5510 } 5511 force_sig(TARGET_SIGSEGV); 5512 } 5513 5514 env->ir[IR_RA] = r26; 5515 env->ir[IR_PV] = env->pc = ka->_sa_handler; 5516 env->ir[IR_A0] = sig; 5517 env->ir[IR_A1] = 0; 5518 env->ir[IR_A2] = frame_addr + offsetof(struct target_sigframe, sc); 5519 env->ir[IR_SP] = frame_addr; 5520 } 5521 5522 static void setup_rt_frame(int sig, struct target_sigaction *ka, 5523 target_siginfo_t *info, 5524 target_sigset_t *set, CPUAlphaState *env) 5525 { 5526 abi_ulong frame_addr, r26; 5527 struct target_rt_sigframe *frame; 5528 int i, err = 0; 5529 5530 frame_addr = get_sigframe(ka, env, sizeof(*frame)); 5531 trace_user_setup_rt_frame(env, frame_addr); 5532 if (!lock_user_struct(VERIFY_WRITE, frame, frame_addr, 0)) { 5533 goto give_sigsegv; 5534 } 5535 5536 tswap_siginfo(&frame->info, info); 5537 5538 __put_user(0, &frame->uc.tuc_flags); 5539 __put_user(0, &frame->uc.tuc_link); 5540 __put_user(set->sig[0], &frame->uc.tuc_osf_sigmask); 5541 __put_user(target_sigaltstack_used.ss_sp, 5542 &frame->uc.tuc_stack.ss_sp); 5543 __put_user(sas_ss_flags(env->ir[IR_SP]), 5544 &frame->uc.tuc_stack.ss_flags); 5545 __put_user(target_sigaltstack_used.ss_size, 5546 &frame->uc.tuc_stack.ss_size); 5547 setup_sigcontext(&frame->uc.tuc_mcontext, env, frame_addr, set); 5548 for (i = 0; i < TARGET_NSIG_WORDS; ++i) { 5549 __put_user(set->sig[i], &frame->uc.tuc_sigmask.sig[i]); 5550 } 5551 5552 if (ka->sa_restorer) { 5553 r26 = ka->sa_restorer; 5554 } else { 5555 __put_user(INSN_MOV_R30_R16, &frame->retcode[0]); 5556 __put_user(INSN_LDI_R0 + TARGET_NR_rt_sigreturn, 5557 &frame->retcode[1]); 5558 __put_user(INSN_CALLSYS, &frame->retcode[2]); 5559 /* imb(); */ 5560 r26 = frame_addr; 5561 } 5562 5563 if (err) { 5564 give_sigsegv: 5565 if (sig == TARGET_SIGSEGV) { 5566 ka->_sa_handler = TARGET_SIG_DFL; 5567 } 5568 force_sig(TARGET_SIGSEGV); 5569 } 5570 5571 env->ir[IR_RA] = r26; 5572 env->ir[IR_PV] = env->pc = ka->_sa_handler; 5573 env->ir[IR_A0] = sig; 5574 env->ir[IR_A1] = frame_addr + offsetof(struct target_rt_sigframe, info); 5575 env->ir[IR_A2] = frame_addr + offsetof(struct target_rt_sigframe, uc); 5576 env->ir[IR_SP] = frame_addr; 5577 } 5578 5579 long do_sigreturn(CPUAlphaState *env) 5580 { 5581 struct target_sigcontext *sc; 5582 abi_ulong sc_addr = env->ir[IR_A0]; 5583 target_sigset_t target_set; 5584 sigset_t set; 5585 5586 if (!lock_user_struct(VERIFY_READ, sc, sc_addr, 1)) { 5587 goto badframe; 5588 } 5589 5590 target_sigemptyset(&target_set); 5591 __get_user(target_set.sig[0], &sc->sc_mask); 5592 5593 target_to_host_sigset_internal(&set, &target_set); 5594 set_sigmask(&set); 5595 5596 restore_sigcontext(env, sc); 5597 unlock_user_struct(sc, sc_addr, 0); 5598 return -TARGET_QEMU_ESIGRETURN; 5599 5600 badframe: 5601 force_sig(TARGET_SIGSEGV); 5602 } 5603 5604 long do_rt_sigreturn(CPUAlphaState *env) 5605 { 5606 abi_ulong frame_addr = env->ir[IR_A0]; 5607 struct target_rt_sigframe *frame; 5608 sigset_t set; 5609 5610 trace_user_do_rt_sigreturn(env, frame_addr); 5611 if (!lock_user_struct(VERIFY_READ, frame, frame_addr, 1)) { 5612 goto badframe; 5613 } 5614 target_to_host_sigset(&set, &frame->uc.tuc_sigmask); 5615 set_sigmask(&set); 5616 5617 restore_sigcontext(env, &frame->uc.tuc_mcontext); 5618 if (do_sigaltstack(frame_addr + offsetof(struct target_rt_sigframe, 5619 uc.tuc_stack), 5620 0, env->ir[IR_SP]) == -EFAULT) { 5621 goto badframe; 5622 } 5623 5624 unlock_user_struct(frame, frame_addr, 0); 5625 return -TARGET_QEMU_ESIGRETURN; 5626 5627 5628 badframe: 5629 unlock_user_struct(frame, frame_addr, 0); 5630 force_sig(TARGET_SIGSEGV); 5631 } 5632 5633 #elif defined(TARGET_TILEGX) 5634 5635 struct target_sigcontext { 5636 union { 5637 /* General-purpose registers. */ 5638 abi_ulong gregs[56]; 5639 struct { 5640 abi_ulong __gregs[53]; 5641 abi_ulong tp; /* Aliases gregs[TREG_TP]. */ 5642 abi_ulong sp; /* Aliases gregs[TREG_SP]. */ 5643 abi_ulong lr; /* Aliases gregs[TREG_LR]. */ 5644 }; 5645 }; 5646 abi_ulong pc; /* Program counter. */ 5647 abi_ulong ics; /* In Interrupt Critical Section? */ 5648 abi_ulong faultnum; /* Fault number. */ 5649 abi_ulong pad[5]; 5650 }; 5651 5652 struct target_ucontext { 5653 abi_ulong tuc_flags; 5654 abi_ulong tuc_link; 5655 target_stack_t tuc_stack; 5656 struct target_sigcontext tuc_mcontext; 5657 target_sigset_t tuc_sigmask; /* mask last for extensibility */ 5658 }; 5659 5660 struct target_rt_sigframe { 5661 unsigned char save_area[16]; /* caller save area */ 5662 struct target_siginfo info; 5663 struct target_ucontext uc; 5664 abi_ulong retcode[2]; 5665 }; 5666 5667 #define INSN_MOVELI_R10_139 0x00045fe551483000ULL /* { moveli r10, 139 } */ 5668 #define INSN_SWINT1 0x286b180051485000ULL /* { swint1 } */ 5669 5670 5671 static void setup_sigcontext(struct target_sigcontext *sc, 5672 CPUArchState *env, int signo) 5673 { 5674 int i; 5675 5676 for (i = 0; i < TILEGX_R_COUNT; ++i) { 5677 __put_user(env->regs[i], &sc->gregs[i]); 5678 } 5679 5680 __put_user(env->pc, &sc->pc); 5681 __put_user(0, &sc->ics); 5682 __put_user(signo, &sc->faultnum); 5683 } 5684 5685 static void restore_sigcontext(CPUTLGState *env, struct target_sigcontext *sc) 5686 { 5687 int i; 5688 5689 for (i = 0; i < TILEGX_R_COUNT; ++i) { 5690 __get_user(env->regs[i], &sc->gregs[i]); 5691 } 5692 5693 __get_user(env->pc, &sc->pc); 5694 } 5695 5696 static abi_ulong get_sigframe(struct target_sigaction *ka, CPUArchState *env, 5697 size_t frame_size) 5698 { 5699 unsigned long sp = env->regs[TILEGX_R_SP]; 5700 5701 if (on_sig_stack(sp) && !likely(on_sig_stack(sp - frame_size))) { 5702 return -1UL; 5703 } 5704 5705 if ((ka->sa_flags & SA_ONSTACK) && !sas_ss_flags(sp)) { 5706 sp = target_sigaltstack_used.ss_sp + target_sigaltstack_used.ss_size; 5707 } 5708 5709 sp -= frame_size; 5710 sp &= -16UL; 5711 return sp; 5712 } 5713 5714 static void setup_rt_frame(int sig, struct target_sigaction *ka, 5715 target_siginfo_t *info, 5716 target_sigset_t *set, CPUArchState *env) 5717 { 5718 abi_ulong frame_addr; 5719 struct target_rt_sigframe *frame; 5720 unsigned long restorer; 5721 5722 frame_addr = get_sigframe(ka, env, sizeof(*frame)); 5723 trace_user_setup_rt_frame(env, frame_addr); 5724 if (!lock_user_struct(VERIFY_WRITE, frame, frame_addr, 0)) { 5725 goto give_sigsegv; 5726 } 5727 5728 /* Always write at least the signal number for the stack backtracer. */ 5729 if (ka->sa_flags & TARGET_SA_SIGINFO) { 5730 /* At sigreturn time, restore the callee-save registers too. */ 5731 tswap_siginfo(&frame->info, info); 5732 /* regs->flags |= PT_FLAGS_RESTORE_REGS; FIXME: we can skip it? */ 5733 } else { 5734 __put_user(info->si_signo, &frame->info.si_signo); 5735 } 5736 5737 /* Create the ucontext. */ 5738 __put_user(0, &frame->uc.tuc_flags); 5739 __put_user(0, &frame->uc.tuc_link); 5740 __put_user(target_sigaltstack_used.ss_sp, &frame->uc.tuc_stack.ss_sp); 5741 __put_user(sas_ss_flags(env->regs[TILEGX_R_SP]), 5742 &frame->uc.tuc_stack.ss_flags); 5743 __put_user(target_sigaltstack_used.ss_size, &frame->uc.tuc_stack.ss_size); 5744 setup_sigcontext(&frame->uc.tuc_mcontext, env, info->si_signo); 5745 5746 if (ka->sa_flags & TARGET_SA_RESTORER) { 5747 restorer = (unsigned long) ka->sa_restorer; 5748 } else { 5749 __put_user(INSN_MOVELI_R10_139, &frame->retcode[0]); 5750 __put_user(INSN_SWINT1, &frame->retcode[1]); 5751 restorer = frame_addr + offsetof(struct target_rt_sigframe, retcode); 5752 } 5753 env->pc = (unsigned long) ka->_sa_handler; 5754 env->regs[TILEGX_R_SP] = (unsigned long) frame; 5755 env->regs[TILEGX_R_LR] = restorer; 5756 env->regs[0] = (unsigned long) sig; 5757 env->regs[1] = (unsigned long) &frame->info; 5758 env->regs[2] = (unsigned long) &frame->uc; 5759 /* regs->flags |= PT_FLAGS_CALLER_SAVES; FIXME: we can skip it? */ 5760 5761 unlock_user_struct(frame, frame_addr, 1); 5762 return; 5763 5764 give_sigsegv: 5765 if (sig == TARGET_SIGSEGV) { 5766 ka->_sa_handler = TARGET_SIG_DFL; 5767 } 5768 force_sig(TARGET_SIGSEGV /* , current */); 5769 } 5770 5771 long do_rt_sigreturn(CPUTLGState *env) 5772 { 5773 abi_ulong frame_addr = env->regs[TILEGX_R_SP]; 5774 struct target_rt_sigframe *frame; 5775 sigset_t set; 5776 5777 trace_user_do_rt_sigreturn(env, frame_addr); 5778 if (!lock_user_struct(VERIFY_READ, frame, frame_addr, 1)) { 5779 goto badframe; 5780 } 5781 target_to_host_sigset(&set, &frame->uc.tuc_sigmask); 5782 set_sigmask(&set); 5783 5784 restore_sigcontext(env, &frame->uc.tuc_mcontext); 5785 if (do_sigaltstack(frame_addr + offsetof(struct target_rt_sigframe, 5786 uc.tuc_stack), 5787 0, env->regs[TILEGX_R_SP]) == -EFAULT) { 5788 goto badframe; 5789 } 5790 5791 unlock_user_struct(frame, frame_addr, 0); 5792 return -TARGET_QEMU_ESIGRETURN; 5793 5794 5795 badframe: 5796 unlock_user_struct(frame, frame_addr, 0); 5797 force_sig(TARGET_SIGSEGV); 5798 } 5799 5800 #else 5801 5802 static void setup_frame(int sig, struct target_sigaction *ka, 5803 target_sigset_t *set, CPUArchState *env) 5804 { 5805 fprintf(stderr, "setup_frame: not implemented\n"); 5806 } 5807 5808 static void setup_rt_frame(int sig, struct target_sigaction *ka, 5809 target_siginfo_t *info, 5810 target_sigset_t *set, CPUArchState *env) 5811 { 5812 fprintf(stderr, "setup_rt_frame: not implemented\n"); 5813 } 5814 5815 long do_sigreturn(CPUArchState *env) 5816 { 5817 fprintf(stderr, "do_sigreturn: not implemented\n"); 5818 return -TARGET_ENOSYS; 5819 } 5820 5821 long do_rt_sigreturn(CPUArchState *env) 5822 { 5823 fprintf(stderr, "do_rt_sigreturn: not implemented\n"); 5824 return -TARGET_ENOSYS; 5825 } 5826 5827 #endif 5828 5829 static void handle_pending_signal(CPUArchState *cpu_env, int sig) 5830 { 5831 CPUState *cpu = ENV_GET_CPU(cpu_env); 5832 abi_ulong handler; 5833 sigset_t set; 5834 target_sigset_t target_old_set; 5835 struct target_sigaction *sa; 5836 TaskState *ts = cpu->opaque; 5837 struct emulated_sigtable *k = &ts->sigtab[sig - 1]; 5838 5839 trace_user_handle_signal(cpu_env, sig); 5840 /* dequeue signal */ 5841 k->pending = 0; 5842 5843 sig = gdb_handlesig(cpu, sig); 5844 if (!sig) { 5845 sa = NULL; 5846 handler = TARGET_SIG_IGN; 5847 } else { 5848 sa = &sigact_table[sig - 1]; 5849 handler = sa->_sa_handler; 5850 } 5851 5852 if (handler == TARGET_SIG_DFL) { 5853 /* default handler : ignore some signal. The other are job control or fatal */ 5854 if (sig == TARGET_SIGTSTP || sig == TARGET_SIGTTIN || sig == TARGET_SIGTTOU) { 5855 kill(getpid(),SIGSTOP); 5856 } else if (sig != TARGET_SIGCHLD && 5857 sig != TARGET_SIGURG && 5858 sig != TARGET_SIGWINCH && 5859 sig != TARGET_SIGCONT) { 5860 force_sig(sig); 5861 } 5862 } else if (handler == TARGET_SIG_IGN) { 5863 /* ignore sig */ 5864 } else if (handler == TARGET_SIG_ERR) { 5865 force_sig(sig); 5866 } else { 5867 /* compute the blocked signals during the handler execution */ 5868 sigset_t *blocked_set; 5869 5870 target_to_host_sigset(&set, &sa->sa_mask); 5871 /* SA_NODEFER indicates that the current signal should not be 5872 blocked during the handler */ 5873 if (!(sa->sa_flags & TARGET_SA_NODEFER)) 5874 sigaddset(&set, target_to_host_signal(sig)); 5875 5876 /* save the previous blocked signal state to restore it at the 5877 end of the signal execution (see do_sigreturn) */ 5878 host_to_target_sigset_internal(&target_old_set, &ts->signal_mask); 5879 5880 /* block signals in the handler */ 5881 blocked_set = ts->in_sigsuspend ? 5882 &ts->sigsuspend_mask : &ts->signal_mask; 5883 sigorset(&ts->signal_mask, blocked_set, &set); 5884 ts->in_sigsuspend = 0; 5885 5886 /* if the CPU is in VM86 mode, we restore the 32 bit values */ 5887 #if defined(TARGET_I386) && !defined(TARGET_X86_64) 5888 { 5889 CPUX86State *env = cpu_env; 5890 if (env->eflags & VM_MASK) 5891 save_v86_state(env); 5892 } 5893 #endif 5894 /* prepare the stack frame of the virtual CPU */ 5895 #if defined(TARGET_ABI_MIPSN32) || defined(TARGET_ABI_MIPSN64) \ 5896 || defined(TARGET_OPENRISC) || defined(TARGET_TILEGX) 5897 /* These targets do not have traditional signals. */ 5898 setup_rt_frame(sig, sa, &k->info, &target_old_set, cpu_env); 5899 #else 5900 if (sa->sa_flags & TARGET_SA_SIGINFO) 5901 setup_rt_frame(sig, sa, &k->info, &target_old_set, cpu_env); 5902 else 5903 setup_frame(sig, sa, &target_old_set, cpu_env); 5904 #endif 5905 if (sa->sa_flags & TARGET_SA_RESETHAND) { 5906 sa->_sa_handler = TARGET_SIG_DFL; 5907 } 5908 } 5909 } 5910 5911 void process_pending_signals(CPUArchState *cpu_env) 5912 { 5913 CPUState *cpu = ENV_GET_CPU(cpu_env); 5914 int sig; 5915 TaskState *ts = cpu->opaque; 5916 sigset_t set; 5917 sigset_t *blocked_set; 5918 5919 while (atomic_read(&ts->signal_pending)) { 5920 /* FIXME: This is not threadsafe. */ 5921 sigfillset(&set); 5922 sigprocmask(SIG_SETMASK, &set, 0); 5923 5924 sig = ts->sync_signal.pending; 5925 if (sig) { 5926 /* Synchronous signals are forced, 5927 * see force_sig_info() and callers in Linux 5928 * Note that not all of our queue_signal() calls in QEMU correspond 5929 * to force_sig_info() calls in Linux (some are send_sig_info()). 5930 * However it seems like a kernel bug to me to allow the process 5931 * to block a synchronous signal since it could then just end up 5932 * looping round and round indefinitely. 5933 */ 5934 if (sigismember(&ts->signal_mask, target_to_host_signal_table[sig]) 5935 || sigact_table[sig - 1]._sa_handler == TARGET_SIG_IGN) { 5936 sigdelset(&ts->signal_mask, target_to_host_signal_table[sig]); 5937 sigact_table[sig - 1]._sa_handler = TARGET_SIG_DFL; 5938 } 5939 5940 handle_pending_signal(cpu_env, sig); 5941 } 5942 5943 for (sig = 1; sig <= TARGET_NSIG; sig++) { 5944 blocked_set = ts->in_sigsuspend ? 5945 &ts->sigsuspend_mask : &ts->signal_mask; 5946 5947 if (ts->sigtab[sig - 1].pending && 5948 (!sigismember(blocked_set, 5949 target_to_host_signal_table[sig]))) { 5950 handle_pending_signal(cpu_env, sig); 5951 /* Restart scan from the beginning */ 5952 sig = 1; 5953 } 5954 } 5955 5956 /* if no signal is pending, unblock signals and recheck (the act 5957 * of unblocking might cause us to take another host signal which 5958 * will set signal_pending again). 5959 */ 5960 atomic_set(&ts->signal_pending, 0); 5961 ts->in_sigsuspend = 0; 5962 set = ts->signal_mask; 5963 sigdelset(&set, SIGSEGV); 5964 sigdelset(&set, SIGBUS); 5965 sigprocmask(SIG_SETMASK, &set, 0); 5966 } 5967 ts->in_sigsuspend = 0; 5968 } 5969