10afacde3Sarnd@arndb.de #define DEBUG 20afacde3Sarnd@arndb.de 3ce8ab854SArnd Bergmann #include <linux/wait.h> 4ce8ab854SArnd Bergmann #include <linux/ptrace.h> 5ce8ab854SArnd Bergmann 6ce8ab854SArnd Bergmann #include <asm/spu.h> 7c6730ed4SJeremy Kerr #include <asm/spu_priv1.h> 8c6730ed4SJeremy Kerr #include <asm/io.h> 9cfff5b23SDave Jones #include <asm/unistd.h> 10ce8ab854SArnd Bergmann 11ce8ab854SArnd Bergmann #include "spufs.h" 12ce8ab854SArnd Bergmann 13ce8ab854SArnd Bergmann /* interrupt-level stop callback function. */ 14ce8ab854SArnd Bergmann void spufs_stop_callback(struct spu *spu) 15ce8ab854SArnd Bergmann { 16ce8ab854SArnd Bergmann struct spu_context *ctx = spu->ctx; 17ce8ab854SArnd Bergmann 18ce8ab854SArnd Bergmann wake_up_all(&ctx->stop_wq); 19ce8ab854SArnd Bergmann } 20ce8ab854SArnd Bergmann 21ce8ab854SArnd Bergmann static inline int spu_stopped(struct spu_context *ctx, u32 * stat) 22ce8ab854SArnd Bergmann { 23ce8ab854SArnd Bergmann struct spu *spu; 24ce8ab854SArnd Bergmann u64 pte_fault; 25ce8ab854SArnd Bergmann 26ce8ab854SArnd Bergmann *stat = ctx->ops->status_read(ctx); 27ce8ab854SArnd Bergmann if (ctx->state != SPU_STATE_RUNNABLE) 28ce8ab854SArnd Bergmann return 1; 29ce8ab854SArnd Bergmann spu = ctx->spu; 30ce8ab854SArnd Bergmann pte_fault = spu->dsisr & 31ce8ab854SArnd Bergmann (MFC_DSISR_PTE_NOT_FOUND | MFC_DSISR_ACCESS_DENIED); 32be703177SSebastian Siewior return (!(*stat & SPU_STATUS_RUNNING) || pte_fault || spu->class_0_pending) ? 33be703177SSebastian Siewior 1 : 0; 34ce8ab854SArnd Bergmann } 35ce8ab854SArnd Bergmann 36c6730ed4SJeremy Kerr static int spu_setup_isolated(struct spu_context *ctx) 37c6730ed4SJeremy Kerr { 38c6730ed4SJeremy Kerr int ret; 39c6730ed4SJeremy Kerr u64 __iomem *mfc_cntl; 40c6730ed4SJeremy Kerr u64 sr1; 41c6730ed4SJeremy Kerr u32 status; 42c6730ed4SJeremy Kerr unsigned long timeout; 43c6730ed4SJeremy Kerr const u32 status_loading = SPU_STATUS_RUNNING 44c6730ed4SJeremy Kerr | SPU_STATUS_ISOLATED_STATE | SPU_STATUS_ISOLATED_LOAD_STATUS; 45c6730ed4SJeremy Kerr 467ec18ab9SChristoph Hellwig ret = -ENODEV; 47c6730ed4SJeremy Kerr if (!isolated_loader) 48c6730ed4SJeremy Kerr goto out; 49c6730ed4SJeremy Kerr 507ec18ab9SChristoph Hellwig /* 517ec18ab9SChristoph Hellwig * We need to exclude userspace access to the context. 527ec18ab9SChristoph Hellwig * 537ec18ab9SChristoph Hellwig * To protect against memory access we invalidate all ptes 547ec18ab9SChristoph Hellwig * and make sure the pagefault handlers block on the mutex. 557ec18ab9SChristoph Hellwig */ 567ec18ab9SChristoph Hellwig spu_unmap_mappings(ctx); 577ec18ab9SChristoph Hellwig 58c6730ed4SJeremy Kerr mfc_cntl = &ctx->spu->priv2->mfc_control_RW; 59c6730ed4SJeremy Kerr 60c6730ed4SJeremy Kerr /* purge the MFC DMA queue to ensure no spurious accesses before we 61c6730ed4SJeremy Kerr * enter kernel mode */ 62c6730ed4SJeremy Kerr timeout = jiffies + HZ; 63c6730ed4SJeremy Kerr out_be64(mfc_cntl, MFC_CNTL_PURGE_DMA_REQUEST); 64c6730ed4SJeremy Kerr while ((in_be64(mfc_cntl) & MFC_CNTL_PURGE_DMA_STATUS_MASK) 65c6730ed4SJeremy Kerr != MFC_CNTL_PURGE_DMA_COMPLETE) { 66c6730ed4SJeremy Kerr if (time_after(jiffies, timeout)) { 67c6730ed4SJeremy Kerr printk(KERN_ERR "%s: timeout flushing MFC DMA queue\n", 68c6730ed4SJeremy Kerr __FUNCTION__); 69c6730ed4SJeremy Kerr ret = -EIO; 707ec18ab9SChristoph Hellwig goto out; 71c6730ed4SJeremy Kerr } 72c6730ed4SJeremy Kerr cond_resched(); 73c6730ed4SJeremy Kerr } 74c6730ed4SJeremy Kerr 75c6730ed4SJeremy Kerr /* put the SPE in kernel mode to allow access to the loader */ 76c6730ed4SJeremy Kerr sr1 = spu_mfc_sr1_get(ctx->spu); 77c6730ed4SJeremy Kerr sr1 &= ~MFC_STATE1_PROBLEM_STATE_MASK; 78c6730ed4SJeremy Kerr spu_mfc_sr1_set(ctx->spu, sr1); 79c6730ed4SJeremy Kerr 80c6730ed4SJeremy Kerr /* start the loader */ 81c6730ed4SJeremy Kerr ctx->ops->signal1_write(ctx, (unsigned long)isolated_loader >> 32); 82c6730ed4SJeremy Kerr ctx->ops->signal2_write(ctx, 83c6730ed4SJeremy Kerr (unsigned long)isolated_loader & 0xffffffff); 84c6730ed4SJeremy Kerr 85c6730ed4SJeremy Kerr ctx->ops->runcntl_write(ctx, 86c6730ed4SJeremy Kerr SPU_RUNCNTL_RUNNABLE | SPU_RUNCNTL_ISOLATE); 87c6730ed4SJeremy Kerr 88c6730ed4SJeremy Kerr ret = 0; 89c6730ed4SJeremy Kerr timeout = jiffies + HZ; 90c6730ed4SJeremy Kerr while (((status = ctx->ops->status_read(ctx)) & status_loading) == 91c6730ed4SJeremy Kerr status_loading) { 92c6730ed4SJeremy Kerr if (time_after(jiffies, timeout)) { 93c6730ed4SJeremy Kerr printk(KERN_ERR "%s: timeout waiting for loader\n", 94c6730ed4SJeremy Kerr __FUNCTION__); 95c6730ed4SJeremy Kerr ret = -EIO; 96c6730ed4SJeremy Kerr goto out_drop_priv; 97c6730ed4SJeremy Kerr } 98c6730ed4SJeremy Kerr cond_resched(); 99c6730ed4SJeremy Kerr } 100c6730ed4SJeremy Kerr 101c6730ed4SJeremy Kerr if (!(status & SPU_STATUS_RUNNING)) { 102c6730ed4SJeremy Kerr /* If isolated LOAD has failed: run SPU, we will get a stop-and 103c6730ed4SJeremy Kerr * signal later. */ 104c6730ed4SJeremy Kerr pr_debug("%s: isolated LOAD failed\n", __FUNCTION__); 105c6730ed4SJeremy Kerr ctx->ops->runcntl_write(ctx, SPU_RUNCNTL_RUNNABLE); 106c6730ed4SJeremy Kerr ret = -EACCES; 1077ec18ab9SChristoph Hellwig goto out_drop_priv; 1087ec18ab9SChristoph Hellwig } 109c6730ed4SJeremy Kerr 1107ec18ab9SChristoph Hellwig if (!(status & SPU_STATUS_ISOLATED_STATE)) { 111c6730ed4SJeremy Kerr /* This isn't allowed by the CBEA, but check anyway */ 112c6730ed4SJeremy Kerr pr_debug("%s: SPU fell out of isolated mode?\n", __FUNCTION__); 113c6730ed4SJeremy Kerr ctx->ops->runcntl_write(ctx, SPU_RUNCNTL_STOP); 114c6730ed4SJeremy Kerr ret = -EINVAL; 1157ec18ab9SChristoph Hellwig goto out_drop_priv; 116c6730ed4SJeremy Kerr } 117c6730ed4SJeremy Kerr 118c6730ed4SJeremy Kerr out_drop_priv: 119c6730ed4SJeremy Kerr /* Finished accessing the loader. Drop kernel mode */ 120c6730ed4SJeremy Kerr sr1 |= MFC_STATE1_PROBLEM_STATE_MASK; 121c6730ed4SJeremy Kerr spu_mfc_sr1_set(ctx->spu, sr1); 122c6730ed4SJeremy Kerr 123c6730ed4SJeremy Kerr out: 124c6730ed4SJeremy Kerr return ret; 125c6730ed4SJeremy Kerr } 126c6730ed4SJeremy Kerr 127aa45e256SChristoph Hellwig static int spu_run_init(struct spu_context *ctx, u32 * npc) 128ce8ab854SArnd Bergmann { 129aa45e256SChristoph Hellwig if (ctx->flags & SPU_CREATE_ISOLATE) { 130aa45e256SChristoph Hellwig unsigned long runcntl; 131ce8ab854SArnd Bergmann 132aa45e256SChristoph Hellwig if (!(ctx->ops->status_read(ctx) & SPU_STATUS_ISOLATED_STATE)) { 133aa45e256SChristoph Hellwig int ret = spu_setup_isolated(ctx); 134c6730ed4SJeremy Kerr if (ret) 135ce8ab854SArnd Bergmann return ret; 1360afacde3Sarnd@arndb.de } 137c6730ed4SJeremy Kerr 138c6730ed4SJeremy Kerr /* if userspace has set the runcntrl register (eg, to issue an 139c6730ed4SJeremy Kerr * isolated exit), we need to re-set it here */ 140c6730ed4SJeremy Kerr runcntl = ctx->ops->runcntl_read(ctx) & 141c6730ed4SJeremy Kerr (SPU_RUNCNTL_RUNNABLE | SPU_RUNCNTL_ISOLATE); 142c6730ed4SJeremy Kerr if (runcntl == 0) 143c6730ed4SJeremy Kerr runcntl = SPU_RUNCNTL_RUNNABLE; 144aa45e256SChristoph Hellwig ctx->ops->runcntl_write(ctx, runcntl); 1452eb1b120SChristoph Hellwig } else { 14605169237SBenjamin Herrenschmidt unsigned long mode = SPU_PRIVCNTL_MODE_NORMAL; 1472eb1b120SChristoph Hellwig spu_start_tick(ctx); 148c6730ed4SJeremy Kerr ctx->ops->npc_write(ctx, *npc); 14905169237SBenjamin Herrenschmidt if (test_thread_flag(TIF_SINGLESTEP)) 15005169237SBenjamin Herrenschmidt mode = SPU_PRIVCNTL_MODE_SINGLE_STEP; 15105169237SBenjamin Herrenschmidt out_be64(&ctx->spu->priv2->spu_privcntl_RW, mode); 152aa45e256SChristoph Hellwig ctx->ops->runcntl_write(ctx, SPU_RUNCNTL_RUNNABLE); 1532eb1b120SChristoph Hellwig } 154c6730ed4SJeremy Kerr 155aa45e256SChristoph Hellwig return 0; 156ce8ab854SArnd Bergmann } 157ce8ab854SArnd Bergmann 158aa45e256SChristoph Hellwig static int spu_run_fini(struct spu_context *ctx, u32 * npc, 159ce8ab854SArnd Bergmann u32 * status) 160ce8ab854SArnd Bergmann { 161ce8ab854SArnd Bergmann int ret = 0; 162ce8ab854SArnd Bergmann 1632eb1b120SChristoph Hellwig spu_stop_tick(ctx); 164ce8ab854SArnd Bergmann *status = ctx->ops->status_read(ctx); 165ce8ab854SArnd Bergmann *npc = ctx->ops->npc_read(ctx); 166ce8ab854SArnd Bergmann spu_release(ctx); 167ce8ab854SArnd Bergmann 168ce8ab854SArnd Bergmann if (signal_pending(current)) 169ce8ab854SArnd Bergmann ret = -ERESTARTSYS; 1702ebb2477SMasato Noguchi 171ce8ab854SArnd Bergmann return ret; 172ce8ab854SArnd Bergmann } 173ce8ab854SArnd Bergmann 174aa45e256SChristoph Hellwig static int spu_reacquire_runnable(struct spu_context *ctx, u32 *npc, 175ce8ab854SArnd Bergmann u32 *status) 176ce8ab854SArnd Bergmann { 177ce8ab854SArnd Bergmann int ret; 178ce8ab854SArnd Bergmann 179aa45e256SChristoph Hellwig ret = spu_run_fini(ctx, npc, status); 180aa45e256SChristoph Hellwig if (ret) 181ce8ab854SArnd Bergmann return ret; 182aa45e256SChristoph Hellwig 183aa45e256SChristoph Hellwig if (*status & (SPU_STATUS_STOPPED_BY_STOP | SPU_STATUS_STOPPED_BY_HALT)) 184ce8ab854SArnd Bergmann return *status; 185aa45e256SChristoph Hellwig 186aa45e256SChristoph Hellwig ret = spu_acquire_runnable(ctx, 0); 187aa45e256SChristoph Hellwig if (ret) 188ce8ab854SArnd Bergmann return ret; 189aa45e256SChristoph Hellwig 190aa45e256SChristoph Hellwig ret = spu_run_init(ctx, npc); 191aa45e256SChristoph Hellwig if (ret) { 192aa45e256SChristoph Hellwig spu_release(ctx); 193aa45e256SChristoph Hellwig return ret; 194aa45e256SChristoph Hellwig } 195ce8ab854SArnd Bergmann return 0; 196ce8ab854SArnd Bergmann } 197ce8ab854SArnd Bergmann 1982dd14934SArnd Bergmann /* 1992dd14934SArnd Bergmann * SPU syscall restarting is tricky because we violate the basic 2002dd14934SArnd Bergmann * assumption that the signal handler is running on the interrupted 2012dd14934SArnd Bergmann * thread. Here instead, the handler runs on PowerPC user space code, 2022dd14934SArnd Bergmann * while the syscall was called from the SPU. 2032dd14934SArnd Bergmann * This means we can only do a very rough approximation of POSIX 2042dd14934SArnd Bergmann * signal semantics. 2052dd14934SArnd Bergmann */ 2062dd14934SArnd Bergmann int spu_handle_restartsys(struct spu_context *ctx, long *spu_ret, 2072dd14934SArnd Bergmann unsigned int *npc) 2082dd14934SArnd Bergmann { 2092dd14934SArnd Bergmann int ret; 2102dd14934SArnd Bergmann 2112dd14934SArnd Bergmann switch (*spu_ret) { 2122dd14934SArnd Bergmann case -ERESTARTSYS: 2132dd14934SArnd Bergmann case -ERESTARTNOINTR: 2142dd14934SArnd Bergmann /* 2152dd14934SArnd Bergmann * Enter the regular syscall restarting for 2162dd14934SArnd Bergmann * sys_spu_run, then restart the SPU syscall 2172dd14934SArnd Bergmann * callback. 2182dd14934SArnd Bergmann */ 2192dd14934SArnd Bergmann *npc -= 8; 2202dd14934SArnd Bergmann ret = -ERESTARTSYS; 2212dd14934SArnd Bergmann break; 2222dd14934SArnd Bergmann case -ERESTARTNOHAND: 2232dd14934SArnd Bergmann case -ERESTART_RESTARTBLOCK: 2242dd14934SArnd Bergmann /* 2252dd14934SArnd Bergmann * Restart block is too hard for now, just return -EINTR 2262dd14934SArnd Bergmann * to the SPU. 2272dd14934SArnd Bergmann * ERESTARTNOHAND comes from sys_pause, we also return 2282dd14934SArnd Bergmann * -EINTR from there. 2292dd14934SArnd Bergmann * Assume that we need to be restarted ourselves though. 2302dd14934SArnd Bergmann */ 2312dd14934SArnd Bergmann *spu_ret = -EINTR; 2322dd14934SArnd Bergmann ret = -ERESTARTSYS; 2332dd14934SArnd Bergmann break; 2342dd14934SArnd Bergmann default: 2352dd14934SArnd Bergmann printk(KERN_WARNING "%s: unexpected return code %ld\n", 2362dd14934SArnd Bergmann __FUNCTION__, *spu_ret); 2372dd14934SArnd Bergmann ret = 0; 2382dd14934SArnd Bergmann } 2392dd14934SArnd Bergmann return ret; 2402dd14934SArnd Bergmann } 2412dd14934SArnd Bergmann 2422dd14934SArnd Bergmann int spu_process_callback(struct spu_context *ctx) 2432dd14934SArnd Bergmann { 2442dd14934SArnd Bergmann struct spu_syscall_block s; 2452dd14934SArnd Bergmann u32 ls_pointer, npc; 2469e2fe2ceSAkinobu Mita void __iomem *ls; 2472dd14934SArnd Bergmann long spu_ret; 2482dd14934SArnd Bergmann int ret; 2492dd14934SArnd Bergmann 2502dd14934SArnd Bergmann /* get syscall block from local store */ 2519e2fe2ceSAkinobu Mita npc = ctx->ops->npc_read(ctx) & ~3; 2529e2fe2ceSAkinobu Mita ls = (void __iomem *)ctx->ops->get_ls(ctx); 2539e2fe2ceSAkinobu Mita ls_pointer = in_be32(ls + npc); 2542dd14934SArnd Bergmann if (ls_pointer > (LS_SIZE - sizeof(s))) 2552dd14934SArnd Bergmann return -EFAULT; 2569e2fe2ceSAkinobu Mita memcpy_fromio(&s, ls + ls_pointer, sizeof(s)); 2572dd14934SArnd Bergmann 2582dd14934SArnd Bergmann /* do actual syscall without pinning the spu */ 2592dd14934SArnd Bergmann ret = 0; 2602dd14934SArnd Bergmann spu_ret = -ENOSYS; 2612dd14934SArnd Bergmann npc += 4; 2622dd14934SArnd Bergmann 2632dd14934SArnd Bergmann if (s.nr_ret < __NR_syscalls) { 2642dd14934SArnd Bergmann spu_release(ctx); 2652dd14934SArnd Bergmann /* do actual system call from here */ 2662dd14934SArnd Bergmann spu_ret = spu_sys_callback(&s); 2672dd14934SArnd Bergmann if (spu_ret <= -ERESTARTSYS) { 2682dd14934SArnd Bergmann ret = spu_handle_restartsys(ctx, &spu_ret, &npc); 2692dd14934SArnd Bergmann } 2702dd14934SArnd Bergmann spu_acquire(ctx); 2712dd14934SArnd Bergmann if (ret == -ERESTARTSYS) 2722dd14934SArnd Bergmann return ret; 2732dd14934SArnd Bergmann } 2742dd14934SArnd Bergmann 2752dd14934SArnd Bergmann /* write result, jump over indirect pointer */ 2769e2fe2ceSAkinobu Mita memcpy_toio(ls + ls_pointer, &spu_ret, sizeof(spu_ret)); 2772dd14934SArnd Bergmann ctx->ops->npc_write(ctx, npc); 2782dd14934SArnd Bergmann ctx->ops->runcntl_write(ctx, SPU_RUNCNTL_RUNNABLE); 2792dd14934SArnd Bergmann return ret; 2802dd14934SArnd Bergmann } 2812dd14934SArnd Bergmann 282ce8ab854SArnd Bergmann static inline int spu_process_events(struct spu_context *ctx) 283ce8ab854SArnd Bergmann { 284ce8ab854SArnd Bergmann struct spu *spu = ctx->spu; 285ce8ab854SArnd Bergmann int ret = 0; 286ce8ab854SArnd Bergmann 287ce8ab854SArnd Bergmann if (spu->class_0_pending) 288ce8ab854SArnd Bergmann ret = spu_irq_class_0_bottom(spu); 289ce8ab854SArnd Bergmann if (!ret && signal_pending(current)) 290ce8ab854SArnd Bergmann ret = -ERESTARTSYS; 291ce8ab854SArnd Bergmann return ret; 292ce8ab854SArnd Bergmann } 293ce8ab854SArnd Bergmann 294ce8ab854SArnd Bergmann long spufs_run_spu(struct file *file, struct spu_context *ctx, 2959add11daSArnd Bergmann u32 *npc, u32 *event) 296ce8ab854SArnd Bergmann { 297ce8ab854SArnd Bergmann int ret; 2989add11daSArnd Bergmann u32 status; 299ce8ab854SArnd Bergmann 300e45d48a3SChristoph Hellwig if (mutex_lock_interruptible(&ctx->run_mutex)) 301ce8ab854SArnd Bergmann return -ERESTARTSYS; 302ce8ab854SArnd Bergmann 303ee2d7340SArnd Bergmann ctx->ops->master_start(ctx); 3049add11daSArnd Bergmann ctx->event_return = 0; 305aa45e256SChristoph Hellwig 306aa45e256SChristoph Hellwig ret = spu_acquire_runnable(ctx, 0); 307ce8ab854SArnd Bergmann if (ret) 308aa45e256SChristoph Hellwig return ret; 309aa45e256SChristoph Hellwig 310aa45e256SChristoph Hellwig ret = spu_run_init(ctx, npc); 311aa45e256SChristoph Hellwig if (ret) { 312aa45e256SChristoph Hellwig spu_release(ctx); 313ce8ab854SArnd Bergmann goto out; 314aa45e256SChristoph Hellwig } 315ce8ab854SArnd Bergmann 316ce8ab854SArnd Bergmann do { 3179add11daSArnd Bergmann ret = spufs_wait(ctx->stop_wq, spu_stopped(ctx, &status)); 318ce8ab854SArnd Bergmann if (unlikely(ret)) 319ce8ab854SArnd Bergmann break; 3209add11daSArnd Bergmann if ((status & SPU_STATUS_STOPPED_BY_STOP) && 3219add11daSArnd Bergmann (status >> SPU_STOP_STATUS_SHIFT == 0x2104)) { 3222dd14934SArnd Bergmann ret = spu_process_callback(ctx); 3232dd14934SArnd Bergmann if (ret) 3242dd14934SArnd Bergmann break; 3259add11daSArnd Bergmann status &= ~SPU_STATUS_STOPPED_BY_STOP; 3262dd14934SArnd Bergmann } 32757dace23SArnd Bergmann ret = spufs_handle_class1(ctx); 32857dace23SArnd Bergmann if (ret) 32957dace23SArnd Bergmann break; 33057dace23SArnd Bergmann 331ce8ab854SArnd Bergmann if (unlikely(ctx->state != SPU_STATE_RUNNABLE)) { 3329add11daSArnd Bergmann ret = spu_reacquire_runnable(ctx, npc, &status); 3332eb1b120SChristoph Hellwig if (ret) { 3342eb1b120SChristoph Hellwig spu_stop_tick(ctx); 3352ebb2477SMasato Noguchi goto out2; 3362eb1b120SChristoph Hellwig } 337ce8ab854SArnd Bergmann continue; 338ce8ab854SArnd Bergmann } 339ce8ab854SArnd Bergmann ret = spu_process_events(ctx); 340ce8ab854SArnd Bergmann 3419add11daSArnd Bergmann } while (!ret && !(status & (SPU_STATUS_STOPPED_BY_STOP | 34205169237SBenjamin Herrenschmidt SPU_STATUS_STOPPED_BY_HALT | 34305169237SBenjamin Herrenschmidt SPU_STATUS_SINGLE_STEP))); 344ce8ab854SArnd Bergmann 345ee2d7340SArnd Bergmann ctx->ops->master_stop(ctx); 3469add11daSArnd Bergmann ret = spu_run_fini(ctx, npc, &status); 347ce8ab854SArnd Bergmann spu_yield(ctx); 348ce8ab854SArnd Bergmann 3492ebb2477SMasato Noguchi out2: 3502ebb2477SMasato Noguchi if ((ret == 0) || 3512ebb2477SMasato Noguchi ((ret == -ERESTARTSYS) && 3522ebb2477SMasato Noguchi ((status & SPU_STATUS_STOPPED_BY_HALT) || 35305169237SBenjamin Herrenschmidt (status & SPU_STATUS_SINGLE_STEP) || 3542ebb2477SMasato Noguchi ((status & SPU_STATUS_STOPPED_BY_STOP) && 3552ebb2477SMasato Noguchi (status >> SPU_STOP_STATUS_SHIFT != 0x2104))))) 3562ebb2477SMasato Noguchi ret = status; 3572ebb2477SMasato Noguchi 35805169237SBenjamin Herrenschmidt /* Note: we don't need to force_sig SIGTRAP on single-step 35905169237SBenjamin Herrenschmidt * since we have TIF_SINGLESTEP set, thus the kernel will do 36005169237SBenjamin Herrenschmidt * it upon return from the syscall anyawy 36105169237SBenjamin Herrenschmidt */ 3622ebb2477SMasato Noguchi if ((status & SPU_STATUS_STOPPED_BY_STOP) 3632ebb2477SMasato Noguchi && (status >> SPU_STOP_STATUS_SHIFT) == 0x3fff) { 3642ebb2477SMasato Noguchi force_sig(SIGTRAP, current); 3652ebb2477SMasato Noguchi ret = -ERESTARTSYS; 3662ebb2477SMasato Noguchi } 3672ebb2477SMasato Noguchi 368ce8ab854SArnd Bergmann out: 3699add11daSArnd Bergmann *event = ctx->event_return; 370e45d48a3SChristoph Hellwig mutex_unlock(&ctx->run_mutex); 371ce8ab854SArnd Bergmann return ret; 372ce8ab854SArnd Bergmann } 373