15473af04SMark Nutter /* 25473af04SMark Nutter * spu_switch.c 35473af04SMark Nutter * 45473af04SMark Nutter * (C) Copyright IBM Corp. 2005 55473af04SMark Nutter * 65473af04SMark Nutter * Author: Mark Nutter <mnutter@us.ibm.com> 75473af04SMark Nutter * 85473af04SMark Nutter * Host-side part of SPU context switch sequence outlined in 95473af04SMark Nutter * Synergistic Processor Element, Book IV. 105473af04SMark Nutter * 115473af04SMark Nutter * A fully premptive switch of an SPE is very expensive in terms 125473af04SMark Nutter * of time and system resources. SPE Book IV indicates that SPE 135473af04SMark Nutter * allocation should follow a "serially reusable device" model, 145473af04SMark Nutter * in which the SPE is assigned a task until it completes. When 155473af04SMark Nutter * this is not possible, this sequence may be used to premptively 165473af04SMark Nutter * save, and then later (optionally) restore the context of a 175473af04SMark Nutter * program executing on an SPE. 185473af04SMark Nutter * 195473af04SMark Nutter * 205473af04SMark Nutter * This program is free software; you can redistribute it and/or modify 215473af04SMark Nutter * it under the terms of the GNU General Public License as published by 225473af04SMark Nutter * the Free Software Foundation; either version 2, or (at your option) 235473af04SMark Nutter * any later version. 245473af04SMark Nutter * 255473af04SMark Nutter * This program is distributed in the hope that it will be useful, 265473af04SMark Nutter * but WITHOUT ANY WARRANTY; without even the implied warranty of 275473af04SMark Nutter * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 285473af04SMark Nutter * GNU General Public License for more details. 295473af04SMark Nutter * 305473af04SMark Nutter * You should have received a copy of the GNU General Public License 315473af04SMark Nutter * along with this program; if not, write to the Free Software 325473af04SMark Nutter * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA. 335473af04SMark Nutter */ 345473af04SMark Nutter 355473af04SMark Nutter #include <linux/module.h> 365473af04SMark Nutter #include <linux/errno.h> 375473af04SMark Nutter #include <linux/sched.h> 385473af04SMark Nutter #include <linux/kernel.h> 395473af04SMark Nutter #include <linux/mm.h> 405473af04SMark Nutter #include <linux/vmalloc.h> 415473af04SMark Nutter #include <linux/smp.h> 425473af04SMark Nutter #include <linux/stddef.h> 435473af04SMark Nutter #include <linux/unistd.h> 445473af04SMark Nutter 455473af04SMark Nutter #include <asm/io.h> 465473af04SMark Nutter #include <asm/spu.h> 47540270d8SGeoff Levand #include <asm/spu_priv1.h> 485473af04SMark Nutter #include <asm/spu_csa.h> 495473af04SMark Nutter #include <asm/mmu_context.h> 505473af04SMark Nutter 515473af04SMark Nutter #include "spu_save_dump.h" 525473af04SMark Nutter #include "spu_restore_dump.h" 535473af04SMark Nutter 547c038749SMark Nutter #if 0 557c038749SMark Nutter #define POLL_WHILE_TRUE(_c) { \ 567c038749SMark Nutter do { \ 577c038749SMark Nutter } while (_c); \ 587c038749SMark Nutter } 597c038749SMark Nutter #else 607c038749SMark Nutter #define RELAX_SPIN_COUNT 1000 617c038749SMark Nutter #define POLL_WHILE_TRUE(_c) { \ 627c038749SMark Nutter do { \ 637c038749SMark Nutter int _i; \ 647c038749SMark Nutter for (_i=0; _i<RELAX_SPIN_COUNT && (_c); _i++) { \ 657c038749SMark Nutter cpu_relax(); \ 667c038749SMark Nutter } \ 677c038749SMark Nutter if (unlikely(_c)) yield(); \ 687c038749SMark Nutter else break; \ 697c038749SMark Nutter } while (_c); \ 707c038749SMark Nutter } 717c038749SMark Nutter #endif /* debug */ 727c038749SMark Nutter 737c038749SMark Nutter #define POLL_WHILE_FALSE(_c) POLL_WHILE_TRUE(!(_c)) 747c038749SMark Nutter 757c038749SMark Nutter static inline void acquire_spu_lock(struct spu *spu) 767c038749SMark Nutter { 777c038749SMark Nutter /* Save, Step 1: 787c038749SMark Nutter * Restore, Step 1: 797c038749SMark Nutter * Acquire SPU-specific mutual exclusion lock. 807c038749SMark Nutter * TBD. 817c038749SMark Nutter */ 827c038749SMark Nutter } 837c038749SMark Nutter 847c038749SMark Nutter static inline void release_spu_lock(struct spu *spu) 857c038749SMark Nutter { 867c038749SMark Nutter /* Restore, Step 76: 877c038749SMark Nutter * Release SPU-specific mutual exclusion lock. 887c038749SMark Nutter * TBD. 897c038749SMark Nutter */ 907c038749SMark Nutter } 917c038749SMark Nutter 927c038749SMark Nutter static inline int check_spu_isolate(struct spu_state *csa, struct spu *spu) 937c038749SMark Nutter { 947c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 957c038749SMark Nutter u32 isolate_state; 967c038749SMark Nutter 977c038749SMark Nutter /* Save, Step 2: 987c038749SMark Nutter * Save, Step 6: 997c038749SMark Nutter * If SPU_Status[E,L,IS] any field is '1', this 1007c038749SMark Nutter * SPU is in isolate state and cannot be context 1017c038749SMark Nutter * saved at this time. 1027c038749SMark Nutter */ 1037c038749SMark Nutter isolate_state = SPU_STATUS_ISOLATED_STATE | 104eb758ce5Sarnd@arndb.de SPU_STATUS_ISOLATED_LOAD_STATUS | SPU_STATUS_ISOLATED_EXIT_STATUS; 1057c038749SMark Nutter return (in_be32(&prob->spu_status_R) & isolate_state) ? 1 : 0; 1067c038749SMark Nutter } 1077c038749SMark Nutter 1087c038749SMark Nutter static inline void disable_interrupts(struct spu_state *csa, struct spu *spu) 1097c038749SMark Nutter { 1107c038749SMark Nutter /* Save, Step 3: 1117c038749SMark Nutter * Restore, Step 2: 1127c038749SMark Nutter * Save INT_Mask_class0 in CSA. 1137c038749SMark Nutter * Write INT_MASK_class0 with value of 0. 1147c038749SMark Nutter * Save INT_Mask_class1 in CSA. 1157c038749SMark Nutter * Write INT_MASK_class1 with value of 0. 1167c038749SMark Nutter * Save INT_Mask_class2 in CSA. 1177c038749SMark Nutter * Write INT_MASK_class2 with value of 0. 1187c038749SMark Nutter */ 1197c038749SMark Nutter spin_lock_irq(&spu->register_lock); 1207c038749SMark Nutter if (csa) { 121f0831accSArnd Bergmann csa->priv1.int_mask_class0_RW = spu_int_mask_get(spu, 0); 122f0831accSArnd Bergmann csa->priv1.int_mask_class1_RW = spu_int_mask_get(spu, 1); 123f0831accSArnd Bergmann csa->priv1.int_mask_class2_RW = spu_int_mask_get(spu, 2); 1247c038749SMark Nutter } 125f0831accSArnd Bergmann spu_int_mask_set(spu, 0, 0ul); 126f0831accSArnd Bergmann spu_int_mask_set(spu, 1, 0ul); 127f0831accSArnd Bergmann spu_int_mask_set(spu, 2, 0ul); 1287c038749SMark Nutter eieio(); 1297c038749SMark Nutter spin_unlock_irq(&spu->register_lock); 1307c038749SMark Nutter } 1317c038749SMark Nutter 1327c038749SMark Nutter static inline void set_watchdog_timer(struct spu_state *csa, struct spu *spu) 1337c038749SMark Nutter { 1347c038749SMark Nutter /* Save, Step 4: 1357c038749SMark Nutter * Restore, Step 25. 1367c038749SMark Nutter * Set a software watchdog timer, which specifies the 1377c038749SMark Nutter * maximum allowable time for a context save sequence. 1387c038749SMark Nutter * 1397c038749SMark Nutter * For present, this implementation will not set a global 1407c038749SMark Nutter * watchdog timer, as virtualization & variable system load 1417c038749SMark Nutter * may cause unpredictable execution times. 1427c038749SMark Nutter */ 1437c038749SMark Nutter } 1447c038749SMark Nutter 1457c038749SMark Nutter static inline void inhibit_user_access(struct spu_state *csa, struct spu *spu) 1467c038749SMark Nutter { 1477c038749SMark Nutter /* Save, Step 5: 1487c038749SMark Nutter * Restore, Step 3: 1497c038749SMark Nutter * Inhibit user-space access (if provided) to this 1507c038749SMark Nutter * SPU by unmapping the virtual pages assigned to 1517c038749SMark Nutter * the SPU memory-mapped I/O (MMIO) for problem 1527c038749SMark Nutter * state. TBD. 1537c038749SMark Nutter */ 1547c038749SMark Nutter } 1557c038749SMark Nutter 1567c038749SMark Nutter static inline void set_switch_pending(struct spu_state *csa, struct spu *spu) 1577c038749SMark Nutter { 1587c038749SMark Nutter /* Save, Step 7: 1597c038749SMark Nutter * Restore, Step 5: 1607c038749SMark Nutter * Set a software context switch pending flag. 1617c038749SMark Nutter */ 1628837d921SArnd Bergmann set_bit(SPU_CONTEXT_SWITCH_PENDING, &spu->flags); 1637c038749SMark Nutter mb(); 1647c038749SMark Nutter } 1657c038749SMark Nutter 1667c038749SMark Nutter static inline void save_mfc_cntl(struct spu_state *csa, struct spu *spu) 1677c038749SMark Nutter { 1687c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 1697c038749SMark Nutter 1707c038749SMark Nutter /* Save, Step 8: 1717f52eb00SGeoff Levand * Suspend DMA and save MFC_CNTL. 1727c038749SMark Nutter */ 1737f52eb00SGeoff Levand switch (in_be64(&priv2->mfc_control_RW) & 1747f52eb00SGeoff Levand MFC_CNTL_SUSPEND_DMA_STATUS_MASK) { 1757f52eb00SGeoff Levand case MFC_CNTL_SUSPEND_IN_PROGRESS: 1767f52eb00SGeoff Levand POLL_WHILE_FALSE((in_be64(&priv2->mfc_control_RW) & 1777f52eb00SGeoff Levand MFC_CNTL_SUSPEND_DMA_STATUS_MASK) == 1787f52eb00SGeoff Levand MFC_CNTL_SUSPEND_COMPLETE); 1797f52eb00SGeoff Levand /* fall through */ 1807f52eb00SGeoff Levand case MFC_CNTL_SUSPEND_COMPLETE: 1817c038749SMark Nutter if (csa) { 1827f52eb00SGeoff Levand csa->priv2.mfc_control_RW = 1837f52eb00SGeoff Levand in_be64(&priv2->mfc_control_RW) | 1847f52eb00SGeoff Levand MFC_CNTL_SUSPEND_DMA_QUEUE; 1857f52eb00SGeoff Levand } 1867f52eb00SGeoff Levand break; 1877f52eb00SGeoff Levand case MFC_CNTL_NORMAL_DMA_QUEUE_OPERATION: 1887f52eb00SGeoff Levand out_be64(&priv2->mfc_control_RW, MFC_CNTL_SUSPEND_DMA_QUEUE); 1897f52eb00SGeoff Levand POLL_WHILE_FALSE((in_be64(&priv2->mfc_control_RW) & 1907f52eb00SGeoff Levand MFC_CNTL_SUSPEND_DMA_STATUS_MASK) == 1917f52eb00SGeoff Levand MFC_CNTL_SUSPEND_COMPLETE); 1927f52eb00SGeoff Levand if (csa) { 1937f52eb00SGeoff Levand csa->priv2.mfc_control_RW = 1947f52eb00SGeoff Levand in_be64(&priv2->mfc_control_RW) & 1957f52eb00SGeoff Levand ~MFC_CNTL_SUSPEND_DMA_QUEUE; 1967f52eb00SGeoff Levand } 1977f52eb00SGeoff Levand break; 1987c038749SMark Nutter } 1997c038749SMark Nutter } 2007c038749SMark Nutter 2017c038749SMark Nutter static inline void save_spu_runcntl(struct spu_state *csa, struct spu *spu) 2027c038749SMark Nutter { 2037c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 2047c038749SMark Nutter 2057c038749SMark Nutter /* Save, Step 9: 2067c038749SMark Nutter * Save SPU_Runcntl in the CSA. This value contains 2077c038749SMark Nutter * the "Application Desired State". 2087c038749SMark Nutter */ 2097c038749SMark Nutter csa->prob.spu_runcntl_RW = in_be32(&prob->spu_runcntl_RW); 2107c038749SMark Nutter } 2117c038749SMark Nutter 2127c038749SMark Nutter static inline void save_mfc_sr1(struct spu_state *csa, struct spu *spu) 2137c038749SMark Nutter { 2147c038749SMark Nutter /* Save, Step 10: 2157c038749SMark Nutter * Save MFC_SR1 in the CSA. 2167c038749SMark Nutter */ 217f0831accSArnd Bergmann csa->priv1.mfc_sr1_RW = spu_mfc_sr1_get(spu); 2187c038749SMark Nutter } 2197c038749SMark Nutter 2207c038749SMark Nutter static inline void save_spu_status(struct spu_state *csa, struct spu *spu) 2217c038749SMark Nutter { 2227c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 2237c038749SMark Nutter 2247c038749SMark Nutter /* Save, Step 11: 2257c038749SMark Nutter * Read SPU_Status[R], and save to CSA. 2267c038749SMark Nutter */ 2277c038749SMark Nutter if ((in_be32(&prob->spu_status_R) & SPU_STATUS_RUNNING) == 0) { 2287c038749SMark Nutter csa->prob.spu_status_R = in_be32(&prob->spu_status_R); 2297c038749SMark Nutter } else { 2307c038749SMark Nutter u32 stopped; 2317c038749SMark Nutter 2327c038749SMark Nutter out_be32(&prob->spu_runcntl_RW, SPU_RUNCNTL_STOP); 2337c038749SMark Nutter eieio(); 2347c038749SMark Nutter POLL_WHILE_TRUE(in_be32(&prob->spu_status_R) & 2357c038749SMark Nutter SPU_STATUS_RUNNING); 2367c038749SMark Nutter stopped = 2377c038749SMark Nutter SPU_STATUS_INVALID_INSTR | SPU_STATUS_SINGLE_STEP | 2387c038749SMark Nutter SPU_STATUS_STOPPED_BY_HALT | SPU_STATUS_STOPPED_BY_STOP; 2397c038749SMark Nutter if ((in_be32(&prob->spu_status_R) & stopped) == 0) 2407c038749SMark Nutter csa->prob.spu_status_R = SPU_STATUS_RUNNING; 2417c038749SMark Nutter else 2427c038749SMark Nutter csa->prob.spu_status_R = in_be32(&prob->spu_status_R); 2437c038749SMark Nutter } 2447c038749SMark Nutter } 2457c038749SMark Nutter 2467c038749SMark Nutter static inline void save_mfc_decr(struct spu_state *csa, struct spu *spu) 2477c038749SMark Nutter { 2487c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 2497c038749SMark Nutter 2507c038749SMark Nutter /* Save, Step 12: 2517c038749SMark Nutter * Read MFC_CNTL[Ds]. Update saved copy of 2527c038749SMark Nutter * CSA.MFC_CNTL[Ds]. 2537c038749SMark Nutter */ 2547c038749SMark Nutter if (in_be64(&priv2->mfc_control_RW) & MFC_CNTL_DECREMENTER_RUNNING) { 2557c038749SMark Nutter csa->priv2.mfc_control_RW |= MFC_CNTL_DECREMENTER_RUNNING; 2567c038749SMark Nutter csa->suspend_time = get_cycles(); 2577c038749SMark Nutter out_be64(&priv2->spu_chnlcntptr_RW, 7ULL); 2587c038749SMark Nutter eieio(); 2597c038749SMark Nutter csa->spu_chnldata_RW[7] = in_be64(&priv2->spu_chnldata_RW); 2607c038749SMark Nutter eieio(); 2617f52eb00SGeoff Levand } else { 2627f52eb00SGeoff Levand csa->priv2.mfc_control_RW &= ~MFC_CNTL_DECREMENTER_RUNNING; 2637c038749SMark Nutter } 2647c038749SMark Nutter } 2657c038749SMark Nutter 2667c038749SMark Nutter static inline void halt_mfc_decr(struct spu_state *csa, struct spu *spu) 2677c038749SMark Nutter { 2687c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 2697c038749SMark Nutter 2707c038749SMark Nutter /* Save, Step 13: 2717c038749SMark Nutter * Write MFC_CNTL[Dh] set to a '1' to halt 2727c038749SMark Nutter * the decrementer. 2737c038749SMark Nutter */ 2747c038749SMark Nutter out_be64(&priv2->mfc_control_RW, MFC_CNTL_DECREMENTER_HALTED); 2757c038749SMark Nutter eieio(); 2767c038749SMark Nutter } 2777c038749SMark Nutter 2787c038749SMark Nutter static inline void save_timebase(struct spu_state *csa, struct spu *spu) 2797c038749SMark Nutter { 2807c038749SMark Nutter /* Save, Step 14: 2817c038749SMark Nutter * Read PPE Timebase High and Timebase low registers 2827c038749SMark Nutter * and save in CSA. TBD. 2837c038749SMark Nutter */ 2847c038749SMark Nutter csa->suspend_time = get_cycles(); 2857c038749SMark Nutter } 2867c038749SMark Nutter 2877c038749SMark Nutter static inline void remove_other_spu_access(struct spu_state *csa, 2887c038749SMark Nutter struct spu *spu) 2897c038749SMark Nutter { 2907c038749SMark Nutter /* Save, Step 15: 2917c038749SMark Nutter * Remove other SPU access to this SPU by unmapping 2927c038749SMark Nutter * this SPU's pages from their address space. TBD. 2937c038749SMark Nutter */ 2947c038749SMark Nutter } 2957c038749SMark Nutter 2967c038749SMark Nutter static inline void do_mfc_mssync(struct spu_state *csa, struct spu *spu) 2977c038749SMark Nutter { 2987c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 2997c038749SMark Nutter 3007c038749SMark Nutter /* Save, Step 16: 3017c038749SMark Nutter * Restore, Step 11. 3027c038749SMark Nutter * Write SPU_MSSync register. Poll SPU_MSSync[P] 3037c038749SMark Nutter * for a value of 0. 3047c038749SMark Nutter */ 3057c038749SMark Nutter out_be64(&prob->spc_mssync_RW, 1UL); 3067c038749SMark Nutter POLL_WHILE_TRUE(in_be64(&prob->spc_mssync_RW) & MS_SYNC_PENDING); 3077c038749SMark Nutter } 3087c038749SMark Nutter 3097c038749SMark Nutter static inline void issue_mfc_tlbie(struct spu_state *csa, struct spu *spu) 3107c038749SMark Nutter { 3117c038749SMark Nutter /* Save, Step 17: 3127c038749SMark Nutter * Restore, Step 12. 3137c038749SMark Nutter * Restore, Step 48. 3147c038749SMark Nutter * Write TLB_Invalidate_Entry[IS,VPN,L,Lp]=0 register. 3157c038749SMark Nutter * Then issue a PPE sync instruction. 3167c038749SMark Nutter */ 317f0831accSArnd Bergmann spu_tlb_invalidate(spu); 3187c038749SMark Nutter mb(); 3197c038749SMark Nutter } 3207c038749SMark Nutter 3217c038749SMark Nutter static inline void handle_pending_interrupts(struct spu_state *csa, 3227c038749SMark Nutter struct spu *spu) 3237c038749SMark Nutter { 3247c038749SMark Nutter /* Save, Step 18: 3257c038749SMark Nutter * Handle any pending interrupts from this SPU 3267c038749SMark Nutter * here. This is OS or hypervisor specific. One 3277c038749SMark Nutter * option is to re-enable interrupts to handle any 3287c038749SMark Nutter * pending interrupts, with the interrupt handlers 3297c038749SMark Nutter * recognizing the software Context Switch Pending 3307c038749SMark Nutter * flag, to ensure the SPU execution or MFC command 3317c038749SMark Nutter * queue is not restarted. TBD. 3327c038749SMark Nutter */ 3337c038749SMark Nutter } 3347c038749SMark Nutter 3357c038749SMark Nutter static inline void save_mfc_queues(struct spu_state *csa, struct spu *spu) 3367c038749SMark Nutter { 3377c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 3387c038749SMark Nutter int i; 3397c038749SMark Nutter 3407c038749SMark Nutter /* Save, Step 19: 3417c038749SMark Nutter * If MFC_Cntl[Se]=0 then save 3427c038749SMark Nutter * MFC command queues. 3437c038749SMark Nutter */ 3447c038749SMark Nutter if ((in_be64(&priv2->mfc_control_RW) & MFC_CNTL_DMA_QUEUES_EMPTY) == 0) { 3457c038749SMark Nutter for (i = 0; i < 8; i++) { 3467c038749SMark Nutter csa->priv2.puq[i].mfc_cq_data0_RW = 3477c038749SMark Nutter in_be64(&priv2->puq[i].mfc_cq_data0_RW); 3487c038749SMark Nutter csa->priv2.puq[i].mfc_cq_data1_RW = 3497c038749SMark Nutter in_be64(&priv2->puq[i].mfc_cq_data1_RW); 3507c038749SMark Nutter csa->priv2.puq[i].mfc_cq_data2_RW = 3517c038749SMark Nutter in_be64(&priv2->puq[i].mfc_cq_data2_RW); 3527c038749SMark Nutter csa->priv2.puq[i].mfc_cq_data3_RW = 3537c038749SMark Nutter in_be64(&priv2->puq[i].mfc_cq_data3_RW); 3547c038749SMark Nutter } 3557c038749SMark Nutter for (i = 0; i < 16; i++) { 3567c038749SMark Nutter csa->priv2.spuq[i].mfc_cq_data0_RW = 3577c038749SMark Nutter in_be64(&priv2->spuq[i].mfc_cq_data0_RW); 3587c038749SMark Nutter csa->priv2.spuq[i].mfc_cq_data1_RW = 3597c038749SMark Nutter in_be64(&priv2->spuq[i].mfc_cq_data1_RW); 3607c038749SMark Nutter csa->priv2.spuq[i].mfc_cq_data2_RW = 3617c038749SMark Nutter in_be64(&priv2->spuq[i].mfc_cq_data2_RW); 3627c038749SMark Nutter csa->priv2.spuq[i].mfc_cq_data3_RW = 3637c038749SMark Nutter in_be64(&priv2->spuq[i].mfc_cq_data3_RW); 3647c038749SMark Nutter } 3657c038749SMark Nutter } 3667c038749SMark Nutter } 3677c038749SMark Nutter 3687c038749SMark Nutter static inline void save_ppu_querymask(struct spu_state *csa, struct spu *spu) 3697c038749SMark Nutter { 3707c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 3717c038749SMark Nutter 3727c038749SMark Nutter /* Save, Step 20: 3737c038749SMark Nutter * Save the PPU_QueryMask register 3747c038749SMark Nutter * in the CSA. 3757c038749SMark Nutter */ 3767c038749SMark Nutter csa->prob.dma_querymask_RW = in_be32(&prob->dma_querymask_RW); 3777c038749SMark Nutter } 3787c038749SMark Nutter 3797c038749SMark Nutter static inline void save_ppu_querytype(struct spu_state *csa, struct spu *spu) 3807c038749SMark Nutter { 3817c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 3827c038749SMark Nutter 3837c038749SMark Nutter /* Save, Step 21: 3847c038749SMark Nutter * Save the PPU_QueryType register 3857c038749SMark Nutter * in the CSA. 3867c038749SMark Nutter */ 3877c038749SMark Nutter csa->prob.dma_querytype_RW = in_be32(&prob->dma_querytype_RW); 3887c038749SMark Nutter } 3897c038749SMark Nutter 3908d038e04SKazunori Asayama static inline void save_ppu_tagstatus(struct spu_state *csa, struct spu *spu) 3918d038e04SKazunori Asayama { 3928d038e04SKazunori Asayama struct spu_problem __iomem *prob = spu->problem; 3938d038e04SKazunori Asayama 3948d038e04SKazunori Asayama /* Save the Prxy_TagStatus register in the CSA. 3958d038e04SKazunori Asayama * 3968d038e04SKazunori Asayama * It is unnecessary to restore dma_tagstatus_R, however, 3978d038e04SKazunori Asayama * dma_tagstatus_R in the CSA is accessed via backing_ops, so 3988d038e04SKazunori Asayama * we must save it. 3998d038e04SKazunori Asayama */ 4008d038e04SKazunori Asayama csa->prob.dma_tagstatus_R = in_be32(&prob->dma_tagstatus_R); 4018d038e04SKazunori Asayama } 4028d038e04SKazunori Asayama 4037c038749SMark Nutter static inline void save_mfc_csr_tsq(struct spu_state *csa, struct spu *spu) 4047c038749SMark Nutter { 4057c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 4067c038749SMark Nutter 4077c038749SMark Nutter /* Save, Step 22: 4087c038749SMark Nutter * Save the MFC_CSR_TSQ register 4097c038749SMark Nutter * in the LSCSA. 4107c038749SMark Nutter */ 4117c038749SMark Nutter csa->priv2.spu_tag_status_query_RW = 4127c038749SMark Nutter in_be64(&priv2->spu_tag_status_query_RW); 4137c038749SMark Nutter } 4147c038749SMark Nutter 4157c038749SMark Nutter static inline void save_mfc_csr_cmd(struct spu_state *csa, struct spu *spu) 4167c038749SMark Nutter { 4177c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 4187c038749SMark Nutter 4197c038749SMark Nutter /* Save, Step 23: 4207c038749SMark Nutter * Save the MFC_CSR_CMD1 and MFC_CSR_CMD2 4217c038749SMark Nutter * registers in the CSA. 4227c038749SMark Nutter */ 4237c038749SMark Nutter csa->priv2.spu_cmd_buf1_RW = in_be64(&priv2->spu_cmd_buf1_RW); 4247c038749SMark Nutter csa->priv2.spu_cmd_buf2_RW = in_be64(&priv2->spu_cmd_buf2_RW); 4257c038749SMark Nutter } 4267c038749SMark Nutter 4277c038749SMark Nutter static inline void save_mfc_csr_ato(struct spu_state *csa, struct spu *spu) 4287c038749SMark Nutter { 4297c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 4307c038749SMark Nutter 4317c038749SMark Nutter /* Save, Step 24: 4327c038749SMark Nutter * Save the MFC_CSR_ATO register in 4337c038749SMark Nutter * the CSA. 4347c038749SMark Nutter */ 4357c038749SMark Nutter csa->priv2.spu_atomic_status_RW = in_be64(&priv2->spu_atomic_status_RW); 4367c038749SMark Nutter } 4377c038749SMark Nutter 4387c038749SMark Nutter static inline void save_mfc_tclass_id(struct spu_state *csa, struct spu *spu) 4397c038749SMark Nutter { 4407c038749SMark Nutter /* Save, Step 25: 4417c038749SMark Nutter * Save the MFC_TCLASS_ID register in 4427c038749SMark Nutter * the CSA. 4437c038749SMark Nutter */ 444f0831accSArnd Bergmann csa->priv1.mfc_tclass_id_RW = spu_mfc_tclass_id_get(spu); 4457c038749SMark Nutter } 4467c038749SMark Nutter 4477c038749SMark Nutter static inline void set_mfc_tclass_id(struct spu_state *csa, struct spu *spu) 4487c038749SMark Nutter { 4497c038749SMark Nutter /* Save, Step 26: 4507c038749SMark Nutter * Restore, Step 23. 4517c038749SMark Nutter * Write the MFC_TCLASS_ID register with 4527c038749SMark Nutter * the value 0x10000000. 4537c038749SMark Nutter */ 454f0831accSArnd Bergmann spu_mfc_tclass_id_set(spu, 0x10000000); 4557c038749SMark Nutter eieio(); 4567c038749SMark Nutter } 4577c038749SMark Nutter 4587c038749SMark Nutter static inline void purge_mfc_queue(struct spu_state *csa, struct spu *spu) 4597c038749SMark Nutter { 4607c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 4617c038749SMark Nutter 4627c038749SMark Nutter /* Save, Step 27: 4637c038749SMark Nutter * Restore, Step 14. 4647c038749SMark Nutter * Write MFC_CNTL[Pc]=1 (purge queue). 4657c038749SMark Nutter */ 4667c038749SMark Nutter out_be64(&priv2->mfc_control_RW, MFC_CNTL_PURGE_DMA_REQUEST); 4677c038749SMark Nutter eieio(); 4687c038749SMark Nutter } 4697c038749SMark Nutter 4707c038749SMark Nutter static inline void wait_purge_complete(struct spu_state *csa, struct spu *spu) 4717c038749SMark Nutter { 4727c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 4737c038749SMark Nutter 4747c038749SMark Nutter /* Save, Step 28: 4757c038749SMark Nutter * Poll MFC_CNTL[Ps] until value '11' is read 4767c038749SMark Nutter * (purge complete). 4777c038749SMark Nutter */ 478910ab66bSBenjamin Herrenschmidt POLL_WHILE_FALSE((in_be64(&priv2->mfc_control_RW) & 479910ab66bSBenjamin Herrenschmidt MFC_CNTL_PURGE_DMA_STATUS_MASK) == 4807c038749SMark Nutter MFC_CNTL_PURGE_DMA_COMPLETE); 4817c038749SMark Nutter } 4827c038749SMark Nutter 4837c038749SMark Nutter static inline void setup_mfc_sr1(struct spu_state *csa, struct spu *spu) 4847c038749SMark Nutter { 4857c038749SMark Nutter /* Save, Step 30: 4867c038749SMark Nutter * Restore, Step 18: 4877c038749SMark Nutter * Write MFC_SR1 with MFC_SR1[D=0,S=1] and 4887c038749SMark Nutter * MFC_SR1[TL,R,Pr,T] set correctly for the 4897c038749SMark Nutter * OS specific environment. 4907c038749SMark Nutter * 4917c038749SMark Nutter * Implementation note: The SPU-side code 4927c038749SMark Nutter * for save/restore is privileged, so the 4937c038749SMark Nutter * MFC_SR1[Pr] bit is not set. 4947c038749SMark Nutter * 4957c038749SMark Nutter */ 496f0831accSArnd Bergmann spu_mfc_sr1_set(spu, (MFC_STATE1_MASTER_RUN_CONTROL_MASK | 4977c038749SMark Nutter MFC_STATE1_RELOCATE_MASK | 4987c038749SMark Nutter MFC_STATE1_BUS_TLBIE_MASK)); 4997c038749SMark Nutter } 5007c038749SMark Nutter 5017c038749SMark Nutter static inline void save_spu_npc(struct spu_state *csa, struct spu *spu) 5027c038749SMark Nutter { 5037c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 5047c038749SMark Nutter 5057c038749SMark Nutter /* Save, Step 31: 5067c038749SMark Nutter * Save SPU_NPC in the CSA. 5077c038749SMark Nutter */ 5087c038749SMark Nutter csa->prob.spu_npc_RW = in_be32(&prob->spu_npc_RW); 5097c038749SMark Nutter } 5107c038749SMark Nutter 5117c038749SMark Nutter static inline void save_spu_privcntl(struct spu_state *csa, struct spu *spu) 5127c038749SMark Nutter { 5137c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 5147c038749SMark Nutter 5157c038749SMark Nutter /* Save, Step 32: 5167c038749SMark Nutter * Save SPU_PrivCntl in the CSA. 5177c038749SMark Nutter */ 5187c038749SMark Nutter csa->priv2.spu_privcntl_RW = in_be64(&priv2->spu_privcntl_RW); 5197c038749SMark Nutter } 5207c038749SMark Nutter 5217c038749SMark Nutter static inline void reset_spu_privcntl(struct spu_state *csa, struct spu *spu) 5227c038749SMark Nutter { 5237c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 5247c038749SMark Nutter 5257c038749SMark Nutter /* Save, Step 33: 5267c038749SMark Nutter * Restore, Step 16: 5277c038749SMark Nutter * Write SPU_PrivCntl[S,Le,A] fields reset to 0. 5287c038749SMark Nutter */ 5297c038749SMark Nutter out_be64(&priv2->spu_privcntl_RW, 0UL); 5307c038749SMark Nutter eieio(); 5317c038749SMark Nutter } 5327c038749SMark Nutter 5337c038749SMark Nutter static inline void save_spu_lslr(struct spu_state *csa, struct spu *spu) 5347c038749SMark Nutter { 5357c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 5367c038749SMark Nutter 5377c038749SMark Nutter /* Save, Step 34: 5387c038749SMark Nutter * Save SPU_LSLR in the CSA. 5397c038749SMark Nutter */ 5407c038749SMark Nutter csa->priv2.spu_lslr_RW = in_be64(&priv2->spu_lslr_RW); 5417c038749SMark Nutter } 5427c038749SMark Nutter 5437c038749SMark Nutter static inline void reset_spu_lslr(struct spu_state *csa, struct spu *spu) 5447c038749SMark Nutter { 5457c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 5467c038749SMark Nutter 5477c038749SMark Nutter /* Save, Step 35: 5487c038749SMark Nutter * Restore, Step 17. 5497c038749SMark Nutter * Reset SPU_LSLR. 5507c038749SMark Nutter */ 5517c038749SMark Nutter out_be64(&priv2->spu_lslr_RW, LS_ADDR_MASK); 5527c038749SMark Nutter eieio(); 5537c038749SMark Nutter } 5547c038749SMark Nutter 5557c038749SMark Nutter static inline void save_spu_cfg(struct spu_state *csa, struct spu *spu) 5567c038749SMark Nutter { 5577c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 5587c038749SMark Nutter 5597c038749SMark Nutter /* Save, Step 36: 5607c038749SMark Nutter * Save SPU_Cfg in the CSA. 5617c038749SMark Nutter */ 5627c038749SMark Nutter csa->priv2.spu_cfg_RW = in_be64(&priv2->spu_cfg_RW); 5637c038749SMark Nutter } 5647c038749SMark Nutter 5657c038749SMark Nutter static inline void save_pm_trace(struct spu_state *csa, struct spu *spu) 5667c038749SMark Nutter { 5677c038749SMark Nutter /* Save, Step 37: 5687c038749SMark Nutter * Save PM_Trace_Tag_Wait_Mask in the CSA. 5697c038749SMark Nutter * Not performed by this implementation. 5707c038749SMark Nutter */ 5717c038749SMark Nutter } 5727c038749SMark Nutter 5737c038749SMark Nutter static inline void save_mfc_rag(struct spu_state *csa, struct spu *spu) 5747c038749SMark Nutter { 5757c038749SMark Nutter /* Save, Step 38: 5767c038749SMark Nutter * Save RA_GROUP_ID register and the 5777c038749SMark Nutter * RA_ENABLE reigster in the CSA. 5787c038749SMark Nutter */ 5797c038749SMark Nutter csa->priv1.resource_allocation_groupID_RW = 580f0831accSArnd Bergmann spu_resource_allocation_groupID_get(spu); 5817c038749SMark Nutter csa->priv1.resource_allocation_enable_RW = 582f0831accSArnd Bergmann spu_resource_allocation_enable_get(spu); 5837c038749SMark Nutter } 5847c038749SMark Nutter 5857c038749SMark Nutter static inline void save_ppu_mb_stat(struct spu_state *csa, struct spu *spu) 5867c038749SMark Nutter { 5877c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 5887c038749SMark Nutter 5897c038749SMark Nutter /* Save, Step 39: 5907c038749SMark Nutter * Save MB_Stat register in the CSA. 5917c038749SMark Nutter */ 5927c038749SMark Nutter csa->prob.mb_stat_R = in_be32(&prob->mb_stat_R); 5937c038749SMark Nutter } 5947c038749SMark Nutter 5957c038749SMark Nutter static inline void save_ppu_mb(struct spu_state *csa, struct spu *spu) 5967c038749SMark Nutter { 5977c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 5987c038749SMark Nutter 5997c038749SMark Nutter /* Save, Step 40: 6007c038749SMark Nutter * Save the PPU_MB register in the CSA. 6017c038749SMark Nutter */ 6027c038749SMark Nutter csa->prob.pu_mb_R = in_be32(&prob->pu_mb_R); 6037c038749SMark Nutter } 6047c038749SMark Nutter 6057c038749SMark Nutter static inline void save_ppuint_mb(struct spu_state *csa, struct spu *spu) 6067c038749SMark Nutter { 6077c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 6087c038749SMark Nutter 6097c038749SMark Nutter /* Save, Step 41: 6107c038749SMark Nutter * Save the PPUINT_MB register in the CSA. 6117c038749SMark Nutter */ 6127c038749SMark Nutter csa->priv2.puint_mb_R = in_be64(&priv2->puint_mb_R); 6137c038749SMark Nutter } 6147c038749SMark Nutter 6157c038749SMark Nutter static inline void save_ch_part1(struct spu_state *csa, struct spu *spu) 6167c038749SMark Nutter { 6177c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 618e46a0237Sarnd@arndb.de u64 idx, ch_indices[7] = { 0UL, 3UL, 4UL, 24UL, 25UL, 27UL }; 6197c038749SMark Nutter int i; 6207c038749SMark Nutter 6217c038749SMark Nutter /* Save, Step 42: 6227c038749SMark Nutter */ 623e46a0237Sarnd@arndb.de 624e46a0237Sarnd@arndb.de /* Save CH 1, without channel count */ 625e46a0237Sarnd@arndb.de out_be64(&priv2->spu_chnlcntptr_RW, 1); 626e46a0237Sarnd@arndb.de csa->spu_chnldata_RW[1] = in_be64(&priv2->spu_chnldata_RW); 627e46a0237Sarnd@arndb.de 628e46a0237Sarnd@arndb.de /* Save the following CH: [0,3,4,24,25,27] */ 6297c038749SMark Nutter for (i = 0; i < 7; i++) { 6307c038749SMark Nutter idx = ch_indices[i]; 6317c038749SMark Nutter out_be64(&priv2->spu_chnlcntptr_RW, idx); 6327c038749SMark Nutter eieio(); 6337c038749SMark Nutter csa->spu_chnldata_RW[idx] = in_be64(&priv2->spu_chnldata_RW); 6347c038749SMark Nutter csa->spu_chnlcnt_RW[idx] = in_be64(&priv2->spu_chnlcnt_RW); 6357c038749SMark Nutter out_be64(&priv2->spu_chnldata_RW, 0UL); 6367c038749SMark Nutter out_be64(&priv2->spu_chnlcnt_RW, 0UL); 6377c038749SMark Nutter eieio(); 6387c038749SMark Nutter } 6397c038749SMark Nutter } 6407c038749SMark Nutter 6417c038749SMark Nutter static inline void save_spu_mb(struct spu_state *csa, struct spu *spu) 6427c038749SMark Nutter { 6437c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 6447c038749SMark Nutter int i; 6457c038749SMark Nutter 6467c038749SMark Nutter /* Save, Step 43: 6477c038749SMark Nutter * Save SPU Read Mailbox Channel. 6487c038749SMark Nutter */ 6497c038749SMark Nutter out_be64(&priv2->spu_chnlcntptr_RW, 29UL); 6507c038749SMark Nutter eieio(); 6517c038749SMark Nutter csa->spu_chnlcnt_RW[29] = in_be64(&priv2->spu_chnlcnt_RW); 6527c038749SMark Nutter for (i = 0; i < 4; i++) { 6538b3d6663SArnd Bergmann csa->spu_mailbox_data[i] = in_be64(&priv2->spu_chnldata_RW); 6547c038749SMark Nutter } 6557c038749SMark Nutter out_be64(&priv2->spu_chnlcnt_RW, 0UL); 6567c038749SMark Nutter eieio(); 6577c038749SMark Nutter } 6587c038749SMark Nutter 6597c038749SMark Nutter static inline void save_mfc_cmd(struct spu_state *csa, struct spu *spu) 6607c038749SMark Nutter { 6617c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 6627c038749SMark Nutter 6637c038749SMark Nutter /* Save, Step 44: 6647c038749SMark Nutter * Save MFC_CMD Channel. 6657c038749SMark Nutter */ 6667c038749SMark Nutter out_be64(&priv2->spu_chnlcntptr_RW, 21UL); 6677c038749SMark Nutter eieio(); 6687c038749SMark Nutter csa->spu_chnlcnt_RW[21] = in_be64(&priv2->spu_chnlcnt_RW); 6697c038749SMark Nutter eieio(); 6707c038749SMark Nutter } 6717c038749SMark Nutter 6727c038749SMark Nutter static inline void reset_ch(struct spu_state *csa, struct spu *spu) 6737c038749SMark Nutter { 6747c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 6757c038749SMark Nutter u64 ch_indices[4] = { 21UL, 23UL, 28UL, 30UL }; 6767c038749SMark Nutter u64 ch_counts[4] = { 16UL, 1UL, 1UL, 1UL }; 6777c038749SMark Nutter u64 idx; 6787c038749SMark Nutter int i; 6797c038749SMark Nutter 6807c038749SMark Nutter /* Save, Step 45: 6817c038749SMark Nutter * Reset the following CH: [21, 23, 28, 30] 6827c038749SMark Nutter */ 6837c038749SMark Nutter for (i = 0; i < 4; i++) { 6847c038749SMark Nutter idx = ch_indices[i]; 6857c038749SMark Nutter out_be64(&priv2->spu_chnlcntptr_RW, idx); 6867c038749SMark Nutter eieio(); 6877c038749SMark Nutter out_be64(&priv2->spu_chnlcnt_RW, ch_counts[i]); 6887c038749SMark Nutter eieio(); 6897c038749SMark Nutter } 6907c038749SMark Nutter } 6917c038749SMark Nutter 6927c038749SMark Nutter static inline void resume_mfc_queue(struct spu_state *csa, struct spu *spu) 6937c038749SMark Nutter { 6947c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 6957c038749SMark Nutter 6967c038749SMark Nutter /* Save, Step 46: 6977c038749SMark Nutter * Restore, Step 25. 6987c038749SMark Nutter * Write MFC_CNTL[Sc]=0 (resume queue processing). 6997c038749SMark Nutter */ 7007c038749SMark Nutter out_be64(&priv2->mfc_control_RW, MFC_CNTL_RESUME_DMA_QUEUE); 7017c038749SMark Nutter } 7027c038749SMark Nutter 7037c038749SMark Nutter static inline void get_kernel_slb(u64 ea, u64 slb[2]) 7047c038749SMark Nutter { 705724bd80eSarnd@arndb.de u64 llp; 7067c038749SMark Nutter 707724bd80eSarnd@arndb.de if (REGION_ID(ea) == KERNEL_REGION_ID) 708724bd80eSarnd@arndb.de llp = mmu_psize_defs[mmu_linear_psize].sllp; 709724bd80eSarnd@arndb.de else 710724bd80eSarnd@arndb.de llp = mmu_psize_defs[mmu_virtual_psize].sllp; 711724bd80eSarnd@arndb.de slb[0] = (get_kernel_vsid(ea) << SLB_VSID_SHIFT) | 712724bd80eSarnd@arndb.de SLB_VSID_KERNEL | llp; 713724bd80eSarnd@arndb.de slb[1] = (ea & ESID_MASK) | SLB_ESID_V; 7147c038749SMark Nutter } 7157c038749SMark Nutter 7167c038749SMark Nutter static inline void load_mfc_slb(struct spu *spu, u64 slb[2], int slbe) 7177c038749SMark Nutter { 7187c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 7197c038749SMark Nutter 7207c038749SMark Nutter out_be64(&priv2->slb_index_W, slbe); 7217c038749SMark Nutter eieio(); 7227c038749SMark Nutter out_be64(&priv2->slb_vsid_RW, slb[0]); 7237c038749SMark Nutter out_be64(&priv2->slb_esid_RW, slb[1]); 7247c038749SMark Nutter eieio(); 7257c038749SMark Nutter } 7267c038749SMark Nutter 7277c038749SMark Nutter static inline void setup_mfc_slbs(struct spu_state *csa, struct spu *spu) 7287c038749SMark Nutter { 7297c038749SMark Nutter u64 code_slb[2]; 7307c038749SMark Nutter u64 lscsa_slb[2]; 7317c038749SMark Nutter 7327c038749SMark Nutter /* Save, Step 47: 7337c038749SMark Nutter * Restore, Step 30. 7347c038749SMark Nutter * If MFC_SR1[R]=1, write 0 to SLB_Invalidate_All 7357c038749SMark Nutter * register, then initialize SLB_VSID and SLB_ESID 7367c038749SMark Nutter * to provide access to SPU context save code and 7377c038749SMark Nutter * LSCSA. 7387c038749SMark Nutter * 7397c038749SMark Nutter * This implementation places both the context 7407c038749SMark Nutter * switch code and LSCSA in kernel address space. 7417c038749SMark Nutter * 7427c038749SMark Nutter * Further this implementation assumes that the 7437c038749SMark Nutter * MFC_SR1[R]=1 (in other words, assume that 7447c038749SMark Nutter * translation is desired by OS environment). 7457c038749SMark Nutter */ 74694b2a439SBenjamin Herrenschmidt spu_invalidate_slbs(spu); 7477c038749SMark Nutter get_kernel_slb((unsigned long)&spu_save_code[0], code_slb); 7487c038749SMark Nutter get_kernel_slb((unsigned long)csa->lscsa, lscsa_slb); 7497c038749SMark Nutter load_mfc_slb(spu, code_slb, 0); 7507c038749SMark Nutter if ((lscsa_slb[0] != code_slb[0]) || (lscsa_slb[1] != code_slb[1])) 7517c038749SMark Nutter load_mfc_slb(spu, lscsa_slb, 1); 7527c038749SMark Nutter } 7537c038749SMark Nutter 7547c038749SMark Nutter static inline void set_switch_active(struct spu_state *csa, struct spu *spu) 7557c038749SMark Nutter { 7567c038749SMark Nutter /* Save, Step 48: 7577c038749SMark Nutter * Restore, Step 23. 7587c038749SMark Nutter * Change the software context switch pending flag 7597c038749SMark Nutter * to context switch active. 7607c038749SMark Nutter */ 7618837d921SArnd Bergmann set_bit(SPU_CONTEXT_SWITCH_ACTIVE, &spu->flags); 7628837d921SArnd Bergmann clear_bit(SPU_CONTEXT_SWITCH_PENDING, &spu->flags); 7637c038749SMark Nutter mb(); 7647c038749SMark Nutter } 7657c038749SMark Nutter 7667c038749SMark Nutter static inline void enable_interrupts(struct spu_state *csa, struct spu *spu) 7677c038749SMark Nutter { 7687c038749SMark Nutter unsigned long class1_mask = CLASS1_ENABLE_SEGMENT_FAULT_INTR | 7697c038749SMark Nutter CLASS1_ENABLE_STORAGE_FAULT_INTR; 7707c038749SMark Nutter 7717c038749SMark Nutter /* Save, Step 49: 7727c038749SMark Nutter * Restore, Step 22: 7737c038749SMark Nutter * Reset and then enable interrupts, as 7747c038749SMark Nutter * needed by OS. 7757c038749SMark Nutter * 7767c038749SMark Nutter * This implementation enables only class1 7777c038749SMark Nutter * (translation) interrupts. 7787c038749SMark Nutter */ 7797c038749SMark Nutter spin_lock_irq(&spu->register_lock); 780f0831accSArnd Bergmann spu_int_stat_clear(spu, 0, ~0ul); 781f0831accSArnd Bergmann spu_int_stat_clear(spu, 1, ~0ul); 782f0831accSArnd Bergmann spu_int_stat_clear(spu, 2, ~0ul); 783f0831accSArnd Bergmann spu_int_mask_set(spu, 0, 0ul); 784f0831accSArnd Bergmann spu_int_mask_set(spu, 1, class1_mask); 785f0831accSArnd Bergmann spu_int_mask_set(spu, 2, 0ul); 7867c038749SMark Nutter spin_unlock_irq(&spu->register_lock); 7877c038749SMark Nutter } 7887c038749SMark Nutter 7897c038749SMark Nutter static inline int send_mfc_dma(struct spu *spu, unsigned long ea, 7907c038749SMark Nutter unsigned int ls_offset, unsigned int size, 7917c038749SMark Nutter unsigned int tag, unsigned int rclass, 7927c038749SMark Nutter unsigned int cmd) 7937c038749SMark Nutter { 7947c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 7957c038749SMark Nutter union mfc_tag_size_class_cmd command; 7967c038749SMark Nutter unsigned int transfer_size; 7977c038749SMark Nutter volatile unsigned int status = 0x0; 7987c038749SMark Nutter 7997c038749SMark Nutter while (size > 0) { 8007c038749SMark Nutter transfer_size = 8017c038749SMark Nutter (size > MFC_MAX_DMA_SIZE) ? MFC_MAX_DMA_SIZE : size; 8027c038749SMark Nutter command.u.mfc_size = transfer_size; 8037c038749SMark Nutter command.u.mfc_tag = tag; 8047c038749SMark Nutter command.u.mfc_rclassid = rclass; 8057c038749SMark Nutter command.u.mfc_cmd = cmd; 8067c038749SMark Nutter do { 8077c038749SMark Nutter out_be32(&prob->mfc_lsa_W, ls_offset); 8087c038749SMark Nutter out_be64(&prob->mfc_ea_W, ea); 8097c038749SMark Nutter out_be64(&prob->mfc_union_W.all64, command.all64); 8107c038749SMark Nutter status = 8117c038749SMark Nutter in_be32(&prob->mfc_union_W.by32.mfc_class_cmd32); 8127c038749SMark Nutter if (unlikely(status & 0x2)) { 8137c038749SMark Nutter cpu_relax(); 8147c038749SMark Nutter } 8157c038749SMark Nutter } while (status & 0x3); 8167c038749SMark Nutter size -= transfer_size; 8177c038749SMark Nutter ea += transfer_size; 8187c038749SMark Nutter ls_offset += transfer_size; 8197c038749SMark Nutter } 8207c038749SMark Nutter return 0; 8217c038749SMark Nutter } 8227c038749SMark Nutter 8237c038749SMark Nutter static inline void save_ls_16kb(struct spu_state *csa, struct spu *spu) 8247c038749SMark Nutter { 8257c038749SMark Nutter unsigned long addr = (unsigned long)&csa->lscsa->ls[0]; 8267c038749SMark Nutter unsigned int ls_offset = 0x0; 8277c038749SMark Nutter unsigned int size = 16384; 8287c038749SMark Nutter unsigned int tag = 0; 8297c038749SMark Nutter unsigned int rclass = 0; 8307c038749SMark Nutter unsigned int cmd = MFC_PUT_CMD; 8317c038749SMark Nutter 8327c038749SMark Nutter /* Save, Step 50: 8337c038749SMark Nutter * Issue a DMA command to copy the first 16K bytes 8347c038749SMark Nutter * of local storage to the CSA. 8357c038749SMark Nutter */ 8367c038749SMark Nutter send_mfc_dma(spu, addr, ls_offset, size, tag, rclass, cmd); 8377c038749SMark Nutter } 8387c038749SMark Nutter 8397c038749SMark Nutter static inline void set_spu_npc(struct spu_state *csa, struct spu *spu) 8407c038749SMark Nutter { 8417c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 8427c038749SMark Nutter 8437c038749SMark Nutter /* Save, Step 51: 8447c038749SMark Nutter * Restore, Step 31. 8457c038749SMark Nutter * Write SPU_NPC[IE]=0 and SPU_NPC[LSA] to entry 8467c038749SMark Nutter * point address of context save code in local 8477c038749SMark Nutter * storage. 8487c038749SMark Nutter * 8497c038749SMark Nutter * This implementation uses SPU-side save/restore 8507c038749SMark Nutter * programs with entry points at LSA of 0. 8517c038749SMark Nutter */ 8527c038749SMark Nutter out_be32(&prob->spu_npc_RW, 0); 8537c038749SMark Nutter eieio(); 8547c038749SMark Nutter } 8557c038749SMark Nutter 8567c038749SMark Nutter static inline void set_signot1(struct spu_state *csa, struct spu *spu) 8577c038749SMark Nutter { 8587c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 8597c038749SMark Nutter union { 8607c038749SMark Nutter u64 ull; 8617c038749SMark Nutter u32 ui[2]; 8627c038749SMark Nutter } addr64; 8637c038749SMark Nutter 8647c038749SMark Nutter /* Save, Step 52: 8657c038749SMark Nutter * Restore, Step 32: 8667c038749SMark Nutter * Write SPU_Sig_Notify_1 register with upper 32-bits 8677c038749SMark Nutter * of the CSA.LSCSA effective address. 8687c038749SMark Nutter */ 8697c038749SMark Nutter addr64.ull = (u64) csa->lscsa; 8707c038749SMark Nutter out_be32(&prob->signal_notify1, addr64.ui[0]); 8717c038749SMark Nutter eieio(); 8727c038749SMark Nutter } 8737c038749SMark Nutter 8747c038749SMark Nutter static inline void set_signot2(struct spu_state *csa, struct spu *spu) 8757c038749SMark Nutter { 8767c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 8777c038749SMark Nutter union { 8787c038749SMark Nutter u64 ull; 8797c038749SMark Nutter u32 ui[2]; 8807c038749SMark Nutter } addr64; 8817c038749SMark Nutter 8827c038749SMark Nutter /* Save, Step 53: 8837c038749SMark Nutter * Restore, Step 33: 8847c038749SMark Nutter * Write SPU_Sig_Notify_2 register with lower 32-bits 8857c038749SMark Nutter * of the CSA.LSCSA effective address. 8867c038749SMark Nutter */ 8877c038749SMark Nutter addr64.ull = (u64) csa->lscsa; 8887c038749SMark Nutter out_be32(&prob->signal_notify2, addr64.ui[1]); 8897c038749SMark Nutter eieio(); 8907c038749SMark Nutter } 8917c038749SMark Nutter 8927c038749SMark Nutter static inline void send_save_code(struct spu_state *csa, struct spu *spu) 8937c038749SMark Nutter { 8947c038749SMark Nutter unsigned long addr = (unsigned long)&spu_save_code[0]; 8957c038749SMark Nutter unsigned int ls_offset = 0x0; 8967c038749SMark Nutter unsigned int size = sizeof(spu_save_code); 8977c038749SMark Nutter unsigned int tag = 0; 8987c038749SMark Nutter unsigned int rclass = 0; 8997c038749SMark Nutter unsigned int cmd = MFC_GETFS_CMD; 9007c038749SMark Nutter 9017c038749SMark Nutter /* Save, Step 54: 9027c038749SMark Nutter * Issue a DMA command to copy context save code 9037c038749SMark Nutter * to local storage and start SPU. 9047c038749SMark Nutter */ 9057c038749SMark Nutter send_mfc_dma(spu, addr, ls_offset, size, tag, rclass, cmd); 9067c038749SMark Nutter } 9077c038749SMark Nutter 9087c038749SMark Nutter static inline void set_ppu_querymask(struct spu_state *csa, struct spu *spu) 9097c038749SMark Nutter { 9107c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 9117c038749SMark Nutter 9127c038749SMark Nutter /* Save, Step 55: 9137c038749SMark Nutter * Restore, Step 38. 9147c038749SMark Nutter * Write PPU_QueryMask=1 (enable Tag Group 0) 9157c038749SMark Nutter * and issue eieio instruction. 9167c038749SMark Nutter */ 9177c038749SMark Nutter out_be32(&prob->dma_querymask_RW, MFC_TAGID_TO_TAGMASK(0)); 9187c038749SMark Nutter eieio(); 9197c038749SMark Nutter } 9207c038749SMark Nutter 9217c038749SMark Nutter static inline void wait_tag_complete(struct spu_state *csa, struct spu *spu) 9227c038749SMark Nutter { 9237c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 9247c038749SMark Nutter u32 mask = MFC_TAGID_TO_TAGMASK(0); 9257c038749SMark Nutter unsigned long flags; 9267c038749SMark Nutter 9277c038749SMark Nutter /* Save, Step 56: 9287c038749SMark Nutter * Restore, Step 39. 9297c038749SMark Nutter * Restore, Step 39. 9307c038749SMark Nutter * Restore, Step 46. 9317c038749SMark Nutter * Poll PPU_TagStatus[gn] until 01 (Tag group 0 complete) 9327c038749SMark Nutter * or write PPU_QueryType[TS]=01 and wait for Tag Group 9337c038749SMark Nutter * Complete Interrupt. Write INT_Stat_Class0 or 9347c038749SMark Nutter * INT_Stat_Class2 with value of 'handled'. 9357c038749SMark Nutter */ 9367c038749SMark Nutter POLL_WHILE_FALSE(in_be32(&prob->dma_tagstatus_R) & mask); 9377c038749SMark Nutter 9387c038749SMark Nutter local_irq_save(flags); 939f0831accSArnd Bergmann spu_int_stat_clear(spu, 0, ~(0ul)); 940f0831accSArnd Bergmann spu_int_stat_clear(spu, 2, ~(0ul)); 9417c038749SMark Nutter local_irq_restore(flags); 9427c038749SMark Nutter } 9437c038749SMark Nutter 9447c038749SMark Nutter static inline void wait_spu_stopped(struct spu_state *csa, struct spu *spu) 9457c038749SMark Nutter { 9467c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 9477c038749SMark Nutter unsigned long flags; 9487c038749SMark Nutter 9497c038749SMark Nutter /* Save, Step 57: 9507c038749SMark Nutter * Restore, Step 40. 9517c038749SMark Nutter * Poll until SPU_Status[R]=0 or wait for SPU Class 0 9527c038749SMark Nutter * or SPU Class 2 interrupt. Write INT_Stat_class0 9537c038749SMark Nutter * or INT_Stat_class2 with value of handled. 9547c038749SMark Nutter */ 9557c038749SMark Nutter POLL_WHILE_TRUE(in_be32(&prob->spu_status_R) & SPU_STATUS_RUNNING); 9567c038749SMark Nutter 9577c038749SMark Nutter local_irq_save(flags); 958f0831accSArnd Bergmann spu_int_stat_clear(spu, 0, ~(0ul)); 959f0831accSArnd Bergmann spu_int_stat_clear(spu, 2, ~(0ul)); 9607c038749SMark Nutter local_irq_restore(flags); 9617c038749SMark Nutter } 9627c038749SMark Nutter 9637c038749SMark Nutter static inline int check_save_status(struct spu_state *csa, struct spu *spu) 9647c038749SMark Nutter { 9657c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 9667c038749SMark Nutter u32 complete; 9677c038749SMark Nutter 9687c038749SMark Nutter /* Save, Step 54: 9697c038749SMark Nutter * If SPU_Status[P]=1 and SPU_Status[SC] = "success", 9707c038749SMark Nutter * context save succeeded, otherwise context save 9717c038749SMark Nutter * failed. 9727c038749SMark Nutter */ 9737c038749SMark Nutter complete = ((SPU_SAVE_COMPLETE << SPU_STOP_STATUS_SHIFT) | 9747c038749SMark Nutter SPU_STATUS_STOPPED_BY_STOP); 9757c038749SMark Nutter return (in_be32(&prob->spu_status_R) != complete) ? 1 : 0; 9767c038749SMark Nutter } 9777c038749SMark Nutter 9787c038749SMark Nutter static inline void terminate_spu_app(struct spu_state *csa, struct spu *spu) 9797c038749SMark Nutter { 9807c038749SMark Nutter /* Restore, Step 4: 9817c038749SMark Nutter * If required, notify the "using application" that 9827c038749SMark Nutter * the SPU task has been terminated. TBD. 9837c038749SMark Nutter */ 9847c038749SMark Nutter } 9857c038749SMark Nutter 9867c038749SMark Nutter static inline void suspend_mfc(struct spu_state *csa, struct spu *spu) 9877c038749SMark Nutter { 9887c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 9897c038749SMark Nutter 9907c038749SMark Nutter /* Restore, Step 7: 9917c038749SMark Nutter * Restore, Step 47. 9927c038749SMark Nutter * Write MFC_Cntl[Dh,Sc]='1','1' to suspend 9937c038749SMark Nutter * the queue and halt the decrementer. 9947c038749SMark Nutter */ 9957c038749SMark Nutter out_be64(&priv2->mfc_control_RW, MFC_CNTL_SUSPEND_DMA_QUEUE | 9967c038749SMark Nutter MFC_CNTL_DECREMENTER_HALTED); 9977c038749SMark Nutter eieio(); 9987c038749SMark Nutter } 9997c038749SMark Nutter 10007c038749SMark Nutter static inline void wait_suspend_mfc_complete(struct spu_state *csa, 10017c038749SMark Nutter struct spu *spu) 10027c038749SMark Nutter { 10037c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 10047c038749SMark Nutter 10057c038749SMark Nutter /* Restore, Step 8: 10067c038749SMark Nutter * Restore, Step 47. 10077c038749SMark Nutter * Poll MFC_CNTL[Ss] until 11 is returned. 10087c038749SMark Nutter */ 1009910ab66bSBenjamin Herrenschmidt POLL_WHILE_FALSE((in_be64(&priv2->mfc_control_RW) & 1010910ab66bSBenjamin Herrenschmidt MFC_CNTL_SUSPEND_DMA_STATUS_MASK) == 10117c038749SMark Nutter MFC_CNTL_SUSPEND_COMPLETE); 10127c038749SMark Nutter } 10137c038749SMark Nutter 10147c038749SMark Nutter static inline int suspend_spe(struct spu_state *csa, struct spu *spu) 10157c038749SMark Nutter { 10167c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 10177c038749SMark Nutter 10187c038749SMark Nutter /* Restore, Step 9: 10197c038749SMark Nutter * If SPU_Status[R]=1, stop SPU execution 10207c038749SMark Nutter * and wait for stop to complete. 10217c038749SMark Nutter * 10227c038749SMark Nutter * Returns 1 if SPU_Status[R]=1 on entry. 10237c038749SMark Nutter * 0 otherwise 10247c038749SMark Nutter */ 10257c038749SMark Nutter if (in_be32(&prob->spu_status_R) & SPU_STATUS_RUNNING) { 10267c038749SMark Nutter if (in_be32(&prob->spu_status_R) & 1027eb758ce5Sarnd@arndb.de SPU_STATUS_ISOLATED_EXIT_STATUS) { 10287c038749SMark Nutter POLL_WHILE_TRUE(in_be32(&prob->spu_status_R) & 10297c038749SMark Nutter SPU_STATUS_RUNNING); 10307c038749SMark Nutter } 10317c038749SMark Nutter if ((in_be32(&prob->spu_status_R) & 1032eb758ce5Sarnd@arndb.de SPU_STATUS_ISOLATED_LOAD_STATUS) 10337c038749SMark Nutter || (in_be32(&prob->spu_status_R) & 10347c038749SMark Nutter SPU_STATUS_ISOLATED_STATE)) { 10357c038749SMark Nutter out_be32(&prob->spu_runcntl_RW, SPU_RUNCNTL_STOP); 10367c038749SMark Nutter eieio(); 10377c038749SMark Nutter POLL_WHILE_TRUE(in_be32(&prob->spu_status_R) & 10387c038749SMark Nutter SPU_STATUS_RUNNING); 10397c038749SMark Nutter out_be32(&prob->spu_runcntl_RW, 0x2); 10407c038749SMark Nutter eieio(); 10417c038749SMark Nutter POLL_WHILE_TRUE(in_be32(&prob->spu_status_R) & 10427c038749SMark Nutter SPU_STATUS_RUNNING); 10437c038749SMark Nutter } 10447c038749SMark Nutter if (in_be32(&prob->spu_status_R) & 10457c038749SMark Nutter SPU_STATUS_WAITING_FOR_CHANNEL) { 10467c038749SMark Nutter out_be32(&prob->spu_runcntl_RW, SPU_RUNCNTL_STOP); 10477c038749SMark Nutter eieio(); 10487c038749SMark Nutter POLL_WHILE_TRUE(in_be32(&prob->spu_status_R) & 10497c038749SMark Nutter SPU_STATUS_RUNNING); 10507c038749SMark Nutter } 10517c038749SMark Nutter return 1; 10527c038749SMark Nutter } 10537c038749SMark Nutter return 0; 10547c038749SMark Nutter } 10557c038749SMark Nutter 10567c038749SMark Nutter static inline void clear_spu_status(struct spu_state *csa, struct spu *spu) 10577c038749SMark Nutter { 10587c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 10597c038749SMark Nutter 10607c038749SMark Nutter /* Restore, Step 10: 10617c038749SMark Nutter * If SPU_Status[R]=0 and SPU_Status[E,L,IS]=1, 10627c038749SMark Nutter * release SPU from isolate state. 10637c038749SMark Nutter */ 10647c038749SMark Nutter if (!(in_be32(&prob->spu_status_R) & SPU_STATUS_RUNNING)) { 10657c038749SMark Nutter if (in_be32(&prob->spu_status_R) & 1066eb758ce5Sarnd@arndb.de SPU_STATUS_ISOLATED_EXIT_STATUS) { 1067f0831accSArnd Bergmann spu_mfc_sr1_set(spu, 10687c038749SMark Nutter MFC_STATE1_MASTER_RUN_CONTROL_MASK); 10697c038749SMark Nutter eieio(); 10707c038749SMark Nutter out_be32(&prob->spu_runcntl_RW, SPU_RUNCNTL_RUNNABLE); 10717c038749SMark Nutter eieio(); 10727c038749SMark Nutter POLL_WHILE_TRUE(in_be32(&prob->spu_status_R) & 10737c038749SMark Nutter SPU_STATUS_RUNNING); 10747c038749SMark Nutter } 10757c038749SMark Nutter if ((in_be32(&prob->spu_status_R) & 1076eb758ce5Sarnd@arndb.de SPU_STATUS_ISOLATED_LOAD_STATUS) 10777c038749SMark Nutter || (in_be32(&prob->spu_status_R) & 10787c038749SMark Nutter SPU_STATUS_ISOLATED_STATE)) { 1079f0831accSArnd Bergmann spu_mfc_sr1_set(spu, 10807c038749SMark Nutter MFC_STATE1_MASTER_RUN_CONTROL_MASK); 10817c038749SMark Nutter eieio(); 10827c038749SMark Nutter out_be32(&prob->spu_runcntl_RW, 0x2); 10837c038749SMark Nutter eieio(); 10847c038749SMark Nutter POLL_WHILE_TRUE(in_be32(&prob->spu_status_R) & 10857c038749SMark Nutter SPU_STATUS_RUNNING); 10867c038749SMark Nutter } 10877c038749SMark Nutter } 10887c038749SMark Nutter } 10897c038749SMark Nutter 10907c038749SMark Nutter static inline void reset_ch_part1(struct spu_state *csa, struct spu *spu) 10917c038749SMark Nutter { 10927c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 1093e46a0237Sarnd@arndb.de u64 ch_indices[7] = { 0UL, 3UL, 4UL, 24UL, 25UL, 27UL }; 10947c038749SMark Nutter u64 idx; 10957c038749SMark Nutter int i; 10967c038749SMark Nutter 10977c038749SMark Nutter /* Restore, Step 20: 10987c038749SMark Nutter */ 1099e46a0237Sarnd@arndb.de 1100e46a0237Sarnd@arndb.de /* Reset CH 1 */ 1101e46a0237Sarnd@arndb.de out_be64(&priv2->spu_chnlcntptr_RW, 1); 1102e46a0237Sarnd@arndb.de out_be64(&priv2->spu_chnldata_RW, 0UL); 1103e46a0237Sarnd@arndb.de 1104e46a0237Sarnd@arndb.de /* Reset the following CH: [0,3,4,24,25,27] */ 11057c038749SMark Nutter for (i = 0; i < 7; i++) { 11067c038749SMark Nutter idx = ch_indices[i]; 11077c038749SMark Nutter out_be64(&priv2->spu_chnlcntptr_RW, idx); 11087c038749SMark Nutter eieio(); 11097c038749SMark Nutter out_be64(&priv2->spu_chnldata_RW, 0UL); 11107c038749SMark Nutter out_be64(&priv2->spu_chnlcnt_RW, 0UL); 11117c038749SMark Nutter eieio(); 11127c038749SMark Nutter } 11137c038749SMark Nutter } 11147c038749SMark Nutter 11157c038749SMark Nutter static inline void reset_ch_part2(struct spu_state *csa, struct spu *spu) 11167c038749SMark Nutter { 11177c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 11187c038749SMark Nutter u64 ch_indices[5] = { 21UL, 23UL, 28UL, 29UL, 30UL }; 11197c038749SMark Nutter u64 ch_counts[5] = { 16UL, 1UL, 1UL, 0UL, 1UL }; 11207c038749SMark Nutter u64 idx; 11217c038749SMark Nutter int i; 11227c038749SMark Nutter 11237c038749SMark Nutter /* Restore, Step 21: 11247c038749SMark Nutter * Reset the following CH: [21, 23, 28, 29, 30] 11257c038749SMark Nutter */ 11267c038749SMark Nutter for (i = 0; i < 5; i++) { 11277c038749SMark Nutter idx = ch_indices[i]; 11287c038749SMark Nutter out_be64(&priv2->spu_chnlcntptr_RW, idx); 11297c038749SMark Nutter eieio(); 11307c038749SMark Nutter out_be64(&priv2->spu_chnlcnt_RW, ch_counts[i]); 11317c038749SMark Nutter eieio(); 11327c038749SMark Nutter } 11337c038749SMark Nutter } 11347c038749SMark Nutter 11357c038749SMark Nutter static inline void setup_spu_status_part1(struct spu_state *csa, 11367c038749SMark Nutter struct spu *spu) 11377c038749SMark Nutter { 11387c038749SMark Nutter u32 status_P = SPU_STATUS_STOPPED_BY_STOP; 11397c038749SMark Nutter u32 status_I = SPU_STATUS_INVALID_INSTR; 11407c038749SMark Nutter u32 status_H = SPU_STATUS_STOPPED_BY_HALT; 11417c038749SMark Nutter u32 status_S = SPU_STATUS_SINGLE_STEP; 11427c038749SMark Nutter u32 status_S_I = SPU_STATUS_SINGLE_STEP | SPU_STATUS_INVALID_INSTR; 11437c038749SMark Nutter u32 status_S_P = SPU_STATUS_SINGLE_STEP | SPU_STATUS_STOPPED_BY_STOP; 11447c038749SMark Nutter u32 status_P_H = SPU_STATUS_STOPPED_BY_HALT |SPU_STATUS_STOPPED_BY_STOP; 11457c038749SMark Nutter u32 status_P_I = SPU_STATUS_STOPPED_BY_STOP |SPU_STATUS_INVALID_INSTR; 11467c038749SMark Nutter u32 status_code; 11477c038749SMark Nutter 11487c038749SMark Nutter /* Restore, Step 27: 11497c038749SMark Nutter * If the CSA.SPU_Status[I,S,H,P]=1 then add the correct 11507c038749SMark Nutter * instruction sequence to the end of the SPU based restore 11517c038749SMark Nutter * code (after the "context restored" stop and signal) to 11527c038749SMark Nutter * restore the correct SPU status. 11537c038749SMark Nutter * 11547c038749SMark Nutter * NOTE: Rather than modifying the SPU executable, we 11557c038749SMark Nutter * instead add a new 'stopped_status' field to the 11567c038749SMark Nutter * LSCSA. The SPU-side restore reads this field and 11577c038749SMark Nutter * takes the appropriate action when exiting. 11587c038749SMark Nutter */ 11597c038749SMark Nutter 11607c038749SMark Nutter status_code = 11617c038749SMark Nutter (csa->prob.spu_status_R >> SPU_STOP_STATUS_SHIFT) & 0xFFFF; 11627c038749SMark Nutter if ((csa->prob.spu_status_R & status_P_I) == status_P_I) { 11637c038749SMark Nutter 11647c038749SMark Nutter /* SPU_Status[P,I]=1 - Illegal Instruction followed 11657c038749SMark Nutter * by Stop and Signal instruction, followed by 'br -4'. 11667c038749SMark Nutter * 11677c038749SMark Nutter */ 11687c038749SMark Nutter csa->lscsa->stopped_status.slot[0] = SPU_STOPPED_STATUS_P_I; 11697c038749SMark Nutter csa->lscsa->stopped_status.slot[1] = status_code; 11707c038749SMark Nutter 11717c038749SMark Nutter } else if ((csa->prob.spu_status_R & status_P_H) == status_P_H) { 11727c038749SMark Nutter 11737c038749SMark Nutter /* SPU_Status[P,H]=1 - Halt Conditional, followed 11747c038749SMark Nutter * by Stop and Signal instruction, followed by 11757c038749SMark Nutter * 'br -4'. 11767c038749SMark Nutter */ 11777c038749SMark Nutter csa->lscsa->stopped_status.slot[0] = SPU_STOPPED_STATUS_P_H; 11787c038749SMark Nutter csa->lscsa->stopped_status.slot[1] = status_code; 11797c038749SMark Nutter 11807c038749SMark Nutter } else if ((csa->prob.spu_status_R & status_S_P) == status_S_P) { 11817c038749SMark Nutter 11827c038749SMark Nutter /* SPU_Status[S,P]=1 - Stop and Signal instruction 11837c038749SMark Nutter * followed by 'br -4'. 11847c038749SMark Nutter */ 11857c038749SMark Nutter csa->lscsa->stopped_status.slot[0] = SPU_STOPPED_STATUS_S_P; 11867c038749SMark Nutter csa->lscsa->stopped_status.slot[1] = status_code; 11877c038749SMark Nutter 11887c038749SMark Nutter } else if ((csa->prob.spu_status_R & status_S_I) == status_S_I) { 11897c038749SMark Nutter 11907c038749SMark Nutter /* SPU_Status[S,I]=1 - Illegal instruction followed 11917c038749SMark Nutter * by 'br -4'. 11927c038749SMark Nutter */ 11937c038749SMark Nutter csa->lscsa->stopped_status.slot[0] = SPU_STOPPED_STATUS_S_I; 11947c038749SMark Nutter csa->lscsa->stopped_status.slot[1] = status_code; 11957c038749SMark Nutter 11967c038749SMark Nutter } else if ((csa->prob.spu_status_R & status_P) == status_P) { 11977c038749SMark Nutter 11987c038749SMark Nutter /* SPU_Status[P]=1 - Stop and Signal instruction 11997c038749SMark Nutter * followed by 'br -4'. 12007c038749SMark Nutter */ 12017c038749SMark Nutter csa->lscsa->stopped_status.slot[0] = SPU_STOPPED_STATUS_P; 12027c038749SMark Nutter csa->lscsa->stopped_status.slot[1] = status_code; 12037c038749SMark Nutter 12047c038749SMark Nutter } else if ((csa->prob.spu_status_R & status_H) == status_H) { 12057c038749SMark Nutter 12067c038749SMark Nutter /* SPU_Status[H]=1 - Halt Conditional, followed 12077c038749SMark Nutter * by 'br -4'. 12087c038749SMark Nutter */ 12097c038749SMark Nutter csa->lscsa->stopped_status.slot[0] = SPU_STOPPED_STATUS_H; 12107c038749SMark Nutter 12117c038749SMark Nutter } else if ((csa->prob.spu_status_R & status_S) == status_S) { 12127c038749SMark Nutter 12137c038749SMark Nutter /* SPU_Status[S]=1 - Two nop instructions. 12147c038749SMark Nutter */ 12157c038749SMark Nutter csa->lscsa->stopped_status.slot[0] = SPU_STOPPED_STATUS_S; 12167c038749SMark Nutter 12177c038749SMark Nutter } else if ((csa->prob.spu_status_R & status_I) == status_I) { 12187c038749SMark Nutter 12197c038749SMark Nutter /* SPU_Status[I]=1 - Illegal instruction followed 12207c038749SMark Nutter * by 'br -4'. 12217c038749SMark Nutter */ 12227c038749SMark Nutter csa->lscsa->stopped_status.slot[0] = SPU_STOPPED_STATUS_I; 12237c038749SMark Nutter 12247c038749SMark Nutter } 12257c038749SMark Nutter } 12267c038749SMark Nutter 12277c038749SMark Nutter static inline void setup_spu_status_part2(struct spu_state *csa, 12287c038749SMark Nutter struct spu *spu) 12297c038749SMark Nutter { 12307c038749SMark Nutter u32 mask; 12317c038749SMark Nutter 12327c038749SMark Nutter /* Restore, Step 28: 12337c038749SMark Nutter * If the CSA.SPU_Status[I,S,H,P,R]=0 then 12347c038749SMark Nutter * add a 'br *' instruction to the end of 12357c038749SMark Nutter * the SPU based restore code. 12367c038749SMark Nutter * 12377c038749SMark Nutter * NOTE: Rather than modifying the SPU executable, we 12387c038749SMark Nutter * instead add a new 'stopped_status' field to the 12397c038749SMark Nutter * LSCSA. The SPU-side restore reads this field and 12407c038749SMark Nutter * takes the appropriate action when exiting. 12417c038749SMark Nutter */ 12427c038749SMark Nutter mask = SPU_STATUS_INVALID_INSTR | 12437c038749SMark Nutter SPU_STATUS_SINGLE_STEP | 12447c038749SMark Nutter SPU_STATUS_STOPPED_BY_HALT | 12457c038749SMark Nutter SPU_STATUS_STOPPED_BY_STOP | SPU_STATUS_RUNNING; 12467c038749SMark Nutter if (!(csa->prob.spu_status_R & mask)) { 12477c038749SMark Nutter csa->lscsa->stopped_status.slot[0] = SPU_STOPPED_STATUS_R; 12487c038749SMark Nutter } 12497c038749SMark Nutter } 12507c038749SMark Nutter 12517c038749SMark Nutter static inline void restore_mfc_rag(struct spu_state *csa, struct spu *spu) 12527c038749SMark Nutter { 12537c038749SMark Nutter /* Restore, Step 29: 12547c038749SMark Nutter * Restore RA_GROUP_ID register and the 12557c038749SMark Nutter * RA_ENABLE reigster from the CSA. 12567c038749SMark Nutter */ 1257f0831accSArnd Bergmann spu_resource_allocation_groupID_set(spu, 12587c038749SMark Nutter csa->priv1.resource_allocation_groupID_RW); 1259f0831accSArnd Bergmann spu_resource_allocation_enable_set(spu, 12607c038749SMark Nutter csa->priv1.resource_allocation_enable_RW); 12617c038749SMark Nutter } 12627c038749SMark Nutter 12637c038749SMark Nutter static inline void send_restore_code(struct spu_state *csa, struct spu *spu) 12647c038749SMark Nutter { 12657c038749SMark Nutter unsigned long addr = (unsigned long)&spu_restore_code[0]; 12667c038749SMark Nutter unsigned int ls_offset = 0x0; 12677c038749SMark Nutter unsigned int size = sizeof(spu_restore_code); 12687c038749SMark Nutter unsigned int tag = 0; 12697c038749SMark Nutter unsigned int rclass = 0; 12707c038749SMark Nutter unsigned int cmd = MFC_GETFS_CMD; 12717c038749SMark Nutter 12727c038749SMark Nutter /* Restore, Step 37: 12737c038749SMark Nutter * Issue MFC DMA command to copy context 12747c038749SMark Nutter * restore code to local storage. 12757c038749SMark Nutter */ 12767c038749SMark Nutter send_mfc_dma(spu, addr, ls_offset, size, tag, rclass, cmd); 12777c038749SMark Nutter } 12787c038749SMark Nutter 12797c038749SMark Nutter static inline void setup_decr(struct spu_state *csa, struct spu *spu) 12807c038749SMark Nutter { 12817c038749SMark Nutter /* Restore, Step 34: 12827c038749SMark Nutter * If CSA.MFC_CNTL[Ds]=1 (decrementer was 12837c038749SMark Nutter * running) then adjust decrementer, set 12847c038749SMark Nutter * decrementer running status in LSCSA, 12857c038749SMark Nutter * and set decrementer "wrapped" status 12867c038749SMark Nutter * in LSCSA. 12877c038749SMark Nutter */ 12887c038749SMark Nutter if (csa->priv2.mfc_control_RW & MFC_CNTL_DECREMENTER_RUNNING) { 12897c038749SMark Nutter cycles_t resume_time = get_cycles(); 12907c038749SMark Nutter cycles_t delta_time = resume_time - csa->suspend_time; 12917c038749SMark Nutter 1292183b73aeSJordi Caubet csa->lscsa->decr.slot[0] -= delta_time; 12937c038749SMark Nutter } 12947c038749SMark Nutter } 12957c038749SMark Nutter 12967c038749SMark Nutter static inline void setup_ppu_mb(struct spu_state *csa, struct spu *spu) 12977c038749SMark Nutter { 12987c038749SMark Nutter /* Restore, Step 35: 12997c038749SMark Nutter * Copy the CSA.PU_MB data into the LSCSA. 13007c038749SMark Nutter */ 13017c038749SMark Nutter csa->lscsa->ppu_mb.slot[0] = csa->prob.pu_mb_R; 13027c038749SMark Nutter } 13037c038749SMark Nutter 13047c038749SMark Nutter static inline void setup_ppuint_mb(struct spu_state *csa, struct spu *spu) 13057c038749SMark Nutter { 13067c038749SMark Nutter /* Restore, Step 36: 13077c038749SMark Nutter * Copy the CSA.PUINT_MB data into the LSCSA. 13087c038749SMark Nutter */ 13097c038749SMark Nutter csa->lscsa->ppuint_mb.slot[0] = csa->priv2.puint_mb_R; 13107c038749SMark Nutter } 13117c038749SMark Nutter 13127c038749SMark Nutter static inline int check_restore_status(struct spu_state *csa, struct spu *spu) 13137c038749SMark Nutter { 13147c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 13157c038749SMark Nutter u32 complete; 13167c038749SMark Nutter 13177c038749SMark Nutter /* Restore, Step 40: 13187c038749SMark Nutter * If SPU_Status[P]=1 and SPU_Status[SC] = "success", 13197c038749SMark Nutter * context restore succeeded, otherwise context restore 13207c038749SMark Nutter * failed. 13217c038749SMark Nutter */ 13227c038749SMark Nutter complete = ((SPU_RESTORE_COMPLETE << SPU_STOP_STATUS_SHIFT) | 13237c038749SMark Nutter SPU_STATUS_STOPPED_BY_STOP); 13247c038749SMark Nutter return (in_be32(&prob->spu_status_R) != complete) ? 1 : 0; 13257c038749SMark Nutter } 13267c038749SMark Nutter 13277c038749SMark Nutter static inline void restore_spu_privcntl(struct spu_state *csa, struct spu *spu) 13287c038749SMark Nutter { 13297c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 13307c038749SMark Nutter 13317c038749SMark Nutter /* Restore, Step 41: 13327c038749SMark Nutter * Restore SPU_PrivCntl from the CSA. 13337c038749SMark Nutter */ 13347c038749SMark Nutter out_be64(&priv2->spu_privcntl_RW, csa->priv2.spu_privcntl_RW); 13357c038749SMark Nutter eieio(); 13367c038749SMark Nutter } 13377c038749SMark Nutter 13387c038749SMark Nutter static inline void restore_status_part1(struct spu_state *csa, struct spu *spu) 13397c038749SMark Nutter { 13407c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 13417c038749SMark Nutter u32 mask; 13427c038749SMark Nutter 13437c038749SMark Nutter /* Restore, Step 42: 13447c038749SMark Nutter * If any CSA.SPU_Status[I,S,H,P]=1, then 13457c038749SMark Nutter * restore the error or single step state. 13467c038749SMark Nutter */ 13477c038749SMark Nutter mask = SPU_STATUS_INVALID_INSTR | 13487c038749SMark Nutter SPU_STATUS_SINGLE_STEP | 13497c038749SMark Nutter SPU_STATUS_STOPPED_BY_HALT | SPU_STATUS_STOPPED_BY_STOP; 13507c038749SMark Nutter if (csa->prob.spu_status_R & mask) { 13517c038749SMark Nutter out_be32(&prob->spu_runcntl_RW, SPU_RUNCNTL_RUNNABLE); 13527c038749SMark Nutter eieio(); 13537c038749SMark Nutter POLL_WHILE_TRUE(in_be32(&prob->spu_status_R) & 13547c038749SMark Nutter SPU_STATUS_RUNNING); 13557c038749SMark Nutter } 13567c038749SMark Nutter } 13577c038749SMark Nutter 13587c038749SMark Nutter static inline void restore_status_part2(struct spu_state *csa, struct spu *spu) 13597c038749SMark Nutter { 13607c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 13617c038749SMark Nutter u32 mask; 13627c038749SMark Nutter 13637c038749SMark Nutter /* Restore, Step 43: 13647c038749SMark Nutter * If all CSA.SPU_Status[I,S,H,P,R]=0 then write 13657c038749SMark Nutter * SPU_RunCntl[R0R1]='01', wait for SPU_Status[R]=1, 13667c038749SMark Nutter * then write '00' to SPU_RunCntl[R0R1] and wait 13677c038749SMark Nutter * for SPU_Status[R]=0. 13687c038749SMark Nutter */ 13697c038749SMark Nutter mask = SPU_STATUS_INVALID_INSTR | 13707c038749SMark Nutter SPU_STATUS_SINGLE_STEP | 13717c038749SMark Nutter SPU_STATUS_STOPPED_BY_HALT | 13727c038749SMark Nutter SPU_STATUS_STOPPED_BY_STOP | SPU_STATUS_RUNNING; 13737c038749SMark Nutter if (!(csa->prob.spu_status_R & mask)) { 13747c038749SMark Nutter out_be32(&prob->spu_runcntl_RW, SPU_RUNCNTL_RUNNABLE); 13757c038749SMark Nutter eieio(); 13767c038749SMark Nutter POLL_WHILE_FALSE(in_be32(&prob->spu_status_R) & 13777c038749SMark Nutter SPU_STATUS_RUNNING); 13787c038749SMark Nutter out_be32(&prob->spu_runcntl_RW, SPU_RUNCNTL_STOP); 13797c038749SMark Nutter eieio(); 13807c038749SMark Nutter POLL_WHILE_TRUE(in_be32(&prob->spu_status_R) & 13817c038749SMark Nutter SPU_STATUS_RUNNING); 13827c038749SMark Nutter } 13837c038749SMark Nutter } 13847c038749SMark Nutter 13857c038749SMark Nutter static inline void restore_ls_16kb(struct spu_state *csa, struct spu *spu) 13867c038749SMark Nutter { 13877c038749SMark Nutter unsigned long addr = (unsigned long)&csa->lscsa->ls[0]; 13887c038749SMark Nutter unsigned int ls_offset = 0x0; 13897c038749SMark Nutter unsigned int size = 16384; 13907c038749SMark Nutter unsigned int tag = 0; 13917c038749SMark Nutter unsigned int rclass = 0; 13927c038749SMark Nutter unsigned int cmd = MFC_GET_CMD; 13937c038749SMark Nutter 13947c038749SMark Nutter /* Restore, Step 44: 13957c038749SMark Nutter * Issue a DMA command to restore the first 13967c038749SMark Nutter * 16kb of local storage from CSA. 13977c038749SMark Nutter */ 13987c038749SMark Nutter send_mfc_dma(spu, addr, ls_offset, size, tag, rclass, cmd); 13997c038749SMark Nutter } 14007c038749SMark Nutter 14017c038749SMark Nutter static inline void clear_interrupts(struct spu_state *csa, struct spu *spu) 14027c038749SMark Nutter { 14037c038749SMark Nutter /* Restore, Step 49: 14047c038749SMark Nutter * Write INT_MASK_class0 with value of 0. 14057c038749SMark Nutter * Write INT_MASK_class1 with value of 0. 14067c038749SMark Nutter * Write INT_MASK_class2 with value of 0. 14077c038749SMark Nutter * Write INT_STAT_class0 with value of -1. 14087c038749SMark Nutter * Write INT_STAT_class1 with value of -1. 14097c038749SMark Nutter * Write INT_STAT_class2 with value of -1. 14107c038749SMark Nutter */ 14117c038749SMark Nutter spin_lock_irq(&spu->register_lock); 1412f0831accSArnd Bergmann spu_int_mask_set(spu, 0, 0ul); 1413f0831accSArnd Bergmann spu_int_mask_set(spu, 1, 0ul); 1414f0831accSArnd Bergmann spu_int_mask_set(spu, 2, 0ul); 1415f0831accSArnd Bergmann spu_int_stat_clear(spu, 0, ~0ul); 1416f0831accSArnd Bergmann spu_int_stat_clear(spu, 1, ~0ul); 1417f0831accSArnd Bergmann spu_int_stat_clear(spu, 2, ~0ul); 14187c038749SMark Nutter spin_unlock_irq(&spu->register_lock); 14197c038749SMark Nutter } 14207c038749SMark Nutter 14217c038749SMark Nutter static inline void restore_mfc_queues(struct spu_state *csa, struct spu *spu) 14227c038749SMark Nutter { 14237c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 14247c038749SMark Nutter int i; 14257c038749SMark Nutter 14267c038749SMark Nutter /* Restore, Step 50: 14277c038749SMark Nutter * If MFC_Cntl[Se]!=0 then restore 14287c038749SMark Nutter * MFC command queues. 14297c038749SMark Nutter */ 14307c038749SMark Nutter if ((csa->priv2.mfc_control_RW & MFC_CNTL_DMA_QUEUES_EMPTY_MASK) == 0) { 14317c038749SMark Nutter for (i = 0; i < 8; i++) { 14327c038749SMark Nutter out_be64(&priv2->puq[i].mfc_cq_data0_RW, 14337c038749SMark Nutter csa->priv2.puq[i].mfc_cq_data0_RW); 14347c038749SMark Nutter out_be64(&priv2->puq[i].mfc_cq_data1_RW, 14357c038749SMark Nutter csa->priv2.puq[i].mfc_cq_data1_RW); 14367c038749SMark Nutter out_be64(&priv2->puq[i].mfc_cq_data2_RW, 14377c038749SMark Nutter csa->priv2.puq[i].mfc_cq_data2_RW); 14387c038749SMark Nutter out_be64(&priv2->puq[i].mfc_cq_data3_RW, 14397c038749SMark Nutter csa->priv2.puq[i].mfc_cq_data3_RW); 14407c038749SMark Nutter } 14417c038749SMark Nutter for (i = 0; i < 16; i++) { 14427c038749SMark Nutter out_be64(&priv2->spuq[i].mfc_cq_data0_RW, 14437c038749SMark Nutter csa->priv2.spuq[i].mfc_cq_data0_RW); 14447c038749SMark Nutter out_be64(&priv2->spuq[i].mfc_cq_data1_RW, 14457c038749SMark Nutter csa->priv2.spuq[i].mfc_cq_data1_RW); 14467c038749SMark Nutter out_be64(&priv2->spuq[i].mfc_cq_data2_RW, 14477c038749SMark Nutter csa->priv2.spuq[i].mfc_cq_data2_RW); 14487c038749SMark Nutter out_be64(&priv2->spuq[i].mfc_cq_data3_RW, 14497c038749SMark Nutter csa->priv2.spuq[i].mfc_cq_data3_RW); 14507c038749SMark Nutter } 14517c038749SMark Nutter } 14527c038749SMark Nutter eieio(); 14537c038749SMark Nutter } 14547c038749SMark Nutter 14557c038749SMark Nutter static inline void restore_ppu_querymask(struct spu_state *csa, struct spu *spu) 14567c038749SMark Nutter { 14577c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 14587c038749SMark Nutter 14597c038749SMark Nutter /* Restore, Step 51: 14607c038749SMark Nutter * Restore the PPU_QueryMask register from CSA. 14617c038749SMark Nutter */ 14627c038749SMark Nutter out_be32(&prob->dma_querymask_RW, csa->prob.dma_querymask_RW); 14637c038749SMark Nutter eieio(); 14647c038749SMark Nutter } 14657c038749SMark Nutter 14667c038749SMark Nutter static inline void restore_ppu_querytype(struct spu_state *csa, struct spu *spu) 14677c038749SMark Nutter { 14687c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 14697c038749SMark Nutter 14707c038749SMark Nutter /* Restore, Step 52: 14717c038749SMark Nutter * Restore the PPU_QueryType register from CSA. 14727c038749SMark Nutter */ 14737c038749SMark Nutter out_be32(&prob->dma_querytype_RW, csa->prob.dma_querytype_RW); 14747c038749SMark Nutter eieio(); 14757c038749SMark Nutter } 14767c038749SMark Nutter 14777c038749SMark Nutter static inline void restore_mfc_csr_tsq(struct spu_state *csa, struct spu *spu) 14787c038749SMark Nutter { 14797c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 14807c038749SMark Nutter 14817c038749SMark Nutter /* Restore, Step 53: 14827c038749SMark Nutter * Restore the MFC_CSR_TSQ register from CSA. 14837c038749SMark Nutter */ 14847c038749SMark Nutter out_be64(&priv2->spu_tag_status_query_RW, 14857c038749SMark Nutter csa->priv2.spu_tag_status_query_RW); 14867c038749SMark Nutter eieio(); 14877c038749SMark Nutter } 14887c038749SMark Nutter 14897c038749SMark Nutter static inline void restore_mfc_csr_cmd(struct spu_state *csa, struct spu *spu) 14907c038749SMark Nutter { 14917c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 14927c038749SMark Nutter 14937c038749SMark Nutter /* Restore, Step 54: 14947c038749SMark Nutter * Restore the MFC_CSR_CMD1 and MFC_CSR_CMD2 14957c038749SMark Nutter * registers from CSA. 14967c038749SMark Nutter */ 14977c038749SMark Nutter out_be64(&priv2->spu_cmd_buf1_RW, csa->priv2.spu_cmd_buf1_RW); 14987c038749SMark Nutter out_be64(&priv2->spu_cmd_buf2_RW, csa->priv2.spu_cmd_buf2_RW); 14997c038749SMark Nutter eieio(); 15007c038749SMark Nutter } 15017c038749SMark Nutter 15027c038749SMark Nutter static inline void restore_mfc_csr_ato(struct spu_state *csa, struct spu *spu) 15037c038749SMark Nutter { 15047c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 15057c038749SMark Nutter 15067c038749SMark Nutter /* Restore, Step 55: 15077c038749SMark Nutter * Restore the MFC_CSR_ATO register from CSA. 15087c038749SMark Nutter */ 15097c038749SMark Nutter out_be64(&priv2->spu_atomic_status_RW, csa->priv2.spu_atomic_status_RW); 15107c038749SMark Nutter } 15117c038749SMark Nutter 15127c038749SMark Nutter static inline void restore_mfc_tclass_id(struct spu_state *csa, struct spu *spu) 15137c038749SMark Nutter { 15147c038749SMark Nutter /* Restore, Step 56: 15157c038749SMark Nutter * Restore the MFC_TCLASS_ID register from CSA. 15167c038749SMark Nutter */ 1517f0831accSArnd Bergmann spu_mfc_tclass_id_set(spu, csa->priv1.mfc_tclass_id_RW); 15187c038749SMark Nutter eieio(); 15197c038749SMark Nutter } 15207c038749SMark Nutter 15217c038749SMark Nutter static inline void set_llr_event(struct spu_state *csa, struct spu *spu) 15227c038749SMark Nutter { 15237c038749SMark Nutter u64 ch0_cnt, ch0_data; 15247c038749SMark Nutter u64 ch1_data; 15257c038749SMark Nutter 15267c038749SMark Nutter /* Restore, Step 57: 15277c038749SMark Nutter * Set the Lock Line Reservation Lost Event by: 15287c038749SMark Nutter * 1. OR CSA.SPU_Event_Status with bit 21 (Lr) set to 1. 15297c038749SMark Nutter * 2. If CSA.SPU_Channel_0_Count=0 and 15307c038749SMark Nutter * CSA.SPU_Wr_Event_Mask[Lr]=1 and 15317c038749SMark Nutter * CSA.SPU_Event_Status[Lr]=0 then set 15327c038749SMark Nutter * CSA.SPU_Event_Status_Count=1. 15337c038749SMark Nutter */ 15347c038749SMark Nutter ch0_cnt = csa->spu_chnlcnt_RW[0]; 15357c038749SMark Nutter ch0_data = csa->spu_chnldata_RW[0]; 15367c038749SMark Nutter ch1_data = csa->spu_chnldata_RW[1]; 15377c038749SMark Nutter csa->spu_chnldata_RW[0] |= MFC_LLR_LOST_EVENT; 15387c038749SMark Nutter if ((ch0_cnt == 0) && !(ch0_data & MFC_LLR_LOST_EVENT) && 15397c038749SMark Nutter (ch1_data & MFC_LLR_LOST_EVENT)) { 15407c038749SMark Nutter csa->spu_chnlcnt_RW[0] = 1; 15417c038749SMark Nutter } 15427c038749SMark Nutter } 15437c038749SMark Nutter 15447c038749SMark Nutter static inline void restore_decr_wrapped(struct spu_state *csa, struct spu *spu) 15457c038749SMark Nutter { 15467c038749SMark Nutter /* Restore, Step 58: 15477c038749SMark Nutter * If the status of the CSA software decrementer 15487c038749SMark Nutter * "wrapped" flag is set, OR in a '1' to 15497c038749SMark Nutter * CSA.SPU_Event_Status[Tm]. 15507c038749SMark Nutter */ 15517c038749SMark Nutter if (csa->lscsa->decr_status.slot[0] == 1) { 15527c038749SMark Nutter csa->spu_chnldata_RW[0] |= 0x20; 15537c038749SMark Nutter } 15547c038749SMark Nutter if ((csa->lscsa->decr_status.slot[0] == 1) && 15557c038749SMark Nutter (csa->spu_chnlcnt_RW[0] == 0 && 15567c038749SMark Nutter ((csa->spu_chnldata_RW[2] & 0x20) == 0x0) && 15577c038749SMark Nutter ((csa->spu_chnldata_RW[0] & 0x20) != 0x1))) { 15587c038749SMark Nutter csa->spu_chnlcnt_RW[0] = 1; 15597c038749SMark Nutter } 15607c038749SMark Nutter } 15617c038749SMark Nutter 15627c038749SMark Nutter static inline void restore_ch_part1(struct spu_state *csa, struct spu *spu) 15637c038749SMark Nutter { 15647c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 1565e46a0237Sarnd@arndb.de u64 idx, ch_indices[7] = { 0UL, 3UL, 4UL, 24UL, 25UL, 27UL }; 15667c038749SMark Nutter int i; 15677c038749SMark Nutter 15687c038749SMark Nutter /* Restore, Step 59: 15697c038749SMark Nutter */ 1570e46a0237Sarnd@arndb.de 1571e46a0237Sarnd@arndb.de /* Restore CH 1 without count */ 1572e46a0237Sarnd@arndb.de out_be64(&priv2->spu_chnlcntptr_RW, 1); 1573e46a0237Sarnd@arndb.de out_be64(&priv2->spu_chnldata_RW, csa->spu_chnldata_RW[1]); 1574e46a0237Sarnd@arndb.de 1575e46a0237Sarnd@arndb.de /* Restore the following CH: [0,3,4,24,25,27] */ 15767c038749SMark Nutter for (i = 0; i < 7; i++) { 15777c038749SMark Nutter idx = ch_indices[i]; 15787c038749SMark Nutter out_be64(&priv2->spu_chnlcntptr_RW, idx); 15797c038749SMark Nutter eieio(); 15807c038749SMark Nutter out_be64(&priv2->spu_chnldata_RW, csa->spu_chnldata_RW[idx]); 15817c038749SMark Nutter out_be64(&priv2->spu_chnlcnt_RW, csa->spu_chnlcnt_RW[idx]); 15827c038749SMark Nutter eieio(); 15837c038749SMark Nutter } 15847c038749SMark Nutter } 15857c038749SMark Nutter 15867c038749SMark Nutter static inline void restore_ch_part2(struct spu_state *csa, struct spu *spu) 15877c038749SMark Nutter { 15887c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 15897c038749SMark Nutter u64 ch_indices[3] = { 9UL, 21UL, 23UL }; 15907c038749SMark Nutter u64 ch_counts[3] = { 1UL, 16UL, 1UL }; 15917c038749SMark Nutter u64 idx; 15927c038749SMark Nutter int i; 15937c038749SMark Nutter 15947c038749SMark Nutter /* Restore, Step 60: 15957c038749SMark Nutter * Restore the following CH: [9,21,23]. 15967c038749SMark Nutter */ 15977c038749SMark Nutter ch_counts[0] = 1UL; 15987c038749SMark Nutter ch_counts[1] = csa->spu_chnlcnt_RW[21]; 15997c038749SMark Nutter ch_counts[2] = 1UL; 16007c038749SMark Nutter for (i = 0; i < 3; i++) { 16017c038749SMark Nutter idx = ch_indices[i]; 16027c038749SMark Nutter out_be64(&priv2->spu_chnlcntptr_RW, idx); 16037c038749SMark Nutter eieio(); 16047c038749SMark Nutter out_be64(&priv2->spu_chnlcnt_RW, ch_counts[i]); 16057c038749SMark Nutter eieio(); 16067c038749SMark Nutter } 16077c038749SMark Nutter } 16087c038749SMark Nutter 16097c038749SMark Nutter static inline void restore_spu_lslr(struct spu_state *csa, struct spu *spu) 16107c038749SMark Nutter { 16117c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 16127c038749SMark Nutter 16137c038749SMark Nutter /* Restore, Step 61: 16147c038749SMark Nutter * Restore the SPU_LSLR register from CSA. 16157c038749SMark Nutter */ 16167c038749SMark Nutter out_be64(&priv2->spu_lslr_RW, csa->priv2.spu_lslr_RW); 16177c038749SMark Nutter eieio(); 16187c038749SMark Nutter } 16197c038749SMark Nutter 16207c038749SMark Nutter static inline void restore_spu_cfg(struct spu_state *csa, struct spu *spu) 16217c038749SMark Nutter { 16227c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 16237c038749SMark Nutter 16247c038749SMark Nutter /* Restore, Step 62: 16257c038749SMark Nutter * Restore the SPU_Cfg register from CSA. 16267c038749SMark Nutter */ 16277c038749SMark Nutter out_be64(&priv2->spu_cfg_RW, csa->priv2.spu_cfg_RW); 16287c038749SMark Nutter eieio(); 16297c038749SMark Nutter } 16307c038749SMark Nutter 16317c038749SMark Nutter static inline void restore_pm_trace(struct spu_state *csa, struct spu *spu) 16327c038749SMark Nutter { 16337c038749SMark Nutter /* Restore, Step 63: 16347c038749SMark Nutter * Restore PM_Trace_Tag_Wait_Mask from CSA. 16357c038749SMark Nutter * Not performed by this implementation. 16367c038749SMark Nutter */ 16377c038749SMark Nutter } 16387c038749SMark Nutter 16397c038749SMark Nutter static inline void restore_spu_npc(struct spu_state *csa, struct spu *spu) 16407c038749SMark Nutter { 16417c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 16427c038749SMark Nutter 16437c038749SMark Nutter /* Restore, Step 64: 16447c038749SMark Nutter * Restore SPU_NPC from CSA. 16457c038749SMark Nutter */ 16467c038749SMark Nutter out_be32(&prob->spu_npc_RW, csa->prob.spu_npc_RW); 16477c038749SMark Nutter eieio(); 16487c038749SMark Nutter } 16497c038749SMark Nutter 16507c038749SMark Nutter static inline void restore_spu_mb(struct spu_state *csa, struct spu *spu) 16517c038749SMark Nutter { 16527c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 16537c038749SMark Nutter int i; 16547c038749SMark Nutter 16557c038749SMark Nutter /* Restore, Step 65: 16567c038749SMark Nutter * Restore MFC_RdSPU_MB from CSA. 16577c038749SMark Nutter */ 16587c038749SMark Nutter out_be64(&priv2->spu_chnlcntptr_RW, 29UL); 16597c038749SMark Nutter eieio(); 16607c038749SMark Nutter out_be64(&priv2->spu_chnlcnt_RW, csa->spu_chnlcnt_RW[29]); 16617c038749SMark Nutter for (i = 0; i < 4; i++) { 16628b3d6663SArnd Bergmann out_be64(&priv2->spu_chnldata_RW, csa->spu_mailbox_data[i]); 16637c038749SMark Nutter } 16647c038749SMark Nutter eieio(); 16657c038749SMark Nutter } 16667c038749SMark Nutter 16677c038749SMark Nutter static inline void check_ppu_mb_stat(struct spu_state *csa, struct spu *spu) 16687c038749SMark Nutter { 16697c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 16707c038749SMark Nutter u32 dummy = 0; 16717c038749SMark Nutter 16727c038749SMark Nutter /* Restore, Step 66: 16737c038749SMark Nutter * If CSA.MB_Stat[P]=0 (mailbox empty) then 16747c038749SMark Nutter * read from the PPU_MB register. 16757c038749SMark Nutter */ 16767c038749SMark Nutter if ((csa->prob.mb_stat_R & 0xFF) == 0) { 16777c038749SMark Nutter dummy = in_be32(&prob->pu_mb_R); 16787c038749SMark Nutter eieio(); 16797c038749SMark Nutter } 16807c038749SMark Nutter } 16817c038749SMark Nutter 16827c038749SMark Nutter static inline void check_ppuint_mb_stat(struct spu_state *csa, struct spu *spu) 16837c038749SMark Nutter { 16847c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 16857c038749SMark Nutter u64 dummy = 0UL; 16867c038749SMark Nutter 16877c038749SMark Nutter /* Restore, Step 66: 16887c038749SMark Nutter * If CSA.MB_Stat[I]=0 (mailbox empty) then 16897c038749SMark Nutter * read from the PPUINT_MB register. 16907c038749SMark Nutter */ 16917c038749SMark Nutter if ((csa->prob.mb_stat_R & 0xFF0000) == 0) { 16927c038749SMark Nutter dummy = in_be64(&priv2->puint_mb_R); 16937c038749SMark Nutter eieio(); 1694f0831accSArnd Bergmann spu_int_stat_clear(spu, 2, CLASS2_ENABLE_MAILBOX_INTR); 16957c038749SMark Nutter eieio(); 16967c038749SMark Nutter } 16977c038749SMark Nutter } 16987c038749SMark Nutter 16997c038749SMark Nutter static inline void restore_mfc_sr1(struct spu_state *csa, struct spu *spu) 17007c038749SMark Nutter { 17017c038749SMark Nutter /* Restore, Step 69: 17027c038749SMark Nutter * Restore the MFC_SR1 register from CSA. 17037c038749SMark Nutter */ 1704f0831accSArnd Bergmann spu_mfc_sr1_set(spu, csa->priv1.mfc_sr1_RW); 17057c038749SMark Nutter eieio(); 17067c038749SMark Nutter } 17077c038749SMark Nutter 17087c038749SMark Nutter static inline void restore_other_spu_access(struct spu_state *csa, 17097c038749SMark Nutter struct spu *spu) 17107c038749SMark Nutter { 17117c038749SMark Nutter /* Restore, Step 70: 17127c038749SMark Nutter * Restore other SPU mappings to this SPU. TBD. 17137c038749SMark Nutter */ 17147c038749SMark Nutter } 17157c038749SMark Nutter 17167c038749SMark Nutter static inline void restore_spu_runcntl(struct spu_state *csa, struct spu *spu) 17177c038749SMark Nutter { 17187c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 17197c038749SMark Nutter 17207c038749SMark Nutter /* Restore, Step 71: 17217c038749SMark Nutter * If CSA.SPU_Status[R]=1 then write 17227c038749SMark Nutter * SPU_RunCntl[R0R1]='01'. 17237c038749SMark Nutter */ 17247c038749SMark Nutter if (csa->prob.spu_status_R & SPU_STATUS_RUNNING) { 17257c038749SMark Nutter out_be32(&prob->spu_runcntl_RW, SPU_RUNCNTL_RUNNABLE); 17267c038749SMark Nutter eieio(); 17277c038749SMark Nutter } 17287c038749SMark Nutter } 17297c038749SMark Nutter 17307c038749SMark Nutter static inline void restore_mfc_cntl(struct spu_state *csa, struct spu *spu) 17317c038749SMark Nutter { 17327c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 17337c038749SMark Nutter 17347c038749SMark Nutter /* Restore, Step 72: 17357c038749SMark Nutter * Restore the MFC_CNTL register for the CSA. 17367c038749SMark Nutter */ 17377c038749SMark Nutter out_be64(&priv2->mfc_control_RW, csa->priv2.mfc_control_RW); 17387c038749SMark Nutter eieio(); 173928347bceSHyeonSeung Jang /* 174028347bceSHyeonSeung Jang * FIXME: this is to restart a DMA that we were processing 174128347bceSHyeonSeung Jang * before the save. better remember the fault information 174228347bceSHyeonSeung Jang * in the csa instead. 174328347bceSHyeonSeung Jang */ 174428347bceSHyeonSeung Jang if ((csa->priv2.mfc_control_RW & MFC_CNTL_SUSPEND_DMA_QUEUE_MASK)) { 174528347bceSHyeonSeung Jang out_be64(&priv2->mfc_control_RW, MFC_CNTL_RESTART_DMA_COMMAND); 174628347bceSHyeonSeung Jang eieio(); 174728347bceSHyeonSeung Jang } 17487c038749SMark Nutter } 17497c038749SMark Nutter 17507c038749SMark Nutter static inline void enable_user_access(struct spu_state *csa, struct spu *spu) 17517c038749SMark Nutter { 17527c038749SMark Nutter /* Restore, Step 73: 17537c038749SMark Nutter * Enable user-space access (if provided) to this 17547c038749SMark Nutter * SPU by mapping the virtual pages assigned to 17557c038749SMark Nutter * the SPU memory-mapped I/O (MMIO) for problem 17567c038749SMark Nutter * state. TBD. 17577c038749SMark Nutter */ 17587c038749SMark Nutter } 17597c038749SMark Nutter 17607c038749SMark Nutter static inline void reset_switch_active(struct spu_state *csa, struct spu *spu) 17617c038749SMark Nutter { 17627c038749SMark Nutter /* Restore, Step 74: 17637c038749SMark Nutter * Reset the "context switch active" flag. 17647c038749SMark Nutter */ 17658837d921SArnd Bergmann clear_bit(SPU_CONTEXT_SWITCH_ACTIVE, &spu->flags); 17667c038749SMark Nutter mb(); 17677c038749SMark Nutter } 17687c038749SMark Nutter 17697c038749SMark Nutter static inline void reenable_interrupts(struct spu_state *csa, struct spu *spu) 17707c038749SMark Nutter { 17717c038749SMark Nutter /* Restore, Step 75: 17727c038749SMark Nutter * Re-enable SPU interrupts. 17737c038749SMark Nutter */ 17747c038749SMark Nutter spin_lock_irq(&spu->register_lock); 1775f0831accSArnd Bergmann spu_int_mask_set(spu, 0, csa->priv1.int_mask_class0_RW); 1776f0831accSArnd Bergmann spu_int_mask_set(spu, 1, csa->priv1.int_mask_class1_RW); 1777f0831accSArnd Bergmann spu_int_mask_set(spu, 2, csa->priv1.int_mask_class2_RW); 17787c038749SMark Nutter spin_unlock_irq(&spu->register_lock); 17797c038749SMark Nutter } 17807c038749SMark Nutter 17817c038749SMark Nutter static int quiece_spu(struct spu_state *prev, struct spu *spu) 17827c038749SMark Nutter { 17837c038749SMark Nutter /* 17847c038749SMark Nutter * Combined steps 2-18 of SPU context save sequence, which 17857c038749SMark Nutter * quiesce the SPU state (disable SPU execution, MFC command 17867c038749SMark Nutter * queues, decrementer, SPU interrupts, etc.). 17877c038749SMark Nutter * 17887c038749SMark Nutter * Returns 0 on success. 17897c038749SMark Nutter * 2 if failed step 2. 17907c038749SMark Nutter * 6 if failed step 6. 17917c038749SMark Nutter */ 17927c038749SMark Nutter 17937c038749SMark Nutter if (check_spu_isolate(prev, spu)) { /* Step 2. */ 17947c038749SMark Nutter return 2; 17957c038749SMark Nutter } 17967c038749SMark Nutter disable_interrupts(prev, spu); /* Step 3. */ 17977c038749SMark Nutter set_watchdog_timer(prev, spu); /* Step 4. */ 17987c038749SMark Nutter inhibit_user_access(prev, spu); /* Step 5. */ 17997c038749SMark Nutter if (check_spu_isolate(prev, spu)) { /* Step 6. */ 18007c038749SMark Nutter return 6; 18017c038749SMark Nutter } 18027c038749SMark Nutter set_switch_pending(prev, spu); /* Step 7. */ 18037c038749SMark Nutter save_mfc_cntl(prev, spu); /* Step 8. */ 18047c038749SMark Nutter save_spu_runcntl(prev, spu); /* Step 9. */ 18057c038749SMark Nutter save_mfc_sr1(prev, spu); /* Step 10. */ 18067c038749SMark Nutter save_spu_status(prev, spu); /* Step 11. */ 18077c038749SMark Nutter save_mfc_decr(prev, spu); /* Step 12. */ 18087c038749SMark Nutter halt_mfc_decr(prev, spu); /* Step 13. */ 18097c038749SMark Nutter save_timebase(prev, spu); /* Step 14. */ 18107c038749SMark Nutter remove_other_spu_access(prev, spu); /* Step 15. */ 18117c038749SMark Nutter do_mfc_mssync(prev, spu); /* Step 16. */ 18127c038749SMark Nutter issue_mfc_tlbie(prev, spu); /* Step 17. */ 18137c038749SMark Nutter handle_pending_interrupts(prev, spu); /* Step 18. */ 18147c038749SMark Nutter 18157c038749SMark Nutter return 0; 18167c038749SMark Nutter } 18177c038749SMark Nutter 18187c038749SMark Nutter static void save_csa(struct spu_state *prev, struct spu *spu) 18197c038749SMark Nutter { 18207c038749SMark Nutter /* 18217c038749SMark Nutter * Combine steps 19-44 of SPU context save sequence, which 18227c038749SMark Nutter * save regions of the privileged & problem state areas. 18237c038749SMark Nutter */ 18247c038749SMark Nutter 18257c038749SMark Nutter save_mfc_queues(prev, spu); /* Step 19. */ 18267c038749SMark Nutter save_ppu_querymask(prev, spu); /* Step 20. */ 18277c038749SMark Nutter save_ppu_querytype(prev, spu); /* Step 21. */ 18288d038e04SKazunori Asayama save_ppu_tagstatus(prev, spu); /* NEW. */ 18297c038749SMark Nutter save_mfc_csr_tsq(prev, spu); /* Step 22. */ 18307c038749SMark Nutter save_mfc_csr_cmd(prev, spu); /* Step 23. */ 18317c038749SMark Nutter save_mfc_csr_ato(prev, spu); /* Step 24. */ 18327c038749SMark Nutter save_mfc_tclass_id(prev, spu); /* Step 25. */ 18337c038749SMark Nutter set_mfc_tclass_id(prev, spu); /* Step 26. */ 18347c038749SMark Nutter purge_mfc_queue(prev, spu); /* Step 27. */ 18357c038749SMark Nutter wait_purge_complete(prev, spu); /* Step 28. */ 18367c038749SMark Nutter setup_mfc_sr1(prev, spu); /* Step 30. */ 18377c038749SMark Nutter save_spu_npc(prev, spu); /* Step 31. */ 18387c038749SMark Nutter save_spu_privcntl(prev, spu); /* Step 32. */ 18397c038749SMark Nutter reset_spu_privcntl(prev, spu); /* Step 33. */ 18407c038749SMark Nutter save_spu_lslr(prev, spu); /* Step 34. */ 18417c038749SMark Nutter reset_spu_lslr(prev, spu); /* Step 35. */ 18427c038749SMark Nutter save_spu_cfg(prev, spu); /* Step 36. */ 18437c038749SMark Nutter save_pm_trace(prev, spu); /* Step 37. */ 18447c038749SMark Nutter save_mfc_rag(prev, spu); /* Step 38. */ 18457c038749SMark Nutter save_ppu_mb_stat(prev, spu); /* Step 39. */ 18467c038749SMark Nutter save_ppu_mb(prev, spu); /* Step 40. */ 18477c038749SMark Nutter save_ppuint_mb(prev, spu); /* Step 41. */ 18487c038749SMark Nutter save_ch_part1(prev, spu); /* Step 42. */ 18497c038749SMark Nutter save_spu_mb(prev, spu); /* Step 43. */ 18507c038749SMark Nutter save_mfc_cmd(prev, spu); /* Step 44. */ 18517c038749SMark Nutter reset_ch(prev, spu); /* Step 45. */ 18527c038749SMark Nutter } 18537c038749SMark Nutter 18547c038749SMark Nutter static void save_lscsa(struct spu_state *prev, struct spu *spu) 18557c038749SMark Nutter { 18567c038749SMark Nutter /* 18577c038749SMark Nutter * Perform steps 46-57 of SPU context save sequence, 18587c038749SMark Nutter * which save regions of the local store and register 18597c038749SMark Nutter * file. 18607c038749SMark Nutter */ 18617c038749SMark Nutter 18627c038749SMark Nutter resume_mfc_queue(prev, spu); /* Step 46. */ 18637c038749SMark Nutter setup_mfc_slbs(prev, spu); /* Step 47. */ 18647c038749SMark Nutter set_switch_active(prev, spu); /* Step 48. */ 18657c038749SMark Nutter enable_interrupts(prev, spu); /* Step 49. */ 18667c038749SMark Nutter save_ls_16kb(prev, spu); /* Step 50. */ 18677c038749SMark Nutter set_spu_npc(prev, spu); /* Step 51. */ 18687c038749SMark Nutter set_signot1(prev, spu); /* Step 52. */ 18697c038749SMark Nutter set_signot2(prev, spu); /* Step 53. */ 18707c038749SMark Nutter send_save_code(prev, spu); /* Step 54. */ 18717c038749SMark Nutter set_ppu_querymask(prev, spu); /* Step 55. */ 18727c038749SMark Nutter wait_tag_complete(prev, spu); /* Step 56. */ 18737c038749SMark Nutter wait_spu_stopped(prev, spu); /* Step 57. */ 18747c038749SMark Nutter } 18757c038749SMark Nutter 18765737edd1SMark Nutter static void force_spu_isolate_exit(struct spu *spu) 18775737edd1SMark Nutter { 18785737edd1SMark Nutter struct spu_problem __iomem *prob = spu->problem; 18795737edd1SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 18805737edd1SMark Nutter 18815737edd1SMark Nutter /* Stop SPE execution and wait for completion. */ 18825737edd1SMark Nutter out_be32(&prob->spu_runcntl_RW, SPU_RUNCNTL_STOP); 18835737edd1SMark Nutter iobarrier_rw(); 18845737edd1SMark Nutter POLL_WHILE_TRUE(in_be32(&prob->spu_status_R) & SPU_STATUS_RUNNING); 18855737edd1SMark Nutter 18865737edd1SMark Nutter /* Restart SPE master runcntl. */ 18875737edd1SMark Nutter spu_mfc_sr1_set(spu, MFC_STATE1_MASTER_RUN_CONTROL_MASK); 18885737edd1SMark Nutter iobarrier_w(); 18895737edd1SMark Nutter 18905737edd1SMark Nutter /* Initiate isolate exit request and wait for completion. */ 18915737edd1SMark Nutter out_be64(&priv2->spu_privcntl_RW, 4LL); 18925737edd1SMark Nutter iobarrier_w(); 18935737edd1SMark Nutter out_be32(&prob->spu_runcntl_RW, 2); 18945737edd1SMark Nutter iobarrier_rw(); 18955737edd1SMark Nutter POLL_WHILE_FALSE((in_be32(&prob->spu_status_R) 18965737edd1SMark Nutter & SPU_STATUS_STOPPED_BY_STOP)); 18975737edd1SMark Nutter 18985737edd1SMark Nutter /* Reset load request to normal. */ 18995737edd1SMark Nutter out_be64(&priv2->spu_privcntl_RW, SPU_PRIVCNT_LOAD_REQUEST_NORMAL); 19005737edd1SMark Nutter iobarrier_w(); 19015737edd1SMark Nutter } 19025737edd1SMark Nutter 19035737edd1SMark Nutter /** 19045737edd1SMark Nutter * stop_spu_isolate 19055737edd1SMark Nutter * Check SPU run-control state and force isolated 19065737edd1SMark Nutter * exit function as necessary. 19075737edd1SMark Nutter */ 19085737edd1SMark Nutter static void stop_spu_isolate(struct spu *spu) 19095737edd1SMark Nutter { 19105737edd1SMark Nutter struct spu_problem __iomem *prob = spu->problem; 19115737edd1SMark Nutter 19125737edd1SMark Nutter if (in_be32(&prob->spu_status_R) & SPU_STATUS_ISOLATED_STATE) { 19135737edd1SMark Nutter /* The SPU is in isolated state; the only way 19145737edd1SMark Nutter * to get it out is to perform an isolated 19155737edd1SMark Nutter * exit (clean) operation. 19165737edd1SMark Nutter */ 19175737edd1SMark Nutter force_spu_isolate_exit(spu); 19185737edd1SMark Nutter } 19195737edd1SMark Nutter } 19205737edd1SMark Nutter 19217c038749SMark Nutter static void harvest(struct spu_state *prev, struct spu *spu) 19227c038749SMark Nutter { 19237c038749SMark Nutter /* 19247c038749SMark Nutter * Perform steps 2-25 of SPU context restore sequence, 19257c038749SMark Nutter * which resets an SPU either after a failed save, or 19267c038749SMark Nutter * when using SPU for first time. 19277c038749SMark Nutter */ 19287c038749SMark Nutter 19297c038749SMark Nutter disable_interrupts(prev, spu); /* Step 2. */ 19307c038749SMark Nutter inhibit_user_access(prev, spu); /* Step 3. */ 19317c038749SMark Nutter terminate_spu_app(prev, spu); /* Step 4. */ 19327c038749SMark Nutter set_switch_pending(prev, spu); /* Step 5. */ 19335737edd1SMark Nutter stop_spu_isolate(spu); /* NEW. */ 19347c038749SMark Nutter remove_other_spu_access(prev, spu); /* Step 6. */ 19357c038749SMark Nutter suspend_mfc(prev, spu); /* Step 7. */ 19367c038749SMark Nutter wait_suspend_mfc_complete(prev, spu); /* Step 8. */ 19377c038749SMark Nutter if (!suspend_spe(prev, spu)) /* Step 9. */ 19387c038749SMark Nutter clear_spu_status(prev, spu); /* Step 10. */ 19397c038749SMark Nutter do_mfc_mssync(prev, spu); /* Step 11. */ 19407c038749SMark Nutter issue_mfc_tlbie(prev, spu); /* Step 12. */ 19417c038749SMark Nutter handle_pending_interrupts(prev, spu); /* Step 13. */ 19427c038749SMark Nutter purge_mfc_queue(prev, spu); /* Step 14. */ 19437c038749SMark Nutter wait_purge_complete(prev, spu); /* Step 15. */ 19447c038749SMark Nutter reset_spu_privcntl(prev, spu); /* Step 16. */ 19457c038749SMark Nutter reset_spu_lslr(prev, spu); /* Step 17. */ 19467c038749SMark Nutter setup_mfc_sr1(prev, spu); /* Step 18. */ 194794b2a439SBenjamin Herrenschmidt spu_invalidate_slbs(spu); /* Step 19. */ 19487c038749SMark Nutter reset_ch_part1(prev, spu); /* Step 20. */ 19497c038749SMark Nutter reset_ch_part2(prev, spu); /* Step 21. */ 19507c038749SMark Nutter enable_interrupts(prev, spu); /* Step 22. */ 19517c038749SMark Nutter set_switch_active(prev, spu); /* Step 23. */ 19527c038749SMark Nutter set_mfc_tclass_id(prev, spu); /* Step 24. */ 19537c038749SMark Nutter resume_mfc_queue(prev, spu); /* Step 25. */ 19547c038749SMark Nutter } 19557c038749SMark Nutter 19567c038749SMark Nutter static void restore_lscsa(struct spu_state *next, struct spu *spu) 19577c038749SMark Nutter { 19587c038749SMark Nutter /* 19597c038749SMark Nutter * Perform steps 26-40 of SPU context restore sequence, 19607c038749SMark Nutter * which restores regions of the local store and register 19617c038749SMark Nutter * file. 19627c038749SMark Nutter */ 19637c038749SMark Nutter 19647c038749SMark Nutter set_watchdog_timer(next, spu); /* Step 26. */ 19657c038749SMark Nutter setup_spu_status_part1(next, spu); /* Step 27. */ 19667c038749SMark Nutter setup_spu_status_part2(next, spu); /* Step 28. */ 19677c038749SMark Nutter restore_mfc_rag(next, spu); /* Step 29. */ 19687c038749SMark Nutter setup_mfc_slbs(next, spu); /* Step 30. */ 19697c038749SMark Nutter set_spu_npc(next, spu); /* Step 31. */ 19707c038749SMark Nutter set_signot1(next, spu); /* Step 32. */ 19717c038749SMark Nutter set_signot2(next, spu); /* Step 33. */ 19727c038749SMark Nutter setup_decr(next, spu); /* Step 34. */ 19737c038749SMark Nutter setup_ppu_mb(next, spu); /* Step 35. */ 19747c038749SMark Nutter setup_ppuint_mb(next, spu); /* Step 36. */ 19757c038749SMark Nutter send_restore_code(next, spu); /* Step 37. */ 19767c038749SMark Nutter set_ppu_querymask(next, spu); /* Step 38. */ 19777c038749SMark Nutter wait_tag_complete(next, spu); /* Step 39. */ 19787c038749SMark Nutter wait_spu_stopped(next, spu); /* Step 40. */ 19797c038749SMark Nutter } 19807c038749SMark Nutter 19817c038749SMark Nutter static void restore_csa(struct spu_state *next, struct spu *spu) 19827c038749SMark Nutter { 19837c038749SMark Nutter /* 19847c038749SMark Nutter * Combine steps 41-76 of SPU context restore sequence, which 19857c038749SMark Nutter * restore regions of the privileged & problem state areas. 19867c038749SMark Nutter */ 19877c038749SMark Nutter 19887c038749SMark Nutter restore_spu_privcntl(next, spu); /* Step 41. */ 19897c038749SMark Nutter restore_status_part1(next, spu); /* Step 42. */ 19907c038749SMark Nutter restore_status_part2(next, spu); /* Step 43. */ 19917c038749SMark Nutter restore_ls_16kb(next, spu); /* Step 44. */ 19927c038749SMark Nutter wait_tag_complete(next, spu); /* Step 45. */ 19937c038749SMark Nutter suspend_mfc(next, spu); /* Step 46. */ 19947c038749SMark Nutter wait_suspend_mfc_complete(next, spu); /* Step 47. */ 19957c038749SMark Nutter issue_mfc_tlbie(next, spu); /* Step 48. */ 19967c038749SMark Nutter clear_interrupts(next, spu); /* Step 49. */ 19977c038749SMark Nutter restore_mfc_queues(next, spu); /* Step 50. */ 19987c038749SMark Nutter restore_ppu_querymask(next, spu); /* Step 51. */ 19997c038749SMark Nutter restore_ppu_querytype(next, spu); /* Step 52. */ 20007c038749SMark Nutter restore_mfc_csr_tsq(next, spu); /* Step 53. */ 20017c038749SMark Nutter restore_mfc_csr_cmd(next, spu); /* Step 54. */ 20027c038749SMark Nutter restore_mfc_csr_ato(next, spu); /* Step 55. */ 20037c038749SMark Nutter restore_mfc_tclass_id(next, spu); /* Step 56. */ 20047c038749SMark Nutter set_llr_event(next, spu); /* Step 57. */ 20057c038749SMark Nutter restore_decr_wrapped(next, spu); /* Step 58. */ 20067c038749SMark Nutter restore_ch_part1(next, spu); /* Step 59. */ 20077c038749SMark Nutter restore_ch_part2(next, spu); /* Step 60. */ 20087c038749SMark Nutter restore_spu_lslr(next, spu); /* Step 61. */ 20097c038749SMark Nutter restore_spu_cfg(next, spu); /* Step 62. */ 20107c038749SMark Nutter restore_pm_trace(next, spu); /* Step 63. */ 20117c038749SMark Nutter restore_spu_npc(next, spu); /* Step 64. */ 20127c038749SMark Nutter restore_spu_mb(next, spu); /* Step 65. */ 20137c038749SMark Nutter check_ppu_mb_stat(next, spu); /* Step 66. */ 20147c038749SMark Nutter check_ppuint_mb_stat(next, spu); /* Step 67. */ 201594b2a439SBenjamin Herrenschmidt spu_invalidate_slbs(spu); /* Modified Step 68. */ 20167c038749SMark Nutter restore_mfc_sr1(next, spu); /* Step 69. */ 20177c038749SMark Nutter restore_other_spu_access(next, spu); /* Step 70. */ 20187c038749SMark Nutter restore_spu_runcntl(next, spu); /* Step 71. */ 20197c038749SMark Nutter restore_mfc_cntl(next, spu); /* Step 72. */ 20207c038749SMark Nutter enable_user_access(next, spu); /* Step 73. */ 20217c038749SMark Nutter reset_switch_active(next, spu); /* Step 74. */ 20227c038749SMark Nutter reenable_interrupts(next, spu); /* Step 75. */ 20237c038749SMark Nutter } 20247c038749SMark Nutter 20257c038749SMark Nutter static int __do_spu_save(struct spu_state *prev, struct spu *spu) 20267c038749SMark Nutter { 20277c038749SMark Nutter int rc; 20287c038749SMark Nutter 20297c038749SMark Nutter /* 20307c038749SMark Nutter * SPU context save can be broken into three phases: 20317c038749SMark Nutter * 20327c038749SMark Nutter * (a) quiesce [steps 2-16]. 20337c038749SMark Nutter * (b) save of CSA, performed by PPE [steps 17-42] 20347c038749SMark Nutter * (c) save of LSCSA, mostly performed by SPU [steps 43-52]. 20357c038749SMark Nutter * 20367c038749SMark Nutter * Returns 0 on success. 20377c038749SMark Nutter * 2,6 if failed to quiece SPU 20387c038749SMark Nutter * 53 if SPU-side of save failed. 20397c038749SMark Nutter */ 20407c038749SMark Nutter 20417c038749SMark Nutter rc = quiece_spu(prev, spu); /* Steps 2-16. */ 20427c038749SMark Nutter switch (rc) { 20437c038749SMark Nutter default: 20447c038749SMark Nutter case 2: 20457c038749SMark Nutter case 6: 20467c038749SMark Nutter harvest(prev, spu); 20477c038749SMark Nutter return rc; 20487c038749SMark Nutter break; 20497c038749SMark Nutter case 0: 20507c038749SMark Nutter break; 20517c038749SMark Nutter } 20527c038749SMark Nutter save_csa(prev, spu); /* Steps 17-43. */ 20537c038749SMark Nutter save_lscsa(prev, spu); /* Steps 44-53. */ 20547c038749SMark Nutter return check_save_status(prev, spu); /* Step 54. */ 20557c038749SMark Nutter } 20567c038749SMark Nutter 20577c038749SMark Nutter static int __do_spu_restore(struct spu_state *next, struct spu *spu) 20587c038749SMark Nutter { 20597c038749SMark Nutter int rc; 20607c038749SMark Nutter 20617c038749SMark Nutter /* 20627c038749SMark Nutter * SPU context restore can be broken into three phases: 20637c038749SMark Nutter * 20647c038749SMark Nutter * (a) harvest (or reset) SPU [steps 2-24]. 20657c038749SMark Nutter * (b) restore LSCSA [steps 25-40], mostly performed by SPU. 20667c038749SMark Nutter * (c) restore CSA [steps 41-76], performed by PPE. 20677c038749SMark Nutter * 20687c038749SMark Nutter * The 'harvest' step is not performed here, but rather 20697c038749SMark Nutter * as needed below. 20707c038749SMark Nutter */ 20717c038749SMark Nutter 20727c038749SMark Nutter restore_lscsa(next, spu); /* Steps 24-39. */ 20737c038749SMark Nutter rc = check_restore_status(next, spu); /* Step 40. */ 20747c038749SMark Nutter switch (rc) { 20757c038749SMark Nutter default: 20767c038749SMark Nutter /* Failed. Return now. */ 20777c038749SMark Nutter return rc; 20787c038749SMark Nutter break; 20797c038749SMark Nutter case 0: 20807c038749SMark Nutter /* Fall through to next step. */ 20817c038749SMark Nutter break; 20827c038749SMark Nutter } 20837c038749SMark Nutter restore_csa(next, spu); 20847c038749SMark Nutter 20857c038749SMark Nutter return 0; 20867c038749SMark Nutter } 20877c038749SMark Nutter 20885473af04SMark Nutter /** 20895473af04SMark Nutter * spu_save - SPU context save, with locking. 20905473af04SMark Nutter * @prev: pointer to SPU context save area, to be saved. 20915473af04SMark Nutter * @spu: pointer to SPU iomem structure. 20925473af04SMark Nutter * 20935473af04SMark Nutter * Acquire locks, perform the save operation then return. 20945473af04SMark Nutter */ 20955473af04SMark Nutter int spu_save(struct spu_state *prev, struct spu *spu) 20965473af04SMark Nutter { 20977c038749SMark Nutter int rc; 20985473af04SMark Nutter 20997c038749SMark Nutter acquire_spu_lock(spu); /* Step 1. */ 210057dace23SArnd Bergmann prev->dar = spu->dar; 210157dace23SArnd Bergmann prev->dsisr = spu->dsisr; 210257dace23SArnd Bergmann spu->dar = 0; 210357dace23SArnd Bergmann spu->dsisr = 0; 21047c038749SMark Nutter rc = __do_spu_save(prev, spu); /* Steps 2-53. */ 21057c038749SMark Nutter release_spu_lock(spu); 21065737edd1SMark Nutter if (rc != 0 && rc != 2 && rc != 6) { 21078b3d6663SArnd Bergmann panic("%s failed on SPU[%d], rc=%d.\n", 21088b3d6663SArnd Bergmann __func__, spu->number, rc); 21098b3d6663SArnd Bergmann } 21105737edd1SMark Nutter return 0; 21115473af04SMark Nutter } 211291edfa49Sarnd@arndb.de EXPORT_SYMBOL_GPL(spu_save); 21135473af04SMark Nutter 21145473af04SMark Nutter /** 21155473af04SMark Nutter * spu_restore - SPU context restore, with harvest and locking. 21165473af04SMark Nutter * @new: pointer to SPU context save area, to be restored. 21175473af04SMark Nutter * @spu: pointer to SPU iomem structure. 21185473af04SMark Nutter * 21195473af04SMark Nutter * Perform harvest + restore, as we may not be coming 2120d6e05edcSAndreas Mohr * from a previous successful save operation, and the 21215473af04SMark Nutter * hardware state is unknown. 21225473af04SMark Nutter */ 21235473af04SMark Nutter int spu_restore(struct spu_state *new, struct spu *spu) 21245473af04SMark Nutter { 21257c038749SMark Nutter int rc; 21265473af04SMark Nutter 21277c038749SMark Nutter acquire_spu_lock(spu); 21287c038749SMark Nutter harvest(NULL, spu); 21298b3d6663SArnd Bergmann spu->slb_replace = 0; 213057dace23SArnd Bergmann new->dar = 0; 213157dace23SArnd Bergmann new->dsisr = 0; 21328b3d6663SArnd Bergmann spu->class_0_pending = 0; 21337c038749SMark Nutter rc = __do_spu_restore(new, spu); 21347c038749SMark Nutter release_spu_lock(spu); 21358b3d6663SArnd Bergmann if (rc) { 21368b3d6663SArnd Bergmann panic("%s failed on SPU[%d] rc=%d.\n", 21378b3d6663SArnd Bergmann __func__, spu->number, rc); 21388b3d6663SArnd Bergmann } 21397c038749SMark Nutter return rc; 21405473af04SMark Nutter } 214191edfa49Sarnd@arndb.de EXPORT_SYMBOL_GPL(spu_restore); 21425473af04SMark Nutter 21435473af04SMark Nutter /** 21448b3d6663SArnd Bergmann * spu_harvest - SPU harvest (reset) operation 21455473af04SMark Nutter * @spu: pointer to SPU iomem structure. 21465473af04SMark Nutter * 21478b3d6663SArnd Bergmann * Perform SPU harvest (reset) operation. 21485473af04SMark Nutter */ 21498b3d6663SArnd Bergmann void spu_harvest(struct spu *spu) 21505473af04SMark Nutter { 21518b3d6663SArnd Bergmann acquire_spu_lock(spu); 21528b3d6663SArnd Bergmann harvest(NULL, spu); 21537c038749SMark Nutter release_spu_lock(spu); 21545473af04SMark Nutter } 21555473af04SMark Nutter 21565473af04SMark Nutter static void init_prob(struct spu_state *csa) 21575473af04SMark Nutter { 21585473af04SMark Nutter csa->spu_chnlcnt_RW[9] = 1; 21595473af04SMark Nutter csa->spu_chnlcnt_RW[21] = 16; 21605473af04SMark Nutter csa->spu_chnlcnt_RW[23] = 1; 21615473af04SMark Nutter csa->spu_chnlcnt_RW[28] = 1; 21625473af04SMark Nutter csa->spu_chnlcnt_RW[30] = 1; 21635473af04SMark Nutter csa->prob.spu_runcntl_RW = SPU_RUNCNTL_STOP; 2164970f1baaSarnd@arndb.de csa->prob.mb_stat_R = 0x000400; 21655473af04SMark Nutter } 21665473af04SMark Nutter 21675473af04SMark Nutter static void init_priv1(struct spu_state *csa) 21685473af04SMark Nutter { 21695473af04SMark Nutter /* Enable decode, relocate, tlbie response, master runcntl. */ 21705473af04SMark Nutter csa->priv1.mfc_sr1_RW = MFC_STATE1_LOCAL_STORAGE_DECODE_MASK | 21715473af04SMark Nutter MFC_STATE1_MASTER_RUN_CONTROL_MASK | 21725473af04SMark Nutter MFC_STATE1_PROBLEM_STATE_MASK | 21735473af04SMark Nutter MFC_STATE1_RELOCATE_MASK | MFC_STATE1_BUS_TLBIE_MASK; 21745473af04SMark Nutter 21755473af04SMark Nutter /* Enable OS-specific set of interrupts. */ 21765473af04SMark Nutter csa->priv1.int_mask_class0_RW = CLASS0_ENABLE_DMA_ALIGNMENT_INTR | 21775473af04SMark Nutter CLASS0_ENABLE_INVALID_DMA_COMMAND_INTR | 21785473af04SMark Nutter CLASS0_ENABLE_SPU_ERROR_INTR; 21795473af04SMark Nutter csa->priv1.int_mask_class1_RW = CLASS1_ENABLE_SEGMENT_FAULT_INTR | 21805473af04SMark Nutter CLASS1_ENABLE_STORAGE_FAULT_INTR; 21813a843d7cSArnd Bergmann csa->priv1.int_mask_class2_RW = CLASS2_ENABLE_SPU_STOP_INTR | 2182a33a7d73SArnd Bergmann CLASS2_ENABLE_SPU_HALT_INTR | 2183a33a7d73SArnd Bergmann CLASS2_ENABLE_SPU_DMA_TAG_GROUP_COMPLETE_INTR; 21845473af04SMark Nutter } 21855473af04SMark Nutter 21865473af04SMark Nutter static void init_priv2(struct spu_state *csa) 21875473af04SMark Nutter { 21885473af04SMark Nutter csa->priv2.spu_lslr_RW = LS_ADDR_MASK; 21895473af04SMark Nutter csa->priv2.mfc_control_RW = MFC_CNTL_RESUME_DMA_QUEUE | 21905473af04SMark Nutter MFC_CNTL_NORMAL_DMA_QUEUE_OPERATION | 21915473af04SMark Nutter MFC_CNTL_DMA_QUEUES_EMPTY_MASK; 21925473af04SMark Nutter } 21935473af04SMark Nutter 21945473af04SMark Nutter /** 21955473af04SMark Nutter * spu_alloc_csa - allocate and initialize an SPU context save area. 21965473af04SMark Nutter * 21975473af04SMark Nutter * Allocate and initialize the contents of an SPU context save area. 21985473af04SMark Nutter * This includes enabling address translation, interrupt masks, etc., 21995473af04SMark Nutter * as appropriate for the given OS environment. 22005473af04SMark Nutter * 22015473af04SMark Nutter * Note that storage for the 'lscsa' is allocated separately, 22025473af04SMark Nutter * as it is by far the largest of the context save regions, 22035473af04SMark Nutter * and may need to be pinned or otherwise specially aligned. 22045473af04SMark Nutter */ 2205f1fa74f4SBenjamin Herrenschmidt int spu_init_csa(struct spu_state *csa) 22065473af04SMark Nutter { 2207f1fa74f4SBenjamin Herrenschmidt int rc; 22085473af04SMark Nutter 22095473af04SMark Nutter if (!csa) 2210f1fa74f4SBenjamin Herrenschmidt return -EINVAL; 22115473af04SMark Nutter memset(csa, 0, sizeof(struct spu_state)); 22125473af04SMark Nutter 2213f1fa74f4SBenjamin Herrenschmidt rc = spu_alloc_lscsa(csa); 2214f1fa74f4SBenjamin Herrenschmidt if (rc) 2215f1fa74f4SBenjamin Herrenschmidt return rc; 22165473af04SMark Nutter 221734af946aSIngo Molnar spin_lock_init(&csa->register_lock); 22188b3d6663SArnd Bergmann 22195473af04SMark Nutter init_prob(csa); 22205473af04SMark Nutter init_priv1(csa); 22215473af04SMark Nutter init_priv2(csa); 2222f1fa74f4SBenjamin Herrenschmidt 2223f1fa74f4SBenjamin Herrenschmidt return 0; 22245473af04SMark Nutter } 222591edfa49Sarnd@arndb.de EXPORT_SYMBOL_GPL(spu_init_csa); 22265473af04SMark Nutter 22275473af04SMark Nutter void spu_fini_csa(struct spu_state *csa) 22285473af04SMark Nutter { 2229f1fa74f4SBenjamin Herrenschmidt spu_free_lscsa(csa); 22305473af04SMark Nutter } 223191edfa49Sarnd@arndb.de EXPORT_SYMBOL_GPL(spu_fini_csa); 2232