15473af04SMark Nutter /* 25473af04SMark Nutter * spu_switch.c 35473af04SMark Nutter * 45473af04SMark Nutter * (C) Copyright IBM Corp. 2005 55473af04SMark Nutter * 65473af04SMark Nutter * Author: Mark Nutter <mnutter@us.ibm.com> 75473af04SMark Nutter * 85473af04SMark Nutter * Host-side part of SPU context switch sequence outlined in 95473af04SMark Nutter * Synergistic Processor Element, Book IV. 105473af04SMark Nutter * 115473af04SMark Nutter * A fully premptive switch of an SPE is very expensive in terms 125473af04SMark Nutter * of time and system resources. SPE Book IV indicates that SPE 135473af04SMark Nutter * allocation should follow a "serially reusable device" model, 145473af04SMark Nutter * in which the SPE is assigned a task until it completes. When 155473af04SMark Nutter * this is not possible, this sequence may be used to premptively 165473af04SMark Nutter * save, and then later (optionally) restore the context of a 175473af04SMark Nutter * program executing on an SPE. 185473af04SMark Nutter * 195473af04SMark Nutter * 205473af04SMark Nutter * This program is free software; you can redistribute it and/or modify 215473af04SMark Nutter * it under the terms of the GNU General Public License as published by 225473af04SMark Nutter * the Free Software Foundation; either version 2, or (at your option) 235473af04SMark Nutter * any later version. 245473af04SMark Nutter * 255473af04SMark Nutter * This program is distributed in the hope that it will be useful, 265473af04SMark Nutter * but WITHOUT ANY WARRANTY; without even the implied warranty of 275473af04SMark Nutter * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 285473af04SMark Nutter * GNU General Public License for more details. 295473af04SMark Nutter * 305473af04SMark Nutter * You should have received a copy of the GNU General Public License 315473af04SMark Nutter * along with this program; if not, write to the Free Software 325473af04SMark Nutter * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA. 335473af04SMark Nutter */ 345473af04SMark Nutter 355473af04SMark Nutter #include <linux/module.h> 365473af04SMark Nutter #include <linux/errno.h> 37fae9ca79SArnd Bergmann #include <linux/hardirq.h> 385473af04SMark Nutter #include <linux/sched.h> 395473af04SMark Nutter #include <linux/kernel.h> 405473af04SMark Nutter #include <linux/mm.h> 415473af04SMark Nutter #include <linux/vmalloc.h> 425473af04SMark Nutter #include <linux/smp.h> 435473af04SMark Nutter #include <linux/stddef.h> 445473af04SMark Nutter #include <linux/unistd.h> 455473af04SMark Nutter 465473af04SMark Nutter #include <asm/io.h> 475473af04SMark Nutter #include <asm/spu.h> 48540270d8SGeoff Levand #include <asm/spu_priv1.h> 495473af04SMark Nutter #include <asm/spu_csa.h> 505473af04SMark Nutter #include <asm/mmu_context.h> 515473af04SMark Nutter 527cd58e43SJeremy Kerr #include "spufs.h" 537cd58e43SJeremy Kerr 545473af04SMark Nutter #include "spu_save_dump.h" 555473af04SMark Nutter #include "spu_restore_dump.h" 565473af04SMark Nutter 577c038749SMark Nutter #if 0 587c038749SMark Nutter #define POLL_WHILE_TRUE(_c) { \ 597c038749SMark Nutter do { \ 607c038749SMark Nutter } while (_c); \ 617c038749SMark Nutter } 627c038749SMark Nutter #else 637c038749SMark Nutter #define RELAX_SPIN_COUNT 1000 647c038749SMark Nutter #define POLL_WHILE_TRUE(_c) { \ 657c038749SMark Nutter do { \ 667c038749SMark Nutter int _i; \ 677c038749SMark Nutter for (_i=0; _i<RELAX_SPIN_COUNT && (_c); _i++) { \ 687c038749SMark Nutter cpu_relax(); \ 697c038749SMark Nutter } \ 707c038749SMark Nutter if (unlikely(_c)) yield(); \ 717c038749SMark Nutter else break; \ 727c038749SMark Nutter } while (_c); \ 737c038749SMark Nutter } 747c038749SMark Nutter #endif /* debug */ 757c038749SMark Nutter 767c038749SMark Nutter #define POLL_WHILE_FALSE(_c) POLL_WHILE_TRUE(!(_c)) 777c038749SMark Nutter 787c038749SMark Nutter static inline void acquire_spu_lock(struct spu *spu) 797c038749SMark Nutter { 807c038749SMark Nutter /* Save, Step 1: 817c038749SMark Nutter * Restore, Step 1: 827c038749SMark Nutter * Acquire SPU-specific mutual exclusion lock. 837c038749SMark Nutter * TBD. 847c038749SMark Nutter */ 857c038749SMark Nutter } 867c038749SMark Nutter 877c038749SMark Nutter static inline void release_spu_lock(struct spu *spu) 887c038749SMark Nutter { 897c038749SMark Nutter /* Restore, Step 76: 907c038749SMark Nutter * Release SPU-specific mutual exclusion lock. 917c038749SMark Nutter * TBD. 927c038749SMark Nutter */ 937c038749SMark Nutter } 947c038749SMark Nutter 957c038749SMark Nutter static inline int check_spu_isolate(struct spu_state *csa, struct spu *spu) 967c038749SMark Nutter { 977c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 987c038749SMark Nutter u32 isolate_state; 997c038749SMark Nutter 1007c038749SMark Nutter /* Save, Step 2: 1017c038749SMark Nutter * Save, Step 6: 1027c038749SMark Nutter * If SPU_Status[E,L,IS] any field is '1', this 1037c038749SMark Nutter * SPU is in isolate state and cannot be context 1047c038749SMark Nutter * saved at this time. 1057c038749SMark Nutter */ 1067c038749SMark Nutter isolate_state = SPU_STATUS_ISOLATED_STATE | 107eb758ce5Sarnd@arndb.de SPU_STATUS_ISOLATED_LOAD_STATUS | SPU_STATUS_ISOLATED_EXIT_STATUS; 1087c038749SMark Nutter return (in_be32(&prob->spu_status_R) & isolate_state) ? 1 : 0; 1097c038749SMark Nutter } 1107c038749SMark Nutter 1117c038749SMark Nutter static inline void disable_interrupts(struct spu_state *csa, struct spu *spu) 1127c038749SMark Nutter { 1137c038749SMark Nutter /* Save, Step 3: 1147c038749SMark Nutter * Restore, Step 2: 1157c038749SMark Nutter * Save INT_Mask_class0 in CSA. 1167c038749SMark Nutter * Write INT_MASK_class0 with value of 0. 1177c038749SMark Nutter * Save INT_Mask_class1 in CSA. 1187c038749SMark Nutter * Write INT_MASK_class1 with value of 0. 1197c038749SMark Nutter * Save INT_Mask_class2 in CSA. 1207c038749SMark Nutter * Write INT_MASK_class2 with value of 0. 121fae9ca79SArnd Bergmann * Synchronize all three interrupts to be sure 122fae9ca79SArnd Bergmann * we no longer execute a handler on another CPU. 1237c038749SMark Nutter */ 1247c038749SMark Nutter spin_lock_irq(&spu->register_lock); 1257c038749SMark Nutter if (csa) { 126f0831accSArnd Bergmann csa->priv1.int_mask_class0_RW = spu_int_mask_get(spu, 0); 127f0831accSArnd Bergmann csa->priv1.int_mask_class1_RW = spu_int_mask_get(spu, 1); 128f0831accSArnd Bergmann csa->priv1.int_mask_class2_RW = spu_int_mask_get(spu, 2); 1297c038749SMark Nutter } 130f0831accSArnd Bergmann spu_int_mask_set(spu, 0, 0ul); 131f0831accSArnd Bergmann spu_int_mask_set(spu, 1, 0ul); 132f0831accSArnd Bergmann spu_int_mask_set(spu, 2, 0ul); 1337c038749SMark Nutter eieio(); 1347c038749SMark Nutter spin_unlock_irq(&spu->register_lock); 135fae9ca79SArnd Bergmann synchronize_irq(spu->irqs[0]); 136fae9ca79SArnd Bergmann synchronize_irq(spu->irqs[1]); 137fae9ca79SArnd Bergmann synchronize_irq(spu->irqs[2]); 1387c038749SMark Nutter } 1397c038749SMark Nutter 1407c038749SMark Nutter static inline void set_watchdog_timer(struct spu_state *csa, struct spu *spu) 1417c038749SMark Nutter { 1427c038749SMark Nutter /* Save, Step 4: 1437c038749SMark Nutter * Restore, Step 25. 1447c038749SMark Nutter * Set a software watchdog timer, which specifies the 1457c038749SMark Nutter * maximum allowable time for a context save sequence. 1467c038749SMark Nutter * 1477c038749SMark Nutter * For present, this implementation will not set a global 1487c038749SMark Nutter * watchdog timer, as virtualization & variable system load 1497c038749SMark Nutter * may cause unpredictable execution times. 1507c038749SMark Nutter */ 1517c038749SMark Nutter } 1527c038749SMark Nutter 1537c038749SMark Nutter static inline void inhibit_user_access(struct spu_state *csa, struct spu *spu) 1547c038749SMark Nutter { 1557c038749SMark Nutter /* Save, Step 5: 1567c038749SMark Nutter * Restore, Step 3: 1577c038749SMark Nutter * Inhibit user-space access (if provided) to this 1587c038749SMark Nutter * SPU by unmapping the virtual pages assigned to 1597c038749SMark Nutter * the SPU memory-mapped I/O (MMIO) for problem 1607c038749SMark Nutter * state. TBD. 1617c038749SMark Nutter */ 1627c038749SMark Nutter } 1637c038749SMark Nutter 1647c038749SMark Nutter static inline void set_switch_pending(struct spu_state *csa, struct spu *spu) 1657c038749SMark Nutter { 1667c038749SMark Nutter /* Save, Step 7: 1677c038749SMark Nutter * Restore, Step 5: 1687c038749SMark Nutter * Set a software context switch pending flag. 1697c038749SMark Nutter */ 1708837d921SArnd Bergmann set_bit(SPU_CONTEXT_SWITCH_PENDING, &spu->flags); 1717c038749SMark Nutter mb(); 1727c038749SMark Nutter } 1737c038749SMark Nutter 1747c038749SMark Nutter static inline void save_mfc_cntl(struct spu_state *csa, struct spu *spu) 1757c038749SMark Nutter { 1767c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 1777c038749SMark Nutter 1787c038749SMark Nutter /* Save, Step 8: 1797f52eb00SGeoff Levand * Suspend DMA and save MFC_CNTL. 1807c038749SMark Nutter */ 1817f52eb00SGeoff Levand switch (in_be64(&priv2->mfc_control_RW) & 1827f52eb00SGeoff Levand MFC_CNTL_SUSPEND_DMA_STATUS_MASK) { 1837f52eb00SGeoff Levand case MFC_CNTL_SUSPEND_IN_PROGRESS: 1847f52eb00SGeoff Levand POLL_WHILE_FALSE((in_be64(&priv2->mfc_control_RW) & 1857f52eb00SGeoff Levand MFC_CNTL_SUSPEND_DMA_STATUS_MASK) == 1867f52eb00SGeoff Levand MFC_CNTL_SUSPEND_COMPLETE); 1877f52eb00SGeoff Levand /* fall through */ 1887f52eb00SGeoff Levand case MFC_CNTL_SUSPEND_COMPLETE: 1891ca4264eSJeremy Kerr if (csa) 1907f52eb00SGeoff Levand csa->priv2.mfc_control_RW = 1911ca4264eSJeremy Kerr in_be64(&priv2->mfc_control_RW) | 1927f52eb00SGeoff Levand MFC_CNTL_SUSPEND_DMA_QUEUE; 1937f52eb00SGeoff Levand break; 1947f52eb00SGeoff Levand case MFC_CNTL_NORMAL_DMA_QUEUE_OPERATION: 1957f52eb00SGeoff Levand out_be64(&priv2->mfc_control_RW, MFC_CNTL_SUSPEND_DMA_QUEUE); 1967f52eb00SGeoff Levand POLL_WHILE_FALSE((in_be64(&priv2->mfc_control_RW) & 1977f52eb00SGeoff Levand MFC_CNTL_SUSPEND_DMA_STATUS_MASK) == 1987f52eb00SGeoff Levand MFC_CNTL_SUSPEND_COMPLETE); 1991ca4264eSJeremy Kerr if (csa) 2001ca4264eSJeremy Kerr csa->priv2.mfc_control_RW = 2011ca4264eSJeremy Kerr in_be64(&priv2->mfc_control_RW) & 2021ca4264eSJeremy Kerr ~MFC_CNTL_SUSPEND_DMA_QUEUE & 2031ca4264eSJeremy Kerr ~MFC_CNTL_SUSPEND_MASK; 2047f52eb00SGeoff Levand break; 2057c038749SMark Nutter } 2067c038749SMark Nutter } 2077c038749SMark Nutter 2087c038749SMark Nutter static inline void save_spu_runcntl(struct spu_state *csa, struct spu *spu) 2097c038749SMark Nutter { 2107c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 2117c038749SMark Nutter 2127c038749SMark Nutter /* Save, Step 9: 2137c038749SMark Nutter * Save SPU_Runcntl in the CSA. This value contains 2147c038749SMark Nutter * the "Application Desired State". 2157c038749SMark Nutter */ 2167c038749SMark Nutter csa->prob.spu_runcntl_RW = in_be32(&prob->spu_runcntl_RW); 2177c038749SMark Nutter } 2187c038749SMark Nutter 2197c038749SMark Nutter static inline void save_mfc_sr1(struct spu_state *csa, struct spu *spu) 2207c038749SMark Nutter { 2217c038749SMark Nutter /* Save, Step 10: 2227c038749SMark Nutter * Save MFC_SR1 in the CSA. 2237c038749SMark Nutter */ 224f0831accSArnd Bergmann csa->priv1.mfc_sr1_RW = spu_mfc_sr1_get(spu); 2257c038749SMark Nutter } 2267c038749SMark Nutter 2277c038749SMark Nutter static inline void save_spu_status(struct spu_state *csa, struct spu *spu) 2287c038749SMark Nutter { 2297c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 2307c038749SMark Nutter 2317c038749SMark Nutter /* Save, Step 11: 2327c038749SMark Nutter * Read SPU_Status[R], and save to CSA. 2337c038749SMark Nutter */ 2347c038749SMark Nutter if ((in_be32(&prob->spu_status_R) & SPU_STATUS_RUNNING) == 0) { 2357c038749SMark Nutter csa->prob.spu_status_R = in_be32(&prob->spu_status_R); 2367c038749SMark Nutter } else { 2377c038749SMark Nutter u32 stopped; 2387c038749SMark Nutter 2397c038749SMark Nutter out_be32(&prob->spu_runcntl_RW, SPU_RUNCNTL_STOP); 2407c038749SMark Nutter eieio(); 2417c038749SMark Nutter POLL_WHILE_TRUE(in_be32(&prob->spu_status_R) & 2427c038749SMark Nutter SPU_STATUS_RUNNING); 2437c038749SMark Nutter stopped = 2447c038749SMark Nutter SPU_STATUS_INVALID_INSTR | SPU_STATUS_SINGLE_STEP | 2457c038749SMark Nutter SPU_STATUS_STOPPED_BY_HALT | SPU_STATUS_STOPPED_BY_STOP; 2467c038749SMark Nutter if ((in_be32(&prob->spu_status_R) & stopped) == 0) 2477c038749SMark Nutter csa->prob.spu_status_R = SPU_STATUS_RUNNING; 2487c038749SMark Nutter else 2497c038749SMark Nutter csa->prob.spu_status_R = in_be32(&prob->spu_status_R); 2507c038749SMark Nutter } 2517c038749SMark Nutter } 2527c038749SMark Nutter 25355d7cd74SJeremy Kerr static inline void save_mfc_stopped_status(struct spu_state *csa, 25455d7cd74SJeremy Kerr struct spu *spu) 2557c038749SMark Nutter { 2567c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 25755d7cd74SJeremy Kerr const u64 mask = MFC_CNTL_DECREMENTER_RUNNING | 25855d7cd74SJeremy Kerr MFC_CNTL_DMA_QUEUES_EMPTY; 2597c038749SMark Nutter 2607c038749SMark Nutter /* Save, Step 12: 2617c038749SMark Nutter * Read MFC_CNTL[Ds]. Update saved copy of 2627c038749SMark Nutter * CSA.MFC_CNTL[Ds]. 26355d7cd74SJeremy Kerr * 26455d7cd74SJeremy Kerr * update: do the same with MFC_CNTL[Q]. 2657c038749SMark Nutter */ 26655d7cd74SJeremy Kerr csa->priv2.mfc_control_RW &= ~mask; 26755d7cd74SJeremy Kerr csa->priv2.mfc_control_RW |= in_be64(&priv2->mfc_control_RW) & mask; 2687c038749SMark Nutter } 2697c038749SMark Nutter 2707c038749SMark Nutter static inline void halt_mfc_decr(struct spu_state *csa, struct spu *spu) 2717c038749SMark Nutter { 2727c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 2737c038749SMark Nutter 2747c038749SMark Nutter /* Save, Step 13: 2757c038749SMark Nutter * Write MFC_CNTL[Dh] set to a '1' to halt 2767c038749SMark Nutter * the decrementer. 2777c038749SMark Nutter */ 27849776d30SKazunori Asayama out_be64(&priv2->mfc_control_RW, 27949776d30SKazunori Asayama MFC_CNTL_DECREMENTER_HALTED | MFC_CNTL_SUSPEND_MASK); 2807c038749SMark Nutter eieio(); 2817c038749SMark Nutter } 2827c038749SMark Nutter 2837c038749SMark Nutter static inline void save_timebase(struct spu_state *csa, struct spu *spu) 2847c038749SMark Nutter { 2857c038749SMark Nutter /* Save, Step 14: 2867c038749SMark Nutter * Read PPE Timebase High and Timebase low registers 2877c038749SMark Nutter * and save in CSA. TBD. 2887c038749SMark Nutter */ 2897c038749SMark Nutter csa->suspend_time = get_cycles(); 2907c038749SMark Nutter } 2917c038749SMark Nutter 2927c038749SMark Nutter static inline void remove_other_spu_access(struct spu_state *csa, 2937c038749SMark Nutter struct spu *spu) 2947c038749SMark Nutter { 2957c038749SMark Nutter /* Save, Step 15: 2967c038749SMark Nutter * Remove other SPU access to this SPU by unmapping 2977c038749SMark Nutter * this SPU's pages from their address space. TBD. 2987c038749SMark Nutter */ 2997c038749SMark Nutter } 3007c038749SMark Nutter 3017c038749SMark Nutter static inline void do_mfc_mssync(struct spu_state *csa, struct spu *spu) 3027c038749SMark Nutter { 3037c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 3047c038749SMark Nutter 3057c038749SMark Nutter /* Save, Step 16: 3067c038749SMark Nutter * Restore, Step 11. 3077c038749SMark Nutter * Write SPU_MSSync register. Poll SPU_MSSync[P] 3087c038749SMark Nutter * for a value of 0. 3097c038749SMark Nutter */ 3107c038749SMark Nutter out_be64(&prob->spc_mssync_RW, 1UL); 3117c038749SMark Nutter POLL_WHILE_TRUE(in_be64(&prob->spc_mssync_RW) & MS_SYNC_PENDING); 3127c038749SMark Nutter } 3137c038749SMark Nutter 3147c038749SMark Nutter static inline void issue_mfc_tlbie(struct spu_state *csa, struct spu *spu) 3157c038749SMark Nutter { 3167c038749SMark Nutter /* Save, Step 17: 3177c038749SMark Nutter * Restore, Step 12. 3187c038749SMark Nutter * Restore, Step 48. 3197c038749SMark Nutter * Write TLB_Invalidate_Entry[IS,VPN,L,Lp]=0 register. 3207c038749SMark Nutter * Then issue a PPE sync instruction. 3217c038749SMark Nutter */ 322f0831accSArnd Bergmann spu_tlb_invalidate(spu); 3237c038749SMark Nutter mb(); 3247c038749SMark Nutter } 3257c038749SMark Nutter 3267c038749SMark Nutter static inline void handle_pending_interrupts(struct spu_state *csa, 3277c038749SMark Nutter struct spu *spu) 3287c038749SMark Nutter { 3297c038749SMark Nutter /* Save, Step 18: 3307c038749SMark Nutter * Handle any pending interrupts from this SPU 3317c038749SMark Nutter * here. This is OS or hypervisor specific. One 3327c038749SMark Nutter * option is to re-enable interrupts to handle any 3337c038749SMark Nutter * pending interrupts, with the interrupt handlers 3347c038749SMark Nutter * recognizing the software Context Switch Pending 3357c038749SMark Nutter * flag, to ensure the SPU execution or MFC command 3367c038749SMark Nutter * queue is not restarted. TBD. 3377c038749SMark Nutter */ 3387c038749SMark Nutter } 3397c038749SMark Nutter 3407c038749SMark Nutter static inline void save_mfc_queues(struct spu_state *csa, struct spu *spu) 3417c038749SMark Nutter { 3427c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 3437c038749SMark Nutter int i; 3447c038749SMark Nutter 3457c038749SMark Nutter /* Save, Step 19: 3467c038749SMark Nutter * If MFC_Cntl[Se]=0 then save 3477c038749SMark Nutter * MFC command queues. 3487c038749SMark Nutter */ 3497c038749SMark Nutter if ((in_be64(&priv2->mfc_control_RW) & MFC_CNTL_DMA_QUEUES_EMPTY) == 0) { 3507c038749SMark Nutter for (i = 0; i < 8; i++) { 3517c038749SMark Nutter csa->priv2.puq[i].mfc_cq_data0_RW = 3527c038749SMark Nutter in_be64(&priv2->puq[i].mfc_cq_data0_RW); 3537c038749SMark Nutter csa->priv2.puq[i].mfc_cq_data1_RW = 3547c038749SMark Nutter in_be64(&priv2->puq[i].mfc_cq_data1_RW); 3557c038749SMark Nutter csa->priv2.puq[i].mfc_cq_data2_RW = 3567c038749SMark Nutter in_be64(&priv2->puq[i].mfc_cq_data2_RW); 3577c038749SMark Nutter csa->priv2.puq[i].mfc_cq_data3_RW = 3587c038749SMark Nutter in_be64(&priv2->puq[i].mfc_cq_data3_RW); 3597c038749SMark Nutter } 3607c038749SMark Nutter for (i = 0; i < 16; i++) { 3617c038749SMark Nutter csa->priv2.spuq[i].mfc_cq_data0_RW = 3627c038749SMark Nutter in_be64(&priv2->spuq[i].mfc_cq_data0_RW); 3637c038749SMark Nutter csa->priv2.spuq[i].mfc_cq_data1_RW = 3647c038749SMark Nutter in_be64(&priv2->spuq[i].mfc_cq_data1_RW); 3657c038749SMark Nutter csa->priv2.spuq[i].mfc_cq_data2_RW = 3667c038749SMark Nutter in_be64(&priv2->spuq[i].mfc_cq_data2_RW); 3677c038749SMark Nutter csa->priv2.spuq[i].mfc_cq_data3_RW = 3687c038749SMark Nutter in_be64(&priv2->spuq[i].mfc_cq_data3_RW); 3697c038749SMark Nutter } 3707c038749SMark Nutter } 3717c038749SMark Nutter } 3727c038749SMark Nutter 3737c038749SMark Nutter static inline void save_ppu_querymask(struct spu_state *csa, struct spu *spu) 3747c038749SMark Nutter { 3757c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 3767c038749SMark Nutter 3777c038749SMark Nutter /* Save, Step 20: 3787c038749SMark Nutter * Save the PPU_QueryMask register 3797c038749SMark Nutter * in the CSA. 3807c038749SMark Nutter */ 3817c038749SMark Nutter csa->prob.dma_querymask_RW = in_be32(&prob->dma_querymask_RW); 3827c038749SMark Nutter } 3837c038749SMark Nutter 3847c038749SMark Nutter static inline void save_ppu_querytype(struct spu_state *csa, struct spu *spu) 3857c038749SMark Nutter { 3867c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 3877c038749SMark Nutter 3887c038749SMark Nutter /* Save, Step 21: 3897c038749SMark Nutter * Save the PPU_QueryType register 3907c038749SMark Nutter * in the CSA. 3917c038749SMark Nutter */ 3927c038749SMark Nutter csa->prob.dma_querytype_RW = in_be32(&prob->dma_querytype_RW); 3937c038749SMark Nutter } 3947c038749SMark Nutter 3958d038e04SKazunori Asayama static inline void save_ppu_tagstatus(struct spu_state *csa, struct spu *spu) 3968d038e04SKazunori Asayama { 3978d038e04SKazunori Asayama struct spu_problem __iomem *prob = spu->problem; 3988d038e04SKazunori Asayama 3998d038e04SKazunori Asayama /* Save the Prxy_TagStatus register in the CSA. 4008d038e04SKazunori Asayama * 4018d038e04SKazunori Asayama * It is unnecessary to restore dma_tagstatus_R, however, 4028d038e04SKazunori Asayama * dma_tagstatus_R in the CSA is accessed via backing_ops, so 4038d038e04SKazunori Asayama * we must save it. 4048d038e04SKazunori Asayama */ 4058d038e04SKazunori Asayama csa->prob.dma_tagstatus_R = in_be32(&prob->dma_tagstatus_R); 4068d038e04SKazunori Asayama } 4078d038e04SKazunori Asayama 4087c038749SMark Nutter static inline void save_mfc_csr_tsq(struct spu_state *csa, struct spu *spu) 4097c038749SMark Nutter { 4107c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 4117c038749SMark Nutter 4127c038749SMark Nutter /* Save, Step 22: 4137c038749SMark Nutter * Save the MFC_CSR_TSQ register 4147c038749SMark Nutter * in the LSCSA. 4157c038749SMark Nutter */ 4167c038749SMark Nutter csa->priv2.spu_tag_status_query_RW = 4177c038749SMark Nutter in_be64(&priv2->spu_tag_status_query_RW); 4187c038749SMark Nutter } 4197c038749SMark Nutter 4207c038749SMark Nutter static inline void save_mfc_csr_cmd(struct spu_state *csa, struct spu *spu) 4217c038749SMark Nutter { 4227c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 4237c038749SMark Nutter 4247c038749SMark Nutter /* Save, Step 23: 4257c038749SMark Nutter * Save the MFC_CSR_CMD1 and MFC_CSR_CMD2 4267c038749SMark Nutter * registers in the CSA. 4277c038749SMark Nutter */ 4287c038749SMark Nutter csa->priv2.spu_cmd_buf1_RW = in_be64(&priv2->spu_cmd_buf1_RW); 4297c038749SMark Nutter csa->priv2.spu_cmd_buf2_RW = in_be64(&priv2->spu_cmd_buf2_RW); 4307c038749SMark Nutter } 4317c038749SMark Nutter 4327c038749SMark Nutter static inline void save_mfc_csr_ato(struct spu_state *csa, struct spu *spu) 4337c038749SMark Nutter { 4347c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 4357c038749SMark Nutter 4367c038749SMark Nutter /* Save, Step 24: 4377c038749SMark Nutter * Save the MFC_CSR_ATO register in 4387c038749SMark Nutter * the CSA. 4397c038749SMark Nutter */ 4407c038749SMark Nutter csa->priv2.spu_atomic_status_RW = in_be64(&priv2->spu_atomic_status_RW); 4417c038749SMark Nutter } 4427c038749SMark Nutter 4437c038749SMark Nutter static inline void save_mfc_tclass_id(struct spu_state *csa, struct spu *spu) 4447c038749SMark Nutter { 4457c038749SMark Nutter /* Save, Step 25: 4467c038749SMark Nutter * Save the MFC_TCLASS_ID register in 4477c038749SMark Nutter * the CSA. 4487c038749SMark Nutter */ 449f0831accSArnd Bergmann csa->priv1.mfc_tclass_id_RW = spu_mfc_tclass_id_get(spu); 4507c038749SMark Nutter } 4517c038749SMark Nutter 4527c038749SMark Nutter static inline void set_mfc_tclass_id(struct spu_state *csa, struct spu *spu) 4537c038749SMark Nutter { 4547c038749SMark Nutter /* Save, Step 26: 4557c038749SMark Nutter * Restore, Step 23. 4567c038749SMark Nutter * Write the MFC_TCLASS_ID register with 4577c038749SMark Nutter * the value 0x10000000. 4587c038749SMark Nutter */ 459f0831accSArnd Bergmann spu_mfc_tclass_id_set(spu, 0x10000000); 4607c038749SMark Nutter eieio(); 4617c038749SMark Nutter } 4627c038749SMark Nutter 4637c038749SMark Nutter static inline void purge_mfc_queue(struct spu_state *csa, struct spu *spu) 4647c038749SMark Nutter { 4657c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 4667c038749SMark Nutter 4677c038749SMark Nutter /* Save, Step 27: 4687c038749SMark Nutter * Restore, Step 14. 4697c038749SMark Nutter * Write MFC_CNTL[Pc]=1 (purge queue). 4707c038749SMark Nutter */ 471943906baSJeremy Kerr out_be64(&priv2->mfc_control_RW, 472943906baSJeremy Kerr MFC_CNTL_PURGE_DMA_REQUEST | 473943906baSJeremy Kerr MFC_CNTL_SUSPEND_MASK); 4747c038749SMark Nutter eieio(); 4757c038749SMark Nutter } 4767c038749SMark Nutter 4777c038749SMark Nutter static inline void wait_purge_complete(struct spu_state *csa, struct spu *spu) 4787c038749SMark Nutter { 4797c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 4807c038749SMark Nutter 4817c038749SMark Nutter /* Save, Step 28: 4827c038749SMark Nutter * Poll MFC_CNTL[Ps] until value '11' is read 4837c038749SMark Nutter * (purge complete). 4847c038749SMark Nutter */ 485910ab66bSBenjamin Herrenschmidt POLL_WHILE_FALSE((in_be64(&priv2->mfc_control_RW) & 486910ab66bSBenjamin Herrenschmidt MFC_CNTL_PURGE_DMA_STATUS_MASK) == 4877c038749SMark Nutter MFC_CNTL_PURGE_DMA_COMPLETE); 4887c038749SMark Nutter } 4897c038749SMark Nutter 4907c038749SMark Nutter static inline void setup_mfc_sr1(struct spu_state *csa, struct spu *spu) 4917c038749SMark Nutter { 4927c038749SMark Nutter /* Save, Step 30: 4937c038749SMark Nutter * Restore, Step 18: 4947c038749SMark Nutter * Write MFC_SR1 with MFC_SR1[D=0,S=1] and 4957c038749SMark Nutter * MFC_SR1[TL,R,Pr,T] set correctly for the 4967c038749SMark Nutter * OS specific environment. 4977c038749SMark Nutter * 4987c038749SMark Nutter * Implementation note: The SPU-side code 4997c038749SMark Nutter * for save/restore is privileged, so the 5007c038749SMark Nutter * MFC_SR1[Pr] bit is not set. 5017c038749SMark Nutter * 5027c038749SMark Nutter */ 503f0831accSArnd Bergmann spu_mfc_sr1_set(spu, (MFC_STATE1_MASTER_RUN_CONTROL_MASK | 5047c038749SMark Nutter MFC_STATE1_RELOCATE_MASK | 5057c038749SMark Nutter MFC_STATE1_BUS_TLBIE_MASK)); 5067c038749SMark Nutter } 5077c038749SMark Nutter 5087c038749SMark Nutter static inline void save_spu_npc(struct spu_state *csa, struct spu *spu) 5097c038749SMark Nutter { 5107c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 5117c038749SMark Nutter 5127c038749SMark Nutter /* Save, Step 31: 5137c038749SMark Nutter * Save SPU_NPC in the CSA. 5147c038749SMark Nutter */ 5157c038749SMark Nutter csa->prob.spu_npc_RW = in_be32(&prob->spu_npc_RW); 5167c038749SMark Nutter } 5177c038749SMark Nutter 5187c038749SMark Nutter static inline void save_spu_privcntl(struct spu_state *csa, struct spu *spu) 5197c038749SMark Nutter { 5207c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 5217c038749SMark Nutter 5227c038749SMark Nutter /* Save, Step 32: 5237c038749SMark Nutter * Save SPU_PrivCntl in the CSA. 5247c038749SMark Nutter */ 5257c038749SMark Nutter csa->priv2.spu_privcntl_RW = in_be64(&priv2->spu_privcntl_RW); 5267c038749SMark Nutter } 5277c038749SMark Nutter 5287c038749SMark Nutter static inline void reset_spu_privcntl(struct spu_state *csa, struct spu *spu) 5297c038749SMark Nutter { 5307c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 5317c038749SMark Nutter 5327c038749SMark Nutter /* Save, Step 33: 5337c038749SMark Nutter * Restore, Step 16: 5347c038749SMark Nutter * Write SPU_PrivCntl[S,Le,A] fields reset to 0. 5357c038749SMark Nutter */ 5367c038749SMark Nutter out_be64(&priv2->spu_privcntl_RW, 0UL); 5377c038749SMark Nutter eieio(); 5387c038749SMark Nutter } 5397c038749SMark Nutter 5407c038749SMark Nutter static inline void save_spu_lslr(struct spu_state *csa, struct spu *spu) 5417c038749SMark Nutter { 5427c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 5437c038749SMark Nutter 5447c038749SMark Nutter /* Save, Step 34: 5457c038749SMark Nutter * Save SPU_LSLR in the CSA. 5467c038749SMark Nutter */ 5477c038749SMark Nutter csa->priv2.spu_lslr_RW = in_be64(&priv2->spu_lslr_RW); 5487c038749SMark Nutter } 5497c038749SMark Nutter 5507c038749SMark Nutter static inline void reset_spu_lslr(struct spu_state *csa, struct spu *spu) 5517c038749SMark Nutter { 5527c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 5537c038749SMark Nutter 5547c038749SMark Nutter /* Save, Step 35: 5557c038749SMark Nutter * Restore, Step 17. 5567c038749SMark Nutter * Reset SPU_LSLR. 5577c038749SMark Nutter */ 5587c038749SMark Nutter out_be64(&priv2->spu_lslr_RW, LS_ADDR_MASK); 5597c038749SMark Nutter eieio(); 5607c038749SMark Nutter } 5617c038749SMark Nutter 5627c038749SMark Nutter static inline void save_spu_cfg(struct spu_state *csa, struct spu *spu) 5637c038749SMark Nutter { 5647c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 5657c038749SMark Nutter 5667c038749SMark Nutter /* Save, Step 36: 5677c038749SMark Nutter * Save SPU_Cfg in the CSA. 5687c038749SMark Nutter */ 5697c038749SMark Nutter csa->priv2.spu_cfg_RW = in_be64(&priv2->spu_cfg_RW); 5707c038749SMark Nutter } 5717c038749SMark Nutter 5727c038749SMark Nutter static inline void save_pm_trace(struct spu_state *csa, struct spu *spu) 5737c038749SMark Nutter { 5747c038749SMark Nutter /* Save, Step 37: 5757c038749SMark Nutter * Save PM_Trace_Tag_Wait_Mask in the CSA. 5767c038749SMark Nutter * Not performed by this implementation. 5777c038749SMark Nutter */ 5787c038749SMark Nutter } 5797c038749SMark Nutter 5807c038749SMark Nutter static inline void save_mfc_rag(struct spu_state *csa, struct spu *spu) 5817c038749SMark Nutter { 5827c038749SMark Nutter /* Save, Step 38: 5837c038749SMark Nutter * Save RA_GROUP_ID register and the 5847c038749SMark Nutter * RA_ENABLE reigster in the CSA. 5857c038749SMark Nutter */ 5867c038749SMark Nutter csa->priv1.resource_allocation_groupID_RW = 587f0831accSArnd Bergmann spu_resource_allocation_groupID_get(spu); 5887c038749SMark Nutter csa->priv1.resource_allocation_enable_RW = 589f0831accSArnd Bergmann spu_resource_allocation_enable_get(spu); 5907c038749SMark Nutter } 5917c038749SMark Nutter 5927c038749SMark Nutter static inline void save_ppu_mb_stat(struct spu_state *csa, struct spu *spu) 5937c038749SMark Nutter { 5947c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 5957c038749SMark Nutter 5967c038749SMark Nutter /* Save, Step 39: 5977c038749SMark Nutter * Save MB_Stat register in the CSA. 5987c038749SMark Nutter */ 5997c038749SMark Nutter csa->prob.mb_stat_R = in_be32(&prob->mb_stat_R); 6007c038749SMark Nutter } 6017c038749SMark Nutter 6027c038749SMark Nutter static inline void save_ppu_mb(struct spu_state *csa, struct spu *spu) 6037c038749SMark Nutter { 6047c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 6057c038749SMark Nutter 6067c038749SMark Nutter /* Save, Step 40: 6077c038749SMark Nutter * Save the PPU_MB register in the CSA. 6087c038749SMark Nutter */ 6097c038749SMark Nutter csa->prob.pu_mb_R = in_be32(&prob->pu_mb_R); 6107c038749SMark Nutter } 6117c038749SMark Nutter 6127c038749SMark Nutter static inline void save_ppuint_mb(struct spu_state *csa, struct spu *spu) 6137c038749SMark Nutter { 6147c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 6157c038749SMark Nutter 6167c038749SMark Nutter /* Save, Step 41: 6177c038749SMark Nutter * Save the PPUINT_MB register in the CSA. 6187c038749SMark Nutter */ 6197c038749SMark Nutter csa->priv2.puint_mb_R = in_be64(&priv2->puint_mb_R); 6207c038749SMark Nutter } 6217c038749SMark Nutter 6227c038749SMark Nutter static inline void save_ch_part1(struct spu_state *csa, struct spu *spu) 6237c038749SMark Nutter { 6247c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 625daced0f7SJeremy Kerr u64 idx, ch_indices[] = { 0UL, 3UL, 4UL, 24UL, 25UL, 27UL }; 6267c038749SMark Nutter int i; 6277c038749SMark Nutter 6287c038749SMark Nutter /* Save, Step 42: 6297c038749SMark Nutter */ 630e46a0237Sarnd@arndb.de 631e46a0237Sarnd@arndb.de /* Save CH 1, without channel count */ 632e46a0237Sarnd@arndb.de out_be64(&priv2->spu_chnlcntptr_RW, 1); 633e46a0237Sarnd@arndb.de csa->spu_chnldata_RW[1] = in_be64(&priv2->spu_chnldata_RW); 634e46a0237Sarnd@arndb.de 635e46a0237Sarnd@arndb.de /* Save the following CH: [0,3,4,24,25,27] */ 636daced0f7SJeremy Kerr for (i = 0; i < ARRAY_SIZE(ch_indices); i++) { 6377c038749SMark Nutter idx = ch_indices[i]; 6387c038749SMark Nutter out_be64(&priv2->spu_chnlcntptr_RW, idx); 6397c038749SMark Nutter eieio(); 6407c038749SMark Nutter csa->spu_chnldata_RW[idx] = in_be64(&priv2->spu_chnldata_RW); 6417c038749SMark Nutter csa->spu_chnlcnt_RW[idx] = in_be64(&priv2->spu_chnlcnt_RW); 6427c038749SMark Nutter out_be64(&priv2->spu_chnldata_RW, 0UL); 6437c038749SMark Nutter out_be64(&priv2->spu_chnlcnt_RW, 0UL); 6447c038749SMark Nutter eieio(); 6457c038749SMark Nutter } 6467c038749SMark Nutter } 6477c038749SMark Nutter 6487c038749SMark Nutter static inline void save_spu_mb(struct spu_state *csa, struct spu *spu) 6497c038749SMark Nutter { 6507c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 6517c038749SMark Nutter int i; 6527c038749SMark Nutter 6537c038749SMark Nutter /* Save, Step 43: 6547c038749SMark Nutter * Save SPU Read Mailbox Channel. 6557c038749SMark Nutter */ 6567c038749SMark Nutter out_be64(&priv2->spu_chnlcntptr_RW, 29UL); 6577c038749SMark Nutter eieio(); 6587c038749SMark Nutter csa->spu_chnlcnt_RW[29] = in_be64(&priv2->spu_chnlcnt_RW); 6597c038749SMark Nutter for (i = 0; i < 4; i++) { 6608b3d6663SArnd Bergmann csa->spu_mailbox_data[i] = in_be64(&priv2->spu_chnldata_RW); 6617c038749SMark Nutter } 6627c038749SMark Nutter out_be64(&priv2->spu_chnlcnt_RW, 0UL); 6637c038749SMark Nutter eieio(); 6647c038749SMark Nutter } 6657c038749SMark Nutter 6667c038749SMark Nutter static inline void save_mfc_cmd(struct spu_state *csa, struct spu *spu) 6677c038749SMark Nutter { 6687c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 6697c038749SMark Nutter 6707c038749SMark Nutter /* Save, Step 44: 6717c038749SMark Nutter * Save MFC_CMD Channel. 6727c038749SMark Nutter */ 6737c038749SMark Nutter out_be64(&priv2->spu_chnlcntptr_RW, 21UL); 6747c038749SMark Nutter eieio(); 6757c038749SMark Nutter csa->spu_chnlcnt_RW[21] = in_be64(&priv2->spu_chnlcnt_RW); 6767c038749SMark Nutter eieio(); 6777c038749SMark Nutter } 6787c038749SMark Nutter 6797c038749SMark Nutter static inline void reset_ch(struct spu_state *csa, struct spu *spu) 6807c038749SMark Nutter { 6817c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 6827c038749SMark Nutter u64 ch_indices[4] = { 21UL, 23UL, 28UL, 30UL }; 6837c038749SMark Nutter u64 ch_counts[4] = { 16UL, 1UL, 1UL, 1UL }; 6847c038749SMark Nutter u64 idx; 6857c038749SMark Nutter int i; 6867c038749SMark Nutter 6877c038749SMark Nutter /* Save, Step 45: 6887c038749SMark Nutter * Reset the following CH: [21, 23, 28, 30] 6897c038749SMark Nutter */ 6907c038749SMark Nutter for (i = 0; i < 4; i++) { 6917c038749SMark Nutter idx = ch_indices[i]; 6927c038749SMark Nutter out_be64(&priv2->spu_chnlcntptr_RW, idx); 6937c038749SMark Nutter eieio(); 6947c038749SMark Nutter out_be64(&priv2->spu_chnlcnt_RW, ch_counts[i]); 6957c038749SMark Nutter eieio(); 6967c038749SMark Nutter } 6977c038749SMark Nutter } 6987c038749SMark Nutter 6997c038749SMark Nutter static inline void resume_mfc_queue(struct spu_state *csa, struct spu *spu) 7007c038749SMark Nutter { 7017c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 7027c038749SMark Nutter 7037c038749SMark Nutter /* Save, Step 46: 7047c038749SMark Nutter * Restore, Step 25. 7057c038749SMark Nutter * Write MFC_CNTL[Sc]=0 (resume queue processing). 7067c038749SMark Nutter */ 7077c038749SMark Nutter out_be64(&priv2->mfc_control_RW, MFC_CNTL_RESUME_DMA_QUEUE); 7087c038749SMark Nutter } 7097c038749SMark Nutter 710684bd614SJeremy Kerr static inline void setup_mfc_slbs(struct spu_state *csa, struct spu *spu, 711684bd614SJeremy Kerr unsigned int *code, int code_size) 7127c038749SMark Nutter { 7137c038749SMark Nutter /* Save, Step 47: 7147c038749SMark Nutter * Restore, Step 30. 7157c038749SMark Nutter * If MFC_SR1[R]=1, write 0 to SLB_Invalidate_All 7167c038749SMark Nutter * register, then initialize SLB_VSID and SLB_ESID 7177c038749SMark Nutter * to provide access to SPU context save code and 7187c038749SMark Nutter * LSCSA. 7197c038749SMark Nutter * 7207c038749SMark Nutter * This implementation places both the context 7217c038749SMark Nutter * switch code and LSCSA in kernel address space. 7227c038749SMark Nutter * 7237c038749SMark Nutter * Further this implementation assumes that the 7247c038749SMark Nutter * MFC_SR1[R]=1 (in other words, assume that 7257c038749SMark Nutter * translation is desired by OS environment). 7267c038749SMark Nutter */ 72794b2a439SBenjamin Herrenschmidt spu_invalidate_slbs(spu); 728684bd614SJeremy Kerr spu_setup_kernel_slbs(spu, csa->lscsa, code, code_size); 7297c038749SMark Nutter } 7307c038749SMark Nutter 7317c038749SMark Nutter static inline void set_switch_active(struct spu_state *csa, struct spu *spu) 7327c038749SMark Nutter { 7337c038749SMark Nutter /* Save, Step 48: 7347c038749SMark Nutter * Restore, Step 23. 7357c038749SMark Nutter * Change the software context switch pending flag 7367c038749SMark Nutter * to context switch active. 73761b36fc1SAndre Detsch * 73861b36fc1SAndre Detsch * This implementation does not uses a switch active flag. 7397c038749SMark Nutter */ 7408837d921SArnd Bergmann clear_bit(SPU_CONTEXT_SWITCH_PENDING, &spu->flags); 7417c038749SMark Nutter mb(); 7427c038749SMark Nutter } 7437c038749SMark Nutter 7447c038749SMark Nutter static inline void enable_interrupts(struct spu_state *csa, struct spu *spu) 7457c038749SMark Nutter { 7467c038749SMark Nutter unsigned long class1_mask = CLASS1_ENABLE_SEGMENT_FAULT_INTR | 7477c038749SMark Nutter CLASS1_ENABLE_STORAGE_FAULT_INTR; 7487c038749SMark Nutter 7497c038749SMark Nutter /* Save, Step 49: 7507c038749SMark Nutter * Restore, Step 22: 7517c038749SMark Nutter * Reset and then enable interrupts, as 7527c038749SMark Nutter * needed by OS. 7537c038749SMark Nutter * 7547c038749SMark Nutter * This implementation enables only class1 7557c038749SMark Nutter * (translation) interrupts. 7567c038749SMark Nutter */ 7577c038749SMark Nutter spin_lock_irq(&spu->register_lock); 7589476141cSMasato Noguchi spu_int_stat_clear(spu, 0, CLASS0_INTR_MASK); 7599476141cSMasato Noguchi spu_int_stat_clear(spu, 1, CLASS1_INTR_MASK); 7609476141cSMasato Noguchi spu_int_stat_clear(spu, 2, CLASS2_INTR_MASK); 761f0831accSArnd Bergmann spu_int_mask_set(spu, 0, 0ul); 762f0831accSArnd Bergmann spu_int_mask_set(spu, 1, class1_mask); 763f0831accSArnd Bergmann spu_int_mask_set(spu, 2, 0ul); 7647c038749SMark Nutter spin_unlock_irq(&spu->register_lock); 7657c038749SMark Nutter } 7667c038749SMark Nutter 7677c038749SMark Nutter static inline int send_mfc_dma(struct spu *spu, unsigned long ea, 7687c038749SMark Nutter unsigned int ls_offset, unsigned int size, 7697c038749SMark Nutter unsigned int tag, unsigned int rclass, 7707c038749SMark Nutter unsigned int cmd) 7717c038749SMark Nutter { 7727c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 7737c038749SMark Nutter union mfc_tag_size_class_cmd command; 7747c038749SMark Nutter unsigned int transfer_size; 7757c038749SMark Nutter volatile unsigned int status = 0x0; 7767c038749SMark Nutter 7777c038749SMark Nutter while (size > 0) { 7787c038749SMark Nutter transfer_size = 7797c038749SMark Nutter (size > MFC_MAX_DMA_SIZE) ? MFC_MAX_DMA_SIZE : size; 7807c038749SMark Nutter command.u.mfc_size = transfer_size; 7817c038749SMark Nutter command.u.mfc_tag = tag; 7827c038749SMark Nutter command.u.mfc_rclassid = rclass; 7837c038749SMark Nutter command.u.mfc_cmd = cmd; 7847c038749SMark Nutter do { 7857c038749SMark Nutter out_be32(&prob->mfc_lsa_W, ls_offset); 7867c038749SMark Nutter out_be64(&prob->mfc_ea_W, ea); 7877c038749SMark Nutter out_be64(&prob->mfc_union_W.all64, command.all64); 7887c038749SMark Nutter status = 7897c038749SMark Nutter in_be32(&prob->mfc_union_W.by32.mfc_class_cmd32); 7907c038749SMark Nutter if (unlikely(status & 0x2)) { 7917c038749SMark Nutter cpu_relax(); 7927c038749SMark Nutter } 7937c038749SMark Nutter } while (status & 0x3); 7947c038749SMark Nutter size -= transfer_size; 7957c038749SMark Nutter ea += transfer_size; 7967c038749SMark Nutter ls_offset += transfer_size; 7977c038749SMark Nutter } 7987c038749SMark Nutter return 0; 7997c038749SMark Nutter } 8007c038749SMark Nutter 8017c038749SMark Nutter static inline void save_ls_16kb(struct spu_state *csa, struct spu *spu) 8027c038749SMark Nutter { 8037c038749SMark Nutter unsigned long addr = (unsigned long)&csa->lscsa->ls[0]; 8047c038749SMark Nutter unsigned int ls_offset = 0x0; 8057c038749SMark Nutter unsigned int size = 16384; 8067c038749SMark Nutter unsigned int tag = 0; 8077c038749SMark Nutter unsigned int rclass = 0; 8087c038749SMark Nutter unsigned int cmd = MFC_PUT_CMD; 8097c038749SMark Nutter 8107c038749SMark Nutter /* Save, Step 50: 8117c038749SMark Nutter * Issue a DMA command to copy the first 16K bytes 8127c038749SMark Nutter * of local storage to the CSA. 8137c038749SMark Nutter */ 8147c038749SMark Nutter send_mfc_dma(spu, addr, ls_offset, size, tag, rclass, cmd); 8157c038749SMark Nutter } 8167c038749SMark Nutter 8177c038749SMark Nutter static inline void set_spu_npc(struct spu_state *csa, struct spu *spu) 8187c038749SMark Nutter { 8197c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 8207c038749SMark Nutter 8217c038749SMark Nutter /* Save, Step 51: 8227c038749SMark Nutter * Restore, Step 31. 8237c038749SMark Nutter * Write SPU_NPC[IE]=0 and SPU_NPC[LSA] to entry 8247c038749SMark Nutter * point address of context save code in local 8257c038749SMark Nutter * storage. 8267c038749SMark Nutter * 8277c038749SMark Nutter * This implementation uses SPU-side save/restore 8287c038749SMark Nutter * programs with entry points at LSA of 0. 8297c038749SMark Nutter */ 8307c038749SMark Nutter out_be32(&prob->spu_npc_RW, 0); 8317c038749SMark Nutter eieio(); 8327c038749SMark Nutter } 8337c038749SMark Nutter 8347c038749SMark Nutter static inline void set_signot1(struct spu_state *csa, struct spu *spu) 8357c038749SMark Nutter { 8367c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 8377c038749SMark Nutter union { 8387c038749SMark Nutter u64 ull; 8397c038749SMark Nutter u32 ui[2]; 8407c038749SMark Nutter } addr64; 8417c038749SMark Nutter 8427c038749SMark Nutter /* Save, Step 52: 8437c038749SMark Nutter * Restore, Step 32: 8447c038749SMark Nutter * Write SPU_Sig_Notify_1 register with upper 32-bits 8457c038749SMark Nutter * of the CSA.LSCSA effective address. 8467c038749SMark Nutter */ 8477c038749SMark Nutter addr64.ull = (u64) csa->lscsa; 8487c038749SMark Nutter out_be32(&prob->signal_notify1, addr64.ui[0]); 8497c038749SMark Nutter eieio(); 8507c038749SMark Nutter } 8517c038749SMark Nutter 8527c038749SMark Nutter static inline void set_signot2(struct spu_state *csa, struct spu *spu) 8537c038749SMark Nutter { 8547c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 8557c038749SMark Nutter union { 8567c038749SMark Nutter u64 ull; 8577c038749SMark Nutter u32 ui[2]; 8587c038749SMark Nutter } addr64; 8597c038749SMark Nutter 8607c038749SMark Nutter /* Save, Step 53: 8617c038749SMark Nutter * Restore, Step 33: 8627c038749SMark Nutter * Write SPU_Sig_Notify_2 register with lower 32-bits 8637c038749SMark Nutter * of the CSA.LSCSA effective address. 8647c038749SMark Nutter */ 8657c038749SMark Nutter addr64.ull = (u64) csa->lscsa; 8667c038749SMark Nutter out_be32(&prob->signal_notify2, addr64.ui[1]); 8677c038749SMark Nutter eieio(); 8687c038749SMark Nutter } 8697c038749SMark Nutter 8707c038749SMark Nutter static inline void send_save_code(struct spu_state *csa, struct spu *spu) 8717c038749SMark Nutter { 8727c038749SMark Nutter unsigned long addr = (unsigned long)&spu_save_code[0]; 8737c038749SMark Nutter unsigned int ls_offset = 0x0; 8747c038749SMark Nutter unsigned int size = sizeof(spu_save_code); 8757c038749SMark Nutter unsigned int tag = 0; 8767c038749SMark Nutter unsigned int rclass = 0; 8777c038749SMark Nutter unsigned int cmd = MFC_GETFS_CMD; 8787c038749SMark Nutter 8797c038749SMark Nutter /* Save, Step 54: 8807c038749SMark Nutter * Issue a DMA command to copy context save code 8817c038749SMark Nutter * to local storage and start SPU. 8827c038749SMark Nutter */ 8837c038749SMark Nutter send_mfc_dma(spu, addr, ls_offset, size, tag, rclass, cmd); 8847c038749SMark Nutter } 8857c038749SMark Nutter 8867c038749SMark Nutter static inline void set_ppu_querymask(struct spu_state *csa, struct spu *spu) 8877c038749SMark Nutter { 8887c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 8897c038749SMark Nutter 8907c038749SMark Nutter /* Save, Step 55: 8917c038749SMark Nutter * Restore, Step 38. 8927c038749SMark Nutter * Write PPU_QueryMask=1 (enable Tag Group 0) 8937c038749SMark Nutter * and issue eieio instruction. 8947c038749SMark Nutter */ 8957c038749SMark Nutter out_be32(&prob->dma_querymask_RW, MFC_TAGID_TO_TAGMASK(0)); 8967c038749SMark Nutter eieio(); 8977c038749SMark Nutter } 8987c038749SMark Nutter 8997c038749SMark Nutter static inline void wait_tag_complete(struct spu_state *csa, struct spu *spu) 9007c038749SMark Nutter { 9017c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 9027c038749SMark Nutter u32 mask = MFC_TAGID_TO_TAGMASK(0); 9037c038749SMark Nutter unsigned long flags; 9047c038749SMark Nutter 9057c038749SMark Nutter /* Save, Step 56: 9067c038749SMark Nutter * Restore, Step 39. 9077c038749SMark Nutter * Restore, Step 39. 9087c038749SMark Nutter * Restore, Step 46. 9097c038749SMark Nutter * Poll PPU_TagStatus[gn] until 01 (Tag group 0 complete) 9107c038749SMark Nutter * or write PPU_QueryType[TS]=01 and wait for Tag Group 9117c038749SMark Nutter * Complete Interrupt. Write INT_Stat_Class0 or 9127c038749SMark Nutter * INT_Stat_Class2 with value of 'handled'. 9137c038749SMark Nutter */ 9147c038749SMark Nutter POLL_WHILE_FALSE(in_be32(&prob->dma_tagstatus_R) & mask); 9157c038749SMark Nutter 9167c038749SMark Nutter local_irq_save(flags); 9179476141cSMasato Noguchi spu_int_stat_clear(spu, 0, CLASS0_INTR_MASK); 9189476141cSMasato Noguchi spu_int_stat_clear(spu, 2, CLASS2_INTR_MASK); 9197c038749SMark Nutter local_irq_restore(flags); 9207c038749SMark Nutter } 9217c038749SMark Nutter 9227c038749SMark Nutter static inline void wait_spu_stopped(struct spu_state *csa, struct spu *spu) 9237c038749SMark Nutter { 9247c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 9257c038749SMark Nutter unsigned long flags; 9267c038749SMark Nutter 9277c038749SMark Nutter /* Save, Step 57: 9287c038749SMark Nutter * Restore, Step 40. 9297c038749SMark Nutter * Poll until SPU_Status[R]=0 or wait for SPU Class 0 9307c038749SMark Nutter * or SPU Class 2 interrupt. Write INT_Stat_class0 9317c038749SMark Nutter * or INT_Stat_class2 with value of handled. 9327c038749SMark Nutter */ 9337c038749SMark Nutter POLL_WHILE_TRUE(in_be32(&prob->spu_status_R) & SPU_STATUS_RUNNING); 9347c038749SMark Nutter 9357c038749SMark Nutter local_irq_save(flags); 9369476141cSMasato Noguchi spu_int_stat_clear(spu, 0, CLASS0_INTR_MASK); 9379476141cSMasato Noguchi spu_int_stat_clear(spu, 2, CLASS2_INTR_MASK); 9387c038749SMark Nutter local_irq_restore(flags); 9397c038749SMark Nutter } 9407c038749SMark Nutter 9417c038749SMark Nutter static inline int check_save_status(struct spu_state *csa, struct spu *spu) 9427c038749SMark Nutter { 9437c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 9447c038749SMark Nutter u32 complete; 9457c038749SMark Nutter 9467c038749SMark Nutter /* Save, Step 54: 9477c038749SMark Nutter * If SPU_Status[P]=1 and SPU_Status[SC] = "success", 9487c038749SMark Nutter * context save succeeded, otherwise context save 9497c038749SMark Nutter * failed. 9507c038749SMark Nutter */ 9517c038749SMark Nutter complete = ((SPU_SAVE_COMPLETE << SPU_STOP_STATUS_SHIFT) | 9527c038749SMark Nutter SPU_STATUS_STOPPED_BY_STOP); 9537c038749SMark Nutter return (in_be32(&prob->spu_status_R) != complete) ? 1 : 0; 9547c038749SMark Nutter } 9557c038749SMark Nutter 9567c038749SMark Nutter static inline void terminate_spu_app(struct spu_state *csa, struct spu *spu) 9577c038749SMark Nutter { 9587c038749SMark Nutter /* Restore, Step 4: 9597c038749SMark Nutter * If required, notify the "using application" that 9607c038749SMark Nutter * the SPU task has been terminated. TBD. 9617c038749SMark Nutter */ 9627c038749SMark Nutter } 9637c038749SMark Nutter 964cf17df22SMasato Noguchi static inline void suspend_mfc_and_halt_decr(struct spu_state *csa, 965cf17df22SMasato Noguchi struct spu *spu) 9667c038749SMark Nutter { 9677c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 9687c038749SMark Nutter 9697c038749SMark Nutter /* Restore, Step 7: 970cf17df22SMasato Noguchi * Write MFC_Cntl[Dh,Sc,Sm]='1','1','0' to suspend 9717c038749SMark Nutter * the queue and halt the decrementer. 9727c038749SMark Nutter */ 9737c038749SMark Nutter out_be64(&priv2->mfc_control_RW, MFC_CNTL_SUSPEND_DMA_QUEUE | 9747c038749SMark Nutter MFC_CNTL_DECREMENTER_HALTED); 9757c038749SMark Nutter eieio(); 9767c038749SMark Nutter } 9777c038749SMark Nutter 9787c038749SMark Nutter static inline void wait_suspend_mfc_complete(struct spu_state *csa, 9797c038749SMark Nutter struct spu *spu) 9807c038749SMark Nutter { 9817c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 9827c038749SMark Nutter 9837c038749SMark Nutter /* Restore, Step 8: 9847c038749SMark Nutter * Restore, Step 47. 9857c038749SMark Nutter * Poll MFC_CNTL[Ss] until 11 is returned. 9867c038749SMark Nutter */ 987910ab66bSBenjamin Herrenschmidt POLL_WHILE_FALSE((in_be64(&priv2->mfc_control_RW) & 988910ab66bSBenjamin Herrenschmidt MFC_CNTL_SUSPEND_DMA_STATUS_MASK) == 9897c038749SMark Nutter MFC_CNTL_SUSPEND_COMPLETE); 9907c038749SMark Nutter } 9917c038749SMark Nutter 9927c038749SMark Nutter static inline int suspend_spe(struct spu_state *csa, struct spu *spu) 9937c038749SMark Nutter { 9947c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 9957c038749SMark Nutter 9967c038749SMark Nutter /* Restore, Step 9: 9977c038749SMark Nutter * If SPU_Status[R]=1, stop SPU execution 9987c038749SMark Nutter * and wait for stop to complete. 9997c038749SMark Nutter * 10007c038749SMark Nutter * Returns 1 if SPU_Status[R]=1 on entry. 10017c038749SMark Nutter * 0 otherwise 10027c038749SMark Nutter */ 10037c038749SMark Nutter if (in_be32(&prob->spu_status_R) & SPU_STATUS_RUNNING) { 10047c038749SMark Nutter if (in_be32(&prob->spu_status_R) & 1005eb758ce5Sarnd@arndb.de SPU_STATUS_ISOLATED_EXIT_STATUS) { 10067c038749SMark Nutter POLL_WHILE_TRUE(in_be32(&prob->spu_status_R) & 10077c038749SMark Nutter SPU_STATUS_RUNNING); 10087c038749SMark Nutter } 10097c038749SMark Nutter if ((in_be32(&prob->spu_status_R) & 1010eb758ce5Sarnd@arndb.de SPU_STATUS_ISOLATED_LOAD_STATUS) 10117c038749SMark Nutter || (in_be32(&prob->spu_status_R) & 10127c038749SMark Nutter SPU_STATUS_ISOLATED_STATE)) { 10137c038749SMark Nutter out_be32(&prob->spu_runcntl_RW, SPU_RUNCNTL_STOP); 10147c038749SMark Nutter eieio(); 10157c038749SMark Nutter POLL_WHILE_TRUE(in_be32(&prob->spu_status_R) & 10167c038749SMark Nutter SPU_STATUS_RUNNING); 10177c038749SMark Nutter out_be32(&prob->spu_runcntl_RW, 0x2); 10187c038749SMark Nutter eieio(); 10197c038749SMark Nutter POLL_WHILE_TRUE(in_be32(&prob->spu_status_R) & 10207c038749SMark Nutter SPU_STATUS_RUNNING); 10217c038749SMark Nutter } 10227c038749SMark Nutter if (in_be32(&prob->spu_status_R) & 10237c038749SMark Nutter SPU_STATUS_WAITING_FOR_CHANNEL) { 10247c038749SMark Nutter out_be32(&prob->spu_runcntl_RW, SPU_RUNCNTL_STOP); 10257c038749SMark Nutter eieio(); 10267c038749SMark Nutter POLL_WHILE_TRUE(in_be32(&prob->spu_status_R) & 10277c038749SMark Nutter SPU_STATUS_RUNNING); 10287c038749SMark Nutter } 10297c038749SMark Nutter return 1; 10307c038749SMark Nutter } 10317c038749SMark Nutter return 0; 10327c038749SMark Nutter } 10337c038749SMark Nutter 10347c038749SMark Nutter static inline void clear_spu_status(struct spu_state *csa, struct spu *spu) 10357c038749SMark Nutter { 10367c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 10377c038749SMark Nutter 10387c038749SMark Nutter /* Restore, Step 10: 10397c038749SMark Nutter * If SPU_Status[R]=0 and SPU_Status[E,L,IS]=1, 10407c038749SMark Nutter * release SPU from isolate state. 10417c038749SMark Nutter */ 10427c038749SMark Nutter if (!(in_be32(&prob->spu_status_R) & SPU_STATUS_RUNNING)) { 10437c038749SMark Nutter if (in_be32(&prob->spu_status_R) & 1044eb758ce5Sarnd@arndb.de SPU_STATUS_ISOLATED_EXIT_STATUS) { 1045f0831accSArnd Bergmann spu_mfc_sr1_set(spu, 10467c038749SMark Nutter MFC_STATE1_MASTER_RUN_CONTROL_MASK); 10477c038749SMark Nutter eieio(); 10487c038749SMark Nutter out_be32(&prob->spu_runcntl_RW, SPU_RUNCNTL_RUNNABLE); 10497c038749SMark Nutter eieio(); 10507c038749SMark Nutter POLL_WHILE_TRUE(in_be32(&prob->spu_status_R) & 10517c038749SMark Nutter SPU_STATUS_RUNNING); 10527c038749SMark Nutter } 10537c038749SMark Nutter if ((in_be32(&prob->spu_status_R) & 1054eb758ce5Sarnd@arndb.de SPU_STATUS_ISOLATED_LOAD_STATUS) 10557c038749SMark Nutter || (in_be32(&prob->spu_status_R) & 10567c038749SMark Nutter SPU_STATUS_ISOLATED_STATE)) { 1057f0831accSArnd Bergmann spu_mfc_sr1_set(spu, 10587c038749SMark Nutter MFC_STATE1_MASTER_RUN_CONTROL_MASK); 10597c038749SMark Nutter eieio(); 10607c038749SMark Nutter out_be32(&prob->spu_runcntl_RW, 0x2); 10617c038749SMark Nutter eieio(); 10627c038749SMark Nutter POLL_WHILE_TRUE(in_be32(&prob->spu_status_R) & 10637c038749SMark Nutter SPU_STATUS_RUNNING); 10647c038749SMark Nutter } 10657c038749SMark Nutter } 10667c038749SMark Nutter } 10677c038749SMark Nutter 10687c038749SMark Nutter static inline void reset_ch_part1(struct spu_state *csa, struct spu *spu) 10697c038749SMark Nutter { 10707c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 1071daced0f7SJeremy Kerr u64 ch_indices[] = { 0UL, 3UL, 4UL, 24UL, 25UL, 27UL }; 10727c038749SMark Nutter u64 idx; 10737c038749SMark Nutter int i; 10747c038749SMark Nutter 10757c038749SMark Nutter /* Restore, Step 20: 10767c038749SMark Nutter */ 1077e46a0237Sarnd@arndb.de 1078e46a0237Sarnd@arndb.de /* Reset CH 1 */ 1079e46a0237Sarnd@arndb.de out_be64(&priv2->spu_chnlcntptr_RW, 1); 1080e46a0237Sarnd@arndb.de out_be64(&priv2->spu_chnldata_RW, 0UL); 1081e46a0237Sarnd@arndb.de 1082e46a0237Sarnd@arndb.de /* Reset the following CH: [0,3,4,24,25,27] */ 1083daced0f7SJeremy Kerr for (i = 0; i < ARRAY_SIZE(ch_indices); i++) { 10847c038749SMark Nutter idx = ch_indices[i]; 10857c038749SMark Nutter out_be64(&priv2->spu_chnlcntptr_RW, idx); 10867c038749SMark Nutter eieio(); 10877c038749SMark Nutter out_be64(&priv2->spu_chnldata_RW, 0UL); 10887c038749SMark Nutter out_be64(&priv2->spu_chnlcnt_RW, 0UL); 10897c038749SMark Nutter eieio(); 10907c038749SMark Nutter } 10917c038749SMark Nutter } 10927c038749SMark Nutter 10937c038749SMark Nutter static inline void reset_ch_part2(struct spu_state *csa, struct spu *spu) 10947c038749SMark Nutter { 10957c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 10967c038749SMark Nutter u64 ch_indices[5] = { 21UL, 23UL, 28UL, 29UL, 30UL }; 10977c038749SMark Nutter u64 ch_counts[5] = { 16UL, 1UL, 1UL, 0UL, 1UL }; 10987c038749SMark Nutter u64 idx; 10997c038749SMark Nutter int i; 11007c038749SMark Nutter 11017c038749SMark Nutter /* Restore, Step 21: 11027c038749SMark Nutter * Reset the following CH: [21, 23, 28, 29, 30] 11037c038749SMark Nutter */ 11047c038749SMark Nutter for (i = 0; i < 5; i++) { 11057c038749SMark Nutter idx = ch_indices[i]; 11067c038749SMark Nutter out_be64(&priv2->spu_chnlcntptr_RW, idx); 11077c038749SMark Nutter eieio(); 11087c038749SMark Nutter out_be64(&priv2->spu_chnlcnt_RW, ch_counts[i]); 11097c038749SMark Nutter eieio(); 11107c038749SMark Nutter } 11117c038749SMark Nutter } 11127c038749SMark Nutter 11137c038749SMark Nutter static inline void setup_spu_status_part1(struct spu_state *csa, 11147c038749SMark Nutter struct spu *spu) 11157c038749SMark Nutter { 11167c038749SMark Nutter u32 status_P = SPU_STATUS_STOPPED_BY_STOP; 11177c038749SMark Nutter u32 status_I = SPU_STATUS_INVALID_INSTR; 11187c038749SMark Nutter u32 status_H = SPU_STATUS_STOPPED_BY_HALT; 11197c038749SMark Nutter u32 status_S = SPU_STATUS_SINGLE_STEP; 11207c038749SMark Nutter u32 status_S_I = SPU_STATUS_SINGLE_STEP | SPU_STATUS_INVALID_INSTR; 11217c038749SMark Nutter u32 status_S_P = SPU_STATUS_SINGLE_STEP | SPU_STATUS_STOPPED_BY_STOP; 11227c038749SMark Nutter u32 status_P_H = SPU_STATUS_STOPPED_BY_HALT |SPU_STATUS_STOPPED_BY_STOP; 11237c038749SMark Nutter u32 status_P_I = SPU_STATUS_STOPPED_BY_STOP |SPU_STATUS_INVALID_INSTR; 11247c038749SMark Nutter u32 status_code; 11257c038749SMark Nutter 11267c038749SMark Nutter /* Restore, Step 27: 11277c038749SMark Nutter * If the CSA.SPU_Status[I,S,H,P]=1 then add the correct 11287c038749SMark Nutter * instruction sequence to the end of the SPU based restore 11297c038749SMark Nutter * code (after the "context restored" stop and signal) to 11307c038749SMark Nutter * restore the correct SPU status. 11317c038749SMark Nutter * 11327c038749SMark Nutter * NOTE: Rather than modifying the SPU executable, we 11337c038749SMark Nutter * instead add a new 'stopped_status' field to the 11347c038749SMark Nutter * LSCSA. The SPU-side restore reads this field and 11357c038749SMark Nutter * takes the appropriate action when exiting. 11367c038749SMark Nutter */ 11377c038749SMark Nutter 11387c038749SMark Nutter status_code = 11397c038749SMark Nutter (csa->prob.spu_status_R >> SPU_STOP_STATUS_SHIFT) & 0xFFFF; 11407c038749SMark Nutter if ((csa->prob.spu_status_R & status_P_I) == status_P_I) { 11417c038749SMark Nutter 11427c038749SMark Nutter /* SPU_Status[P,I]=1 - Illegal Instruction followed 11437c038749SMark Nutter * by Stop and Signal instruction, followed by 'br -4'. 11447c038749SMark Nutter * 11457c038749SMark Nutter */ 11467c038749SMark Nutter csa->lscsa->stopped_status.slot[0] = SPU_STOPPED_STATUS_P_I; 11477c038749SMark Nutter csa->lscsa->stopped_status.slot[1] = status_code; 11487c038749SMark Nutter 11497c038749SMark Nutter } else if ((csa->prob.spu_status_R & status_P_H) == status_P_H) { 11507c038749SMark Nutter 11517c038749SMark Nutter /* SPU_Status[P,H]=1 - Halt Conditional, followed 11527c038749SMark Nutter * by Stop and Signal instruction, followed by 11537c038749SMark Nutter * 'br -4'. 11547c038749SMark Nutter */ 11557c038749SMark Nutter csa->lscsa->stopped_status.slot[0] = SPU_STOPPED_STATUS_P_H; 11567c038749SMark Nutter csa->lscsa->stopped_status.slot[1] = status_code; 11577c038749SMark Nutter 11587c038749SMark Nutter } else if ((csa->prob.spu_status_R & status_S_P) == status_S_P) { 11597c038749SMark Nutter 11607c038749SMark Nutter /* SPU_Status[S,P]=1 - Stop and Signal instruction 11617c038749SMark Nutter * followed by 'br -4'. 11627c038749SMark Nutter */ 11637c038749SMark Nutter csa->lscsa->stopped_status.slot[0] = SPU_STOPPED_STATUS_S_P; 11647c038749SMark Nutter csa->lscsa->stopped_status.slot[1] = status_code; 11657c038749SMark Nutter 11667c038749SMark Nutter } else if ((csa->prob.spu_status_R & status_S_I) == status_S_I) { 11677c038749SMark Nutter 11687c038749SMark Nutter /* SPU_Status[S,I]=1 - Illegal instruction followed 11697c038749SMark Nutter * by 'br -4'. 11707c038749SMark Nutter */ 11717c038749SMark Nutter csa->lscsa->stopped_status.slot[0] = SPU_STOPPED_STATUS_S_I; 11727c038749SMark Nutter csa->lscsa->stopped_status.slot[1] = status_code; 11737c038749SMark Nutter 11747c038749SMark Nutter } else if ((csa->prob.spu_status_R & status_P) == status_P) { 11757c038749SMark Nutter 11767c038749SMark Nutter /* SPU_Status[P]=1 - Stop and Signal instruction 11777c038749SMark Nutter * followed by 'br -4'. 11787c038749SMark Nutter */ 11797c038749SMark Nutter csa->lscsa->stopped_status.slot[0] = SPU_STOPPED_STATUS_P; 11807c038749SMark Nutter csa->lscsa->stopped_status.slot[1] = status_code; 11817c038749SMark Nutter 11827c038749SMark Nutter } else if ((csa->prob.spu_status_R & status_H) == status_H) { 11837c038749SMark Nutter 11847c038749SMark Nutter /* SPU_Status[H]=1 - Halt Conditional, followed 11857c038749SMark Nutter * by 'br -4'. 11867c038749SMark Nutter */ 11877c038749SMark Nutter csa->lscsa->stopped_status.slot[0] = SPU_STOPPED_STATUS_H; 11887c038749SMark Nutter 11897c038749SMark Nutter } else if ((csa->prob.spu_status_R & status_S) == status_S) { 11907c038749SMark Nutter 11917c038749SMark Nutter /* SPU_Status[S]=1 - Two nop instructions. 11927c038749SMark Nutter */ 11937c038749SMark Nutter csa->lscsa->stopped_status.slot[0] = SPU_STOPPED_STATUS_S; 11947c038749SMark Nutter 11957c038749SMark Nutter } else if ((csa->prob.spu_status_R & status_I) == status_I) { 11967c038749SMark Nutter 11977c038749SMark Nutter /* SPU_Status[I]=1 - Illegal instruction followed 11987c038749SMark Nutter * by 'br -4'. 11997c038749SMark Nutter */ 12007c038749SMark Nutter csa->lscsa->stopped_status.slot[0] = SPU_STOPPED_STATUS_I; 12017c038749SMark Nutter 12027c038749SMark Nutter } 12037c038749SMark Nutter } 12047c038749SMark Nutter 12057c038749SMark Nutter static inline void setup_spu_status_part2(struct spu_state *csa, 12067c038749SMark Nutter struct spu *spu) 12077c038749SMark Nutter { 12087c038749SMark Nutter u32 mask; 12097c038749SMark Nutter 12107c038749SMark Nutter /* Restore, Step 28: 12117c038749SMark Nutter * If the CSA.SPU_Status[I,S,H,P,R]=0 then 12127c038749SMark Nutter * add a 'br *' instruction to the end of 12137c038749SMark Nutter * the SPU based restore code. 12147c038749SMark Nutter * 12157c038749SMark Nutter * NOTE: Rather than modifying the SPU executable, we 12167c038749SMark Nutter * instead add a new 'stopped_status' field to the 12177c038749SMark Nutter * LSCSA. The SPU-side restore reads this field and 12187c038749SMark Nutter * takes the appropriate action when exiting. 12197c038749SMark Nutter */ 12207c038749SMark Nutter mask = SPU_STATUS_INVALID_INSTR | 12217c038749SMark Nutter SPU_STATUS_SINGLE_STEP | 12227c038749SMark Nutter SPU_STATUS_STOPPED_BY_HALT | 12237c038749SMark Nutter SPU_STATUS_STOPPED_BY_STOP | SPU_STATUS_RUNNING; 12247c038749SMark Nutter if (!(csa->prob.spu_status_R & mask)) { 12257c038749SMark Nutter csa->lscsa->stopped_status.slot[0] = SPU_STOPPED_STATUS_R; 12267c038749SMark Nutter } 12277c038749SMark Nutter } 12287c038749SMark Nutter 12297c038749SMark Nutter static inline void restore_mfc_rag(struct spu_state *csa, struct spu *spu) 12307c038749SMark Nutter { 12317c038749SMark Nutter /* Restore, Step 29: 12327c038749SMark Nutter * Restore RA_GROUP_ID register and the 12337c038749SMark Nutter * RA_ENABLE reigster from the CSA. 12347c038749SMark Nutter */ 1235f0831accSArnd Bergmann spu_resource_allocation_groupID_set(spu, 12367c038749SMark Nutter csa->priv1.resource_allocation_groupID_RW); 1237f0831accSArnd Bergmann spu_resource_allocation_enable_set(spu, 12387c038749SMark Nutter csa->priv1.resource_allocation_enable_RW); 12397c038749SMark Nutter } 12407c038749SMark Nutter 12417c038749SMark Nutter static inline void send_restore_code(struct spu_state *csa, struct spu *spu) 12427c038749SMark Nutter { 12437c038749SMark Nutter unsigned long addr = (unsigned long)&spu_restore_code[0]; 12447c038749SMark Nutter unsigned int ls_offset = 0x0; 12457c038749SMark Nutter unsigned int size = sizeof(spu_restore_code); 12467c038749SMark Nutter unsigned int tag = 0; 12477c038749SMark Nutter unsigned int rclass = 0; 12487c038749SMark Nutter unsigned int cmd = MFC_GETFS_CMD; 12497c038749SMark Nutter 12507c038749SMark Nutter /* Restore, Step 37: 12517c038749SMark Nutter * Issue MFC DMA command to copy context 12527c038749SMark Nutter * restore code to local storage. 12537c038749SMark Nutter */ 12547c038749SMark Nutter send_mfc_dma(spu, addr, ls_offset, size, tag, rclass, cmd); 12557c038749SMark Nutter } 12567c038749SMark Nutter 12577c038749SMark Nutter static inline void setup_decr(struct spu_state *csa, struct spu *spu) 12587c038749SMark Nutter { 12597c038749SMark Nutter /* Restore, Step 34: 12607c038749SMark Nutter * If CSA.MFC_CNTL[Ds]=1 (decrementer was 12617c038749SMark Nutter * running) then adjust decrementer, set 12627c038749SMark Nutter * decrementer running status in LSCSA, 12637c038749SMark Nutter * and set decrementer "wrapped" status 12647c038749SMark Nutter * in LSCSA. 12657c038749SMark Nutter */ 12667c038749SMark Nutter if (csa->priv2.mfc_control_RW & MFC_CNTL_DECREMENTER_RUNNING) { 12677c038749SMark Nutter cycles_t resume_time = get_cycles(); 12687c038749SMark Nutter cycles_t delta_time = resume_time - csa->suspend_time; 12697c038749SMark Nutter 12701cfc0f86SMasato Noguchi csa->lscsa->decr_status.slot[0] = SPU_DECR_STATUS_RUNNING; 12711cfc0f86SMasato Noguchi if (csa->lscsa->decr.slot[0] < delta_time) { 12721cfc0f86SMasato Noguchi csa->lscsa->decr_status.slot[0] |= 12731cfc0f86SMasato Noguchi SPU_DECR_STATUS_WRAPPED; 12741cfc0f86SMasato Noguchi } 12751cfc0f86SMasato Noguchi 1276183b73aeSJordi Caubet csa->lscsa->decr.slot[0] -= delta_time; 12771cfc0f86SMasato Noguchi } else { 12781cfc0f86SMasato Noguchi csa->lscsa->decr_status.slot[0] = 0; 12797c038749SMark Nutter } 12807c038749SMark Nutter } 12817c038749SMark Nutter 12827c038749SMark Nutter static inline void setup_ppu_mb(struct spu_state *csa, struct spu *spu) 12837c038749SMark Nutter { 12847c038749SMark Nutter /* Restore, Step 35: 12857c038749SMark Nutter * Copy the CSA.PU_MB data into the LSCSA. 12867c038749SMark Nutter */ 12877c038749SMark Nutter csa->lscsa->ppu_mb.slot[0] = csa->prob.pu_mb_R; 12887c038749SMark Nutter } 12897c038749SMark Nutter 12907c038749SMark Nutter static inline void setup_ppuint_mb(struct spu_state *csa, struct spu *spu) 12917c038749SMark Nutter { 12927c038749SMark Nutter /* Restore, Step 36: 12937c038749SMark Nutter * Copy the CSA.PUINT_MB data into the LSCSA. 12947c038749SMark Nutter */ 12957c038749SMark Nutter csa->lscsa->ppuint_mb.slot[0] = csa->priv2.puint_mb_R; 12967c038749SMark Nutter } 12977c038749SMark Nutter 12987c038749SMark Nutter static inline int check_restore_status(struct spu_state *csa, struct spu *spu) 12997c038749SMark Nutter { 13007c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 13017c038749SMark Nutter u32 complete; 13027c038749SMark Nutter 13037c038749SMark Nutter /* Restore, Step 40: 13047c038749SMark Nutter * If SPU_Status[P]=1 and SPU_Status[SC] = "success", 13057c038749SMark Nutter * context restore succeeded, otherwise context restore 13067c038749SMark Nutter * failed. 13077c038749SMark Nutter */ 13087c038749SMark Nutter complete = ((SPU_RESTORE_COMPLETE << SPU_STOP_STATUS_SHIFT) | 13097c038749SMark Nutter SPU_STATUS_STOPPED_BY_STOP); 13107c038749SMark Nutter return (in_be32(&prob->spu_status_R) != complete) ? 1 : 0; 13117c038749SMark Nutter } 13127c038749SMark Nutter 13137c038749SMark Nutter static inline void restore_spu_privcntl(struct spu_state *csa, struct spu *spu) 13147c038749SMark Nutter { 13157c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 13167c038749SMark Nutter 13177c038749SMark Nutter /* Restore, Step 41: 13187c038749SMark Nutter * Restore SPU_PrivCntl from the CSA. 13197c038749SMark Nutter */ 13207c038749SMark Nutter out_be64(&priv2->spu_privcntl_RW, csa->priv2.spu_privcntl_RW); 13217c038749SMark Nutter eieio(); 13227c038749SMark Nutter } 13237c038749SMark Nutter 13247c038749SMark Nutter static inline void restore_status_part1(struct spu_state *csa, struct spu *spu) 13257c038749SMark Nutter { 13267c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 13277c038749SMark Nutter u32 mask; 13287c038749SMark Nutter 13297c038749SMark Nutter /* Restore, Step 42: 13307c038749SMark Nutter * If any CSA.SPU_Status[I,S,H,P]=1, then 13317c038749SMark Nutter * restore the error or single step state. 13327c038749SMark Nutter */ 13337c038749SMark Nutter mask = SPU_STATUS_INVALID_INSTR | 13347c038749SMark Nutter SPU_STATUS_SINGLE_STEP | 13357c038749SMark Nutter SPU_STATUS_STOPPED_BY_HALT | SPU_STATUS_STOPPED_BY_STOP; 13367c038749SMark Nutter if (csa->prob.spu_status_R & mask) { 13377c038749SMark Nutter out_be32(&prob->spu_runcntl_RW, SPU_RUNCNTL_RUNNABLE); 13387c038749SMark Nutter eieio(); 13397c038749SMark Nutter POLL_WHILE_TRUE(in_be32(&prob->spu_status_R) & 13407c038749SMark Nutter SPU_STATUS_RUNNING); 13417c038749SMark Nutter } 13427c038749SMark Nutter } 13437c038749SMark Nutter 13447c038749SMark Nutter static inline void restore_status_part2(struct spu_state *csa, struct spu *spu) 13457c038749SMark Nutter { 13467c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 13477c038749SMark Nutter u32 mask; 13487c038749SMark Nutter 13497c038749SMark Nutter /* Restore, Step 43: 13507c038749SMark Nutter * If all CSA.SPU_Status[I,S,H,P,R]=0 then write 13517c038749SMark Nutter * SPU_RunCntl[R0R1]='01', wait for SPU_Status[R]=1, 13527c038749SMark Nutter * then write '00' to SPU_RunCntl[R0R1] and wait 13537c038749SMark Nutter * for SPU_Status[R]=0. 13547c038749SMark Nutter */ 13557c038749SMark Nutter mask = SPU_STATUS_INVALID_INSTR | 13567c038749SMark Nutter SPU_STATUS_SINGLE_STEP | 13577c038749SMark Nutter SPU_STATUS_STOPPED_BY_HALT | 13587c038749SMark Nutter SPU_STATUS_STOPPED_BY_STOP | SPU_STATUS_RUNNING; 13597c038749SMark Nutter if (!(csa->prob.spu_status_R & mask)) { 13607c038749SMark Nutter out_be32(&prob->spu_runcntl_RW, SPU_RUNCNTL_RUNNABLE); 13617c038749SMark Nutter eieio(); 13627c038749SMark Nutter POLL_WHILE_FALSE(in_be32(&prob->spu_status_R) & 13637c038749SMark Nutter SPU_STATUS_RUNNING); 13647c038749SMark Nutter out_be32(&prob->spu_runcntl_RW, SPU_RUNCNTL_STOP); 13657c038749SMark Nutter eieio(); 13667c038749SMark Nutter POLL_WHILE_TRUE(in_be32(&prob->spu_status_R) & 13677c038749SMark Nutter SPU_STATUS_RUNNING); 13687c038749SMark Nutter } 13697c038749SMark Nutter } 13707c038749SMark Nutter 13717c038749SMark Nutter static inline void restore_ls_16kb(struct spu_state *csa, struct spu *spu) 13727c038749SMark Nutter { 13737c038749SMark Nutter unsigned long addr = (unsigned long)&csa->lscsa->ls[0]; 13747c038749SMark Nutter unsigned int ls_offset = 0x0; 13757c038749SMark Nutter unsigned int size = 16384; 13767c038749SMark Nutter unsigned int tag = 0; 13777c038749SMark Nutter unsigned int rclass = 0; 13787c038749SMark Nutter unsigned int cmd = MFC_GET_CMD; 13797c038749SMark Nutter 13807c038749SMark Nutter /* Restore, Step 44: 13817c038749SMark Nutter * Issue a DMA command to restore the first 13827c038749SMark Nutter * 16kb of local storage from CSA. 13837c038749SMark Nutter */ 13847c038749SMark Nutter send_mfc_dma(spu, addr, ls_offset, size, tag, rclass, cmd); 13857c038749SMark Nutter } 13867c038749SMark Nutter 1387cf17df22SMasato Noguchi static inline void suspend_mfc(struct spu_state *csa, struct spu *spu) 1388cf17df22SMasato Noguchi { 1389cf17df22SMasato Noguchi struct spu_priv2 __iomem *priv2 = spu->priv2; 1390cf17df22SMasato Noguchi 1391cf17df22SMasato Noguchi /* Restore, Step 47. 1392cf17df22SMasato Noguchi * Write MFC_Cntl[Sc,Sm]='1','0' to suspend 1393cf17df22SMasato Noguchi * the queue. 1394cf17df22SMasato Noguchi */ 1395cf17df22SMasato Noguchi out_be64(&priv2->mfc_control_RW, MFC_CNTL_SUSPEND_DMA_QUEUE); 1396cf17df22SMasato Noguchi eieio(); 1397cf17df22SMasato Noguchi } 1398cf17df22SMasato Noguchi 13997c038749SMark Nutter static inline void clear_interrupts(struct spu_state *csa, struct spu *spu) 14007c038749SMark Nutter { 14017c038749SMark Nutter /* Restore, Step 49: 14027c038749SMark Nutter * Write INT_MASK_class0 with value of 0. 14037c038749SMark Nutter * Write INT_MASK_class1 with value of 0. 14047c038749SMark Nutter * Write INT_MASK_class2 with value of 0. 14057c038749SMark Nutter * Write INT_STAT_class0 with value of -1. 14067c038749SMark Nutter * Write INT_STAT_class1 with value of -1. 14077c038749SMark Nutter * Write INT_STAT_class2 with value of -1. 14087c038749SMark Nutter */ 14097c038749SMark Nutter spin_lock_irq(&spu->register_lock); 1410f0831accSArnd Bergmann spu_int_mask_set(spu, 0, 0ul); 1411f0831accSArnd Bergmann spu_int_mask_set(spu, 1, 0ul); 1412f0831accSArnd Bergmann spu_int_mask_set(spu, 2, 0ul); 14139476141cSMasato Noguchi spu_int_stat_clear(spu, 0, CLASS0_INTR_MASK); 14149476141cSMasato Noguchi spu_int_stat_clear(spu, 1, CLASS1_INTR_MASK); 14159476141cSMasato Noguchi spu_int_stat_clear(spu, 2, CLASS2_INTR_MASK); 14167c038749SMark Nutter spin_unlock_irq(&spu->register_lock); 14177c038749SMark Nutter } 14187c038749SMark Nutter 14197c038749SMark Nutter static inline void restore_mfc_queues(struct spu_state *csa, struct spu *spu) 14207c038749SMark Nutter { 14217c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 14227c038749SMark Nutter int i; 14237c038749SMark Nutter 14247c038749SMark Nutter /* Restore, Step 50: 14257c038749SMark Nutter * If MFC_Cntl[Se]!=0 then restore 14267c038749SMark Nutter * MFC command queues. 14277c038749SMark Nutter */ 14287c038749SMark Nutter if ((csa->priv2.mfc_control_RW & MFC_CNTL_DMA_QUEUES_EMPTY_MASK) == 0) { 14297c038749SMark Nutter for (i = 0; i < 8; i++) { 14307c038749SMark Nutter out_be64(&priv2->puq[i].mfc_cq_data0_RW, 14317c038749SMark Nutter csa->priv2.puq[i].mfc_cq_data0_RW); 14327c038749SMark Nutter out_be64(&priv2->puq[i].mfc_cq_data1_RW, 14337c038749SMark Nutter csa->priv2.puq[i].mfc_cq_data1_RW); 14347c038749SMark Nutter out_be64(&priv2->puq[i].mfc_cq_data2_RW, 14357c038749SMark Nutter csa->priv2.puq[i].mfc_cq_data2_RW); 14367c038749SMark Nutter out_be64(&priv2->puq[i].mfc_cq_data3_RW, 14377c038749SMark Nutter csa->priv2.puq[i].mfc_cq_data3_RW); 14387c038749SMark Nutter } 14397c038749SMark Nutter for (i = 0; i < 16; i++) { 14407c038749SMark Nutter out_be64(&priv2->spuq[i].mfc_cq_data0_RW, 14417c038749SMark Nutter csa->priv2.spuq[i].mfc_cq_data0_RW); 14427c038749SMark Nutter out_be64(&priv2->spuq[i].mfc_cq_data1_RW, 14437c038749SMark Nutter csa->priv2.spuq[i].mfc_cq_data1_RW); 14447c038749SMark Nutter out_be64(&priv2->spuq[i].mfc_cq_data2_RW, 14457c038749SMark Nutter csa->priv2.spuq[i].mfc_cq_data2_RW); 14467c038749SMark Nutter out_be64(&priv2->spuq[i].mfc_cq_data3_RW, 14477c038749SMark Nutter csa->priv2.spuq[i].mfc_cq_data3_RW); 14487c038749SMark Nutter } 14497c038749SMark Nutter } 14507c038749SMark Nutter eieio(); 14517c038749SMark Nutter } 14527c038749SMark Nutter 14537c038749SMark Nutter static inline void restore_ppu_querymask(struct spu_state *csa, struct spu *spu) 14547c038749SMark Nutter { 14557c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 14567c038749SMark Nutter 14577c038749SMark Nutter /* Restore, Step 51: 14587c038749SMark Nutter * Restore the PPU_QueryMask register from CSA. 14597c038749SMark Nutter */ 14607c038749SMark Nutter out_be32(&prob->dma_querymask_RW, csa->prob.dma_querymask_RW); 14617c038749SMark Nutter eieio(); 14627c038749SMark Nutter } 14637c038749SMark Nutter 14647c038749SMark Nutter static inline void restore_ppu_querytype(struct spu_state *csa, struct spu *spu) 14657c038749SMark Nutter { 14667c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 14677c038749SMark Nutter 14687c038749SMark Nutter /* Restore, Step 52: 14697c038749SMark Nutter * Restore the PPU_QueryType register from CSA. 14707c038749SMark Nutter */ 14717c038749SMark Nutter out_be32(&prob->dma_querytype_RW, csa->prob.dma_querytype_RW); 14727c038749SMark Nutter eieio(); 14737c038749SMark Nutter } 14747c038749SMark Nutter 14757c038749SMark Nutter static inline void restore_mfc_csr_tsq(struct spu_state *csa, struct spu *spu) 14767c038749SMark Nutter { 14777c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 14787c038749SMark Nutter 14797c038749SMark Nutter /* Restore, Step 53: 14807c038749SMark Nutter * Restore the MFC_CSR_TSQ register from CSA. 14817c038749SMark Nutter */ 14827c038749SMark Nutter out_be64(&priv2->spu_tag_status_query_RW, 14837c038749SMark Nutter csa->priv2.spu_tag_status_query_RW); 14847c038749SMark Nutter eieio(); 14857c038749SMark Nutter } 14867c038749SMark Nutter 14877c038749SMark Nutter static inline void restore_mfc_csr_cmd(struct spu_state *csa, struct spu *spu) 14887c038749SMark Nutter { 14897c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 14907c038749SMark Nutter 14917c038749SMark Nutter /* Restore, Step 54: 14927c038749SMark Nutter * Restore the MFC_CSR_CMD1 and MFC_CSR_CMD2 14937c038749SMark Nutter * registers from CSA. 14947c038749SMark Nutter */ 14957c038749SMark Nutter out_be64(&priv2->spu_cmd_buf1_RW, csa->priv2.spu_cmd_buf1_RW); 14967c038749SMark Nutter out_be64(&priv2->spu_cmd_buf2_RW, csa->priv2.spu_cmd_buf2_RW); 14977c038749SMark Nutter eieio(); 14987c038749SMark Nutter } 14997c038749SMark Nutter 15007c038749SMark Nutter static inline void restore_mfc_csr_ato(struct spu_state *csa, struct spu *spu) 15017c038749SMark Nutter { 15027c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 15037c038749SMark Nutter 15047c038749SMark Nutter /* Restore, Step 55: 15057c038749SMark Nutter * Restore the MFC_CSR_ATO register from CSA. 15067c038749SMark Nutter */ 15077c038749SMark Nutter out_be64(&priv2->spu_atomic_status_RW, csa->priv2.spu_atomic_status_RW); 15087c038749SMark Nutter } 15097c038749SMark Nutter 15107c038749SMark Nutter static inline void restore_mfc_tclass_id(struct spu_state *csa, struct spu *spu) 15117c038749SMark Nutter { 15127c038749SMark Nutter /* Restore, Step 56: 15137c038749SMark Nutter * Restore the MFC_TCLASS_ID register from CSA. 15147c038749SMark Nutter */ 1515f0831accSArnd Bergmann spu_mfc_tclass_id_set(spu, csa->priv1.mfc_tclass_id_RW); 15167c038749SMark Nutter eieio(); 15177c038749SMark Nutter } 15187c038749SMark Nutter 15197c038749SMark Nutter static inline void set_llr_event(struct spu_state *csa, struct spu *spu) 15207c038749SMark Nutter { 15217c038749SMark Nutter u64 ch0_cnt, ch0_data; 15227c038749SMark Nutter u64 ch1_data; 15237c038749SMark Nutter 15247c038749SMark Nutter /* Restore, Step 57: 15257c038749SMark Nutter * Set the Lock Line Reservation Lost Event by: 15267c038749SMark Nutter * 1. OR CSA.SPU_Event_Status with bit 21 (Lr) set to 1. 15277c038749SMark Nutter * 2. If CSA.SPU_Channel_0_Count=0 and 15287c038749SMark Nutter * CSA.SPU_Wr_Event_Mask[Lr]=1 and 15297c038749SMark Nutter * CSA.SPU_Event_Status[Lr]=0 then set 15307c038749SMark Nutter * CSA.SPU_Event_Status_Count=1. 15317c038749SMark Nutter */ 15327c038749SMark Nutter ch0_cnt = csa->spu_chnlcnt_RW[0]; 15337c038749SMark Nutter ch0_data = csa->spu_chnldata_RW[0]; 15347c038749SMark Nutter ch1_data = csa->spu_chnldata_RW[1]; 15357c038749SMark Nutter csa->spu_chnldata_RW[0] |= MFC_LLR_LOST_EVENT; 15367c038749SMark Nutter if ((ch0_cnt == 0) && !(ch0_data & MFC_LLR_LOST_EVENT) && 15377c038749SMark Nutter (ch1_data & MFC_LLR_LOST_EVENT)) { 15387c038749SMark Nutter csa->spu_chnlcnt_RW[0] = 1; 15397c038749SMark Nutter } 15407c038749SMark Nutter } 15417c038749SMark Nutter 15427c038749SMark Nutter static inline void restore_decr_wrapped(struct spu_state *csa, struct spu *spu) 15437c038749SMark Nutter { 15447c038749SMark Nutter /* Restore, Step 58: 15457c038749SMark Nutter * If the status of the CSA software decrementer 15467c038749SMark Nutter * "wrapped" flag is set, OR in a '1' to 15477c038749SMark Nutter * CSA.SPU_Event_Status[Tm]. 15487c038749SMark Nutter */ 154905a059f3SJeremy Kerr if (!(csa->lscsa->decr_status.slot[0] & SPU_DECR_STATUS_WRAPPED)) 155005a059f3SJeremy Kerr return; 155105a059f3SJeremy Kerr 155205a059f3SJeremy Kerr if ((csa->spu_chnlcnt_RW[0] == 0) && 155305a059f3SJeremy Kerr (csa->spu_chnldata_RW[1] & 0x20) && 155405a059f3SJeremy Kerr !(csa->spu_chnldata_RW[0] & 0x20)) 15557c038749SMark Nutter csa->spu_chnlcnt_RW[0] = 1; 155605a059f3SJeremy Kerr 155705a059f3SJeremy Kerr csa->spu_chnldata_RW[0] |= 0x20; 15587c038749SMark Nutter } 15597c038749SMark Nutter 15607c038749SMark Nutter static inline void restore_ch_part1(struct spu_state *csa, struct spu *spu) 15617c038749SMark Nutter { 15627c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 1563daced0f7SJeremy Kerr u64 idx, ch_indices[] = { 0UL, 3UL, 4UL, 24UL, 25UL, 27UL }; 15647c038749SMark Nutter int i; 15657c038749SMark Nutter 15667c038749SMark Nutter /* Restore, Step 59: 1567cfd529b2SMasato Noguchi * Restore the following CH: [0,3,4,24,25,27] 15687c038749SMark Nutter */ 1569daced0f7SJeremy Kerr for (i = 0; i < ARRAY_SIZE(ch_indices); i++) { 15707c038749SMark Nutter idx = ch_indices[i]; 15717c038749SMark Nutter out_be64(&priv2->spu_chnlcntptr_RW, idx); 15727c038749SMark Nutter eieio(); 15737c038749SMark Nutter out_be64(&priv2->spu_chnldata_RW, csa->spu_chnldata_RW[idx]); 15747c038749SMark Nutter out_be64(&priv2->spu_chnlcnt_RW, csa->spu_chnlcnt_RW[idx]); 15757c038749SMark Nutter eieio(); 15767c038749SMark Nutter } 15777c038749SMark Nutter } 15787c038749SMark Nutter 15797c038749SMark Nutter static inline void restore_ch_part2(struct spu_state *csa, struct spu *spu) 15807c038749SMark Nutter { 15817c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 15827c038749SMark Nutter u64 ch_indices[3] = { 9UL, 21UL, 23UL }; 15837c038749SMark Nutter u64 ch_counts[3] = { 1UL, 16UL, 1UL }; 15847c038749SMark Nutter u64 idx; 15857c038749SMark Nutter int i; 15867c038749SMark Nutter 15877c038749SMark Nutter /* Restore, Step 60: 15887c038749SMark Nutter * Restore the following CH: [9,21,23]. 15897c038749SMark Nutter */ 15907c038749SMark Nutter ch_counts[0] = 1UL; 15917c038749SMark Nutter ch_counts[1] = csa->spu_chnlcnt_RW[21]; 15927c038749SMark Nutter ch_counts[2] = 1UL; 15937c038749SMark Nutter for (i = 0; i < 3; i++) { 15947c038749SMark Nutter idx = ch_indices[i]; 15957c038749SMark Nutter out_be64(&priv2->spu_chnlcntptr_RW, idx); 15967c038749SMark Nutter eieio(); 15977c038749SMark Nutter out_be64(&priv2->spu_chnlcnt_RW, ch_counts[i]); 15987c038749SMark Nutter eieio(); 15997c038749SMark Nutter } 16007c038749SMark Nutter } 16017c038749SMark Nutter 16027c038749SMark Nutter static inline void restore_spu_lslr(struct spu_state *csa, struct spu *spu) 16037c038749SMark Nutter { 16047c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 16057c038749SMark Nutter 16067c038749SMark Nutter /* Restore, Step 61: 16077c038749SMark Nutter * Restore the SPU_LSLR register from CSA. 16087c038749SMark Nutter */ 16097c038749SMark Nutter out_be64(&priv2->spu_lslr_RW, csa->priv2.spu_lslr_RW); 16107c038749SMark Nutter eieio(); 16117c038749SMark Nutter } 16127c038749SMark Nutter 16137c038749SMark Nutter static inline void restore_spu_cfg(struct spu_state *csa, struct spu *spu) 16147c038749SMark Nutter { 16157c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 16167c038749SMark Nutter 16177c038749SMark Nutter /* Restore, Step 62: 16187c038749SMark Nutter * Restore the SPU_Cfg register from CSA. 16197c038749SMark Nutter */ 16207c038749SMark Nutter out_be64(&priv2->spu_cfg_RW, csa->priv2.spu_cfg_RW); 16217c038749SMark Nutter eieio(); 16227c038749SMark Nutter } 16237c038749SMark Nutter 16247c038749SMark Nutter static inline void restore_pm_trace(struct spu_state *csa, struct spu *spu) 16257c038749SMark Nutter { 16267c038749SMark Nutter /* Restore, Step 63: 16277c038749SMark Nutter * Restore PM_Trace_Tag_Wait_Mask from CSA. 16287c038749SMark Nutter * Not performed by this implementation. 16297c038749SMark Nutter */ 16307c038749SMark Nutter } 16317c038749SMark Nutter 16327c038749SMark Nutter static inline void restore_spu_npc(struct spu_state *csa, struct spu *spu) 16337c038749SMark Nutter { 16347c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 16357c038749SMark Nutter 16367c038749SMark Nutter /* Restore, Step 64: 16377c038749SMark Nutter * Restore SPU_NPC from CSA. 16387c038749SMark Nutter */ 16397c038749SMark Nutter out_be32(&prob->spu_npc_RW, csa->prob.spu_npc_RW); 16407c038749SMark Nutter eieio(); 16417c038749SMark Nutter } 16427c038749SMark Nutter 16437c038749SMark Nutter static inline void restore_spu_mb(struct spu_state *csa, struct spu *spu) 16447c038749SMark Nutter { 16457c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 16467c038749SMark Nutter int i; 16477c038749SMark Nutter 16487c038749SMark Nutter /* Restore, Step 65: 16497c038749SMark Nutter * Restore MFC_RdSPU_MB from CSA. 16507c038749SMark Nutter */ 16517c038749SMark Nutter out_be64(&priv2->spu_chnlcntptr_RW, 29UL); 16527c038749SMark Nutter eieio(); 16537c038749SMark Nutter out_be64(&priv2->spu_chnlcnt_RW, csa->spu_chnlcnt_RW[29]); 16547c038749SMark Nutter for (i = 0; i < 4; i++) { 16558b3d6663SArnd Bergmann out_be64(&priv2->spu_chnldata_RW, csa->spu_mailbox_data[i]); 16567c038749SMark Nutter } 16577c038749SMark Nutter eieio(); 16587c038749SMark Nutter } 16597c038749SMark Nutter 16607c038749SMark Nutter static inline void check_ppu_mb_stat(struct spu_state *csa, struct spu *spu) 16617c038749SMark Nutter { 16627c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 16637c038749SMark Nutter u32 dummy = 0; 16647c038749SMark Nutter 16657c038749SMark Nutter /* Restore, Step 66: 16667c038749SMark Nutter * If CSA.MB_Stat[P]=0 (mailbox empty) then 16677c038749SMark Nutter * read from the PPU_MB register. 16687c038749SMark Nutter */ 16697c038749SMark Nutter if ((csa->prob.mb_stat_R & 0xFF) == 0) { 16707c038749SMark Nutter dummy = in_be32(&prob->pu_mb_R); 16717c038749SMark Nutter eieio(); 16727c038749SMark Nutter } 16737c038749SMark Nutter } 16747c038749SMark Nutter 16757c038749SMark Nutter static inline void check_ppuint_mb_stat(struct spu_state *csa, struct spu *spu) 16767c038749SMark Nutter { 16777c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 16787c038749SMark Nutter u64 dummy = 0UL; 16797c038749SMark Nutter 16807c038749SMark Nutter /* Restore, Step 66: 16817c038749SMark Nutter * If CSA.MB_Stat[I]=0 (mailbox empty) then 16827c038749SMark Nutter * read from the PPUINT_MB register. 16837c038749SMark Nutter */ 16847c038749SMark Nutter if ((csa->prob.mb_stat_R & 0xFF0000) == 0) { 16857c038749SMark Nutter dummy = in_be64(&priv2->puint_mb_R); 16867c038749SMark Nutter eieio(); 1687f0831accSArnd Bergmann spu_int_stat_clear(spu, 2, CLASS2_ENABLE_MAILBOX_INTR); 16887c038749SMark Nutter eieio(); 16897c038749SMark Nutter } 16907c038749SMark Nutter } 16917c038749SMark Nutter 16927c038749SMark Nutter static inline void restore_mfc_sr1(struct spu_state *csa, struct spu *spu) 16937c038749SMark Nutter { 16947c038749SMark Nutter /* Restore, Step 69: 16957c038749SMark Nutter * Restore the MFC_SR1 register from CSA. 16967c038749SMark Nutter */ 1697f0831accSArnd Bergmann spu_mfc_sr1_set(spu, csa->priv1.mfc_sr1_RW); 16987c038749SMark Nutter eieio(); 16997c038749SMark Nutter } 17007c038749SMark Nutter 17017c038749SMark Nutter static inline void restore_other_spu_access(struct spu_state *csa, 17027c038749SMark Nutter struct spu *spu) 17037c038749SMark Nutter { 17047c038749SMark Nutter /* Restore, Step 70: 17057c038749SMark Nutter * Restore other SPU mappings to this SPU. TBD. 17067c038749SMark Nutter */ 17077c038749SMark Nutter } 17087c038749SMark Nutter 17097c038749SMark Nutter static inline void restore_spu_runcntl(struct spu_state *csa, struct spu *spu) 17107c038749SMark Nutter { 17117c038749SMark Nutter struct spu_problem __iomem *prob = spu->problem; 17127c038749SMark Nutter 17137c038749SMark Nutter /* Restore, Step 71: 17147c038749SMark Nutter * If CSA.SPU_Status[R]=1 then write 17157c038749SMark Nutter * SPU_RunCntl[R0R1]='01'. 17167c038749SMark Nutter */ 17177c038749SMark Nutter if (csa->prob.spu_status_R & SPU_STATUS_RUNNING) { 17187c038749SMark Nutter out_be32(&prob->spu_runcntl_RW, SPU_RUNCNTL_RUNNABLE); 17197c038749SMark Nutter eieio(); 17207c038749SMark Nutter } 17217c038749SMark Nutter } 17227c038749SMark Nutter 17237c038749SMark Nutter static inline void restore_mfc_cntl(struct spu_state *csa, struct spu *spu) 17247c038749SMark Nutter { 17257c038749SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 17267c038749SMark Nutter 17277c038749SMark Nutter /* Restore, Step 72: 17287c038749SMark Nutter * Restore the MFC_CNTL register for the CSA. 17297c038749SMark Nutter */ 17307c038749SMark Nutter out_be64(&priv2->mfc_control_RW, csa->priv2.mfc_control_RW); 17317c038749SMark Nutter eieio(); 173228347bceSHyeonSeung Jang /* 173328347bceSHyeonSeung Jang * FIXME: this is to restart a DMA that we were processing 173428347bceSHyeonSeung Jang * before the save. better remember the fault information 173528347bceSHyeonSeung Jang * in the csa instead. 173628347bceSHyeonSeung Jang */ 173728347bceSHyeonSeung Jang if ((csa->priv2.mfc_control_RW & MFC_CNTL_SUSPEND_DMA_QUEUE_MASK)) { 173828347bceSHyeonSeung Jang out_be64(&priv2->mfc_control_RW, MFC_CNTL_RESTART_DMA_COMMAND); 173928347bceSHyeonSeung Jang eieio(); 174028347bceSHyeonSeung Jang } 17417c038749SMark Nutter } 17427c038749SMark Nutter 17437c038749SMark Nutter static inline void enable_user_access(struct spu_state *csa, struct spu *spu) 17447c038749SMark Nutter { 17457c038749SMark Nutter /* Restore, Step 73: 17467c038749SMark Nutter * Enable user-space access (if provided) to this 17477c038749SMark Nutter * SPU by mapping the virtual pages assigned to 17487c038749SMark Nutter * the SPU memory-mapped I/O (MMIO) for problem 17497c038749SMark Nutter * state. TBD. 17507c038749SMark Nutter */ 17517c038749SMark Nutter } 17527c038749SMark Nutter 17537c038749SMark Nutter static inline void reset_switch_active(struct spu_state *csa, struct spu *spu) 17547c038749SMark Nutter { 17557c038749SMark Nutter /* Restore, Step 74: 17567c038749SMark Nutter * Reset the "context switch active" flag. 175761b36fc1SAndre Detsch * Not performed by this implementation. 17587c038749SMark Nutter */ 17597c038749SMark Nutter } 17607c038749SMark Nutter 17617c038749SMark Nutter static inline void reenable_interrupts(struct spu_state *csa, struct spu *spu) 17627c038749SMark Nutter { 17637c038749SMark Nutter /* Restore, Step 75: 17647c038749SMark Nutter * Re-enable SPU interrupts. 17657c038749SMark Nutter */ 17667c038749SMark Nutter spin_lock_irq(&spu->register_lock); 1767f0831accSArnd Bergmann spu_int_mask_set(spu, 0, csa->priv1.int_mask_class0_RW); 1768f0831accSArnd Bergmann spu_int_mask_set(spu, 1, csa->priv1.int_mask_class1_RW); 1769f0831accSArnd Bergmann spu_int_mask_set(spu, 2, csa->priv1.int_mask_class2_RW); 17707c038749SMark Nutter spin_unlock_irq(&spu->register_lock); 17717c038749SMark Nutter } 17727c038749SMark Nutter 17737c038749SMark Nutter static int quiece_spu(struct spu_state *prev, struct spu *spu) 17747c038749SMark Nutter { 17757c038749SMark Nutter /* 17767c038749SMark Nutter * Combined steps 2-18 of SPU context save sequence, which 17777c038749SMark Nutter * quiesce the SPU state (disable SPU execution, MFC command 17787c038749SMark Nutter * queues, decrementer, SPU interrupts, etc.). 17797c038749SMark Nutter * 17807c038749SMark Nutter * Returns 0 on success. 17817c038749SMark Nutter * 2 if failed step 2. 17827c038749SMark Nutter * 6 if failed step 6. 17837c038749SMark Nutter */ 17847c038749SMark Nutter 17857c038749SMark Nutter if (check_spu_isolate(prev, spu)) { /* Step 2. */ 17867c038749SMark Nutter return 2; 17877c038749SMark Nutter } 17887c038749SMark Nutter disable_interrupts(prev, spu); /* Step 3. */ 17897c038749SMark Nutter set_watchdog_timer(prev, spu); /* Step 4. */ 17907c038749SMark Nutter inhibit_user_access(prev, spu); /* Step 5. */ 17917c038749SMark Nutter if (check_spu_isolate(prev, spu)) { /* Step 6. */ 17927c038749SMark Nutter return 6; 17937c038749SMark Nutter } 17947c038749SMark Nutter set_switch_pending(prev, spu); /* Step 7. */ 17957c038749SMark Nutter save_mfc_cntl(prev, spu); /* Step 8. */ 17967c038749SMark Nutter save_spu_runcntl(prev, spu); /* Step 9. */ 17977c038749SMark Nutter save_mfc_sr1(prev, spu); /* Step 10. */ 17987c038749SMark Nutter save_spu_status(prev, spu); /* Step 11. */ 179955d7cd74SJeremy Kerr save_mfc_stopped_status(prev, spu); /* Step 12. */ 18007c038749SMark Nutter halt_mfc_decr(prev, spu); /* Step 13. */ 18017c038749SMark Nutter save_timebase(prev, spu); /* Step 14. */ 18027c038749SMark Nutter remove_other_spu_access(prev, spu); /* Step 15. */ 18037c038749SMark Nutter do_mfc_mssync(prev, spu); /* Step 16. */ 18047c038749SMark Nutter issue_mfc_tlbie(prev, spu); /* Step 17. */ 18057c038749SMark Nutter handle_pending_interrupts(prev, spu); /* Step 18. */ 18067c038749SMark Nutter 18077c038749SMark Nutter return 0; 18087c038749SMark Nutter } 18097c038749SMark Nutter 18107c038749SMark Nutter static void save_csa(struct spu_state *prev, struct spu *spu) 18117c038749SMark Nutter { 18127c038749SMark Nutter /* 18137c038749SMark Nutter * Combine steps 19-44 of SPU context save sequence, which 18147c038749SMark Nutter * save regions of the privileged & problem state areas. 18157c038749SMark Nutter */ 18167c038749SMark Nutter 18177c038749SMark Nutter save_mfc_queues(prev, spu); /* Step 19. */ 18187c038749SMark Nutter save_ppu_querymask(prev, spu); /* Step 20. */ 18197c038749SMark Nutter save_ppu_querytype(prev, spu); /* Step 21. */ 18208d038e04SKazunori Asayama save_ppu_tagstatus(prev, spu); /* NEW. */ 18217c038749SMark Nutter save_mfc_csr_tsq(prev, spu); /* Step 22. */ 18227c038749SMark Nutter save_mfc_csr_cmd(prev, spu); /* Step 23. */ 18237c038749SMark Nutter save_mfc_csr_ato(prev, spu); /* Step 24. */ 18247c038749SMark Nutter save_mfc_tclass_id(prev, spu); /* Step 25. */ 18257c038749SMark Nutter set_mfc_tclass_id(prev, spu); /* Step 26. */ 182636d29edbSJeremy Kerr save_mfc_cmd(prev, spu); /* Step 26a - moved from 44. */ 18277c038749SMark Nutter purge_mfc_queue(prev, spu); /* Step 27. */ 18287c038749SMark Nutter wait_purge_complete(prev, spu); /* Step 28. */ 18297c038749SMark Nutter setup_mfc_sr1(prev, spu); /* Step 30. */ 18307c038749SMark Nutter save_spu_npc(prev, spu); /* Step 31. */ 18317c038749SMark Nutter save_spu_privcntl(prev, spu); /* Step 32. */ 18327c038749SMark Nutter reset_spu_privcntl(prev, spu); /* Step 33. */ 18337c038749SMark Nutter save_spu_lslr(prev, spu); /* Step 34. */ 18347c038749SMark Nutter reset_spu_lslr(prev, spu); /* Step 35. */ 18357c038749SMark Nutter save_spu_cfg(prev, spu); /* Step 36. */ 18367c038749SMark Nutter save_pm_trace(prev, spu); /* Step 37. */ 18377c038749SMark Nutter save_mfc_rag(prev, spu); /* Step 38. */ 18387c038749SMark Nutter save_ppu_mb_stat(prev, spu); /* Step 39. */ 18397c038749SMark Nutter save_ppu_mb(prev, spu); /* Step 40. */ 18407c038749SMark Nutter save_ppuint_mb(prev, spu); /* Step 41. */ 18417c038749SMark Nutter save_ch_part1(prev, spu); /* Step 42. */ 18427c038749SMark Nutter save_spu_mb(prev, spu); /* Step 43. */ 18437c038749SMark Nutter reset_ch(prev, spu); /* Step 45. */ 18447c038749SMark Nutter } 18457c038749SMark Nutter 18467c038749SMark Nutter static void save_lscsa(struct spu_state *prev, struct spu *spu) 18477c038749SMark Nutter { 18487c038749SMark Nutter /* 18497c038749SMark Nutter * Perform steps 46-57 of SPU context save sequence, 18507c038749SMark Nutter * which save regions of the local store and register 18517c038749SMark Nutter * file. 18527c038749SMark Nutter */ 18537c038749SMark Nutter 18547c038749SMark Nutter resume_mfc_queue(prev, spu); /* Step 46. */ 1855684bd614SJeremy Kerr /* Step 47. */ 1856684bd614SJeremy Kerr setup_mfc_slbs(prev, spu, spu_save_code, sizeof(spu_save_code)); 18577c038749SMark Nutter set_switch_active(prev, spu); /* Step 48. */ 18587c038749SMark Nutter enable_interrupts(prev, spu); /* Step 49. */ 18597c038749SMark Nutter save_ls_16kb(prev, spu); /* Step 50. */ 18607c038749SMark Nutter set_spu_npc(prev, spu); /* Step 51. */ 18617c038749SMark Nutter set_signot1(prev, spu); /* Step 52. */ 18627c038749SMark Nutter set_signot2(prev, spu); /* Step 53. */ 18637c038749SMark Nutter send_save_code(prev, spu); /* Step 54. */ 18647c038749SMark Nutter set_ppu_querymask(prev, spu); /* Step 55. */ 18657c038749SMark Nutter wait_tag_complete(prev, spu); /* Step 56. */ 18667c038749SMark Nutter wait_spu_stopped(prev, spu); /* Step 57. */ 18677c038749SMark Nutter } 18687c038749SMark Nutter 18695737edd1SMark Nutter static void force_spu_isolate_exit(struct spu *spu) 18705737edd1SMark Nutter { 18715737edd1SMark Nutter struct spu_problem __iomem *prob = spu->problem; 18725737edd1SMark Nutter struct spu_priv2 __iomem *priv2 = spu->priv2; 18735737edd1SMark Nutter 18745737edd1SMark Nutter /* Stop SPE execution and wait for completion. */ 18755737edd1SMark Nutter out_be32(&prob->spu_runcntl_RW, SPU_RUNCNTL_STOP); 18765737edd1SMark Nutter iobarrier_rw(); 18775737edd1SMark Nutter POLL_WHILE_TRUE(in_be32(&prob->spu_status_R) & SPU_STATUS_RUNNING); 18785737edd1SMark Nutter 18795737edd1SMark Nutter /* Restart SPE master runcntl. */ 18805737edd1SMark Nutter spu_mfc_sr1_set(spu, MFC_STATE1_MASTER_RUN_CONTROL_MASK); 18815737edd1SMark Nutter iobarrier_w(); 18825737edd1SMark Nutter 18835737edd1SMark Nutter /* Initiate isolate exit request and wait for completion. */ 18845737edd1SMark Nutter out_be64(&priv2->spu_privcntl_RW, 4LL); 18855737edd1SMark Nutter iobarrier_w(); 18865737edd1SMark Nutter out_be32(&prob->spu_runcntl_RW, 2); 18875737edd1SMark Nutter iobarrier_rw(); 18885737edd1SMark Nutter POLL_WHILE_FALSE((in_be32(&prob->spu_status_R) 18895737edd1SMark Nutter & SPU_STATUS_STOPPED_BY_STOP)); 18905737edd1SMark Nutter 18915737edd1SMark Nutter /* Reset load request to normal. */ 18925737edd1SMark Nutter out_be64(&priv2->spu_privcntl_RW, SPU_PRIVCNT_LOAD_REQUEST_NORMAL); 18935737edd1SMark Nutter iobarrier_w(); 18945737edd1SMark Nutter } 18955737edd1SMark Nutter 18965737edd1SMark Nutter /** 18975737edd1SMark Nutter * stop_spu_isolate 18985737edd1SMark Nutter * Check SPU run-control state and force isolated 18995737edd1SMark Nutter * exit function as necessary. 19005737edd1SMark Nutter */ 19015737edd1SMark Nutter static void stop_spu_isolate(struct spu *spu) 19025737edd1SMark Nutter { 19035737edd1SMark Nutter struct spu_problem __iomem *prob = spu->problem; 19045737edd1SMark Nutter 19055737edd1SMark Nutter if (in_be32(&prob->spu_status_R) & SPU_STATUS_ISOLATED_STATE) { 19065737edd1SMark Nutter /* The SPU is in isolated state; the only way 19075737edd1SMark Nutter * to get it out is to perform an isolated 19085737edd1SMark Nutter * exit (clean) operation. 19095737edd1SMark Nutter */ 19105737edd1SMark Nutter force_spu_isolate_exit(spu); 19115737edd1SMark Nutter } 19125737edd1SMark Nutter } 19135737edd1SMark Nutter 19147c038749SMark Nutter static void harvest(struct spu_state *prev, struct spu *spu) 19157c038749SMark Nutter { 19167c038749SMark Nutter /* 19177c038749SMark Nutter * Perform steps 2-25 of SPU context restore sequence, 19187c038749SMark Nutter * which resets an SPU either after a failed save, or 19197c038749SMark Nutter * when using SPU for first time. 19207c038749SMark Nutter */ 19217c038749SMark Nutter 19227c038749SMark Nutter disable_interrupts(prev, spu); /* Step 2. */ 19237c038749SMark Nutter inhibit_user_access(prev, spu); /* Step 3. */ 19247c038749SMark Nutter terminate_spu_app(prev, spu); /* Step 4. */ 19257c038749SMark Nutter set_switch_pending(prev, spu); /* Step 5. */ 19265737edd1SMark Nutter stop_spu_isolate(spu); /* NEW. */ 19277c038749SMark Nutter remove_other_spu_access(prev, spu); /* Step 6. */ 1928cf17df22SMasato Noguchi suspend_mfc_and_halt_decr(prev, spu); /* Step 7. */ 19297c038749SMark Nutter wait_suspend_mfc_complete(prev, spu); /* Step 8. */ 19307c038749SMark Nutter if (!suspend_spe(prev, spu)) /* Step 9. */ 19317c038749SMark Nutter clear_spu_status(prev, spu); /* Step 10. */ 19327c038749SMark Nutter do_mfc_mssync(prev, spu); /* Step 11. */ 19337c038749SMark Nutter issue_mfc_tlbie(prev, spu); /* Step 12. */ 19347c038749SMark Nutter handle_pending_interrupts(prev, spu); /* Step 13. */ 19357c038749SMark Nutter purge_mfc_queue(prev, spu); /* Step 14. */ 19367c038749SMark Nutter wait_purge_complete(prev, spu); /* Step 15. */ 19377c038749SMark Nutter reset_spu_privcntl(prev, spu); /* Step 16. */ 19387c038749SMark Nutter reset_spu_lslr(prev, spu); /* Step 17. */ 19397c038749SMark Nutter setup_mfc_sr1(prev, spu); /* Step 18. */ 194094b2a439SBenjamin Herrenschmidt spu_invalidate_slbs(spu); /* Step 19. */ 19417c038749SMark Nutter reset_ch_part1(prev, spu); /* Step 20. */ 19427c038749SMark Nutter reset_ch_part2(prev, spu); /* Step 21. */ 19437c038749SMark Nutter enable_interrupts(prev, spu); /* Step 22. */ 19447c038749SMark Nutter set_switch_active(prev, spu); /* Step 23. */ 19457c038749SMark Nutter set_mfc_tclass_id(prev, spu); /* Step 24. */ 19467c038749SMark Nutter resume_mfc_queue(prev, spu); /* Step 25. */ 19477c038749SMark Nutter } 19487c038749SMark Nutter 19497c038749SMark Nutter static void restore_lscsa(struct spu_state *next, struct spu *spu) 19507c038749SMark Nutter { 19517c038749SMark Nutter /* 19527c038749SMark Nutter * Perform steps 26-40 of SPU context restore sequence, 19537c038749SMark Nutter * which restores regions of the local store and register 19547c038749SMark Nutter * file. 19557c038749SMark Nutter */ 19567c038749SMark Nutter 19577c038749SMark Nutter set_watchdog_timer(next, spu); /* Step 26. */ 19587c038749SMark Nutter setup_spu_status_part1(next, spu); /* Step 27. */ 19597c038749SMark Nutter setup_spu_status_part2(next, spu); /* Step 28. */ 19607c038749SMark Nutter restore_mfc_rag(next, spu); /* Step 29. */ 1961684bd614SJeremy Kerr /* Step 30. */ 1962684bd614SJeremy Kerr setup_mfc_slbs(next, spu, spu_restore_code, sizeof(spu_restore_code)); 19637c038749SMark Nutter set_spu_npc(next, spu); /* Step 31. */ 19647c038749SMark Nutter set_signot1(next, spu); /* Step 32. */ 19657c038749SMark Nutter set_signot2(next, spu); /* Step 33. */ 19667c038749SMark Nutter setup_decr(next, spu); /* Step 34. */ 19677c038749SMark Nutter setup_ppu_mb(next, spu); /* Step 35. */ 19687c038749SMark Nutter setup_ppuint_mb(next, spu); /* Step 36. */ 19697c038749SMark Nutter send_restore_code(next, spu); /* Step 37. */ 19707c038749SMark Nutter set_ppu_querymask(next, spu); /* Step 38. */ 19717c038749SMark Nutter wait_tag_complete(next, spu); /* Step 39. */ 19727c038749SMark Nutter wait_spu_stopped(next, spu); /* Step 40. */ 19737c038749SMark Nutter } 19747c038749SMark Nutter 19757c038749SMark Nutter static void restore_csa(struct spu_state *next, struct spu *spu) 19767c038749SMark Nutter { 19777c038749SMark Nutter /* 19787c038749SMark Nutter * Combine steps 41-76 of SPU context restore sequence, which 19797c038749SMark Nutter * restore regions of the privileged & problem state areas. 19807c038749SMark Nutter */ 19817c038749SMark Nutter 19827c038749SMark Nutter restore_spu_privcntl(next, spu); /* Step 41. */ 19837c038749SMark Nutter restore_status_part1(next, spu); /* Step 42. */ 19847c038749SMark Nutter restore_status_part2(next, spu); /* Step 43. */ 19857c038749SMark Nutter restore_ls_16kb(next, spu); /* Step 44. */ 19867c038749SMark Nutter wait_tag_complete(next, spu); /* Step 45. */ 19877c038749SMark Nutter suspend_mfc(next, spu); /* Step 46. */ 19887c038749SMark Nutter wait_suspend_mfc_complete(next, spu); /* Step 47. */ 19897c038749SMark Nutter issue_mfc_tlbie(next, spu); /* Step 48. */ 19907c038749SMark Nutter clear_interrupts(next, spu); /* Step 49. */ 19917c038749SMark Nutter restore_mfc_queues(next, spu); /* Step 50. */ 19927c038749SMark Nutter restore_ppu_querymask(next, spu); /* Step 51. */ 19937c038749SMark Nutter restore_ppu_querytype(next, spu); /* Step 52. */ 19947c038749SMark Nutter restore_mfc_csr_tsq(next, spu); /* Step 53. */ 19957c038749SMark Nutter restore_mfc_csr_cmd(next, spu); /* Step 54. */ 19967c038749SMark Nutter restore_mfc_csr_ato(next, spu); /* Step 55. */ 19977c038749SMark Nutter restore_mfc_tclass_id(next, spu); /* Step 56. */ 19987c038749SMark Nutter set_llr_event(next, spu); /* Step 57. */ 19997c038749SMark Nutter restore_decr_wrapped(next, spu); /* Step 58. */ 20007c038749SMark Nutter restore_ch_part1(next, spu); /* Step 59. */ 20017c038749SMark Nutter restore_ch_part2(next, spu); /* Step 60. */ 20027c038749SMark Nutter restore_spu_lslr(next, spu); /* Step 61. */ 20037c038749SMark Nutter restore_spu_cfg(next, spu); /* Step 62. */ 20047c038749SMark Nutter restore_pm_trace(next, spu); /* Step 63. */ 20057c038749SMark Nutter restore_spu_npc(next, spu); /* Step 64. */ 20067c038749SMark Nutter restore_spu_mb(next, spu); /* Step 65. */ 20077c038749SMark Nutter check_ppu_mb_stat(next, spu); /* Step 66. */ 20087c038749SMark Nutter check_ppuint_mb_stat(next, spu); /* Step 67. */ 200994b2a439SBenjamin Herrenschmidt spu_invalidate_slbs(spu); /* Modified Step 68. */ 20107c038749SMark Nutter restore_mfc_sr1(next, spu); /* Step 69. */ 20117c038749SMark Nutter restore_other_spu_access(next, spu); /* Step 70. */ 20127c038749SMark Nutter restore_spu_runcntl(next, spu); /* Step 71. */ 20137c038749SMark Nutter restore_mfc_cntl(next, spu); /* Step 72. */ 20147c038749SMark Nutter enable_user_access(next, spu); /* Step 73. */ 20157c038749SMark Nutter reset_switch_active(next, spu); /* Step 74. */ 20167c038749SMark Nutter reenable_interrupts(next, spu); /* Step 75. */ 20177c038749SMark Nutter } 20187c038749SMark Nutter 20197c038749SMark Nutter static int __do_spu_save(struct spu_state *prev, struct spu *spu) 20207c038749SMark Nutter { 20217c038749SMark Nutter int rc; 20227c038749SMark Nutter 20237c038749SMark Nutter /* 20247c038749SMark Nutter * SPU context save can be broken into three phases: 20257c038749SMark Nutter * 20267c038749SMark Nutter * (a) quiesce [steps 2-16]. 20277c038749SMark Nutter * (b) save of CSA, performed by PPE [steps 17-42] 20287c038749SMark Nutter * (c) save of LSCSA, mostly performed by SPU [steps 43-52]. 20297c038749SMark Nutter * 20307c038749SMark Nutter * Returns 0 on success. 20317c038749SMark Nutter * 2,6 if failed to quiece SPU 20327c038749SMark Nutter * 53 if SPU-side of save failed. 20337c038749SMark Nutter */ 20347c038749SMark Nutter 20357c038749SMark Nutter rc = quiece_spu(prev, spu); /* Steps 2-16. */ 20367c038749SMark Nutter switch (rc) { 20377c038749SMark Nutter default: 20387c038749SMark Nutter case 2: 20397c038749SMark Nutter case 6: 20407c038749SMark Nutter harvest(prev, spu); 20417c038749SMark Nutter return rc; 20427c038749SMark Nutter break; 20437c038749SMark Nutter case 0: 20447c038749SMark Nutter break; 20457c038749SMark Nutter } 20467c038749SMark Nutter save_csa(prev, spu); /* Steps 17-43. */ 20477c038749SMark Nutter save_lscsa(prev, spu); /* Steps 44-53. */ 20487c038749SMark Nutter return check_save_status(prev, spu); /* Step 54. */ 20497c038749SMark Nutter } 20507c038749SMark Nutter 20517c038749SMark Nutter static int __do_spu_restore(struct spu_state *next, struct spu *spu) 20527c038749SMark Nutter { 20537c038749SMark Nutter int rc; 20547c038749SMark Nutter 20557c038749SMark Nutter /* 20567c038749SMark Nutter * SPU context restore can be broken into three phases: 20577c038749SMark Nutter * 20587c038749SMark Nutter * (a) harvest (or reset) SPU [steps 2-24]. 20597c038749SMark Nutter * (b) restore LSCSA [steps 25-40], mostly performed by SPU. 20607c038749SMark Nutter * (c) restore CSA [steps 41-76], performed by PPE. 20617c038749SMark Nutter * 20627c038749SMark Nutter * The 'harvest' step is not performed here, but rather 20637c038749SMark Nutter * as needed below. 20647c038749SMark Nutter */ 20657c038749SMark Nutter 20667c038749SMark Nutter restore_lscsa(next, spu); /* Steps 24-39. */ 20677c038749SMark Nutter rc = check_restore_status(next, spu); /* Step 40. */ 20687c038749SMark Nutter switch (rc) { 20697c038749SMark Nutter default: 20707c038749SMark Nutter /* Failed. Return now. */ 20717c038749SMark Nutter return rc; 20727c038749SMark Nutter break; 20737c038749SMark Nutter case 0: 20747c038749SMark Nutter /* Fall through to next step. */ 20757c038749SMark Nutter break; 20767c038749SMark Nutter } 20777c038749SMark Nutter restore_csa(next, spu); 20787c038749SMark Nutter 20797c038749SMark Nutter return 0; 20807c038749SMark Nutter } 20817c038749SMark Nutter 20825473af04SMark Nutter /** 20835473af04SMark Nutter * spu_save - SPU context save, with locking. 20845473af04SMark Nutter * @prev: pointer to SPU context save area, to be saved. 20855473af04SMark Nutter * @spu: pointer to SPU iomem structure. 20865473af04SMark Nutter * 20875473af04SMark Nutter * Acquire locks, perform the save operation then return. 20885473af04SMark Nutter */ 20895473af04SMark Nutter int spu_save(struct spu_state *prev, struct spu *spu) 20905473af04SMark Nutter { 20917c038749SMark Nutter int rc; 20925473af04SMark Nutter 20937c038749SMark Nutter acquire_spu_lock(spu); /* Step 1. */ 20947c038749SMark Nutter rc = __do_spu_save(prev, spu); /* Steps 2-53. */ 20957c038749SMark Nutter release_spu_lock(spu); 20965737edd1SMark Nutter if (rc != 0 && rc != 2 && rc != 6) { 20978b3d6663SArnd Bergmann panic("%s failed on SPU[%d], rc=%d.\n", 20988b3d6663SArnd Bergmann __func__, spu->number, rc); 20998b3d6663SArnd Bergmann } 21005737edd1SMark Nutter return 0; 21015473af04SMark Nutter } 210291edfa49Sarnd@arndb.de EXPORT_SYMBOL_GPL(spu_save); 21035473af04SMark Nutter 21045473af04SMark Nutter /** 21055473af04SMark Nutter * spu_restore - SPU context restore, with harvest and locking. 21065473af04SMark Nutter * @new: pointer to SPU context save area, to be restored. 21075473af04SMark Nutter * @spu: pointer to SPU iomem structure. 21085473af04SMark Nutter * 21095473af04SMark Nutter * Perform harvest + restore, as we may not be coming 2110d6e05edcSAndreas Mohr * from a previous successful save operation, and the 21115473af04SMark Nutter * hardware state is unknown. 21125473af04SMark Nutter */ 21135473af04SMark Nutter int spu_restore(struct spu_state *new, struct spu *spu) 21145473af04SMark Nutter { 21157c038749SMark Nutter int rc; 21165473af04SMark Nutter 21177c038749SMark Nutter acquire_spu_lock(spu); 21187c038749SMark Nutter harvest(NULL, spu); 21198b3d6663SArnd Bergmann spu->slb_replace = 0; 21207c038749SMark Nutter rc = __do_spu_restore(new, spu); 21217c038749SMark Nutter release_spu_lock(spu); 21228b3d6663SArnd Bergmann if (rc) { 21238b3d6663SArnd Bergmann panic("%s failed on SPU[%d] rc=%d.\n", 21248b3d6663SArnd Bergmann __func__, spu->number, rc); 21258b3d6663SArnd Bergmann } 21267c038749SMark Nutter return rc; 21275473af04SMark Nutter } 212891edfa49Sarnd@arndb.de EXPORT_SYMBOL_GPL(spu_restore); 21295473af04SMark Nutter 21305473af04SMark Nutter static void init_prob(struct spu_state *csa) 21315473af04SMark Nutter { 21325473af04SMark Nutter csa->spu_chnlcnt_RW[9] = 1; 21335473af04SMark Nutter csa->spu_chnlcnt_RW[21] = 16; 21345473af04SMark Nutter csa->spu_chnlcnt_RW[23] = 1; 21355473af04SMark Nutter csa->spu_chnlcnt_RW[28] = 1; 21365473af04SMark Nutter csa->spu_chnlcnt_RW[30] = 1; 21375473af04SMark Nutter csa->prob.spu_runcntl_RW = SPU_RUNCNTL_STOP; 2138970f1baaSarnd@arndb.de csa->prob.mb_stat_R = 0x000400; 21395473af04SMark Nutter } 21405473af04SMark Nutter 21415473af04SMark Nutter static void init_priv1(struct spu_state *csa) 21425473af04SMark Nutter { 21435473af04SMark Nutter /* Enable decode, relocate, tlbie response, master runcntl. */ 21445473af04SMark Nutter csa->priv1.mfc_sr1_RW = MFC_STATE1_LOCAL_STORAGE_DECODE_MASK | 21455473af04SMark Nutter MFC_STATE1_MASTER_RUN_CONTROL_MASK | 21465473af04SMark Nutter MFC_STATE1_PROBLEM_STATE_MASK | 21475473af04SMark Nutter MFC_STATE1_RELOCATE_MASK | MFC_STATE1_BUS_TLBIE_MASK; 21485473af04SMark Nutter 21495473af04SMark Nutter /* Enable OS-specific set of interrupts. */ 21505473af04SMark Nutter csa->priv1.int_mask_class0_RW = CLASS0_ENABLE_DMA_ALIGNMENT_INTR | 21515473af04SMark Nutter CLASS0_ENABLE_INVALID_DMA_COMMAND_INTR | 21525473af04SMark Nutter CLASS0_ENABLE_SPU_ERROR_INTR; 21535473af04SMark Nutter csa->priv1.int_mask_class1_RW = CLASS1_ENABLE_SEGMENT_FAULT_INTR | 21545473af04SMark Nutter CLASS1_ENABLE_STORAGE_FAULT_INTR; 21553a843d7cSArnd Bergmann csa->priv1.int_mask_class2_RW = CLASS2_ENABLE_SPU_STOP_INTR | 2156a33a7d73SArnd Bergmann CLASS2_ENABLE_SPU_HALT_INTR | 2157a33a7d73SArnd Bergmann CLASS2_ENABLE_SPU_DMA_TAG_GROUP_COMPLETE_INTR; 21585473af04SMark Nutter } 21595473af04SMark Nutter 21605473af04SMark Nutter static void init_priv2(struct spu_state *csa) 21615473af04SMark Nutter { 21625473af04SMark Nutter csa->priv2.spu_lslr_RW = LS_ADDR_MASK; 21635473af04SMark Nutter csa->priv2.mfc_control_RW = MFC_CNTL_RESUME_DMA_QUEUE | 21645473af04SMark Nutter MFC_CNTL_NORMAL_DMA_QUEUE_OPERATION | 21655473af04SMark Nutter MFC_CNTL_DMA_QUEUES_EMPTY_MASK; 21665473af04SMark Nutter } 21675473af04SMark Nutter 21685473af04SMark Nutter /** 21695473af04SMark Nutter * spu_alloc_csa - allocate and initialize an SPU context save area. 21705473af04SMark Nutter * 21715473af04SMark Nutter * Allocate and initialize the contents of an SPU context save area. 21725473af04SMark Nutter * This includes enabling address translation, interrupt masks, etc., 21735473af04SMark Nutter * as appropriate for the given OS environment. 21745473af04SMark Nutter * 21755473af04SMark Nutter * Note that storage for the 'lscsa' is allocated separately, 21765473af04SMark Nutter * as it is by far the largest of the context save regions, 21775473af04SMark Nutter * and may need to be pinned or otherwise specially aligned. 21785473af04SMark Nutter */ 2179f1fa74f4SBenjamin Herrenschmidt int spu_init_csa(struct spu_state *csa) 21805473af04SMark Nutter { 2181f1fa74f4SBenjamin Herrenschmidt int rc; 21825473af04SMark Nutter 21835473af04SMark Nutter if (!csa) 2184f1fa74f4SBenjamin Herrenschmidt return -EINVAL; 21855473af04SMark Nutter memset(csa, 0, sizeof(struct spu_state)); 21865473af04SMark Nutter 2187f1fa74f4SBenjamin Herrenschmidt rc = spu_alloc_lscsa(csa); 2188f1fa74f4SBenjamin Herrenschmidt if (rc) 2189f1fa74f4SBenjamin Herrenschmidt return rc; 21905473af04SMark Nutter 219134af946aSIngo Molnar spin_lock_init(&csa->register_lock); 21928b3d6663SArnd Bergmann 21935473af04SMark Nutter init_prob(csa); 21945473af04SMark Nutter init_priv1(csa); 21955473af04SMark Nutter init_priv2(csa); 2196f1fa74f4SBenjamin Herrenschmidt 2197f1fa74f4SBenjamin Herrenschmidt return 0; 21985473af04SMark Nutter } 21995473af04SMark Nutter 22005473af04SMark Nutter void spu_fini_csa(struct spu_state *csa) 22015473af04SMark Nutter { 2202f1fa74f4SBenjamin Herrenschmidt spu_free_lscsa(csa); 22035473af04SMark Nutter } 2204