1 /* SPDX-License-Identifier: GPL-2.0 */ 2 #ifndef _ASM_POWERPC_PLPAR_WRAPPERS_H 3 #define _ASM_POWERPC_PLPAR_WRAPPERS_H 4 5 #include <linux/string.h> 6 #include <linux/irqflags.h> 7 8 #include <asm/hvcall.h> 9 #include <asm/paca.h> 10 #include <asm/page.h> 11 12 /* Get state of physical CPU from query_cpu_stopped */ 13 int smp_query_cpu_stopped(unsigned int pcpu); 14 #define QCSS_STOPPED 0 15 #define QCSS_STOPPING 1 16 #define QCSS_NOT_STOPPED 2 17 #define QCSS_HARDWARE_ERROR -1 18 #define QCSS_HARDWARE_BUSY -2 19 20 static inline long poll_pending(void) 21 { 22 return plpar_hcall_norets(H_POLL_PENDING); 23 } 24 25 static inline u8 get_cede_latency_hint(void) 26 { 27 return get_lppaca()->cede_latency_hint; 28 } 29 30 static inline void set_cede_latency_hint(u8 latency_hint) 31 { 32 get_lppaca()->cede_latency_hint = latency_hint; 33 } 34 35 static inline long cede_processor(void) 36 { 37 return plpar_hcall_norets(H_CEDE); 38 } 39 40 static inline long extended_cede_processor(unsigned long latency_hint) 41 { 42 long rc; 43 u8 old_latency_hint = get_cede_latency_hint(); 44 45 set_cede_latency_hint(latency_hint); 46 47 rc = cede_processor(); 48 #ifdef CONFIG_TRACE_IRQFLAGS 49 /* Ensure that H_CEDE returns with IRQs on */ 50 if (WARN_ON(!(mfmsr() & MSR_EE))) 51 __hard_irq_enable(); 52 #endif 53 54 set_cede_latency_hint(old_latency_hint); 55 56 return rc; 57 } 58 59 static inline long vpa_call(unsigned long flags, unsigned long cpu, 60 unsigned long vpa) 61 { 62 flags = flags << H_VPA_FUNC_SHIFT; 63 64 return plpar_hcall_norets(H_REGISTER_VPA, flags, cpu, vpa); 65 } 66 67 static inline long unregister_vpa(unsigned long cpu) 68 { 69 return vpa_call(H_VPA_DEREG_VPA, cpu, 0); 70 } 71 72 static inline long register_vpa(unsigned long cpu, unsigned long vpa) 73 { 74 return vpa_call(H_VPA_REG_VPA, cpu, vpa); 75 } 76 77 static inline long unregister_slb_shadow(unsigned long cpu) 78 { 79 return vpa_call(H_VPA_DEREG_SLB, cpu, 0); 80 } 81 82 static inline long register_slb_shadow(unsigned long cpu, unsigned long vpa) 83 { 84 return vpa_call(H_VPA_REG_SLB, cpu, vpa); 85 } 86 87 static inline long unregister_dtl(unsigned long cpu) 88 { 89 return vpa_call(H_VPA_DEREG_DTL, cpu, 0); 90 } 91 92 static inline long register_dtl(unsigned long cpu, unsigned long vpa) 93 { 94 return vpa_call(H_VPA_REG_DTL, cpu, vpa); 95 } 96 97 extern void vpa_init(int cpu); 98 99 static inline long plpar_pte_enter(unsigned long flags, 100 unsigned long hpte_group, unsigned long hpte_v, 101 unsigned long hpte_r, unsigned long *slot) 102 { 103 long rc; 104 unsigned long retbuf[PLPAR_HCALL_BUFSIZE]; 105 106 rc = plpar_hcall(H_ENTER, retbuf, flags, hpte_group, hpte_v, hpte_r); 107 108 *slot = retbuf[0]; 109 110 return rc; 111 } 112 113 static inline long plpar_pte_remove(unsigned long flags, unsigned long ptex, 114 unsigned long avpn, unsigned long *old_pteh_ret, 115 unsigned long *old_ptel_ret) 116 { 117 long rc; 118 unsigned long retbuf[PLPAR_HCALL_BUFSIZE]; 119 120 rc = plpar_hcall(H_REMOVE, retbuf, flags, ptex, avpn); 121 122 *old_pteh_ret = retbuf[0]; 123 *old_ptel_ret = retbuf[1]; 124 125 return rc; 126 } 127 128 /* plpar_pte_remove_raw can be called in real mode. It calls plpar_hcall_raw */ 129 static inline long plpar_pte_remove_raw(unsigned long flags, unsigned long ptex, 130 unsigned long avpn, unsigned long *old_pteh_ret, 131 unsigned long *old_ptel_ret) 132 { 133 long rc; 134 unsigned long retbuf[PLPAR_HCALL_BUFSIZE]; 135 136 rc = plpar_hcall_raw(H_REMOVE, retbuf, flags, ptex, avpn); 137 138 *old_pteh_ret = retbuf[0]; 139 *old_ptel_ret = retbuf[1]; 140 141 return rc; 142 } 143 144 static inline long plpar_pte_read(unsigned long flags, unsigned long ptex, 145 unsigned long *old_pteh_ret, unsigned long *old_ptel_ret) 146 { 147 long rc; 148 unsigned long retbuf[PLPAR_HCALL_BUFSIZE]; 149 150 rc = plpar_hcall(H_READ, retbuf, flags, ptex); 151 152 *old_pteh_ret = retbuf[0]; 153 *old_ptel_ret = retbuf[1]; 154 155 return rc; 156 } 157 158 /* plpar_pte_read_raw can be called in real mode. It calls plpar_hcall_raw */ 159 static inline long plpar_pte_read_raw(unsigned long flags, unsigned long ptex, 160 unsigned long *old_pteh_ret, unsigned long *old_ptel_ret) 161 { 162 long rc; 163 unsigned long retbuf[PLPAR_HCALL_BUFSIZE]; 164 165 rc = plpar_hcall_raw(H_READ, retbuf, flags, ptex); 166 167 *old_pteh_ret = retbuf[0]; 168 *old_ptel_ret = retbuf[1]; 169 170 return rc; 171 } 172 173 /* 174 * ptes must be 8*sizeof(unsigned long) 175 */ 176 static inline long plpar_pte_read_4(unsigned long flags, unsigned long ptex, 177 unsigned long *ptes) 178 179 { 180 long rc; 181 unsigned long retbuf[PLPAR_HCALL9_BUFSIZE]; 182 183 rc = plpar_hcall9(H_READ, retbuf, flags | H_READ_4, ptex); 184 185 memcpy(ptes, retbuf, 8*sizeof(unsigned long)); 186 187 return rc; 188 } 189 190 /* 191 * plpar_pte_read_4_raw can be called in real mode. 192 * ptes must be 8*sizeof(unsigned long) 193 */ 194 static inline long plpar_pte_read_4_raw(unsigned long flags, unsigned long ptex, 195 unsigned long *ptes) 196 197 { 198 long rc; 199 unsigned long retbuf[PLPAR_HCALL9_BUFSIZE]; 200 201 rc = plpar_hcall9_raw(H_READ, retbuf, flags | H_READ_4, ptex); 202 203 memcpy(ptes, retbuf, 8*sizeof(unsigned long)); 204 205 return rc; 206 } 207 208 static inline long plpar_pte_protect(unsigned long flags, unsigned long ptex, 209 unsigned long avpn) 210 { 211 return plpar_hcall_norets(H_PROTECT, flags, ptex, avpn); 212 } 213 214 static inline long plpar_resize_hpt_prepare(unsigned long flags, 215 unsigned long shift) 216 { 217 return plpar_hcall_norets(H_RESIZE_HPT_PREPARE, flags, shift); 218 } 219 220 static inline long plpar_resize_hpt_commit(unsigned long flags, 221 unsigned long shift) 222 { 223 return plpar_hcall_norets(H_RESIZE_HPT_COMMIT, flags, shift); 224 } 225 226 static inline long plpar_tce_get(unsigned long liobn, unsigned long ioba, 227 unsigned long *tce_ret) 228 { 229 long rc; 230 unsigned long retbuf[PLPAR_HCALL_BUFSIZE]; 231 232 rc = plpar_hcall(H_GET_TCE, retbuf, liobn, ioba); 233 234 *tce_ret = retbuf[0]; 235 236 return rc; 237 } 238 239 static inline long plpar_tce_put(unsigned long liobn, unsigned long ioba, 240 unsigned long tceval) 241 { 242 return plpar_hcall_norets(H_PUT_TCE, liobn, ioba, tceval); 243 } 244 245 static inline long plpar_tce_put_indirect(unsigned long liobn, 246 unsigned long ioba, unsigned long page, unsigned long count) 247 { 248 return plpar_hcall_norets(H_PUT_TCE_INDIRECT, liobn, ioba, page, count); 249 } 250 251 static inline long plpar_tce_stuff(unsigned long liobn, unsigned long ioba, 252 unsigned long tceval, unsigned long count) 253 { 254 return plpar_hcall_norets(H_STUFF_TCE, liobn, ioba, tceval, count); 255 } 256 257 /* Set various resource mode parameters */ 258 static inline long plpar_set_mode(unsigned long mflags, unsigned long resource, 259 unsigned long value1, unsigned long value2) 260 { 261 return plpar_hcall_norets(H_SET_MODE, mflags, resource, value1, value2); 262 } 263 264 /* 265 * Enable relocation on exceptions on this partition 266 * 267 * Note: this call has a partition wide scope and can take a while to complete. 268 * If it returns H_LONG_BUSY_* it should be retried periodically until it 269 * returns H_SUCCESS. 270 */ 271 static inline long enable_reloc_on_exceptions(void) 272 { 273 /* mflags = 3: Exceptions at 0xC000000000004000 */ 274 return plpar_set_mode(3, H_SET_MODE_RESOURCE_ADDR_TRANS_MODE, 0, 0); 275 } 276 277 /* 278 * Disable relocation on exceptions on this partition 279 * 280 * Note: this call has a partition wide scope and can take a while to complete. 281 * If it returns H_LONG_BUSY_* it should be retried periodically until it 282 * returns H_SUCCESS. 283 */ 284 static inline long disable_reloc_on_exceptions(void) { 285 return plpar_set_mode(0, H_SET_MODE_RESOURCE_ADDR_TRANS_MODE, 0, 0); 286 } 287 288 /* 289 * Take exceptions in big endian mode on this partition 290 * 291 * Note: this call has a partition wide scope and can take a while to complete. 292 * If it returns H_LONG_BUSY_* it should be retried periodically until it 293 * returns H_SUCCESS. 294 */ 295 static inline long enable_big_endian_exceptions(void) 296 { 297 /* mflags = 0: big endian exceptions */ 298 return plpar_set_mode(0, H_SET_MODE_RESOURCE_LE, 0, 0); 299 } 300 301 /* 302 * Take exceptions in little endian mode on this partition 303 * 304 * Note: this call has a partition wide scope and can take a while to complete. 305 * If it returns H_LONG_BUSY_* it should be retried periodically until it 306 * returns H_SUCCESS. 307 */ 308 static inline long enable_little_endian_exceptions(void) 309 { 310 /* mflags = 1: little endian exceptions */ 311 return plpar_set_mode(1, H_SET_MODE_RESOURCE_LE, 0, 0); 312 } 313 314 static inline long plapr_set_ciabr(unsigned long ciabr) 315 { 316 return plpar_set_mode(0, H_SET_MODE_RESOURCE_SET_CIABR, ciabr, 0); 317 } 318 319 static inline long plapr_set_watchpoint0(unsigned long dawr0, unsigned long dawrx0) 320 { 321 return plpar_set_mode(0, H_SET_MODE_RESOURCE_SET_DAWR, dawr0, dawrx0); 322 } 323 324 static inline long plapr_signal_sys_reset(long cpu) 325 { 326 return plpar_hcall_norets(H_SIGNAL_SYS_RESET, cpu); 327 } 328 329 static inline long plpar_get_cpu_characteristics(struct h_cpu_char_result *p) 330 { 331 unsigned long retbuf[PLPAR_HCALL_BUFSIZE]; 332 long rc; 333 334 rc = plpar_hcall(H_GET_CPU_CHARACTERISTICS, retbuf); 335 if (rc == H_SUCCESS) { 336 p->character = retbuf[0]; 337 p->behaviour = retbuf[1]; 338 } 339 340 return rc; 341 } 342 343 #endif /* _ASM_POWERPC_PLPAR_WRAPPERS_H */ 344