1 // SPDX-License-Identifier: GPL-2.0-or-later 2 /* 3 * Copyright 2016-17 IBM Corp. 4 */ 5 6 #define pr_fmt(fmt) "vas: " fmt 7 8 #include <linux/types.h> 9 #include <linux/mutex.h> 10 #include <linux/slab.h> 11 #include <linux/io.h> 12 #include <linux/log2.h> 13 #include <linux/rcupdate.h> 14 #include <linux/cred.h> 15 #include <linux/sched/mm.h> 16 #include <linux/mmu_context.h> 17 #include <asm/switch_to.h> 18 #include <asm/ppc-opcode.h> 19 #include "vas.h" 20 #include "copy-paste.h" 21 22 #define CREATE_TRACE_POINTS 23 #include "vas-trace.h" 24 25 /* 26 * Compute the paste address region for the window @window using the 27 * ->paste_base_addr and ->paste_win_id_shift we got from device tree. 28 */ 29 static void compute_paste_address(struct vas_window *window, u64 *addr, int *len) 30 { 31 int winid; 32 u64 base, shift; 33 34 base = window->vinst->paste_base_addr; 35 shift = window->vinst->paste_win_id_shift; 36 winid = window->winid; 37 38 *addr = base + (winid << shift); 39 if (len) 40 *len = PAGE_SIZE; 41 42 pr_debug("Txwin #%d: Paste addr 0x%llx\n", winid, *addr); 43 } 44 45 static inline void get_hvwc_mmio_bar(struct vas_window *window, 46 u64 *start, int *len) 47 { 48 u64 pbaddr; 49 50 pbaddr = window->vinst->hvwc_bar_start; 51 *start = pbaddr + window->winid * VAS_HVWC_SIZE; 52 *len = VAS_HVWC_SIZE; 53 } 54 55 static inline void get_uwc_mmio_bar(struct vas_window *window, 56 u64 *start, int *len) 57 { 58 u64 pbaddr; 59 60 pbaddr = window->vinst->uwc_bar_start; 61 *start = pbaddr + window->winid * VAS_UWC_SIZE; 62 *len = VAS_UWC_SIZE; 63 } 64 65 /* 66 * Map the paste bus address of the given send window into kernel address 67 * space. Unlike MMIO regions (map_mmio_region() below), paste region must 68 * be mapped cache-able and is only applicable to send windows. 69 */ 70 static void *map_paste_region(struct vas_window *txwin) 71 { 72 int len; 73 void *map; 74 char *name; 75 u64 start; 76 77 name = kasprintf(GFP_KERNEL, "window-v%d-w%d", txwin->vinst->vas_id, 78 txwin->winid); 79 if (!name) 80 goto free_name; 81 82 txwin->paste_addr_name = name; 83 compute_paste_address(txwin, &start, &len); 84 85 if (!request_mem_region(start, len, name)) { 86 pr_devel("%s(): request_mem_region(0x%llx, %d) failed\n", 87 __func__, start, len); 88 goto free_name; 89 } 90 91 map = ioremap_cache(start, len); 92 if (!map) { 93 pr_devel("%s(): ioremap_cache(0x%llx, %d) failed\n", __func__, 94 start, len); 95 goto free_name; 96 } 97 98 pr_devel("Mapped paste addr 0x%llx to kaddr 0x%p\n", start, map); 99 return map; 100 101 free_name: 102 kfree(name); 103 return ERR_PTR(-ENOMEM); 104 } 105 106 static void *map_mmio_region(char *name, u64 start, int len) 107 { 108 void *map; 109 110 if (!request_mem_region(start, len, name)) { 111 pr_devel("%s(): request_mem_region(0x%llx, %d) failed\n", 112 __func__, start, len); 113 return NULL; 114 } 115 116 map = ioremap(start, len); 117 if (!map) { 118 pr_devel("%s(): ioremap(0x%llx, %d) failed\n", __func__, start, 119 len); 120 return NULL; 121 } 122 123 return map; 124 } 125 126 static void unmap_region(void *addr, u64 start, int len) 127 { 128 iounmap(addr); 129 release_mem_region((phys_addr_t)start, len); 130 } 131 132 /* 133 * Unmap the paste address region for a window. 134 */ 135 static void unmap_paste_region(struct vas_window *window) 136 { 137 int len; 138 u64 busaddr_start; 139 140 if (window->paste_kaddr) { 141 compute_paste_address(window, &busaddr_start, &len); 142 unmap_region(window->paste_kaddr, busaddr_start, len); 143 window->paste_kaddr = NULL; 144 kfree(window->paste_addr_name); 145 window->paste_addr_name = NULL; 146 } 147 } 148 149 /* 150 * Unmap the MMIO regions for a window. Hold the vas_mutex so we don't 151 * unmap when the window's debugfs dir is in use. This serializes close 152 * of a window even on another VAS instance but since its not a critical 153 * path, just minimize the time we hold the mutex for now. We can add 154 * a per-instance mutex later if necessary. 155 */ 156 static void unmap_winctx_mmio_bars(struct vas_window *window) 157 { 158 int len; 159 void *uwc_map; 160 void *hvwc_map; 161 u64 busaddr_start; 162 163 mutex_lock(&vas_mutex); 164 165 hvwc_map = window->hvwc_map; 166 window->hvwc_map = NULL; 167 168 uwc_map = window->uwc_map; 169 window->uwc_map = NULL; 170 171 mutex_unlock(&vas_mutex); 172 173 if (hvwc_map) { 174 get_hvwc_mmio_bar(window, &busaddr_start, &len); 175 unmap_region(hvwc_map, busaddr_start, len); 176 } 177 178 if (uwc_map) { 179 get_uwc_mmio_bar(window, &busaddr_start, &len); 180 unmap_region(uwc_map, busaddr_start, len); 181 } 182 } 183 184 /* 185 * Find the Hypervisor Window Context (HVWC) MMIO Base Address Region and the 186 * OS/User Window Context (UWC) MMIO Base Address Region for the given window. 187 * Map these bus addresses and save the mapped kernel addresses in @window. 188 */ 189 int map_winctx_mmio_bars(struct vas_window *window) 190 { 191 int len; 192 u64 start; 193 194 get_hvwc_mmio_bar(window, &start, &len); 195 window->hvwc_map = map_mmio_region("HVWCM_Window", start, len); 196 197 get_uwc_mmio_bar(window, &start, &len); 198 window->uwc_map = map_mmio_region("UWCM_Window", start, len); 199 200 if (!window->hvwc_map || !window->uwc_map) { 201 unmap_winctx_mmio_bars(window); 202 return -1; 203 } 204 205 return 0; 206 } 207 208 /* 209 * Reset all valid registers in the HV and OS/User Window Contexts for 210 * the window identified by @window. 211 * 212 * NOTE: We cannot really use a for loop to reset window context. Not all 213 * offsets in a window context are valid registers and the valid 214 * registers are not sequential. And, we can only write to offsets 215 * with valid registers. 216 */ 217 void reset_window_regs(struct vas_window *window) 218 { 219 write_hvwc_reg(window, VREG(LPID), 0ULL); 220 write_hvwc_reg(window, VREG(PID), 0ULL); 221 write_hvwc_reg(window, VREG(XLATE_MSR), 0ULL); 222 write_hvwc_reg(window, VREG(XLATE_LPCR), 0ULL); 223 write_hvwc_reg(window, VREG(XLATE_CTL), 0ULL); 224 write_hvwc_reg(window, VREG(AMR), 0ULL); 225 write_hvwc_reg(window, VREG(SEIDR), 0ULL); 226 write_hvwc_reg(window, VREG(FAULT_TX_WIN), 0ULL); 227 write_hvwc_reg(window, VREG(OSU_INTR_SRC_RA), 0ULL); 228 write_hvwc_reg(window, VREG(HV_INTR_SRC_RA), 0ULL); 229 write_hvwc_reg(window, VREG(PSWID), 0ULL); 230 write_hvwc_reg(window, VREG(LFIFO_BAR), 0ULL); 231 write_hvwc_reg(window, VREG(LDATA_STAMP_CTL), 0ULL); 232 write_hvwc_reg(window, VREG(LDMA_CACHE_CTL), 0ULL); 233 write_hvwc_reg(window, VREG(LRFIFO_PUSH), 0ULL); 234 write_hvwc_reg(window, VREG(CURR_MSG_COUNT), 0ULL); 235 write_hvwc_reg(window, VREG(LNOTIFY_AFTER_COUNT), 0ULL); 236 write_hvwc_reg(window, VREG(LRX_WCRED), 0ULL); 237 write_hvwc_reg(window, VREG(LRX_WCRED_ADDER), 0ULL); 238 write_hvwc_reg(window, VREG(TX_WCRED), 0ULL); 239 write_hvwc_reg(window, VREG(TX_WCRED_ADDER), 0ULL); 240 write_hvwc_reg(window, VREG(LFIFO_SIZE), 0ULL); 241 write_hvwc_reg(window, VREG(WINCTL), 0ULL); 242 write_hvwc_reg(window, VREG(WIN_STATUS), 0ULL); 243 write_hvwc_reg(window, VREG(WIN_CTX_CACHING_CTL), 0ULL); 244 write_hvwc_reg(window, VREG(TX_RSVD_BUF_COUNT), 0ULL); 245 write_hvwc_reg(window, VREG(LRFIFO_WIN_PTR), 0ULL); 246 write_hvwc_reg(window, VREG(LNOTIFY_CTL), 0ULL); 247 write_hvwc_reg(window, VREG(LNOTIFY_PID), 0ULL); 248 write_hvwc_reg(window, VREG(LNOTIFY_LPID), 0ULL); 249 write_hvwc_reg(window, VREG(LNOTIFY_TID), 0ULL); 250 write_hvwc_reg(window, VREG(LNOTIFY_SCOPE), 0ULL); 251 write_hvwc_reg(window, VREG(NX_UTIL_ADDER), 0ULL); 252 253 /* Skip read-only registers: NX_UTIL and NX_UTIL_SE */ 254 255 /* 256 * The send and receive window credit adder registers are also 257 * accessible from HVWC and have been initialized above. We don't 258 * need to initialize from the OS/User Window Context, so skip 259 * following calls: 260 * 261 * write_uwc_reg(window, VREG(TX_WCRED_ADDER), 0ULL); 262 * write_uwc_reg(window, VREG(LRX_WCRED_ADDER), 0ULL); 263 */ 264 } 265 266 /* 267 * Initialize window context registers related to Address Translation. 268 * These registers are common to send/receive windows although they 269 * differ for user/kernel windows. As we resolve the TODOs we may 270 * want to add fields to vas_winctx and move the initialization to 271 * init_vas_winctx_regs(). 272 */ 273 static void init_xlate_regs(struct vas_window *window, bool user_win) 274 { 275 u64 lpcr, val; 276 277 /* 278 * MSR_TA, MSR_US are false for both kernel and user. 279 * MSR_DR and MSR_PR are false for kernel. 280 */ 281 val = 0ULL; 282 val = SET_FIELD(VAS_XLATE_MSR_HV, val, 1); 283 val = SET_FIELD(VAS_XLATE_MSR_SF, val, 1); 284 if (user_win) { 285 val = SET_FIELD(VAS_XLATE_MSR_DR, val, 1); 286 val = SET_FIELD(VAS_XLATE_MSR_PR, val, 1); 287 } 288 write_hvwc_reg(window, VREG(XLATE_MSR), val); 289 290 lpcr = mfspr(SPRN_LPCR); 291 val = 0ULL; 292 /* 293 * NOTE: From Section 5.7.8.1 Segment Lookaside Buffer of the 294 * Power ISA, v3.0B, Page size encoding is 0 = 4KB, 5 = 64KB. 295 * 296 * NOTE: From Section 1.3.1, Address Translation Context of the 297 * Nest MMU Workbook, LPCR_SC should be 0 for Power9. 298 */ 299 val = SET_FIELD(VAS_XLATE_LPCR_PAGE_SIZE, val, 5); 300 val = SET_FIELD(VAS_XLATE_LPCR_ISL, val, lpcr & LPCR_ISL); 301 val = SET_FIELD(VAS_XLATE_LPCR_TC, val, lpcr & LPCR_TC); 302 val = SET_FIELD(VAS_XLATE_LPCR_SC, val, 0); 303 write_hvwc_reg(window, VREG(XLATE_LPCR), val); 304 305 /* 306 * Section 1.3.1 (Address translation Context) of NMMU workbook. 307 * 0b00 Hashed Page Table mode 308 * 0b01 Reserved 309 * 0b10 Radix on HPT 310 * 0b11 Radix on Radix 311 */ 312 val = 0ULL; 313 val = SET_FIELD(VAS_XLATE_MODE, val, radix_enabled() ? 3 : 2); 314 write_hvwc_reg(window, VREG(XLATE_CTL), val); 315 316 /* 317 * TODO: Can we mfspr(AMR) even for user windows? 318 */ 319 val = 0ULL; 320 val = SET_FIELD(VAS_AMR, val, mfspr(SPRN_AMR)); 321 write_hvwc_reg(window, VREG(AMR), val); 322 323 val = 0ULL; 324 val = SET_FIELD(VAS_SEIDR, val, 0); 325 write_hvwc_reg(window, VREG(SEIDR), val); 326 } 327 328 /* 329 * Initialize Reserved Send Buffer Count for the send window. It involves 330 * writing to the register, reading it back to confirm that the hardware 331 * has enough buffers to reserve. See section 1.3.1.2.1 of VAS workbook. 332 * 333 * Since we can only make a best-effort attempt to fulfill the request, 334 * we don't return any errors if we cannot. 335 * 336 * TODO: Reserved (aka dedicated) send buffers are not supported yet. 337 */ 338 static void init_rsvd_tx_buf_count(struct vas_window *txwin, 339 struct vas_winctx *winctx) 340 { 341 write_hvwc_reg(txwin, VREG(TX_RSVD_BUF_COUNT), 0ULL); 342 } 343 344 /* 345 * init_winctx_regs() 346 * Initialize window context registers for a receive window. 347 * Except for caching control and marking window open, the registers 348 * are initialized in the order listed in Section 3.1.4 (Window Context 349 * Cache Register Details) of the VAS workbook although they don't need 350 * to be. 351 * 352 * Design note: For NX receive windows, NX allocates the FIFO buffer in OPAL 353 * (so that it can get a large contiguous area) and passes that buffer 354 * to kernel via device tree. We now write that buffer address to the 355 * FIFO BAR. Would it make sense to do this all in OPAL? i.e have OPAL 356 * write the per-chip RX FIFO addresses to the windows during boot-up 357 * as a one-time task? That could work for NX but what about other 358 * receivers? Let the receivers tell us the rx-fifo buffers for now. 359 */ 360 int init_winctx_regs(struct vas_window *window, struct vas_winctx *winctx) 361 { 362 u64 val; 363 int fifo_size; 364 365 reset_window_regs(window); 366 367 val = 0ULL; 368 val = SET_FIELD(VAS_LPID, val, winctx->lpid); 369 write_hvwc_reg(window, VREG(LPID), val); 370 371 val = 0ULL; 372 val = SET_FIELD(VAS_PID_ID, val, winctx->pidr); 373 write_hvwc_reg(window, VREG(PID), val); 374 375 init_xlate_regs(window, winctx->user_win); 376 377 val = 0ULL; 378 val = SET_FIELD(VAS_FAULT_TX_WIN, val, winctx->fault_win_id); 379 write_hvwc_reg(window, VREG(FAULT_TX_WIN), val); 380 381 /* In PowerNV, interrupts go to HV. */ 382 write_hvwc_reg(window, VREG(OSU_INTR_SRC_RA), 0ULL); 383 384 val = 0ULL; 385 val = SET_FIELD(VAS_HV_INTR_SRC_RA, val, winctx->irq_port); 386 write_hvwc_reg(window, VREG(HV_INTR_SRC_RA), val); 387 388 val = 0ULL; 389 val = SET_FIELD(VAS_PSWID_EA_HANDLE, val, winctx->pswid); 390 write_hvwc_reg(window, VREG(PSWID), val); 391 392 write_hvwc_reg(window, VREG(SPARE1), 0ULL); 393 write_hvwc_reg(window, VREG(SPARE2), 0ULL); 394 write_hvwc_reg(window, VREG(SPARE3), 0ULL); 395 396 /* 397 * NOTE: VAS expects the FIFO address to be copied into the LFIFO_BAR 398 * register as is - do NOT shift the address into VAS_LFIFO_BAR 399 * bit fields! Ok to set the page migration select fields - 400 * VAS ignores the lower 10+ bits in the address anyway, because 401 * the minimum FIFO size is 1K? 402 * 403 * See also: Design note in function header. 404 */ 405 val = __pa(winctx->rx_fifo); 406 val = SET_FIELD(VAS_PAGE_MIGRATION_SELECT, val, 0); 407 write_hvwc_reg(window, VREG(LFIFO_BAR), val); 408 409 val = 0ULL; 410 val = SET_FIELD(VAS_LDATA_STAMP, val, winctx->data_stamp); 411 write_hvwc_reg(window, VREG(LDATA_STAMP_CTL), val); 412 413 val = 0ULL; 414 val = SET_FIELD(VAS_LDMA_TYPE, val, winctx->dma_type); 415 val = SET_FIELD(VAS_LDMA_FIFO_DISABLE, val, winctx->fifo_disable); 416 write_hvwc_reg(window, VREG(LDMA_CACHE_CTL), val); 417 418 write_hvwc_reg(window, VREG(LRFIFO_PUSH), 0ULL); 419 write_hvwc_reg(window, VREG(CURR_MSG_COUNT), 0ULL); 420 write_hvwc_reg(window, VREG(LNOTIFY_AFTER_COUNT), 0ULL); 421 422 val = 0ULL; 423 val = SET_FIELD(VAS_LRX_WCRED, val, winctx->wcreds_max); 424 write_hvwc_reg(window, VREG(LRX_WCRED), val); 425 426 val = 0ULL; 427 val = SET_FIELD(VAS_TX_WCRED, val, winctx->wcreds_max); 428 write_hvwc_reg(window, VREG(TX_WCRED), val); 429 430 write_hvwc_reg(window, VREG(LRX_WCRED_ADDER), 0ULL); 431 write_hvwc_reg(window, VREG(TX_WCRED_ADDER), 0ULL); 432 433 fifo_size = winctx->rx_fifo_size / 1024; 434 435 val = 0ULL; 436 val = SET_FIELD(VAS_LFIFO_SIZE, val, ilog2(fifo_size)); 437 write_hvwc_reg(window, VREG(LFIFO_SIZE), val); 438 439 /* Update window control and caching control registers last so 440 * we mark the window open only after fully initializing it and 441 * pushing context to cache. 442 */ 443 444 write_hvwc_reg(window, VREG(WIN_STATUS), 0ULL); 445 446 init_rsvd_tx_buf_count(window, winctx); 447 448 /* for a send window, point to the matching receive window */ 449 val = 0ULL; 450 val = SET_FIELD(VAS_LRX_WIN_ID, val, winctx->rx_win_id); 451 write_hvwc_reg(window, VREG(LRFIFO_WIN_PTR), val); 452 453 write_hvwc_reg(window, VREG(SPARE4), 0ULL); 454 455 val = 0ULL; 456 val = SET_FIELD(VAS_NOTIFY_DISABLE, val, winctx->notify_disable); 457 val = SET_FIELD(VAS_INTR_DISABLE, val, winctx->intr_disable); 458 val = SET_FIELD(VAS_NOTIFY_EARLY, val, winctx->notify_early); 459 val = SET_FIELD(VAS_NOTIFY_OSU_INTR, val, winctx->notify_os_intr_reg); 460 write_hvwc_reg(window, VREG(LNOTIFY_CTL), val); 461 462 val = 0ULL; 463 val = SET_FIELD(VAS_LNOTIFY_PID, val, winctx->lnotify_pid); 464 write_hvwc_reg(window, VREG(LNOTIFY_PID), val); 465 466 val = 0ULL; 467 val = SET_FIELD(VAS_LNOTIFY_LPID, val, winctx->lnotify_lpid); 468 write_hvwc_reg(window, VREG(LNOTIFY_LPID), val); 469 470 val = 0ULL; 471 val = SET_FIELD(VAS_LNOTIFY_TID, val, winctx->lnotify_tid); 472 write_hvwc_reg(window, VREG(LNOTIFY_TID), val); 473 474 val = 0ULL; 475 val = SET_FIELD(VAS_LNOTIFY_MIN_SCOPE, val, winctx->min_scope); 476 val = SET_FIELD(VAS_LNOTIFY_MAX_SCOPE, val, winctx->max_scope); 477 write_hvwc_reg(window, VREG(LNOTIFY_SCOPE), val); 478 479 /* Skip read-only registers NX_UTIL and NX_UTIL_SE */ 480 481 write_hvwc_reg(window, VREG(SPARE5), 0ULL); 482 write_hvwc_reg(window, VREG(NX_UTIL_ADDER), 0ULL); 483 write_hvwc_reg(window, VREG(SPARE6), 0ULL); 484 485 /* Finally, push window context to memory and... */ 486 val = 0ULL; 487 val = SET_FIELD(VAS_PUSH_TO_MEM, val, 1); 488 write_hvwc_reg(window, VREG(WIN_CTX_CACHING_CTL), val); 489 490 /* ... mark the window open for business */ 491 val = 0ULL; 492 val = SET_FIELD(VAS_WINCTL_REJ_NO_CREDIT, val, winctx->rej_no_credit); 493 val = SET_FIELD(VAS_WINCTL_PIN, val, winctx->pin_win); 494 val = SET_FIELD(VAS_WINCTL_TX_WCRED_MODE, val, winctx->tx_wcred_mode); 495 val = SET_FIELD(VAS_WINCTL_RX_WCRED_MODE, val, winctx->rx_wcred_mode); 496 val = SET_FIELD(VAS_WINCTL_TX_WORD_MODE, val, winctx->tx_word_mode); 497 val = SET_FIELD(VAS_WINCTL_RX_WORD_MODE, val, winctx->rx_word_mode); 498 val = SET_FIELD(VAS_WINCTL_FAULT_WIN, val, winctx->fault_win); 499 val = SET_FIELD(VAS_WINCTL_NX_WIN, val, winctx->nx_win); 500 val = SET_FIELD(VAS_WINCTL_OPEN, val, 1); 501 write_hvwc_reg(window, VREG(WINCTL), val); 502 503 return 0; 504 } 505 506 static void vas_release_window_id(struct ida *ida, int winid) 507 { 508 ida_free(ida, winid); 509 } 510 511 static int vas_assign_window_id(struct ida *ida) 512 { 513 int winid = ida_alloc_max(ida, VAS_WINDOWS_PER_CHIP - 1, GFP_KERNEL); 514 515 if (winid == -ENOSPC) { 516 pr_err("Too many (%d) open windows\n", VAS_WINDOWS_PER_CHIP); 517 return -EAGAIN; 518 } 519 520 return winid; 521 } 522 523 static void vas_window_free(struct vas_window *window) 524 { 525 int winid = window->winid; 526 struct vas_instance *vinst = window->vinst; 527 528 unmap_winctx_mmio_bars(window); 529 530 vas_window_free_dbgdir(window); 531 532 kfree(window); 533 534 vas_release_window_id(&vinst->ida, winid); 535 } 536 537 static struct vas_window *vas_window_alloc(struct vas_instance *vinst) 538 { 539 int winid; 540 struct vas_window *window; 541 542 winid = vas_assign_window_id(&vinst->ida); 543 if (winid < 0) 544 return ERR_PTR(winid); 545 546 window = kzalloc(sizeof(*window), GFP_KERNEL); 547 if (!window) 548 goto out_free; 549 550 window->vinst = vinst; 551 window->winid = winid; 552 553 if (map_winctx_mmio_bars(window)) 554 goto out_free; 555 556 vas_window_init_dbgdir(window); 557 558 return window; 559 560 out_free: 561 kfree(window); 562 vas_release_window_id(&vinst->ida, winid); 563 return ERR_PTR(-ENOMEM); 564 } 565 566 static void put_rx_win(struct vas_window *rxwin) 567 { 568 /* Better not be a send window! */ 569 WARN_ON_ONCE(rxwin->tx_win); 570 571 atomic_dec(&rxwin->num_txwins); 572 } 573 574 /* 575 * Find the user space receive window given the @pswid. 576 * - We must have a valid vasid and it must belong to this instance. 577 * (so both send and receive windows are on the same VAS instance) 578 * - The window must refer to an OPEN, FTW, RECEIVE window. 579 * 580 * NOTE: We access ->windows[] table and assume that vinst->mutex is held. 581 */ 582 static struct vas_window *get_user_rxwin(struct vas_instance *vinst, u32 pswid) 583 { 584 int vasid, winid; 585 struct vas_window *rxwin; 586 587 decode_pswid(pswid, &vasid, &winid); 588 589 if (vinst->vas_id != vasid) 590 return ERR_PTR(-EINVAL); 591 592 rxwin = vinst->windows[winid]; 593 594 if (!rxwin || rxwin->tx_win || rxwin->cop != VAS_COP_TYPE_FTW) 595 return ERR_PTR(-EINVAL); 596 597 return rxwin; 598 } 599 600 /* 601 * Get the VAS receive window associated with NX engine identified 602 * by @cop and if applicable, @pswid. 603 * 604 * See also function header of set_vinst_win(). 605 */ 606 static struct vas_window *get_vinst_rxwin(struct vas_instance *vinst, 607 enum vas_cop_type cop, u32 pswid) 608 { 609 struct vas_window *rxwin; 610 611 mutex_lock(&vinst->mutex); 612 613 if (cop == VAS_COP_TYPE_FTW) 614 rxwin = get_user_rxwin(vinst, pswid); 615 else 616 rxwin = vinst->rxwin[cop] ?: ERR_PTR(-EINVAL); 617 618 if (!IS_ERR(rxwin)) 619 atomic_inc(&rxwin->num_txwins); 620 621 mutex_unlock(&vinst->mutex); 622 623 return rxwin; 624 } 625 626 /* 627 * We have two tables of windows in a VAS instance. The first one, 628 * ->windows[], contains all the windows in the instance and allows 629 * looking up a window by its id. It is used to look up send windows 630 * during fault handling and receive windows when pairing user space 631 * send/receive windows. 632 * 633 * The second table, ->rxwin[], contains receive windows that are 634 * associated with NX engines. This table has VAS_COP_TYPE_MAX 635 * entries and is used to look up a receive window by its 636 * coprocessor type. 637 * 638 * Here, we save @window in the ->windows[] table. If it is a receive 639 * window, we also save the window in the ->rxwin[] table. 640 */ 641 static void set_vinst_win(struct vas_instance *vinst, 642 struct vas_window *window) 643 { 644 int id = window->winid; 645 646 mutex_lock(&vinst->mutex); 647 648 /* 649 * There should only be one receive window for a coprocessor type 650 * unless its a user (FTW) window. 651 */ 652 if (!window->user_win && !window->tx_win) { 653 WARN_ON_ONCE(vinst->rxwin[window->cop]); 654 vinst->rxwin[window->cop] = window; 655 } 656 657 WARN_ON_ONCE(vinst->windows[id] != NULL); 658 vinst->windows[id] = window; 659 660 mutex_unlock(&vinst->mutex); 661 } 662 663 /* 664 * Clear this window from the table(s) of windows for this VAS instance. 665 * See also function header of set_vinst_win(). 666 */ 667 static void clear_vinst_win(struct vas_window *window) 668 { 669 int id = window->winid; 670 struct vas_instance *vinst = window->vinst; 671 672 mutex_lock(&vinst->mutex); 673 674 if (!window->user_win && !window->tx_win) { 675 WARN_ON_ONCE(!vinst->rxwin[window->cop]); 676 vinst->rxwin[window->cop] = NULL; 677 } 678 679 WARN_ON_ONCE(vinst->windows[id] != window); 680 vinst->windows[id] = NULL; 681 682 mutex_unlock(&vinst->mutex); 683 } 684 685 static void init_winctx_for_rxwin(struct vas_window *rxwin, 686 struct vas_rx_win_attr *rxattr, 687 struct vas_winctx *winctx) 688 { 689 /* 690 * We first zero (memset()) all fields and only set non-zero fields. 691 * Following fields are 0/false but maybe deserve a comment: 692 * 693 * ->notify_os_intr_reg In powerNV, send intrs to HV 694 * ->notify_disable False for NX windows 695 * ->intr_disable False for Fault Windows 696 * ->xtra_write False for NX windows 697 * ->notify_early NA for NX windows 698 * ->rsvd_txbuf_count NA for Rx windows 699 * ->lpid, ->pid, ->tid NA for Rx windows 700 */ 701 702 memset(winctx, 0, sizeof(struct vas_winctx)); 703 704 winctx->rx_fifo = rxattr->rx_fifo; 705 winctx->rx_fifo_size = rxattr->rx_fifo_size; 706 winctx->wcreds_max = rxwin->wcreds_max; 707 winctx->pin_win = rxattr->pin_win; 708 709 winctx->nx_win = rxattr->nx_win; 710 winctx->fault_win = rxattr->fault_win; 711 winctx->user_win = rxattr->user_win; 712 winctx->rej_no_credit = rxattr->rej_no_credit; 713 winctx->rx_word_mode = rxattr->rx_win_ord_mode; 714 winctx->tx_word_mode = rxattr->tx_win_ord_mode; 715 winctx->rx_wcred_mode = rxattr->rx_wcred_mode; 716 winctx->tx_wcred_mode = rxattr->tx_wcred_mode; 717 winctx->notify_early = rxattr->notify_early; 718 719 if (winctx->nx_win) { 720 winctx->data_stamp = true; 721 winctx->intr_disable = true; 722 winctx->pin_win = true; 723 724 WARN_ON_ONCE(winctx->fault_win); 725 WARN_ON_ONCE(!winctx->rx_word_mode); 726 WARN_ON_ONCE(!winctx->tx_word_mode); 727 WARN_ON_ONCE(winctx->notify_after_count); 728 } else if (winctx->fault_win) { 729 winctx->notify_disable = true; 730 } else if (winctx->user_win) { 731 /* 732 * Section 1.8.1 Low Latency Core-Core Wake up of 733 * the VAS workbook: 734 * 735 * - disable credit checks ([tr]x_wcred_mode = false) 736 * - disable FIFO writes 737 * - enable ASB_Notify, disable interrupt 738 */ 739 winctx->fifo_disable = true; 740 winctx->intr_disable = true; 741 winctx->rx_fifo = NULL; 742 } 743 744 winctx->lnotify_lpid = rxattr->lnotify_lpid; 745 winctx->lnotify_pid = rxattr->lnotify_pid; 746 winctx->lnotify_tid = rxattr->lnotify_tid; 747 winctx->pswid = rxattr->pswid; 748 winctx->dma_type = VAS_DMA_TYPE_INJECT; 749 winctx->tc_mode = rxattr->tc_mode; 750 751 winctx->min_scope = VAS_SCOPE_LOCAL; 752 winctx->max_scope = VAS_SCOPE_VECTORED_GROUP; 753 if (rxwin->vinst->virq) 754 winctx->irq_port = rxwin->vinst->irq_port; 755 } 756 757 static bool rx_win_args_valid(enum vas_cop_type cop, 758 struct vas_rx_win_attr *attr) 759 { 760 pr_debug("Rxattr: fault %d, notify %d, intr %d, early %d, fifo %d\n", 761 attr->fault_win, attr->notify_disable, 762 attr->intr_disable, attr->notify_early, 763 attr->rx_fifo_size); 764 765 if (cop >= VAS_COP_TYPE_MAX) 766 return false; 767 768 if (cop != VAS_COP_TYPE_FTW && 769 attr->rx_fifo_size < VAS_RX_FIFO_SIZE_MIN) 770 return false; 771 772 if (attr->rx_fifo_size > VAS_RX_FIFO_SIZE_MAX) 773 return false; 774 775 if (!attr->wcreds_max) 776 return false; 777 778 if (attr->nx_win) { 779 /* cannot be fault or user window if it is nx */ 780 if (attr->fault_win || attr->user_win) 781 return false; 782 /* 783 * Section 3.1.4.32: NX Windows must not disable notification, 784 * and must not enable interrupts or early notification. 785 */ 786 if (attr->notify_disable || !attr->intr_disable || 787 attr->notify_early) 788 return false; 789 } else if (attr->fault_win) { 790 /* cannot be both fault and user window */ 791 if (attr->user_win) 792 return false; 793 794 /* 795 * Section 3.1.4.32: Fault windows must disable notification 796 * but not interrupts. 797 */ 798 if (!attr->notify_disable || attr->intr_disable) 799 return false; 800 801 } else if (attr->user_win) { 802 /* 803 * User receive windows are only for fast-thread-wakeup 804 * (FTW). They don't need a FIFO and must disable interrupts 805 */ 806 if (attr->rx_fifo || attr->rx_fifo_size || !attr->intr_disable) 807 return false; 808 } else { 809 /* Rx window must be one of NX or Fault or User window. */ 810 return false; 811 } 812 813 return true; 814 } 815 816 void vas_init_rx_win_attr(struct vas_rx_win_attr *rxattr, enum vas_cop_type cop) 817 { 818 memset(rxattr, 0, sizeof(*rxattr)); 819 820 if (cop == VAS_COP_TYPE_842 || cop == VAS_COP_TYPE_842_HIPRI) { 821 rxattr->pin_win = true; 822 rxattr->nx_win = true; 823 rxattr->fault_win = false; 824 rxattr->intr_disable = true; 825 rxattr->rx_wcred_mode = true; 826 rxattr->tx_wcred_mode = true; 827 rxattr->rx_win_ord_mode = true; 828 rxattr->tx_win_ord_mode = true; 829 } else if (cop == VAS_COP_TYPE_FAULT) { 830 rxattr->pin_win = true; 831 rxattr->fault_win = true; 832 rxattr->notify_disable = true; 833 rxattr->rx_wcred_mode = true; 834 rxattr->rx_win_ord_mode = true; 835 rxattr->rej_no_credit = true; 836 rxattr->tc_mode = VAS_THRESH_DISABLED; 837 } else if (cop == VAS_COP_TYPE_FTW) { 838 rxattr->user_win = true; 839 rxattr->intr_disable = true; 840 841 /* 842 * As noted in the VAS Workbook we disable credit checks. 843 * If we enable credit checks in the future, we must also 844 * implement a mechanism to return the user credits or new 845 * paste operations will fail. 846 */ 847 } 848 } 849 EXPORT_SYMBOL_GPL(vas_init_rx_win_attr); 850 851 struct vas_window *vas_rx_win_open(int vasid, enum vas_cop_type cop, 852 struct vas_rx_win_attr *rxattr) 853 { 854 struct vas_window *rxwin; 855 struct vas_winctx winctx; 856 struct vas_instance *vinst; 857 858 trace_vas_rx_win_open(current, vasid, cop, rxattr); 859 860 if (!rx_win_args_valid(cop, rxattr)) 861 return ERR_PTR(-EINVAL); 862 863 vinst = find_vas_instance(vasid); 864 if (!vinst) { 865 pr_devel("vasid %d not found!\n", vasid); 866 return ERR_PTR(-EINVAL); 867 } 868 pr_devel("Found instance %d\n", vasid); 869 870 rxwin = vas_window_alloc(vinst); 871 if (IS_ERR(rxwin)) { 872 pr_devel("Unable to allocate memory for Rx window\n"); 873 return rxwin; 874 } 875 876 rxwin->tx_win = false; 877 rxwin->nx_win = rxattr->nx_win; 878 rxwin->user_win = rxattr->user_win; 879 rxwin->cop = cop; 880 rxwin->wcreds_max = rxattr->wcreds_max; 881 882 init_winctx_for_rxwin(rxwin, rxattr, &winctx); 883 init_winctx_regs(rxwin, &winctx); 884 885 set_vinst_win(vinst, rxwin); 886 887 return rxwin; 888 } 889 EXPORT_SYMBOL_GPL(vas_rx_win_open); 890 891 void vas_init_tx_win_attr(struct vas_tx_win_attr *txattr, enum vas_cop_type cop) 892 { 893 memset(txattr, 0, sizeof(*txattr)); 894 895 if (cop == VAS_COP_TYPE_842 || cop == VAS_COP_TYPE_842_HIPRI) { 896 txattr->rej_no_credit = false; 897 txattr->rx_wcred_mode = true; 898 txattr->tx_wcred_mode = true; 899 txattr->rx_win_ord_mode = true; 900 txattr->tx_win_ord_mode = true; 901 } else if (cop == VAS_COP_TYPE_FTW) { 902 txattr->user_win = true; 903 } 904 } 905 EXPORT_SYMBOL_GPL(vas_init_tx_win_attr); 906 907 static void init_winctx_for_txwin(struct vas_window *txwin, 908 struct vas_tx_win_attr *txattr, 909 struct vas_winctx *winctx) 910 { 911 /* 912 * We first zero all fields and only set non-zero ones. Following 913 * are some fields set to 0/false for the stated reason: 914 * 915 * ->notify_os_intr_reg In powernv, send intrs to HV 916 * ->rsvd_txbuf_count Not supported yet. 917 * ->notify_disable False for NX windows 918 * ->xtra_write False for NX windows 919 * ->notify_early NA for NX windows 920 * ->lnotify_lpid NA for Tx windows 921 * ->lnotify_pid NA for Tx windows 922 * ->lnotify_tid NA for Tx windows 923 * ->tx_win_cred_mode Ignore for now for NX windows 924 * ->rx_win_cred_mode Ignore for now for NX windows 925 */ 926 memset(winctx, 0, sizeof(struct vas_winctx)); 927 928 winctx->wcreds_max = txwin->wcreds_max; 929 930 winctx->user_win = txattr->user_win; 931 winctx->nx_win = txwin->rxwin->nx_win; 932 winctx->pin_win = txattr->pin_win; 933 winctx->rej_no_credit = txattr->rej_no_credit; 934 winctx->rsvd_txbuf_enable = txattr->rsvd_txbuf_enable; 935 936 winctx->rx_wcred_mode = txattr->rx_wcred_mode; 937 winctx->tx_wcred_mode = txattr->tx_wcred_mode; 938 winctx->rx_word_mode = txattr->rx_win_ord_mode; 939 winctx->tx_word_mode = txattr->tx_win_ord_mode; 940 winctx->rsvd_txbuf_count = txattr->rsvd_txbuf_count; 941 942 winctx->intr_disable = true; 943 if (winctx->nx_win) 944 winctx->data_stamp = true; 945 946 winctx->lpid = txattr->lpid; 947 winctx->pidr = txattr->pidr; 948 winctx->rx_win_id = txwin->rxwin->winid; 949 /* 950 * IRQ and fault window setup is successful. Set fault window 951 * for the send window so that ready to handle faults. 952 */ 953 if (txwin->vinst->virq) 954 winctx->fault_win_id = txwin->vinst->fault_win->winid; 955 956 winctx->dma_type = VAS_DMA_TYPE_INJECT; 957 winctx->tc_mode = txattr->tc_mode; 958 winctx->min_scope = VAS_SCOPE_LOCAL; 959 winctx->max_scope = VAS_SCOPE_VECTORED_GROUP; 960 if (txwin->vinst->virq) 961 winctx->irq_port = txwin->vinst->irq_port; 962 963 winctx->pswid = txattr->pswid ? txattr->pswid : 964 encode_pswid(txwin->vinst->vas_id, txwin->winid); 965 } 966 967 static bool tx_win_args_valid(enum vas_cop_type cop, 968 struct vas_tx_win_attr *attr) 969 { 970 if (attr->tc_mode != VAS_THRESH_DISABLED) 971 return false; 972 973 if (cop > VAS_COP_TYPE_MAX) 974 return false; 975 976 if (attr->wcreds_max > VAS_TX_WCREDS_MAX) 977 return false; 978 979 if (attr->user_win && 980 (cop != VAS_COP_TYPE_FTW || attr->rsvd_txbuf_count)) 981 return false; 982 983 return true; 984 } 985 986 struct vas_window *vas_tx_win_open(int vasid, enum vas_cop_type cop, 987 struct vas_tx_win_attr *attr) 988 { 989 int rc; 990 struct vas_window *txwin; 991 struct vas_window *rxwin; 992 struct vas_winctx winctx; 993 struct vas_instance *vinst; 994 995 trace_vas_tx_win_open(current, vasid, cop, attr); 996 997 if (!tx_win_args_valid(cop, attr)) 998 return ERR_PTR(-EINVAL); 999 1000 /* 1001 * If caller did not specify a vasid but specified the PSWID of a 1002 * receive window (applicable only to FTW windows), use the vasid 1003 * from that receive window. 1004 */ 1005 if (vasid == -1 && attr->pswid) 1006 decode_pswid(attr->pswid, &vasid, NULL); 1007 1008 vinst = find_vas_instance(vasid); 1009 if (!vinst) { 1010 pr_devel("vasid %d not found!\n", vasid); 1011 return ERR_PTR(-EINVAL); 1012 } 1013 1014 rxwin = get_vinst_rxwin(vinst, cop, attr->pswid); 1015 if (IS_ERR(rxwin)) { 1016 pr_devel("No RxWin for vasid %d, cop %d\n", vasid, cop); 1017 return rxwin; 1018 } 1019 1020 txwin = vas_window_alloc(vinst); 1021 if (IS_ERR(txwin)) { 1022 rc = PTR_ERR(txwin); 1023 goto put_rxwin; 1024 } 1025 1026 txwin->cop = cop; 1027 txwin->tx_win = 1; 1028 txwin->rxwin = rxwin; 1029 txwin->nx_win = txwin->rxwin->nx_win; 1030 txwin->user_win = attr->user_win; 1031 txwin->wcreds_max = attr->wcreds_max ?: VAS_WCREDS_DEFAULT; 1032 1033 init_winctx_for_txwin(txwin, attr, &winctx); 1034 1035 init_winctx_regs(txwin, &winctx); 1036 1037 /* 1038 * If its a kernel send window, map the window address into the 1039 * kernel's address space. For user windows, user must issue an 1040 * mmap() to map the window into their address space. 1041 * 1042 * NOTE: If kernel ever resubmits a user CRB after handling a page 1043 * fault, we will need to map this into kernel as well. 1044 */ 1045 if (!txwin->user_win) { 1046 txwin->paste_kaddr = map_paste_region(txwin); 1047 if (IS_ERR(txwin->paste_kaddr)) { 1048 rc = PTR_ERR(txwin->paste_kaddr); 1049 goto free_window; 1050 } 1051 } else { 1052 /* 1053 * Interrupt hanlder or fault window setup failed. Means 1054 * NX can not generate fault for page fault. So not 1055 * opening for user space tx window. 1056 */ 1057 if (!vinst->virq) { 1058 rc = -ENODEV; 1059 goto free_window; 1060 } 1061 /* 1062 * A user mapping must ensure that context switch issues 1063 * CP_ABORT for this thread. 1064 */ 1065 rc = set_thread_uses_vas(); 1066 if (rc) 1067 goto free_window; 1068 1069 /* 1070 * Window opened by a child thread may not be closed when 1071 * it exits. So take reference to its pid and release it 1072 * when the window is free by parent thread. 1073 * Acquire a reference to the task's pid to make sure 1074 * pid will not be re-used - needed only for multithread 1075 * applications. 1076 */ 1077 txwin->pid = get_task_pid(current, PIDTYPE_PID); 1078 /* 1079 * Acquire a reference to the task's mm. 1080 */ 1081 txwin->mm = get_task_mm(current); 1082 1083 if (!txwin->mm) { 1084 put_pid(txwin->pid); 1085 pr_err("VAS: pid(%d): mm_struct is not found\n", 1086 current->pid); 1087 rc = -EPERM; 1088 goto free_window; 1089 } 1090 1091 mmgrab(txwin->mm); 1092 mmput(txwin->mm); 1093 mm_context_add_copro(txwin->mm); 1094 /* 1095 * Process closes window during exit. In the case of 1096 * multithread application, the child thread can open 1097 * window and can exit without closing it. Expects parent 1098 * thread to use and close the window. So do not need 1099 * to take pid reference for parent thread. 1100 */ 1101 txwin->tgid = find_get_pid(task_tgid_vnr(current)); 1102 } 1103 1104 set_vinst_win(vinst, txwin); 1105 1106 return txwin; 1107 1108 free_window: 1109 vas_window_free(txwin); 1110 1111 put_rxwin: 1112 put_rx_win(rxwin); 1113 return ERR_PTR(rc); 1114 1115 } 1116 EXPORT_SYMBOL_GPL(vas_tx_win_open); 1117 1118 int vas_copy_crb(void *crb, int offset) 1119 { 1120 return vas_copy(crb, offset); 1121 } 1122 EXPORT_SYMBOL_GPL(vas_copy_crb); 1123 1124 #define RMA_LSMP_REPORT_ENABLE PPC_BIT(53) 1125 int vas_paste_crb(struct vas_window *txwin, int offset, bool re) 1126 { 1127 int rc; 1128 void *addr; 1129 uint64_t val; 1130 1131 trace_vas_paste_crb(current, txwin); 1132 1133 /* 1134 * Only NX windows are supported for now and hardware assumes 1135 * report-enable flag is set for NX windows. Ensure software 1136 * complies too. 1137 */ 1138 WARN_ON_ONCE(txwin->nx_win && !re); 1139 1140 addr = txwin->paste_kaddr; 1141 if (re) { 1142 /* 1143 * Set the REPORT_ENABLE bit (equivalent to writing 1144 * to 1K offset of the paste address) 1145 */ 1146 val = SET_FIELD(RMA_LSMP_REPORT_ENABLE, 0ULL, 1); 1147 addr += val; 1148 } 1149 1150 /* 1151 * Map the raw CR value from vas_paste() to an error code (there 1152 * is just pass or fail for now though). 1153 */ 1154 rc = vas_paste(addr, offset); 1155 if (rc == 2) 1156 rc = 0; 1157 else 1158 rc = -EINVAL; 1159 1160 pr_debug("Txwin #%d: Msg count %llu\n", txwin->winid, 1161 read_hvwc_reg(txwin, VREG(LRFIFO_PUSH))); 1162 1163 return rc; 1164 } 1165 EXPORT_SYMBOL_GPL(vas_paste_crb); 1166 1167 /* 1168 * If credit checking is enabled for this window, poll for the return 1169 * of window credits (i.e for NX engines to process any outstanding CRBs). 1170 * Since NX-842 waits for the CRBs to be processed before closing the 1171 * window, we should not have to wait for too long. 1172 * 1173 * TODO: We retry in 10ms intervals now. We could/should probably peek at 1174 * the VAS_LRFIFO_PUSH_OFFSET register to get an estimate of pending 1175 * CRBs on the FIFO and compute the delay dynamically on each retry. 1176 * But that is not really needed until we support NX-GZIP access from 1177 * user space. (NX-842 driver waits for CSB and Fast thread-wakeup 1178 * doesn't use credit checking). 1179 */ 1180 static void poll_window_credits(struct vas_window *window) 1181 { 1182 u64 val; 1183 int creds, mode; 1184 int count = 0; 1185 1186 val = read_hvwc_reg(window, VREG(WINCTL)); 1187 if (window->tx_win) 1188 mode = GET_FIELD(VAS_WINCTL_TX_WCRED_MODE, val); 1189 else 1190 mode = GET_FIELD(VAS_WINCTL_RX_WCRED_MODE, val); 1191 1192 if (!mode) 1193 return; 1194 retry: 1195 if (window->tx_win) { 1196 val = read_hvwc_reg(window, VREG(TX_WCRED)); 1197 creds = GET_FIELD(VAS_TX_WCRED, val); 1198 } else { 1199 val = read_hvwc_reg(window, VREG(LRX_WCRED)); 1200 creds = GET_FIELD(VAS_LRX_WCRED, val); 1201 } 1202 1203 /* 1204 * Takes around few milliseconds to complete all pending requests 1205 * and return credits. 1206 * TODO: Scan fault FIFO and invalidate CRBs points to this window 1207 * and issue CRB Kill to stop all pending requests. Need only 1208 * if there is a bug in NX or fault handling in kernel. 1209 */ 1210 if (creds < window->wcreds_max) { 1211 val = 0; 1212 set_current_state(TASK_UNINTERRUPTIBLE); 1213 schedule_timeout(msecs_to_jiffies(10)); 1214 count++; 1215 /* 1216 * Process can not close send window until all credits are 1217 * returned. 1218 */ 1219 if (!(count % 1000)) 1220 pr_warn_ratelimited("VAS: pid %d stuck. Waiting for credits returned for Window(%d). creds %d, Retries %d\n", 1221 vas_window_pid(window), window->winid, 1222 creds, count); 1223 1224 goto retry; 1225 } 1226 } 1227 1228 /* 1229 * Wait for the window to go to "not-busy" state. It should only take a 1230 * short time to queue a CRB, so window should not be busy for too long. 1231 * Trying 5ms intervals. 1232 */ 1233 static void poll_window_busy_state(struct vas_window *window) 1234 { 1235 int busy; 1236 u64 val; 1237 int count = 0; 1238 1239 retry: 1240 val = read_hvwc_reg(window, VREG(WIN_STATUS)); 1241 busy = GET_FIELD(VAS_WIN_BUSY, val); 1242 if (busy) { 1243 val = 0; 1244 set_current_state(TASK_UNINTERRUPTIBLE); 1245 schedule_timeout(msecs_to_jiffies(10)); 1246 count++; 1247 /* 1248 * Takes around few milliseconds to process all pending 1249 * requests. 1250 */ 1251 if (!(count % 1000)) 1252 pr_warn_ratelimited("VAS: pid %d stuck. Window (ID=%d) is in busy state. Retries %d\n", 1253 vas_window_pid(window), window->winid, count); 1254 1255 goto retry; 1256 } 1257 } 1258 1259 /* 1260 * Have the hardware cast a window out of cache and wait for it to 1261 * be completed. 1262 * 1263 * NOTE: It can take a relatively long time to cast the window context 1264 * out of the cache. It is not strictly necessary to cast out if: 1265 * 1266 * - we clear the "Pin Window" bit (so hardware is free to evict) 1267 * 1268 * - we re-initialize the window context when it is reassigned. 1269 * 1270 * We do the former in vas_win_close() and latter in vas_win_open(). 1271 * So, ignoring the cast-out for now. We can add it as needed. If 1272 * casting out becomes necessary we should consider offloading the 1273 * job to a worker thread, so the window close can proceed quickly. 1274 */ 1275 static void poll_window_castout(struct vas_window *window) 1276 { 1277 /* stub for now */ 1278 } 1279 1280 /* 1281 * Unpin and close a window so no new requests are accepted and the 1282 * hardware can evict this window from cache if necessary. 1283 */ 1284 static void unpin_close_window(struct vas_window *window) 1285 { 1286 u64 val; 1287 1288 val = read_hvwc_reg(window, VREG(WINCTL)); 1289 val = SET_FIELD(VAS_WINCTL_PIN, val, 0); 1290 val = SET_FIELD(VAS_WINCTL_OPEN, val, 0); 1291 write_hvwc_reg(window, VREG(WINCTL), val); 1292 } 1293 1294 /* 1295 * Close a window. 1296 * 1297 * See Section 1.12.1 of VAS workbook v1.05 for details on closing window: 1298 * - Disable new paste operations (unmap paste address) 1299 * - Poll for the "Window Busy" bit to be cleared 1300 * - Clear the Open/Enable bit for the Window. 1301 * - Poll for return of window Credits (implies FIFO empty for Rx win?) 1302 * - Unpin and cast window context out of cache 1303 * 1304 * Besides the hardware, kernel has some bookkeeping of course. 1305 */ 1306 int vas_win_close(struct vas_window *window) 1307 { 1308 if (!window) 1309 return 0; 1310 1311 if (!window->tx_win && atomic_read(&window->num_txwins) != 0) { 1312 pr_devel("Attempting to close an active Rx window!\n"); 1313 WARN_ON_ONCE(1); 1314 return -EBUSY; 1315 } 1316 1317 unmap_paste_region(window); 1318 1319 clear_vinst_win(window); 1320 1321 poll_window_busy_state(window); 1322 1323 unpin_close_window(window); 1324 1325 poll_window_credits(window); 1326 1327 poll_window_castout(window); 1328 1329 /* if send window, drop reference to matching receive window */ 1330 if (window->tx_win) { 1331 if (window->user_win) { 1332 /* Drop references to pid and mm */ 1333 put_pid(window->pid); 1334 if (window->mm) { 1335 mm_context_remove_copro(window->mm); 1336 mmdrop(window->mm); 1337 } 1338 } 1339 put_rx_win(window->rxwin); 1340 } 1341 1342 vas_window_free(window); 1343 1344 return 0; 1345 } 1346 EXPORT_SYMBOL_GPL(vas_win_close); 1347 1348 /* 1349 * Return credit for the given window. 1350 * Send windows and fault window uses credit mechanism as follows: 1351 * 1352 * Send windows: 1353 * - The default number of credits available for each send window is 1354 * 1024. It means 1024 requests can be issued asynchronously at the 1355 * same time. If the credit is not available, that request will be 1356 * returned with RMA_Busy. 1357 * - One credit is taken when NX request is issued. 1358 * - This credit is returned after NX processed that request. 1359 * - If NX encounters translation error, kernel will return the 1360 * credit on the specific send window after processing the fault CRB. 1361 * 1362 * Fault window: 1363 * - The total number credits available is FIFO_SIZE/CRB_SIZE. 1364 * Means 4MB/128 in the current implementation. If credit is not 1365 * available, RMA_Reject is returned. 1366 * - A credit is taken when NX pastes CRB in fault FIFO. 1367 * - The kernel with return credit on fault window after reading entry 1368 * from fault FIFO. 1369 */ 1370 void vas_return_credit(struct vas_window *window, bool tx) 1371 { 1372 uint64_t val; 1373 1374 val = 0ULL; 1375 if (tx) { /* send window */ 1376 val = SET_FIELD(VAS_TX_WCRED, val, 1); 1377 write_hvwc_reg(window, VREG(TX_WCRED_ADDER), val); 1378 } else { 1379 val = SET_FIELD(VAS_LRX_WCRED, val, 1); 1380 write_hvwc_reg(window, VREG(LRX_WCRED_ADDER), val); 1381 } 1382 } 1383 1384 struct vas_window *vas_pswid_to_window(struct vas_instance *vinst, 1385 uint32_t pswid) 1386 { 1387 struct vas_window *window; 1388 int winid; 1389 1390 if (!pswid) { 1391 pr_devel("%s: called for pswid 0!\n", __func__); 1392 return ERR_PTR(-ESRCH); 1393 } 1394 1395 decode_pswid(pswid, NULL, &winid); 1396 1397 if (winid >= VAS_WINDOWS_PER_CHIP) 1398 return ERR_PTR(-ESRCH); 1399 1400 /* 1401 * If application closes the window before the hardware 1402 * returns the fault CRB, we should wait in vas_win_close() 1403 * for the pending requests. so the window must be active 1404 * and the process alive. 1405 * 1406 * If its a kernel process, we should not get any faults and 1407 * should not get here. 1408 */ 1409 window = vinst->windows[winid]; 1410 1411 if (!window) { 1412 pr_err("PSWID decode: Could not find window for winid %d pswid %d vinst 0x%p\n", 1413 winid, pswid, vinst); 1414 return NULL; 1415 } 1416 1417 /* 1418 * Do some sanity checks on the decoded window. Window should be 1419 * NX GZIP user send window. FTW windows should not incur faults 1420 * since their CRBs are ignored (not queued on FIFO or processed 1421 * by NX). 1422 */ 1423 if (!window->tx_win || !window->user_win || !window->nx_win || 1424 window->cop == VAS_COP_TYPE_FAULT || 1425 window->cop == VAS_COP_TYPE_FTW) { 1426 pr_err("PSWID decode: id %d, tx %d, user %d, nx %d, cop %d\n", 1427 winid, window->tx_win, window->user_win, 1428 window->nx_win, window->cop); 1429 WARN_ON(1); 1430 } 1431 1432 return window; 1433 } 1434