1 /* 2 * SN Platform GRU Driver 3 * 4 * GRU HANDLE DEFINITION 5 * 6 * Copyright (c) 2008 Silicon Graphics, Inc. All Rights Reserved. 7 * 8 * This program is free software; you can redistribute it and/or modify 9 * it under the terms of the GNU General Public License as published by 10 * the Free Software Foundation; either version 2 of the License, or 11 * (at your option) any later version. 12 * 13 * This program is distributed in the hope that it will be useful, 14 * but WITHOUT ANY WARRANTY; without even the implied warranty of 15 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 16 * GNU General Public License for more details. 17 * 18 * You should have received a copy of the GNU General Public License 19 * along with this program; if not, write to the Free Software 20 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA 21 */ 22 23 #ifndef __GRUHANDLES_H__ 24 #define __GRUHANDLES_H__ 25 #include "gru_instructions.h" 26 27 /* 28 * Manifest constants for GRU Memory Map 29 */ 30 #define GRU_GSEG0_BASE 0 31 #define GRU_MCS_BASE (64 * 1024 * 1024) 32 #define GRU_SIZE (128UL * 1024 * 1024) 33 34 /* Handle & resource counts */ 35 #define GRU_NUM_CB 128 36 #define GRU_NUM_DSR_BYTES (32 * 1024) 37 #define GRU_NUM_TFM 16 38 #define GRU_NUM_TGH 24 39 #define GRU_NUM_CBE 128 40 #define GRU_NUM_TFH 128 41 #define GRU_NUM_CCH 16 42 43 /* Maximum resource counts that can be reserved by user programs */ 44 #define GRU_NUM_USER_CBR GRU_NUM_CBE 45 #define GRU_NUM_USER_DSR_BYTES GRU_NUM_DSR_BYTES 46 47 /* Bytes per handle & handle stride. Code assumes all cb, tfh, cbe handles 48 * are the same */ 49 #define GRU_HANDLE_BYTES 64 50 #define GRU_HANDLE_STRIDE 256 51 52 /* Base addresses of handles */ 53 #define GRU_TFM_BASE (GRU_MCS_BASE + 0x00000) 54 #define GRU_TGH_BASE (GRU_MCS_BASE + 0x08000) 55 #define GRU_CBE_BASE (GRU_MCS_BASE + 0x10000) 56 #define GRU_TFH_BASE (GRU_MCS_BASE + 0x18000) 57 #define GRU_CCH_BASE (GRU_MCS_BASE + 0x20000) 58 59 /* User gseg constants */ 60 #define GRU_GSEG_STRIDE (4 * 1024 * 1024) 61 #define GSEG_BASE(a) ((a) & ~(GRU_GSEG_PAGESIZE - 1)) 62 63 /* Data segment constants */ 64 #define GRU_DSR_AU_BYTES 1024 65 #define GRU_DSR_CL (GRU_NUM_DSR_BYTES / GRU_CACHE_LINE_BYTES) 66 #define GRU_DSR_AU_CL (GRU_DSR_AU_BYTES / GRU_CACHE_LINE_BYTES) 67 #define GRU_DSR_AU (GRU_NUM_DSR_BYTES / GRU_DSR_AU_BYTES) 68 69 /* Control block constants */ 70 #define GRU_CBR_AU_SIZE 2 71 #define GRU_CBR_AU (GRU_NUM_CBE / GRU_CBR_AU_SIZE) 72 73 /* Convert resource counts to the number of AU */ 74 #define GRU_DS_BYTES_TO_AU(n) DIV_ROUND_UP(n, GRU_DSR_AU_BYTES) 75 #define GRU_CB_COUNT_TO_AU(n) DIV_ROUND_UP(n, GRU_CBR_AU_SIZE) 76 77 /* UV limits */ 78 #define GRU_CHIPLETS_PER_HUB 2 79 #define GRU_HUBS_PER_BLADE 1 80 #define GRU_CHIPLETS_PER_BLADE (GRU_HUBS_PER_BLADE * GRU_CHIPLETS_PER_HUB) 81 82 /* User GRU Gseg offsets */ 83 #define GRU_CB_BASE 0 84 #define GRU_CB_LIMIT (GRU_CB_BASE + GRU_HANDLE_STRIDE * GRU_NUM_CBE) 85 #define GRU_DS_BASE 0x20000 86 #define GRU_DS_LIMIT (GRU_DS_BASE + GRU_NUM_DSR_BYTES) 87 88 /* Convert a GRU physical address to the chiplet offset */ 89 #define GSEGPOFF(h) ((h) & (GRU_SIZE - 1)) 90 91 /* Convert an arbitrary handle address to the beginning of the GRU segment */ 92 #define GRUBASE(h) ((void *)((unsigned long)(h) & ~(GRU_SIZE - 1))) 93 94 /* Test a valid handle address to determine the type */ 95 #define TYPE_IS(hn, h) ((h) >= GRU_##hn##_BASE && (h) < \ 96 GRU_##hn##_BASE + GRU_NUM_##hn * GRU_HANDLE_STRIDE && \ 97 (((h) & (GRU_HANDLE_STRIDE - 1)) == 0)) 98 99 100 /* General addressing macros. */ 101 static inline void *get_gseg_base_address(void *base, int ctxnum) 102 { 103 return (void *)(base + GRU_GSEG0_BASE + GRU_GSEG_STRIDE * ctxnum); 104 } 105 106 static inline void *get_gseg_base_address_cb(void *base, int ctxnum, int line) 107 { 108 return (void *)(get_gseg_base_address(base, ctxnum) + 109 GRU_CB_BASE + GRU_HANDLE_STRIDE * line); 110 } 111 112 static inline void *get_gseg_base_address_ds(void *base, int ctxnum, int line) 113 { 114 return (void *)(get_gseg_base_address(base, ctxnum) + GRU_DS_BASE + 115 GRU_CACHE_LINE_BYTES * line); 116 } 117 118 static inline struct gru_tlb_fault_map *get_tfm(void *base, int ctxnum) 119 { 120 return (struct gru_tlb_fault_map *)(base + GRU_TFM_BASE + 121 ctxnum * GRU_HANDLE_STRIDE); 122 } 123 124 static inline struct gru_tlb_global_handle *get_tgh(void *base, int ctxnum) 125 { 126 return (struct gru_tlb_global_handle *)(base + GRU_TGH_BASE + 127 ctxnum * GRU_HANDLE_STRIDE); 128 } 129 130 static inline struct gru_control_block_extended *get_cbe(void *base, int ctxnum) 131 { 132 return (struct gru_control_block_extended *)(base + GRU_CBE_BASE + 133 ctxnum * GRU_HANDLE_STRIDE); 134 } 135 136 static inline struct gru_tlb_fault_handle *get_tfh(void *base, int ctxnum) 137 { 138 return (struct gru_tlb_fault_handle *)(base + GRU_TFH_BASE + 139 ctxnum * GRU_HANDLE_STRIDE); 140 } 141 142 static inline struct gru_context_configuration_handle *get_cch(void *base, 143 int ctxnum) 144 { 145 return (struct gru_context_configuration_handle *)(base + 146 GRU_CCH_BASE + ctxnum * GRU_HANDLE_STRIDE); 147 } 148 149 static inline unsigned long get_cb_number(void *cb) 150 { 151 return (((unsigned long)cb - GRU_CB_BASE) % GRU_GSEG_PAGESIZE) / 152 GRU_HANDLE_STRIDE; 153 } 154 155 /* byte offset to a specific GRU chiplet. (p=pnode, c=chiplet (0 or 1)*/ 156 static inline unsigned long gru_chiplet_paddr(unsigned long paddr, int pnode, 157 int chiplet) 158 { 159 return paddr + GRU_SIZE * (2 * pnode + chiplet); 160 } 161 162 static inline void *gru_chiplet_vaddr(void *vaddr, int pnode, int chiplet) 163 { 164 return vaddr + GRU_SIZE * (2 * pnode + chiplet); 165 } 166 167 static inline struct gru_control_block_extended *gru_tfh_to_cbe( 168 struct gru_tlb_fault_handle *tfh) 169 { 170 unsigned long cbe; 171 172 cbe = (unsigned long)tfh - GRU_TFH_BASE + GRU_CBE_BASE; 173 return (struct gru_control_block_extended*)cbe; 174 } 175 176 177 178 179 /* 180 * Global TLB Fault Map 181 * Bitmap of outstanding TLB misses needing interrupt/polling service. 182 * 183 */ 184 struct gru_tlb_fault_map { 185 unsigned long fault_bits[BITS_TO_LONGS(GRU_NUM_CBE)]; 186 unsigned long fill0[2]; 187 unsigned long done_bits[BITS_TO_LONGS(GRU_NUM_CBE)]; 188 unsigned long fill1[2]; 189 }; 190 191 /* 192 * TGH - TLB Global Handle 193 * Used for TLB flushing. 194 * 195 */ 196 struct gru_tlb_global_handle { 197 unsigned int cmd:1; /* DW 0 */ 198 unsigned int delresp:1; 199 unsigned int opc:1; 200 unsigned int fill1:5; 201 202 unsigned int fill2:8; 203 204 unsigned int status:2; 205 unsigned long fill3:2; 206 unsigned int state:3; 207 unsigned long fill4:1; 208 209 unsigned int cause:3; 210 unsigned long fill5:37; 211 212 unsigned long vaddr:64; /* DW 1 */ 213 214 unsigned int asid:24; /* DW 2 */ 215 unsigned int fill6:8; 216 217 unsigned int pagesize:5; 218 unsigned int fill7:11; 219 220 unsigned int global:1; 221 unsigned int fill8:15; 222 223 unsigned long vaddrmask:39; /* DW 3 */ 224 unsigned int fill9:9; 225 unsigned int n:10; 226 unsigned int fill10:6; 227 228 unsigned int ctxbitmap:16; /* DW4 */ 229 unsigned long fill11[3]; 230 }; 231 232 enum gru_tgh_cmd { 233 TGHCMD_START 234 }; 235 236 enum gru_tgh_opc { 237 TGHOP_TLBNOP, 238 TGHOP_TLBINV 239 }; 240 241 enum gru_tgh_status { 242 TGHSTATUS_IDLE, 243 TGHSTATUS_EXCEPTION, 244 TGHSTATUS_ACTIVE 245 }; 246 247 enum gru_tgh_state { 248 TGHSTATE_IDLE, 249 TGHSTATE_PE_INVAL, 250 TGHSTATE_INTERRUPT_INVAL, 251 TGHSTATE_WAITDONE, 252 TGHSTATE_RESTART_CTX, 253 }; 254 255 enum gru_tgh_cause { 256 TGHCAUSE_RR_ECC, 257 TGHCAUSE_TLB_ECC, 258 TGHCAUSE_LRU_ECC, 259 TGHCAUSE_PS_ECC, 260 TGHCAUSE_MUL_ERR, 261 TGHCAUSE_DATA_ERR, 262 TGHCAUSE_SW_FORCE 263 }; 264 265 266 /* 267 * TFH - TLB Global Handle 268 * Used for TLB dropins into the GRU TLB. 269 * 270 */ 271 struct gru_tlb_fault_handle { 272 unsigned int cmd:1; /* DW 0 - low 32*/ 273 unsigned int delresp:1; 274 unsigned int fill0:2; 275 unsigned int opc:3; 276 unsigned int fill1:9; 277 278 unsigned int status:2; 279 unsigned int fill2:2; 280 unsigned int state:3; 281 unsigned int fill3:1; 282 283 unsigned int cause:6; 284 unsigned int cb_int:1; 285 unsigned int fill4:1; 286 287 unsigned int indexway:12; /* DW 0 - high 32 */ 288 unsigned int fill5:4; 289 290 unsigned int ctxnum:4; 291 unsigned int fill6:12; 292 293 unsigned long missvaddr:64; /* DW 1 */ 294 295 unsigned int missasid:24; /* DW 2 */ 296 unsigned int fill7:8; 297 unsigned int fillasid:24; 298 unsigned int dirty:1; 299 unsigned int gaa:2; 300 unsigned long fill8:5; 301 302 unsigned long pfn:41; /* DW 3 */ 303 unsigned int fill9:7; 304 unsigned int pagesize:5; 305 unsigned int fill10:11; 306 307 unsigned long fillvaddr:64; /* DW 4 */ 308 309 unsigned long fill11[3]; 310 }; 311 312 enum gru_tfh_opc { 313 TFHOP_NOOP, 314 TFHOP_RESTART, 315 TFHOP_WRITE_ONLY, 316 TFHOP_WRITE_RESTART, 317 TFHOP_EXCEPTION, 318 TFHOP_USER_POLLING_MODE = 7, 319 }; 320 321 enum tfh_status { 322 TFHSTATUS_IDLE, 323 TFHSTATUS_EXCEPTION, 324 TFHSTATUS_ACTIVE, 325 }; 326 327 enum tfh_state { 328 TFHSTATE_INACTIVE, 329 TFHSTATE_IDLE, 330 TFHSTATE_MISS_UPM, 331 TFHSTATE_MISS_FMM, 332 TFHSTATE_HW_ERR, 333 TFHSTATE_WRITE_TLB, 334 TFHSTATE_RESTART_CBR, 335 }; 336 337 /* TFH cause bits */ 338 enum tfh_cause { 339 TFHCAUSE_NONE, 340 TFHCAUSE_TLB_MISS, 341 TFHCAUSE_TLB_MOD, 342 TFHCAUSE_HW_ERROR_RR, 343 TFHCAUSE_HW_ERROR_MAIN_ARRAY, 344 TFHCAUSE_HW_ERROR_VALID, 345 TFHCAUSE_HW_ERROR_PAGESIZE, 346 TFHCAUSE_INSTRUCTION_EXCEPTION, 347 TFHCAUSE_UNCORRECTIBLE_ERROR, 348 }; 349 350 /* GAA values */ 351 #define GAA_RAM 0x0 352 #define GAA_NCRAM 0x2 353 #define GAA_MMIO 0x1 354 #define GAA_REGISTER 0x3 355 356 /* GRU paddr shift for pfn. (NOTE: shift is NOT by actual pagesize) */ 357 #define GRU_PADDR_SHIFT 12 358 359 /* 360 * Context Configuration handle 361 * Used to allocate resources to a GSEG context. 362 * 363 */ 364 struct gru_context_configuration_handle { 365 unsigned int cmd:1; /* DW0 */ 366 unsigned int delresp:1; 367 unsigned int opc:3; 368 unsigned int unmap_enable:1; 369 unsigned int req_slice_set_enable:1; 370 unsigned int req_slice:2; 371 unsigned int cb_int_enable:1; 372 unsigned int tlb_int_enable:1; 373 unsigned int tfm_fault_bit_enable:1; 374 unsigned int tlb_int_select:4; 375 376 unsigned int status:2; 377 unsigned int state:2; 378 unsigned int reserved2:4; 379 380 unsigned int cause:4; 381 unsigned int tfm_done_bit_enable:1; 382 unsigned int unused:3; 383 384 unsigned int dsr_allocation_map; 385 386 unsigned long cbr_allocation_map; /* DW1 */ 387 388 unsigned int asid[8]; /* DW 2 - 5 */ 389 unsigned short sizeavail[8]; /* DW 6 - 7 */ 390 } __attribute__ ((packed)); 391 392 enum gru_cch_opc { 393 CCHOP_START = 1, 394 CCHOP_ALLOCATE, 395 CCHOP_INTERRUPT, 396 CCHOP_DEALLOCATE, 397 CCHOP_INTERRUPT_SYNC, 398 }; 399 400 enum gru_cch_status { 401 CCHSTATUS_IDLE, 402 CCHSTATUS_EXCEPTION, 403 CCHSTATUS_ACTIVE, 404 }; 405 406 enum gru_cch_state { 407 CCHSTATE_INACTIVE, 408 CCHSTATE_MAPPED, 409 CCHSTATE_ACTIVE, 410 CCHSTATE_INTERRUPTED, 411 }; 412 413 /* CCH Exception cause */ 414 enum gru_cch_cause { 415 CCHCAUSE_REGION_REGISTER_WRITE_ERROR = 1, 416 CCHCAUSE_ILLEGAL_OPCODE = 2, 417 CCHCAUSE_INVALID_START_REQUEST = 3, 418 CCHCAUSE_INVALID_ALLOCATION_REQUEST = 4, 419 CCHCAUSE_INVALID_DEALLOCATION_REQUEST = 5, 420 CCHCAUSE_INVALID_INTERRUPT_REQUEST = 6, 421 CCHCAUSE_CCH_BUSY = 7, 422 CCHCAUSE_NO_CBRS_TO_ALLOCATE = 8, 423 CCHCAUSE_BAD_TFM_CONFIG = 9, 424 CCHCAUSE_CBR_RESOURCES_OVERSUBSCRIPED = 10, 425 CCHCAUSE_DSR_RESOURCES_OVERSUBSCRIPED = 11, 426 CCHCAUSE_CBR_DEALLOCATION_ERROR = 12, 427 }; 428 /* 429 * CBE - Control Block Extended 430 * Maintains internal GRU state for active CBs. 431 * 432 */ 433 struct gru_control_block_extended { 434 unsigned int reserved0:1; /* DW 0 - low */ 435 unsigned int imacpy:3; 436 unsigned int reserved1:4; 437 unsigned int xtypecpy:3; 438 unsigned int iaa0cpy:2; 439 unsigned int iaa1cpy:2; 440 unsigned int reserved2:1; 441 unsigned int opccpy:8; 442 unsigned int exopccpy:8; 443 444 unsigned int idef2cpy:22; /* DW 0 - high */ 445 unsigned int reserved3:10; 446 447 unsigned int idef4cpy:22; /* DW 1 */ 448 unsigned int reserved4:10; 449 unsigned int idef4upd:22; 450 unsigned int reserved5:10; 451 452 unsigned long idef1upd:64; /* DW 2 */ 453 454 unsigned long idef5cpy:64; /* DW 3 */ 455 456 unsigned long idef6cpy:64; /* DW 4 */ 457 458 unsigned long idef3upd:64; /* DW 5 */ 459 460 unsigned long idef5upd:64; /* DW 6 */ 461 462 unsigned int idef2upd:22; /* DW 7 */ 463 unsigned int reserved6:10; 464 465 unsigned int ecause:20; 466 unsigned int cbrstate:4; 467 unsigned int cbrexecstatus:8; 468 }; 469 470 /* CBE fields for active BCOPY instructions */ 471 #define cbe_baddr0 idef1upd 472 #define cbe_baddr1 idef3upd 473 #define cbe_src_cl idef6cpy 474 #define cbe_nelemcur idef5upd 475 476 enum gru_cbr_state { 477 CBRSTATE_INACTIVE, 478 CBRSTATE_IDLE, 479 CBRSTATE_PE_CHECK, 480 CBRSTATE_QUEUED, 481 CBRSTATE_WAIT_RESPONSE, 482 CBRSTATE_INTERRUPTED, 483 CBRSTATE_INTERRUPTED_MISS_FMM, 484 CBRSTATE_BUSY_INTERRUPT_MISS_FMM, 485 CBRSTATE_INTERRUPTED_MISS_UPM, 486 CBRSTATE_BUSY_INTERRUPTED_MISS_UPM, 487 CBRSTATE_REQUEST_ISSUE, 488 CBRSTATE_BUSY_INTERRUPT, 489 }; 490 491 /* CBE cbrexecstatus bits - defined in gru_instructions.h*/ 492 /* CBE ecause bits - defined in gru_instructions.h */ 493 494 /* 495 * Convert a processor pagesize into the strange encoded pagesize used by the 496 * GRU. Processor pagesize is encoded as log of bytes per page. (or PAGE_SHIFT) 497 * pagesize log pagesize grupagesize 498 * 4k 12 0 499 * 16k 14 1 500 * 64k 16 2 501 * 256k 18 3 502 * 1m 20 4 503 * 2m 21 5 504 * 4m 22 6 505 * 16m 24 7 506 * 64m 26 8 507 * ... 508 */ 509 #define GRU_PAGESIZE(sh) ((((sh) > 20 ? (sh) + 2 : (sh)) >> 1) - 6) 510 #define GRU_SIZEAVAIL(sh) (1UL << GRU_PAGESIZE(sh)) 511 512 /* minimum TLB purge count to ensure a full purge */ 513 #define GRUMAXINVAL 1024UL 514 515 int cch_allocate(struct gru_context_configuration_handle *cch); 516 int cch_start(struct gru_context_configuration_handle *cch); 517 int cch_interrupt(struct gru_context_configuration_handle *cch); 518 int cch_deallocate(struct gru_context_configuration_handle *cch); 519 int cch_interrupt_sync(struct gru_context_configuration_handle *cch); 520 int tgh_invalidate(struct gru_tlb_global_handle *tgh, unsigned long vaddr, 521 unsigned long vaddrmask, int asid, int pagesize, int global, int n, 522 unsigned short ctxbitmap); 523 int tfh_write_only(struct gru_tlb_fault_handle *tfh, unsigned long paddr, 524 int gaa, unsigned long vaddr, int asid, int dirty, int pagesize); 525 void tfh_write_restart(struct gru_tlb_fault_handle *tfh, unsigned long paddr, 526 int gaa, unsigned long vaddr, int asid, int dirty, int pagesize); 527 void tfh_restart(struct gru_tlb_fault_handle *tfh); 528 void tfh_user_polling_mode(struct gru_tlb_fault_handle *tfh); 529 void tfh_exception(struct gru_tlb_fault_handle *tfh); 530 531 #endif /* __GRUHANDLES_H__ */ 532