1 /* 2 * Intel e7xxx Memory Controller kernel module 3 * (C) 2003 Linux Networx (http://lnxi.com) 4 * This file may be distributed under the terms of the 5 * GNU General Public License. 6 * 7 * See "enum e7xxx_chips" below for supported chipsets 8 * 9 * Written by Thayne Harbaugh 10 * Based on work by Dan Hollis <goemon at anime dot net> and others. 11 * http://www.anime.net/~goemon/linux-ecc/ 12 * 13 * Datasheet: 14 * http://www.intel.com/content/www/us/en/chipsets/e7501-chipset-memory-controller-hub-datasheet.html 15 * 16 * Contributors: 17 * Eric Biederman (Linux Networx) 18 * Tom Zimmerman (Linux Networx) 19 * Jim Garlick (Lawrence Livermore National Labs) 20 * Dave Peterson (Lawrence Livermore National Labs) 21 * That One Guy (Some other place) 22 * Wang Zhenyu (intel.com) 23 * 24 * $Id: edac_e7xxx.c,v 1.5.2.9 2005/10/05 00:43:44 dsp_llnl Exp $ 25 * 26 */ 27 28 #include <linux/module.h> 29 #include <linux/init.h> 30 #include <linux/pci.h> 31 #include <linux/pci_ids.h> 32 #include <linux/edac.h> 33 #include "edac_module.h" 34 35 #define EDAC_MOD_STR "e7xxx_edac" 36 37 #define e7xxx_printk(level, fmt, arg...) \ 38 edac_printk(level, "e7xxx", fmt, ##arg) 39 40 #define e7xxx_mc_printk(mci, level, fmt, arg...) \ 41 edac_mc_chipset_printk(mci, level, "e7xxx", fmt, ##arg) 42 43 #ifndef PCI_DEVICE_ID_INTEL_7205_0 44 #define PCI_DEVICE_ID_INTEL_7205_0 0x255d 45 #endif /* PCI_DEVICE_ID_INTEL_7205_0 */ 46 47 #ifndef PCI_DEVICE_ID_INTEL_7205_1_ERR 48 #define PCI_DEVICE_ID_INTEL_7205_1_ERR 0x2551 49 #endif /* PCI_DEVICE_ID_INTEL_7205_1_ERR */ 50 51 #ifndef PCI_DEVICE_ID_INTEL_7500_0 52 #define PCI_DEVICE_ID_INTEL_7500_0 0x2540 53 #endif /* PCI_DEVICE_ID_INTEL_7500_0 */ 54 55 #ifndef PCI_DEVICE_ID_INTEL_7500_1_ERR 56 #define PCI_DEVICE_ID_INTEL_7500_1_ERR 0x2541 57 #endif /* PCI_DEVICE_ID_INTEL_7500_1_ERR */ 58 59 #ifndef PCI_DEVICE_ID_INTEL_7501_0 60 #define PCI_DEVICE_ID_INTEL_7501_0 0x254c 61 #endif /* PCI_DEVICE_ID_INTEL_7501_0 */ 62 63 #ifndef PCI_DEVICE_ID_INTEL_7501_1_ERR 64 #define PCI_DEVICE_ID_INTEL_7501_1_ERR 0x2541 65 #endif /* PCI_DEVICE_ID_INTEL_7501_1_ERR */ 66 67 #ifndef PCI_DEVICE_ID_INTEL_7505_0 68 #define PCI_DEVICE_ID_INTEL_7505_0 0x2550 69 #endif /* PCI_DEVICE_ID_INTEL_7505_0 */ 70 71 #ifndef PCI_DEVICE_ID_INTEL_7505_1_ERR 72 #define PCI_DEVICE_ID_INTEL_7505_1_ERR 0x2551 73 #endif /* PCI_DEVICE_ID_INTEL_7505_1_ERR */ 74 75 #define E7XXX_NR_CSROWS 8 /* number of csrows */ 76 #define E7XXX_NR_DIMMS 8 /* 2 channels, 4 dimms/channel */ 77 78 /* E7XXX register addresses - device 0 function 0 */ 79 #define E7XXX_DRB 0x60 /* DRAM row boundary register (8b) */ 80 #define E7XXX_DRA 0x70 /* DRAM row attribute register (8b) */ 81 /* 82 * 31 Device width row 7 0=x8 1=x4 83 * 27 Device width row 6 84 * 23 Device width row 5 85 * 19 Device width row 4 86 * 15 Device width row 3 87 * 11 Device width row 2 88 * 7 Device width row 1 89 * 3 Device width row 0 90 */ 91 #define E7XXX_DRC 0x7C /* DRAM controller mode reg (32b) */ 92 /* 93 * 22 Number channels 0=1,1=2 94 * 19:18 DRB Granularity 32/64MB 95 */ 96 #define E7XXX_TOLM 0xC4 /* DRAM top of low memory reg (16b) */ 97 #define E7XXX_REMAPBASE 0xC6 /* DRAM remap base address reg (16b) */ 98 #define E7XXX_REMAPLIMIT 0xC8 /* DRAM remap limit address reg (16b) */ 99 100 /* E7XXX register addresses - device 0 function 1 */ 101 #define E7XXX_DRAM_FERR 0x80 /* DRAM first error register (8b) */ 102 #define E7XXX_DRAM_NERR 0x82 /* DRAM next error register (8b) */ 103 #define E7XXX_DRAM_CELOG_ADD 0xA0 /* DRAM first correctable memory */ 104 /* error address register (32b) */ 105 /* 106 * 31:28 Reserved 107 * 27:6 CE address (4k block 33:12) 108 * 5:0 Reserved 109 */ 110 #define E7XXX_DRAM_UELOG_ADD 0xB0 /* DRAM first uncorrectable memory */ 111 /* error address register (32b) */ 112 /* 113 * 31:28 Reserved 114 * 27:6 CE address (4k block 33:12) 115 * 5:0 Reserved 116 */ 117 #define E7XXX_DRAM_CELOG_SYNDROME 0xD0 /* DRAM first correctable memory */ 118 /* error syndrome register (16b) */ 119 120 enum e7xxx_chips { 121 E7500 = 0, 122 E7501, 123 E7505, 124 E7205, 125 }; 126 127 struct e7xxx_pvt { 128 struct pci_dev *bridge_ck; 129 u32 tolm; 130 u32 remapbase; 131 u32 remaplimit; 132 const struct e7xxx_dev_info *dev_info; 133 }; 134 135 struct e7xxx_dev_info { 136 u16 err_dev; 137 const char *ctl_name; 138 }; 139 140 struct e7xxx_error_info { 141 u8 dram_ferr; 142 u8 dram_nerr; 143 u32 dram_celog_add; 144 u16 dram_celog_syndrome; 145 u32 dram_uelog_add; 146 }; 147 148 static struct edac_pci_ctl_info *e7xxx_pci; 149 150 static const struct e7xxx_dev_info e7xxx_devs[] = { 151 [E7500] = { 152 .err_dev = PCI_DEVICE_ID_INTEL_7500_1_ERR, 153 .ctl_name = "E7500"}, 154 [E7501] = { 155 .err_dev = PCI_DEVICE_ID_INTEL_7501_1_ERR, 156 .ctl_name = "E7501"}, 157 [E7505] = { 158 .err_dev = PCI_DEVICE_ID_INTEL_7505_1_ERR, 159 .ctl_name = "E7505"}, 160 [E7205] = { 161 .err_dev = PCI_DEVICE_ID_INTEL_7205_1_ERR, 162 .ctl_name = "E7205"}, 163 }; 164 165 /* FIXME - is this valid for both SECDED and S4ECD4ED? */ 166 static inline int e7xxx_find_channel(u16 syndrome) 167 { 168 edac_dbg(3, "\n"); 169 170 if ((syndrome & 0xff00) == 0) 171 return 0; 172 173 if ((syndrome & 0x00ff) == 0) 174 return 1; 175 176 if ((syndrome & 0xf000) == 0 || (syndrome & 0x0f00) == 0) 177 return 0; 178 179 return 1; 180 } 181 182 static unsigned long ctl_page_to_phys(struct mem_ctl_info *mci, 183 unsigned long page) 184 { 185 u32 remap; 186 struct e7xxx_pvt *pvt = (struct e7xxx_pvt *)mci->pvt_info; 187 188 edac_dbg(3, "\n"); 189 190 if ((page < pvt->tolm) || 191 ((page >= 0x100000) && (page < pvt->remapbase))) 192 return page; 193 194 remap = (page - pvt->tolm) + pvt->remapbase; 195 196 if (remap < pvt->remaplimit) 197 return remap; 198 199 e7xxx_printk(KERN_ERR, "Invalid page %lx - out of range\n", page); 200 return pvt->tolm - 1; 201 } 202 203 static void process_ce(struct mem_ctl_info *mci, struct e7xxx_error_info *info) 204 { 205 u32 error_1b, page; 206 u16 syndrome; 207 int row; 208 int channel; 209 210 edac_dbg(3, "\n"); 211 /* read the error address */ 212 error_1b = info->dram_celog_add; 213 /* FIXME - should use PAGE_SHIFT */ 214 page = error_1b >> 6; /* convert the address to 4k page */ 215 /* read the syndrome */ 216 syndrome = info->dram_celog_syndrome; 217 /* FIXME - check for -1 */ 218 row = edac_mc_find_csrow_by_page(mci, page); 219 /* convert syndrome to channel */ 220 channel = e7xxx_find_channel(syndrome); 221 edac_mc_handle_error(HW_EVENT_ERR_CORRECTED, mci, 1, page, 0, syndrome, 222 row, channel, -1, "e7xxx CE", ""); 223 } 224 225 static void process_ce_no_info(struct mem_ctl_info *mci) 226 { 227 edac_dbg(3, "\n"); 228 edac_mc_handle_error(HW_EVENT_ERR_CORRECTED, mci, 1, 0, 0, 0, -1, -1, -1, 229 "e7xxx CE log register overflow", ""); 230 } 231 232 static void process_ue(struct mem_ctl_info *mci, struct e7xxx_error_info *info) 233 { 234 u32 error_2b, block_page; 235 int row; 236 237 edac_dbg(3, "\n"); 238 /* read the error address */ 239 error_2b = info->dram_uelog_add; 240 /* FIXME - should use PAGE_SHIFT */ 241 block_page = error_2b >> 6; /* convert to 4k address */ 242 row = edac_mc_find_csrow_by_page(mci, block_page); 243 244 edac_mc_handle_error(HW_EVENT_ERR_UNCORRECTED, mci, 1, block_page, 0, 0, 245 row, -1, -1, "e7xxx UE", ""); 246 } 247 248 static void process_ue_no_info(struct mem_ctl_info *mci) 249 { 250 edac_dbg(3, "\n"); 251 252 edac_mc_handle_error(HW_EVENT_ERR_UNCORRECTED, mci, 1, 0, 0, 0, -1, -1, -1, 253 "e7xxx UE log register overflow", ""); 254 } 255 256 static void e7xxx_get_error_info(struct mem_ctl_info *mci, 257 struct e7xxx_error_info *info) 258 { 259 struct e7xxx_pvt *pvt; 260 261 pvt = (struct e7xxx_pvt *)mci->pvt_info; 262 pci_read_config_byte(pvt->bridge_ck, E7XXX_DRAM_FERR, &info->dram_ferr); 263 pci_read_config_byte(pvt->bridge_ck, E7XXX_DRAM_NERR, &info->dram_nerr); 264 265 if ((info->dram_ferr & 1) || (info->dram_nerr & 1)) { 266 pci_read_config_dword(pvt->bridge_ck, E7XXX_DRAM_CELOG_ADD, 267 &info->dram_celog_add); 268 pci_read_config_word(pvt->bridge_ck, 269 E7XXX_DRAM_CELOG_SYNDROME, 270 &info->dram_celog_syndrome); 271 } 272 273 if ((info->dram_ferr & 2) || (info->dram_nerr & 2)) 274 pci_read_config_dword(pvt->bridge_ck, E7XXX_DRAM_UELOG_ADD, 275 &info->dram_uelog_add); 276 277 if (info->dram_ferr & 3) 278 pci_write_bits8(pvt->bridge_ck, E7XXX_DRAM_FERR, 0x03, 0x03); 279 280 if (info->dram_nerr & 3) 281 pci_write_bits8(pvt->bridge_ck, E7XXX_DRAM_NERR, 0x03, 0x03); 282 } 283 284 static int e7xxx_process_error_info(struct mem_ctl_info *mci, 285 struct e7xxx_error_info *info, 286 int handle_errors) 287 { 288 int error_found; 289 290 error_found = 0; 291 292 /* decode and report errors */ 293 if (info->dram_ferr & 1) { /* check first error correctable */ 294 error_found = 1; 295 296 if (handle_errors) 297 process_ce(mci, info); 298 } 299 300 if (info->dram_ferr & 2) { /* check first error uncorrectable */ 301 error_found = 1; 302 303 if (handle_errors) 304 process_ue(mci, info); 305 } 306 307 if (info->dram_nerr & 1) { /* check next error correctable */ 308 error_found = 1; 309 310 if (handle_errors) { 311 if (info->dram_ferr & 1) 312 process_ce_no_info(mci); 313 else 314 process_ce(mci, info); 315 } 316 } 317 318 if (info->dram_nerr & 2) { /* check next error uncorrectable */ 319 error_found = 1; 320 321 if (handle_errors) { 322 if (info->dram_ferr & 2) 323 process_ue_no_info(mci); 324 else 325 process_ue(mci, info); 326 } 327 } 328 329 return error_found; 330 } 331 332 static void e7xxx_check(struct mem_ctl_info *mci) 333 { 334 struct e7xxx_error_info info; 335 336 e7xxx_get_error_info(mci, &info); 337 e7xxx_process_error_info(mci, &info, 1); 338 } 339 340 /* Return 1 if dual channel mode is active. Else return 0. */ 341 static inline int dual_channel_active(u32 drc, int dev_idx) 342 { 343 return (dev_idx == E7501) ? ((drc >> 22) & 0x1) : 1; 344 } 345 346 /* Return DRB granularity (0=32mb, 1=64mb). */ 347 static inline int drb_granularity(u32 drc, int dev_idx) 348 { 349 /* only e7501 can be single channel */ 350 return (dev_idx == E7501) ? ((drc >> 18) & 0x3) : 1; 351 } 352 353 static void e7xxx_init_csrows(struct mem_ctl_info *mci, struct pci_dev *pdev, 354 int dev_idx, u32 drc) 355 { 356 unsigned long last_cumul_size; 357 int index, j; 358 u8 value; 359 u32 dra, cumul_size, nr_pages; 360 int drc_chan, drc_drbg, drc_ddim, mem_dev; 361 struct csrow_info *csrow; 362 struct dimm_info *dimm; 363 enum edac_type edac_mode; 364 365 pci_read_config_dword(pdev, E7XXX_DRA, &dra); 366 drc_chan = dual_channel_active(drc, dev_idx); 367 drc_drbg = drb_granularity(drc, dev_idx); 368 drc_ddim = (drc >> 20) & 0x3; 369 last_cumul_size = 0; 370 371 /* The dram row boundary (DRB) reg values are boundary address 372 * for each DRAM row with a granularity of 32 or 64MB (single/dual 373 * channel operation). DRB regs are cumulative; therefore DRB7 will 374 * contain the total memory contained in all eight rows. 375 */ 376 for (index = 0; index < mci->nr_csrows; index++) { 377 /* mem_dev 0=x8, 1=x4 */ 378 mem_dev = (dra >> (index * 4 + 3)) & 0x1; 379 csrow = mci->csrows[index]; 380 381 pci_read_config_byte(pdev, E7XXX_DRB + index, &value); 382 /* convert a 64 or 32 MiB DRB to a page size. */ 383 cumul_size = value << (25 + drc_drbg - PAGE_SHIFT); 384 edac_dbg(3, "(%d) cumul_size 0x%x\n", index, cumul_size); 385 if (cumul_size == last_cumul_size) 386 continue; /* not populated */ 387 388 csrow->first_page = last_cumul_size; 389 csrow->last_page = cumul_size - 1; 390 nr_pages = cumul_size - last_cumul_size; 391 last_cumul_size = cumul_size; 392 393 /* 394 * if single channel or x8 devices then SECDED 395 * if dual channel and x4 then S4ECD4ED 396 */ 397 if (drc_ddim) { 398 if (drc_chan && mem_dev) { 399 edac_mode = EDAC_S4ECD4ED; 400 mci->edac_cap |= EDAC_FLAG_S4ECD4ED; 401 } else { 402 edac_mode = EDAC_SECDED; 403 mci->edac_cap |= EDAC_FLAG_SECDED; 404 } 405 } else 406 edac_mode = EDAC_NONE; 407 408 for (j = 0; j < drc_chan + 1; j++) { 409 dimm = csrow->channels[j]->dimm; 410 411 dimm->nr_pages = nr_pages / (drc_chan + 1); 412 dimm->grain = 1 << 12; /* 4KiB - resolution of CELOG */ 413 dimm->mtype = MEM_RDDR; /* only one type supported */ 414 dimm->dtype = mem_dev ? DEV_X4 : DEV_X8; 415 dimm->edac_mode = edac_mode; 416 } 417 } 418 } 419 420 static int e7xxx_probe1(struct pci_dev *pdev, int dev_idx) 421 { 422 u16 pci_data; 423 struct mem_ctl_info *mci = NULL; 424 struct edac_mc_layer layers[2]; 425 struct e7xxx_pvt *pvt = NULL; 426 u32 drc; 427 int drc_chan; 428 struct e7xxx_error_info discard; 429 430 edac_dbg(0, "mci\n"); 431 432 pci_read_config_dword(pdev, E7XXX_DRC, &drc); 433 434 drc_chan = dual_channel_active(drc, dev_idx); 435 /* 436 * According with the datasheet, this device has a maximum of 437 * 4 DIMMS per channel, either single-rank or dual-rank. So, the 438 * total amount of dimms is 8 (E7XXX_NR_DIMMS). 439 * That means that the DIMM is mapped as CSROWs, and the channel 440 * will map the rank. So, an error to either channel should be 441 * attributed to the same dimm. 442 */ 443 layers[0].type = EDAC_MC_LAYER_CHIP_SELECT; 444 layers[0].size = E7XXX_NR_CSROWS; 445 layers[0].is_virt_csrow = true; 446 layers[1].type = EDAC_MC_LAYER_CHANNEL; 447 layers[1].size = drc_chan + 1; 448 layers[1].is_virt_csrow = false; 449 mci = edac_mc_alloc(0, ARRAY_SIZE(layers), layers, sizeof(*pvt)); 450 if (mci == NULL) 451 return -ENOMEM; 452 453 edac_dbg(3, "init mci\n"); 454 mci->mtype_cap = MEM_FLAG_RDDR; 455 mci->edac_ctl_cap = EDAC_FLAG_NONE | EDAC_FLAG_SECDED | 456 EDAC_FLAG_S4ECD4ED; 457 /* FIXME - what if different memory types are in different csrows? */ 458 mci->mod_name = EDAC_MOD_STR; 459 mci->pdev = &pdev->dev; 460 edac_dbg(3, "init pvt\n"); 461 pvt = (struct e7xxx_pvt *)mci->pvt_info; 462 pvt->dev_info = &e7xxx_devs[dev_idx]; 463 pvt->bridge_ck = pci_get_device(PCI_VENDOR_ID_INTEL, 464 pvt->dev_info->err_dev, pvt->bridge_ck); 465 466 if (!pvt->bridge_ck) { 467 e7xxx_printk(KERN_ERR, "error reporting device not found:" 468 "vendor %x device 0x%x (broken BIOS?)\n", 469 PCI_VENDOR_ID_INTEL, e7xxx_devs[dev_idx].err_dev); 470 goto fail0; 471 } 472 473 edac_dbg(3, "more mci init\n"); 474 mci->ctl_name = pvt->dev_info->ctl_name; 475 mci->dev_name = pci_name(pdev); 476 mci->edac_check = e7xxx_check; 477 mci->ctl_page_to_phys = ctl_page_to_phys; 478 e7xxx_init_csrows(mci, pdev, dev_idx, drc); 479 mci->edac_cap |= EDAC_FLAG_NONE; 480 edac_dbg(3, "tolm, remapbase, remaplimit\n"); 481 /* load the top of low memory, remap base, and remap limit vars */ 482 pci_read_config_word(pdev, E7XXX_TOLM, &pci_data); 483 pvt->tolm = ((u32) pci_data) << 4; 484 pci_read_config_word(pdev, E7XXX_REMAPBASE, &pci_data); 485 pvt->remapbase = ((u32) pci_data) << 14; 486 pci_read_config_word(pdev, E7XXX_REMAPLIMIT, &pci_data); 487 pvt->remaplimit = ((u32) pci_data) << 14; 488 e7xxx_printk(KERN_INFO, 489 "tolm = %x, remapbase = %x, remaplimit = %x\n", pvt->tolm, 490 pvt->remapbase, pvt->remaplimit); 491 492 /* clear any pending errors, or initial state bits */ 493 e7xxx_get_error_info(mci, &discard); 494 495 /* Here we assume that we will never see multiple instances of this 496 * type of memory controller. The ID is therefore hardcoded to 0. 497 */ 498 if (edac_mc_add_mc(mci)) { 499 edac_dbg(3, "failed edac_mc_add_mc()\n"); 500 goto fail1; 501 } 502 503 /* allocating generic PCI control info */ 504 e7xxx_pci = edac_pci_create_generic_ctl(&pdev->dev, EDAC_MOD_STR); 505 if (!e7xxx_pci) { 506 printk(KERN_WARNING 507 "%s(): Unable to create PCI control\n", 508 __func__); 509 printk(KERN_WARNING 510 "%s(): PCI error report via EDAC not setup\n", 511 __func__); 512 } 513 514 /* get this far and it's successful */ 515 edac_dbg(3, "success\n"); 516 return 0; 517 518 fail1: 519 pci_dev_put(pvt->bridge_ck); 520 521 fail0: 522 edac_mc_free(mci); 523 524 return -ENODEV; 525 } 526 527 /* returns count (>= 0), or negative on error */ 528 static int e7xxx_init_one(struct pci_dev *pdev, const struct pci_device_id *ent) 529 { 530 edac_dbg(0, "\n"); 531 532 /* wake up and enable device */ 533 return pci_enable_device(pdev) ? 534 -EIO : e7xxx_probe1(pdev, ent->driver_data); 535 } 536 537 static void e7xxx_remove_one(struct pci_dev *pdev) 538 { 539 struct mem_ctl_info *mci; 540 struct e7xxx_pvt *pvt; 541 542 edac_dbg(0, "\n"); 543 544 if (e7xxx_pci) 545 edac_pci_release_generic_ctl(e7xxx_pci); 546 547 if ((mci = edac_mc_del_mc(&pdev->dev)) == NULL) 548 return; 549 550 pvt = (struct e7xxx_pvt *)mci->pvt_info; 551 pci_dev_put(pvt->bridge_ck); 552 edac_mc_free(mci); 553 } 554 555 static const struct pci_device_id e7xxx_pci_tbl[] = { 556 { 557 PCI_VEND_DEV(INTEL, 7205_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0, 558 E7205}, 559 { 560 PCI_VEND_DEV(INTEL, 7500_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0, 561 E7500}, 562 { 563 PCI_VEND_DEV(INTEL, 7501_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0, 564 E7501}, 565 { 566 PCI_VEND_DEV(INTEL, 7505_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0, 567 E7505}, 568 { 569 0, 570 } /* 0 terminated list. */ 571 }; 572 573 MODULE_DEVICE_TABLE(pci, e7xxx_pci_tbl); 574 575 static struct pci_driver e7xxx_driver = { 576 .name = EDAC_MOD_STR, 577 .probe = e7xxx_init_one, 578 .remove = e7xxx_remove_one, 579 .id_table = e7xxx_pci_tbl, 580 }; 581 582 static int __init e7xxx_init(void) 583 { 584 /* Ensure that the OPSTATE is set correctly for POLL or NMI */ 585 opstate_init(); 586 587 return pci_register_driver(&e7xxx_driver); 588 } 589 590 static void __exit e7xxx_exit(void) 591 { 592 pci_unregister_driver(&e7xxx_driver); 593 } 594 595 module_init(e7xxx_init); 596 module_exit(e7xxx_exit); 597 598 MODULE_LICENSE("GPL"); 599 MODULE_AUTHOR("Linux Networx (http://lnxi.com) Thayne Harbaugh et al"); 600 MODULE_DESCRIPTION("MC support for Intel e7xxx memory controllers"); 601 module_param(edac_op_state, int, 0444); 602 MODULE_PARM_DESC(edac_op_state, "EDAC Error Reporting state: 0=Poll,1=NMI"); 603