1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 /* 22 * Copyright 2009 Sun Microsystems, Inc. All rights reserved. 23 * Use is subject to license terms. 24 */ 25 26 /* 27 * Driver to retire/unretire L2/L3 cachelines on panther 28 */ 29 #include <sys/types.h> 30 #include <sys/types32.h> 31 #include <sys/time.h> 32 #include <sys/errno.h> 33 #include <sys/cmn_err.h> 34 #include <sys/param.h> 35 #include <sys/modctl.h> 36 #include <sys/conf.h> 37 #include <sys/open.h> 38 #include <sys/stat.h> 39 #include <sys/ddi.h> 40 #include <sys/sunddi.h> 41 #include <sys/file.h> 42 #include <sys/cpuvar.h> 43 #include <sys/x_call.h> 44 #include <sys/cheetahregs.h> 45 #include <sys/mem_cache.h> 46 #include <sys/mem_cache_ioctl.h> 47 48 extern int retire_l2(uint64_t, uint64_t); 49 extern int retire_l2_alternate(uint64_t, uint64_t); 50 extern int unretire_l2(uint64_t, uint64_t); 51 extern int unretire_l2_alternate(uint64_t, uint64_t); 52 extern int retire_l3(uint64_t, uint64_t); 53 extern int retire_l3_alternate(uint64_t, uint64_t); 54 extern int unretire_l3(uint64_t, uint64_t); 55 extern int unretire_l3_alternate(uint64_t, uint64_t); 56 57 extern void retire_l2_start(uint64_t, uint64_t); 58 extern void retire_l2_end(uint64_t, uint64_t); 59 extern void unretire_l2_start(uint64_t, uint64_t); 60 extern void unretire_l2_end(uint64_t, uint64_t); 61 extern void retire_l3_start(uint64_t, uint64_t); 62 extern void retire_l3_end(uint64_t, uint64_t); 63 extern void unretire_l3_start(uint64_t, uint64_t); 64 extern void unretire_l3_end(uint64_t, uint64_t); 65 66 extern void get_ecache_dtags_tl1(uint64_t, ch_cpu_logout_t *); 67 extern void get_l2_tag_tl1(uint64_t, uint64_t); 68 extern void get_l3_tag_tl1(uint64_t, uint64_t); 69 extern const int _ncpu; 70 71 /* Macro for putting 64-bit onto stack as two 32-bit ints */ 72 #define PRTF_64_TO_32(x) (uint32_t)((x)>>32), (uint32_t)(x) 73 74 75 uint_t l2_flush_retries_done = 0; 76 int mem_cache_debug = 0x0; 77 uint64_t pattern = 0; 78 uint32_t retire_failures = 0; 79 #ifdef DEBUG 80 int inject_anonymous_tag_error = 0; 81 int32_t last_error_injected_way = 0; 82 uint8_t last_error_injected_bit = 0; 83 int32_t last_l3tag_error_injected_way; 84 uint8_t last_l3tag_error_injected_bit; 85 int32_t last_l2tag_error_injected_way; 86 uint8_t last_l2tag_error_injected_bit; 87 #endif 88 89 /* dev_ops and cb_ops entry point function declarations */ 90 static int mem_cache_attach(dev_info_t *, ddi_attach_cmd_t); 91 static int mem_cache_detach(dev_info_t *, ddi_detach_cmd_t); 92 static int mem_cache_getinfo(dev_info_t *, ddi_info_cmd_t, void *, 93 void **); 94 static int mem_cache_open(dev_t *, int, int, cred_t *); 95 static int mem_cache_close(dev_t, int, int, cred_t *); 96 static int mem_cache_ioctl_ops(int, int, cache_info_t *); 97 static int mem_cache_ioctl(dev_t, int, intptr_t, int, cred_t *, int *); 98 99 struct cb_ops mem_cache_cb_ops = { 100 mem_cache_open, 101 mem_cache_close, 102 nodev, 103 nodev, 104 nodev, /* dump */ 105 nodev, 106 nodev, 107 mem_cache_ioctl, 108 nodev, /* devmap */ 109 nodev, 110 ddi_segmap, /* segmap */ 111 nochpoll, 112 ddi_prop_op, 113 NULL, /* for STREAMS drivers */ 114 D_NEW | D_MP /* driver compatibility flag */ 115 }; 116 117 static struct dev_ops mem_cache_dev_ops = { 118 DEVO_REV, /* driver build version */ 119 0, /* device reference count */ 120 mem_cache_getinfo, 121 nulldev, 122 nulldev, /* probe */ 123 mem_cache_attach, 124 mem_cache_detach, 125 nulldev, /* reset */ 126 &mem_cache_cb_ops, 127 (struct bus_ops *)NULL, 128 nulldev, /* power */ 129 ddi_quiesce_not_needed, /* quiesce */ 130 }; 131 132 /* 133 * Soft state 134 */ 135 struct mem_cache_softc { 136 dev_info_t *dip; 137 kmutex_t mutex; 138 }; 139 #define getsoftc(inst) ((struct mem_cache_softc *)ddi_get_soft_state(statep,\ 140 (inst))) 141 142 /* module configuration stuff */ 143 static void *statep; 144 extern struct mod_ops mod_driverops; 145 146 static struct modldrv modldrv = { 147 &mod_driverops, 148 "mem_cache_driver (08/01/30) ", 149 &mem_cache_dev_ops 150 }; 151 152 static struct modlinkage modlinkage = { 153 MODREV_1, 154 &modldrv, 155 0 156 }; 157 158 extern const int _ncpu; /* Pull the kernel's global _ncpu definition */ 159 160 int 161 _init(void) 162 { 163 int e; 164 165 if (e = ddi_soft_state_init(&statep, sizeof (struct mem_cache_softc), 166 MAX_MEM_CACHE_INSTANCES)) { 167 return (e); 168 } 169 170 if ((e = mod_install(&modlinkage)) != 0) 171 ddi_soft_state_fini(&statep); 172 173 return (e); 174 } 175 176 int 177 _fini(void) 178 { 179 int e; 180 181 if ((e = mod_remove(&modlinkage)) != 0) 182 return (e); 183 184 ddi_soft_state_fini(&statep); 185 186 return (DDI_SUCCESS); 187 } 188 189 int 190 _info(struct modinfo *modinfop) 191 { 192 return (mod_info(&modlinkage, modinfop)); 193 } 194 195 /*ARGSUSED*/ 196 static int 197 mem_cache_getinfo(dev_info_t *dip, ddi_info_cmd_t cmd, void *arg, void **result) 198 { 199 int inst; 200 int retval = DDI_SUCCESS; 201 struct mem_cache_softc *softc; 202 203 inst = getminor((dev_t)arg); 204 205 switch (cmd) { 206 case DDI_INFO_DEVT2DEVINFO: 207 if ((softc = getsoftc(inst)) == NULL) { 208 *result = (void *)NULL; 209 retval = DDI_FAILURE; 210 } else 211 *result = (void *)softc->dip; 212 break; 213 214 case DDI_INFO_DEVT2INSTANCE: 215 *result = (void *)((uintptr_t)inst); 216 break; 217 218 default: 219 retval = DDI_FAILURE; 220 } 221 222 return (retval); 223 } 224 225 static int 226 mem_cache_attach(dev_info_t *dip, ddi_attach_cmd_t cmd) 227 { 228 int inst; 229 struct mem_cache_softc *softc = NULL; 230 char name[80]; 231 232 switch (cmd) { 233 case DDI_ATTACH: 234 inst = ddi_get_instance(dip); 235 if (inst >= MAX_MEM_CACHE_INSTANCES) { 236 cmn_err(CE_WARN, "attach failed, too many instances\n"); 237 return (DDI_FAILURE); 238 } 239 (void) sprintf(name, MEM_CACHE_DRIVER_NAME"%d", inst); 240 if (ddi_create_priv_minor_node(dip, name, 241 S_IFCHR, 242 inst, 243 DDI_PSEUDO, 244 0, NULL, "all", 0640) == 245 DDI_FAILURE) { 246 ddi_remove_minor_node(dip, NULL); 247 return (DDI_FAILURE); 248 } 249 250 /* Allocate a soft state structure for this instance */ 251 if (ddi_soft_state_zalloc(statep, inst) != DDI_SUCCESS) { 252 cmn_err(CE_WARN, " ddi_soft_state_zalloc() failed " 253 "for inst %d\n", inst); 254 goto attach_failed; 255 } 256 257 /* Setup soft state */ 258 softc = getsoftc(inst); 259 softc->dip = dip; 260 mutex_init(&softc->mutex, NULL, MUTEX_DRIVER, NULL); 261 262 /* Create main environmental node */ 263 ddi_report_dev(dip); 264 265 return (DDI_SUCCESS); 266 267 case DDI_RESUME: 268 return (DDI_SUCCESS); 269 270 default: 271 return (DDI_FAILURE); 272 } 273 274 attach_failed: 275 276 /* Free soft state, if allocated. remove minor node if added earlier */ 277 if (softc) 278 ddi_soft_state_free(statep, inst); 279 280 ddi_remove_minor_node(dip, NULL); 281 282 return (DDI_FAILURE); 283 } 284 285 static int 286 mem_cache_detach(dev_info_t *dip, ddi_detach_cmd_t cmd) 287 { 288 int inst; 289 struct mem_cache_softc *softc; 290 291 switch (cmd) { 292 case DDI_DETACH: 293 inst = ddi_get_instance(dip); 294 if ((softc = getsoftc(inst)) == NULL) 295 return (ENXIO); 296 297 /* Free the soft state and remove minor node added earlier */ 298 mutex_destroy(&softc->mutex); 299 ddi_soft_state_free(statep, inst); 300 ddi_remove_minor_node(dip, NULL); 301 return (DDI_SUCCESS); 302 303 case DDI_SUSPEND: 304 return (DDI_SUCCESS); 305 306 default: 307 return (DDI_FAILURE); 308 } 309 } 310 311 /*ARGSUSED*/ 312 static int 313 mem_cache_open(dev_t *devp, int flag, int otyp, cred_t *credp) 314 { 315 int inst = getminor(*devp); 316 317 return (getsoftc(inst) == NULL ? ENXIO : 0); 318 } 319 320 /*ARGSUSED*/ 321 static int 322 mem_cache_close(dev_t dev, int flag, int otyp, cred_t *credp) 323 { 324 int inst = getminor(dev); 325 326 return (getsoftc(inst) == NULL ? ENXIO : 0); 327 } 328 329 static char *tstate_to_desc[] = { 330 "Invalid", /* 0 */ 331 "Shared", /* 1 */ 332 "Exclusive", /* 2 */ 333 "Owner", /* 3 */ 334 "Modified", /* 4 */ 335 "NA", /* 5 */ 336 "Owner/Shared", /* 6 */ 337 "Reserved(7)", /* 7 */ 338 }; 339 340 static char * 341 tag_state_to_desc(uint8_t tagstate) 342 { 343 return (tstate_to_desc[tagstate & CH_ECSTATE_MASK]); 344 } 345 346 void 347 print_l2_tag(uint64_t tag_addr, uint64_t l2_tag) 348 { 349 uint64_t l2_subaddr; 350 uint8_t l2_state; 351 352 l2_subaddr = PN_L2TAG_TO_PA(l2_tag); 353 l2_subaddr |= (tag_addr & PN_L2_INDEX_MASK); 354 355 l2_state = (l2_tag & CH_ECSTATE_MASK); 356 cmn_err(CE_CONT, 357 "PA=0x%08x.%08x E$tag 0x%08x.%08x E$state %s\n", 358 PRTF_64_TO_32(l2_subaddr), 359 PRTF_64_TO_32(l2_tag), 360 tag_state_to_desc(l2_state)); 361 } 362 363 void 364 print_l2cache_line(ch_cpu_logout_t *clop) 365 { 366 uint64_t l2_subaddr; 367 int i, offset; 368 uint8_t way, l2_state; 369 ch_ec_data_t *ecp; 370 371 372 for (way = 0; way < PN_CACHE_NWAYS; way++) { 373 ecp = &clop->clo_data.chd_l2_data[way]; 374 l2_subaddr = PN_L2TAG_TO_PA(ecp->ec_tag); 375 l2_subaddr |= (ecp->ec_idx & PN_L2_INDEX_MASK); 376 377 l2_state = (ecp->ec_tag & CH_ECSTATE_MASK); 378 cmn_err(CE_CONT, 379 "\nWAY = %d index = 0x%08x PA=0x%08x.%08x\n" 380 "E$tag 0x%08x.%08x E$state %s", 381 way, (uint32_t)ecp->ec_idx, PRTF_64_TO_32(l2_subaddr), 382 PRTF_64_TO_32(ecp->ec_tag), 383 tag_state_to_desc(l2_state)); 384 /* 385 * Dump out Ecache subblock data captured. 386 * For Cheetah, we need to compute the ECC for each 16-byte 387 * chunk and compare it with the captured chunk ECC to figure 388 * out which chunk is bad. 389 */ 390 for (i = 0; i < (CH_ECACHE_SUBBLK_SIZE/16); i++) { 391 ec_data_elm_t *ecdptr; 392 uint64_t d_low, d_high; 393 uint32_t ecc; 394 int l2_data_idx = (i/2); 395 396 offset = i * 16; 397 ecdptr = &clop->clo_data.chd_l2_data[way].ec_data 398 [l2_data_idx]; 399 if ((i & 1) == 0) { 400 ecc = (ecdptr->ec_eccd >> 9) & 0x1ff; 401 d_high = ecdptr->ec_d8[0]; 402 d_low = ecdptr->ec_d8[1]; 403 } else { 404 ecc = ecdptr->ec_eccd & 0x1ff; 405 d_high = ecdptr->ec_d8[2]; 406 d_low = ecdptr->ec_d8[3]; 407 } 408 409 cmn_err(CE_CONT, 410 "\nE$Data (0x%02x) 0x%08x.%08x 0x%08x.%08x" 411 " ECC 0x%03x", 412 offset, PRTF_64_TO_32(d_high), 413 PRTF_64_TO_32(d_low), ecc); 414 } 415 } /* end of for way loop */ 416 } 417 418 void 419 print_ecache_line(ch_cpu_logout_t *clop) 420 { 421 uint64_t ec_subaddr; 422 int i, offset; 423 uint8_t way, ec_state; 424 ch_ec_data_t *ecp; 425 426 427 for (way = 0; way < PN_CACHE_NWAYS; way++) { 428 ecp = &clop->clo_data.chd_ec_data[way]; 429 ec_subaddr = PN_L3TAG_TO_PA(ecp->ec_tag); 430 ec_subaddr |= (ecp->ec_idx & PN_L3_TAG_RD_MASK); 431 432 ec_state = (ecp->ec_tag & CH_ECSTATE_MASK); 433 cmn_err(CE_CONT, 434 "\nWAY = %d index = 0x%08x PA=0x%08x.%08x\n" 435 "E$tag 0x%08x.%08x E$state %s", 436 way, (uint32_t)ecp->ec_idx, PRTF_64_TO_32(ec_subaddr), 437 PRTF_64_TO_32(ecp->ec_tag), 438 tag_state_to_desc(ec_state)); 439 /* 440 * Dump out Ecache subblock data captured. 441 * For Cheetah, we need to compute the ECC for each 16-byte 442 * chunk and compare it with the captured chunk ECC to figure 443 * out which chunk is bad. 444 */ 445 for (i = 0; i < (CH_ECACHE_SUBBLK_SIZE/16); i++) { 446 ec_data_elm_t *ecdptr; 447 uint64_t d_low, d_high; 448 uint32_t ecc; 449 int ec_data_idx = (i/2); 450 451 offset = i * 16; 452 ecdptr = 453 &clop->clo_data.chd_ec_data[way].ec_data 454 [ec_data_idx]; 455 if ((i & 1) == 0) { 456 ecc = (ecdptr->ec_eccd >> 9) & 0x1ff; 457 d_high = ecdptr->ec_d8[0]; 458 d_low = ecdptr->ec_d8[1]; 459 } else { 460 ecc = ecdptr->ec_eccd & 0x1ff; 461 d_high = ecdptr->ec_d8[2]; 462 d_low = ecdptr->ec_d8[3]; 463 } 464 465 cmn_err(CE_CONT, 466 "\nE$Data (0x%02x) 0x%08x.%08x 0x%08x.%08x" 467 " ECC 0x%03x", 468 offset, PRTF_64_TO_32(d_high), 469 PRTF_64_TO_32(d_low), ecc); 470 } 471 } 472 } 473 474 static boolean_t 475 tag_addr_collides(uint64_t tag_addr, cache_id_t type, 476 retire_func_t start_of_func, retire_func_t end_of_func) 477 { 478 uint64_t start_paddr, end_paddr; 479 char *type_str; 480 481 start_paddr = va_to_pa((void *)start_of_func); 482 end_paddr = va_to_pa((void *)end_of_func); 483 switch (type) { 484 case L2_CACHE_TAG: 485 case L2_CACHE_DATA: 486 tag_addr &= PN_L2_INDEX_MASK; 487 start_paddr &= PN_L2_INDEX_MASK; 488 end_paddr &= PN_L2_INDEX_MASK; 489 type_str = "L2:"; 490 break; 491 case L3_CACHE_TAG: 492 case L3_CACHE_DATA: 493 tag_addr &= PN_L3_TAG_RD_MASK; 494 start_paddr &= PN_L3_TAG_RD_MASK; 495 end_paddr &= PN_L3_TAG_RD_MASK; 496 type_str = "L3:"; 497 break; 498 default: 499 /* 500 * Should never reach here. 501 */ 502 ASSERT(0); 503 return (B_FALSE); 504 } 505 if ((tag_addr > (start_paddr - 0x100)) && 506 (tag_addr < (end_paddr + 0x100))) { 507 if (mem_cache_debug & 0x1) 508 cmn_err(CE_CONT, 509 "%s collision detected tag_addr = 0x%08x" 510 " start_paddr = 0x%08x end_paddr = 0x%08x\n", 511 type_str, (uint32_t)tag_addr, (uint32_t)start_paddr, 512 (uint32_t)end_paddr); 513 return (B_TRUE); 514 } 515 else 516 return (B_FALSE); 517 } 518 519 static uint64_t 520 get_tag_addr(cache_info_t *cache_info) 521 { 522 uint64_t tag_addr, scratch; 523 524 switch (cache_info->cache) { 525 case L2_CACHE_TAG: 526 case L2_CACHE_DATA: 527 tag_addr = (uint64_t)(cache_info->index << 528 PN_CACHE_LINE_SHIFT); 529 scratch = (uint64_t)(cache_info->way << 530 PN_L2_WAY_SHIFT); 531 tag_addr |= scratch; 532 tag_addr |= PN_L2_IDX_HW_ECC_EN; 533 break; 534 case L3_CACHE_TAG: 535 case L3_CACHE_DATA: 536 tag_addr = (uint64_t)(cache_info->index << 537 PN_CACHE_LINE_SHIFT); 538 scratch = (uint64_t)(cache_info->way << 539 PN_L3_WAY_SHIFT); 540 tag_addr |= scratch; 541 tag_addr |= PN_L3_IDX_HW_ECC_EN; 542 break; 543 default: 544 /* 545 * Should never reach here. 546 */ 547 ASSERT(0); 548 return (uint64_t)(0); 549 } 550 return (tag_addr); 551 } 552 553 static int 554 mem_cache_ioctl_ops(int cmd, int mode, cache_info_t *cache_info) 555 { 556 int ret_val = 0; 557 uint64_t afar, tag_addr; 558 ch_cpu_logout_t clop; 559 uint64_t Lxcache_tag_data[PN_CACHE_NWAYS]; 560 int i, retire_retry_count; 561 cpu_t *cpu; 562 uint64_t tag_data; 563 uint8_t state; 564 565 if (cache_info->way >= PN_CACHE_NWAYS) 566 return (EINVAL); 567 switch (cache_info->cache) { 568 case L2_CACHE_TAG: 569 case L2_CACHE_DATA: 570 if (cache_info->index >= 571 (PN_L2_SET_SIZE/PN_L2_LINESIZE)) 572 return (EINVAL); 573 break; 574 case L3_CACHE_TAG: 575 case L3_CACHE_DATA: 576 if (cache_info->index >= 577 (PN_L3_SET_SIZE/PN_L3_LINESIZE)) 578 return (EINVAL); 579 break; 580 default: 581 return (ENOTSUP); 582 } 583 /* 584 * Check if we have a valid cpu ID and that 585 * CPU is ONLINE. 586 */ 587 mutex_enter(&cpu_lock); 588 cpu = cpu_get(cache_info->cpu_id); 589 if ((cpu == NULL) || (!cpu_is_online(cpu))) { 590 mutex_exit(&cpu_lock); 591 return (EINVAL); 592 } 593 mutex_exit(&cpu_lock); 594 pattern = 0; /* default value of TAG PA when cacheline is retired. */ 595 switch (cmd) { 596 case MEM_CACHE_RETIRE: 597 tag_addr = get_tag_addr(cache_info); 598 pattern |= PN_ECSTATE_NA; 599 retire_retry_count = 0; 600 affinity_set(cache_info->cpu_id); 601 switch (cache_info->cache) { 602 case L2_CACHE_DATA: 603 case L2_CACHE_TAG: 604 if ((cache_info->bit & MSB_BIT_MASK) == 605 MSB_BIT_MASK) 606 pattern |= PN_L2TAG_PA_MASK; 607 retry_l2_retire: 608 if (tag_addr_collides(tag_addr, 609 cache_info->cache, 610 retire_l2_start, retire_l2_end)) 611 ret_val = 612 retire_l2_alternate( 613 tag_addr, pattern); 614 else 615 ret_val = retire_l2(tag_addr, 616 pattern); 617 if (ret_val == 1) { 618 /* 619 * cacheline was in retired 620 * STATE already. 621 * so return success. 622 */ 623 ret_val = 0; 624 } 625 if (ret_val < 0) { 626 cmn_err(CE_WARN, 627 "retire_l2() failed. index = 0x%x way %d. Retrying...\n", 628 cache_info->index, 629 cache_info->way); 630 if (retire_retry_count >= 2) { 631 retire_failures++; 632 affinity_clear(); 633 return (EIO); 634 } 635 retire_retry_count++; 636 goto retry_l2_retire; 637 } 638 if (ret_val == 2) 639 l2_flush_retries_done++; 640 /* 641 * We bind ourself to a CPU and send cross trap to 642 * ourself. On return from xt_one we can rely on the 643 * data in tag_data being filled in. Normally one would 644 * do a xt_sync to make sure that the CPU has completed 645 * the cross trap call xt_one. 646 */ 647 xt_one(cache_info->cpu_id, 648 (xcfunc_t *)(get_l2_tag_tl1), 649 tag_addr, (uint64_t)(&tag_data)); 650 state = tag_data & CH_ECSTATE_MASK; 651 if (state != PN_ECSTATE_NA) { 652 retire_failures++; 653 print_l2_tag(tag_addr, 654 tag_data); 655 cmn_err(CE_WARN, 656 "L2 RETIRE:failed for index 0x%x way %d. Retrying...\n", 657 cache_info->index, 658 cache_info->way); 659 if (retire_retry_count >= 2) { 660 retire_failures++; 661 affinity_clear(); 662 return (EIO); 663 } 664 retire_retry_count++; 665 goto retry_l2_retire; 666 } 667 break; 668 case L3_CACHE_TAG: 669 case L3_CACHE_DATA: 670 if ((cache_info->bit & MSB_BIT_MASK) == 671 MSB_BIT_MASK) 672 pattern |= PN_L3TAG_PA_MASK; 673 if (tag_addr_collides(tag_addr, 674 cache_info->cache, 675 retire_l3_start, retire_l3_end)) 676 ret_val = 677 retire_l3_alternate( 678 tag_addr, pattern); 679 else 680 ret_val = retire_l3(tag_addr, 681 pattern); 682 if (ret_val == 1) { 683 /* 684 * cacheline was in retired 685 * STATE already. 686 * so return success. 687 */ 688 ret_val = 0; 689 } 690 if (ret_val < 0) { 691 cmn_err(CE_WARN, 692 "retire_l3() failed. ret_val = %d index = 0x%x\n", 693 ret_val, 694 cache_info->index); 695 retire_failures++; 696 affinity_clear(); 697 return (EIO); 698 } 699 /* 700 * We bind ourself to a CPU and send cross trap to 701 * ourself. On return from xt_one we can rely on the 702 * data in tag_data being filled in. Normally one would 703 * do a xt_sync to make sure that the CPU has completed 704 * the cross trap call xt_one. 705 */ 706 xt_one(cache_info->cpu_id, 707 (xcfunc_t *)(get_l3_tag_tl1), 708 tag_addr, (uint64_t)(&tag_data)); 709 state = tag_data & CH_ECSTATE_MASK; 710 if (state != PN_ECSTATE_NA) { 711 cmn_err(CE_WARN, 712 "L3 RETIRE failed for index 0x%x\n", 713 cache_info->index); 714 retire_failures++; 715 affinity_clear(); 716 return (EIO); 717 } 718 719 break; 720 } 721 affinity_clear(); 722 break; 723 case MEM_CACHE_UNRETIRE: 724 tag_addr = get_tag_addr(cache_info); 725 pattern = PN_ECSTATE_INV; 726 affinity_set(cache_info->cpu_id); 727 switch (cache_info->cache) { 728 case L2_CACHE_DATA: 729 case L2_CACHE_TAG: 730 /* 731 * We bind ourself to a CPU and send cross trap to 732 * ourself. On return from xt_one we can rely on the 733 * data in tag_data being filled in. Normally one would 734 * do a xt_sync to make sure that the CPU has completed 735 * the cross trap call xt_one. 736 */ 737 xt_one(cache_info->cpu_id, 738 (xcfunc_t *)(get_l2_tag_tl1), 739 tag_addr, (uint64_t)(&tag_data)); 740 state = tag_data & CH_ECSTATE_MASK; 741 if (state != PN_ECSTATE_NA) { 742 affinity_clear(); 743 return (EINVAL); 744 } 745 if (tag_addr_collides(tag_addr, 746 cache_info->cache, 747 unretire_l2_start, unretire_l2_end)) 748 ret_val = 749 unretire_l2_alternate( 750 tag_addr, pattern); 751 else 752 ret_val = 753 unretire_l2(tag_addr, 754 pattern); 755 if (ret_val != 0) { 756 cmn_err(CE_WARN, 757 "unretire_l2() failed. ret_val = %d index = 0x%x\n", 758 ret_val, 759 cache_info->index); 760 retire_failures++; 761 affinity_clear(); 762 return (EIO); 763 } 764 break; 765 case L3_CACHE_TAG: 766 case L3_CACHE_DATA: 767 /* 768 * We bind ourself to a CPU and send cross trap to 769 * ourself. On return from xt_one we can rely on the 770 * data in tag_data being filled in. Normally one would 771 * do a xt_sync to make sure that the CPU has completed 772 * the cross trap call xt_one. 773 */ 774 xt_one(cache_info->cpu_id, 775 (xcfunc_t *)(get_l3_tag_tl1), 776 tag_addr, (uint64_t)(&tag_data)); 777 state = tag_data & CH_ECSTATE_MASK; 778 if (state != PN_ECSTATE_NA) { 779 affinity_clear(); 780 return (EINVAL); 781 } 782 if (tag_addr_collides(tag_addr, 783 cache_info->cache, 784 unretire_l3_start, unretire_l3_end)) 785 ret_val = 786 unretire_l3_alternate( 787 tag_addr, pattern); 788 else 789 ret_val = 790 unretire_l3(tag_addr, 791 pattern); 792 if (ret_val != 0) { 793 cmn_err(CE_WARN, 794 "unretire_l3() failed. ret_val = %d index = 0x%x\n", 795 ret_val, 796 cache_info->index); 797 affinity_clear(); 798 return (EIO); 799 } 800 break; 801 } 802 affinity_clear(); 803 break; 804 case MEM_CACHE_ISRETIRED: 805 case MEM_CACHE_STATE: 806 return (ENOTSUP); 807 case MEM_CACHE_READ_TAGS: 808 #ifdef DEBUG 809 case MEM_CACHE_READ_ERROR_INJECTED_TAGS: 810 #endif 811 /* 812 * Read tag and data for all the ways at a given afar 813 */ 814 afar = (uint64_t)(cache_info->index 815 << PN_CACHE_LINE_SHIFT); 816 mutex_enter(&cpu_lock); 817 affinity_set(cache_info->cpu_id); 818 (void) pause_cpus(NULL, NULL); 819 mutex_exit(&cpu_lock); 820 /* 821 * We bind ourself to a CPU and send cross trap to 822 * ourself. On return from xt_one we can rely on the 823 * data in clop being filled in. Normally one would 824 * do a xt_sync to make sure that the CPU has completed 825 * the cross trap call xt_one. 826 */ 827 xt_one(cache_info->cpu_id, 828 (xcfunc_t *)(get_ecache_dtags_tl1), 829 afar, (uint64_t)(&clop)); 830 mutex_enter(&cpu_lock); 831 (void) start_cpus(); 832 mutex_exit(&cpu_lock); 833 affinity_clear(); 834 switch (cache_info->cache) { 835 case L2_CACHE_TAG: 836 for (i = 0; i < PN_CACHE_NWAYS; i++) { 837 Lxcache_tag_data[i] = 838 clop.clo_data.chd_l2_data 839 [i].ec_tag; 840 } 841 #ifdef DEBUG 842 last_error_injected_bit = 843 last_l2tag_error_injected_bit; 844 last_error_injected_way = 845 last_l2tag_error_injected_way; 846 #endif 847 break; 848 case L3_CACHE_TAG: 849 for (i = 0; i < PN_CACHE_NWAYS; i++) { 850 Lxcache_tag_data[i] = 851 clop.clo_data.chd_ec_data 852 [i].ec_tag; 853 } 854 #ifdef DEBUG 855 last_error_injected_bit = 856 last_l3tag_error_injected_bit; 857 last_error_injected_way = 858 last_l3tag_error_injected_way; 859 #endif 860 break; 861 default: 862 return (ENOTSUP); 863 } /* end if switch(cache) */ 864 #ifdef DEBUG 865 if ((cmd == MEM_CACHE_READ_ERROR_INJECTED_TAGS) && 866 (inject_anonymous_tag_error == 0) && 867 (last_error_injected_way >= 0) && 868 (last_error_injected_way <= 3)) { 869 pattern = ((uint64_t)1 << 870 last_error_injected_bit); 871 /* 872 * If error bit is ECC we need to make sure 873 * ECC on all all WAYS are corrupted. 874 */ 875 if ((last_error_injected_bit >= 6) && 876 (last_error_injected_bit <= 14)) { 877 for (i = 0; i < PN_CACHE_NWAYS; i++) 878 Lxcache_tag_data[i] ^= 879 pattern; 880 } else 881 Lxcache_tag_data 882 [last_error_injected_way] ^= 883 pattern; 884 } 885 #endif 886 if (ddi_copyout((caddr_t)Lxcache_tag_data, 887 (caddr_t)cache_info->datap, 888 sizeof (Lxcache_tag_data), mode) 889 != DDI_SUCCESS) { 890 return (EFAULT); 891 } 892 break; /* end of READ_TAGS */ 893 default: 894 return (ENOTSUP); 895 } /* end if switch(cmd) */ 896 return (ret_val); 897 } 898 899 /*ARGSUSED*/ 900 static int 901 mem_cache_ioctl(dev_t dev, int cmd, intptr_t arg, int mode, cred_t *credp, 902 int *rvalp) 903 { 904 int inst; 905 struct mem_cache_softc *softc; 906 cache_info_t cache_info; 907 cache_info32_t cache_info32; 908 int ret_val; 909 int is_panther; 910 911 inst = getminor(dev); 912 if ((softc = getsoftc(inst)) == NULL) 913 return (ENXIO); 914 915 mutex_enter(&softc->mutex); 916 917 #ifdef _MULTI_DATAMODEL 918 if (ddi_model_convert_from(mode & FMODELS) == DDI_MODEL_ILP32) { 919 if (ddi_copyin((cache_info32_t *)arg, &cache_info32, 920 sizeof (cache_info32), mode) != DDI_SUCCESS) { 921 mutex_exit(&softc->mutex); 922 return (EFAULT); 923 } 924 cache_info.cache = cache_info32.cache; 925 cache_info.index = cache_info32.index; 926 cache_info.way = cache_info32.way; 927 cache_info.cpu_id = cache_info32.cpu_id; 928 cache_info.bit = cache_info32.bit; 929 cache_info.datap = (void *)((uint64_t)cache_info32.datap); 930 } else 931 #endif 932 if (ddi_copyin((cache_info_t *)arg, &cache_info, 933 sizeof (cache_info), mode) != DDI_SUCCESS) { 934 mutex_exit(&softc->mutex); 935 return (EFAULT); 936 } 937 938 if ((cache_info.cpu_id < 0) || (cache_info.cpu_id >= _ncpu)) { 939 mutex_exit(&softc->mutex); 940 return (EINVAL); 941 } 942 is_panther = IS_PANTHER(cpunodes[cache_info.cpu_id].implementation); 943 if (!is_panther) { 944 mutex_exit(&softc->mutex); 945 return (ENOTSUP); 946 } 947 switch (cmd) { 948 case MEM_CACHE_RETIRE: 949 case MEM_CACHE_UNRETIRE: 950 if ((mode & FWRITE) == 0) { 951 ret_val = EBADF; 952 break; 953 } 954 /*FALLTHROUGH*/ 955 case MEM_CACHE_ISRETIRED: 956 case MEM_CACHE_STATE: 957 case MEM_CACHE_READ_TAGS: 958 #ifdef DEBUG 959 case MEM_CACHE_READ_ERROR_INJECTED_TAGS: 960 #endif 961 ret_val = mem_cache_ioctl_ops(cmd, mode, &cache_info); 962 break; 963 default: 964 ret_val = ENOTSUP; 965 break; 966 } 967 mutex_exit(&softc->mutex); 968 return (ret_val); 969 }