1 /* 2 * Emulated CXL Switch Upstream Port 3 * 4 * Copyright (c) 2022 Huawei Technologies. 5 * 6 * Based on xio3130_upstream.c 7 * 8 * SPDX-License-Identifier: GPL-2.0-or-later 9 */ 10 11 #include "qemu/osdep.h" 12 #include "qemu/log.h" 13 #include "hw/qdev-properties.h" 14 #include "hw/pci/msi.h" 15 #include "hw/pci/pcie.h" 16 #include "hw/pci/pcie_port.h" 17 #include "hw/pci-bridge/cxl_upstream_port.h" 18 /* 19 * Null value of all Fs suggested by IEEE RA guidelines for use of 20 * EU, OUI and CID 21 */ 22 #define UI64_NULL (~0ULL) 23 24 #define CXL_UPSTREAM_PORT_MSI_NR_VECTOR 2 25 26 #define CXL_UPSTREAM_PORT_MSI_OFFSET 0x70 27 #define CXL_UPSTREAM_PORT_PCIE_CAP_OFFSET 0x90 28 #define CXL_UPSTREAM_PORT_AER_OFFSET 0x100 29 #define CXL_UPSTREAM_PORT_SN_OFFSET \ 30 (CXL_UPSTREAM_PORT_AER_OFFSET + PCI_ERR_SIZEOF) 31 #define CXL_UPSTREAM_PORT_DVSEC_OFFSET \ 32 (CXL_UPSTREAM_PORT_SN_OFFSET + PCI_EXT_CAP_DSN_SIZEOF) 33 34 CXLComponentState *cxl_usp_to_cstate(CXLUpstreamPort *usp) 35 { 36 return &usp->cxl_cstate; 37 } 38 39 static void cxl_usp_dvsec_write_config(PCIDevice *dev, uint32_t addr, 40 uint32_t val, int len) 41 { 42 CXLUpstreamPort *usp = CXL_USP(dev); 43 44 if (range_contains(&usp->cxl_cstate.dvsecs[EXTENSIONS_PORT_DVSEC], addr)) { 45 uint8_t *reg = &dev->config[addr]; 46 addr -= usp->cxl_cstate.dvsecs[EXTENSIONS_PORT_DVSEC].lob; 47 if (addr == PORT_CONTROL_OFFSET) { 48 if (pci_get_word(reg) & PORT_CONTROL_UNMASK_SBR) { 49 /* unmask SBR */ 50 qemu_log_mask(LOG_UNIMP, "SBR mask control is not supported\n"); 51 } 52 if (pci_get_word(reg) & PORT_CONTROL_ALT_MEMID_EN) { 53 /* Alt Memory & ID Space Enable */ 54 qemu_log_mask(LOG_UNIMP, 55 "Alt Memory & ID space is not supported\n"); 56 } 57 } 58 } 59 } 60 61 static void cxl_usp_write_config(PCIDevice *d, uint32_t address, 62 uint32_t val, int len) 63 { 64 CXLUpstreamPort *usp = CXL_USP(d); 65 66 pcie_doe_write_config(&usp->doe_cdat, address, val, len); 67 pci_bridge_write_config(d, address, val, len); 68 pcie_cap_flr_write_config(d, address, val, len); 69 pcie_aer_write_config(d, address, val, len); 70 71 cxl_usp_dvsec_write_config(d, address, val, len); 72 } 73 74 static uint32_t cxl_usp_read_config(PCIDevice *d, uint32_t address, int len) 75 { 76 CXLUpstreamPort *usp = CXL_USP(d); 77 uint32_t val; 78 79 if (pcie_doe_read_config(&usp->doe_cdat, address, len, &val)) { 80 return val; 81 } 82 83 return pci_default_read_config(d, address, len); 84 } 85 86 static void latch_registers(CXLUpstreamPort *usp) 87 { 88 uint32_t *reg_state = usp->cxl_cstate.crb.cache_mem_registers; 89 uint32_t *write_msk = usp->cxl_cstate.crb.cache_mem_regs_write_mask; 90 91 cxl_component_register_init_common(reg_state, write_msk, 92 CXL2_UPSTREAM_PORT); 93 ARRAY_FIELD_DP32(reg_state, CXL_HDM_DECODER_CAPABILITY, TARGET_COUNT, 8); 94 } 95 96 static void cxl_usp_reset(DeviceState *qdev) 97 { 98 PCIDevice *d = PCI_DEVICE(qdev); 99 CXLUpstreamPort *usp = CXL_USP(qdev); 100 101 pci_bridge_reset(qdev); 102 pcie_cap_deverr_reset(d); 103 latch_registers(usp); 104 } 105 106 static void build_dvsecs(CXLComponentState *cxl) 107 { 108 uint8_t *dvsec; 109 110 dvsec = (uint8_t *)&(CXLDVSECPortExt){ 111 .status = 0x1, /* Port Power Management Init Complete */ 112 }; 113 cxl_component_create_dvsec(cxl, CXL2_UPSTREAM_PORT, 114 EXTENSIONS_PORT_DVSEC_LENGTH, 115 EXTENSIONS_PORT_DVSEC, 116 EXTENSIONS_PORT_DVSEC_REVID, dvsec); 117 dvsec = (uint8_t *)&(CXLDVSECPortFlexBus){ 118 .cap = 0x27, /* Cache, IO, Mem, non-MLD */ 119 .ctrl = 0x27, /* Cache, IO, Mem */ 120 .status = 0x26, /* same */ 121 .rcvd_mod_ts_data_phase1 = 0xef, /* WTF? */ 122 }; 123 cxl_component_create_dvsec(cxl, CXL2_UPSTREAM_PORT, 124 PCIE_CXL3_FLEXBUS_PORT_DVSEC_LENGTH, 125 PCIE_FLEXBUS_PORT_DVSEC, 126 PCIE_CXL3_FLEXBUS_PORT_DVSEC_REVID, dvsec); 127 128 dvsec = (uint8_t *)&(CXLDVSECRegisterLocator){ 129 .rsvd = 0, 130 .reg0_base_lo = RBI_COMPONENT_REG | CXL_COMPONENT_REG_BAR_IDX, 131 .reg0_base_hi = 0, 132 }; 133 cxl_component_create_dvsec(cxl, CXL2_UPSTREAM_PORT, 134 REG_LOC_DVSEC_LENGTH, REG_LOC_DVSEC, 135 REG_LOC_DVSEC_REVID, dvsec); 136 } 137 138 static bool cxl_doe_cdat_rsp(DOECap *doe_cap) 139 { 140 CDATObject *cdat = &CXL_USP(doe_cap->pdev)->cxl_cstate.cdat; 141 uint16_t ent; 142 void *base; 143 uint32_t len; 144 CDATReq *req = pcie_doe_get_write_mbox_ptr(doe_cap); 145 CDATRsp rsp; 146 147 cxl_doe_cdat_update(&CXL_USP(doe_cap->pdev)->cxl_cstate, &error_fatal); 148 assert(cdat->entry_len); 149 150 /* Discard if request length mismatched */ 151 if (pcie_doe_get_obj_len(req) < 152 DIV_ROUND_UP(sizeof(CDATReq), sizeof(uint32_t))) { 153 return false; 154 } 155 156 ent = req->entry_handle; 157 base = cdat->entry[ent].base; 158 len = cdat->entry[ent].length; 159 160 rsp = (CDATRsp) { 161 .header = { 162 .vendor_id = CXL_VENDOR_ID, 163 .data_obj_type = CXL_DOE_TABLE_ACCESS, 164 .reserved = 0x0, 165 .length = DIV_ROUND_UP((sizeof(rsp) + len), sizeof(uint32_t)), 166 }, 167 .rsp_code = CXL_DOE_TAB_RSP, 168 .table_type = CXL_DOE_TAB_TYPE_CDAT, 169 .entry_handle = (ent < cdat->entry_len - 1) ? 170 ent + 1 : CXL_DOE_TAB_ENT_MAX, 171 }; 172 173 memcpy(doe_cap->read_mbox, &rsp, sizeof(rsp)); 174 memcpy(doe_cap->read_mbox + DIV_ROUND_UP(sizeof(rsp), sizeof(uint32_t)), 175 base, len); 176 177 doe_cap->read_mbox_len += rsp.header.length; 178 179 return true; 180 } 181 182 static DOEProtocol doe_cdat_prot[] = { 183 { CXL_VENDOR_ID, CXL_DOE_TABLE_ACCESS, cxl_doe_cdat_rsp }, 184 { } 185 }; 186 187 enum { 188 CXL_USP_CDAT_SSLBIS_LAT, 189 CXL_USP_CDAT_SSLBIS_BW, 190 CXL_USP_CDAT_NUM_ENTRIES 191 }; 192 193 static int build_cdat_table(CDATSubHeader ***cdat_table, void *priv) 194 { 195 CDATSslbis *sslbis_latency; 196 CDATSslbis *sslbis_bandwidth; 197 CXLUpstreamPort *us = CXL_USP(priv); 198 PCIBus *bus = &PCI_BRIDGE(us)->sec_bus; 199 int devfn, sslbis_size, i; 200 int count = 0; 201 uint16_t port_ids[256]; 202 203 for (devfn = 0; devfn < ARRAY_SIZE(bus->devices); devfn++) { 204 PCIDevice *d = bus->devices[devfn]; 205 PCIEPort *port; 206 207 if (!d || !pci_is_express(d) || !d->exp.exp_cap) { 208 continue; 209 } 210 211 /* 212 * Whilst the PCI express spec doesn't allow anything other than 213 * downstream ports on this bus, let us be a little paranoid 214 */ 215 if (!object_dynamic_cast(OBJECT(d), TYPE_PCIE_PORT)) { 216 continue; 217 } 218 219 port = PCIE_PORT(d); 220 port_ids[count] = port->port; 221 count++; 222 } 223 224 /* May not yet have any ports - try again later */ 225 if (count == 0) { 226 return 0; 227 } 228 229 sslbis_size = sizeof(CDATSslbis) + sizeof(*sslbis_latency->sslbe) * count; 230 sslbis_latency = g_malloc(sslbis_size); 231 *sslbis_latency = (CDATSslbis) { 232 .sslbis_header = { 233 .header = { 234 .type = CDAT_TYPE_SSLBIS, 235 .length = sslbis_size, 236 }, 237 .data_type = HMATLB_DATA_TYPE_ACCESS_LATENCY, 238 .entry_base_unit = 10000, 239 }, 240 }; 241 242 for (i = 0; i < count; i++) { 243 sslbis_latency->sslbe[i] = (CDATSslbe) { 244 .port_x_id = CDAT_PORT_ID_USP, 245 .port_y_id = port_ids[i], 246 .latency_bandwidth = 15, /* 150ns */ 247 }; 248 } 249 250 sslbis_bandwidth = g_malloc(sslbis_size); 251 *sslbis_bandwidth = (CDATSslbis) { 252 .sslbis_header = { 253 .header = { 254 .type = CDAT_TYPE_SSLBIS, 255 .length = sslbis_size, 256 }, 257 .data_type = HMATLB_DATA_TYPE_ACCESS_BANDWIDTH, 258 .entry_base_unit = 1024, 259 }, 260 }; 261 262 for (i = 0; i < count; i++) { 263 sslbis_bandwidth->sslbe[i] = (CDATSslbe) { 264 .port_x_id = CDAT_PORT_ID_USP, 265 .port_y_id = port_ids[i], 266 .latency_bandwidth = 16, /* 16 GB/s */ 267 }; 268 } 269 270 *cdat_table = g_new0(CDATSubHeader *, CXL_USP_CDAT_NUM_ENTRIES); 271 272 /* Header always at start of structure */ 273 (*cdat_table)[CXL_USP_CDAT_SSLBIS_LAT] = (CDATSubHeader *)sslbis_latency; 274 (*cdat_table)[CXL_USP_CDAT_SSLBIS_BW] = (CDATSubHeader *)sslbis_bandwidth; 275 276 return CXL_USP_CDAT_NUM_ENTRIES; 277 } 278 279 static void free_default_cdat_table(CDATSubHeader **cdat_table, int num, 280 void *priv) 281 { 282 int i; 283 284 for (i = 0; i < num; i++) { 285 g_free(cdat_table[i]); 286 } 287 g_free(cdat_table); 288 } 289 290 static void cxl_usp_realize(PCIDevice *d, Error **errp) 291 { 292 ERRP_GUARD(); 293 PCIEPort *p = PCIE_PORT(d); 294 CXLUpstreamPort *usp = CXL_USP(d); 295 CXLComponentState *cxl_cstate = &usp->cxl_cstate; 296 ComponentRegisters *cregs = &cxl_cstate->crb; 297 MemoryRegion *component_bar = &cregs->component_registers; 298 int rc; 299 300 pci_bridge_initfn(d, TYPE_PCIE_BUS); 301 pcie_port_init_reg(d); 302 303 rc = msi_init(d, CXL_UPSTREAM_PORT_MSI_OFFSET, 304 CXL_UPSTREAM_PORT_MSI_NR_VECTOR, true, true, errp); 305 if (rc) { 306 assert(rc == -ENOTSUP); 307 goto err_bridge; 308 } 309 310 rc = pcie_cap_init(d, CXL_UPSTREAM_PORT_PCIE_CAP_OFFSET, 311 PCI_EXP_TYPE_UPSTREAM, p->port, errp); 312 if (rc < 0) { 313 goto err_msi; 314 } 315 316 pcie_cap_flr_init(d); 317 pcie_cap_deverr_init(d); 318 rc = pcie_aer_init(d, PCI_ERR_VER, CXL_UPSTREAM_PORT_AER_OFFSET, 319 PCI_ERR_SIZEOF, errp); 320 if (rc) { 321 goto err_cap; 322 } 323 if (usp->sn != UI64_NULL) { 324 pcie_dev_ser_num_init(d, CXL_UPSTREAM_PORT_SN_OFFSET, usp->sn); 325 } 326 cxl_cstate->dvsec_offset = CXL_UPSTREAM_PORT_DVSEC_OFFSET; 327 cxl_cstate->pdev = d; 328 build_dvsecs(cxl_cstate); 329 cxl_component_register_block_init(OBJECT(d), cxl_cstate, TYPE_CXL_USP); 330 pci_register_bar(d, CXL_COMPONENT_REG_BAR_IDX, 331 PCI_BASE_ADDRESS_SPACE_MEMORY | 332 PCI_BASE_ADDRESS_MEM_TYPE_64, 333 component_bar); 334 335 pcie_doe_init(d, &usp->doe_cdat, cxl_cstate->dvsec_offset, doe_cdat_prot, 336 true, 1); 337 338 cxl_cstate->cdat.build_cdat_table = build_cdat_table; 339 cxl_cstate->cdat.free_cdat_table = free_default_cdat_table; 340 cxl_cstate->cdat.private = d; 341 cxl_doe_cdat_init(cxl_cstate, errp); 342 if (*errp) { 343 goto err_cap; 344 } 345 346 return; 347 348 err_cap: 349 pcie_cap_exit(d); 350 err_msi: 351 msi_uninit(d); 352 err_bridge: 353 pci_bridge_exitfn(d); 354 } 355 356 static void cxl_usp_exitfn(PCIDevice *d) 357 { 358 pcie_aer_exit(d); 359 pcie_cap_exit(d); 360 msi_uninit(d); 361 pci_bridge_exitfn(d); 362 } 363 364 static Property cxl_upstream_props[] = { 365 DEFINE_PROP_UINT64("sn", CXLUpstreamPort, sn, UI64_NULL), 366 DEFINE_PROP_STRING("cdat", CXLUpstreamPort, cxl_cstate.cdat.filename), 367 DEFINE_PROP_END_OF_LIST() 368 }; 369 370 static void cxl_upstream_class_init(ObjectClass *oc, void *data) 371 { 372 DeviceClass *dc = DEVICE_CLASS(oc); 373 PCIDeviceClass *k = PCI_DEVICE_CLASS(oc); 374 375 k->config_write = cxl_usp_write_config; 376 k->config_read = cxl_usp_read_config; 377 k->realize = cxl_usp_realize; 378 k->exit = cxl_usp_exitfn; 379 k->vendor_id = 0x19e5; /* Huawei */ 380 k->device_id = 0xa128; /* Emulated CXL Switch Upstream Port */ 381 k->revision = 0; 382 set_bit(DEVICE_CATEGORY_BRIDGE, dc->categories); 383 dc->desc = "CXL Switch Upstream Port"; 384 dc->reset = cxl_usp_reset; 385 device_class_set_props(dc, cxl_upstream_props); 386 } 387 388 static const TypeInfo cxl_usp_info = { 389 .name = TYPE_CXL_USP, 390 .parent = TYPE_PCIE_PORT, 391 .instance_size = sizeof(CXLUpstreamPort), 392 .class_init = cxl_upstream_class_init, 393 .interfaces = (InterfaceInfo[]) { 394 { INTERFACE_PCIE_DEVICE }, 395 { INTERFACE_CXL_DEVICE }, 396 { } 397 }, 398 }; 399 400 static void cxl_usp_register_type(void) 401 { 402 type_register_static(&cxl_usp_info); 403 } 404 405 type_init(cxl_usp_register_type); 406