/* * CXL Utility library for mailbox interface * * Copyright(C) 2020 Intel Corporation. * * This work is licensed under the terms of the GNU GPL, version 2. See the * COPYING file in the top-level directory. */ #include "qemu/osdep.h" #include "hw/pci/msi.h" #include "hw/pci/msix.h" #include "hw/cxl/cxl.h" #include "hw/cxl/cxl_events.h" #include "hw/cxl/cxl_mailbox.h" #include "hw/pci/pci.h" #include "hw/pci-bridge/cxl_upstream_port.h" #include "qemu/cutils.h" #include "qemu/log.h" #include "qemu/units.h" #include "qemu/uuid.h" #include "sysemu/hostmem.h" #include "qemu/range.h" #define CXL_CAPACITY_MULTIPLIER (256 * MiB) #define CXL_DC_EVENT_LOG_SIZE 8 #define CXL_NUM_EXTENTS_SUPPORTED 512 #define CXL_NUM_TAGS_SUPPORTED 0 /* * How to add a new command, example. The command set FOO, with cmd BAR. * 1. Add the command set and cmd to the enum. * FOO = 0x7f, * #define BAR 0 * 2. Implement the handler * static CXLRetCode cmd_foo_bar(struct cxl_cmd *cmd, * CXLDeviceState *cxl_dstate, uint16_t *len) * 3. Add the command to the cxl_cmd_set[][] * [FOO][BAR] = { "FOO_BAR", cmd_foo_bar, x, y }, * 4. Implement your handler * define_mailbox_handler(FOO_BAR) { ... return CXL_MBOX_SUCCESS; } * * * Writing the handler: * The handler will provide the &struct cxl_cmd, the &CXLDeviceState, and the * in/out length of the payload. The handler is responsible for consuming the * payload from cmd->payload and operating upon it as necessary. It must then * fill the output data into cmd->payload (overwriting what was there), * setting the length, and returning a valid return code. * * XXX: The handler need not worry about endianness. The payload is read out of * a register interface that already deals with it. */ enum { INFOSTAT = 0x00, #define IS_IDENTIFY 0x1 #define BACKGROUND_OPERATION_STATUS 0x2 EVENTS = 0x01, #define GET_RECORDS 0x0 #define CLEAR_RECORDS 0x1 #define GET_INTERRUPT_POLICY 0x2 #define SET_INTERRUPT_POLICY 0x3 FIRMWARE_UPDATE = 0x02, #define GET_INFO 0x0 #define TRANSFER 0x1 #define ACTIVATE 0x2 TIMESTAMP = 0x03, #define GET 0x0 #define SET 0x1 LOGS = 0x04, #define GET_SUPPORTED 0x0 #define GET_LOG 0x1 FEATURES = 0x05, #define GET_SUPPORTED 0x0 #define GET_FEATURE 0x1 #define SET_FEATURE 0x2 IDENTIFY = 0x40, #define MEMORY_DEVICE 0x0 CCLS = 0x41, #define GET_PARTITION_INFO 0x0 #define GET_LSA 0x2 #define SET_LSA 0x3 SANITIZE = 0x44, #define OVERWRITE 0x0 #define SECURE_ERASE 0x1 PERSISTENT_MEM = 0x45, #define GET_SECURITY_STATE 0x0 MEDIA_AND_POISON = 0x43, #define GET_POISON_LIST 0x0 #define INJECT_POISON 0x1 #define CLEAR_POISON 0x2 #define GET_SCAN_MEDIA_CAPABILITIES 0x3 #define SCAN_MEDIA 0x4 #define GET_SCAN_MEDIA_RESULTS 0x5 DCD_CONFIG = 0x48, #define GET_DC_CONFIG 0x0 #define GET_DYN_CAP_EXT_LIST 0x1 #define ADD_DYN_CAP_RSP 0x2 #define RELEASE_DYN_CAP 0x3 PHYSICAL_SWITCH = 0x51, #define IDENTIFY_SWITCH_DEVICE 0x0 #define GET_PHYSICAL_PORT_STATE 0x1 TUNNEL = 0x53, #define MANAGEMENT_COMMAND 0x0 }; /* CCI Message Format CXL r3.1 Figure 7-19 */ typedef struct CXLCCIMessage { uint8_t category; #define CXL_CCI_CAT_REQ 0 #define CXL_CCI_CAT_RSP 1 uint8_t tag; uint8_t resv1; uint8_t command; uint8_t command_set; uint8_t pl_length[3]; uint16_t rc; uint16_t vendor_specific; uint8_t payload[]; } QEMU_PACKED CXLCCIMessage; /* This command is only defined to an MLD FM Owned LD or an MHD */ static CXLRetCode cmd_tunnel_management_cmd(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { PCIDevice *tunnel_target; CXLCCI *target_cci; struct { uint8_t port_or_ld_id; uint8_t target_type; uint16_t size; CXLCCIMessage ccimessage; } QEMU_PACKED *in; struct { uint16_t resp_len; uint8_t resv[2]; CXLCCIMessage ccimessage; } QEMU_PACKED *out; size_t pl_length, length_out; bool bg_started; int rc; if (cmd->in < sizeof(*in)) { return CXL_MBOX_INVALID_INPUT; } in = (void *)payload_in; out = (void *)payload_out; /* Enough room for minimum sized message - no payload */ if (in->size < sizeof(in->ccimessage)) { return CXL_MBOX_INVALID_PAYLOAD_LENGTH; } /* Length of input payload should be in->size + a wrapping tunnel header */ if (in->size != len_in - offsetof(typeof(*out), ccimessage)) { return CXL_MBOX_INVALID_PAYLOAD_LENGTH; } if (in->ccimessage.category != CXL_CCI_CAT_REQ) { return CXL_MBOX_INVALID_INPUT; } if (in->target_type != 0) { qemu_log_mask(LOG_UNIMP, "Tunneled Command sent to non existent FM-LD"); return CXL_MBOX_INVALID_INPUT; } /* * Target of a tunnel unfortunately depends on type of CCI readint * the message. * If in a switch, then it's the port number. * If in an MLD it is the ld number. * If in an MHD target type indicate where we are going. */ if (object_dynamic_cast(OBJECT(cci->d), TYPE_CXL_TYPE3)) { CXLType3Dev *ct3d = CXL_TYPE3(cci->d); if (in->port_or_ld_id != 0) { /* Only pretending to have one for now! */ return CXL_MBOX_INVALID_INPUT; } target_cci = &ct3d->ld0_cci; } else if (object_dynamic_cast(OBJECT(cci->d), TYPE_CXL_USP)) { CXLUpstreamPort *usp = CXL_USP(cci->d); tunnel_target = pcie_find_port_by_pn(&PCI_BRIDGE(usp)->sec_bus, in->port_or_ld_id); if (!tunnel_target) { return CXL_MBOX_INVALID_INPUT; } tunnel_target = pci_bridge_get_sec_bus(PCI_BRIDGE(tunnel_target))->devices[0]; if (!tunnel_target) { return CXL_MBOX_INVALID_INPUT; } if (object_dynamic_cast(OBJECT(tunnel_target), TYPE_CXL_TYPE3)) { CXLType3Dev *ct3d = CXL_TYPE3(tunnel_target); /* Tunneled VDMs always land on FM Owned LD */ target_cci = &ct3d->vdm_fm_owned_ld_mctp_cci; } else { return CXL_MBOX_INVALID_INPUT; } } else { return CXL_MBOX_INVALID_INPUT; } pl_length = in->ccimessage.pl_length[2] << 16 | in->ccimessage.pl_length[1] << 8 | in->ccimessage.pl_length[0]; rc = cxl_process_cci_message(target_cci, in->ccimessage.command_set, in->ccimessage.command, pl_length, in->ccimessage.payload, &length_out, out->ccimessage.payload, &bg_started); /* Payload should be in place. Rest of CCI header and needs filling */ out->resp_len = length_out + sizeof(CXLCCIMessage); st24_le_p(out->ccimessage.pl_length, length_out); out->ccimessage.rc = rc; out->ccimessage.category = CXL_CCI_CAT_RSP; out->ccimessage.command = in->ccimessage.command; out->ccimessage.command_set = in->ccimessage.command_set; out->ccimessage.tag = in->ccimessage.tag; *len_out = length_out + sizeof(*out); return CXL_MBOX_SUCCESS; } static CXLRetCode cmd_events_get_records(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { CXLDeviceState *cxlds = &CXL_TYPE3(cci->d)->cxl_dstate; CXLGetEventPayload *pl; uint8_t log_type; int max_recs; if (cmd->in < sizeof(log_type)) { return CXL_MBOX_INVALID_INPUT; } log_type = payload_in[0]; pl = (CXLGetEventPayload *)payload_out; max_recs = (cxlds->payload_size - CXL_EVENT_PAYLOAD_HDR_SIZE) / CXL_EVENT_RECORD_SIZE; if (max_recs > 0xFFFF) { max_recs = 0xFFFF; } return cxl_event_get_records(cxlds, pl, log_type, max_recs, len_out); } static CXLRetCode cmd_events_clear_records(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { CXLDeviceState *cxlds = &CXL_TYPE3(cci->d)->cxl_dstate; CXLClearEventPayload *pl; pl = (CXLClearEventPayload *)payload_in; *len_out = 0; return cxl_event_clear_records(cxlds, pl); } static CXLRetCode cmd_events_get_interrupt_policy(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { CXLDeviceState *cxlds = &CXL_TYPE3(cci->d)->cxl_dstate; CXLEventInterruptPolicy *policy; CXLEventLog *log; policy = (CXLEventInterruptPolicy *)payload_out; log = &cxlds->event_logs[CXL_EVENT_TYPE_INFO]; if (log->irq_enabled) { policy->info_settings = CXL_EVENT_INT_SETTING(log->irq_vec); } log = &cxlds->event_logs[CXL_EVENT_TYPE_WARN]; if (log->irq_enabled) { policy->warn_settings = CXL_EVENT_INT_SETTING(log->irq_vec); } log = &cxlds->event_logs[CXL_EVENT_TYPE_FAIL]; if (log->irq_enabled) { policy->failure_settings = CXL_EVENT_INT_SETTING(log->irq_vec); } log = &cxlds->event_logs[CXL_EVENT_TYPE_FATAL]; if (log->irq_enabled) { policy->fatal_settings = CXL_EVENT_INT_SETTING(log->irq_vec); } log = &cxlds->event_logs[CXL_EVENT_TYPE_DYNAMIC_CAP]; if (log->irq_enabled) { /* Dynamic Capacity borrows the same vector as info */ policy->dyn_cap_settings = CXL_INT_MSI_MSIX; } *len_out = sizeof(*policy); return CXL_MBOX_SUCCESS; } static CXLRetCode cmd_events_set_interrupt_policy(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { CXLDeviceState *cxlds = &CXL_TYPE3(cci->d)->cxl_dstate; CXLEventInterruptPolicy *policy; CXLEventLog *log; if (len_in < CXL_EVENT_INT_SETTING_MIN_LEN) { return CXL_MBOX_INVALID_PAYLOAD_LENGTH; } policy = (CXLEventInterruptPolicy *)payload_in; log = &cxlds->event_logs[CXL_EVENT_TYPE_INFO]; log->irq_enabled = (policy->info_settings & CXL_EVENT_INT_MODE_MASK) == CXL_INT_MSI_MSIX; log = &cxlds->event_logs[CXL_EVENT_TYPE_WARN]; log->irq_enabled = (policy->warn_settings & CXL_EVENT_INT_MODE_MASK) == CXL_INT_MSI_MSIX; log = &cxlds->event_logs[CXL_EVENT_TYPE_FAIL]; log->irq_enabled = (policy->failure_settings & CXL_EVENT_INT_MODE_MASK) == CXL_INT_MSI_MSIX; log = &cxlds->event_logs[CXL_EVENT_TYPE_FATAL]; log->irq_enabled = (policy->fatal_settings & CXL_EVENT_INT_MODE_MASK) == CXL_INT_MSI_MSIX; /* DCD is optional */ if (len_in < sizeof(*policy)) { return CXL_MBOX_SUCCESS; } log = &cxlds->event_logs[CXL_EVENT_TYPE_DYNAMIC_CAP]; log->irq_enabled = (policy->dyn_cap_settings & CXL_EVENT_INT_MODE_MASK) == CXL_INT_MSI_MSIX; *len_out = 0; return CXL_MBOX_SUCCESS; } /* CXL r3.1 section 8.2.9.1.1: Identify (Opcode 0001h) */ static CXLRetCode cmd_infostat_identify(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { PCIDeviceClass *class = PCI_DEVICE_GET_CLASS(cci->d); struct { uint16_t pcie_vid; uint16_t pcie_did; uint16_t pcie_subsys_vid; uint16_t pcie_subsys_id; uint64_t sn; uint8_t max_message_size; uint8_t component_type; } QEMU_PACKED *is_identify; QEMU_BUILD_BUG_ON(sizeof(*is_identify) != 18); is_identify = (void *)payload_out; is_identify->pcie_vid = class->vendor_id; is_identify->pcie_did = class->device_id; if (object_dynamic_cast(OBJECT(cci->d), TYPE_CXL_USP)) { is_identify->sn = CXL_USP(cci->d)->sn; /* Subsystem info not defined for a USP */ is_identify->pcie_subsys_vid = 0; is_identify->pcie_subsys_id = 0; is_identify->component_type = 0x0; /* Switch */ } else if (object_dynamic_cast(OBJECT(cci->d), TYPE_CXL_TYPE3)) { PCIDevice *pci_dev = PCI_DEVICE(cci->d); is_identify->sn = CXL_TYPE3(cci->d)->sn; /* * We can't always use class->subsystem_vendor_id as * it is not set if the defaults are used. */ is_identify->pcie_subsys_vid = pci_get_word(pci_dev->config + PCI_SUBSYSTEM_VENDOR_ID); is_identify->pcie_subsys_id = pci_get_word(pci_dev->config + PCI_SUBSYSTEM_ID); is_identify->component_type = 0x3; /* Type 3 */ } /* TODO: Allow this to vary across different CCIs */ is_identify->max_message_size = 9; /* 512 bytes - MCTP_CXL_MAILBOX_BYTES */ *len_out = sizeof(*is_identify); return CXL_MBOX_SUCCESS; } static void cxl_set_dsp_active_bm(PCIBus *b, PCIDevice *d, void *private) { uint8_t *bm = private; if (object_dynamic_cast(OBJECT(d), TYPE_CXL_DSP)) { uint8_t port = PCIE_PORT(d)->port; bm[port / 8] |= 1 << (port % 8); } } /* CXL r3.1 Section 7.6.7.1.1: Identify Switch Device (Opcode 5100h) */ static CXLRetCode cmd_identify_switch_device(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { PCIEPort *usp = PCIE_PORT(cci->d); PCIBus *bus = &PCI_BRIDGE(cci->d)->sec_bus; int num_phys_ports = pcie_count_ds_ports(bus); struct cxl_fmapi_ident_switch_dev_resp_pl { uint8_t ingress_port_id; uint8_t rsvd; uint8_t num_physical_ports; uint8_t num_vcss; uint8_t active_port_bitmask[0x20]; uint8_t active_vcs_bitmask[0x20]; uint16_t total_vppbs; uint16_t bound_vppbs; uint8_t num_hdm_decoders_per_usp; } QEMU_PACKED *out; QEMU_BUILD_BUG_ON(sizeof(*out) != 0x49); out = (struct cxl_fmapi_ident_switch_dev_resp_pl *)payload_out; *out = (struct cxl_fmapi_ident_switch_dev_resp_pl) { .num_physical_ports = num_phys_ports + 1, /* 1 USP */ .num_vcss = 1, /* Not yet support multiple VCS - potentially tricky */ .active_vcs_bitmask[0] = 0x1, .total_vppbs = num_phys_ports + 1, .bound_vppbs = num_phys_ports + 1, .num_hdm_decoders_per_usp = 4, }; /* Depends on the CCI type */ if (object_dynamic_cast(OBJECT(cci->intf), TYPE_PCIE_PORT)) { out->ingress_port_id = PCIE_PORT(cci->intf)->port; } else { /* MCTP? */ out->ingress_port_id = 0; } pci_for_each_device_under_bus(bus, cxl_set_dsp_active_bm, out->active_port_bitmask); out->active_port_bitmask[usp->port / 8] |= (1 << usp->port % 8); *len_out = sizeof(*out); return CXL_MBOX_SUCCESS; } /* CXL r3.1 Section 7.6.7.1.2: Get Physical Port State (Opcode 5101h) */ static CXLRetCode cmd_get_physical_port_state(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { /* CXL r3.1 Table 7-17: Get Physical Port State Request Payload */ struct cxl_fmapi_get_phys_port_state_req_pl { uint8_t num_ports; uint8_t ports[]; } QEMU_PACKED *in; /* * CXL r3.1 Table 7-19: Get Physical Port State Port Information Block * Format */ struct cxl_fmapi_port_state_info_block { uint8_t port_id; uint8_t config_state; uint8_t connected_device_cxl_version; uint8_t rsv1; uint8_t connected_device_type; uint8_t port_cxl_version_bitmask; uint8_t max_link_width; uint8_t negotiated_link_width; uint8_t supported_link_speeds_vector; uint8_t max_link_speed; uint8_t current_link_speed; uint8_t ltssm_state; uint8_t first_lane_num; uint16_t link_state; uint8_t supported_ld_count; } QEMU_PACKED; /* CXL r3.1 Table 7-18: Get Physical Port State Response Payload */ struct cxl_fmapi_get_phys_port_state_resp_pl { uint8_t num_ports; uint8_t rsv1[3]; struct cxl_fmapi_port_state_info_block ports[]; } QEMU_PACKED *out; PCIBus *bus = &PCI_BRIDGE(cci->d)->sec_bus; PCIEPort *usp = PCIE_PORT(cci->d); size_t pl_size; int i; in = (struct cxl_fmapi_get_phys_port_state_req_pl *)payload_in; out = (struct cxl_fmapi_get_phys_port_state_resp_pl *)payload_out; /* Check if what was requested can fit */ if (sizeof(*out) + sizeof(*out->ports) * in->num_ports > cci->payload_max) { return CXL_MBOX_INVALID_INPUT; } /* For success there should be a match for each requested */ out->num_ports = in->num_ports; for (i = 0; i < in->num_ports; i++) { struct cxl_fmapi_port_state_info_block *port; /* First try to match on downstream port */ PCIDevice *port_dev; uint16_t lnkcap, lnkcap2, lnksta; port = &out->ports[i]; port_dev = pcie_find_port_by_pn(bus, in->ports[i]); if (port_dev) { /* DSP */ PCIDevice *ds_dev = pci_bridge_get_sec_bus(PCI_BRIDGE(port_dev)) ->devices[0]; port->config_state = 3; if (ds_dev) { if (object_dynamic_cast(OBJECT(ds_dev), TYPE_CXL_TYPE3)) { port->connected_device_type = 5; /* Assume MLD for now */ } else { port->connected_device_type = 1; } } else { port->connected_device_type = 0; } port->supported_ld_count = 3; } else if (usp->port == in->ports[i]) { /* USP */ port_dev = PCI_DEVICE(usp); port->config_state = 4; port->connected_device_type = 0; } else { return CXL_MBOX_INVALID_INPUT; } port->port_id = in->ports[i]; /* Information on status of this port in lnksta, lnkcap */ if (!port_dev->exp.exp_cap) { return CXL_MBOX_INTERNAL_ERROR; } lnksta = port_dev->config_read(port_dev, port_dev->exp.exp_cap + PCI_EXP_LNKSTA, sizeof(lnksta)); lnkcap = port_dev->config_read(port_dev, port_dev->exp.exp_cap + PCI_EXP_LNKCAP, sizeof(lnkcap)); lnkcap2 = port_dev->config_read(port_dev, port_dev->exp.exp_cap + PCI_EXP_LNKCAP2, sizeof(lnkcap2)); port->max_link_width = (lnkcap & PCI_EXP_LNKCAP_MLW) >> 4; port->negotiated_link_width = (lnksta & PCI_EXP_LNKSTA_NLW) >> 4; /* No definition for SLS field in linux/pci_regs.h */ port->supported_link_speeds_vector = (lnkcap2 & 0xFE) >> 1; port->max_link_speed = lnkcap & PCI_EXP_LNKCAP_SLS; port->current_link_speed = lnksta & PCI_EXP_LNKSTA_CLS; /* TODO: Track down if we can get the rest of the info */ port->ltssm_state = 0x7; port->first_lane_num = 0; port->link_state = 0; port->port_cxl_version_bitmask = 0x2; port->connected_device_cxl_version = 0x2; } pl_size = sizeof(*out) + sizeof(*out->ports) * in->num_ports; *len_out = pl_size; return CXL_MBOX_SUCCESS; } /* CXL r3.1 Section 8.2.9.1.2: Background Operation Status (Opcode 0002h) */ static CXLRetCode cmd_infostat_bg_op_sts(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { struct { uint8_t status; uint8_t rsvd; uint16_t opcode; uint16_t returncode; uint16_t vendor_ext_status; } QEMU_PACKED *bg_op_status; QEMU_BUILD_BUG_ON(sizeof(*bg_op_status) != 8); bg_op_status = (void *)payload_out; bg_op_status->status = cci->bg.complete_pct << 1; if (cci->bg.runtime > 0) { bg_op_status->status |= 1U << 0; } bg_op_status->opcode = cci->bg.opcode; bg_op_status->returncode = cci->bg.ret_code; *len_out = sizeof(*bg_op_status); return CXL_MBOX_SUCCESS; } #define CXL_FW_SLOTS 2 #define CXL_FW_SIZE 0x02000000 /* 32 mb */ /* CXL r3.1 Section 8.2.9.3.1: Get FW Info (Opcode 0200h) */ static CXLRetCode cmd_firmware_update_get_info(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { CXLType3Dev *ct3d = CXL_TYPE3(cci->d); CXLDeviceState *cxl_dstate = &ct3d->cxl_dstate; struct { uint8_t slots_supported; uint8_t slot_info; uint8_t caps; uint8_t rsvd[0xd]; char fw_rev1[0x10]; char fw_rev2[0x10]; char fw_rev3[0x10]; char fw_rev4[0x10]; } QEMU_PACKED *fw_info; QEMU_BUILD_BUG_ON(sizeof(*fw_info) != 0x50); if ((cxl_dstate->vmem_size < CXL_CAPACITY_MULTIPLIER) || (cxl_dstate->pmem_size < CXL_CAPACITY_MULTIPLIER) || (ct3d->dc.total_capacity < CXL_CAPACITY_MULTIPLIER)) { return CXL_MBOX_INTERNAL_ERROR; } fw_info = (void *)payload_out; fw_info->slots_supported = CXL_FW_SLOTS; fw_info->slot_info = (cci->fw.active_slot & 0x7) | ((cci->fw.staged_slot & 0x7) << 3); fw_info->caps = BIT(0); /* online update supported */ if (cci->fw.slot[0]) { pstrcpy(fw_info->fw_rev1, sizeof(fw_info->fw_rev1), "BWFW VERSION 0"); } if (cci->fw.slot[1]) { pstrcpy(fw_info->fw_rev2, sizeof(fw_info->fw_rev2), "BWFW VERSION 1"); } *len_out = sizeof(*fw_info); return CXL_MBOX_SUCCESS; } /* CXL r3.1 section 8.2.9.3.2: Transfer FW (Opcode 0201h) */ #define CXL_FW_XFER_ALIGNMENT 128 #define CXL_FW_XFER_ACTION_FULL 0x0 #define CXL_FW_XFER_ACTION_INIT 0x1 #define CXL_FW_XFER_ACTION_CONTINUE 0x2 #define CXL_FW_XFER_ACTION_END 0x3 #define CXL_FW_XFER_ACTION_ABORT 0x4 static CXLRetCode cmd_firmware_update_transfer(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { struct { uint8_t action; uint8_t slot; uint8_t rsvd1[2]; uint32_t offset; uint8_t rsvd2[0x78]; uint8_t data[]; } QEMU_PACKED *fw_transfer = (void *)payload_in; size_t offset, length; if (fw_transfer->action == CXL_FW_XFER_ACTION_ABORT) { /* * At this point there aren't any on-going transfers * running in the bg - this is serialized before this * call altogether. Just mark the state machine and * disregard any other input. */ cci->fw.transferring = false; return CXL_MBOX_SUCCESS; } offset = fw_transfer->offset * CXL_FW_XFER_ALIGNMENT; length = len - sizeof(*fw_transfer); if (offset + length > CXL_FW_SIZE) { return CXL_MBOX_INVALID_INPUT; } if (cci->fw.transferring) { if (fw_transfer->action == CXL_FW_XFER_ACTION_FULL || fw_transfer->action == CXL_FW_XFER_ACTION_INIT) { return CXL_MBOX_FW_XFER_IN_PROGRESS; } /* * Abort partitioned package transfer if over 30 secs * between parts. As opposed to the explicit ABORT action, * semantically treat this condition as an error - as * if a part action were passed without a previous INIT. */ if (difftime(time(NULL), cci->fw.last_partxfer) > 30.0) { cci->fw.transferring = false; return CXL_MBOX_INVALID_INPUT; } } else if (fw_transfer->action == CXL_FW_XFER_ACTION_CONTINUE || fw_transfer->action == CXL_FW_XFER_ACTION_END) { return CXL_MBOX_INVALID_INPUT; } /* allow back-to-back retransmission */ if ((offset != cci->fw.prev_offset || length != cci->fw.prev_len) && (fw_transfer->action == CXL_FW_XFER_ACTION_CONTINUE || fw_transfer->action == CXL_FW_XFER_ACTION_END)) { /* verify no overlaps */ if (offset < cci->fw.prev_offset + cci->fw.prev_len) { return CXL_MBOX_FW_XFER_OUT_OF_ORDER; } } switch (fw_transfer->action) { case CXL_FW_XFER_ACTION_FULL: /* ignores offset */ case CXL_FW_XFER_ACTION_END: if (fw_transfer->slot == 0 || fw_transfer->slot == cci->fw.active_slot || fw_transfer->slot > CXL_FW_SLOTS) { return CXL_MBOX_FW_INVALID_SLOT; } /* mark the slot used upon bg completion */ break; case CXL_FW_XFER_ACTION_INIT: if (offset != 0) { return CXL_MBOX_INVALID_INPUT; } cci->fw.transferring = true; cci->fw.prev_offset = offset; cci->fw.prev_len = length; break; case CXL_FW_XFER_ACTION_CONTINUE: cci->fw.prev_offset = offset; cci->fw.prev_len = length; break; default: return CXL_MBOX_INVALID_INPUT; } if (fw_transfer->action == CXL_FW_XFER_ACTION_FULL) { cci->bg.runtime = 10 * 1000UL; } else { cci->bg.runtime = 2 * 1000UL; } /* keep relevant context for bg completion */ cci->fw.curr_action = fw_transfer->action; cci->fw.curr_slot = fw_transfer->slot; *len_out = 0; return CXL_MBOX_BG_STARTED; } static void __do_firmware_xfer(CXLCCI *cci) { switch (cci->fw.curr_action) { case CXL_FW_XFER_ACTION_FULL: case CXL_FW_XFER_ACTION_END: cci->fw.slot[cci->fw.curr_slot - 1] = true; cci->fw.transferring = false; break; case CXL_FW_XFER_ACTION_INIT: case CXL_FW_XFER_ACTION_CONTINUE: time(&cci->fw.last_partxfer); break; default: break; } } /* CXL r3.1 section 8.2.9.3.3: Activate FW (Opcode 0202h) */ static CXLRetCode cmd_firmware_update_activate(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { struct { uint8_t action; uint8_t slot; } QEMU_PACKED *fw_activate = (void *)payload_in; QEMU_BUILD_BUG_ON(sizeof(*fw_activate) != 0x2); if (fw_activate->slot == 0 || fw_activate->slot == cci->fw.active_slot || fw_activate->slot > CXL_FW_SLOTS) { return CXL_MBOX_FW_INVALID_SLOT; } /* ensure that an actual fw package is there */ if (!cci->fw.slot[fw_activate->slot - 1]) { return CXL_MBOX_FW_INVALID_SLOT; } switch (fw_activate->action) { case 0: /* online */ cci->fw.active_slot = fw_activate->slot; break; case 1: /* reset */ cci->fw.staged_slot = fw_activate->slot; break; default: return CXL_MBOX_INVALID_INPUT; } return CXL_MBOX_SUCCESS; } /* CXL r3.1 Section 8.2.9.4.1: Get Timestamp (Opcode 0300h) */ static CXLRetCode cmd_timestamp_get(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { CXLDeviceState *cxl_dstate = &CXL_TYPE3(cci->d)->cxl_dstate; uint64_t final_time = cxl_device_get_timestamp(cxl_dstate); stq_le_p(payload_out, final_time); *len_out = 8; return CXL_MBOX_SUCCESS; } /* CXL r3.1 Section 8.2.9.4.2: Set Timestamp (Opcode 0301h) */ static CXLRetCode cmd_timestamp_set(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { CXLDeviceState *cxl_dstate = &CXL_TYPE3(cci->d)->cxl_dstate; cxl_dstate->timestamp.set = true; cxl_dstate->timestamp.last_set = qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL); cxl_dstate->timestamp.host_set = le64_to_cpu(*(uint64_t *)payload_in); *len_out = 0; return CXL_MBOX_SUCCESS; } /* CXL r3.1 Section 8.2.9.5.2.1: Command Effects Log (CEL) */ static const QemuUUID cel_uuid = { .data = UUID(0x0da9c0b5, 0xbf41, 0x4b78, 0x8f, 0x79, 0x96, 0xb1, 0x62, 0x3b, 0x3f, 0x17) }; /* CXL r3.1 Section 8.2.9.5.1: Get Supported Logs (Opcode 0400h) */ static CXLRetCode cmd_logs_get_supported(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { struct { uint16_t entries; uint8_t rsvd[6]; struct { QemuUUID uuid; uint32_t size; } log_entries[1]; } QEMU_PACKED *supported_logs = (void *)payload_out; QEMU_BUILD_BUG_ON(sizeof(*supported_logs) != 0x1c); supported_logs->entries = 1; supported_logs->log_entries[0].uuid = cel_uuid; supported_logs->log_entries[0].size = 4 * cci->cel_size; *len_out = sizeof(*supported_logs); return CXL_MBOX_SUCCESS; } /* CXL r3.1 Section 8.2.9.5.2: Get Log (Opcode 0401h) */ static CXLRetCode cmd_logs_get_log(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { struct { QemuUUID uuid; uint32_t offset; uint32_t length; } QEMU_PACKED QEMU_ALIGNED(16) *get_log; get_log = (void *)payload_in; /* * CXL r3.1 Section 8.2.9.5.2: Get Log (Opcode 0401h) * The device shall return Invalid Input if the Offset or Length * fields attempt to access beyond the size of the log as reported by Get * Supported Logs. * * The CEL buffer is large enough to fit all commands in the emulation, so * the only possible failure would be if the mailbox itself isn't big * enough. */ if (get_log->offset + get_log->length > cci->payload_max) { return CXL_MBOX_INVALID_INPUT; } if (!qemu_uuid_is_equal(&get_log->uuid, &cel_uuid)) { return CXL_MBOX_INVALID_LOG; } /* Store off everything to local variables so we can wipe out the payload */ *len_out = get_log->length; memmove(payload_out, cci->cel_log + get_log->offset, get_log->length); return CXL_MBOX_SUCCESS; } /* CXL r3.1 section 8.2.9.6: Features */ /* * Get Supported Features output payload * CXL r3.1 section 8.2.9.6.1 Table 8-96 */ typedef struct CXLSupportedFeatureHeader { uint16_t entries; uint16_t nsuppfeats_dev; uint32_t reserved; } QEMU_PACKED CXLSupportedFeatureHeader; /* * Get Supported Features Supported Feature Entry * CXL r3.1 section 8.2.9.6.1 Table 8-97 */ typedef struct CXLSupportedFeatureEntry { QemuUUID uuid; uint16_t feat_index; uint16_t get_feat_size; uint16_t set_feat_size; uint32_t attr_flags; uint8_t get_feat_version; uint8_t set_feat_version; uint16_t set_feat_effects; uint8_t rsvd[18]; } QEMU_PACKED CXLSupportedFeatureEntry; /* * Get Supported Features Supported Feature Entry * CXL rev 3.1 section 8.2.9.6.1 Table 8-97 */ /* Supported Feature Entry : attribute flags */ #define CXL_FEAT_ENTRY_ATTR_FLAG_CHANGABLE BIT(0) #define CXL_FEAT_ENTRY_ATTR_FLAG_DEEPEST_RESET_PERSISTENCE_MASK GENMASK(3, 1) #define CXL_FEAT_ENTRY_ATTR_FLAG_PERSIST_ACROSS_FIRMWARE_UPDATE BIT(4) #define CXL_FEAT_ENTRY_ATTR_FLAG_SUPPORT_DEFAULT_SELECTION BIT(5) #define CXL_FEAT_ENTRY_ATTR_FLAG_SUPPORT_SAVED_SELECTION BIT(6) /* Supported Feature Entry : set feature effects */ #define CXL_FEAT_ENTRY_SFE_CONFIG_CHANGE_COLD_RESET BIT(0) #define CXL_FEAT_ENTRY_SFE_IMMEDIATE_CONFIG_CHANGE BIT(1) #define CXL_FEAT_ENTRY_SFE_IMMEDIATE_DATA_CHANGE BIT(2) #define CXL_FEAT_ENTRY_SFE_IMMEDIATE_POLICY_CHANGE BIT(3) #define CXL_FEAT_ENTRY_SFE_IMMEDIATE_LOG_CHANGE BIT(4) #define CXL_FEAT_ENTRY_SFE_SECURITY_STATE_CHANGE BIT(5) #define CXL_FEAT_ENTRY_SFE_BACKGROUND_OPERATION BIT(6) #define CXL_FEAT_ENTRY_SFE_SUPPORT_SECONDARY_MAILBOX BIT(7) #define CXL_FEAT_ENTRY_SFE_SUPPORT_ABORT_BACKGROUND_OPERATION BIT(8) #define CXL_FEAT_ENTRY_SFE_CEL_VALID BIT(9) #define CXL_FEAT_ENTRY_SFE_CONFIG_CHANGE_CONV_RESET BIT(10) #define CXL_FEAT_ENTRY_SFE_CONFIG_CHANGE_CXL_RESET BIT(11) enum CXL_SUPPORTED_FEATURES_LIST { CXL_FEATURE_PATROL_SCRUB = 0, CXL_FEATURE_ECS, CXL_FEATURE_MAX }; /* Get Feature CXL 3.1 Spec 8.2.9.6.2 */ /* * Get Feature input payload * CXL r3.1 section 8.2.9.6.2 Table 8-99 */ /* Get Feature : Payload in selection */ enum CXL_GET_FEATURE_SELECTION { CXL_GET_FEATURE_SEL_CURRENT_VALUE, CXL_GET_FEATURE_SEL_DEFAULT_VALUE, CXL_GET_FEATURE_SEL_SAVED_VALUE, CXL_GET_FEATURE_SEL_MAX }; /* Set Feature CXL 3.1 Spec 8.2.9.6.3 */ /* * Set Feature input payload * CXL r3.1 section 8.2.9.6.3 Table 8-101 */ typedef struct CXLSetFeatureInHeader { QemuUUID uuid; uint32_t flags; uint16_t offset; uint8_t version; uint8_t rsvd[9]; } QEMU_PACKED QEMU_ALIGNED(16) CXLSetFeatureInHeader; /* Set Feature : Payload in flags */ #define CXL_SET_FEATURE_FLAG_DATA_TRANSFER_MASK 0x7 enum CXL_SET_FEATURE_FLAG_DATA_TRANSFER { CXL_SET_FEATURE_FLAG_FULL_DATA_TRANSFER, CXL_SET_FEATURE_FLAG_INITIATE_DATA_TRANSFER, CXL_SET_FEATURE_FLAG_CONTINUE_DATA_TRANSFER, CXL_SET_FEATURE_FLAG_FINISH_DATA_TRANSFER, CXL_SET_FEATURE_FLAG_ABORT_DATA_TRANSFER, CXL_SET_FEATURE_FLAG_DATA_TRANSFER_MAX }; #define CXL_SET_FEAT_DATA_SAVED_ACROSS_RESET BIT(3) /* CXL r3.1 section 8.2.9.9.11.1: Device Patrol Scrub Control Feature */ static const QemuUUID patrol_scrub_uuid = { .data = UUID(0x96dad7d6, 0xfde8, 0x482b, 0xa7, 0x33, 0x75, 0x77, 0x4e, 0x06, 0xdb, 0x8a) }; typedef struct CXLMemPatrolScrubSetFeature { CXLSetFeatureInHeader hdr; CXLMemPatrolScrubWriteAttrs feat_data; } QEMU_PACKED QEMU_ALIGNED(16) CXLMemPatrolScrubSetFeature; /* * CXL r3.1 section 8.2.9.9.11.2: * DDR5 Error Check Scrub (ECS) Control Feature */ static const QemuUUID ecs_uuid = { .data = UUID(0xe5b13f22, 0x2328, 0x4a14, 0xb8, 0xba, 0xb9, 0x69, 0x1e, 0x89, 0x33, 0x86) }; typedef struct CXLMemECSSetFeature { CXLSetFeatureInHeader hdr; CXLMemECSWriteAttrs feat_data[]; } QEMU_PACKED QEMU_ALIGNED(16) CXLMemECSSetFeature; /* CXL r3.1 section 8.2.9.6.1: Get Supported Features (Opcode 0500h) */ static CXLRetCode cmd_features_get_supported(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { struct { uint32_t count; uint16_t start_index; uint16_t reserved; } QEMU_PACKED QEMU_ALIGNED(16) * get_feats_in = (void *)payload_in; struct { CXLSupportedFeatureHeader hdr; CXLSupportedFeatureEntry feat_entries[]; } QEMU_PACKED QEMU_ALIGNED(16) * get_feats_out = (void *)payload_out; uint16_t index, req_entries; uint16_t entry; if (!object_dynamic_cast(OBJECT(cci->d), TYPE_CXL_TYPE3)) { return CXL_MBOX_UNSUPPORTED; } if (get_feats_in->count < sizeof(CXLSupportedFeatureHeader) || get_feats_in->start_index >= CXL_FEATURE_MAX) { return CXL_MBOX_INVALID_INPUT; } req_entries = (get_feats_in->count - sizeof(CXLSupportedFeatureHeader)) / sizeof(CXLSupportedFeatureEntry); req_entries = MIN(req_entries, (CXL_FEATURE_MAX - get_feats_in->start_index)); for (entry = 0, index = get_feats_in->start_index; entry < req_entries; index++) { switch (index) { case CXL_FEATURE_PATROL_SCRUB: /* Fill supported feature entry for device patrol scrub control */ get_feats_out->feat_entries[entry++] = (struct CXLSupportedFeatureEntry) { .uuid = patrol_scrub_uuid, .feat_index = index, .get_feat_size = sizeof(CXLMemPatrolScrubReadAttrs), .set_feat_size = sizeof(CXLMemPatrolScrubWriteAttrs), .attr_flags = CXL_FEAT_ENTRY_ATTR_FLAG_CHANGABLE, .get_feat_version = CXL_MEMDEV_PS_GET_FEATURE_VERSION, .set_feat_version = CXL_MEMDEV_PS_SET_FEATURE_VERSION, .set_feat_effects = CXL_FEAT_ENTRY_SFE_IMMEDIATE_CONFIG_CHANGE | CXL_FEAT_ENTRY_SFE_CEL_VALID, }; break; case CXL_FEATURE_ECS: /* Fill supported feature entry for device DDR5 ECS control */ get_feats_out->feat_entries[entry++] = (struct CXLSupportedFeatureEntry) { .uuid = ecs_uuid, .feat_index = index, .get_feat_size = CXL_ECS_NUM_MEDIA_FRUS * sizeof(CXLMemECSReadAttrs), .set_feat_size = CXL_ECS_NUM_MEDIA_FRUS * sizeof(CXLMemECSWriteAttrs), .attr_flags = CXL_FEAT_ENTRY_ATTR_FLAG_CHANGABLE, .get_feat_version = CXL_ECS_GET_FEATURE_VERSION, .set_feat_version = CXL_ECS_SET_FEATURE_VERSION, .set_feat_effects = CXL_FEAT_ENTRY_SFE_IMMEDIATE_CONFIG_CHANGE | CXL_FEAT_ENTRY_SFE_CEL_VALID, }; break; default: __builtin_unreachable(); } } get_feats_out->hdr.nsuppfeats_dev = CXL_FEATURE_MAX; get_feats_out->hdr.entries = req_entries; *len_out = sizeof(CXLSupportedFeatureHeader) + req_entries * sizeof(CXLSupportedFeatureEntry); return CXL_MBOX_SUCCESS; } /* CXL r3.1 section 8.2.9.6.2: Get Feature (Opcode 0501h) */ static CXLRetCode cmd_features_get_feature(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { struct { QemuUUID uuid; uint16_t offset; uint16_t count; uint8_t selection; } QEMU_PACKED QEMU_ALIGNED(16) * get_feature; uint16_t bytes_to_copy = 0; CXLType3Dev *ct3d; CXLSetFeatureInfo *set_feat_info; if (!object_dynamic_cast(OBJECT(cci->d), TYPE_CXL_TYPE3)) { return CXL_MBOX_UNSUPPORTED; } ct3d = CXL_TYPE3(cci->d); get_feature = (void *)payload_in; set_feat_info = &ct3d->set_feat_info; if (qemu_uuid_is_equal(&get_feature->uuid, &set_feat_info->uuid)) { return CXL_MBOX_FEATURE_TRANSFER_IN_PROGRESS; } if (get_feature->selection != CXL_GET_FEATURE_SEL_CURRENT_VALUE) { return CXL_MBOX_UNSUPPORTED; } if (get_feature->offset + get_feature->count > cci->payload_max) { return CXL_MBOX_INVALID_INPUT; } if (qemu_uuid_is_equal(&get_feature->uuid, &patrol_scrub_uuid)) { if (get_feature->offset >= sizeof(CXLMemPatrolScrubReadAttrs)) { return CXL_MBOX_INVALID_INPUT; } bytes_to_copy = sizeof(CXLMemPatrolScrubReadAttrs) - get_feature->offset; bytes_to_copy = MIN(bytes_to_copy, get_feature->count); memcpy(payload_out, (uint8_t *)&ct3d->patrol_scrub_attrs + get_feature->offset, bytes_to_copy); } else if (qemu_uuid_is_equal(&get_feature->uuid, &ecs_uuid)) { if (get_feature->offset >= CXL_ECS_NUM_MEDIA_FRUS * sizeof(CXLMemECSReadAttrs)) { return CXL_MBOX_INVALID_INPUT; } bytes_to_copy = CXL_ECS_NUM_MEDIA_FRUS * sizeof(CXLMemECSReadAttrs) - get_feature->offset; bytes_to_copy = MIN(bytes_to_copy, get_feature->count); memcpy(payload_out, (uint8_t *)&ct3d->ecs_attrs + get_feature->offset, bytes_to_copy); } else { return CXL_MBOX_UNSUPPORTED; } *len_out = bytes_to_copy; return CXL_MBOX_SUCCESS; } /* CXL r3.1 section 8.2.9.6.3: Set Feature (Opcode 0502h) */ static CXLRetCode cmd_features_set_feature(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { CXLSetFeatureInHeader *hdr = (void *)payload_in; CXLMemPatrolScrubWriteAttrs *ps_write_attrs; CXLMemPatrolScrubSetFeature *ps_set_feature; CXLMemECSWriteAttrs *ecs_write_attrs; CXLMemECSSetFeature *ecs_set_feature; CXLSetFeatureInfo *set_feat_info; uint16_t bytes_to_copy = 0; uint8_t data_transfer_flag; CXLType3Dev *ct3d; uint16_t count; if (!object_dynamic_cast(OBJECT(cci->d), TYPE_CXL_TYPE3)) { return CXL_MBOX_UNSUPPORTED; } ct3d = CXL_TYPE3(cci->d); set_feat_info = &ct3d->set_feat_info; if (!qemu_uuid_is_null(&set_feat_info->uuid) && !qemu_uuid_is_equal(&hdr->uuid, &set_feat_info->uuid)) { return CXL_MBOX_FEATURE_TRANSFER_IN_PROGRESS; } if (hdr->flags & CXL_SET_FEAT_DATA_SAVED_ACROSS_RESET) { set_feat_info->data_saved_across_reset = true; } else { set_feat_info->data_saved_across_reset = false; } data_transfer_flag = hdr->flags & CXL_SET_FEATURE_FLAG_DATA_TRANSFER_MASK; if (data_transfer_flag == CXL_SET_FEATURE_FLAG_INITIATE_DATA_TRANSFER) { set_feat_info->uuid = hdr->uuid; set_feat_info->data_size = 0; } set_feat_info->data_transfer_flag = data_transfer_flag; set_feat_info->data_offset = hdr->offset; bytes_to_copy = len_in - sizeof(CXLSetFeatureInHeader); if (qemu_uuid_is_equal(&hdr->uuid, &patrol_scrub_uuid)) { if (hdr->version != CXL_MEMDEV_PS_SET_FEATURE_VERSION) { return CXL_MBOX_UNSUPPORTED; } ps_set_feature = (void *)payload_in; ps_write_attrs = &ps_set_feature->feat_data; memcpy((uint8_t *)&ct3d->patrol_scrub_wr_attrs + hdr->offset, ps_write_attrs, bytes_to_copy); set_feat_info->data_size += bytes_to_copy; if (data_transfer_flag == CXL_SET_FEATURE_FLAG_FULL_DATA_TRANSFER || data_transfer_flag == CXL_SET_FEATURE_FLAG_FINISH_DATA_TRANSFER) { ct3d->patrol_scrub_attrs.scrub_cycle &= ~0xFF; ct3d->patrol_scrub_attrs.scrub_cycle |= ct3d->patrol_scrub_wr_attrs.scrub_cycle_hr & 0xFF; ct3d->patrol_scrub_attrs.scrub_flags &= ~0x1; ct3d->patrol_scrub_attrs.scrub_flags |= ct3d->patrol_scrub_wr_attrs.scrub_flags & 0x1; } } else if (qemu_uuid_is_equal(&hdr->uuid, &ecs_uuid)) { if (hdr->version != CXL_ECS_SET_FEATURE_VERSION) { return CXL_MBOX_UNSUPPORTED; } ecs_set_feature = (void *)payload_in; ecs_write_attrs = ecs_set_feature->feat_data; memcpy((uint8_t *)ct3d->ecs_wr_attrs + hdr->offset, ecs_write_attrs, bytes_to_copy); set_feat_info->data_size += bytes_to_copy; if (data_transfer_flag == CXL_SET_FEATURE_FLAG_FULL_DATA_TRANSFER || data_transfer_flag == CXL_SET_FEATURE_FLAG_FINISH_DATA_TRANSFER) { for (count = 0; count < CXL_ECS_NUM_MEDIA_FRUS; count++) { ct3d->ecs_attrs[count].ecs_log_cap = ct3d->ecs_wr_attrs[count].ecs_log_cap; ct3d->ecs_attrs[count].ecs_config = ct3d->ecs_wr_attrs[count].ecs_config & 0x1F; } } } else { return CXL_MBOX_UNSUPPORTED; } if (data_transfer_flag == CXL_SET_FEATURE_FLAG_FULL_DATA_TRANSFER || data_transfer_flag == CXL_SET_FEATURE_FLAG_FINISH_DATA_TRANSFER || data_transfer_flag == CXL_SET_FEATURE_FLAG_ABORT_DATA_TRANSFER) { memset(&set_feat_info->uuid, 0, sizeof(QemuUUID)); if (qemu_uuid_is_equal(&hdr->uuid, &patrol_scrub_uuid)) { memset(&ct3d->patrol_scrub_wr_attrs, 0, set_feat_info->data_size); } else if (qemu_uuid_is_equal(&hdr->uuid, &ecs_uuid)) { memset(ct3d->ecs_wr_attrs, 0, set_feat_info->data_size); } set_feat_info->data_transfer_flag = 0; set_feat_info->data_saved_across_reset = false; set_feat_info->data_offset = 0; set_feat_info->data_size = 0; } return CXL_MBOX_SUCCESS; } /* CXL r3.1 Section 8.2.9.9.1.1: Identify Memory Device (Opcode 4000h) */ static CXLRetCode cmd_identify_memory_device(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { struct { char fw_revision[0x10]; uint64_t total_capacity; uint64_t volatile_capacity; uint64_t persistent_capacity; uint64_t partition_align; uint16_t info_event_log_size; uint16_t warning_event_log_size; uint16_t failure_event_log_size; uint16_t fatal_event_log_size; uint32_t lsa_size; uint8_t poison_list_max_mer[3]; uint16_t inject_poison_limit; uint8_t poison_caps; uint8_t qos_telemetry_caps; uint16_t dc_event_log_size; } QEMU_PACKED *id; QEMU_BUILD_BUG_ON(sizeof(*id) != 0x45); CXLType3Dev *ct3d = CXL_TYPE3(cci->d); CXLType3Class *cvc = CXL_TYPE3_GET_CLASS(ct3d); CXLDeviceState *cxl_dstate = &ct3d->cxl_dstate; if ((!QEMU_IS_ALIGNED(cxl_dstate->vmem_size, CXL_CAPACITY_MULTIPLIER)) || (!QEMU_IS_ALIGNED(cxl_dstate->pmem_size, CXL_CAPACITY_MULTIPLIER)) || (!QEMU_IS_ALIGNED(ct3d->dc.total_capacity, CXL_CAPACITY_MULTIPLIER))) { return CXL_MBOX_INTERNAL_ERROR; } id = (void *)payload_out; snprintf(id->fw_revision, 0x10, "BWFW VERSION %02d", 0); stq_le_p(&id->total_capacity, cxl_dstate->static_mem_size / CXL_CAPACITY_MULTIPLIER); stq_le_p(&id->persistent_capacity, cxl_dstate->pmem_size / CXL_CAPACITY_MULTIPLIER); stq_le_p(&id->volatile_capacity, cxl_dstate->vmem_size / CXL_CAPACITY_MULTIPLIER); stl_le_p(&id->lsa_size, cvc->get_lsa_size(ct3d)); /* 256 poison records */ st24_le_p(id->poison_list_max_mer, 256); /* No limit - so limited by main poison record limit */ stw_le_p(&id->inject_poison_limit, 0); stw_le_p(&id->dc_event_log_size, CXL_DC_EVENT_LOG_SIZE); *len_out = sizeof(*id); return CXL_MBOX_SUCCESS; } /* CXL r3.1 Section 8.2.9.9.2.1: Get Partition Info (Opcode 4100h) */ static CXLRetCode cmd_ccls_get_partition_info(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { CXLDeviceState *cxl_dstate = &CXL_TYPE3(cci->d)->cxl_dstate; struct { uint64_t active_vmem; uint64_t active_pmem; uint64_t next_vmem; uint64_t next_pmem; } QEMU_PACKED *part_info = (void *)payload_out; QEMU_BUILD_BUG_ON(sizeof(*part_info) != 0x20); CXLType3Dev *ct3d = container_of(cxl_dstate, CXLType3Dev, cxl_dstate); if ((!QEMU_IS_ALIGNED(cxl_dstate->vmem_size, CXL_CAPACITY_MULTIPLIER)) || (!QEMU_IS_ALIGNED(cxl_dstate->pmem_size, CXL_CAPACITY_MULTIPLIER)) || (!QEMU_IS_ALIGNED(ct3d->dc.total_capacity, CXL_CAPACITY_MULTIPLIER))) { return CXL_MBOX_INTERNAL_ERROR; } stq_le_p(&part_info->active_vmem, cxl_dstate->vmem_size / CXL_CAPACITY_MULTIPLIER); /* * When both next_vmem and next_pmem are 0, there is no pending change to * partitioning. */ stq_le_p(&part_info->next_vmem, 0); stq_le_p(&part_info->active_pmem, cxl_dstate->pmem_size / CXL_CAPACITY_MULTIPLIER); stq_le_p(&part_info->next_pmem, 0); *len_out = sizeof(*part_info); return CXL_MBOX_SUCCESS; } /* CXL r3.1 Section 8.2.9.9.2.3: Get LSA (Opcode 4102h) */ static CXLRetCode cmd_ccls_get_lsa(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { struct { uint32_t offset; uint32_t length; } QEMU_PACKED *get_lsa; CXLType3Dev *ct3d = CXL_TYPE3(cci->d); CXLType3Class *cvc = CXL_TYPE3_GET_CLASS(ct3d); uint32_t offset, length; get_lsa = (void *)payload_in; offset = get_lsa->offset; length = get_lsa->length; if (offset + length > cvc->get_lsa_size(ct3d)) { *len_out = 0; return CXL_MBOX_INVALID_INPUT; } *len_out = cvc->get_lsa(ct3d, payload_out, length, offset); return CXL_MBOX_SUCCESS; } /* CXL r3.1 Section 8.2.9.9.2.4: Set LSA (Opcode 4103h) */ static CXLRetCode cmd_ccls_set_lsa(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { struct set_lsa_pl { uint32_t offset; uint32_t rsvd; uint8_t data[]; } QEMU_PACKED; struct set_lsa_pl *set_lsa_payload = (void *)payload_in; CXLType3Dev *ct3d = CXL_TYPE3(cci->d); CXLType3Class *cvc = CXL_TYPE3_GET_CLASS(ct3d); const size_t hdr_len = offsetof(struct set_lsa_pl, data); *len_out = 0; if (!len_in) { return CXL_MBOX_SUCCESS; } if (set_lsa_payload->offset + len_in > cvc->get_lsa_size(ct3d) + hdr_len) { return CXL_MBOX_INVALID_INPUT; } len_in -= hdr_len; cvc->set_lsa(ct3d, set_lsa_payload->data, len_in, set_lsa_payload->offset); return CXL_MBOX_SUCCESS; } /* Perform the actual device zeroing */ static void __do_sanitization(CXLType3Dev *ct3d) { MemoryRegion *mr; if (ct3d->hostvmem) { mr = host_memory_backend_get_memory(ct3d->hostvmem); if (mr) { void *hostmem = memory_region_get_ram_ptr(mr); memset(hostmem, 0, memory_region_size(mr)); } } if (ct3d->hostpmem) { mr = host_memory_backend_get_memory(ct3d->hostpmem); if (mr) { void *hostmem = memory_region_get_ram_ptr(mr); memset(hostmem, 0, memory_region_size(mr)); } } if (ct3d->lsa) { mr = host_memory_backend_get_memory(ct3d->lsa); if (mr) { void *lsa = memory_region_get_ram_ptr(mr); memset(lsa, 0, memory_region_size(mr)); } } cxl_discard_all_event_records(&ct3d->cxl_dstate); } /* * CXL r3.1 Section 8.2.9.9.5.1: Sanitize (Opcode 4400h) * * Once the Sanitize command has started successfully, the device shall be * placed in the media disabled state. If the command fails or is interrupted * by a reset or power failure, it shall remain in the media disabled state * until a successful Sanitize command has been completed. During this state: * * 1. Memory writes to the device will have no effect, and all memory reads * will return random values (no user data returned, even for locations that * the failed Sanitize operation didn’t sanitize yet). * * 2. Mailbox commands shall still be processed in the disabled state, except * that commands that access Sanitized areas shall fail with the Media Disabled * error code. */ static CXLRetCode cmd_sanitize_overwrite(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { CXLType3Dev *ct3d = CXL_TYPE3(cci->d); uint64_t total_mem; /* in Mb */ int secs; total_mem = (ct3d->cxl_dstate.vmem_size + ct3d->cxl_dstate.pmem_size) >> 20; if (total_mem <= 512) { secs = 4; } else if (total_mem <= 1024) { secs = 8; } else if (total_mem <= 2 * 1024) { secs = 15; } else if (total_mem <= 4 * 1024) { secs = 30; } else if (total_mem <= 8 * 1024) { secs = 60; } else if (total_mem <= 16 * 1024) { secs = 2 * 60; } else if (total_mem <= 32 * 1024) { secs = 4 * 60; } else if (total_mem <= 64 * 1024) { secs = 8 * 60; } else if (total_mem <= 128 * 1024) { secs = 15 * 60; } else if (total_mem <= 256 * 1024) { secs = 30 * 60; } else if (total_mem <= 512 * 1024) { secs = 60 * 60; } else if (total_mem <= 1024 * 1024) { secs = 120 * 60; } else { secs = 240 * 60; /* max 4 hrs */ } /* EBUSY other bg cmds as of now */ cci->bg.runtime = secs * 1000UL; *len_out = 0; cxl_dev_disable_media(&ct3d->cxl_dstate); /* sanitize when done */ return CXL_MBOX_BG_STARTED; } static CXLRetCode cmd_get_security_state(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { uint32_t *state = (uint32_t *)payload_out; *state = 0; *len_out = 4; return CXL_MBOX_SUCCESS; } /* * CXL r3.1 Section 8.2.9.9.4.1: Get Poison List (Opcode 4300h) * * This is very inefficient, but good enough for now! * Also the payload will always fit, so no need to handle the MORE flag and * make this stateful. We may want to allow longer poison lists to aid * testing that kernel functionality. */ static CXLRetCode cmd_media_get_poison_list(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { struct get_poison_list_pl { uint64_t pa; uint64_t length; } QEMU_PACKED; struct get_poison_list_out_pl { uint8_t flags; uint8_t rsvd1; uint64_t overflow_timestamp; uint16_t count; uint8_t rsvd2[0x14]; struct { uint64_t addr; uint32_t length; uint32_t resv; } QEMU_PACKED records[]; } QEMU_PACKED; struct get_poison_list_pl *in = (void *)payload_in; struct get_poison_list_out_pl *out = (void *)payload_out; CXLType3Dev *ct3d = CXL_TYPE3(cci->d); uint16_t record_count = 0, i = 0; uint64_t query_start, query_length; CXLPoisonList *poison_list = &ct3d->poison_list; CXLPoison *ent; uint16_t out_pl_len; query_start = ldq_le_p(&in->pa); /* 64 byte alignment required */ if (query_start & 0x3f) { return CXL_MBOX_INVALID_INPUT; } query_length = ldq_le_p(&in->length) * CXL_CACHE_LINE_SIZE; QLIST_FOREACH(ent, poison_list, node) { /* Check for no overlap */ if (!ranges_overlap(ent->start, ent->length, query_start, query_length)) { continue; } record_count++; } out_pl_len = sizeof(*out) + record_count * sizeof(out->records[0]); assert(out_pl_len <= CXL_MAILBOX_MAX_PAYLOAD_SIZE); QLIST_FOREACH(ent, poison_list, node) { uint64_t start, stop; /* Check for no overlap */ if (!ranges_overlap(ent->start, ent->length, query_start, query_length)) { continue; } /* Deal with overlap */ start = MAX(ROUND_DOWN(ent->start, 64ull), query_start); stop = MIN(ROUND_DOWN(ent->start, 64ull) + ent->length, query_start + query_length); stq_le_p(&out->records[i].addr, start | (ent->type & 0x7)); stl_le_p(&out->records[i].length, (stop - start) / CXL_CACHE_LINE_SIZE); i++; } if (ct3d->poison_list_overflowed) { out->flags = (1 << 1); stq_le_p(&out->overflow_timestamp, ct3d->poison_list_overflow_ts); } if (scan_media_running(cci)) { out->flags |= (1 << 2); } stw_le_p(&out->count, record_count); *len_out = out_pl_len; return CXL_MBOX_SUCCESS; } /* CXL r3.1 Section 8.2.9.9.4.2: Inject Poison (Opcode 4301h) */ static CXLRetCode cmd_media_inject_poison(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { CXLType3Dev *ct3d = CXL_TYPE3(cci->d); CXLPoisonList *poison_list = &ct3d->poison_list; CXLPoison *ent; struct inject_poison_pl { uint64_t dpa; }; struct inject_poison_pl *in = (void *)payload_in; uint64_t dpa = ldq_le_p(&in->dpa); CXLPoison *p; QLIST_FOREACH(ent, poison_list, node) { if (dpa >= ent->start && dpa + CXL_CACHE_LINE_SIZE <= ent->start + ent->length) { return CXL_MBOX_SUCCESS; } } /* * Freeze the list if there is an on-going scan media operation. */ if (scan_media_running(cci)) { /* * XXX: Spec is ambiguous - is this case considered * a successful return despite not adding to the list? */ goto success; } if (ct3d->poison_list_cnt == CXL_POISON_LIST_LIMIT) { return CXL_MBOX_INJECT_POISON_LIMIT; } p = g_new0(CXLPoison, 1); p->length = CXL_CACHE_LINE_SIZE; p->start = dpa; p->type = CXL_POISON_TYPE_INJECTED; /* * Possible todo: Merge with existing entry if next to it and if same type */ QLIST_INSERT_HEAD(poison_list, p, node); ct3d->poison_list_cnt++; success: *len_out = 0; return CXL_MBOX_SUCCESS; } /* CXL r3.1 Section 8.2.9.9.4.3: Clear Poison (Opcode 4302h */ static CXLRetCode cmd_media_clear_poison(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { CXLType3Dev *ct3d = CXL_TYPE3(cci->d); CXLDeviceState *cxl_dstate = &ct3d->cxl_dstate; CXLPoisonList *poison_list = &ct3d->poison_list; CXLType3Class *cvc = CXL_TYPE3_GET_CLASS(ct3d); struct clear_poison_pl { uint64_t dpa; uint8_t data[64]; }; CXLPoison *ent; uint64_t dpa; struct clear_poison_pl *in = (void *)payload_in; dpa = ldq_le_p(&in->dpa); if (dpa + CXL_CACHE_LINE_SIZE > cxl_dstate->static_mem_size + ct3d->dc.total_capacity) { return CXL_MBOX_INVALID_PA; } /* Clearing a region with no poison is not an error so always do so */ if (cvc->set_cacheline) { if (!cvc->set_cacheline(ct3d, dpa, in->data)) { return CXL_MBOX_INTERNAL_ERROR; } } /* * Freeze the list if there is an on-going scan media operation. */ if (scan_media_running(cci)) { /* * XXX: Spec is ambiguous - is this case considered * a successful return despite not removing from the list? */ goto success; } QLIST_FOREACH(ent, poison_list, node) { /* * Test for contained in entry. Simpler than general case * as clearing 64 bytes and entries 64 byte aligned */ if ((dpa >= ent->start) && (dpa < ent->start + ent->length)) { break; } } if (!ent) { goto success; } QLIST_REMOVE(ent, node); ct3d->poison_list_cnt--; if (dpa > ent->start) { CXLPoison *frag; /* Cannot overflow as replacing existing entry */ frag = g_new0(CXLPoison, 1); frag->start = ent->start; frag->length = dpa - ent->start; frag->type = ent->type; QLIST_INSERT_HEAD(poison_list, frag, node); ct3d->poison_list_cnt++; } if (dpa + CXL_CACHE_LINE_SIZE < ent->start + ent->length) { CXLPoison *frag; if (ct3d->poison_list_cnt == CXL_POISON_LIST_LIMIT) { cxl_set_poison_list_overflowed(ct3d); } else { frag = g_new0(CXLPoison, 1); frag->start = dpa + CXL_CACHE_LINE_SIZE; frag->length = ent->start + ent->length - frag->start; frag->type = ent->type; QLIST_INSERT_HEAD(poison_list, frag, node); ct3d->poison_list_cnt++; } } /* Any fragments have been added, free original entry */ g_free(ent); success: *len_out = 0; return CXL_MBOX_SUCCESS; } /* * CXL r3.1 section 8.2.9.9.4.4: Get Scan Media Capabilities */ static CXLRetCode cmd_media_get_scan_media_capabilities(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { struct get_scan_media_capabilities_pl { uint64_t pa; uint64_t length; } QEMU_PACKED; struct get_scan_media_capabilities_out_pl { uint32_t estimated_runtime_ms; }; CXLType3Dev *ct3d = CXL_TYPE3(cci->d); CXLDeviceState *cxl_dstate = &ct3d->cxl_dstate; struct get_scan_media_capabilities_pl *in = (void *)payload_in; struct get_scan_media_capabilities_out_pl *out = (void *)payload_out; uint64_t query_start; uint64_t query_length; query_start = ldq_le_p(&in->pa); /* 64 byte alignment required */ if (query_start & 0x3f) { return CXL_MBOX_INVALID_INPUT; } query_length = ldq_le_p(&in->length) * CXL_CACHE_LINE_SIZE; if (query_start + query_length > cxl_dstate->static_mem_size) { return CXL_MBOX_INVALID_PA; } /* * Just use 400 nanosecond access/read latency + 100 ns for * the cost of updating the poison list. For small enough * chunks return at least 1 ms. */ stl_le_p(&out->estimated_runtime_ms, MAX(1, query_length * (0.0005L / 64))); *len_out = sizeof(*out); return CXL_MBOX_SUCCESS; } static void __do_scan_media(CXLType3Dev *ct3d) { CXLPoison *ent; unsigned int results_cnt = 0; QLIST_FOREACH(ent, &ct3d->scan_media_results, node) { results_cnt++; } /* only scan media may clear the overflow */ if (ct3d->poison_list_overflowed && ct3d->poison_list_cnt == results_cnt) { cxl_clear_poison_list_overflowed(ct3d); } /* scan media has run since last conventional reset */ ct3d->scan_media_hasrun = true; } /* * CXL r3.1 section 8.2.9.9.4.5: Scan Media */ static CXLRetCode cmd_media_scan_media(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { struct scan_media_pl { uint64_t pa; uint64_t length; uint8_t flags; } QEMU_PACKED; struct scan_media_pl *in = (void *)payload_in; CXLType3Dev *ct3d = CXL_TYPE3(cci->d); CXLDeviceState *cxl_dstate = &ct3d->cxl_dstate; uint64_t query_start; uint64_t query_length; CXLPoison *ent, *next; query_start = ldq_le_p(&in->pa); /* 64 byte alignment required */ if (query_start & 0x3f) { return CXL_MBOX_INVALID_INPUT; } query_length = ldq_le_p(&in->length) * CXL_CACHE_LINE_SIZE; if (query_start + query_length > cxl_dstate->static_mem_size) { return CXL_MBOX_INVALID_PA; } if (ct3d->dc.num_regions && query_start + query_length >= cxl_dstate->static_mem_size + ct3d->dc.total_capacity) { return CXL_MBOX_INVALID_PA; } if (in->flags == 0) { /* TODO */ qemu_log_mask(LOG_UNIMP, "Scan Media Event Log is unsupported\n"); } /* any previous results are discarded upon a new Scan Media */ QLIST_FOREACH_SAFE(ent, &ct3d->scan_media_results, node, next) { QLIST_REMOVE(ent, node); g_free(ent); } /* kill the poison list - it will be recreated */ if (ct3d->poison_list_overflowed) { QLIST_FOREACH_SAFE(ent, &ct3d->poison_list, node, next) { QLIST_REMOVE(ent, node); g_free(ent); ct3d->poison_list_cnt--; } } /* * Scan the backup list and move corresponding entries * into the results list, updating the poison list * when possible. */ QLIST_FOREACH_SAFE(ent, &ct3d->poison_list_bkp, node, next) { CXLPoison *res; if (ent->start >= query_start + query_length || ent->start + ent->length <= query_start) { continue; } /* * If a Get Poison List cmd comes in while this * scan is being done, it will see the new complete * list, while setting the respective flag. */ if (ct3d->poison_list_cnt < CXL_POISON_LIST_LIMIT) { CXLPoison *p = g_new0(CXLPoison, 1); p->start = ent->start; p->length = ent->length; p->type = ent->type; QLIST_INSERT_HEAD(&ct3d->poison_list, p, node); ct3d->poison_list_cnt++; } res = g_new0(CXLPoison, 1); res->start = ent->start; res->length = ent->length; res->type = ent->type; QLIST_INSERT_HEAD(&ct3d->scan_media_results, res, node); QLIST_REMOVE(ent, node); g_free(ent); } cci->bg.runtime = MAX(1, query_length * (0.0005L / 64)); *len_out = 0; return CXL_MBOX_BG_STARTED; } /* * CXL r3.1 section 8.2.9.9.4.6: Get Scan Media Results */ static CXLRetCode cmd_media_get_scan_media_results(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { struct get_scan_media_results_out_pl { uint64_t dpa_restart; uint64_t length; uint8_t flags; uint8_t rsvd1; uint16_t count; uint8_t rsvd2[0xc]; struct { uint64_t addr; uint32_t length; uint32_t resv; } QEMU_PACKED records[]; } QEMU_PACKED; struct get_scan_media_results_out_pl *out = (void *)payload_out; CXLType3Dev *ct3d = CXL_TYPE3(cci->d); CXLPoisonList *scan_media_results = &ct3d->scan_media_results; CXLPoison *ent, *next; uint16_t total_count = 0, record_count = 0, i = 0; uint16_t out_pl_len; if (!ct3d->scan_media_hasrun) { return CXL_MBOX_UNSUPPORTED; } /* * Calculate limits, all entries are within the same address range of the * last scan media call. */ QLIST_FOREACH(ent, scan_media_results, node) { size_t rec_size = record_count * sizeof(out->records[0]); if (sizeof(*out) + rec_size < CXL_MAILBOX_MAX_PAYLOAD_SIZE) { record_count++; } total_count++; } out_pl_len = sizeof(*out) + record_count * sizeof(out->records[0]); assert(out_pl_len <= CXL_MAILBOX_MAX_PAYLOAD_SIZE); memset(out, 0, out_pl_len); QLIST_FOREACH_SAFE(ent, scan_media_results, node, next) { uint64_t start, stop; if (i == record_count) { break; } start = ROUND_DOWN(ent->start, 64ull); stop = ROUND_DOWN(ent->start, 64ull) + ent->length; stq_le_p(&out->records[i].addr, start); stl_le_p(&out->records[i].length, (stop - start) / CXL_CACHE_LINE_SIZE); i++; /* consume the returning entry */ QLIST_REMOVE(ent, node); g_free(ent); } stw_le_p(&out->count, record_count); if (total_count > record_count) { out->flags = (1 << 0); /* More Media Error Records */ } *len_out = out_pl_len; return CXL_MBOX_SUCCESS; } /* * CXL r3.1 section 8.2.9.9.9.1: Get Dynamic Capacity Configuration * (Opcode: 4800h) */ static CXLRetCode cmd_dcd_get_dyn_cap_config(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { CXLType3Dev *ct3d = CXL_TYPE3(cci->d); struct { uint8_t region_cnt; uint8_t start_rid; } QEMU_PACKED *in = (void *)payload_in; struct { uint8_t num_regions; uint8_t regions_returned; uint8_t rsvd1[6]; struct { uint64_t base; uint64_t decode_len; uint64_t region_len; uint64_t block_size; uint32_t dsmadhandle; uint8_t flags; uint8_t rsvd2[3]; } QEMU_PACKED records[]; } QEMU_PACKED *out = (void *)payload_out; struct { uint32_t num_extents_supported; uint32_t num_extents_available; uint32_t num_tags_supported; uint32_t num_tags_available; } QEMU_PACKED *extra_out; uint16_t record_count; uint16_t i; uint16_t out_pl_len; uint8_t start_rid; start_rid = in->start_rid; if (start_rid >= ct3d->dc.num_regions) { return CXL_MBOX_INVALID_INPUT; } record_count = MIN(ct3d->dc.num_regions - in->start_rid, in->region_cnt); out_pl_len = sizeof(*out) + record_count * sizeof(out->records[0]); extra_out = (void *)(payload_out + out_pl_len); out_pl_len += sizeof(*extra_out); assert(out_pl_len <= CXL_MAILBOX_MAX_PAYLOAD_SIZE); out->num_regions = ct3d->dc.num_regions; out->regions_returned = record_count; for (i = 0; i < record_count; i++) { stq_le_p(&out->records[i].base, ct3d->dc.regions[start_rid + i].base); stq_le_p(&out->records[i].decode_len, ct3d->dc.regions[start_rid + i].decode_len / CXL_CAPACITY_MULTIPLIER); stq_le_p(&out->records[i].region_len, ct3d->dc.regions[start_rid + i].len); stq_le_p(&out->records[i].block_size, ct3d->dc.regions[start_rid + i].block_size); stl_le_p(&out->records[i].dsmadhandle, ct3d->dc.regions[start_rid + i].dsmadhandle); out->records[i].flags = ct3d->dc.regions[start_rid + i].flags; } /* * TODO: Assign values once extents and tags are introduced * to use. */ stl_le_p(&extra_out->num_extents_supported, CXL_NUM_EXTENTS_SUPPORTED); stl_le_p(&extra_out->num_extents_available, CXL_NUM_EXTENTS_SUPPORTED - ct3d->dc.total_extent_count); stl_le_p(&extra_out->num_tags_supported, CXL_NUM_TAGS_SUPPORTED); stl_le_p(&extra_out->num_tags_available, CXL_NUM_TAGS_SUPPORTED); *len_out = out_pl_len; return CXL_MBOX_SUCCESS; } /* * CXL r3.1 section 8.2.9.9.9.2: * Get Dynamic Capacity Extent List (Opcode 4801h) */ static CXLRetCode cmd_dcd_get_dyn_cap_ext_list(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { CXLType3Dev *ct3d = CXL_TYPE3(cci->d); struct { uint32_t extent_cnt; uint32_t start_extent_id; } QEMU_PACKED *in = (void *)payload_in; struct { uint32_t count; uint32_t total_extents; uint32_t generation_num; uint8_t rsvd[4]; CXLDCExtentRaw records[]; } QEMU_PACKED *out = (void *)payload_out; uint32_t start_extent_id = in->start_extent_id; CXLDCExtentList *extent_list = &ct3d->dc.extents; uint16_t record_count = 0, i = 0, record_done = 0; uint16_t out_pl_len, size; CXLDCExtent *ent; if (start_extent_id > ct3d->dc.total_extent_count) { return CXL_MBOX_INVALID_INPUT; } record_count = MIN(in->extent_cnt, ct3d->dc.total_extent_count - start_extent_id); size = CXL_MAILBOX_MAX_PAYLOAD_SIZE - sizeof(*out); record_count = MIN(record_count, size / sizeof(out->records[0])); out_pl_len = sizeof(*out) + record_count * sizeof(out->records[0]); stl_le_p(&out->count, record_count); stl_le_p(&out->total_extents, ct3d->dc.total_extent_count); stl_le_p(&out->generation_num, ct3d->dc.ext_list_gen_seq); if (record_count > 0) { CXLDCExtentRaw *out_rec = &out->records[record_done]; QTAILQ_FOREACH(ent, extent_list, node) { if (i++ < start_extent_id) { continue; } stq_le_p(&out_rec->start_dpa, ent->start_dpa); stq_le_p(&out_rec->len, ent->len); memcpy(&out_rec->tag, ent->tag, 0x10); stw_le_p(&out_rec->shared_seq, ent->shared_seq); record_done++; if (record_done == record_count) { break; } } } *len_out = out_pl_len; return CXL_MBOX_SUCCESS; } /* * Check whether any bit between addr[nr, nr+size) is set, * return true if any bit is set, otherwise return false */ bool test_any_bits_set(const unsigned long *addr, unsigned long nr, unsigned long size) { unsigned long res = find_next_bit(addr, size + nr, nr); return res < nr + size; } CXLDCRegion *cxl_find_dc_region(CXLType3Dev *ct3d, uint64_t dpa, uint64_t len) { int i; CXLDCRegion *region = &ct3d->dc.regions[0]; if (dpa < region->base || dpa >= region->base + ct3d->dc.total_capacity) { return NULL; } /* * CXL r3.1 section 9.13.3: Dynamic Capacity Device (DCD) * * Regions are used in increasing-DPA order, with Region 0 being used for * the lowest DPA of Dynamic Capacity and Region 7 for the highest DPA. * So check from the last region to find where the dpa belongs. Extents that * cross multiple regions are not allowed. */ for (i = ct3d->dc.num_regions - 1; i >= 0; i--) { region = &ct3d->dc.regions[i]; if (dpa >= region->base) { if (dpa + len > region->base + region->len) { return NULL; } return region; } } return NULL; } void cxl_insert_extent_to_extent_list(CXLDCExtentList *list, uint64_t dpa, uint64_t len, uint8_t *tag, uint16_t shared_seq) { CXLDCExtent *extent; extent = g_new0(CXLDCExtent, 1); extent->start_dpa = dpa; extent->len = len; if (tag) { memcpy(extent->tag, tag, 0x10); } extent->shared_seq = shared_seq; QTAILQ_INSERT_TAIL(list, extent, node); } void cxl_remove_extent_from_extent_list(CXLDCExtentList *list, CXLDCExtent *extent) { QTAILQ_REMOVE(list, extent, node); g_free(extent); } /* * Add a new extent to the extent "group" if group exists; * otherwise, create a new group * Return value: the extent group where the extent is inserted. */ CXLDCExtentGroup *cxl_insert_extent_to_extent_group(CXLDCExtentGroup *group, uint64_t dpa, uint64_t len, uint8_t *tag, uint16_t shared_seq) { if (!group) { group = g_new0(CXLDCExtentGroup, 1); QTAILQ_INIT(&group->list); } cxl_insert_extent_to_extent_list(&group->list, dpa, len, tag, shared_seq); return group; } void cxl_extent_group_list_insert_tail(CXLDCExtentGroupList *list, CXLDCExtentGroup *group) { QTAILQ_INSERT_TAIL(list, group, node); } void cxl_extent_group_list_delete_front(CXLDCExtentGroupList *list) { CXLDCExtent *ent, *ent_next; CXLDCExtentGroup *group = QTAILQ_FIRST(list); QTAILQ_REMOVE(list, group, node); QTAILQ_FOREACH_SAFE(ent, &group->list, node, ent_next) { cxl_remove_extent_from_extent_list(&group->list, ent); } g_free(group); } /* * CXL r3.1 Table 8-168: Add Dynamic Capacity Response Input Payload * CXL r3.1 Table 8-170: Release Dynamic Capacity Input Payload */ typedef struct CXLUpdateDCExtentListInPl { uint32_t num_entries_updated; uint8_t flags; uint8_t rsvd[3]; /* CXL r3.1 Table 8-169: Updated Extent */ struct { uint64_t start_dpa; uint64_t len; uint8_t rsvd[8]; } QEMU_PACKED updated_entries[]; } QEMU_PACKED CXLUpdateDCExtentListInPl; /* * For the extents in the extent list to operate, check whether they are valid * 1. The extent should be in the range of a valid DC region; * 2. The extent should not cross multiple regions; * 3. The start DPA and the length of the extent should align with the block * size of the region; * 4. The address range of multiple extents in the list should not overlap. */ static CXLRetCode cxl_detect_malformed_extent_list(CXLType3Dev *ct3d, const CXLUpdateDCExtentListInPl *in) { uint64_t min_block_size = UINT64_MAX; CXLDCRegion *region; CXLDCRegion *lastregion = &ct3d->dc.regions[ct3d->dc.num_regions - 1]; g_autofree unsigned long *blk_bitmap = NULL; uint64_t dpa, len; uint32_t i; for (i = 0; i < ct3d->dc.num_regions; i++) { region = &ct3d->dc.regions[i]; min_block_size = MIN(min_block_size, region->block_size); } blk_bitmap = bitmap_new((lastregion->base + lastregion->len - ct3d->dc.regions[0].base) / min_block_size); for (i = 0; i < in->num_entries_updated; i++) { dpa = in->updated_entries[i].start_dpa; len = in->updated_entries[i].len; region = cxl_find_dc_region(ct3d, dpa, len); if (!region) { return CXL_MBOX_INVALID_PA; } dpa -= ct3d->dc.regions[0].base; if (dpa % region->block_size || len % region->block_size) { return CXL_MBOX_INVALID_EXTENT_LIST; } /* the dpa range already covered by some other extents in the list */ if (test_any_bits_set(blk_bitmap, dpa / min_block_size, len / min_block_size)) { return CXL_MBOX_INVALID_EXTENT_LIST; } bitmap_set(blk_bitmap, dpa / min_block_size, len / min_block_size); } return CXL_MBOX_SUCCESS; } static CXLRetCode cxl_dcd_add_dyn_cap_rsp_dry_run(CXLType3Dev *ct3d, const CXLUpdateDCExtentListInPl *in) { uint32_t i; CXLDCExtent *ent; CXLDCExtentGroup *ext_group; uint64_t dpa, len; Range range1, range2; for (i = 0; i < in->num_entries_updated; i++) { dpa = in->updated_entries[i].start_dpa; len = in->updated_entries[i].len; range_init_nofail(&range1, dpa, len); /* * The host-accepted DPA range must be contained by the first extent * group in the pending list */ ext_group = QTAILQ_FIRST(&ct3d->dc.extents_pending); if (!cxl_extents_contains_dpa_range(&ext_group->list, dpa, len)) { return CXL_MBOX_INVALID_PA; } /* to-be-added range should not overlap with range already accepted */ QTAILQ_FOREACH(ent, &ct3d->dc.extents, node) { range_init_nofail(&range2, ent->start_dpa, ent->len); if (range_overlaps_range(&range1, &range2)) { return CXL_MBOX_INVALID_PA; } } } return CXL_MBOX_SUCCESS; } /* * CXL r3.1 section 8.2.9.9.9.3: Add Dynamic Capacity Response (Opcode 4802h) * An extent is added to the extent list and becomes usable only after the * response is processed successfully. */ static CXLRetCode cmd_dcd_add_dyn_cap_rsp(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { CXLUpdateDCExtentListInPl *in = (void *)payload_in; CXLType3Dev *ct3d = CXL_TYPE3(cci->d); CXLDCExtentList *extent_list = &ct3d->dc.extents; uint32_t i; uint64_t dpa, len; CXLRetCode ret; if (in->num_entries_updated == 0) { cxl_extent_group_list_delete_front(&ct3d->dc.extents_pending); return CXL_MBOX_SUCCESS; } /* Adding extents causes exceeding device's extent tracking ability. */ if (in->num_entries_updated + ct3d->dc.total_extent_count > CXL_NUM_EXTENTS_SUPPORTED) { return CXL_MBOX_RESOURCES_EXHAUSTED; } ret = cxl_detect_malformed_extent_list(ct3d, in); if (ret != CXL_MBOX_SUCCESS) { return ret; } ret = cxl_dcd_add_dyn_cap_rsp_dry_run(ct3d, in); if (ret != CXL_MBOX_SUCCESS) { return ret; } for (i = 0; i < in->num_entries_updated; i++) { dpa = in->updated_entries[i].start_dpa; len = in->updated_entries[i].len; cxl_insert_extent_to_extent_list(extent_list, dpa, len, NULL, 0); ct3d->dc.total_extent_count += 1; ct3_set_region_block_backed(ct3d, dpa, len); } /* Remove the first extent group in the pending list */ cxl_extent_group_list_delete_front(&ct3d->dc.extents_pending); return CXL_MBOX_SUCCESS; } /* * Copy extent list from src to dst * Return value: number of extents copied */ static uint32_t copy_extent_list(CXLDCExtentList *dst, const CXLDCExtentList *src) { uint32_t cnt = 0; CXLDCExtent *ent; if (!dst || !src) { return 0; } QTAILQ_FOREACH(ent, src, node) { cxl_insert_extent_to_extent_list(dst, ent->start_dpa, ent->len, ent->tag, ent->shared_seq); cnt++; } return cnt; } static CXLRetCode cxl_dc_extent_release_dry_run(CXLType3Dev *ct3d, const CXLUpdateDCExtentListInPl *in, CXLDCExtentList *updated_list, uint32_t *updated_list_size) { CXLDCExtent *ent, *ent_next; uint64_t dpa, len; uint32_t i; int cnt_delta = 0; CXLRetCode ret = CXL_MBOX_SUCCESS; QTAILQ_INIT(updated_list); copy_extent_list(updated_list, &ct3d->dc.extents); for (i = 0; i < in->num_entries_updated; i++) { Range range; dpa = in->updated_entries[i].start_dpa; len = in->updated_entries[i].len; /* Check if the DPA range is not fully backed with valid extents */ if (!ct3_test_region_block_backed(ct3d, dpa, len)) { ret = CXL_MBOX_INVALID_PA; goto free_and_exit; } /* After this point, extent overflow is the only error can happen */ while (len > 0) { QTAILQ_FOREACH(ent, updated_list, node) { range_init_nofail(&range, ent->start_dpa, ent->len); if (range_contains(&range, dpa)) { uint64_t len1, len2 = 0, len_done = 0; uint64_t ent_start_dpa = ent->start_dpa; uint64_t ent_len = ent->len; len1 = dpa - ent->start_dpa; /* Found the extent or the subset of an existing extent */ if (range_contains(&range, dpa + len - 1)) { len2 = ent_start_dpa + ent_len - dpa - len; } else { dpa = ent_start_dpa + ent_len; } len_done = ent_len - len1 - len2; cxl_remove_extent_from_extent_list(updated_list, ent); cnt_delta--; if (len1) { cxl_insert_extent_to_extent_list(updated_list, ent_start_dpa, len1, NULL, 0); cnt_delta++; } if (len2) { cxl_insert_extent_to_extent_list(updated_list, dpa + len, len2, NULL, 0); cnt_delta++; } if (cnt_delta + ct3d->dc.total_extent_count > CXL_NUM_EXTENTS_SUPPORTED) { ret = CXL_MBOX_RESOURCES_EXHAUSTED; goto free_and_exit; } len -= len_done; break; } } } } free_and_exit: if (ret != CXL_MBOX_SUCCESS) { QTAILQ_FOREACH_SAFE(ent, updated_list, node, ent_next) { cxl_remove_extent_from_extent_list(updated_list, ent); } *updated_list_size = 0; } else { *updated_list_size = ct3d->dc.total_extent_count + cnt_delta; } return ret; } /* * CXL r3.1 section 8.2.9.9.9.4: Release Dynamic Capacity (Opcode 4803h) */ static CXLRetCode cmd_dcd_release_dyn_cap(const struct cxl_cmd *cmd, uint8_t *payload_in, size_t len_in, uint8_t *payload_out, size_t *len_out, CXLCCI *cci) { CXLUpdateDCExtentListInPl *in = (void *)payload_in; CXLType3Dev *ct3d = CXL_TYPE3(cci->d); CXLDCExtentList updated_list; CXLDCExtent *ent, *ent_next; uint32_t updated_list_size; CXLRetCode ret; if (in->num_entries_updated == 0) { return CXL_MBOX_INVALID_INPUT; } ret = cxl_detect_malformed_extent_list(ct3d, in); if (ret != CXL_MBOX_SUCCESS) { return ret; } ret = cxl_dc_extent_release_dry_run(ct3d, in, &updated_list, &updated_list_size); if (ret != CXL_MBOX_SUCCESS) { return ret; } /* * If the dry run release passes, the returned updated_list will * be the updated extent list and we just need to clear the extents * in the accepted list and copy extents in the updated_list to accepted * list and update the extent count; */ QTAILQ_FOREACH_SAFE(ent, &ct3d->dc.extents, node, ent_next) { ct3_clear_region_block_backed(ct3d, ent->start_dpa, ent->len); cxl_remove_extent_from_extent_list(&ct3d->dc.extents, ent); } copy_extent_list(&ct3d->dc.extents, &updated_list); QTAILQ_FOREACH_SAFE(ent, &updated_list, node, ent_next) { ct3_set_region_block_backed(ct3d, ent->start_dpa, ent->len); cxl_remove_extent_from_extent_list(&updated_list, ent); } ct3d->dc.total_extent_count = updated_list_size; return CXL_MBOX_SUCCESS; } static const struct cxl_cmd cxl_cmd_set[256][256] = { [EVENTS][GET_RECORDS] = { "EVENTS_GET_RECORDS", cmd_events_get_records, 1, 0 }, [EVENTS][CLEAR_RECORDS] = { "EVENTS_CLEAR_RECORDS", cmd_events_clear_records, ~0, CXL_MBOX_IMMEDIATE_LOG_CHANGE }, [EVENTS][GET_INTERRUPT_POLICY] = { "EVENTS_GET_INTERRUPT_POLICY", cmd_events_get_interrupt_policy, 0, 0 }, [EVENTS][SET_INTERRUPT_POLICY] = { "EVENTS_SET_INTERRUPT_POLICY", cmd_events_set_interrupt_policy, ~0, CXL_MBOX_IMMEDIATE_CONFIG_CHANGE }, [FIRMWARE_UPDATE][GET_INFO] = { "FIRMWARE_UPDATE_GET_INFO", cmd_firmware_update_get_info, 0, 0 }, [FIRMWARE_UPDATE][TRANSFER] = { "FIRMWARE_UPDATE_TRANSFER", cmd_firmware_update_transfer, ~0, CXL_MBOX_BACKGROUND_OPERATION }, [FIRMWARE_UPDATE][ACTIVATE] = { "FIRMWARE_UPDATE_ACTIVATE", cmd_firmware_update_activate, 2, CXL_MBOX_BACKGROUND_OPERATION }, [TIMESTAMP][GET] = { "TIMESTAMP_GET", cmd_timestamp_get, 0, 0 }, [TIMESTAMP][SET] = { "TIMESTAMP_SET", cmd_timestamp_set, 8, CXL_MBOX_IMMEDIATE_POLICY_CHANGE }, [LOGS][GET_SUPPORTED] = { "LOGS_GET_SUPPORTED", cmd_logs_get_supported, 0, 0 }, [LOGS][GET_LOG] = { "LOGS_GET_LOG", cmd_logs_get_log, 0x18, 0 }, [FEATURES][GET_SUPPORTED] = { "FEATURES_GET_SUPPORTED", cmd_features_get_supported, 0x8, 0 }, [FEATURES][GET_FEATURE] = { "FEATURES_GET_FEATURE", cmd_features_get_feature, 0x15, 0 }, [FEATURES][SET_FEATURE] = { "FEATURES_SET_FEATURE", cmd_features_set_feature, ~0, (CXL_MBOX_IMMEDIATE_CONFIG_CHANGE | CXL_MBOX_IMMEDIATE_DATA_CHANGE | CXL_MBOX_IMMEDIATE_POLICY_CHANGE | CXL_MBOX_IMMEDIATE_LOG_CHANGE | CXL_MBOX_SECURITY_STATE_CHANGE)}, [IDENTIFY][MEMORY_DEVICE] = { "IDENTIFY_MEMORY_DEVICE", cmd_identify_memory_device, 0, 0 }, [CCLS][GET_PARTITION_INFO] = { "CCLS_GET_PARTITION_INFO", cmd_ccls_get_partition_info, 0, 0 }, [CCLS][GET_LSA] = { "CCLS_GET_LSA", cmd_ccls_get_lsa, 8, 0 }, [CCLS][SET_LSA] = { "CCLS_SET_LSA", cmd_ccls_set_lsa, ~0, CXL_MBOX_IMMEDIATE_CONFIG_CHANGE | CXL_MBOX_IMMEDIATE_DATA_CHANGE }, [SANITIZE][OVERWRITE] = { "SANITIZE_OVERWRITE", cmd_sanitize_overwrite, 0, (CXL_MBOX_IMMEDIATE_DATA_CHANGE | CXL_MBOX_SECURITY_STATE_CHANGE | CXL_MBOX_BACKGROUND_OPERATION)}, [PERSISTENT_MEM][GET_SECURITY_STATE] = { "GET_SECURITY_STATE", cmd_get_security_state, 0, 0 }, [MEDIA_AND_POISON][GET_POISON_LIST] = { "MEDIA_AND_POISON_GET_POISON_LIST", cmd_media_get_poison_list, 16, 0 }, [MEDIA_AND_POISON][INJECT_POISON] = { "MEDIA_AND_POISON_INJECT_POISON", cmd_media_inject_poison, 8, 0 }, [MEDIA_AND_POISON][CLEAR_POISON] = { "MEDIA_AND_POISON_CLEAR_POISON", cmd_media_clear_poison, 72, 0 }, [MEDIA_AND_POISON][GET_SCAN_MEDIA_CAPABILITIES] = { "MEDIA_AND_POISON_GET_SCAN_MEDIA_CAPABILITIES", cmd_media_get_scan_media_capabilities, 16, 0 }, [MEDIA_AND_POISON][SCAN_MEDIA] = { "MEDIA_AND_POISON_SCAN_MEDIA", cmd_media_scan_media, 17, CXL_MBOX_BACKGROUND_OPERATION }, [MEDIA_AND_POISON][GET_SCAN_MEDIA_RESULTS] = { "MEDIA_AND_POISON_GET_SCAN_MEDIA_RESULTS", cmd_media_get_scan_media_results, 0, 0 }, }; static const struct cxl_cmd cxl_cmd_set_dcd[256][256] = { [DCD_CONFIG][GET_DC_CONFIG] = { "DCD_GET_DC_CONFIG", cmd_dcd_get_dyn_cap_config, 2, 0 }, [DCD_CONFIG][GET_DYN_CAP_EXT_LIST] = { "DCD_GET_DYNAMIC_CAPACITY_EXTENT_LIST", cmd_dcd_get_dyn_cap_ext_list, 8, 0 }, [DCD_CONFIG][ADD_DYN_CAP_RSP] = { "DCD_ADD_DYNAMIC_CAPACITY_RESPONSE", cmd_dcd_add_dyn_cap_rsp, ~0, CXL_MBOX_IMMEDIATE_DATA_CHANGE }, [DCD_CONFIG][RELEASE_DYN_CAP] = { "DCD_RELEASE_DYNAMIC_CAPACITY", cmd_dcd_release_dyn_cap, ~0, CXL_MBOX_IMMEDIATE_DATA_CHANGE }, }; static const struct cxl_cmd cxl_cmd_set_sw[256][256] = { [INFOSTAT][IS_IDENTIFY] = { "IDENTIFY", cmd_infostat_identify, 0, 0 }, [INFOSTAT][BACKGROUND_OPERATION_STATUS] = { "BACKGROUND_OPERATION_STATUS", cmd_infostat_bg_op_sts, 0, 0 }, [TIMESTAMP][GET] = { "TIMESTAMP_GET", cmd_timestamp_get, 0, 0 }, [TIMESTAMP][SET] = { "TIMESTAMP_SET", cmd_timestamp_set, 8, CXL_MBOX_IMMEDIATE_POLICY_CHANGE }, [LOGS][GET_SUPPORTED] = { "LOGS_GET_SUPPORTED", cmd_logs_get_supported, 0, 0 }, [LOGS][GET_LOG] = { "LOGS_GET_LOG", cmd_logs_get_log, 0x18, 0 }, [PHYSICAL_SWITCH][IDENTIFY_SWITCH_DEVICE] = { "IDENTIFY_SWITCH_DEVICE", cmd_identify_switch_device, 0, 0 }, [PHYSICAL_SWITCH][GET_PHYSICAL_PORT_STATE] = { "SWITCH_PHYSICAL_PORT_STATS", cmd_get_physical_port_state, ~0, 0 }, [TUNNEL][MANAGEMENT_COMMAND] = { "TUNNEL_MANAGEMENT_COMMAND", cmd_tunnel_management_cmd, ~0, 0 }, }; /* * While the command is executing in the background, the device should * update the percentage complete in the Background Command Status Register * at least once per second. */ #define CXL_MBOX_BG_UPDATE_FREQ 1000UL int cxl_process_cci_message(CXLCCI *cci, uint8_t set, uint8_t cmd, size_t len_in, uint8_t *pl_in, size_t *len_out, uint8_t *pl_out, bool *bg_started) { int ret; const struct cxl_cmd *cxl_cmd; opcode_handler h; CXLDeviceState *cxl_dstate; *len_out = 0; cxl_cmd = &cci->cxl_cmd_set[set][cmd]; h = cxl_cmd->handler; if (!h) { qemu_log_mask(LOG_UNIMP, "Command %04xh not implemented\n", set << 8 | cmd); return CXL_MBOX_UNSUPPORTED; } if (len_in != cxl_cmd->in && cxl_cmd->in != ~0) { return CXL_MBOX_INVALID_PAYLOAD_LENGTH; } /* Only one bg command at a time */ if ((cxl_cmd->effect & CXL_MBOX_BACKGROUND_OPERATION) && cci->bg.runtime > 0) { return CXL_MBOX_BUSY; } /* forbid any selected commands while the media is disabled */ if (object_dynamic_cast(OBJECT(cci->d), TYPE_CXL_TYPE3)) { cxl_dstate = &CXL_TYPE3(cci->d)->cxl_dstate; if (cxl_dev_media_disabled(cxl_dstate)) { if (h == cmd_events_get_records || h == cmd_ccls_get_partition_info || h == cmd_ccls_set_lsa || h == cmd_ccls_get_lsa || h == cmd_logs_get_log || h == cmd_media_get_poison_list || h == cmd_media_inject_poison || h == cmd_media_clear_poison || h == cmd_sanitize_overwrite || h == cmd_firmware_update_transfer || h == cmd_firmware_update_activate) { return CXL_MBOX_MEDIA_DISABLED; } } } ret = (*h)(cxl_cmd, pl_in, len_in, pl_out, len_out, cci); if ((cxl_cmd->effect & CXL_MBOX_BACKGROUND_OPERATION) && ret == CXL_MBOX_BG_STARTED) { *bg_started = true; } else { *bg_started = false; } /* Set bg and the return code */ if (*bg_started) { uint64_t now; cci->bg.opcode = (set << 8) | cmd; cci->bg.complete_pct = 0; cci->bg.ret_code = 0; now = qemu_clock_get_ms(QEMU_CLOCK_VIRTUAL); cci->bg.starttime = now; timer_mod(cci->bg.timer, now + CXL_MBOX_BG_UPDATE_FREQ); } return ret; } static void bg_timercb(void *opaque) { CXLCCI *cci = opaque; uint64_t now = qemu_clock_get_ms(QEMU_CLOCK_VIRTUAL); uint64_t total_time = cci->bg.starttime + cci->bg.runtime; assert(cci->bg.runtime > 0); if (now >= total_time) { /* we are done */ uint16_t ret = CXL_MBOX_SUCCESS; cci->bg.complete_pct = 100; cci->bg.ret_code = ret; switch (cci->bg.opcode) { case 0x0201: /* fw transfer */ __do_firmware_xfer(cci); break; case 0x4400: /* sanitize */ { CXLType3Dev *ct3d = CXL_TYPE3(cci->d); __do_sanitization(ct3d); cxl_dev_enable_media(&ct3d->cxl_dstate); } break; case 0x4304: /* scan media */ { CXLType3Dev *ct3d = CXL_TYPE3(cci->d); __do_scan_media(ct3d); break; } default: __builtin_unreachable(); break; } } else { /* estimate only */ cci->bg.complete_pct = 100 * now / total_time; timer_mod(cci->bg.timer, now + CXL_MBOX_BG_UPDATE_FREQ); } if (cci->bg.complete_pct == 100) { /* TODO: generalize to switch CCI */ CXLType3Dev *ct3d = CXL_TYPE3(cci->d); CXLDeviceState *cxl_dstate = &ct3d->cxl_dstate; PCIDevice *pdev = PCI_DEVICE(cci->d); cci->bg.starttime = 0; /* registers are updated, allow new bg-capable cmds */ cci->bg.runtime = 0; if (msix_enabled(pdev)) { msix_notify(pdev, cxl_dstate->mbox_msi_n); } else if (msi_enabled(pdev)) { msi_notify(pdev, cxl_dstate->mbox_msi_n); } } } static void cxl_rebuild_cel(CXLCCI *cci) { cci->cel_size = 0; /* Reset for a fresh build */ for (int set = 0; set < 256; set++) { for (int cmd = 0; cmd < 256; cmd++) { if (cci->cxl_cmd_set[set][cmd].handler) { const struct cxl_cmd *c = &cci->cxl_cmd_set[set][cmd]; struct cel_log *log = &cci->cel_log[cci->cel_size]; log->opcode = (set << 8) | cmd; log->effect = c->effect; cci->cel_size++; } } } } void cxl_init_cci(CXLCCI *cci, size_t payload_max) { cci->payload_max = payload_max; cxl_rebuild_cel(cci); cci->bg.complete_pct = 0; cci->bg.starttime = 0; cci->bg.runtime = 0; cci->bg.timer = timer_new_ms(QEMU_CLOCK_VIRTUAL, bg_timercb, cci); memset(&cci->fw, 0, sizeof(cci->fw)); cci->fw.active_slot = 1; cci->fw.slot[cci->fw.active_slot - 1] = true; } static void cxl_copy_cci_commands(CXLCCI *cci, const struct cxl_cmd (*cxl_cmds)[256]) { for (int set = 0; set < 256; set++) { for (int cmd = 0; cmd < 256; cmd++) { if (cxl_cmds[set][cmd].handler) { cci->cxl_cmd_set[set][cmd] = cxl_cmds[set][cmd]; } } } } void cxl_add_cci_commands(CXLCCI *cci, const struct cxl_cmd (*cxl_cmd_set)[256], size_t payload_max) { cci->payload_max = MAX(payload_max, cci->payload_max); cxl_copy_cci_commands(cci, cxl_cmd_set); cxl_rebuild_cel(cci); } void cxl_initialize_mailbox_swcci(CXLCCI *cci, DeviceState *intf, DeviceState *d, size_t payload_max) { cxl_copy_cci_commands(cci, cxl_cmd_set_sw); cci->d = d; cci->intf = intf; cxl_init_cci(cci, payload_max); } void cxl_initialize_mailbox_t3(CXLCCI *cci, DeviceState *d, size_t payload_max) { CXLType3Dev *ct3d = CXL_TYPE3(d); cxl_copy_cci_commands(cci, cxl_cmd_set); if (ct3d->dc.num_regions) { cxl_copy_cci_commands(cci, cxl_cmd_set_dcd); } cci->d = d; /* No separation for PCI MB as protocol handled in PCI device */ cci->intf = d; cxl_init_cci(cci, payload_max); } static const struct cxl_cmd cxl_cmd_set_t3_ld[256][256] = { [INFOSTAT][IS_IDENTIFY] = { "IDENTIFY", cmd_infostat_identify, 0, 0 }, [LOGS][GET_SUPPORTED] = { "LOGS_GET_SUPPORTED", cmd_logs_get_supported, 0, 0 }, [LOGS][GET_LOG] = { "LOGS_GET_LOG", cmd_logs_get_log, 0x18, 0 }, }; void cxl_initialize_t3_ld_cci(CXLCCI *cci, DeviceState *d, DeviceState *intf, size_t payload_max) { cxl_copy_cci_commands(cci, cxl_cmd_set_t3_ld); cci->d = d; cci->intf = intf; cxl_init_cci(cci, payload_max); } static const struct cxl_cmd cxl_cmd_set_t3_fm_owned_ld_mctp[256][256] = { [INFOSTAT][IS_IDENTIFY] = { "IDENTIFY", cmd_infostat_identify, 0, 0}, [LOGS][GET_SUPPORTED] = { "LOGS_GET_SUPPORTED", cmd_logs_get_supported, 0, 0 }, [LOGS][GET_LOG] = { "LOGS_GET_LOG", cmd_logs_get_log, 0x18, 0 }, [TIMESTAMP][GET] = { "TIMESTAMP_GET", cmd_timestamp_get, 0, 0 }, [TUNNEL][MANAGEMENT_COMMAND] = { "TUNNEL_MANAGEMENT_COMMAND", cmd_tunnel_management_cmd, ~0, 0 }, }; void cxl_initialize_t3_fm_owned_ld_mctpcci(CXLCCI *cci, DeviceState *d, DeviceState *intf, size_t payload_max) { cxl_copy_cci_commands(cci, cxl_cmd_set_t3_fm_owned_ld_mctp); cci->d = d; cci->intf = intf; cxl_init_cci(cci, payload_max); }