/* * QEMU PowerPC pSeries Logical Partition (aka sPAPR) hardware System Emulator * * RTAS events handling * * Copyright (c) 2012 David Gibson, IBM Corporation. * * Permission is hereby granted, free of charge, to any person obtaining a copy * of this software and associated documentation files (the "Software"), to deal * in the Software without restriction, including without limitation the rights * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell * copies of the Software, and to permit persons to whom the Software is * furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice shall be included in * all copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN * THE SOFTWARE. * */ #include "qemu/osdep.h" #include "qapi/error.h" #include "cpu.h" #include "sysemu/sysemu.h" #include "hw/qdev.h" #include "sysemu/device_tree.h" #include "hw/ppc/fdt.h" #include "hw/ppc/spapr.h" #include "hw/ppc/spapr_vio.h" #include "hw/pci/pci.h" #include "hw/pci-host/spapr.h" #include "hw/ppc/spapr_drc.h" #include "qemu/help_option.h" #include "qemu/bcd.h" #include "hw/ppc/spapr_ovec.h" #include #define RTAS_LOG_VERSION_MASK 0xff000000 #define RTAS_LOG_VERSION_6 0x06000000 #define RTAS_LOG_SEVERITY_MASK 0x00e00000 #define RTAS_LOG_SEVERITY_ALREADY_REPORTED 0x00c00000 #define RTAS_LOG_SEVERITY_FATAL 0x00a00000 #define RTAS_LOG_SEVERITY_ERROR 0x00800000 #define RTAS_LOG_SEVERITY_ERROR_SYNC 0x00600000 #define RTAS_LOG_SEVERITY_WARNING 0x00400000 #define RTAS_LOG_SEVERITY_EVENT 0x00200000 #define RTAS_LOG_SEVERITY_NO_ERROR 0x00000000 #define RTAS_LOG_DISPOSITION_MASK 0x00180000 #define RTAS_LOG_DISPOSITION_FULLY_RECOVERED 0x00000000 #define RTAS_LOG_DISPOSITION_LIMITED_RECOVERY 0x00080000 #define RTAS_LOG_DISPOSITION_NOT_RECOVERED 0x00100000 #define RTAS_LOG_OPTIONAL_PART_PRESENT 0x00040000 #define RTAS_LOG_INITIATOR_MASK 0x0000f000 #define RTAS_LOG_INITIATOR_UNKNOWN 0x00000000 #define RTAS_LOG_INITIATOR_CPU 0x00001000 #define RTAS_LOG_INITIATOR_PCI 0x00002000 #define RTAS_LOG_INITIATOR_MEMORY 0x00004000 #define RTAS_LOG_INITIATOR_HOTPLUG 0x00006000 #define RTAS_LOG_TARGET_MASK 0x00000f00 #define RTAS_LOG_TARGET_UNKNOWN 0x00000000 #define RTAS_LOG_TARGET_CPU 0x00000100 #define RTAS_LOG_TARGET_PCI 0x00000200 #define RTAS_LOG_TARGET_MEMORY 0x00000400 #define RTAS_LOG_TARGET_HOTPLUG 0x00000600 #define RTAS_LOG_TYPE_MASK 0x000000ff #define RTAS_LOG_TYPE_OTHER 0x00000000 #define RTAS_LOG_TYPE_RETRY 0x00000001 #define RTAS_LOG_TYPE_TCE_ERR 0x00000002 #define RTAS_LOG_TYPE_INTERN_DEV_FAIL 0x00000003 #define RTAS_LOG_TYPE_TIMEOUT 0x00000004 #define RTAS_LOG_TYPE_DATA_PARITY 0x00000005 #define RTAS_LOG_TYPE_ADDR_PARITY 0x00000006 #define RTAS_LOG_TYPE_CACHE_PARITY 0x00000007 #define RTAS_LOG_TYPE_ADDR_INVALID 0x00000008 #define RTAS_LOG_TYPE_ECC_UNCORR 0x00000009 #define RTAS_LOG_TYPE_ECC_CORR 0x0000000a #define RTAS_LOG_TYPE_EPOW 0x00000040 #define RTAS_LOG_TYPE_HOTPLUG 0x000000e5 struct rtas_error_log { uint32_t summary; uint32_t extended_length; } QEMU_PACKED; struct rtas_event_log_v6 { uint8_t b0; #define RTAS_LOG_V6_B0_VALID 0x80 #define RTAS_LOG_V6_B0_UNRECOVERABLE_ERROR 0x40 #define RTAS_LOG_V6_B0_RECOVERABLE_ERROR 0x20 #define RTAS_LOG_V6_B0_DEGRADED_OPERATION 0x10 #define RTAS_LOG_V6_B0_PREDICTIVE_ERROR 0x08 #define RTAS_LOG_V6_B0_NEW_LOG 0x04 #define RTAS_LOG_V6_B0_BIGENDIAN 0x02 uint8_t _resv1; uint8_t b2; #define RTAS_LOG_V6_B2_POWERPC_FORMAT 0x80 #define RTAS_LOG_V6_B2_LOG_FORMAT_MASK 0x0f #define RTAS_LOG_V6_B2_LOG_FORMAT_PLATFORM_EVENT 0x0e uint8_t _resv2[9]; uint32_t company; #define RTAS_LOG_V6_COMPANY_IBM 0x49424d00 /* IBM */ } QEMU_PACKED; struct rtas_event_log_v6_section_header { uint16_t section_id; uint16_t section_length; uint8_t section_version; uint8_t section_subtype; uint16_t creator_component_id; } QEMU_PACKED; struct rtas_event_log_v6_maina { #define RTAS_LOG_V6_SECTION_ID_MAINA 0x5048 /* PH */ struct rtas_event_log_v6_section_header hdr; uint32_t creation_date; /* BCD: YYYYMMDD */ uint32_t creation_time; /* BCD: HHMMSS00 */ uint8_t _platform1[8]; char creator_id; uint8_t _resv1[2]; uint8_t section_count; uint8_t _resv2[4]; uint8_t _platform2[8]; uint32_t plid; uint8_t _platform3[4]; } QEMU_PACKED; struct rtas_event_log_v6_mainb { #define RTAS_LOG_V6_SECTION_ID_MAINB 0x5548 /* UH */ struct rtas_event_log_v6_section_header hdr; uint8_t subsystem_id; uint8_t _platform1; uint8_t event_severity; uint8_t event_subtype; uint8_t _platform2[4]; uint8_t _resv1[2]; uint16_t action_flags; uint8_t _resv2[4]; } QEMU_PACKED; struct rtas_event_log_v6_epow { #define RTAS_LOG_V6_SECTION_ID_EPOW 0x4550 /* EP */ struct rtas_event_log_v6_section_header hdr; uint8_t sensor_value; #define RTAS_LOG_V6_EPOW_ACTION_RESET 0 #define RTAS_LOG_V6_EPOW_ACTION_WARN_COOLING 1 #define RTAS_LOG_V6_EPOW_ACTION_WARN_POWER 2 #define RTAS_LOG_V6_EPOW_ACTION_SYSTEM_SHUTDOWN 3 #define RTAS_LOG_V6_EPOW_ACTION_SYSTEM_HALT 4 #define RTAS_LOG_V6_EPOW_ACTION_MAIN_ENCLOSURE 5 #define RTAS_LOG_V6_EPOW_ACTION_POWER_OFF 7 uint8_t event_modifier; #define RTAS_LOG_V6_EPOW_MODIFIER_NORMAL 1 #define RTAS_LOG_V6_EPOW_MODIFIER_ON_UPS 2 #define RTAS_LOG_V6_EPOW_MODIFIER_CRITICAL 3 #define RTAS_LOG_V6_EPOW_MODIFIER_TEMPERATURE 4 uint8_t extended_modifier; #define RTAS_LOG_V6_EPOW_XMODIFIER_SYSTEM_WIDE 0 #define RTAS_LOG_V6_EPOW_XMODIFIER_PARTITION_SPECIFIC 1 uint8_t _resv; uint64_t reason_code; } QEMU_PACKED; struct epow_extended_log { struct rtas_event_log_v6 v6hdr; struct rtas_event_log_v6_maina maina; struct rtas_event_log_v6_mainb mainb; struct rtas_event_log_v6_epow epow; } QEMU_PACKED; union drc_identifier { uint32_t index; uint32_t count; struct { uint32_t count; uint32_t index; } count_indexed; char name[1]; } QEMU_PACKED; struct rtas_event_log_v6_hp { #define RTAS_LOG_V6_SECTION_ID_HOTPLUG 0x4850 /* HP */ struct rtas_event_log_v6_section_header hdr; uint8_t hotplug_type; #define RTAS_LOG_V6_HP_TYPE_CPU 1 #define RTAS_LOG_V6_HP_TYPE_MEMORY 2 #define RTAS_LOG_V6_HP_TYPE_SLOT 3 #define RTAS_LOG_V6_HP_TYPE_PHB 4 #define RTAS_LOG_V6_HP_TYPE_PCI 5 uint8_t hotplug_action; #define RTAS_LOG_V6_HP_ACTION_ADD 1 #define RTAS_LOG_V6_HP_ACTION_REMOVE 2 uint8_t hotplug_identifier; #define RTAS_LOG_V6_HP_ID_DRC_NAME 1 #define RTAS_LOG_V6_HP_ID_DRC_INDEX 2 #define RTAS_LOG_V6_HP_ID_DRC_COUNT 3 #define RTAS_LOG_V6_HP_ID_DRC_COUNT_INDEXED 4 uint8_t reserved; union drc_identifier drc_id; } QEMU_PACKED; struct hp_extended_log { struct rtas_event_log_v6 v6hdr; struct rtas_event_log_v6_maina maina; struct rtas_event_log_v6_mainb mainb; struct rtas_event_log_v6_hp hp; } QEMU_PACKED; typedef enum EventClass { EVENT_CLASS_INTERNAL_ERRORS = 0, EVENT_CLASS_EPOW = 1, EVENT_CLASS_RESERVED = 2, EVENT_CLASS_HOT_PLUG = 3, EVENT_CLASS_IO = 4, EVENT_CLASS_MAX } EventClassIndex; #define EVENT_CLASS_MASK(index) (1 << (31 - index)) static const char * const event_names[EVENT_CLASS_MAX] = { [EVENT_CLASS_INTERNAL_ERRORS] = "internal-errors", [EVENT_CLASS_EPOW] = "epow-events", [EVENT_CLASS_HOT_PLUG] = "hot-plug-events", [EVENT_CLASS_IO] = "ibm,io-events", }; struct sPAPREventSource { int irq; uint32_t mask; bool enabled; }; static sPAPREventSource *spapr_event_sources_new(void) { return g_new0(sPAPREventSource, EVENT_CLASS_MAX); } static void spapr_event_sources_register(sPAPREventSource *event_sources, EventClassIndex index, int irq) { /* we only support 1 irq per event class at the moment */ g_assert(event_sources); g_assert(!event_sources[index].enabled); event_sources[index].irq = irq; event_sources[index].mask = EVENT_CLASS_MASK(index); event_sources[index].enabled = true; } static const sPAPREventSource * spapr_event_sources_get_source(sPAPREventSource *event_sources, EventClassIndex index) { g_assert(index < EVENT_CLASS_MAX); g_assert(event_sources); return &event_sources[index]; } void spapr_dt_events(sPAPRMachineState *spapr, void *fdt) { uint32_t irq_ranges[EVENT_CLASS_MAX * 2]; int i, count = 0, event_sources; sPAPREventSource *events = spapr->event_sources; g_assert(events); _FDT(event_sources = fdt_add_subnode(fdt, 0, "event-sources")); for (i = 0, count = 0; i < EVENT_CLASS_MAX; i++) { int node_offset; uint32_t interrupts[2]; const sPAPREventSource *source = spapr_event_sources_get_source(events, i); const char *source_name = event_names[i]; if (!source->enabled) { continue; } spapr_dt_xics_irq(interrupts, source->irq, false); _FDT(node_offset = fdt_add_subnode(fdt, event_sources, source_name)); _FDT(fdt_setprop(fdt, node_offset, "interrupts", interrupts, sizeof(interrupts))); irq_ranges[count++] = interrupts[0]; irq_ranges[count++] = cpu_to_be32(1); } _FDT((fdt_setprop(fdt, event_sources, "interrupt-controller", NULL, 0))); _FDT((fdt_setprop_cell(fdt, event_sources, "#interrupt-cells", 2))); _FDT((fdt_setprop(fdt, event_sources, "interrupt-ranges", irq_ranges, count * sizeof(uint32_t)))); } static const sPAPREventSource * rtas_event_log_to_source(sPAPRMachineState *spapr, int log_type) { const sPAPREventSource *source; g_assert(spapr->event_sources); switch (log_type) { case RTAS_LOG_TYPE_HOTPLUG: source = spapr_event_sources_get_source(spapr->event_sources, EVENT_CLASS_HOT_PLUG); if (spapr_ovec_test(spapr->ov5_cas, OV5_HP_EVT)) { g_assert(source->enabled); break; } /* fall back to epow for legacy hotplug interrupt source */ case RTAS_LOG_TYPE_EPOW: source = spapr_event_sources_get_source(spapr->event_sources, EVENT_CLASS_EPOW); break; default: source = NULL; } return source; } static int rtas_event_log_to_irq(sPAPRMachineState *spapr, int log_type) { const sPAPREventSource *source; source = rtas_event_log_to_source(spapr, log_type); g_assert(source); g_assert(source->enabled); return source->irq; } static uint32_t spapr_event_log_entry_type(sPAPREventLogEntry *entry) { return entry->summary & RTAS_LOG_TYPE_MASK; } static void rtas_event_log_queue(sPAPRMachineState *spapr, sPAPREventLogEntry *entry) { QTAILQ_INSERT_TAIL(&spapr->pending_events, entry, next); } static sPAPREventLogEntry *rtas_event_log_dequeue(sPAPRMachineState *spapr, uint32_t event_mask) { sPAPREventLogEntry *entry = NULL; QTAILQ_FOREACH(entry, &spapr->pending_events, next) { const sPAPREventSource *source = rtas_event_log_to_source(spapr, spapr_event_log_entry_type(entry)); if (source->mask & event_mask) { break; } } if (entry) { QTAILQ_REMOVE(&spapr->pending_events, entry, next); } return entry; } static bool rtas_event_log_contains(uint32_t event_mask) { sPAPRMachineState *spapr = SPAPR_MACHINE(qdev_get_machine()); sPAPREventLogEntry *entry = NULL; QTAILQ_FOREACH(entry, &spapr->pending_events, next) { const sPAPREventSource *source = rtas_event_log_to_source(spapr, spapr_event_log_entry_type(entry)); if (source->mask & event_mask) { return true; } } return false; } static uint32_t next_plid; static void spapr_init_v6hdr(struct rtas_event_log_v6 *v6hdr) { v6hdr->b0 = RTAS_LOG_V6_B0_VALID | RTAS_LOG_V6_B0_NEW_LOG | RTAS_LOG_V6_B0_BIGENDIAN; v6hdr->b2 = RTAS_LOG_V6_B2_POWERPC_FORMAT | RTAS_LOG_V6_B2_LOG_FORMAT_PLATFORM_EVENT; v6hdr->company = cpu_to_be32(RTAS_LOG_V6_COMPANY_IBM); } static void spapr_init_maina(struct rtas_event_log_v6_maina *maina, int section_count) { sPAPRMachineState *spapr = SPAPR_MACHINE(qdev_get_machine()); struct tm tm; int year; maina->hdr.section_id = cpu_to_be16(RTAS_LOG_V6_SECTION_ID_MAINA); maina->hdr.section_length = cpu_to_be16(sizeof(*maina)); /* FIXME: section version, subtype and creator id? */ spapr_rtc_read(&spapr->rtc, &tm, NULL); year = tm.tm_year + 1900; maina->creation_date = cpu_to_be32((to_bcd(year / 100) << 24) | (to_bcd(year % 100) << 16) | (to_bcd(tm.tm_mon + 1) << 8) | to_bcd(tm.tm_mday)); maina->creation_time = cpu_to_be32((to_bcd(tm.tm_hour) << 24) | (to_bcd(tm.tm_min) << 16) | (to_bcd(tm.tm_sec) << 8)); maina->creator_id = 'H'; /* Hypervisor */ maina->section_count = section_count; maina->plid = next_plid++; } static void spapr_powerdown_req(Notifier *n, void *opaque) { sPAPRMachineState *spapr = SPAPR_MACHINE(qdev_get_machine()); sPAPREventLogEntry *entry; struct rtas_event_log_v6 *v6hdr; struct rtas_event_log_v6_maina *maina; struct rtas_event_log_v6_mainb *mainb; struct rtas_event_log_v6_epow *epow; struct epow_extended_log *new_epow; entry = g_new(sPAPREventLogEntry, 1); new_epow = g_malloc0(sizeof(*new_epow)); entry->extended_log = new_epow; v6hdr = &new_epow->v6hdr; maina = &new_epow->maina; mainb = &new_epow->mainb; epow = &new_epow->epow; entry->summary = RTAS_LOG_VERSION_6 | RTAS_LOG_SEVERITY_EVENT | RTAS_LOG_DISPOSITION_NOT_RECOVERED | RTAS_LOG_OPTIONAL_PART_PRESENT | RTAS_LOG_TYPE_EPOW; entry->extended_length = sizeof(*new_epow); spapr_init_v6hdr(v6hdr); spapr_init_maina(maina, 3 /* Main-A, Main-B and EPOW */); mainb->hdr.section_id = cpu_to_be16(RTAS_LOG_V6_SECTION_ID_MAINB); mainb->hdr.section_length = cpu_to_be16(sizeof(*mainb)); /* FIXME: section version, subtype and creator id? */ mainb->subsystem_id = 0xa0; /* External environment */ mainb->event_severity = 0x00; /* Informational / non-error */ mainb->event_subtype = 0xd0; /* Normal shutdown */ epow->hdr.section_id = cpu_to_be16(RTAS_LOG_V6_SECTION_ID_EPOW); epow->hdr.section_length = cpu_to_be16(sizeof(*epow)); epow->hdr.section_version = 2; /* includes extended modifier */ /* FIXME: section subtype and creator id? */ epow->sensor_value = RTAS_LOG_V6_EPOW_ACTION_SYSTEM_SHUTDOWN; epow->event_modifier = RTAS_LOG_V6_EPOW_MODIFIER_NORMAL; epow->extended_modifier = RTAS_LOG_V6_EPOW_XMODIFIER_PARTITION_SPECIFIC; rtas_event_log_queue(spapr, entry); qemu_irq_pulse(spapr_qirq(spapr, rtas_event_log_to_irq(spapr, RTAS_LOG_TYPE_EPOW))); } static void spapr_hotplug_req_event(uint8_t hp_id, uint8_t hp_action, sPAPRDRConnectorType drc_type, union drc_identifier *drc_id) { sPAPRMachineState *spapr = SPAPR_MACHINE(qdev_get_machine()); sPAPREventLogEntry *entry; struct hp_extended_log *new_hp; struct rtas_event_log_v6 *v6hdr; struct rtas_event_log_v6_maina *maina; struct rtas_event_log_v6_mainb *mainb; struct rtas_event_log_v6_hp *hp; entry = g_new(sPAPREventLogEntry, 1); new_hp = g_malloc0(sizeof(struct hp_extended_log)); entry->extended_log = new_hp; v6hdr = &new_hp->v6hdr; maina = &new_hp->maina; mainb = &new_hp->mainb; hp = &new_hp->hp; entry->summary = RTAS_LOG_VERSION_6 | RTAS_LOG_SEVERITY_EVENT | RTAS_LOG_DISPOSITION_NOT_RECOVERED | RTAS_LOG_OPTIONAL_PART_PRESENT | RTAS_LOG_INITIATOR_HOTPLUG | RTAS_LOG_TYPE_HOTPLUG; entry->extended_length = sizeof(*new_hp); spapr_init_v6hdr(v6hdr); spapr_init_maina(maina, 3 /* Main-A, Main-B, HP */); mainb->hdr.section_id = cpu_to_be16(RTAS_LOG_V6_SECTION_ID_MAINB); mainb->hdr.section_length = cpu_to_be16(sizeof(*mainb)); mainb->subsystem_id = 0x80; /* External environment */ mainb->event_severity = 0x00; /* Informational / non-error */ mainb->event_subtype = 0x00; /* Normal shutdown */ hp->hdr.section_id = cpu_to_be16(RTAS_LOG_V6_SECTION_ID_HOTPLUG); hp->hdr.section_length = cpu_to_be16(sizeof(*hp)); hp->hdr.section_version = 1; /* includes extended modifier */ hp->hotplug_action = hp_action; hp->hotplug_identifier = hp_id; switch (drc_type) { case SPAPR_DR_CONNECTOR_TYPE_PCI: hp->hotplug_type = RTAS_LOG_V6_HP_TYPE_PCI; break; case SPAPR_DR_CONNECTOR_TYPE_LMB: hp->hotplug_type = RTAS_LOG_V6_HP_TYPE_MEMORY; break; case SPAPR_DR_CONNECTOR_TYPE_CPU: hp->hotplug_type = RTAS_LOG_V6_HP_TYPE_CPU; break; default: /* we shouldn't be signaling hotplug events for resources * that don't support them */ g_assert(false); return; } if (hp_id == RTAS_LOG_V6_HP_ID_DRC_COUNT) { hp->drc_id.count = cpu_to_be32(drc_id->count); } else if (hp_id == RTAS_LOG_V6_HP_ID_DRC_INDEX) { hp->drc_id.index = cpu_to_be32(drc_id->index); } else if (hp_id == RTAS_LOG_V6_HP_ID_DRC_COUNT_INDEXED) { /* we should not be using count_indexed value unless the guest * supports dedicated hotplug event source */ g_assert(spapr_ovec_test(spapr->ov5_cas, OV5_HP_EVT)); hp->drc_id.count_indexed.count = cpu_to_be32(drc_id->count_indexed.count); hp->drc_id.count_indexed.index = cpu_to_be32(drc_id->count_indexed.index); } rtas_event_log_queue(spapr, entry); qemu_irq_pulse(spapr_qirq(spapr, rtas_event_log_to_irq(spapr, RTAS_LOG_TYPE_HOTPLUG))); } void spapr_hotplug_req_add_by_index(sPAPRDRConnector *drc) { sPAPRDRConnectorType drc_type = spapr_drc_type(drc); union drc_identifier drc_id; drc_id.index = spapr_drc_index(drc); spapr_hotplug_req_event(RTAS_LOG_V6_HP_ID_DRC_INDEX, RTAS_LOG_V6_HP_ACTION_ADD, drc_type, &drc_id); } void spapr_hotplug_req_remove_by_index(sPAPRDRConnector *drc) { sPAPRDRConnectorType drc_type = spapr_drc_type(drc); union drc_identifier drc_id; drc_id.index = spapr_drc_index(drc); spapr_hotplug_req_event(RTAS_LOG_V6_HP_ID_DRC_INDEX, RTAS_LOG_V6_HP_ACTION_REMOVE, drc_type, &drc_id); } void spapr_hotplug_req_add_by_count(sPAPRDRConnectorType drc_type, uint32_t count) { union drc_identifier drc_id; drc_id.count = count; spapr_hotplug_req_event(RTAS_LOG_V6_HP_ID_DRC_COUNT, RTAS_LOG_V6_HP_ACTION_ADD, drc_type, &drc_id); } void spapr_hotplug_req_remove_by_count(sPAPRDRConnectorType drc_type, uint32_t count) { union drc_identifier drc_id; drc_id.count = count; spapr_hotplug_req_event(RTAS_LOG_V6_HP_ID_DRC_COUNT, RTAS_LOG_V6_HP_ACTION_REMOVE, drc_type, &drc_id); } void spapr_hotplug_req_add_by_count_indexed(sPAPRDRConnectorType drc_type, uint32_t count, uint32_t index) { union drc_identifier drc_id; drc_id.count_indexed.count = count; drc_id.count_indexed.index = index; spapr_hotplug_req_event(RTAS_LOG_V6_HP_ID_DRC_COUNT_INDEXED, RTAS_LOG_V6_HP_ACTION_ADD, drc_type, &drc_id); } void spapr_hotplug_req_remove_by_count_indexed(sPAPRDRConnectorType drc_type, uint32_t count, uint32_t index) { union drc_identifier drc_id; drc_id.count_indexed.count = count; drc_id.count_indexed.index = index; spapr_hotplug_req_event(RTAS_LOG_V6_HP_ID_DRC_COUNT_INDEXED, RTAS_LOG_V6_HP_ACTION_REMOVE, drc_type, &drc_id); } static void check_exception(PowerPCCPU *cpu, sPAPRMachineState *spapr, uint32_t token, uint32_t nargs, target_ulong args, uint32_t nret, target_ulong rets) { uint32_t mask, buf, len, event_len; uint64_t xinfo; sPAPREventLogEntry *event; struct rtas_error_log header; int i; if ((nargs < 6) || (nargs > 7) || nret != 1) { rtas_st(rets, 0, RTAS_OUT_PARAM_ERROR); return; } xinfo = rtas_ld(args, 1); mask = rtas_ld(args, 2); buf = rtas_ld(args, 4); len = rtas_ld(args, 5); if (nargs == 7) { xinfo |= (uint64_t)rtas_ld(args, 6) << 32; } event = rtas_event_log_dequeue(spapr, mask); if (!event) { goto out_no_events; } event_len = event->extended_length + sizeof(header); if (event_len < len) { len = event_len; } header.summary = cpu_to_be32(event->summary); header.extended_length = cpu_to_be32(event->extended_length); cpu_physical_memory_write(buf, &header, sizeof(header)); cpu_physical_memory_write(buf + sizeof(header), event->extended_log, event->extended_length); rtas_st(rets, 0, RTAS_OUT_SUCCESS); g_free(event->extended_log); g_free(event); /* according to PAPR+, the IRQ must be left asserted, or re-asserted, if * there are still pending events to be fetched via check-exception. We * do the latter here, since our code relies on edge-triggered * interrupts. */ for (i = 0; i < EVENT_CLASS_MAX; i++) { if (rtas_event_log_contains(EVENT_CLASS_MASK(i))) { const sPAPREventSource *source = spapr_event_sources_get_source(spapr->event_sources, i); g_assert(source->enabled); qemu_irq_pulse(spapr_qirq(spapr, source->irq)); } } return; out_no_events: rtas_st(rets, 0, RTAS_OUT_NO_ERRORS_FOUND); } static void event_scan(PowerPCCPU *cpu, sPAPRMachineState *spapr, uint32_t token, uint32_t nargs, target_ulong args, uint32_t nret, target_ulong rets) { if (nargs != 4 || nret != 1) { rtas_st(rets, 0, RTAS_OUT_PARAM_ERROR); return; } rtas_st(rets, 0, RTAS_OUT_NO_ERRORS_FOUND); } void spapr_clear_pending_events(sPAPRMachineState *spapr) { sPAPREventLogEntry *entry = NULL, *next_entry; QTAILQ_FOREACH_SAFE(entry, &spapr->pending_events, next, next_entry) { QTAILQ_REMOVE(&spapr->pending_events, entry, next); g_free(entry->extended_log); g_free(entry); } } void spapr_events_init(sPAPRMachineState *spapr) { QTAILQ_INIT(&spapr->pending_events); spapr->event_sources = spapr_event_sources_new(); spapr_event_sources_register(spapr->event_sources, EVENT_CLASS_EPOW, spapr_irq_alloc(spapr, 0, false, &error_fatal)); /* NOTE: if machine supports modern/dedicated hotplug event source, * we add it to the device-tree unconditionally. This means we may * have cases where the source is enabled in QEMU, but unused by the * guest because it does not support modern hotplug events, so we * take care to rely on checking for negotiation of OV5_HP_EVT option * before attempting to use it to signal events, rather than simply * checking that it's enabled. */ if (spapr->use_hotplug_event_source) { spapr_event_sources_register(spapr->event_sources, EVENT_CLASS_HOT_PLUG, spapr_irq_alloc(spapr, 0, false, &error_fatal)); } spapr->epow_notifier.notify = spapr_powerdown_req; qemu_register_powerdown_notifier(&spapr->epow_notifier); spapr_rtas_register(RTAS_CHECK_EXCEPTION, "check-exception", check_exception); spapr_rtas_register(RTAS_EVENT_SCAN, "event-scan", event_scan); }