1 /*
2 * ASPEED XDMA Controller
3 * Eddie James <eajames@linux.ibm.com>
4 *
5 * Copyright (C) 2019 IBM Corp
6 * SPDX-License-Identifier: GPL-2.0-or-later
7 */
8
9 #include "qemu/osdep.h"
10 #include "qemu/log.h"
11 #include "qemu/error-report.h"
12 #include "hw/irq.h"
13 #include "hw/misc/aspeed_xdma.h"
14 #include "migration/vmstate.h"
15 #include "qapi/error.h"
16
17 #include "trace.h"
18
19 #define XDMA_BMC_CMDQ_ADDR 0x10
20 #define XDMA_BMC_CMDQ_ENDP 0x14
21 #define XDMA_BMC_CMDQ_WRP 0x18
22 #define XDMA_BMC_CMDQ_W_MASK 0x0003FFFF
23 #define XDMA_BMC_CMDQ_RDP 0x1C
24 #define XDMA_BMC_CMDQ_RDP_MAGIC 0xEE882266
25 #define XDMA_IRQ_ENG_CTRL 0x20
26 #define XDMA_IRQ_ENG_CTRL_US_COMP BIT(4)
27 #define XDMA_IRQ_ENG_CTRL_DS_COMP BIT(5)
28 #define XDMA_IRQ_ENG_CTRL_W_MASK 0xBFEFF07F
29 #define XDMA_IRQ_ENG_STAT 0x24
30 #define XDMA_IRQ_ENG_STAT_US_COMP BIT(4)
31 #define XDMA_IRQ_ENG_STAT_DS_COMP BIT(5)
32 #define XDMA_IRQ_ENG_STAT_RESET 0xF8000000
33
34 #define XDMA_AST2600_BMC_CMDQ_ADDR 0x14
35 #define XDMA_AST2600_BMC_CMDQ_ENDP 0x18
36 #define XDMA_AST2600_BMC_CMDQ_WRP 0x1c
37 #define XDMA_AST2600_BMC_CMDQ_RDP 0x20
38 #define XDMA_AST2600_IRQ_CTRL 0x38
39 #define XDMA_AST2600_IRQ_CTRL_US_COMP BIT(16)
40 #define XDMA_AST2600_IRQ_CTRL_DS_COMP BIT(17)
41 #define XDMA_AST2600_IRQ_CTRL_W_MASK 0x017003FF
42 #define XDMA_AST2600_IRQ_STATUS 0x3c
43 #define XDMA_AST2600_IRQ_STATUS_US_COMP BIT(16)
44 #define XDMA_AST2600_IRQ_STATUS_DS_COMP BIT(17)
45
46 #define XDMA_MEM_SIZE 0x1000
47
48 #define TO_REG(addr) ((addr) / sizeof(uint32_t))
49
aspeed_xdma_read(void * opaque,hwaddr addr,unsigned int size)50 static uint64_t aspeed_xdma_read(void *opaque, hwaddr addr, unsigned int size)
51 {
52 uint32_t val = 0;
53 AspeedXDMAState *xdma = opaque;
54
55 if (addr < ASPEED_XDMA_REG_SIZE) {
56 val = xdma->regs[TO_REG(addr)];
57 }
58
59 return (uint64_t)val;
60 }
61
aspeed_xdma_write(void * opaque,hwaddr addr,uint64_t val,unsigned int size)62 static void aspeed_xdma_write(void *opaque, hwaddr addr, uint64_t val,
63 unsigned int size)
64 {
65 unsigned int idx;
66 uint32_t val32 = (uint32_t)val;
67 AspeedXDMAState *xdma = opaque;
68 AspeedXDMAClass *axc = ASPEED_XDMA_GET_CLASS(xdma);
69
70 if (addr >= ASPEED_XDMA_REG_SIZE) {
71 return;
72 }
73
74 if (addr == axc->cmdq_endp) {
75 xdma->regs[TO_REG(addr)] = val32 & XDMA_BMC_CMDQ_W_MASK;
76 } else if (addr == axc->cmdq_wrp) {
77 idx = TO_REG(addr);
78 xdma->regs[idx] = val32 & XDMA_BMC_CMDQ_W_MASK;
79 xdma->regs[TO_REG(axc->cmdq_rdp)] = xdma->regs[idx];
80
81 trace_aspeed_xdma_write(addr, val);
82
83 if (xdma->bmc_cmdq_readp_set) {
84 xdma->bmc_cmdq_readp_set = 0;
85 } else {
86 xdma->regs[TO_REG(axc->intr_status)] |= axc->intr_complete;
87
88 if (xdma->regs[TO_REG(axc->intr_ctrl)] & axc->intr_complete) {
89 qemu_irq_raise(xdma->irq);
90 }
91 }
92 } else if (addr == axc->cmdq_rdp) {
93 trace_aspeed_xdma_write(addr, val);
94
95 if (val32 == XDMA_BMC_CMDQ_RDP_MAGIC) {
96 xdma->bmc_cmdq_readp_set = 1;
97 }
98 } else if (addr == axc->intr_ctrl) {
99 xdma->regs[TO_REG(addr)] = val32 & axc->intr_ctrl_mask;
100 } else if (addr == axc->intr_status) {
101 trace_aspeed_xdma_write(addr, val);
102
103 idx = TO_REG(addr);
104 if (val32 & axc->intr_complete) {
105 xdma->regs[idx] &= ~axc->intr_complete;
106 qemu_irq_lower(xdma->irq);
107 }
108 } else {
109 xdma->regs[TO_REG(addr)] = val32;
110 }
111 }
112
113 static const MemoryRegionOps aspeed_xdma_ops = {
114 .read = aspeed_xdma_read,
115 .write = aspeed_xdma_write,
116 .endianness = DEVICE_NATIVE_ENDIAN,
117 .valid.min_access_size = 4,
118 .valid.max_access_size = 4,
119 };
120
aspeed_xdma_realize(DeviceState * dev,Error ** errp)121 static void aspeed_xdma_realize(DeviceState *dev, Error **errp)
122 {
123 SysBusDevice *sbd = SYS_BUS_DEVICE(dev);
124 AspeedXDMAState *xdma = ASPEED_XDMA(dev);
125
126 sysbus_init_irq(sbd, &xdma->irq);
127 memory_region_init_io(&xdma->iomem, OBJECT(xdma), &aspeed_xdma_ops, xdma,
128 TYPE_ASPEED_XDMA, XDMA_MEM_SIZE);
129 sysbus_init_mmio(sbd, &xdma->iomem);
130 }
131
aspeed_xdma_reset(DeviceState * dev)132 static void aspeed_xdma_reset(DeviceState *dev)
133 {
134 AspeedXDMAState *xdma = ASPEED_XDMA(dev);
135 AspeedXDMAClass *axc = ASPEED_XDMA_GET_CLASS(xdma);
136
137 xdma->bmc_cmdq_readp_set = 0;
138 memset(xdma->regs, 0, ASPEED_XDMA_REG_SIZE);
139 xdma->regs[TO_REG(axc->intr_status)] = XDMA_IRQ_ENG_STAT_RESET;
140
141 qemu_irq_lower(xdma->irq);
142 }
143
144 static const VMStateDescription aspeed_xdma_vmstate = {
145 .name = TYPE_ASPEED_XDMA,
146 .version_id = 1,
147 .fields = (const VMStateField[]) {
148 VMSTATE_UINT32_ARRAY(regs, AspeedXDMAState, ASPEED_XDMA_NUM_REGS),
149 VMSTATE_END_OF_LIST(),
150 },
151 };
152
aspeed_2600_xdma_class_init(ObjectClass * klass,void * data)153 static void aspeed_2600_xdma_class_init(ObjectClass *klass, void *data)
154 {
155 DeviceClass *dc = DEVICE_CLASS(klass);
156 AspeedXDMAClass *axc = ASPEED_XDMA_CLASS(klass);
157
158 dc->desc = "ASPEED 2600 XDMA Controller";
159
160 axc->cmdq_endp = XDMA_AST2600_BMC_CMDQ_ENDP;
161 axc->cmdq_wrp = XDMA_AST2600_BMC_CMDQ_WRP;
162 axc->cmdq_rdp = XDMA_AST2600_BMC_CMDQ_RDP;
163 axc->intr_ctrl = XDMA_AST2600_IRQ_CTRL;
164 axc->intr_ctrl_mask = XDMA_AST2600_IRQ_CTRL_W_MASK;
165 axc->intr_status = XDMA_AST2600_IRQ_STATUS;
166 axc->intr_complete = XDMA_AST2600_IRQ_STATUS_US_COMP |
167 XDMA_AST2600_IRQ_STATUS_DS_COMP;
168 }
169
170 static const TypeInfo aspeed_2600_xdma_info = {
171 .name = TYPE_ASPEED_2600_XDMA,
172 .parent = TYPE_ASPEED_XDMA,
173 .class_init = aspeed_2600_xdma_class_init,
174 };
175
aspeed_2500_xdma_class_init(ObjectClass * klass,void * data)176 static void aspeed_2500_xdma_class_init(ObjectClass *klass, void *data)
177 {
178 DeviceClass *dc = DEVICE_CLASS(klass);
179 AspeedXDMAClass *axc = ASPEED_XDMA_CLASS(klass);
180
181 dc->desc = "ASPEED 2500 XDMA Controller";
182
183 axc->cmdq_endp = XDMA_BMC_CMDQ_ENDP;
184 axc->cmdq_wrp = XDMA_BMC_CMDQ_WRP;
185 axc->cmdq_rdp = XDMA_BMC_CMDQ_RDP;
186 axc->intr_ctrl = XDMA_IRQ_ENG_CTRL;
187 axc->intr_ctrl_mask = XDMA_IRQ_ENG_CTRL_W_MASK;
188 axc->intr_status = XDMA_IRQ_ENG_STAT;
189 axc->intr_complete = XDMA_IRQ_ENG_STAT_US_COMP | XDMA_IRQ_ENG_STAT_DS_COMP;
190 };
191
192 static const TypeInfo aspeed_2500_xdma_info = {
193 .name = TYPE_ASPEED_2500_XDMA,
194 .parent = TYPE_ASPEED_XDMA,
195 .class_init = aspeed_2500_xdma_class_init,
196 };
197
aspeed_2400_xdma_class_init(ObjectClass * klass,void * data)198 static void aspeed_2400_xdma_class_init(ObjectClass *klass, void *data)
199 {
200 DeviceClass *dc = DEVICE_CLASS(klass);
201 AspeedXDMAClass *axc = ASPEED_XDMA_CLASS(klass);
202
203 dc->desc = "ASPEED 2400 XDMA Controller";
204
205 axc->cmdq_endp = XDMA_BMC_CMDQ_ENDP;
206 axc->cmdq_wrp = XDMA_BMC_CMDQ_WRP;
207 axc->cmdq_rdp = XDMA_BMC_CMDQ_RDP;
208 axc->intr_ctrl = XDMA_IRQ_ENG_CTRL;
209 axc->intr_ctrl_mask = XDMA_IRQ_ENG_CTRL_W_MASK;
210 axc->intr_status = XDMA_IRQ_ENG_STAT;
211 axc->intr_complete = XDMA_IRQ_ENG_STAT_US_COMP | XDMA_IRQ_ENG_STAT_DS_COMP;
212 };
213
214 static const TypeInfo aspeed_2400_xdma_info = {
215 .name = TYPE_ASPEED_2400_XDMA,
216 .parent = TYPE_ASPEED_XDMA,
217 .class_init = aspeed_2400_xdma_class_init,
218 };
219
aspeed_xdma_class_init(ObjectClass * classp,void * data)220 static void aspeed_xdma_class_init(ObjectClass *classp, void *data)
221 {
222 DeviceClass *dc = DEVICE_CLASS(classp);
223
224 dc->realize = aspeed_xdma_realize;
225 dc->reset = aspeed_xdma_reset;
226 dc->vmsd = &aspeed_xdma_vmstate;
227 }
228
229 static const TypeInfo aspeed_xdma_info = {
230 .name = TYPE_ASPEED_XDMA,
231 .parent = TYPE_SYS_BUS_DEVICE,
232 .instance_size = sizeof(AspeedXDMAState),
233 .class_init = aspeed_xdma_class_init,
234 .class_size = sizeof(AspeedXDMAClass),
235 .abstract = true,
236 };
237
aspeed_xdma_register_type(void)238 static void aspeed_xdma_register_type(void)
239 {
240 type_register_static(&aspeed_xdma_info);
241 type_register_static(&aspeed_2400_xdma_info);
242 type_register_static(&aspeed_2500_xdma_info);
243 type_register_static(&aspeed_2600_xdma_info);
244 }
245 type_init(aspeed_xdma_register_type);
246