xref: /openbmc/qemu/hw/dma/pl080.c (revision cde3c425)
1 /*
2  * Arm PrimeCell PL080/PL081 DMA controller
3  *
4  * Copyright (c) 2006 CodeSourcery.
5  * Written by Paul Brook
6  *
7  * This code is licensed under the GPL.
8  */
9 
10 #include "qemu/osdep.h"
11 #include "hw/sysbus.h"
12 #include "migration/vmstate.h"
13 #include "qemu/log.h"
14 #include "qemu/module.h"
15 #include "hw/dma/pl080.h"
16 #include "hw/hw.h"
17 #include "hw/irq.h"
18 #include "hw/qdev-properties.h"
19 #include "qapi/error.h"
20 
21 #define PL080_CONF_E    0x1
22 #define PL080_CONF_M1   0x2
23 #define PL080_CONF_M2   0x4
24 
25 #define PL080_CCONF_H   0x40000
26 #define PL080_CCONF_A   0x20000
27 #define PL080_CCONF_L   0x10000
28 #define PL080_CCONF_ITC 0x08000
29 #define PL080_CCONF_IE  0x04000
30 #define PL080_CCONF_E   0x00001
31 
32 #define PL080_CCTRL_I   0x80000000
33 #define PL080_CCTRL_DI  0x08000000
34 #define PL080_CCTRL_SI  0x04000000
35 #define PL080_CCTRL_D   0x02000000
36 #define PL080_CCTRL_S   0x01000000
37 
38 static const VMStateDescription vmstate_pl080_channel = {
39     .name = "pl080_channel",
40     .version_id = 1,
41     .minimum_version_id = 1,
42     .fields = (const VMStateField[]) {
43         VMSTATE_UINT32(src, pl080_channel),
44         VMSTATE_UINT32(dest, pl080_channel),
45         VMSTATE_UINT32(lli, pl080_channel),
46         VMSTATE_UINT32(ctrl, pl080_channel),
47         VMSTATE_UINT32(conf, pl080_channel),
48         VMSTATE_END_OF_LIST()
49     }
50 };
51 
52 static const VMStateDescription vmstate_pl080 = {
53     .name = "pl080",
54     .version_id = 1,
55     .minimum_version_id = 1,
56     .fields = (const VMStateField[]) {
57         VMSTATE_UINT8(tc_int, PL080State),
58         VMSTATE_UINT8(tc_mask, PL080State),
59         VMSTATE_UINT8(err_int, PL080State),
60         VMSTATE_UINT8(err_mask, PL080State),
61         VMSTATE_UINT32(conf, PL080State),
62         VMSTATE_UINT32(sync, PL080State),
63         VMSTATE_UINT32(req_single, PL080State),
64         VMSTATE_UINT32(req_burst, PL080State),
65         VMSTATE_UINT8(tc_int, PL080State),
66         VMSTATE_UINT8(tc_int, PL080State),
67         VMSTATE_UINT8(tc_int, PL080State),
68         VMSTATE_STRUCT_ARRAY(chan, PL080State, PL080_MAX_CHANNELS,
69                              1, vmstate_pl080_channel, pl080_channel),
70         VMSTATE_INT32(running, PL080State),
71         VMSTATE_END_OF_LIST()
72     }
73 };
74 
75 static const unsigned char pl080_id[] =
76 { 0x80, 0x10, 0x04, 0x0a, 0x0d, 0xf0, 0x05, 0xb1 };
77 
78 static const unsigned char pl081_id[] =
79 { 0x81, 0x10, 0x04, 0x0a, 0x0d, 0xf0, 0x05, 0xb1 };
80 
81 static void pl080_update(PL080State *s)
82 {
83     bool tclevel = (s->tc_int & s->tc_mask);
84     bool errlevel = (s->err_int & s->err_mask);
85 
86     qemu_set_irq(s->interr, errlevel);
87     qemu_set_irq(s->inttc, tclevel);
88     qemu_set_irq(s->irq, errlevel || tclevel);
89 }
90 
91 static void pl080_run(PL080State *s)
92 {
93     int c;
94     int flow;
95     pl080_channel *ch;
96     int swidth;
97     int dwidth;
98     int xsize;
99     int n;
100     int src_id;
101     int dest_id;
102     int size;
103     uint8_t buff[4];
104     uint32_t req;
105 
106     s->tc_mask = 0;
107     for (c = 0; c < s->nchannels; c++) {
108         if (s->chan[c].conf & PL080_CCONF_ITC)
109             s->tc_mask |= 1 << c;
110         if (s->chan[c].conf & PL080_CCONF_IE)
111             s->err_mask |= 1 << c;
112     }
113 
114     if ((s->conf & PL080_CONF_E) == 0)
115         return;
116 
117     /* If we are already in the middle of a DMA operation then indicate that
118        there may be new DMA requests and return immediately.  */
119     if (s->running) {
120         s->running++;
121         return;
122     }
123     s->running = 1;
124     while (s->running) {
125         for (c = 0; c < s->nchannels; c++) {
126             ch = &s->chan[c];
127 again:
128             /* Test if thiws channel has any pending DMA requests.  */
129             if ((ch->conf & (PL080_CCONF_H | PL080_CCONF_E))
130                     != PL080_CCONF_E)
131                 continue;
132             flow = (ch->conf >> 11) & 7;
133             if (flow >= 4) {
134                 hw_error(
135                     "pl080_run: Peripheral flow control not implemented\n");
136             }
137             src_id = (ch->conf >> 1) & 0x1f;
138             dest_id = (ch->conf >> 6) & 0x1f;
139             size = ch->ctrl & 0xfff;
140             req = s->req_single | s->req_burst;
141             switch (flow) {
142             case 0:
143                 break;
144             case 1:
145                 if ((req & (1u << dest_id)) == 0)
146                     size = 0;
147                 break;
148             case 2:
149                 if ((req & (1u << src_id)) == 0)
150                     size = 0;
151                 break;
152             case 3:
153                 if ((req & (1u << src_id)) == 0
154                         || (req & (1u << dest_id)) == 0)
155                     size = 0;
156                 break;
157             }
158             if (!size)
159                 continue;
160 
161             /* Transfer one element.  */
162             /* ??? Should transfer multiple elements for a burst request.  */
163             /* ??? Unclear what the proper behavior is when source and
164                destination widths are different.  */
165             swidth = 1 << ((ch->ctrl >> 18) & 7);
166             dwidth = 1 << ((ch->ctrl >> 21) & 7);
167             for (n = 0; n < dwidth; n+= swidth) {
168                 address_space_read(&s->downstream_as, ch->src,
169                                    MEMTXATTRS_UNSPECIFIED, buff + n, swidth);
170                 if (ch->ctrl & PL080_CCTRL_SI)
171                     ch->src += swidth;
172             }
173             xsize = (dwidth < swidth) ? swidth : dwidth;
174             /* ??? This may pad the value incorrectly for dwidth < 32.  */
175             for (n = 0; n < xsize; n += dwidth) {
176                 address_space_write(&s->downstream_as, ch->dest + n,
177                                     MEMTXATTRS_UNSPECIFIED, buff + n, dwidth);
178                 if (ch->ctrl & PL080_CCTRL_DI)
179                     ch->dest += swidth;
180             }
181 
182             size--;
183             ch->ctrl = (ch->ctrl & 0xfffff000) | size;
184             if (size == 0) {
185                 /* Transfer complete.  */
186                 if (ch->lli) {
187                     ch->src = address_space_ldl_le(&s->downstream_as,
188                                                    ch->lli,
189                                                    MEMTXATTRS_UNSPECIFIED,
190                                                    NULL);
191                     ch->dest = address_space_ldl_le(&s->downstream_as,
192                                                     ch->lli + 4,
193                                                     MEMTXATTRS_UNSPECIFIED,
194                                                     NULL);
195                     ch->ctrl = address_space_ldl_le(&s->downstream_as,
196                                                     ch->lli + 12,
197                                                     MEMTXATTRS_UNSPECIFIED,
198                                                     NULL);
199                     ch->lli = address_space_ldl_le(&s->downstream_as,
200                                                    ch->lli + 8,
201                                                    MEMTXATTRS_UNSPECIFIED,
202                                                    NULL);
203                 } else {
204                     ch->conf &= ~PL080_CCONF_E;
205                 }
206                 if (ch->ctrl & PL080_CCTRL_I) {
207                     s->tc_int |= 1 << c;
208                 }
209             }
210             goto again;
211         }
212         if (--s->running)
213             s->running = 1;
214     }
215 }
216 
217 static uint64_t pl080_read(void *opaque, hwaddr offset,
218                            unsigned size)
219 {
220     PL080State *s = (PL080State *)opaque;
221     uint32_t i;
222     uint32_t mask;
223 
224     if (offset >= 0xfe0 && offset < 0x1000) {
225         if (s->nchannels == 8) {
226             return pl080_id[(offset - 0xfe0) >> 2];
227         } else {
228             return pl081_id[(offset - 0xfe0) >> 2];
229         }
230     }
231     if (offset >= 0x100 && offset < 0x200) {
232         i = (offset & 0xe0) >> 5;
233         if (i >= s->nchannels)
234             goto bad_offset;
235         switch ((offset >> 2) & 7) {
236         case 0: /* SrcAddr */
237             return s->chan[i].src;
238         case 1: /* DestAddr */
239             return s->chan[i].dest;
240         case 2: /* LLI */
241             return s->chan[i].lli;
242         case 3: /* Control */
243             return s->chan[i].ctrl;
244         case 4: /* Configuration */
245             return s->chan[i].conf;
246         default:
247             goto bad_offset;
248         }
249     }
250     switch (offset >> 2) {
251     case 0: /* IntStatus */
252         return (s->tc_int & s->tc_mask) | (s->err_int & s->err_mask);
253     case 1: /* IntTCStatus */
254         return (s->tc_int & s->tc_mask);
255     case 3: /* IntErrorStatus */
256         return (s->err_int & s->err_mask);
257     case 5: /* RawIntTCStatus */
258         return s->tc_int;
259     case 6: /* RawIntErrorStatus */
260         return s->err_int;
261     case 7: /* EnbldChns */
262         mask = 0;
263         for (i = 0; i < s->nchannels; i++) {
264             if (s->chan[i].conf & PL080_CCONF_E)
265                 mask |= 1 << i;
266         }
267         return mask;
268     case 8: /* SoftBReq */
269     case 9: /* SoftSReq */
270     case 10: /* SoftLBReq */
271     case 11: /* SoftLSReq */
272         /* ??? Implement these. */
273         return 0;
274     case 12: /* Configuration */
275         return s->conf;
276     case 13: /* Sync */
277         return s->sync;
278     default:
279     bad_offset:
280         qemu_log_mask(LOG_GUEST_ERROR,
281                       "pl080_read: Bad offset %x\n", (int)offset);
282         return 0;
283     }
284 }
285 
286 static void pl080_write(void *opaque, hwaddr offset,
287                         uint64_t value, unsigned size)
288 {
289     PL080State *s = (PL080State *)opaque;
290     int i;
291 
292     if (offset >= 0x100 && offset < 0x200) {
293         i = (offset & 0xe0) >> 5;
294         if (i >= s->nchannels)
295             goto bad_offset;
296         switch ((offset >> 2) & 7) {
297         case 0: /* SrcAddr */
298             s->chan[i].src = value;
299             break;
300         case 1: /* DestAddr */
301             s->chan[i].dest = value;
302             break;
303         case 2: /* LLI */
304             s->chan[i].lli = value;
305             break;
306         case 3: /* Control */
307             s->chan[i].ctrl = value;
308             break;
309         case 4: /* Configuration */
310             s->chan[i].conf = value;
311             pl080_run(s);
312             break;
313         }
314         return;
315     }
316     switch (offset >> 2) {
317     case 2: /* IntTCClear */
318         s->tc_int &= ~value;
319         break;
320     case 4: /* IntErrorClear */
321         s->err_int &= ~value;
322         break;
323     case 8: /* SoftBReq */
324     case 9: /* SoftSReq */
325     case 10: /* SoftLBReq */
326     case 11: /* SoftLSReq */
327         /* ??? Implement these.  */
328         qemu_log_mask(LOG_UNIMP, "pl080_write: Soft DMA not implemented\n");
329         break;
330     case 12: /* Configuration */
331         s->conf = value;
332         if (s->conf & (PL080_CONF_M1 | PL080_CONF_M2)) {
333             qemu_log_mask(LOG_UNIMP,
334                           "pl080_write: Big-endian DMA not implemented\n");
335         }
336         pl080_run(s);
337         break;
338     case 13: /* Sync */
339         s->sync = value;
340         break;
341     default:
342     bad_offset:
343         qemu_log_mask(LOG_GUEST_ERROR,
344                       "pl080_write: Bad offset %x\n", (int)offset);
345     }
346     pl080_update(s);
347 }
348 
349 static const MemoryRegionOps pl080_ops = {
350     .read = pl080_read,
351     .write = pl080_write,
352     .endianness = DEVICE_NATIVE_ENDIAN,
353 };
354 
355 static void pl080_reset(DeviceState *dev)
356 {
357     PL080State *s = PL080(dev);
358     int i;
359 
360     s->tc_int = 0;
361     s->tc_mask = 0;
362     s->err_int = 0;
363     s->err_mask = 0;
364     s->conf = 0;
365     s->sync = 0;
366     s->req_single = 0;
367     s->req_burst = 0;
368     s->running = 0;
369 
370     for (i = 0; i < s->nchannels; i++) {
371         s->chan[i].src = 0;
372         s->chan[i].dest = 0;
373         s->chan[i].lli = 0;
374         s->chan[i].ctrl = 0;
375         s->chan[i].conf = 0;
376     }
377 }
378 
379 static void pl080_init(Object *obj)
380 {
381     SysBusDevice *sbd = SYS_BUS_DEVICE(obj);
382     PL080State *s = PL080(obj);
383 
384     memory_region_init_io(&s->iomem, OBJECT(s), &pl080_ops, s, "pl080", 0x1000);
385     sysbus_init_mmio(sbd, &s->iomem);
386     sysbus_init_irq(sbd, &s->irq);
387     sysbus_init_irq(sbd, &s->interr);
388     sysbus_init_irq(sbd, &s->inttc);
389     s->nchannels = 8;
390 }
391 
392 static void pl080_realize(DeviceState *dev, Error **errp)
393 {
394     PL080State *s = PL080(dev);
395 
396     if (!s->downstream) {
397         error_setg(errp, "PL080 'downstream' link not set");
398         return;
399     }
400 
401     address_space_init(&s->downstream_as, s->downstream, "pl080-downstream");
402 }
403 
404 static void pl081_init(Object *obj)
405 {
406     PL080State *s = PL080(obj);
407 
408     s->nchannels = 2;
409 }
410 
411 static Property pl080_properties[] = {
412     DEFINE_PROP_LINK("downstream", PL080State, downstream,
413                      TYPE_MEMORY_REGION, MemoryRegion *),
414     DEFINE_PROP_END_OF_LIST(),
415 };
416 
417 static void pl080_class_init(ObjectClass *oc, void *data)
418 {
419     DeviceClass *dc = DEVICE_CLASS(oc);
420 
421     dc->vmsd = &vmstate_pl080;
422     dc->realize = pl080_realize;
423     device_class_set_props(dc, pl080_properties);
424     device_class_set_legacy_reset(dc, pl080_reset);
425 }
426 
427 static const TypeInfo pl080_info = {
428     .name          = TYPE_PL080,
429     .parent        = TYPE_SYS_BUS_DEVICE,
430     .instance_size = sizeof(PL080State),
431     .instance_init = pl080_init,
432     .class_init    = pl080_class_init,
433 };
434 
435 static const TypeInfo pl081_info = {
436     .name          = TYPE_PL081,
437     .parent        = TYPE_PL080,
438     .instance_init = pl081_init,
439 };
440 
441 /* The PL080 and PL081 are the same except for the number of channels
442    they implement (8 and 2 respectively).  */
443 static void pl080_register_types(void)
444 {
445     type_register_static(&pl080_info);
446     type_register_static(&pl081_info);
447 }
448 
449 type_init(pl080_register_types)
450