1 /* 2 * Arm PrimeCell PL080/PL081 DMA controller 3 * 4 * Copyright (c) 2006 CodeSourcery. 5 * Written by Paul Brook 6 * 7 * This code is licensed under the GPL. 8 */ 9 10 #include "hw/sysbus.h" 11 12 #define PL080_MAX_CHANNELS 8 13 #define PL080_CONF_E 0x1 14 #define PL080_CONF_M1 0x2 15 #define PL080_CONF_M2 0x4 16 17 #define PL080_CCONF_H 0x40000 18 #define PL080_CCONF_A 0x20000 19 #define PL080_CCONF_L 0x10000 20 #define PL080_CCONF_ITC 0x08000 21 #define PL080_CCONF_IE 0x04000 22 #define PL080_CCONF_E 0x00001 23 24 #define PL080_CCTRL_I 0x80000000 25 #define PL080_CCTRL_DI 0x08000000 26 #define PL080_CCTRL_SI 0x04000000 27 #define PL080_CCTRL_D 0x02000000 28 #define PL080_CCTRL_S 0x01000000 29 30 typedef struct { 31 uint32_t src; 32 uint32_t dest; 33 uint32_t lli; 34 uint32_t ctrl; 35 uint32_t conf; 36 } pl080_channel; 37 38 #define TYPE_PL080 "pl080" 39 #define PL080(obj) OBJECT_CHECK(PL080State, (obj), TYPE_PL080) 40 41 typedef struct PL080State { 42 SysBusDevice parent_obj; 43 44 MemoryRegion iomem; 45 uint8_t tc_int; 46 uint8_t tc_mask; 47 uint8_t err_int; 48 uint8_t err_mask; 49 uint32_t conf; 50 uint32_t sync; 51 uint32_t req_single; 52 uint32_t req_burst; 53 pl080_channel chan[PL080_MAX_CHANNELS]; 54 int nchannels; 55 /* Flag to avoid recursive DMA invocations. */ 56 int running; 57 qemu_irq irq; 58 } PL080State; 59 60 static const VMStateDescription vmstate_pl080_channel = { 61 .name = "pl080_channel", 62 .version_id = 1, 63 .minimum_version_id = 1, 64 .fields = (VMStateField[]) { 65 VMSTATE_UINT32(src, pl080_channel), 66 VMSTATE_UINT32(dest, pl080_channel), 67 VMSTATE_UINT32(lli, pl080_channel), 68 VMSTATE_UINT32(ctrl, pl080_channel), 69 VMSTATE_UINT32(conf, pl080_channel), 70 VMSTATE_END_OF_LIST() 71 } 72 }; 73 74 static const VMStateDescription vmstate_pl080 = { 75 .name = "pl080", 76 .version_id = 1, 77 .minimum_version_id = 1, 78 .fields = (VMStateField[]) { 79 VMSTATE_UINT8(tc_int, PL080State), 80 VMSTATE_UINT8(tc_mask, PL080State), 81 VMSTATE_UINT8(err_int, PL080State), 82 VMSTATE_UINT8(err_mask, PL080State), 83 VMSTATE_UINT32(conf, PL080State), 84 VMSTATE_UINT32(sync, PL080State), 85 VMSTATE_UINT32(req_single, PL080State), 86 VMSTATE_UINT32(req_burst, PL080State), 87 VMSTATE_UINT8(tc_int, PL080State), 88 VMSTATE_UINT8(tc_int, PL080State), 89 VMSTATE_UINT8(tc_int, PL080State), 90 VMSTATE_STRUCT_ARRAY(chan, PL080State, PL080_MAX_CHANNELS, 91 1, vmstate_pl080_channel, pl080_channel), 92 VMSTATE_INT32(running, PL080State), 93 VMSTATE_END_OF_LIST() 94 } 95 }; 96 97 static const unsigned char pl080_id[] = 98 { 0x80, 0x10, 0x04, 0x0a, 0x0d, 0xf0, 0x05, 0xb1 }; 99 100 static const unsigned char pl081_id[] = 101 { 0x81, 0x10, 0x04, 0x0a, 0x0d, 0xf0, 0x05, 0xb1 }; 102 103 static void pl080_update(PL080State *s) 104 { 105 if ((s->tc_int & s->tc_mask) 106 || (s->err_int & s->err_mask)) 107 qemu_irq_raise(s->irq); 108 else 109 qemu_irq_lower(s->irq); 110 } 111 112 static void pl080_run(PL080State *s) 113 { 114 int c; 115 int flow; 116 pl080_channel *ch; 117 int swidth; 118 int dwidth; 119 int xsize; 120 int n; 121 int src_id; 122 int dest_id; 123 int size; 124 uint8_t buff[4]; 125 uint32_t req; 126 127 s->tc_mask = 0; 128 for (c = 0; c < s->nchannels; c++) { 129 if (s->chan[c].conf & PL080_CCONF_ITC) 130 s->tc_mask |= 1 << c; 131 if (s->chan[c].conf & PL080_CCONF_IE) 132 s->err_mask |= 1 << c; 133 } 134 135 if ((s->conf & PL080_CONF_E) == 0) 136 return; 137 138 hw_error("DMA active\n"); 139 /* If we are already in the middle of a DMA operation then indicate that 140 there may be new DMA requests and return immediately. */ 141 if (s->running) { 142 s->running++; 143 return; 144 } 145 s->running = 1; 146 while (s->running) { 147 for (c = 0; c < s->nchannels; c++) { 148 ch = &s->chan[c]; 149 again: 150 /* Test if thiws channel has any pending DMA requests. */ 151 if ((ch->conf & (PL080_CCONF_H | PL080_CCONF_E)) 152 != PL080_CCONF_E) 153 continue; 154 flow = (ch->conf >> 11) & 7; 155 if (flow >= 4) { 156 hw_error( 157 "pl080_run: Peripheral flow control not implemented\n"); 158 } 159 src_id = (ch->conf >> 1) & 0x1f; 160 dest_id = (ch->conf >> 6) & 0x1f; 161 size = ch->ctrl & 0xfff; 162 req = s->req_single | s->req_burst; 163 switch (flow) { 164 case 0: 165 break; 166 case 1: 167 if ((req & (1u << dest_id)) == 0) 168 size = 0; 169 break; 170 case 2: 171 if ((req & (1u << src_id)) == 0) 172 size = 0; 173 break; 174 case 3: 175 if ((req & (1u << src_id)) == 0 176 || (req & (1u << dest_id)) == 0) 177 size = 0; 178 break; 179 } 180 if (!size) 181 continue; 182 183 /* Transfer one element. */ 184 /* ??? Should transfer multiple elements for a burst request. */ 185 /* ??? Unclear what the proper behavior is when source and 186 destination widths are different. */ 187 swidth = 1 << ((ch->ctrl >> 18) & 7); 188 dwidth = 1 << ((ch->ctrl >> 21) & 7); 189 for (n = 0; n < dwidth; n+= swidth) { 190 cpu_physical_memory_read(ch->src, buff + n, swidth); 191 if (ch->ctrl & PL080_CCTRL_SI) 192 ch->src += swidth; 193 } 194 xsize = (dwidth < swidth) ? swidth : dwidth; 195 /* ??? This may pad the value incorrectly for dwidth < 32. */ 196 for (n = 0; n < xsize; n += dwidth) { 197 cpu_physical_memory_write(ch->dest + n, buff + n, dwidth); 198 if (ch->ctrl & PL080_CCTRL_DI) 199 ch->dest += swidth; 200 } 201 202 size--; 203 ch->ctrl = (ch->ctrl & 0xfffff000) | size; 204 if (size == 0) { 205 /* Transfer complete. */ 206 if (ch->lli) { 207 ch->src = ldl_le_phys(ch->lli); 208 ch->dest = ldl_le_phys(ch->lli + 4); 209 ch->ctrl = ldl_le_phys(ch->lli + 12); 210 ch->lli = ldl_le_phys(ch->lli + 8); 211 } else { 212 ch->conf &= ~PL080_CCONF_E; 213 } 214 if (ch->ctrl & PL080_CCTRL_I) { 215 s->tc_int |= 1 << c; 216 } 217 } 218 goto again; 219 } 220 if (--s->running) 221 s->running = 1; 222 } 223 } 224 225 static uint64_t pl080_read(void *opaque, hwaddr offset, 226 unsigned size) 227 { 228 PL080State *s = (PL080State *)opaque; 229 uint32_t i; 230 uint32_t mask; 231 232 if (offset >= 0xfe0 && offset < 0x1000) { 233 if (s->nchannels == 8) { 234 return pl080_id[(offset - 0xfe0) >> 2]; 235 } else { 236 return pl081_id[(offset - 0xfe0) >> 2]; 237 } 238 } 239 if (offset >= 0x100 && offset < 0x200) { 240 i = (offset & 0xe0) >> 5; 241 if (i >= s->nchannels) 242 goto bad_offset; 243 switch (offset >> 2) { 244 case 0: /* SrcAddr */ 245 return s->chan[i].src; 246 case 1: /* DestAddr */ 247 return s->chan[i].dest; 248 case 2: /* LLI */ 249 return s->chan[i].lli; 250 case 3: /* Control */ 251 return s->chan[i].ctrl; 252 case 4: /* Configuration */ 253 return s->chan[i].conf; 254 default: 255 goto bad_offset; 256 } 257 } 258 switch (offset >> 2) { 259 case 0: /* IntStatus */ 260 return (s->tc_int & s->tc_mask) | (s->err_int & s->err_mask); 261 case 1: /* IntTCStatus */ 262 return (s->tc_int & s->tc_mask); 263 case 3: /* IntErrorStatus */ 264 return (s->err_int & s->err_mask); 265 case 5: /* RawIntTCStatus */ 266 return s->tc_int; 267 case 6: /* RawIntErrorStatus */ 268 return s->err_int; 269 case 7: /* EnbldChns */ 270 mask = 0; 271 for (i = 0; i < s->nchannels; i++) { 272 if (s->chan[i].conf & PL080_CCONF_E) 273 mask |= 1 << i; 274 } 275 return mask; 276 case 8: /* SoftBReq */ 277 case 9: /* SoftSReq */ 278 case 10: /* SoftLBReq */ 279 case 11: /* SoftLSReq */ 280 /* ??? Implement these. */ 281 return 0; 282 case 12: /* Configuration */ 283 return s->conf; 284 case 13: /* Sync */ 285 return s->sync; 286 default: 287 bad_offset: 288 qemu_log_mask(LOG_GUEST_ERROR, 289 "pl080_read: Bad offset %x\n", (int)offset); 290 return 0; 291 } 292 } 293 294 static void pl080_write(void *opaque, hwaddr offset, 295 uint64_t value, unsigned size) 296 { 297 PL080State *s = (PL080State *)opaque; 298 int i; 299 300 if (offset >= 0x100 && offset < 0x200) { 301 i = (offset & 0xe0) >> 5; 302 if (i >= s->nchannels) 303 goto bad_offset; 304 switch (offset >> 2) { 305 case 0: /* SrcAddr */ 306 s->chan[i].src = value; 307 break; 308 case 1: /* DestAddr */ 309 s->chan[i].dest = value; 310 break; 311 case 2: /* LLI */ 312 s->chan[i].lli = value; 313 break; 314 case 3: /* Control */ 315 s->chan[i].ctrl = value; 316 break; 317 case 4: /* Configuration */ 318 s->chan[i].conf = value; 319 pl080_run(s); 320 break; 321 } 322 } 323 switch (offset >> 2) { 324 case 2: /* IntTCClear */ 325 s->tc_int &= ~value; 326 break; 327 case 4: /* IntErrorClear */ 328 s->err_int &= ~value; 329 break; 330 case 8: /* SoftBReq */ 331 case 9: /* SoftSReq */ 332 case 10: /* SoftLBReq */ 333 case 11: /* SoftLSReq */ 334 /* ??? Implement these. */ 335 qemu_log_mask(LOG_UNIMP, "pl080_write: Soft DMA not implemented\n"); 336 break; 337 case 12: /* Configuration */ 338 s->conf = value; 339 if (s->conf & (PL080_CONF_M1 | PL080_CONF_M1)) { 340 qemu_log_mask(LOG_UNIMP, 341 "pl080_write: Big-endian DMA not implemented\n"); 342 } 343 pl080_run(s); 344 break; 345 case 13: /* Sync */ 346 s->sync = value; 347 break; 348 default: 349 bad_offset: 350 qemu_log_mask(LOG_GUEST_ERROR, 351 "pl080_write: Bad offset %x\n", (int)offset); 352 } 353 pl080_update(s); 354 } 355 356 static const MemoryRegionOps pl080_ops = { 357 .read = pl080_read, 358 .write = pl080_write, 359 .endianness = DEVICE_NATIVE_ENDIAN, 360 }; 361 362 static void pl080_init(Object *obj) 363 { 364 SysBusDevice *sbd = SYS_BUS_DEVICE(obj); 365 PL080State *s = PL080(obj); 366 367 memory_region_init_io(&s->iomem, OBJECT(s), &pl080_ops, s, "pl080", 0x1000); 368 sysbus_init_mmio(sbd, &s->iomem); 369 sysbus_init_irq(sbd, &s->irq); 370 s->nchannels = 8; 371 } 372 373 static void pl081_init(Object *obj) 374 { 375 PL080State *s = PL080(obj); 376 377 s->nchannels = 2; 378 } 379 380 static void pl080_class_init(ObjectClass *oc, void *data) 381 { 382 DeviceClass *dc = DEVICE_CLASS(oc); 383 384 dc->no_user = 1; 385 dc->vmsd = &vmstate_pl080; 386 } 387 388 static const TypeInfo pl080_info = { 389 .name = TYPE_PL080, 390 .parent = TYPE_SYS_BUS_DEVICE, 391 .instance_size = sizeof(PL080State), 392 .instance_init = pl080_init, 393 .class_init = pl080_class_init, 394 }; 395 396 static const TypeInfo pl081_info = { 397 .name = "pl081", 398 .parent = TYPE_PL080, 399 .instance_init = pl081_init, 400 }; 401 402 /* The PL080 and PL081 are the same except for the number of channels 403 they implement (8 and 2 respectively). */ 404 static void pl080_register_types(void) 405 { 406 type_register_static(&pl080_info); 407 type_register_static(&pl081_info); 408 } 409 410 type_init(pl080_register_types) 411