1 /* 2 * linux/arch/arm/mach-rpc/dma.c 3 * 4 * Copyright (C) 1998 Russell King 5 * 6 * This program is free software; you can redistribute it and/or modify 7 * it under the terms of the GNU General Public License version 2 as 8 * published by the Free Software Foundation. 9 * 10 * DMA functions specific to RiscPC architecture 11 */ 12 #include <linux/mman.h> 13 #include <linux/init.h> 14 #include <linux/interrupt.h> 15 #include <linux/dma-mapping.h> 16 #include <linux/io.h> 17 18 #include <asm/page.h> 19 #include <asm/dma.h> 20 #include <asm/fiq.h> 21 #include <asm/irq.h> 22 #include <mach/hardware.h> 23 #include <linux/uaccess.h> 24 25 #include <asm/mach/dma.h> 26 #include <asm/hardware/iomd.h> 27 28 struct iomd_dma { 29 struct dma_struct dma; 30 unsigned int state; 31 unsigned long base; /* Controller base address */ 32 int irq; /* Controller IRQ */ 33 struct scatterlist cur_sg; /* Current controller buffer */ 34 dma_addr_t dma_addr; 35 unsigned int dma_len; 36 }; 37 38 #if 0 39 typedef enum { 40 dma_size_8 = 1, 41 dma_size_16 = 2, 42 dma_size_32 = 4, 43 dma_size_128 = 16 44 } dma_size_t; 45 #endif 46 47 #define TRANSFER_SIZE 2 48 49 #define CURA (0) 50 #define ENDA (IOMD_IO0ENDA - IOMD_IO0CURA) 51 #define CURB (IOMD_IO0CURB - IOMD_IO0CURA) 52 #define ENDB (IOMD_IO0ENDB - IOMD_IO0CURA) 53 #define CR (IOMD_IO0CR - IOMD_IO0CURA) 54 #define ST (IOMD_IO0ST - IOMD_IO0CURA) 55 56 static void iomd_get_next_sg(struct scatterlist *sg, struct iomd_dma *idma) 57 { 58 unsigned long end, offset, flags = 0; 59 60 if (idma->dma.sg) { 61 sg->dma_address = idma->dma_addr; 62 offset = sg->dma_address & ~PAGE_MASK; 63 64 end = offset + idma->dma_len; 65 66 if (end > PAGE_SIZE) 67 end = PAGE_SIZE; 68 69 if (offset + TRANSFER_SIZE >= end) 70 flags |= DMA_END_L; 71 72 sg->length = end - TRANSFER_SIZE; 73 74 idma->dma_len -= end - offset; 75 idma->dma_addr += end - offset; 76 77 if (idma->dma_len == 0) { 78 if (idma->dma.sgcount > 1) { 79 idma->dma.sg = sg_next(idma->dma.sg); 80 idma->dma_addr = idma->dma.sg->dma_address; 81 idma->dma_len = idma->dma.sg->length; 82 idma->dma.sgcount--; 83 } else { 84 idma->dma.sg = NULL; 85 flags |= DMA_END_S; 86 } 87 } 88 } else { 89 flags = DMA_END_S | DMA_END_L; 90 sg->dma_address = 0; 91 sg->length = 0; 92 } 93 94 sg->length |= flags; 95 } 96 97 static irqreturn_t iomd_dma_handle(int irq, void *dev_id) 98 { 99 struct iomd_dma *idma = dev_id; 100 unsigned long base = idma->base; 101 102 do { 103 unsigned int status; 104 105 status = iomd_readb(base + ST); 106 if (!(status & DMA_ST_INT)) 107 return IRQ_HANDLED; 108 109 if ((idma->state ^ status) & DMA_ST_AB) 110 iomd_get_next_sg(&idma->cur_sg, idma); 111 112 switch (status & (DMA_ST_OFL | DMA_ST_AB)) { 113 case DMA_ST_OFL: /* OIA */ 114 case DMA_ST_AB: /* .IB */ 115 iomd_writel(idma->cur_sg.dma_address, base + CURA); 116 iomd_writel(idma->cur_sg.length, base + ENDA); 117 idma->state = DMA_ST_AB; 118 break; 119 120 case DMA_ST_OFL | DMA_ST_AB: /* OIB */ 121 case 0: /* .IA */ 122 iomd_writel(idma->cur_sg.dma_address, base + CURB); 123 iomd_writel(idma->cur_sg.length, base + ENDB); 124 idma->state = 0; 125 break; 126 } 127 128 if (status & DMA_ST_OFL && 129 idma->cur_sg.length == (DMA_END_S|DMA_END_L)) 130 break; 131 } while (1); 132 133 idma->state = ~DMA_ST_AB; 134 disable_irq(irq); 135 136 return IRQ_HANDLED; 137 } 138 139 static int iomd_request_dma(unsigned int chan, dma_t *dma) 140 { 141 struct iomd_dma *idma = container_of(dma, struct iomd_dma, dma); 142 143 return request_irq(idma->irq, iomd_dma_handle, 144 0, idma->dma.device_id, idma); 145 } 146 147 static void iomd_free_dma(unsigned int chan, dma_t *dma) 148 { 149 struct iomd_dma *idma = container_of(dma, struct iomd_dma, dma); 150 151 free_irq(idma->irq, idma); 152 } 153 154 static struct device isa_dma_dev = { 155 .init_name = "fallback device", 156 .coherent_dma_mask = ~(dma_addr_t)0, 157 .dma_mask = &isa_dma_dev.coherent_dma_mask, 158 }; 159 160 static void iomd_enable_dma(unsigned int chan, dma_t *dma) 161 { 162 struct iomd_dma *idma = container_of(dma, struct iomd_dma, dma); 163 unsigned long dma_base = idma->base; 164 unsigned int ctrl = TRANSFER_SIZE | DMA_CR_E; 165 166 if (idma->dma.invalid) { 167 idma->dma.invalid = 0; 168 169 /* 170 * Cope with ISA-style drivers which expect cache 171 * coherence. 172 */ 173 if (!idma->dma.sg) { 174 idma->dma.sg = &idma->dma.buf; 175 idma->dma.sgcount = 1; 176 idma->dma.buf.length = idma->dma.count; 177 idma->dma.buf.dma_address = dma_map_single(&isa_dma_dev, 178 idma->dma.addr, idma->dma.count, 179 idma->dma.dma_mode == DMA_MODE_READ ? 180 DMA_FROM_DEVICE : DMA_TO_DEVICE); 181 } 182 183 iomd_writeb(DMA_CR_C, dma_base + CR); 184 idma->state = DMA_ST_AB; 185 } 186 187 if (idma->dma.dma_mode == DMA_MODE_READ) 188 ctrl |= DMA_CR_D; 189 190 iomd_writeb(ctrl, dma_base + CR); 191 enable_irq(idma->irq); 192 } 193 194 static void iomd_disable_dma(unsigned int chan, dma_t *dma) 195 { 196 struct iomd_dma *idma = container_of(dma, struct iomd_dma, dma); 197 unsigned long dma_base = idma->base; 198 unsigned long flags; 199 200 local_irq_save(flags); 201 if (idma->state != ~DMA_ST_AB) 202 disable_irq(idma->irq); 203 iomd_writeb(0, dma_base + CR); 204 local_irq_restore(flags); 205 } 206 207 static int iomd_set_dma_speed(unsigned int chan, dma_t *dma, int cycle) 208 { 209 int tcr, speed; 210 211 if (cycle < 188) 212 speed = 3; 213 else if (cycle <= 250) 214 speed = 2; 215 else if (cycle < 438) 216 speed = 1; 217 else 218 speed = 0; 219 220 tcr = iomd_readb(IOMD_DMATCR); 221 speed &= 3; 222 223 switch (chan) { 224 case DMA_0: 225 tcr = (tcr & ~0x03) | speed; 226 break; 227 228 case DMA_1: 229 tcr = (tcr & ~0x0c) | (speed << 2); 230 break; 231 232 case DMA_2: 233 tcr = (tcr & ~0x30) | (speed << 4); 234 break; 235 236 case DMA_3: 237 tcr = (tcr & ~0xc0) | (speed << 6); 238 break; 239 240 default: 241 break; 242 } 243 244 iomd_writeb(tcr, IOMD_DMATCR); 245 246 return speed; 247 } 248 249 static struct dma_ops iomd_dma_ops = { 250 .type = "IOMD", 251 .request = iomd_request_dma, 252 .free = iomd_free_dma, 253 .enable = iomd_enable_dma, 254 .disable = iomd_disable_dma, 255 .setspeed = iomd_set_dma_speed, 256 }; 257 258 static struct fiq_handler fh = { 259 .name = "floppydma" 260 }; 261 262 struct floppy_dma { 263 struct dma_struct dma; 264 unsigned int fiq; 265 }; 266 267 static void floppy_enable_dma(unsigned int chan, dma_t *dma) 268 { 269 struct floppy_dma *fdma = container_of(dma, struct floppy_dma, dma); 270 void *fiqhandler_start; 271 unsigned int fiqhandler_length; 272 struct pt_regs regs; 273 274 if (fdma->dma.sg) 275 BUG(); 276 277 if (fdma->dma.dma_mode == DMA_MODE_READ) { 278 extern unsigned char floppy_fiqin_start, floppy_fiqin_end; 279 fiqhandler_start = &floppy_fiqin_start; 280 fiqhandler_length = &floppy_fiqin_end - &floppy_fiqin_start; 281 } else { 282 extern unsigned char floppy_fiqout_start, floppy_fiqout_end; 283 fiqhandler_start = &floppy_fiqout_start; 284 fiqhandler_length = &floppy_fiqout_end - &floppy_fiqout_start; 285 } 286 287 regs.ARM_r9 = fdma->dma.count; 288 regs.ARM_r10 = (unsigned long)fdma->dma.addr; 289 regs.ARM_fp = (unsigned long)FLOPPYDMA_BASE; 290 291 if (claim_fiq(&fh)) { 292 printk("floppydma: couldn't claim FIQ.\n"); 293 return; 294 } 295 296 set_fiq_handler(fiqhandler_start, fiqhandler_length); 297 set_fiq_regs(®s); 298 enable_fiq(fdma->fiq); 299 } 300 301 static void floppy_disable_dma(unsigned int chan, dma_t *dma) 302 { 303 struct floppy_dma *fdma = container_of(dma, struct floppy_dma, dma); 304 disable_fiq(fdma->fiq); 305 release_fiq(&fh); 306 } 307 308 static int floppy_get_residue(unsigned int chan, dma_t *dma) 309 { 310 struct pt_regs regs; 311 get_fiq_regs(®s); 312 return regs.ARM_r9; 313 } 314 315 static struct dma_ops floppy_dma_ops = { 316 .type = "FIQDMA", 317 .enable = floppy_enable_dma, 318 .disable = floppy_disable_dma, 319 .residue = floppy_get_residue, 320 }; 321 322 /* 323 * This is virtual DMA - we don't need anything here. 324 */ 325 static void sound_enable_disable_dma(unsigned int chan, dma_t *dma) 326 { 327 } 328 329 static struct dma_ops sound_dma_ops = { 330 .type = "VIRTUAL", 331 .enable = sound_enable_disable_dma, 332 .disable = sound_enable_disable_dma, 333 }; 334 335 static struct iomd_dma iomd_dma[6]; 336 337 static struct floppy_dma floppy_dma = { 338 .dma = { 339 .d_ops = &floppy_dma_ops, 340 }, 341 .fiq = FIQ_FLOPPYDATA, 342 }; 343 344 static dma_t sound_dma = { 345 .d_ops = &sound_dma_ops, 346 }; 347 348 static int __init rpc_dma_init(void) 349 { 350 unsigned int i; 351 int ret; 352 353 iomd_writeb(0, IOMD_IO0CR); 354 iomd_writeb(0, IOMD_IO1CR); 355 iomd_writeb(0, IOMD_IO2CR); 356 iomd_writeb(0, IOMD_IO3CR); 357 358 iomd_writeb(0xa0, IOMD_DMATCR); 359 360 /* 361 * Setup DMA channels 2,3 to be for podules 362 * and channels 0,1 for internal devices 363 */ 364 iomd_writeb(DMA_EXT_IO3|DMA_EXT_IO2, IOMD_DMAEXT); 365 366 iomd_dma[DMA_0].base = IOMD_IO0CURA; 367 iomd_dma[DMA_0].irq = IRQ_DMA0; 368 iomd_dma[DMA_1].base = IOMD_IO1CURA; 369 iomd_dma[DMA_1].irq = IRQ_DMA1; 370 iomd_dma[DMA_2].base = IOMD_IO2CURA; 371 iomd_dma[DMA_2].irq = IRQ_DMA2; 372 iomd_dma[DMA_3].base = IOMD_IO3CURA; 373 iomd_dma[DMA_3].irq = IRQ_DMA3; 374 iomd_dma[DMA_S0].base = IOMD_SD0CURA; 375 iomd_dma[DMA_S0].irq = IRQ_DMAS0; 376 iomd_dma[DMA_S1].base = IOMD_SD1CURA; 377 iomd_dma[DMA_S1].irq = IRQ_DMAS1; 378 379 for (i = DMA_0; i <= DMA_S1; i++) { 380 iomd_dma[i].dma.d_ops = &iomd_dma_ops; 381 382 ret = isa_dma_add(i, &iomd_dma[i].dma); 383 if (ret) 384 printk("IOMDDMA%u: unable to register: %d\n", i, ret); 385 } 386 387 ret = isa_dma_add(DMA_VIRTUAL_FLOPPY, &floppy_dma.dma); 388 if (ret) 389 printk("IOMDFLOPPY: unable to register: %d\n", ret); 390 ret = isa_dma_add(DMA_VIRTUAL_SOUND, &sound_dma); 391 if (ret) 392 printk("IOMDSOUND: unable to register: %d\n", ret); 393 return 0; 394 } 395 core_initcall(rpc_dma_init); 396