1 /* 2 * Copyright (c) by Jaroslav Kysela <perex@suse.cz> 3 * Copyright (c) by Takashi Iwai <tiwai@suse.de> 4 * Copyright (c) by Scott McNab <sdm@fractalgraphics.com.au> 5 * 6 * Trident 4DWave-NX memory page allocation (TLB area) 7 * Trident chip can handle only 16MByte of the memory at the same time. 8 * 9 * 10 * This program is free software; you can redistribute it and/or modify 11 * it under the terms of the GNU General Public License as published by 12 * the Free Software Foundation; either version 2 of the License, or 13 * (at your option) any later version. 14 * 15 * This program is distributed in the hope that it will be useful, 16 * but WITHOUT ANY WARRANTY; without even the implied warranty of 17 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 18 * GNU General Public License for more details. 19 * 20 * You should have received a copy of the GNU General Public License 21 * along with this program; if not, write to the Free Software 22 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA 23 * 24 */ 25 26 #include <sound/driver.h> 27 #include <asm/io.h> 28 #include <linux/pci.h> 29 #include <linux/time.h> 30 #include <linux/mutex.h> 31 32 #include <sound/core.h> 33 #include <sound/trident.h> 34 35 /* page arguments of these two macros are Trident page (4096 bytes), not like 36 * aligned pages in others 37 */ 38 #define __set_tlb_bus(trident,page,ptr,addr) \ 39 do { (trident)->tlb.entries[page] = cpu_to_le32((addr) & ~(SNDRV_TRIDENT_PAGE_SIZE-1)); \ 40 (trident)->tlb.shadow_entries[page] = (ptr); } while (0) 41 #define __tlb_to_ptr(trident,page) \ 42 (void*)((trident)->tlb.shadow_entries[page]) 43 #define __tlb_to_addr(trident,page) \ 44 (dma_addr_t)le32_to_cpu((trident->tlb.entries[page]) & ~(SNDRV_TRIDENT_PAGE_SIZE - 1)) 45 46 #if PAGE_SIZE == 4096 47 /* page size == SNDRV_TRIDENT_PAGE_SIZE */ 48 #define ALIGN_PAGE_SIZE PAGE_SIZE /* minimum page size for allocation */ 49 #define MAX_ALIGN_PAGES SNDRV_TRIDENT_MAX_PAGES /* maxmium aligned pages */ 50 /* fill TLB entrie(s) corresponding to page with ptr */ 51 #define set_tlb_bus(trident,page,ptr,addr) __set_tlb_bus(trident,page,ptr,addr) 52 /* fill TLB entrie(s) corresponding to page with silence pointer */ 53 #define set_silent_tlb(trident,page) __set_tlb_bus(trident, page, (unsigned long)trident->tlb.silent_page.area, trident->tlb.silent_page.addr) 54 /* get aligned page from offset address */ 55 #define get_aligned_page(offset) ((offset) >> 12) 56 /* get offset address from aligned page */ 57 #define aligned_page_offset(page) ((page) << 12) 58 /* get buffer address from aligned page */ 59 #define page_to_ptr(trident,page) __tlb_to_ptr(trident, page) 60 /* get PCI physical address from aligned page */ 61 #define page_to_addr(trident,page) __tlb_to_addr(trident, page) 62 63 #elif PAGE_SIZE == 8192 64 /* page size == SNDRV_TRIDENT_PAGE_SIZE x 2*/ 65 #define ALIGN_PAGE_SIZE PAGE_SIZE 66 #define MAX_ALIGN_PAGES (SNDRV_TRIDENT_MAX_PAGES / 2) 67 #define get_aligned_page(offset) ((offset) >> 13) 68 #define aligned_page_offset(page) ((page) << 13) 69 #define page_to_ptr(trident,page) __tlb_to_ptr(trident, (page) << 1) 70 #define page_to_addr(trident,page) __tlb_to_addr(trident, (page) << 1) 71 72 /* fill TLB entries -- we need to fill two entries */ 73 static inline void set_tlb_bus(struct snd_trident *trident, int page, 74 unsigned long ptr, dma_addr_t addr) 75 { 76 page <<= 1; 77 __set_tlb_bus(trident, page, ptr, addr); 78 __set_tlb_bus(trident, page+1, ptr + SNDRV_TRIDENT_PAGE_SIZE, addr + SNDRV_TRIDENT_PAGE_SIZE); 79 } 80 static inline void set_silent_tlb(struct snd_trident *trident, int page) 81 { 82 page <<= 1; 83 __set_tlb_bus(trident, page, (unsigned long)trident->tlb.silent_page.area, trident->tlb.silent_page.addr); 84 __set_tlb_bus(trident, page+1, (unsigned long)trident->tlb.silent_page.area, trident->tlb.silent_page.addr); 85 } 86 87 #else 88 /* arbitrary size */ 89 #define UNIT_PAGES (PAGE_SIZE / SNDRV_TRIDENT_PAGE_SIZE) 90 #define ALIGN_PAGE_SIZE (SNDRV_TRIDENT_PAGE_SIZE * UNIT_PAGES) 91 #define MAX_ALIGN_PAGES (SNDRV_TRIDENT_MAX_PAGES / UNIT_PAGES) 92 /* Note: if alignment doesn't match to the maximum size, the last few blocks 93 * become unusable. To use such blocks, you'll need to check the validity 94 * of accessing page in set_tlb_bus and set_silent_tlb. search_empty() 95 * should also check it, too. 96 */ 97 #define get_aligned_page(offset) ((offset) / ALIGN_PAGE_SIZE) 98 #define aligned_page_offset(page) ((page) * ALIGN_PAGE_SIZE) 99 #define page_to_ptr(trident,page) __tlb_to_ptr(trident, (page) * UNIT_PAGES) 100 #define page_to_addr(trident,page) __tlb_to_addr(trident, (page) * UNIT_PAGES) 101 102 /* fill TLB entries -- UNIT_PAGES entries must be filled */ 103 static inline void set_tlb_bus(struct snd_trident *trident, int page, 104 unsigned long ptr, dma_addr_t addr) 105 { 106 int i; 107 page *= UNIT_PAGES; 108 for (i = 0; i < UNIT_PAGES; i++, page++) { 109 __set_tlb_bus(trident, page, ptr, addr); 110 ptr += SNDRV_TRIDENT_PAGE_SIZE; 111 addr += SNDRV_TRIDENT_PAGE_SIZE; 112 } 113 } 114 static inline void set_silent_tlb(struct snd_trident *trident, int page) 115 { 116 int i; 117 page *= UNIT_PAGES; 118 for (i = 0; i < UNIT_PAGES; i++, page++) 119 __set_tlb_bus(trident, page, (unsigned long)trident->tlb.silent_page.area, trident->tlb.silent_page.addr); 120 } 121 122 #endif /* PAGE_SIZE */ 123 124 /* calculate buffer pointer from offset address */ 125 static inline void *offset_ptr(struct snd_trident *trident, int offset) 126 { 127 char *ptr; 128 ptr = page_to_ptr(trident, get_aligned_page(offset)); 129 ptr += offset % ALIGN_PAGE_SIZE; 130 return (void*)ptr; 131 } 132 133 /* first and last (aligned) pages of memory block */ 134 #define firstpg(blk) (((struct snd_trident_memblk_arg *)snd_util_memblk_argptr(blk))->first_page) 135 #define lastpg(blk) (((struct snd_trident_memblk_arg *)snd_util_memblk_argptr(blk))->last_page) 136 137 /* 138 * search empty pages which may contain given size 139 */ 140 static struct snd_util_memblk * 141 search_empty(struct snd_util_memhdr *hdr, int size) 142 { 143 struct snd_util_memblk *blk, *prev; 144 int page, psize; 145 struct list_head *p; 146 147 psize = get_aligned_page(size + ALIGN_PAGE_SIZE -1); 148 prev = NULL; 149 page = 0; 150 list_for_each(p, &hdr->block) { 151 blk = list_entry(p, struct snd_util_memblk, list); 152 if (page + psize <= firstpg(blk)) 153 goto __found_pages; 154 page = lastpg(blk) + 1; 155 } 156 if (page + psize > MAX_ALIGN_PAGES) 157 return NULL; 158 159 __found_pages: 160 /* create a new memory block */ 161 blk = __snd_util_memblk_new(hdr, psize * ALIGN_PAGE_SIZE, p->prev); 162 if (blk == NULL) 163 return NULL; 164 blk->offset = aligned_page_offset(page); /* set aligned offset */ 165 firstpg(blk) = page; 166 lastpg(blk) = page + psize - 1; 167 return blk; 168 } 169 170 171 /* 172 * check if the given pointer is valid for pages 173 */ 174 static int is_valid_page(unsigned long ptr) 175 { 176 if (ptr & ~0x3fffffffUL) { 177 snd_printk(KERN_ERR "max memory size is 1GB!!\n"); 178 return 0; 179 } 180 if (ptr & (SNDRV_TRIDENT_PAGE_SIZE-1)) { 181 snd_printk(KERN_ERR "page is not aligned\n"); 182 return 0; 183 } 184 return 1; 185 } 186 187 /* 188 * page allocation for DMA (Scatter-Gather version) 189 */ 190 static struct snd_util_memblk * 191 snd_trident_alloc_sg_pages(struct snd_trident *trident, 192 struct snd_pcm_substream *substream) 193 { 194 struct snd_util_memhdr *hdr; 195 struct snd_util_memblk *blk; 196 struct snd_pcm_runtime *runtime = substream->runtime; 197 int idx, page; 198 struct snd_sg_buf *sgbuf = snd_pcm_substream_sgbuf(substream); 199 200 snd_assert(runtime->dma_bytes > 0 && runtime->dma_bytes <= SNDRV_TRIDENT_MAX_PAGES * SNDRV_TRIDENT_PAGE_SIZE, return NULL); 201 hdr = trident->tlb.memhdr; 202 snd_assert(hdr != NULL, return NULL); 203 204 205 206 mutex_lock(&hdr->block_mutex); 207 blk = search_empty(hdr, runtime->dma_bytes); 208 if (blk == NULL) { 209 mutex_unlock(&hdr->block_mutex); 210 return NULL; 211 } 212 if (lastpg(blk) - firstpg(blk) >= sgbuf->pages) { 213 snd_printk(KERN_ERR "page calculation doesn't match: allocated pages = %d, trident = %d/%d\n", sgbuf->pages, firstpg(blk), lastpg(blk)); 214 __snd_util_mem_free(hdr, blk); 215 mutex_unlock(&hdr->block_mutex); 216 return NULL; 217 } 218 219 /* set TLB entries */ 220 idx = 0; 221 for (page = firstpg(blk); page <= lastpg(blk); page++, idx++) { 222 dma_addr_t addr = sgbuf->table[idx].addr; 223 unsigned long ptr = (unsigned long)sgbuf->table[idx].buf; 224 if (! is_valid_page(addr)) { 225 __snd_util_mem_free(hdr, blk); 226 mutex_unlock(&hdr->block_mutex); 227 return NULL; 228 } 229 set_tlb_bus(trident, page, ptr, addr); 230 } 231 mutex_unlock(&hdr->block_mutex); 232 return blk; 233 } 234 235 /* 236 * page allocation for DMA (contiguous version) 237 */ 238 static struct snd_util_memblk * 239 snd_trident_alloc_cont_pages(struct snd_trident *trident, 240 struct snd_pcm_substream *substream) 241 { 242 struct snd_util_memhdr *hdr; 243 struct snd_util_memblk *blk; 244 int page; 245 struct snd_pcm_runtime *runtime = substream->runtime; 246 dma_addr_t addr; 247 unsigned long ptr; 248 249 snd_assert(runtime->dma_bytes> 0 && runtime->dma_bytes <= SNDRV_TRIDENT_MAX_PAGES * SNDRV_TRIDENT_PAGE_SIZE, return NULL); 250 hdr = trident->tlb.memhdr; 251 snd_assert(hdr != NULL, return NULL); 252 253 mutex_lock(&hdr->block_mutex); 254 blk = search_empty(hdr, runtime->dma_bytes); 255 if (blk == NULL) { 256 mutex_unlock(&hdr->block_mutex); 257 return NULL; 258 } 259 260 /* set TLB entries */ 261 addr = runtime->dma_addr; 262 ptr = (unsigned long)runtime->dma_area; 263 for (page = firstpg(blk); page <= lastpg(blk); page++, 264 ptr += SNDRV_TRIDENT_PAGE_SIZE, addr += SNDRV_TRIDENT_PAGE_SIZE) { 265 if (! is_valid_page(addr)) { 266 __snd_util_mem_free(hdr, blk); 267 mutex_unlock(&hdr->block_mutex); 268 return NULL; 269 } 270 set_tlb_bus(trident, page, ptr, addr); 271 } 272 mutex_unlock(&hdr->block_mutex); 273 return blk; 274 } 275 276 /* 277 * page allocation for DMA 278 */ 279 struct snd_util_memblk * 280 snd_trident_alloc_pages(struct snd_trident *trident, 281 struct snd_pcm_substream *substream) 282 { 283 snd_assert(trident != NULL, return NULL); 284 snd_assert(substream != NULL, return NULL); 285 if (substream->dma_buffer.dev.type == SNDRV_DMA_TYPE_DEV_SG) 286 return snd_trident_alloc_sg_pages(trident, substream); 287 else 288 return snd_trident_alloc_cont_pages(trident, substream); 289 } 290 291 292 /* 293 * release DMA buffer from page table 294 */ 295 int snd_trident_free_pages(struct snd_trident *trident, 296 struct snd_util_memblk *blk) 297 { 298 struct snd_util_memhdr *hdr; 299 int page; 300 301 snd_assert(trident != NULL, return -EINVAL); 302 snd_assert(blk != NULL, return -EINVAL); 303 304 hdr = trident->tlb.memhdr; 305 mutex_lock(&hdr->block_mutex); 306 /* reset TLB entries */ 307 for (page = firstpg(blk); page <= lastpg(blk); page++) 308 set_silent_tlb(trident, page); 309 /* free memory block */ 310 __snd_util_mem_free(hdr, blk); 311 mutex_unlock(&hdr->block_mutex); 312 return 0; 313 } 314 315 316 /*---------------------------------------------------------------- 317 * memory allocation using multiple pages (for synth) 318 *---------------------------------------------------------------- 319 * Unlike the DMA allocation above, non-contiguous pages are 320 * assigned to TLB. 321 *----------------------------------------------------------------*/ 322 323 /* 324 */ 325 static int synth_alloc_pages(struct snd_trident *hw, struct snd_util_memblk *blk); 326 static int synth_free_pages(struct snd_trident *hw, struct snd_util_memblk *blk); 327 328 /* 329 * allocate a synth sample area 330 */ 331 struct snd_util_memblk * 332 snd_trident_synth_alloc(struct snd_trident *hw, unsigned int size) 333 { 334 struct snd_util_memblk *blk; 335 struct snd_util_memhdr *hdr = hw->tlb.memhdr; 336 337 mutex_lock(&hdr->block_mutex); 338 blk = __snd_util_mem_alloc(hdr, size); 339 if (blk == NULL) { 340 mutex_unlock(&hdr->block_mutex); 341 return NULL; 342 } 343 if (synth_alloc_pages(hw, blk)) { 344 __snd_util_mem_free(hdr, blk); 345 mutex_unlock(&hdr->block_mutex); 346 return NULL; 347 } 348 mutex_unlock(&hdr->block_mutex); 349 return blk; 350 } 351 352 EXPORT_SYMBOL(snd_trident_synth_alloc); 353 354 /* 355 * free a synth sample area 356 */ 357 int 358 snd_trident_synth_free(struct snd_trident *hw, struct snd_util_memblk *blk) 359 { 360 struct snd_util_memhdr *hdr = hw->tlb.memhdr; 361 362 mutex_lock(&hdr->block_mutex); 363 synth_free_pages(hw, blk); 364 __snd_util_mem_free(hdr, blk); 365 mutex_unlock(&hdr->block_mutex); 366 return 0; 367 } 368 369 EXPORT_SYMBOL(snd_trident_synth_free); 370 371 /* 372 * reset TLB entry and free kernel page 373 */ 374 static void clear_tlb(struct snd_trident *trident, int page) 375 { 376 void *ptr = page_to_ptr(trident, page); 377 dma_addr_t addr = page_to_addr(trident, page); 378 set_silent_tlb(trident, page); 379 if (ptr) { 380 struct snd_dma_buffer dmab; 381 dmab.dev.type = SNDRV_DMA_TYPE_DEV; 382 dmab.dev.dev = snd_dma_pci_data(trident->pci); 383 dmab.area = ptr; 384 dmab.addr = addr; 385 dmab.bytes = ALIGN_PAGE_SIZE; 386 snd_dma_free_pages(&dmab); 387 } 388 } 389 390 /* check new allocation range */ 391 static void get_single_page_range(struct snd_util_memhdr *hdr, 392 struct snd_util_memblk *blk, 393 int *first_page_ret, int *last_page_ret) 394 { 395 struct list_head *p; 396 struct snd_util_memblk *q; 397 int first_page, last_page; 398 first_page = firstpg(blk); 399 if ((p = blk->list.prev) != &hdr->block) { 400 q = list_entry(p, struct snd_util_memblk, list); 401 if (lastpg(q) == first_page) 402 first_page++; /* first page was already allocated */ 403 } 404 last_page = lastpg(blk); 405 if ((p = blk->list.next) != &hdr->block) { 406 q = list_entry(p, struct snd_util_memblk, list); 407 if (firstpg(q) == last_page) 408 last_page--; /* last page was already allocated */ 409 } 410 *first_page_ret = first_page; 411 *last_page_ret = last_page; 412 } 413 414 /* 415 * allocate kernel pages and assign them to TLB 416 */ 417 static int synth_alloc_pages(struct snd_trident *hw, struct snd_util_memblk *blk) 418 { 419 int page, first_page, last_page; 420 struct snd_dma_buffer dmab; 421 422 firstpg(blk) = get_aligned_page(blk->offset); 423 lastpg(blk) = get_aligned_page(blk->offset + blk->size - 1); 424 get_single_page_range(hw->tlb.memhdr, blk, &first_page, &last_page); 425 426 /* allocate a kernel page for each Trident page - 427 * fortunately Trident page size and kernel PAGE_SIZE is identical! 428 */ 429 for (page = first_page; page <= last_page; page++) { 430 if (snd_dma_alloc_pages(SNDRV_DMA_TYPE_DEV, snd_dma_pci_data(hw->pci), 431 ALIGN_PAGE_SIZE, &dmab) < 0) 432 goto __fail; 433 if (! is_valid_page(dmab.addr)) { 434 snd_dma_free_pages(&dmab); 435 goto __fail; 436 } 437 set_tlb_bus(hw, page, (unsigned long)dmab.area, dmab.addr); 438 } 439 return 0; 440 441 __fail: 442 /* release allocated pages */ 443 last_page = page - 1; 444 for (page = first_page; page <= last_page; page++) 445 clear_tlb(hw, page); 446 447 return -ENOMEM; 448 } 449 450 /* 451 * free pages 452 */ 453 static int synth_free_pages(struct snd_trident *trident, struct snd_util_memblk *blk) 454 { 455 int page, first_page, last_page; 456 457 get_single_page_range(trident->tlb.memhdr, blk, &first_page, &last_page); 458 for (page = first_page; page <= last_page; page++) 459 clear_tlb(trident, page); 460 461 return 0; 462 } 463 464 /* 465 * copy_from_user(blk + offset, data, size) 466 */ 467 int snd_trident_synth_copy_from_user(struct snd_trident *trident, 468 struct snd_util_memblk *blk, 469 int offset, const char __user *data, int size) 470 { 471 int page, nextofs, end_offset, temp, temp1; 472 473 offset += blk->offset; 474 end_offset = offset + size; 475 page = get_aligned_page(offset) + 1; 476 do { 477 nextofs = aligned_page_offset(page); 478 temp = nextofs - offset; 479 temp1 = end_offset - offset; 480 if (temp1 < temp) 481 temp = temp1; 482 if (copy_from_user(offset_ptr(trident, offset), data, temp)) 483 return -EFAULT; 484 offset = nextofs; 485 data += temp; 486 page++; 487 } while (offset < end_offset); 488 return 0; 489 } 490 491 EXPORT_SYMBOL(snd_trident_synth_copy_from_user); 492