xref: /openbmc/linux/arch/alpha/kernel/pci_iommu.c (revision 808643ea)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  *	linux/arch/alpha/kernel/pci_iommu.c
4  */
5 
6 #include <linux/kernel.h>
7 #include <linux/mm.h>
8 #include <linux/pci.h>
9 #include <linux/gfp.h>
10 #include <linux/memblock.h>
11 #include <linux/export.h>
12 #include <linux/scatterlist.h>
13 #include <linux/log2.h>
14 #include <linux/dma-map-ops.h>
15 #include <linux/iommu-helper.h>
16 
17 #include <asm/io.h>
18 #include <asm/hwrpb.h>
19 
20 #include "proto.h"
21 #include "pci_impl.h"
22 
23 
24 #define DEBUG_ALLOC 0
25 #if DEBUG_ALLOC > 0
26 # define DBGA(args...)		printk(KERN_DEBUG args)
27 #else
28 # define DBGA(args...)
29 #endif
30 #if DEBUG_ALLOC > 1
31 # define DBGA2(args...)		printk(KERN_DEBUG args)
32 #else
33 # define DBGA2(args...)
34 #endif
35 
36 #define DEBUG_NODIRECT 0
37 
38 #define ISA_DMA_MASK		0x00ffffff
39 
40 static inline unsigned long
41 mk_iommu_pte(unsigned long paddr)
42 {
43 	return (paddr >> (PAGE_SHIFT-1)) | 1;
44 }
45 
46 /* Return the minimum of MAX or the first power of two larger
47    than main memory.  */
48 
49 unsigned long
50 size_for_memory(unsigned long max)
51 {
52 	unsigned long mem = max_low_pfn << PAGE_SHIFT;
53 	if (mem < max)
54 		max = roundup_pow_of_two(mem);
55 	return max;
56 }
57 
58 struct pci_iommu_arena * __init
59 iommu_arena_new_node(int nid, struct pci_controller *hose, dma_addr_t base,
60 		     unsigned long window_size, unsigned long align)
61 {
62 	unsigned long mem_size;
63 	struct pci_iommu_arena *arena;
64 
65 	mem_size = window_size / (PAGE_SIZE / sizeof(unsigned long));
66 
67 	/* Note that the TLB lookup logic uses bitwise concatenation,
68 	   not addition, so the required arena alignment is based on
69 	   the size of the window.  Retain the align parameter so that
70 	   particular systems can over-align the arena.  */
71 	if (align < mem_size)
72 		align = mem_size;
73 
74 	arena = memblock_alloc(sizeof(*arena), SMP_CACHE_BYTES);
75 	if (!arena)
76 		panic("%s: Failed to allocate %zu bytes\n", __func__,
77 		      sizeof(*arena));
78 	arena->ptes = memblock_alloc(mem_size, align);
79 	if (!arena->ptes)
80 		panic("%s: Failed to allocate %lu bytes align=0x%lx\n",
81 		      __func__, mem_size, align);
82 
83 	spin_lock_init(&arena->lock);
84 	arena->hose = hose;
85 	arena->dma_base = base;
86 	arena->size = window_size;
87 	arena->next_entry = 0;
88 
89 	/* Align allocations to a multiple of a page size.  Not needed
90 	   unless there are chip bugs.  */
91 	arena->align_entry = 1;
92 
93 	return arena;
94 }
95 
96 struct pci_iommu_arena * __init
97 iommu_arena_new(struct pci_controller *hose, dma_addr_t base,
98 		unsigned long window_size, unsigned long align)
99 {
100 	return iommu_arena_new_node(0, hose, base, window_size, align);
101 }
102 
103 /* Must be called with the arena lock held */
104 static long
105 iommu_arena_find_pages(struct device *dev, struct pci_iommu_arena *arena,
106 		       long n, long mask)
107 {
108 	unsigned long *ptes;
109 	long i, p, nent;
110 	int pass = 0;
111 	unsigned long base;
112 	unsigned long boundary_size;
113 
114 	base = arena->dma_base >> PAGE_SHIFT;
115 	boundary_size = dma_get_seg_boundary_nr_pages(dev, PAGE_SHIFT);
116 
117 	/* Search forward for the first mask-aligned sequence of N free ptes */
118 	ptes = arena->ptes;
119 	nent = arena->size >> PAGE_SHIFT;
120 	p = ALIGN(arena->next_entry, mask + 1);
121 	i = 0;
122 
123 again:
124 	while (i < n && p+i < nent) {
125 		if (!i && iommu_is_span_boundary(p, n, base, boundary_size)) {
126 			p = ALIGN(p + 1, mask + 1);
127 			goto again;
128 		}
129 
130 		if (ptes[p+i])
131 			p = ALIGN(p + i + 1, mask + 1), i = 0;
132 		else
133 			i = i + 1;
134 	}
135 
136 	if (i < n) {
137 		if (pass < 1) {
138 			/*
139 			 * Reached the end.  Flush the TLB and restart
140 			 * the search from the beginning.
141 			*/
142 			alpha_mv.mv_pci_tbi(arena->hose, 0, -1);
143 
144 			pass++;
145 			p = 0;
146 			i = 0;
147 			goto again;
148 		} else
149 			return -1;
150 	}
151 
152 	/* Success. It's the responsibility of the caller to mark them
153 	   in use before releasing the lock */
154 	return p;
155 }
156 
157 static long
158 iommu_arena_alloc(struct device *dev, struct pci_iommu_arena *arena, long n,
159 		  unsigned int align)
160 {
161 	unsigned long flags;
162 	unsigned long *ptes;
163 	long i, p, mask;
164 
165 	spin_lock_irqsave(&arena->lock, flags);
166 
167 	/* Search for N empty ptes */
168 	ptes = arena->ptes;
169 	mask = max(align, arena->align_entry) - 1;
170 	p = iommu_arena_find_pages(dev, arena, n, mask);
171 	if (p < 0) {
172 		spin_unlock_irqrestore(&arena->lock, flags);
173 		return -1;
174 	}
175 
176 	/* Success.  Mark them all in use, ie not zero and invalid
177 	   for the iommu tlb that could load them from under us.
178 	   The chip specific bits will fill this in with something
179 	   kosher when we return.  */
180 	for (i = 0; i < n; ++i)
181 		ptes[p+i] = IOMMU_INVALID_PTE;
182 
183 	arena->next_entry = p + n;
184 	spin_unlock_irqrestore(&arena->lock, flags);
185 
186 	return p;
187 }
188 
189 static void
190 iommu_arena_free(struct pci_iommu_arena *arena, long ofs, long n)
191 {
192 	unsigned long *p;
193 	long i;
194 
195 	p = arena->ptes + ofs;
196 	for (i = 0; i < n; ++i)
197 		p[i] = 0;
198 }
199 
200 /*
201  * True if the machine supports DAC addressing, and DEV can
202  * make use of it given MASK.
203  */
204 static int pci_dac_dma_supported(struct pci_dev *dev, u64 mask)
205 {
206 	dma_addr_t dac_offset = alpha_mv.pci_dac_offset;
207 	int ok = 1;
208 
209 	/* If this is not set, the machine doesn't support DAC at all.  */
210 	if (dac_offset == 0)
211 		ok = 0;
212 
213 	/* The device has to be able to address our DAC bit.  */
214 	if ((dac_offset & dev->dma_mask) != dac_offset)
215 		ok = 0;
216 
217 	/* If both conditions above are met, we are fine. */
218 	DBGA("pci_dac_dma_supported %s from %ps\n",
219 	     ok ? "yes" : "no", __builtin_return_address(0));
220 
221 	return ok;
222 }
223 
224 /* Map a single buffer of the indicated size for PCI DMA in streaming
225    mode.  The 32-bit PCI bus mastering address to use is returned.
226    Once the device is given the dma address, the device owns this memory
227    until either pci_unmap_single or pci_dma_sync_single is performed.  */
228 
229 static dma_addr_t
230 pci_map_single_1(struct pci_dev *pdev, void *cpu_addr, size_t size,
231 		 int dac_allowed)
232 {
233 	struct pci_controller *hose = pdev ? pdev->sysdata : pci_isa_hose;
234 	dma_addr_t max_dma = pdev ? pdev->dma_mask : ISA_DMA_MASK;
235 	struct pci_iommu_arena *arena;
236 	long npages, dma_ofs, i;
237 	unsigned long paddr;
238 	dma_addr_t ret;
239 	unsigned int align = 0;
240 	struct device *dev = pdev ? &pdev->dev : NULL;
241 
242 	paddr = __pa(cpu_addr);
243 
244 #if !DEBUG_NODIRECT
245 	/* First check to see if we can use the direct map window.  */
246 	if (paddr + size + __direct_map_base - 1 <= max_dma
247 	    && paddr + size <= __direct_map_size) {
248 		ret = paddr + __direct_map_base;
249 
250 		DBGA2("pci_map_single: [%p,%zx] -> direct %llx from %ps\n",
251 		      cpu_addr, size, ret, __builtin_return_address(0));
252 
253 		return ret;
254 	}
255 #endif
256 
257 	/* Next, use DAC if selected earlier.  */
258 	if (dac_allowed) {
259 		ret = paddr + alpha_mv.pci_dac_offset;
260 
261 		DBGA2("pci_map_single: [%p,%zx] -> DAC %llx from %ps\n",
262 		      cpu_addr, size, ret, __builtin_return_address(0));
263 
264 		return ret;
265 	}
266 
267 	/* If the machine doesn't define a pci_tbi routine, we have to
268 	   assume it doesn't support sg mapping, and, since we tried to
269 	   use direct_map above, it now must be considered an error. */
270 	if (! alpha_mv.mv_pci_tbi) {
271 		printk_once(KERN_WARNING "pci_map_single: no HW sg\n");
272 		return DMA_MAPPING_ERROR;
273 	}
274 
275 	arena = hose->sg_pci;
276 	if (!arena || arena->dma_base + arena->size - 1 > max_dma)
277 		arena = hose->sg_isa;
278 
279 	npages = iommu_num_pages(paddr, size, PAGE_SIZE);
280 
281 	/* Force allocation to 64KB boundary for ISA bridges. */
282 	if (pdev && pdev == isa_bridge)
283 		align = 8;
284 	dma_ofs = iommu_arena_alloc(dev, arena, npages, align);
285 	if (dma_ofs < 0) {
286 		printk(KERN_WARNING "pci_map_single failed: "
287 		       "could not allocate dma page tables\n");
288 		return DMA_MAPPING_ERROR;
289 	}
290 
291 	paddr &= PAGE_MASK;
292 	for (i = 0; i < npages; ++i, paddr += PAGE_SIZE)
293 		arena->ptes[i + dma_ofs] = mk_iommu_pte(paddr);
294 
295 	ret = arena->dma_base + dma_ofs * PAGE_SIZE;
296 	ret += (unsigned long)cpu_addr & ~PAGE_MASK;
297 
298 	DBGA2("pci_map_single: [%p,%zx] np %ld -> sg %llx from %ps\n",
299 	      cpu_addr, size, npages, ret, __builtin_return_address(0));
300 
301 	return ret;
302 }
303 
304 /* Helper for generic DMA-mapping functions. */
305 static struct pci_dev *alpha_gendev_to_pci(struct device *dev)
306 {
307 	if (dev && dev_is_pci(dev))
308 		return to_pci_dev(dev);
309 
310 	/* Assume that non-PCI devices asking for DMA are either ISA or EISA,
311 	   BUG() otherwise. */
312 	BUG_ON(!isa_bridge);
313 
314 	/* Assume non-busmaster ISA DMA when dma_mask is not set (the ISA
315 	   bridge is bus master then). */
316 	if (!dev || !dev->dma_mask || !*dev->dma_mask)
317 		return isa_bridge;
318 
319 	/* For EISA bus masters, return isa_bridge (it might have smaller
320 	   dma_mask due to wiring limitations). */
321 	if (*dev->dma_mask >= isa_bridge->dma_mask)
322 		return isa_bridge;
323 
324 	/* This assumes ISA bus master with dma_mask 0xffffff. */
325 	return NULL;
326 }
327 
328 static dma_addr_t alpha_pci_map_page(struct device *dev, struct page *page,
329 				     unsigned long offset, size_t size,
330 				     enum dma_data_direction dir,
331 				     unsigned long attrs)
332 {
333 	struct pci_dev *pdev = alpha_gendev_to_pci(dev);
334 	int dac_allowed;
335 
336 	BUG_ON(dir == PCI_DMA_NONE);
337 
338 	dac_allowed = pdev ? pci_dac_dma_supported(pdev, pdev->dma_mask) : 0;
339 	return pci_map_single_1(pdev, (char *)page_address(page) + offset,
340 				size, dac_allowed);
341 }
342 
343 /* Unmap a single streaming mode DMA translation.  The DMA_ADDR and
344    SIZE must match what was provided for in a previous pci_map_single
345    call.  All other usages are undefined.  After this call, reads by
346    the cpu to the buffer are guaranteed to see whatever the device
347    wrote there.  */
348 
349 static void alpha_pci_unmap_page(struct device *dev, dma_addr_t dma_addr,
350 				 size_t size, enum dma_data_direction dir,
351 				 unsigned long attrs)
352 {
353 	unsigned long flags;
354 	struct pci_dev *pdev = alpha_gendev_to_pci(dev);
355 	struct pci_controller *hose = pdev ? pdev->sysdata : pci_isa_hose;
356 	struct pci_iommu_arena *arena;
357 	long dma_ofs, npages;
358 
359 	BUG_ON(dir == PCI_DMA_NONE);
360 
361 	if (dma_addr >= __direct_map_base
362 	    && dma_addr < __direct_map_base + __direct_map_size) {
363 		/* Nothing to do.  */
364 
365 		DBGA2("pci_unmap_single: direct [%llx,%zx] from %ps\n",
366 		      dma_addr, size, __builtin_return_address(0));
367 
368 		return;
369 	}
370 
371 	if (dma_addr > 0xffffffff) {
372 		DBGA2("pci64_unmap_single: DAC [%llx,%zx] from %ps\n",
373 		      dma_addr, size, __builtin_return_address(0));
374 		return;
375 	}
376 
377 	arena = hose->sg_pci;
378 	if (!arena || dma_addr < arena->dma_base)
379 		arena = hose->sg_isa;
380 
381 	dma_ofs = (dma_addr - arena->dma_base) >> PAGE_SHIFT;
382 	if (dma_ofs * PAGE_SIZE >= arena->size) {
383 		printk(KERN_ERR "Bogus pci_unmap_single: dma_addr %llx "
384 		       " base %llx size %x\n",
385 		       dma_addr, arena->dma_base, arena->size);
386 		return;
387 		BUG();
388 	}
389 
390 	npages = iommu_num_pages(dma_addr, size, PAGE_SIZE);
391 
392 	spin_lock_irqsave(&arena->lock, flags);
393 
394 	iommu_arena_free(arena, dma_ofs, npages);
395 
396         /* If we're freeing ptes above the `next_entry' pointer (they
397            may have snuck back into the TLB since the last wrap flush),
398            we need to flush the TLB before reallocating the latter.  */
399 	if (dma_ofs >= arena->next_entry)
400 		alpha_mv.mv_pci_tbi(hose, dma_addr, dma_addr + size - 1);
401 
402 	spin_unlock_irqrestore(&arena->lock, flags);
403 
404 	DBGA2("pci_unmap_single: sg [%llx,%zx] np %ld from %ps\n",
405 	      dma_addr, size, npages, __builtin_return_address(0));
406 }
407 
408 /* Allocate and map kernel buffer using consistent mode DMA for PCI
409    device.  Returns non-NULL cpu-view pointer to the buffer if
410    successful and sets *DMA_ADDRP to the pci side dma address as well,
411    else DMA_ADDRP is undefined.  */
412 
413 static void *alpha_pci_alloc_coherent(struct device *dev, size_t size,
414 				      dma_addr_t *dma_addrp, gfp_t gfp,
415 				      unsigned long attrs)
416 {
417 	struct pci_dev *pdev = alpha_gendev_to_pci(dev);
418 	void *cpu_addr;
419 	long order = get_order(size);
420 
421 	gfp &= ~GFP_DMA;
422 
423 try_again:
424 	cpu_addr = (void *)__get_free_pages(gfp | __GFP_ZERO, order);
425 	if (! cpu_addr) {
426 		printk(KERN_INFO "pci_alloc_consistent: "
427 		       "get_free_pages failed from %ps\n",
428 			__builtin_return_address(0));
429 		/* ??? Really atomic allocation?  Otherwise we could play
430 		   with vmalloc and sg if we can't find contiguous memory.  */
431 		return NULL;
432 	}
433 	memset(cpu_addr, 0, size);
434 
435 	*dma_addrp = pci_map_single_1(pdev, cpu_addr, size, 0);
436 	if (*dma_addrp == DMA_MAPPING_ERROR) {
437 		free_pages((unsigned long)cpu_addr, order);
438 		if (alpha_mv.mv_pci_tbi || (gfp & GFP_DMA))
439 			return NULL;
440 		/* The address doesn't fit required mask and we
441 		   do not have iommu. Try again with GFP_DMA. */
442 		gfp |= GFP_DMA;
443 		goto try_again;
444 	}
445 
446 	DBGA2("pci_alloc_consistent: %zx -> [%p,%llx] from %ps\n",
447 	      size, cpu_addr, *dma_addrp, __builtin_return_address(0));
448 
449 	return cpu_addr;
450 }
451 
452 /* Free and unmap a consistent DMA buffer.  CPU_ADDR and DMA_ADDR must
453    be values that were returned from pci_alloc_consistent.  SIZE must
454    be the same as what as passed into pci_alloc_consistent.
455    References to the memory and mappings associated with CPU_ADDR or
456    DMA_ADDR past this call are illegal.  */
457 
458 static void alpha_pci_free_coherent(struct device *dev, size_t size,
459 				    void *cpu_addr, dma_addr_t dma_addr,
460 				    unsigned long attrs)
461 {
462 	struct pci_dev *pdev = alpha_gendev_to_pci(dev);
463 	pci_unmap_single(pdev, dma_addr, size, PCI_DMA_BIDIRECTIONAL);
464 	free_pages((unsigned long)cpu_addr, get_order(size));
465 
466 	DBGA2("pci_free_consistent: [%llx,%zx] from %ps\n",
467 	      dma_addr, size, __builtin_return_address(0));
468 }
469 
470 /* Classify the elements of the scatterlist.  Write dma_address
471    of each element with:
472 	0   : Followers all physically adjacent.
473 	1   : Followers all virtually adjacent.
474 	-1  : Not leader, physically adjacent to previous.
475 	-2  : Not leader, virtually adjacent to previous.
476    Write dma_length of each leader with the combined lengths of
477    the mergable followers.  */
478 
479 #define SG_ENT_VIRT_ADDRESS(SG) (sg_virt((SG)))
480 #define SG_ENT_PHYS_ADDRESS(SG) __pa(SG_ENT_VIRT_ADDRESS(SG))
481 
482 static void
483 sg_classify(struct device *dev, struct scatterlist *sg, struct scatterlist *end,
484 	    int virt_ok)
485 {
486 	unsigned long next_paddr;
487 	struct scatterlist *leader;
488 	long leader_flag, leader_length;
489 	unsigned int max_seg_size;
490 
491 	leader = sg;
492 	leader_flag = 0;
493 	leader_length = leader->length;
494 	next_paddr = SG_ENT_PHYS_ADDRESS(leader) + leader_length;
495 
496 	/* we will not marge sg without device. */
497 	max_seg_size = dev ? dma_get_max_seg_size(dev) : 0;
498 	for (++sg; sg < end; ++sg) {
499 		unsigned long addr, len;
500 		addr = SG_ENT_PHYS_ADDRESS(sg);
501 		len = sg->length;
502 
503 		if (leader_length + len > max_seg_size)
504 			goto new_segment;
505 
506 		if (next_paddr == addr) {
507 			sg->dma_address = -1;
508 			leader_length += len;
509 		} else if (((next_paddr | addr) & ~PAGE_MASK) == 0 && virt_ok) {
510 			sg->dma_address = -2;
511 			leader_flag = 1;
512 			leader_length += len;
513 		} else {
514 new_segment:
515 			leader->dma_address = leader_flag;
516 			leader->dma_length = leader_length;
517 			leader = sg;
518 			leader_flag = 0;
519 			leader_length = len;
520 		}
521 
522 		next_paddr = addr + len;
523 	}
524 
525 	leader->dma_address = leader_flag;
526 	leader->dma_length = leader_length;
527 }
528 
529 /* Given a scatterlist leader, choose an allocation method and fill
530    in the blanks.  */
531 
532 static int
533 sg_fill(struct device *dev, struct scatterlist *leader, struct scatterlist *end,
534 	struct scatterlist *out, struct pci_iommu_arena *arena,
535 	dma_addr_t max_dma, int dac_allowed)
536 {
537 	unsigned long paddr = SG_ENT_PHYS_ADDRESS(leader);
538 	long size = leader->dma_length;
539 	struct scatterlist *sg;
540 	unsigned long *ptes;
541 	long npages, dma_ofs, i;
542 
543 #if !DEBUG_NODIRECT
544 	/* If everything is physically contiguous, and the addresses
545 	   fall into the direct-map window, use it.  */
546 	if (leader->dma_address == 0
547 	    && paddr + size + __direct_map_base - 1 <= max_dma
548 	    && paddr + size <= __direct_map_size) {
549 		out->dma_address = paddr + __direct_map_base;
550 		out->dma_length = size;
551 
552 		DBGA("    sg_fill: [%p,%lx] -> direct %llx\n",
553 		     __va(paddr), size, out->dma_address);
554 
555 		return 0;
556 	}
557 #endif
558 
559 	/* If physically contiguous and DAC is available, use it.  */
560 	if (leader->dma_address == 0 && dac_allowed) {
561 		out->dma_address = paddr + alpha_mv.pci_dac_offset;
562 		out->dma_length = size;
563 
564 		DBGA("    sg_fill: [%p,%lx] -> DAC %llx\n",
565 		     __va(paddr), size, out->dma_address);
566 
567 		return 0;
568 	}
569 
570 	/* Otherwise, we'll use the iommu to make the pages virtually
571 	   contiguous.  */
572 
573 	paddr &= ~PAGE_MASK;
574 	npages = iommu_num_pages(paddr, size, PAGE_SIZE);
575 	dma_ofs = iommu_arena_alloc(dev, arena, npages, 0);
576 	if (dma_ofs < 0) {
577 		/* If we attempted a direct map above but failed, die.  */
578 		if (leader->dma_address == 0)
579 			return -1;
580 
581 		/* Otherwise, break up the remaining virtually contiguous
582 		   hunks into individual direct maps and retry.  */
583 		sg_classify(dev, leader, end, 0);
584 		return sg_fill(dev, leader, end, out, arena, max_dma, dac_allowed);
585 	}
586 
587 	out->dma_address = arena->dma_base + dma_ofs*PAGE_SIZE + paddr;
588 	out->dma_length = size;
589 
590 	DBGA("    sg_fill: [%p,%lx] -> sg %llx np %ld\n",
591 	     __va(paddr), size, out->dma_address, npages);
592 
593 	/* All virtually contiguous.  We need to find the length of each
594 	   physically contiguous subsegment to fill in the ptes.  */
595 	ptes = &arena->ptes[dma_ofs];
596 	sg = leader;
597 	do {
598 #if DEBUG_ALLOC > 0
599 		struct scatterlist *last_sg = sg;
600 #endif
601 
602 		size = sg->length;
603 		paddr = SG_ENT_PHYS_ADDRESS(sg);
604 
605 		while (sg+1 < end && (int) sg[1].dma_address == -1) {
606 			size += sg[1].length;
607 			sg = sg_next(sg);
608 		}
609 
610 		npages = iommu_num_pages(paddr, size, PAGE_SIZE);
611 
612 		paddr &= PAGE_MASK;
613 		for (i = 0; i < npages; ++i, paddr += PAGE_SIZE)
614 			*ptes++ = mk_iommu_pte(paddr);
615 
616 #if DEBUG_ALLOC > 0
617 		DBGA("    (%ld) [%p,%x] np %ld\n",
618 		     last_sg - leader, SG_ENT_VIRT_ADDRESS(last_sg),
619 		     last_sg->length, npages);
620 		while (++last_sg <= sg) {
621 			DBGA("        (%ld) [%p,%x] cont\n",
622 			     last_sg - leader, SG_ENT_VIRT_ADDRESS(last_sg),
623 			     last_sg->length);
624 		}
625 #endif
626 	} while (++sg < end && (int) sg->dma_address < 0);
627 
628 	return 1;
629 }
630 
631 static int alpha_pci_map_sg(struct device *dev, struct scatterlist *sg,
632 			    int nents, enum dma_data_direction dir,
633 			    unsigned long attrs)
634 {
635 	struct pci_dev *pdev = alpha_gendev_to_pci(dev);
636 	struct scatterlist *start, *end, *out;
637 	struct pci_controller *hose;
638 	struct pci_iommu_arena *arena;
639 	dma_addr_t max_dma;
640 	int dac_allowed;
641 
642 	BUG_ON(dir == PCI_DMA_NONE);
643 
644 	dac_allowed = dev ? pci_dac_dma_supported(pdev, pdev->dma_mask) : 0;
645 
646 	/* Fast path single entry scatterlists.  */
647 	if (nents == 1) {
648 		sg->dma_length = sg->length;
649 		sg->dma_address
650 		  = pci_map_single_1(pdev, SG_ENT_VIRT_ADDRESS(sg),
651 				     sg->length, dac_allowed);
652 		return sg->dma_address != DMA_MAPPING_ERROR;
653 	}
654 
655 	start = sg;
656 	end = sg + nents;
657 
658 	/* First, prepare information about the entries.  */
659 	sg_classify(dev, sg, end, alpha_mv.mv_pci_tbi != 0);
660 
661 	/* Second, figure out where we're going to map things.  */
662 	if (alpha_mv.mv_pci_tbi) {
663 		hose = pdev ? pdev->sysdata : pci_isa_hose;
664 		max_dma = pdev ? pdev->dma_mask : ISA_DMA_MASK;
665 		arena = hose->sg_pci;
666 		if (!arena || arena->dma_base + arena->size - 1 > max_dma)
667 			arena = hose->sg_isa;
668 	} else {
669 		max_dma = -1;
670 		arena = NULL;
671 		hose = NULL;
672 	}
673 
674 	/* Third, iterate over the scatterlist leaders and allocate
675 	   dma space as needed.  */
676 	for (out = sg; sg < end; ++sg) {
677 		if ((int) sg->dma_address < 0)
678 			continue;
679 		if (sg_fill(dev, sg, end, out, arena, max_dma, dac_allowed) < 0)
680 			goto error;
681 		out++;
682 	}
683 
684 	/* Mark the end of the list for pci_unmap_sg.  */
685 	if (out < end)
686 		out->dma_length = 0;
687 
688 	if (out - start == 0)
689 		printk(KERN_WARNING "pci_map_sg failed: no entries?\n");
690 	DBGA("pci_map_sg: %ld entries\n", out - start);
691 
692 	return out - start;
693 
694  error:
695 	printk(KERN_WARNING "pci_map_sg failed: "
696 	       "could not allocate dma page tables\n");
697 
698 	/* Some allocation failed while mapping the scatterlist
699 	   entries.  Unmap them now.  */
700 	if (out > start)
701 		pci_unmap_sg(pdev, start, out - start, dir);
702 	return 0;
703 }
704 
705 /* Unmap a set of streaming mode DMA translations.  Again, cpu read
706    rules concerning calls here are the same as for pci_unmap_single()
707    above.  */
708 
709 static void alpha_pci_unmap_sg(struct device *dev, struct scatterlist *sg,
710 			       int nents, enum dma_data_direction dir,
711 			       unsigned long attrs)
712 {
713 	struct pci_dev *pdev = alpha_gendev_to_pci(dev);
714 	unsigned long flags;
715 	struct pci_controller *hose;
716 	struct pci_iommu_arena *arena;
717 	struct scatterlist *end;
718 	dma_addr_t max_dma;
719 	dma_addr_t fbeg, fend;
720 
721 	BUG_ON(dir == PCI_DMA_NONE);
722 
723 	if (! alpha_mv.mv_pci_tbi)
724 		return;
725 
726 	hose = pdev ? pdev->sysdata : pci_isa_hose;
727 	max_dma = pdev ? pdev->dma_mask : ISA_DMA_MASK;
728 	arena = hose->sg_pci;
729 	if (!arena || arena->dma_base + arena->size - 1 > max_dma)
730 		arena = hose->sg_isa;
731 
732 	fbeg = -1, fend = 0;
733 
734 	spin_lock_irqsave(&arena->lock, flags);
735 
736 	for (end = sg + nents; sg < end; ++sg) {
737 		dma_addr_t addr;
738 		size_t size;
739 		long npages, ofs;
740 		dma_addr_t tend;
741 
742 		addr = sg->dma_address;
743 		size = sg->dma_length;
744 		if (!size)
745 			break;
746 
747 		if (addr > 0xffffffff) {
748 			/* It's a DAC address -- nothing to do.  */
749 			DBGA("    (%ld) DAC [%llx,%zx]\n",
750 			      sg - end + nents, addr, size);
751 			continue;
752 		}
753 
754 		if (addr >= __direct_map_base
755 		    && addr < __direct_map_base + __direct_map_size) {
756 			/* Nothing to do.  */
757 			DBGA("    (%ld) direct [%llx,%zx]\n",
758 			      sg - end + nents, addr, size);
759 			continue;
760 		}
761 
762 		DBGA("    (%ld) sg [%llx,%zx]\n",
763 		     sg - end + nents, addr, size);
764 
765 		npages = iommu_num_pages(addr, size, PAGE_SIZE);
766 		ofs = (addr - arena->dma_base) >> PAGE_SHIFT;
767 		iommu_arena_free(arena, ofs, npages);
768 
769 		tend = addr + size - 1;
770 		if (fbeg > addr) fbeg = addr;
771 		if (fend < tend) fend = tend;
772 	}
773 
774         /* If we're freeing ptes above the `next_entry' pointer (they
775            may have snuck back into the TLB since the last wrap flush),
776            we need to flush the TLB before reallocating the latter.  */
777 	if ((fend - arena->dma_base) >> PAGE_SHIFT >= arena->next_entry)
778 		alpha_mv.mv_pci_tbi(hose, fbeg, fend);
779 
780 	spin_unlock_irqrestore(&arena->lock, flags);
781 
782 	DBGA("pci_unmap_sg: %ld entries\n", nents - (end - sg));
783 }
784 
785 /* Return whether the given PCI device DMA address mask can be
786    supported properly.  */
787 
788 static int alpha_pci_supported(struct device *dev, u64 mask)
789 {
790 	struct pci_dev *pdev = alpha_gendev_to_pci(dev);
791 	struct pci_controller *hose;
792 	struct pci_iommu_arena *arena;
793 
794 	/* If there exists a direct map, and the mask fits either
795 	   the entire direct mapped space or the total system memory as
796 	   shifted by the map base */
797 	if (__direct_map_size != 0
798 	    && (__direct_map_base + __direct_map_size - 1 <= mask ||
799 		__direct_map_base + (max_low_pfn << PAGE_SHIFT) - 1 <= mask))
800 		return 1;
801 
802 	/* Check that we have a scatter-gather arena that fits.  */
803 	hose = pdev ? pdev->sysdata : pci_isa_hose;
804 	arena = hose->sg_isa;
805 	if (arena && arena->dma_base + arena->size - 1 <= mask)
806 		return 1;
807 	arena = hose->sg_pci;
808 	if (arena && arena->dma_base + arena->size - 1 <= mask)
809 		return 1;
810 
811 	/* As last resort try ZONE_DMA.  */
812 	if (!__direct_map_base && MAX_DMA_ADDRESS - IDENT_ADDR - 1 <= mask)
813 		return 1;
814 
815 	return 0;
816 }
817 
818 
819 /*
820  * AGP GART extensions to the IOMMU
821  */
822 int
823 iommu_reserve(struct pci_iommu_arena *arena, long pg_count, long align_mask)
824 {
825 	unsigned long flags;
826 	unsigned long *ptes;
827 	long i, p;
828 
829 	if (!arena) return -EINVAL;
830 
831 	spin_lock_irqsave(&arena->lock, flags);
832 
833 	/* Search for N empty ptes.  */
834 	ptes = arena->ptes;
835 	p = iommu_arena_find_pages(NULL, arena, pg_count, align_mask);
836 	if (p < 0) {
837 		spin_unlock_irqrestore(&arena->lock, flags);
838 		return -1;
839 	}
840 
841 	/* Success.  Mark them all reserved (ie not zero and invalid)
842 	   for the iommu tlb that could load them from under us.
843 	   They will be filled in with valid bits by _bind() */
844 	for (i = 0; i < pg_count; ++i)
845 		ptes[p+i] = IOMMU_RESERVED_PTE;
846 
847 	arena->next_entry = p + pg_count;
848 	spin_unlock_irqrestore(&arena->lock, flags);
849 
850 	return p;
851 }
852 
853 int
854 iommu_release(struct pci_iommu_arena *arena, long pg_start, long pg_count)
855 {
856 	unsigned long *ptes;
857 	long i;
858 
859 	if (!arena) return -EINVAL;
860 
861 	ptes = arena->ptes;
862 
863 	/* Make sure they're all reserved first... */
864 	for(i = pg_start; i < pg_start + pg_count; i++)
865 		if (ptes[i] != IOMMU_RESERVED_PTE)
866 			return -EBUSY;
867 
868 	iommu_arena_free(arena, pg_start, pg_count);
869 	return 0;
870 }
871 
872 int
873 iommu_bind(struct pci_iommu_arena *arena, long pg_start, long pg_count,
874 	   struct page **pages)
875 {
876 	unsigned long flags;
877 	unsigned long *ptes;
878 	long i, j;
879 
880 	if (!arena) return -EINVAL;
881 
882 	spin_lock_irqsave(&arena->lock, flags);
883 
884 	ptes = arena->ptes;
885 
886 	for(j = pg_start; j < pg_start + pg_count; j++) {
887 		if (ptes[j] != IOMMU_RESERVED_PTE) {
888 			spin_unlock_irqrestore(&arena->lock, flags);
889 			return -EBUSY;
890 		}
891 	}
892 
893 	for(i = 0, j = pg_start; i < pg_count; i++, j++)
894 		ptes[j] = mk_iommu_pte(page_to_phys(pages[i]));
895 
896 	spin_unlock_irqrestore(&arena->lock, flags);
897 
898 	return 0;
899 }
900 
901 int
902 iommu_unbind(struct pci_iommu_arena *arena, long pg_start, long pg_count)
903 {
904 	unsigned long *p;
905 	long i;
906 
907 	if (!arena) return -EINVAL;
908 
909 	p = arena->ptes + pg_start;
910 	for(i = 0; i < pg_count; i++)
911 		p[i] = IOMMU_RESERVED_PTE;
912 
913 	return 0;
914 }
915 
916 const struct dma_map_ops alpha_pci_ops = {
917 	.alloc			= alpha_pci_alloc_coherent,
918 	.free			= alpha_pci_free_coherent,
919 	.map_page		= alpha_pci_map_page,
920 	.unmap_page		= alpha_pci_unmap_page,
921 	.map_sg			= alpha_pci_map_sg,
922 	.unmap_sg		= alpha_pci_unmap_sg,
923 	.dma_supported		= alpha_pci_supported,
924 	.mmap			= dma_common_mmap,
925 	.get_sgtable		= dma_common_get_sgtable,
926 	.alloc_pages		= dma_common_alloc_pages,
927 	.free_pages		= dma_common_free_pages,
928 };
929 EXPORT_SYMBOL(alpha_pci_ops);
930