aboutsummaryrefslogtreecommitdiffstats
path: root/arch
diff options
context:
space:
mode:
authorJoerg Roedel <joerg.roedel@amd.com>2008-07-11 11:14:22 -0400
committerIngo Molnar <mingo@elte.hu>2008-07-11 12:01:10 -0400
commit431b2a2015337533f1a9e39a840266a8a2c93144 (patch)
tree300bc72db7e995ecaf8b3c00981773acb9289067 /arch
parentb65233a9c1da587bf19ee161982f4f0ec59941c0 (diff)
x86, AMD IOMMU: add comments to core code
This patch adds comments about how the AMD IOMMU core code works for the DMA remapping functionality. Signed-off-by: Joerg Roedel <joerg.roedel@amd.com> Cc: iommu@lists.linux-foundation.org Cc: bhavna.sarathy@amd.com Cc: robert.richter@amd.com Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'arch')
-rw-r--r--arch/x86/kernel/amd_iommu.c201
1 files changed, 199 insertions, 2 deletions
diff --git a/arch/x86/kernel/amd_iommu.c b/arch/x86/kernel/amd_iommu.c
index f2766d84c7a0..4bae96ca7c11 100644
--- a/arch/x86/kernel/amd_iommu.c
+++ b/arch/x86/kernel/amd_iommu.c
@@ -34,6 +34,9 @@
34 34
35static DEFINE_RWLOCK(amd_iommu_devtable_lock); 35static DEFINE_RWLOCK(amd_iommu_devtable_lock);
36 36
37/*
38 * general struct to manage commands send to an IOMMU
39 */
37struct command { 40struct command {
38 u32 data[4]; 41 u32 data[4];
39}; 42};
@@ -41,11 +44,22 @@ struct command {
41static int dma_ops_unity_map(struct dma_ops_domain *dma_dom, 44static int dma_ops_unity_map(struct dma_ops_domain *dma_dom,
42 struct unity_map_entry *e); 45 struct unity_map_entry *e);
43 46
47/* returns !0 if the IOMMU is caching non-present entries in its TLB */
44static int iommu_has_npcache(struct amd_iommu *iommu) 48static int iommu_has_npcache(struct amd_iommu *iommu)
45{ 49{
46 return iommu->cap & IOMMU_CAP_NPCACHE; 50 return iommu->cap & IOMMU_CAP_NPCACHE;
47} 51}
48 52
53/****************************************************************************
54 *
55 * IOMMU command queuing functions
56 *
57 ****************************************************************************/
58
59/*
60 * Writes the command to the IOMMUs command buffer and informs the
61 * hardware about the new command. Must be called with iommu->lock held.
62 */
49static int __iommu_queue_command(struct amd_iommu *iommu, struct command *cmd) 63static int __iommu_queue_command(struct amd_iommu *iommu, struct command *cmd)
50{ 64{
51 u32 tail, head; 65 u32 tail, head;
@@ -63,6 +77,10 @@ static int __iommu_queue_command(struct amd_iommu *iommu, struct command *cmd)
63 return 0; 77 return 0;
64} 78}
65 79
80/*
81 * General queuing function for commands. Takes iommu->lock and calls
82 * __iommu_queue_command().
83 */
66static int iommu_queue_command(struct amd_iommu *iommu, struct command *cmd) 84static int iommu_queue_command(struct amd_iommu *iommu, struct command *cmd)
67{ 85{
68 unsigned long flags; 86 unsigned long flags;
@@ -75,6 +93,13 @@ static int iommu_queue_command(struct amd_iommu *iommu, struct command *cmd)
75 return ret; 93 return ret;
76} 94}
77 95
96/*
97 * This function is called whenever we need to ensure that the IOMMU has
98 * completed execution of all commands we sent. It sends a
99 * COMPLETION_WAIT command and waits for it to finish. The IOMMU informs
100 * us about that by writing a value to a physical address we pass with
101 * the command.
102 */
78static int iommu_completion_wait(struct amd_iommu *iommu) 103static int iommu_completion_wait(struct amd_iommu *iommu)
79{ 104{
80 int ret; 105 int ret;
@@ -101,6 +126,9 @@ static int iommu_completion_wait(struct amd_iommu *iommu)
101 return 0; 126 return 0;
102} 127}
103 128
129/*
130 * Command send function for invalidating a device table entry
131 */
104static int iommu_queue_inv_dev_entry(struct amd_iommu *iommu, u16 devid) 132static int iommu_queue_inv_dev_entry(struct amd_iommu *iommu, u16 devid)
105{ 133{
106 struct command cmd; 134 struct command cmd;
@@ -116,6 +144,9 @@ static int iommu_queue_inv_dev_entry(struct amd_iommu *iommu, u16 devid)
116 return iommu_queue_command(iommu, &cmd); 144 return iommu_queue_command(iommu, &cmd);
117} 145}
118 146
147/*
148 * Generic command send function for invalidaing TLB entries
149 */
119static int iommu_queue_inv_iommu_pages(struct amd_iommu *iommu, 150static int iommu_queue_inv_iommu_pages(struct amd_iommu *iommu,
120 u64 address, u16 domid, int pde, int s) 151 u64 address, u16 domid, int pde, int s)
121{ 152{
@@ -127,9 +158,9 @@ static int iommu_queue_inv_iommu_pages(struct amd_iommu *iommu,
127 cmd.data[1] |= domid; 158 cmd.data[1] |= domid;
128 cmd.data[2] = LOW_U32(address); 159 cmd.data[2] = LOW_U32(address);
129 cmd.data[3] = HIGH_U32(address); 160 cmd.data[3] = HIGH_U32(address);
130 if (s) 161 if (s) /* size bit - we flush more than one 4kb page */
131 cmd.data[2] |= CMD_INV_IOMMU_PAGES_SIZE_MASK; 162 cmd.data[2] |= CMD_INV_IOMMU_PAGES_SIZE_MASK;
132 if (pde) 163 if (pde) /* PDE bit - we wan't flush everything not only the PTEs */
133 cmd.data[2] |= CMD_INV_IOMMU_PAGES_PDE_MASK; 164 cmd.data[2] |= CMD_INV_IOMMU_PAGES_PDE_MASK;
134 165
135 iommu->need_sync = 1; 166 iommu->need_sync = 1;
@@ -137,6 +168,11 @@ static int iommu_queue_inv_iommu_pages(struct amd_iommu *iommu,
137 return iommu_queue_command(iommu, &cmd); 168 return iommu_queue_command(iommu, &cmd);
138} 169}
139 170
171/*
172 * TLB invalidation function which is called from the mapping functions.
173 * It invalidates a single PTE if the range to flush is within a single
174 * page. Otherwise it flushes the whole TLB of the IOMMU.
175 */
140static int iommu_flush_pages(struct amd_iommu *iommu, u16 domid, 176static int iommu_flush_pages(struct amd_iommu *iommu, u16 domid,
141 u64 address, size_t size) 177 u64 address, size_t size)
142{ 178{
@@ -159,6 +195,20 @@ static int iommu_flush_pages(struct amd_iommu *iommu, u16 domid,
159 return 0; 195 return 0;
160} 196}
161 197
198/****************************************************************************
199 *
200 * The functions below are used the create the page table mappings for
201 * unity mapped regions.
202 *
203 ****************************************************************************/
204
205/*
206 * Generic mapping functions. It maps a physical address into a DMA
207 * address space. It allocates the page table pages if necessary.
208 * In the future it can be extended to a generic mapping function
209 * supporting all features of AMD IOMMU page tables like level skipping
210 * and full 64 bit address spaces.
211 */
162static int iommu_map(struct protection_domain *dom, 212static int iommu_map(struct protection_domain *dom,
163 unsigned long bus_addr, 213 unsigned long bus_addr,
164 unsigned long phys_addr, 214 unsigned long phys_addr,
@@ -209,6 +259,10 @@ static int iommu_map(struct protection_domain *dom,
209 return 0; 259 return 0;
210} 260}
211 261
262/*
263 * This function checks if a specific unity mapping entry is needed for
264 * this specific IOMMU.
265 */
212static int iommu_for_unity_map(struct amd_iommu *iommu, 266static int iommu_for_unity_map(struct amd_iommu *iommu,
213 struct unity_map_entry *entry) 267 struct unity_map_entry *entry)
214{ 268{
@@ -223,6 +277,12 @@ static int iommu_for_unity_map(struct amd_iommu *iommu,
223 return 0; 277 return 0;
224} 278}
225 279
280/*
281 * Init the unity mappings for a specific IOMMU in the system
282 *
283 * Basically iterates over all unity mapping entries and applies them to
284 * the default domain DMA of that IOMMU if necessary.
285 */
226static int iommu_init_unity_mappings(struct amd_iommu *iommu) 286static int iommu_init_unity_mappings(struct amd_iommu *iommu)
227{ 287{
228 struct unity_map_entry *entry; 288 struct unity_map_entry *entry;
@@ -239,6 +299,10 @@ static int iommu_init_unity_mappings(struct amd_iommu *iommu)
239 return 0; 299 return 0;
240} 300}
241 301
302/*
303 * This function actually applies the mapping to the page table of the
304 * dma_ops domain.
305 */
242static int dma_ops_unity_map(struct dma_ops_domain *dma_dom, 306static int dma_ops_unity_map(struct dma_ops_domain *dma_dom,
243 struct unity_map_entry *e) 307 struct unity_map_entry *e)
244{ 308{
@@ -261,6 +325,9 @@ static int dma_ops_unity_map(struct dma_ops_domain *dma_dom,
261 return 0; 325 return 0;
262} 326}
263 327
328/*
329 * Inits the unity mappings required for a specific device
330 */
264static int init_unity_mappings_for_device(struct dma_ops_domain *dma_dom, 331static int init_unity_mappings_for_device(struct dma_ops_domain *dma_dom,
265 u16 devid) 332 u16 devid)
266{ 333{
@@ -278,12 +345,26 @@ static int init_unity_mappings_for_device(struct dma_ops_domain *dma_dom,
278 return 0; 345 return 0;
279} 346}
280 347
348/****************************************************************************
349 *
350 * The next functions belong to the address allocator for the dma_ops
351 * interface functions. They work like the allocators in the other IOMMU
352 * drivers. Its basically a bitmap which marks the allocated pages in
353 * the aperture. Maybe it could be enhanced in the future to a more
354 * efficient allocator.
355 *
356 ****************************************************************************/
281static unsigned long dma_mask_to_pages(unsigned long mask) 357static unsigned long dma_mask_to_pages(unsigned long mask)
282{ 358{
283 return (mask >> PAGE_SHIFT) + 359 return (mask >> PAGE_SHIFT) +
284 (PAGE_ALIGN(mask & ~PAGE_MASK) >> PAGE_SHIFT); 360 (PAGE_ALIGN(mask & ~PAGE_MASK) >> PAGE_SHIFT);
285} 361}
286 362
363/*
364 * The address allocator core function.
365 *
366 * called with domain->lock held
367 */
287static unsigned long dma_ops_alloc_addresses(struct device *dev, 368static unsigned long dma_ops_alloc_addresses(struct device *dev,
288 struct dma_ops_domain *dom, 369 struct dma_ops_domain *dom,
289 unsigned int pages) 370 unsigned int pages)
@@ -317,6 +398,11 @@ static unsigned long dma_ops_alloc_addresses(struct device *dev,
317 return address; 398 return address;
318} 399}
319 400
401/*
402 * The address free function.
403 *
404 * called with domain->lock held
405 */
320static void dma_ops_free_addresses(struct dma_ops_domain *dom, 406static void dma_ops_free_addresses(struct dma_ops_domain *dom,
321 unsigned long address, 407 unsigned long address,
322 unsigned int pages) 408 unsigned int pages)
@@ -325,6 +411,16 @@ static void dma_ops_free_addresses(struct dma_ops_domain *dom,
325 iommu_area_free(dom->bitmap, address, pages); 411 iommu_area_free(dom->bitmap, address, pages);
326} 412}
327 413
414/****************************************************************************
415 *
416 * The next functions belong to the domain allocation. A domain is
417 * allocated for every IOMMU as the default domain. If device isolation
418 * is enabled, every device get its own domain. The most important thing
419 * about domains is the page table mapping the DMA address space they
420 * contain.
421 *
422 ****************************************************************************/
423
328static u16 domain_id_alloc(void) 424static u16 domain_id_alloc(void)
329{ 425{
330 unsigned long flags; 426 unsigned long flags;
@@ -342,6 +438,10 @@ static u16 domain_id_alloc(void)
342 return id; 438 return id;
343} 439}
344 440
441/*
442 * Used to reserve address ranges in the aperture (e.g. for exclusion
443 * ranges.
444 */
345static void dma_ops_reserve_addresses(struct dma_ops_domain *dom, 445static void dma_ops_reserve_addresses(struct dma_ops_domain *dom,
346 unsigned long start_page, 446 unsigned long start_page,
347 unsigned int pages) 447 unsigned int pages)
@@ -382,6 +482,10 @@ static void dma_ops_free_pagetable(struct dma_ops_domain *dma_dom)
382 free_page((unsigned long)p1); 482 free_page((unsigned long)p1);
383} 483}
384 484
485/*
486 * Free a domain, only used if something went wrong in the
487 * allocation path and we need to free an already allocated page table
488 */
385static void dma_ops_domain_free(struct dma_ops_domain *dom) 489static void dma_ops_domain_free(struct dma_ops_domain *dom)
386{ 490{
387 if (!dom) 491 if (!dom)
@@ -396,6 +500,11 @@ static void dma_ops_domain_free(struct dma_ops_domain *dom)
396 kfree(dom); 500 kfree(dom);
397} 501}
398 502
503/*
504 * Allocates a new protection domain usable for the dma_ops functions.
505 * It also intializes the page table and the address allocator data
506 * structures required for the dma_ops interface
507 */
399static struct dma_ops_domain *dma_ops_domain_alloc(struct amd_iommu *iommu, 508static struct dma_ops_domain *dma_ops_domain_alloc(struct amd_iommu *iommu,
400 unsigned order) 509 unsigned order)
401{ 510{
@@ -436,6 +545,7 @@ static struct dma_ops_domain *dma_ops_domain_alloc(struct amd_iommu *iommu,
436 dma_dom->bitmap[0] = 1; 545 dma_dom->bitmap[0] = 1;
437 dma_dom->next_bit = 0; 546 dma_dom->next_bit = 0;
438 547
548 /* Intialize the exclusion range if necessary */
439 if (iommu->exclusion_start && 549 if (iommu->exclusion_start &&
440 iommu->exclusion_start < dma_dom->aperture_size) { 550 iommu->exclusion_start < dma_dom->aperture_size) {
441 unsigned long startpage = iommu->exclusion_start >> PAGE_SHIFT; 551 unsigned long startpage = iommu->exclusion_start >> PAGE_SHIFT;
@@ -444,6 +554,11 @@ static struct dma_ops_domain *dma_ops_domain_alloc(struct amd_iommu *iommu,
444 dma_ops_reserve_addresses(dma_dom, startpage, pages); 554 dma_ops_reserve_addresses(dma_dom, startpage, pages);
445 } 555 }
446 556
557 /*
558 * At the last step, build the page tables so we don't need to
559 * allocate page table pages in the dma_ops mapping/unmapping
560 * path.
561 */
447 num_pte_pages = dma_dom->aperture_size / (PAGE_SIZE * 512); 562 num_pte_pages = dma_dom->aperture_size / (PAGE_SIZE * 512);
448 dma_dom->pte_pages = kzalloc(num_pte_pages * sizeof(void *), 563 dma_dom->pte_pages = kzalloc(num_pte_pages * sizeof(void *),
449 GFP_KERNEL); 564 GFP_KERNEL);
@@ -472,6 +587,10 @@ free_dma_dom:
472 return NULL; 587 return NULL;
473} 588}
474 589
590/*
591 * Find out the protection domain structure for a given PCI device. This
592 * will give us the pointer to the page table root for example.
593 */
475static struct protection_domain *domain_for_device(u16 devid) 594static struct protection_domain *domain_for_device(u16 devid)
476{ 595{
477 struct protection_domain *dom; 596 struct protection_domain *dom;
@@ -484,6 +603,10 @@ static struct protection_domain *domain_for_device(u16 devid)
484 return dom; 603 return dom;
485} 604}
486 605
606/*
607 * If a device is not yet associated with a domain, this function does
608 * assigns it visible for the hardware
609 */
487static void set_device_domain(struct amd_iommu *iommu, 610static void set_device_domain(struct amd_iommu *iommu,
488 struct protection_domain *domain, 611 struct protection_domain *domain,
489 u16 devid) 612 u16 devid)
@@ -508,6 +631,19 @@ static void set_device_domain(struct amd_iommu *iommu,
508 iommu->need_sync = 1; 631 iommu->need_sync = 1;
509} 632}
510 633
634/*****************************************************************************
635 *
636 * The next functions belong to the dma_ops mapping/unmapping code.
637 *
638 *****************************************************************************/
639
640/*
641 * In the dma_ops path we only have the struct device. This function
642 * finds the corresponding IOMMU, the protection domain and the
643 * requestor id for a given device.
644 * If the device is not yet associated with a domain this is also done
645 * in this function.
646 */
511static int get_device_resources(struct device *dev, 647static int get_device_resources(struct device *dev,
512 struct amd_iommu **iommu, 648 struct amd_iommu **iommu,
513 struct protection_domain **domain, 649 struct protection_domain **domain,
@@ -522,6 +658,7 @@ static int get_device_resources(struct device *dev,
522 pcidev = to_pci_dev(dev); 658 pcidev = to_pci_dev(dev);
523 _bdf = (pcidev->bus->number << 8) | pcidev->devfn; 659 _bdf = (pcidev->bus->number << 8) | pcidev->devfn;
524 660
661 /* device not translated by any IOMMU in the system? */
525 if (_bdf >= amd_iommu_last_bdf) { 662 if (_bdf >= amd_iommu_last_bdf) {
526 *iommu = NULL; 663 *iommu = NULL;
527 *domain = NULL; 664 *domain = NULL;
@@ -547,6 +684,10 @@ static int get_device_resources(struct device *dev,
547 return 1; 684 return 1;
548} 685}
549 686
687/*
688 * This is the generic map function. It maps one 4kb page at paddr to
689 * the given address in the DMA address space for the domain.
690 */
550static dma_addr_t dma_ops_domain_map(struct amd_iommu *iommu, 691static dma_addr_t dma_ops_domain_map(struct amd_iommu *iommu,
551 struct dma_ops_domain *dom, 692 struct dma_ops_domain *dom,
552 unsigned long address, 693 unsigned long address,
@@ -578,6 +719,9 @@ static dma_addr_t dma_ops_domain_map(struct amd_iommu *iommu,
578 return (dma_addr_t)address; 719 return (dma_addr_t)address;
579} 720}
580 721
722/*
723 * The generic unmapping function for on page in the DMA address space.
724 */
581static void dma_ops_domain_unmap(struct amd_iommu *iommu, 725static void dma_ops_domain_unmap(struct amd_iommu *iommu,
582 struct dma_ops_domain *dom, 726 struct dma_ops_domain *dom,
583 unsigned long address) 727 unsigned long address)
@@ -597,6 +741,12 @@ static void dma_ops_domain_unmap(struct amd_iommu *iommu,
597 *pte = 0ULL; 741 *pte = 0ULL;
598} 742}
599 743
744/*
745 * This function contains common code for mapping of a physically
746 * contiguous memory region into DMA address space. It is uses by all
747 * mapping functions provided by this IOMMU driver.
748 * Must be called with the domain lock held.
749 */
600static dma_addr_t __map_single(struct device *dev, 750static dma_addr_t __map_single(struct device *dev,
601 struct amd_iommu *iommu, 751 struct amd_iommu *iommu,
602 struct dma_ops_domain *dma_dom, 752 struct dma_ops_domain *dma_dom,
@@ -628,6 +778,10 @@ out:
628 return address; 778 return address;
629} 779}
630 780
781/*
782 * Does the reverse of the __map_single function. Must be called with
783 * the domain lock held too
784 */
631static void __unmap_single(struct amd_iommu *iommu, 785static void __unmap_single(struct amd_iommu *iommu,
632 struct dma_ops_domain *dma_dom, 786 struct dma_ops_domain *dma_dom,
633 dma_addr_t dma_addr, 787 dma_addr_t dma_addr,
@@ -652,6 +806,9 @@ static void __unmap_single(struct amd_iommu *iommu,
652 dma_ops_free_addresses(dma_dom, dma_addr, pages); 806 dma_ops_free_addresses(dma_dom, dma_addr, pages);
653} 807}
654 808
809/*
810 * The exported map_single function for dma_ops.
811 */
655static dma_addr_t map_single(struct device *dev, phys_addr_t paddr, 812static dma_addr_t map_single(struct device *dev, phys_addr_t paddr,
656 size_t size, int dir) 813 size_t size, int dir)
657{ 814{
@@ -664,6 +821,7 @@ static dma_addr_t map_single(struct device *dev, phys_addr_t paddr,
664 get_device_resources(dev, &iommu, &domain, &devid); 821 get_device_resources(dev, &iommu, &domain, &devid);
665 822
666 if (iommu == NULL || domain == NULL) 823 if (iommu == NULL || domain == NULL)
824 /* device not handled by any AMD IOMMU */
667 return (dma_addr_t)paddr; 825 return (dma_addr_t)paddr;
668 826
669 spin_lock_irqsave(&domain->lock, flags); 827 spin_lock_irqsave(&domain->lock, flags);
@@ -683,6 +841,9 @@ out:
683 return addr; 841 return addr;
684} 842}
685 843
844/*
845 * The exported unmap_single function for dma_ops.
846 */
686static void unmap_single(struct device *dev, dma_addr_t dma_addr, 847static void unmap_single(struct device *dev, dma_addr_t dma_addr,
687 size_t size, int dir) 848 size_t size, int dir)
688{ 849{
@@ -692,6 +853,7 @@ static void unmap_single(struct device *dev, dma_addr_t dma_addr,
692 u16 devid; 853 u16 devid;
693 854
694 if (!get_device_resources(dev, &iommu, &domain, &devid)) 855 if (!get_device_resources(dev, &iommu, &domain, &devid))
856 /* device not handled by any AMD IOMMU */
695 return; 857 return;
696 858
697 spin_lock_irqsave(&domain->lock, flags); 859 spin_lock_irqsave(&domain->lock, flags);
@@ -706,6 +868,10 @@ static void unmap_single(struct device *dev, dma_addr_t dma_addr,
706 spin_unlock_irqrestore(&domain->lock, flags); 868 spin_unlock_irqrestore(&domain->lock, flags);
707} 869}
708 870
871/*
872 * This is a special map_sg function which is used if we should map a
873 * device which is not handled by an AMD IOMMU in the system.
874 */
709static int map_sg_no_iommu(struct device *dev, struct scatterlist *sglist, 875static int map_sg_no_iommu(struct device *dev, struct scatterlist *sglist,
710 int nelems, int dir) 876 int nelems, int dir)
711{ 877{
@@ -720,6 +886,10 @@ static int map_sg_no_iommu(struct device *dev, struct scatterlist *sglist,
720 return nelems; 886 return nelems;
721} 887}
722 888
889/*
890 * The exported map_sg function for dma_ops (handles scatter-gather
891 * lists).
892 */
723static int map_sg(struct device *dev, struct scatterlist *sglist, 893static int map_sg(struct device *dev, struct scatterlist *sglist,
724 int nelems, int dir) 894 int nelems, int dir)
725{ 895{
@@ -775,6 +945,10 @@ unmap:
775 goto out; 945 goto out;
776} 946}
777 947
948/*
949 * The exported map_sg function for dma_ops (handles scatter-gather
950 * lists).
951 */
778static void unmap_sg(struct device *dev, struct scatterlist *sglist, 952static void unmap_sg(struct device *dev, struct scatterlist *sglist,
779 int nelems, int dir) 953 int nelems, int dir)
780{ 954{
@@ -804,6 +978,9 @@ static void unmap_sg(struct device *dev, struct scatterlist *sglist,
804 spin_unlock_irqrestore(&domain->lock, flags); 978 spin_unlock_irqrestore(&domain->lock, flags);
805} 979}
806 980
981/*
982 * The exported alloc_coherent function for dma_ops.
983 */
807static void *alloc_coherent(struct device *dev, size_t size, 984static void *alloc_coherent(struct device *dev, size_t size,
808 dma_addr_t *dma_addr, gfp_t flag) 985 dma_addr_t *dma_addr, gfp_t flag)
809{ 986{
@@ -851,6 +1028,11 @@ out:
851 return virt_addr; 1028 return virt_addr;
852} 1029}
853 1030
1031/*
1032 * The exported free_coherent function for dma_ops.
1033 * FIXME: fix the generic x86 DMA layer so that it actually calls that
1034 * function.
1035 */
854static void free_coherent(struct device *dev, size_t size, 1036static void free_coherent(struct device *dev, size_t size,
855 void *virt_addr, dma_addr_t dma_addr) 1037 void *virt_addr, dma_addr_t dma_addr)
856{ 1038{
@@ -879,6 +1061,8 @@ free_mem:
879} 1061}
880 1062
881/* 1063/*
1064 * The function for pre-allocating protection domains.
1065 *
882 * If the driver core informs the DMA layer if a driver grabs a device 1066 * If the driver core informs the DMA layer if a driver grabs a device
883 * we don't need to preallocate the protection domains anymore. 1067 * we don't need to preallocate the protection domains anymore.
884 * For now we have to. 1068 * For now we have to.
@@ -921,12 +1105,20 @@ static struct dma_mapping_ops amd_iommu_dma_ops = {
921 .unmap_sg = unmap_sg, 1105 .unmap_sg = unmap_sg,
922}; 1106};
923 1107
1108/*
1109 * The function which clues the AMD IOMMU driver into dma_ops.
1110 */
924int __init amd_iommu_init_dma_ops(void) 1111int __init amd_iommu_init_dma_ops(void)
925{ 1112{
926 struct amd_iommu *iommu; 1113 struct amd_iommu *iommu;
927 int order = amd_iommu_aperture_order; 1114 int order = amd_iommu_aperture_order;
928 int ret; 1115 int ret;
929 1116
1117 /*
1118 * first allocate a default protection domain for every IOMMU we
1119 * found in the system. Devices not assigned to any other
1120 * protection domain will be assigned to the default one.
1121 */
930 list_for_each_entry(iommu, &amd_iommu_list, list) { 1122 list_for_each_entry(iommu, &amd_iommu_list, list) {
931 iommu->default_dom = dma_ops_domain_alloc(iommu, order); 1123 iommu->default_dom = dma_ops_domain_alloc(iommu, order);
932 if (iommu->default_dom == NULL) 1124 if (iommu->default_dom == NULL)
@@ -936,6 +1128,10 @@ int __init amd_iommu_init_dma_ops(void)
936 goto free_domains; 1128 goto free_domains;
937 } 1129 }
938 1130
1131 /*
1132 * If device isolation is enabled, pre-allocate the protection
1133 * domains for each device.
1134 */
939 if (amd_iommu_isolate) 1135 if (amd_iommu_isolate)
940 prealloc_protection_domains(); 1136 prealloc_protection_domains();
941 1137
@@ -947,6 +1143,7 @@ int __init amd_iommu_init_dma_ops(void)
947 gart_iommu_aperture = 0; 1143 gart_iommu_aperture = 0;
948#endif 1144#endif
949 1145
1146 /* Make the driver finally visible to the drivers */
950 dma_ops = &amd_iommu_dma_ops; 1147 dma_ops = &amd_iommu_dma_ops;
951 1148
952 return 0; 1149 return 0;