aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorFUJITA Tomonori <fujita.tomonori@lab.ntt.co.jp>2009-01-05 09:47:26 -0500
committerIngo Molnar <mingo@elte.hu>2009-01-06 08:06:55 -0500
commitffbbef5c0639dbf31a0511d8f65f574bb3e30758 (patch)
tree47ab8be07ea12212e35460c56dc140e998892c98
parent51491367c2541c51a9d435eec88b0e846223fb59 (diff)
intel-iommu: add map_page and unmap_page
This is a preparation of struct dma_mapping_ops unification. We use map_page and unmap_page instead of map_single and unmap_single. This uses a temporary workaround, ifdef X86_64 to avoid IA64 build. The workaround will be removed after the unification. Well, changing x86's struct dma_mapping_ops could break IA64. It's just wrong. It's one of problems that this patchset fixes. We will remove map_single and unmap_single hooks in the last patch in this patchset. Signed-off-by: FUJITA Tomonori <fujita.tomonori@lab.ntt.co.jp> Signed-off-by: Ingo Molnar <mingo@elte.hu>
-rw-r--r--drivers/pci/intel-iommu.c24
1 files changed, 22 insertions, 2 deletions
diff --git a/drivers/pci/intel-iommu.c b/drivers/pci/intel-iommu.c
index 235fb7a5a8a5..60258ecbcb4c 100644
--- a/drivers/pci/intel-iommu.c
+++ b/drivers/pci/intel-iommu.c
@@ -2273,6 +2273,15 @@ error:
2273 return 0; 2273 return 0;
2274} 2274}
2275 2275
2276static dma_addr_t intel_map_page(struct device *dev, struct page *page,
2277 unsigned long offset, size_t size,
2278 enum dma_data_direction dir,
2279 struct dma_attrs *attrs)
2280{
2281 return __intel_map_single(dev, page_to_phys(page) + offset, size,
2282 dir, to_pci_dev(dev)->dma_mask);
2283}
2284
2276dma_addr_t intel_map_single(struct device *hwdev, phys_addr_t paddr, 2285dma_addr_t intel_map_single(struct device *hwdev, phys_addr_t paddr,
2277 size_t size, int dir) 2286 size_t size, int dir)
2278{ 2287{
@@ -2341,8 +2350,9 @@ static void add_unmap(struct dmar_domain *dom, struct iova *iova)
2341 spin_unlock_irqrestore(&async_umap_flush_lock, flags); 2350 spin_unlock_irqrestore(&async_umap_flush_lock, flags);
2342} 2351}
2343 2352
2344void intel_unmap_single(struct device *dev, dma_addr_t dev_addr, size_t size, 2353static void intel_unmap_page(struct device *dev, dma_addr_t dev_addr,
2345 int dir) 2354 size_t size, enum dma_data_direction dir,
2355 struct dma_attrs *attrs)
2346{ 2356{
2347 struct pci_dev *pdev = to_pci_dev(dev); 2357 struct pci_dev *pdev = to_pci_dev(dev);
2348 struct dmar_domain *domain; 2358 struct dmar_domain *domain;
@@ -2386,6 +2396,12 @@ void intel_unmap_single(struct device *dev, dma_addr_t dev_addr, size_t size,
2386 } 2396 }
2387} 2397}
2388 2398
2399void intel_unmap_single(struct device *dev, dma_addr_t dev_addr, size_t size,
2400 int dir)
2401{
2402 intel_unmap_page(dev, dev_addr, size, dir, NULL);
2403}
2404
2389void *intel_alloc_coherent(struct device *hwdev, size_t size, 2405void *intel_alloc_coherent(struct device *hwdev, size_t size,
2390 dma_addr_t *dma_handle, gfp_t flags) 2406 dma_addr_t *dma_handle, gfp_t flags)
2391{ 2407{
@@ -2570,6 +2586,10 @@ static struct dma_mapping_ops intel_dma_ops = {
2570 .unmap_single = intel_unmap_single, 2586 .unmap_single = intel_unmap_single,
2571 .map_sg = intel_map_sg, 2587 .map_sg = intel_map_sg,
2572 .unmap_sg = intel_unmap_sg, 2588 .unmap_sg = intel_unmap_sg,
2589#ifdef CONFIG_X86_64
2590 .map_page = intel_map_page,
2591 .unmap_page = intel_unmap_page,
2592#endif
2573}; 2593};
2574 2594
2575static inline int iommu_domain_cache_init(void) 2595static inline int iommu_domain_cache_init(void)