diff options
-rw-r--r-- | arch/x86/Kconfig | 1 | ||||
-rw-r--r-- | arch/x86/include/asm/amd_iommu_proto.h | 12 | ||||
-rw-r--r-- | arch/x86/include/asm/amd_iommu_types.h | 28 | ||||
-rw-r--r-- | arch/x86/kernel/amd_iommu.c | 519 | ||||
-rw-r--r-- | arch/x86/kernel/amd_iommu_init.c | 39 | ||||
-rw-r--r-- | drivers/pci/intel-iommu.c | 1 | ||||
-rw-r--r-- | drivers/pci/iov.c | 1 | ||||
-rw-r--r-- | drivers/pci/pci.h | 37 | ||||
-rw-r--r-- | include/linux/pci-ats.h | 52 | ||||
-rw-r--r-- | lib/dma-debug.c | 18 |
10 files changed, 438 insertions, 270 deletions
diff --git a/arch/x86/Kconfig b/arch/x86/Kconfig index cc6c53a95bfd..8cc29da2d689 100644 --- a/arch/x86/Kconfig +++ b/arch/x86/Kconfig | |||
@@ -690,6 +690,7 @@ config AMD_IOMMU | |||
690 | bool "AMD IOMMU support" | 690 | bool "AMD IOMMU support" |
691 | select SWIOTLB | 691 | select SWIOTLB |
692 | select PCI_MSI | 692 | select PCI_MSI |
693 | select PCI_IOV | ||
693 | depends on X86_64 && PCI && ACPI | 694 | depends on X86_64 && PCI && ACPI |
694 | ---help--- | 695 | ---help--- |
695 | With this option you can enable support for AMD IOMMU hardware in | 696 | With this option you can enable support for AMD IOMMU hardware in |
diff --git a/arch/x86/include/asm/amd_iommu_proto.h b/arch/x86/include/asm/amd_iommu_proto.h index 916bc8111a01..a4ae6c3875eb 100644 --- a/arch/x86/include/asm/amd_iommu_proto.h +++ b/arch/x86/include/asm/amd_iommu_proto.h | |||
@@ -19,13 +19,11 @@ | |||
19 | #ifndef _ASM_X86_AMD_IOMMU_PROTO_H | 19 | #ifndef _ASM_X86_AMD_IOMMU_PROTO_H |
20 | #define _ASM_X86_AMD_IOMMU_PROTO_H | 20 | #define _ASM_X86_AMD_IOMMU_PROTO_H |
21 | 21 | ||
22 | struct amd_iommu; | 22 | #include <asm/amd_iommu_types.h> |
23 | 23 | ||
24 | extern int amd_iommu_init_dma_ops(void); | 24 | extern int amd_iommu_init_dma_ops(void); |
25 | extern int amd_iommu_init_passthrough(void); | 25 | extern int amd_iommu_init_passthrough(void); |
26 | extern irqreturn_t amd_iommu_int_handler(int irq, void *data); | 26 | extern irqreturn_t amd_iommu_int_handler(int irq, void *data); |
27 | extern void amd_iommu_flush_all_domains(void); | ||
28 | extern void amd_iommu_flush_all_devices(void); | ||
29 | extern void amd_iommu_apply_erratum_63(u16 devid); | 27 | extern void amd_iommu_apply_erratum_63(u16 devid); |
30 | extern void amd_iommu_reset_cmd_buffer(struct amd_iommu *iommu); | 28 | extern void amd_iommu_reset_cmd_buffer(struct amd_iommu *iommu); |
31 | extern int amd_iommu_init_devices(void); | 29 | extern int amd_iommu_init_devices(void); |
@@ -44,4 +42,12 @@ static inline bool is_rd890_iommu(struct pci_dev *pdev) | |||
44 | (pdev->device == PCI_DEVICE_ID_RD890_IOMMU); | 42 | (pdev->device == PCI_DEVICE_ID_RD890_IOMMU); |
45 | } | 43 | } |
46 | 44 | ||
45 | static inline bool iommu_feature(struct amd_iommu *iommu, u64 f) | ||
46 | { | ||
47 | if (!(iommu->cap & (1 << IOMMU_CAP_EFR))) | ||
48 | return false; | ||
49 | |||
50 | return !!(iommu->features & f); | ||
51 | } | ||
52 | |||
47 | #endif /* _ASM_X86_AMD_IOMMU_PROTO_H */ | 53 | #endif /* _ASM_X86_AMD_IOMMU_PROTO_H */ |
diff --git a/arch/x86/include/asm/amd_iommu_types.h b/arch/x86/include/asm/amd_iommu_types.h index e3509fc303bf..4c9982995414 100644 --- a/arch/x86/include/asm/amd_iommu_types.h +++ b/arch/x86/include/asm/amd_iommu_types.h | |||
@@ -68,12 +68,25 @@ | |||
68 | #define MMIO_CONTROL_OFFSET 0x0018 | 68 | #define MMIO_CONTROL_OFFSET 0x0018 |
69 | #define MMIO_EXCL_BASE_OFFSET 0x0020 | 69 | #define MMIO_EXCL_BASE_OFFSET 0x0020 |
70 | #define MMIO_EXCL_LIMIT_OFFSET 0x0028 | 70 | #define MMIO_EXCL_LIMIT_OFFSET 0x0028 |
71 | #define MMIO_EXT_FEATURES 0x0030 | ||
71 | #define MMIO_CMD_HEAD_OFFSET 0x2000 | 72 | #define MMIO_CMD_HEAD_OFFSET 0x2000 |
72 | #define MMIO_CMD_TAIL_OFFSET 0x2008 | 73 | #define MMIO_CMD_TAIL_OFFSET 0x2008 |
73 | #define MMIO_EVT_HEAD_OFFSET 0x2010 | 74 | #define MMIO_EVT_HEAD_OFFSET 0x2010 |
74 | #define MMIO_EVT_TAIL_OFFSET 0x2018 | 75 | #define MMIO_EVT_TAIL_OFFSET 0x2018 |
75 | #define MMIO_STATUS_OFFSET 0x2020 | 76 | #define MMIO_STATUS_OFFSET 0x2020 |
76 | 77 | ||
78 | |||
79 | /* Extended Feature Bits */ | ||
80 | #define FEATURE_PREFETCH (1ULL<<0) | ||
81 | #define FEATURE_PPR (1ULL<<1) | ||
82 | #define FEATURE_X2APIC (1ULL<<2) | ||
83 | #define FEATURE_NX (1ULL<<3) | ||
84 | #define FEATURE_GT (1ULL<<4) | ||
85 | #define FEATURE_IA (1ULL<<6) | ||
86 | #define FEATURE_GA (1ULL<<7) | ||
87 | #define FEATURE_HE (1ULL<<8) | ||
88 | #define FEATURE_PC (1ULL<<9) | ||
89 | |||
77 | /* MMIO status bits */ | 90 | /* MMIO status bits */ |
78 | #define MMIO_STATUS_COM_WAIT_INT_MASK 0x04 | 91 | #define MMIO_STATUS_COM_WAIT_INT_MASK 0x04 |
79 | 92 | ||
@@ -113,7 +126,9 @@ | |||
113 | /* command specific defines */ | 126 | /* command specific defines */ |
114 | #define CMD_COMPL_WAIT 0x01 | 127 | #define CMD_COMPL_WAIT 0x01 |
115 | #define CMD_INV_DEV_ENTRY 0x02 | 128 | #define CMD_INV_DEV_ENTRY 0x02 |
116 | #define CMD_INV_IOMMU_PAGES 0x03 | 129 | #define CMD_INV_IOMMU_PAGES 0x03 |
130 | #define CMD_INV_IOTLB_PAGES 0x04 | ||
131 | #define CMD_INV_ALL 0x08 | ||
117 | 132 | ||
118 | #define CMD_COMPL_WAIT_STORE_MASK 0x01 | 133 | #define CMD_COMPL_WAIT_STORE_MASK 0x01 |
119 | #define CMD_COMPL_WAIT_INT_MASK 0x02 | 134 | #define CMD_COMPL_WAIT_INT_MASK 0x02 |
@@ -215,6 +230,8 @@ | |||
215 | #define IOMMU_PTE_IR (1ULL << 61) | 230 | #define IOMMU_PTE_IR (1ULL << 61) |
216 | #define IOMMU_PTE_IW (1ULL << 62) | 231 | #define IOMMU_PTE_IW (1ULL << 62) |
217 | 232 | ||
233 | #define DTE_FLAG_IOTLB 0x01 | ||
234 | |||
218 | #define IOMMU_PAGE_MASK (((1ULL << 52) - 1) & ~0xfffULL) | 235 | #define IOMMU_PAGE_MASK (((1ULL << 52) - 1) & ~0xfffULL) |
219 | #define IOMMU_PTE_PRESENT(pte) ((pte) & IOMMU_PTE_P) | 236 | #define IOMMU_PTE_PRESENT(pte) ((pte) & IOMMU_PTE_P) |
220 | #define IOMMU_PTE_PAGE(pte) (phys_to_virt((pte) & IOMMU_PAGE_MASK)) | 237 | #define IOMMU_PTE_PAGE(pte) (phys_to_virt((pte) & IOMMU_PAGE_MASK)) |
@@ -227,6 +244,7 @@ | |||
227 | /* IOMMU capabilities */ | 244 | /* IOMMU capabilities */ |
228 | #define IOMMU_CAP_IOTLB 24 | 245 | #define IOMMU_CAP_IOTLB 24 |
229 | #define IOMMU_CAP_NPCACHE 26 | 246 | #define IOMMU_CAP_NPCACHE 26 |
247 | #define IOMMU_CAP_EFR 27 | ||
230 | 248 | ||
231 | #define MAX_DOMAIN_ID 65536 | 249 | #define MAX_DOMAIN_ID 65536 |
232 | 250 | ||
@@ -249,6 +267,8 @@ extern bool amd_iommu_dump; | |||
249 | 267 | ||
250 | /* global flag if IOMMUs cache non-present entries */ | 268 | /* global flag if IOMMUs cache non-present entries */ |
251 | extern bool amd_iommu_np_cache; | 269 | extern bool amd_iommu_np_cache; |
270 | /* Only true if all IOMMUs support device IOTLBs */ | ||
271 | extern bool amd_iommu_iotlb_sup; | ||
252 | 272 | ||
253 | /* | 273 | /* |
254 | * Make iterating over all IOMMUs easier | 274 | * Make iterating over all IOMMUs easier |
@@ -371,6 +391,9 @@ struct amd_iommu { | |||
371 | /* flags read from acpi table */ | 391 | /* flags read from acpi table */ |
372 | u8 acpi_flags; | 392 | u8 acpi_flags; |
373 | 393 | ||
394 | /* Extended features */ | ||
395 | u64 features; | ||
396 | |||
374 | /* | 397 | /* |
375 | * Capability pointer. There could be more than one IOMMU per PCI | 398 | * Capability pointer. There could be more than one IOMMU per PCI |
376 | * device function if there are more than one AMD IOMMU capability | 399 | * device function if there are more than one AMD IOMMU capability |
@@ -409,9 +432,6 @@ struct amd_iommu { | |||
409 | /* if one, we need to send a completion wait command */ | 432 | /* if one, we need to send a completion wait command */ |
410 | bool need_sync; | 433 | bool need_sync; |
411 | 434 | ||
412 | /* becomes true if a command buffer reset is running */ | ||
413 | bool reset_in_progress; | ||
414 | |||
415 | /* default dma_ops domain for that IOMMU */ | 435 | /* default dma_ops domain for that IOMMU */ |
416 | struct dma_ops_domain *default_dom; | 436 | struct dma_ops_domain *default_dom; |
417 | 437 | ||
diff --git a/arch/x86/kernel/amd_iommu.c b/arch/x86/kernel/amd_iommu.c index 57ca77787220..dc5dddafe5c2 100644 --- a/arch/x86/kernel/amd_iommu.c +++ b/arch/x86/kernel/amd_iommu.c | |||
@@ -18,6 +18,7 @@ | |||
18 | */ | 18 | */ |
19 | 19 | ||
20 | #include <linux/pci.h> | 20 | #include <linux/pci.h> |
21 | #include <linux/pci-ats.h> | ||
21 | #include <linux/bitmap.h> | 22 | #include <linux/bitmap.h> |
22 | #include <linux/slab.h> | 23 | #include <linux/slab.h> |
23 | #include <linux/debugfs.h> | 24 | #include <linux/debugfs.h> |
@@ -25,6 +26,7 @@ | |||
25 | #include <linux/dma-mapping.h> | 26 | #include <linux/dma-mapping.h> |
26 | #include <linux/iommu-helper.h> | 27 | #include <linux/iommu-helper.h> |
27 | #include <linux/iommu.h> | 28 | #include <linux/iommu.h> |
29 | #include <linux/delay.h> | ||
28 | #include <asm/proto.h> | 30 | #include <asm/proto.h> |
29 | #include <asm/iommu.h> | 31 | #include <asm/iommu.h> |
30 | #include <asm/gart.h> | 32 | #include <asm/gart.h> |
@@ -34,7 +36,7 @@ | |||
34 | 36 | ||
35 | #define CMD_SET_TYPE(cmd, t) ((cmd)->data[1] |= ((t) << 28)) | 37 | #define CMD_SET_TYPE(cmd, t) ((cmd)->data[1] |= ((t) << 28)) |
36 | 38 | ||
37 | #define EXIT_LOOP_COUNT 10000000 | 39 | #define LOOP_TIMEOUT 100000 |
38 | 40 | ||
39 | static DEFINE_RWLOCK(amd_iommu_devtable_lock); | 41 | static DEFINE_RWLOCK(amd_iommu_devtable_lock); |
40 | 42 | ||
@@ -57,7 +59,6 @@ struct iommu_cmd { | |||
57 | u32 data[4]; | 59 | u32 data[4]; |
58 | }; | 60 | }; |
59 | 61 | ||
60 | static void reset_iommu_command_buffer(struct amd_iommu *iommu); | ||
61 | static void update_domain(struct protection_domain *domain); | 62 | static void update_domain(struct protection_domain *domain); |
62 | 63 | ||
63 | /**************************************************************************** | 64 | /**************************************************************************** |
@@ -322,8 +323,6 @@ static void iommu_print_event(struct amd_iommu *iommu, void *__evt) | |||
322 | break; | 323 | break; |
323 | case EVENT_TYPE_ILL_CMD: | 324 | case EVENT_TYPE_ILL_CMD: |
324 | printk("ILLEGAL_COMMAND_ERROR address=0x%016llx]\n", address); | 325 | printk("ILLEGAL_COMMAND_ERROR address=0x%016llx]\n", address); |
325 | iommu->reset_in_progress = true; | ||
326 | reset_iommu_command_buffer(iommu); | ||
327 | dump_command(address); | 326 | dump_command(address); |
328 | break; | 327 | break; |
329 | case EVENT_TYPE_CMD_HARD_ERR: | 328 | case EVENT_TYPE_CMD_HARD_ERR: |
@@ -383,186 +382,289 @@ irqreturn_t amd_iommu_int_handler(int irq, void *data) | |||
383 | * | 382 | * |
384 | ****************************************************************************/ | 383 | ****************************************************************************/ |
385 | 384 | ||
386 | /* | 385 | static int wait_on_sem(volatile u64 *sem) |
387 | * Writes the command to the IOMMUs command buffer and informs the | 386 | { |
388 | * hardware about the new command. Must be called with iommu->lock held. | 387 | int i = 0; |
389 | */ | 388 | |
390 | static int __iommu_queue_command(struct amd_iommu *iommu, struct iommu_cmd *cmd) | 389 | while (*sem == 0 && i < LOOP_TIMEOUT) { |
390 | udelay(1); | ||
391 | i += 1; | ||
392 | } | ||
393 | |||
394 | if (i == LOOP_TIMEOUT) { | ||
395 | pr_alert("AMD-Vi: Completion-Wait loop timed out\n"); | ||
396 | return -EIO; | ||
397 | } | ||
398 | |||
399 | return 0; | ||
400 | } | ||
401 | |||
402 | static void copy_cmd_to_buffer(struct amd_iommu *iommu, | ||
403 | struct iommu_cmd *cmd, | ||
404 | u32 tail) | ||
391 | { | 405 | { |
392 | u32 tail, head; | ||
393 | u8 *target; | 406 | u8 *target; |
394 | 407 | ||
395 | WARN_ON(iommu->cmd_buf_size & CMD_BUFFER_UNINITIALIZED); | ||
396 | tail = readl(iommu->mmio_base + MMIO_CMD_TAIL_OFFSET); | ||
397 | target = iommu->cmd_buf + tail; | 408 | target = iommu->cmd_buf + tail; |
398 | memcpy_toio(target, cmd, sizeof(*cmd)); | 409 | tail = (tail + sizeof(*cmd)) % iommu->cmd_buf_size; |
399 | tail = (tail + sizeof(*cmd)) % iommu->cmd_buf_size; | 410 | |
400 | head = readl(iommu->mmio_base + MMIO_CMD_HEAD_OFFSET); | 411 | /* Copy command to buffer */ |
401 | if (tail == head) | 412 | memcpy(target, cmd, sizeof(*cmd)); |
402 | return -ENOMEM; | 413 | |
414 | /* Tell the IOMMU about it */ | ||
403 | writel(tail, iommu->mmio_base + MMIO_CMD_TAIL_OFFSET); | 415 | writel(tail, iommu->mmio_base + MMIO_CMD_TAIL_OFFSET); |
416 | } | ||
404 | 417 | ||
405 | return 0; | 418 | static void build_completion_wait(struct iommu_cmd *cmd, u64 address) |
419 | { | ||
420 | WARN_ON(address & 0x7ULL); | ||
421 | |||
422 | memset(cmd, 0, sizeof(*cmd)); | ||
423 | cmd->data[0] = lower_32_bits(__pa(address)) | CMD_COMPL_WAIT_STORE_MASK; | ||
424 | cmd->data[1] = upper_32_bits(__pa(address)); | ||
425 | cmd->data[2] = 1; | ||
426 | CMD_SET_TYPE(cmd, CMD_COMPL_WAIT); | ||
427 | } | ||
428 | |||
429 | static void build_inv_dte(struct iommu_cmd *cmd, u16 devid) | ||
430 | { | ||
431 | memset(cmd, 0, sizeof(*cmd)); | ||
432 | cmd->data[0] = devid; | ||
433 | CMD_SET_TYPE(cmd, CMD_INV_DEV_ENTRY); | ||
434 | } | ||
435 | |||
436 | static void build_inv_iommu_pages(struct iommu_cmd *cmd, u64 address, | ||
437 | size_t size, u16 domid, int pde) | ||
438 | { | ||
439 | u64 pages; | ||
440 | int s; | ||
441 | |||
442 | pages = iommu_num_pages(address, size, PAGE_SIZE); | ||
443 | s = 0; | ||
444 | |||
445 | if (pages > 1) { | ||
446 | /* | ||
447 | * If we have to flush more than one page, flush all | ||
448 | * TLB entries for this domain | ||
449 | */ | ||
450 | address = CMD_INV_IOMMU_ALL_PAGES_ADDRESS; | ||
451 | s = 1; | ||
452 | } | ||
453 | |||
454 | address &= PAGE_MASK; | ||
455 | |||
456 | memset(cmd, 0, sizeof(*cmd)); | ||
457 | cmd->data[1] |= domid; | ||
458 | cmd->data[2] = lower_32_bits(address); | ||
459 | cmd->data[3] = upper_32_bits(address); | ||
460 | CMD_SET_TYPE(cmd, CMD_INV_IOMMU_PAGES); | ||
461 | if (s) /* size bit - we flush more than one 4kb page */ | ||
462 | cmd->data[2] |= CMD_INV_IOMMU_PAGES_SIZE_MASK; | ||
463 | if (pde) /* PDE bit - we wan't flush everything not only the PTEs */ | ||
464 | cmd->data[2] |= CMD_INV_IOMMU_PAGES_PDE_MASK; | ||
465 | } | ||
466 | |||
467 | static void build_inv_iotlb_pages(struct iommu_cmd *cmd, u16 devid, int qdep, | ||
468 | u64 address, size_t size) | ||
469 | { | ||
470 | u64 pages; | ||
471 | int s; | ||
472 | |||
473 | pages = iommu_num_pages(address, size, PAGE_SIZE); | ||
474 | s = 0; | ||
475 | |||
476 | if (pages > 1) { | ||
477 | /* | ||
478 | * If we have to flush more than one page, flush all | ||
479 | * TLB entries for this domain | ||
480 | */ | ||
481 | address = CMD_INV_IOMMU_ALL_PAGES_ADDRESS; | ||
482 | s = 1; | ||
483 | } | ||
484 | |||
485 | address &= PAGE_MASK; | ||
486 | |||
487 | memset(cmd, 0, sizeof(*cmd)); | ||
488 | cmd->data[0] = devid; | ||
489 | cmd->data[0] |= (qdep & 0xff) << 24; | ||
490 | cmd->data[1] = devid; | ||
491 | cmd->data[2] = lower_32_bits(address); | ||
492 | cmd->data[3] = upper_32_bits(address); | ||
493 | CMD_SET_TYPE(cmd, CMD_INV_IOTLB_PAGES); | ||
494 | if (s) | ||
495 | cmd->data[2] |= CMD_INV_IOMMU_PAGES_SIZE_MASK; | ||
496 | } | ||
497 | |||
498 | static void build_inv_all(struct iommu_cmd *cmd) | ||
499 | { | ||
500 | memset(cmd, 0, sizeof(*cmd)); | ||
501 | CMD_SET_TYPE(cmd, CMD_INV_ALL); | ||
406 | } | 502 | } |
407 | 503 | ||
408 | /* | 504 | /* |
409 | * General queuing function for commands. Takes iommu->lock and calls | 505 | * Writes the command to the IOMMUs command buffer and informs the |
410 | * __iommu_queue_command(). | 506 | * hardware about the new command. |
411 | */ | 507 | */ |
412 | static int iommu_queue_command(struct amd_iommu *iommu, struct iommu_cmd *cmd) | 508 | static int iommu_queue_command(struct amd_iommu *iommu, struct iommu_cmd *cmd) |
413 | { | 509 | { |
510 | u32 left, tail, head, next_tail; | ||
414 | unsigned long flags; | 511 | unsigned long flags; |
415 | int ret; | ||
416 | 512 | ||
513 | WARN_ON(iommu->cmd_buf_size & CMD_BUFFER_UNINITIALIZED); | ||
514 | |||
515 | again: | ||
417 | spin_lock_irqsave(&iommu->lock, flags); | 516 | spin_lock_irqsave(&iommu->lock, flags); |
418 | ret = __iommu_queue_command(iommu, cmd); | ||
419 | if (!ret) | ||
420 | iommu->need_sync = true; | ||
421 | spin_unlock_irqrestore(&iommu->lock, flags); | ||
422 | 517 | ||
423 | return ret; | 518 | head = readl(iommu->mmio_base + MMIO_CMD_HEAD_OFFSET); |
424 | } | 519 | tail = readl(iommu->mmio_base + MMIO_CMD_TAIL_OFFSET); |
520 | next_tail = (tail + sizeof(*cmd)) % iommu->cmd_buf_size; | ||
521 | left = (head - next_tail) % iommu->cmd_buf_size; | ||
425 | 522 | ||
426 | /* | 523 | if (left <= 2) { |
427 | * This function waits until an IOMMU has completed a completion | 524 | struct iommu_cmd sync_cmd; |
428 | * wait command | 525 | volatile u64 sem = 0; |
429 | */ | 526 | int ret; |
430 | static void __iommu_wait_for_completion(struct amd_iommu *iommu) | ||
431 | { | ||
432 | int ready = 0; | ||
433 | unsigned status = 0; | ||
434 | unsigned long i = 0; | ||
435 | 527 | ||
436 | INC_STATS_COUNTER(compl_wait); | 528 | build_completion_wait(&sync_cmd, (u64)&sem); |
529 | copy_cmd_to_buffer(iommu, &sync_cmd, tail); | ||
437 | 530 | ||
438 | while (!ready && (i < EXIT_LOOP_COUNT)) { | 531 | spin_unlock_irqrestore(&iommu->lock, flags); |
439 | ++i; | 532 | |
440 | /* wait for the bit to become one */ | 533 | if ((ret = wait_on_sem(&sem)) != 0) |
441 | status = readl(iommu->mmio_base + MMIO_STATUS_OFFSET); | 534 | return ret; |
442 | ready = status & MMIO_STATUS_COM_WAIT_INT_MASK; | 535 | |
536 | goto again; | ||
443 | } | 537 | } |
444 | 538 | ||
445 | /* set bit back to zero */ | 539 | copy_cmd_to_buffer(iommu, cmd, tail); |
446 | status &= ~MMIO_STATUS_COM_WAIT_INT_MASK; | 540 | |
447 | writel(status, iommu->mmio_base + MMIO_STATUS_OFFSET); | 541 | /* We need to sync now to make sure all commands are processed */ |
542 | iommu->need_sync = true; | ||
543 | |||
544 | spin_unlock_irqrestore(&iommu->lock, flags); | ||
448 | 545 | ||
449 | if (unlikely(i == EXIT_LOOP_COUNT)) | 546 | return 0; |
450 | iommu->reset_in_progress = true; | ||
451 | } | 547 | } |
452 | 548 | ||
453 | /* | 549 | /* |
454 | * This function queues a completion wait command into the command | 550 | * This function queues a completion wait command into the command |
455 | * buffer of an IOMMU | 551 | * buffer of an IOMMU |
456 | */ | 552 | */ |
457 | static int __iommu_completion_wait(struct amd_iommu *iommu) | 553 | static int iommu_completion_wait(struct amd_iommu *iommu) |
458 | { | 554 | { |
459 | struct iommu_cmd cmd; | 555 | struct iommu_cmd cmd; |
556 | volatile u64 sem = 0; | ||
557 | int ret; | ||
460 | 558 | ||
461 | memset(&cmd, 0, sizeof(cmd)); | 559 | if (!iommu->need_sync) |
462 | cmd.data[0] = CMD_COMPL_WAIT_INT_MASK; | 560 | return 0; |
463 | CMD_SET_TYPE(&cmd, CMD_COMPL_WAIT); | 561 | |
562 | build_completion_wait(&cmd, (u64)&sem); | ||
563 | |||
564 | ret = iommu_queue_command(iommu, &cmd); | ||
565 | if (ret) | ||
566 | return ret; | ||
464 | 567 | ||
465 | return __iommu_queue_command(iommu, &cmd); | 568 | return wait_on_sem(&sem); |
466 | } | 569 | } |
467 | 570 | ||
468 | /* | 571 | static int iommu_flush_dte(struct amd_iommu *iommu, u16 devid) |
469 | * This function is called whenever we need to ensure that the IOMMU has | ||
470 | * completed execution of all commands we sent. It sends a | ||
471 | * COMPLETION_WAIT command and waits for it to finish. The IOMMU informs | ||
472 | * us about that by writing a value to a physical address we pass with | ||
473 | * the command. | ||
474 | */ | ||
475 | static int iommu_completion_wait(struct amd_iommu *iommu) | ||
476 | { | 572 | { |
477 | int ret = 0; | 573 | struct iommu_cmd cmd; |
478 | unsigned long flags; | ||
479 | 574 | ||
480 | spin_lock_irqsave(&iommu->lock, flags); | 575 | build_inv_dte(&cmd, devid); |
481 | 576 | ||
482 | if (!iommu->need_sync) | 577 | return iommu_queue_command(iommu, &cmd); |
483 | goto out; | 578 | } |
484 | 579 | ||
485 | ret = __iommu_completion_wait(iommu); | 580 | static void iommu_flush_dte_all(struct amd_iommu *iommu) |
581 | { | ||
582 | u32 devid; | ||
486 | 583 | ||
487 | iommu->need_sync = false; | 584 | for (devid = 0; devid <= 0xffff; ++devid) |
585 | iommu_flush_dte(iommu, devid); | ||
488 | 586 | ||
489 | if (ret) | 587 | iommu_completion_wait(iommu); |
490 | goto out; | 588 | } |
491 | |||
492 | __iommu_wait_for_completion(iommu); | ||
493 | 589 | ||
494 | out: | 590 | /* |
495 | spin_unlock_irqrestore(&iommu->lock, flags); | 591 | * This function uses heavy locking and may disable irqs for some time. But |
592 | * this is no issue because it is only called during resume. | ||
593 | */ | ||
594 | static void iommu_flush_tlb_all(struct amd_iommu *iommu) | ||
595 | { | ||
596 | u32 dom_id; | ||
496 | 597 | ||
497 | if (iommu->reset_in_progress) | 598 | for (dom_id = 0; dom_id <= 0xffff; ++dom_id) { |
498 | reset_iommu_command_buffer(iommu); | 599 | struct iommu_cmd cmd; |
600 | build_inv_iommu_pages(&cmd, 0, CMD_INV_IOMMU_ALL_PAGES_ADDRESS, | ||
601 | dom_id, 1); | ||
602 | iommu_queue_command(iommu, &cmd); | ||
603 | } | ||
499 | 604 | ||
500 | return 0; | 605 | iommu_completion_wait(iommu); |
501 | } | 606 | } |
502 | 607 | ||
503 | static void iommu_flush_complete(struct protection_domain *domain) | 608 | static void iommu_flush_all(struct amd_iommu *iommu) |
504 | { | 609 | { |
505 | int i; | 610 | struct iommu_cmd cmd; |
506 | 611 | ||
507 | for (i = 0; i < amd_iommus_present; ++i) { | 612 | build_inv_all(&cmd); |
508 | if (!domain->dev_iommu[i]) | ||
509 | continue; | ||
510 | 613 | ||
511 | /* | 614 | iommu_queue_command(iommu, &cmd); |
512 | * Devices of this domain are behind this IOMMU | 615 | iommu_completion_wait(iommu); |
513 | * We need to wait for completion of all commands. | 616 | } |
514 | */ | 617 | |
515 | iommu_completion_wait(amd_iommus[i]); | 618 | void iommu_flush_all_caches(struct amd_iommu *iommu) |
619 | { | ||
620 | if (iommu_feature(iommu, FEATURE_IA)) { | ||
621 | iommu_flush_all(iommu); | ||
622 | } else { | ||
623 | iommu_flush_dte_all(iommu); | ||
624 | iommu_flush_tlb_all(iommu); | ||
516 | } | 625 | } |
517 | } | 626 | } |
518 | 627 | ||
519 | /* | 628 | /* |
520 | * Command send function for invalidating a device table entry | 629 | * Command send function for flushing on-device TLB |
521 | */ | 630 | */ |
522 | static int iommu_flush_device(struct device *dev) | 631 | static int device_flush_iotlb(struct device *dev, u64 address, size_t size) |
523 | { | 632 | { |
633 | struct pci_dev *pdev = to_pci_dev(dev); | ||
524 | struct amd_iommu *iommu; | 634 | struct amd_iommu *iommu; |
525 | struct iommu_cmd cmd; | 635 | struct iommu_cmd cmd; |
526 | u16 devid; | 636 | u16 devid; |
637 | int qdep; | ||
527 | 638 | ||
639 | qdep = pci_ats_queue_depth(pdev); | ||
528 | devid = get_device_id(dev); | 640 | devid = get_device_id(dev); |
529 | iommu = amd_iommu_rlookup_table[devid]; | 641 | iommu = amd_iommu_rlookup_table[devid]; |
530 | 642 | ||
531 | /* Build command */ | 643 | build_inv_iotlb_pages(&cmd, devid, qdep, address, size); |
532 | memset(&cmd, 0, sizeof(cmd)); | ||
533 | CMD_SET_TYPE(&cmd, CMD_INV_DEV_ENTRY); | ||
534 | cmd.data[0] = devid; | ||
535 | 644 | ||
536 | return iommu_queue_command(iommu, &cmd); | 645 | return iommu_queue_command(iommu, &cmd); |
537 | } | 646 | } |
538 | 647 | ||
539 | static void __iommu_build_inv_iommu_pages(struct iommu_cmd *cmd, u64 address, | ||
540 | u16 domid, int pde, int s) | ||
541 | { | ||
542 | memset(cmd, 0, sizeof(*cmd)); | ||
543 | address &= PAGE_MASK; | ||
544 | CMD_SET_TYPE(cmd, CMD_INV_IOMMU_PAGES); | ||
545 | cmd->data[1] |= domid; | ||
546 | cmd->data[2] = lower_32_bits(address); | ||
547 | cmd->data[3] = upper_32_bits(address); | ||
548 | if (s) /* size bit - we flush more than one 4kb page */ | ||
549 | cmd->data[2] |= CMD_INV_IOMMU_PAGES_SIZE_MASK; | ||
550 | if (pde) /* PDE bit - we wan't flush everything not only the PTEs */ | ||
551 | cmd->data[2] |= CMD_INV_IOMMU_PAGES_PDE_MASK; | ||
552 | } | ||
553 | |||
554 | /* | 648 | /* |
555 | * Generic command send function for invalidaing TLB entries | 649 | * Command send function for invalidating a device table entry |
556 | */ | 650 | */ |
557 | static int iommu_queue_inv_iommu_pages(struct amd_iommu *iommu, | 651 | static int device_flush_dte(struct device *dev) |
558 | u64 address, u16 domid, int pde, int s) | ||
559 | { | 652 | { |
560 | struct iommu_cmd cmd; | 653 | struct amd_iommu *iommu; |
654 | struct pci_dev *pdev; | ||
655 | u16 devid; | ||
561 | int ret; | 656 | int ret; |
562 | 657 | ||
563 | __iommu_build_inv_iommu_pages(&cmd, address, domid, pde, s); | 658 | pdev = to_pci_dev(dev); |
659 | devid = get_device_id(dev); | ||
660 | iommu = amd_iommu_rlookup_table[devid]; | ||
564 | 661 | ||
565 | ret = iommu_queue_command(iommu, &cmd); | 662 | ret = iommu_flush_dte(iommu, devid); |
663 | if (ret) | ||
664 | return ret; | ||
665 | |||
666 | if (pci_ats_enabled(pdev)) | ||
667 | ret = device_flush_iotlb(dev, 0, ~0UL); | ||
566 | 668 | ||
567 | return ret; | 669 | return ret; |
568 | } | 670 | } |
@@ -572,23 +674,14 @@ static int iommu_queue_inv_iommu_pages(struct amd_iommu *iommu, | |||
572 | * It invalidates a single PTE if the range to flush is within a single | 674 | * It invalidates a single PTE if the range to flush is within a single |
573 | * page. Otherwise it flushes the whole TLB of the IOMMU. | 675 | * page. Otherwise it flushes the whole TLB of the IOMMU. |
574 | */ | 676 | */ |
575 | static void __iommu_flush_pages(struct protection_domain *domain, | 677 | static void __domain_flush_pages(struct protection_domain *domain, |
576 | u64 address, size_t size, int pde) | 678 | u64 address, size_t size, int pde) |
577 | { | 679 | { |
578 | int s = 0, i; | 680 | struct iommu_dev_data *dev_data; |
579 | unsigned long pages = iommu_num_pages(address, size, PAGE_SIZE); | 681 | struct iommu_cmd cmd; |
580 | 682 | int ret = 0, i; | |
581 | address &= PAGE_MASK; | ||
582 | |||
583 | if (pages > 1) { | ||
584 | /* | ||
585 | * If we have to flush more than one page, flush all | ||
586 | * TLB entries for this domain | ||
587 | */ | ||
588 | address = CMD_INV_IOMMU_ALL_PAGES_ADDRESS; | ||
589 | s = 1; | ||
590 | } | ||
591 | 683 | ||
684 | build_inv_iommu_pages(&cmd, address, size, domain->id, pde); | ||
592 | 685 | ||
593 | for (i = 0; i < amd_iommus_present; ++i) { | 686 | for (i = 0; i < amd_iommus_present; ++i) { |
594 | if (!domain->dev_iommu[i]) | 687 | if (!domain->dev_iommu[i]) |
@@ -598,101 +691,70 @@ static void __iommu_flush_pages(struct protection_domain *domain, | |||
598 | * Devices of this domain are behind this IOMMU | 691 | * Devices of this domain are behind this IOMMU |
599 | * We need a TLB flush | 692 | * We need a TLB flush |
600 | */ | 693 | */ |
601 | iommu_queue_inv_iommu_pages(amd_iommus[i], address, | 694 | ret |= iommu_queue_command(amd_iommus[i], &cmd); |
602 | domain->id, pde, s); | ||
603 | } | 695 | } |
604 | 696 | ||
605 | return; | 697 | list_for_each_entry(dev_data, &domain->dev_list, list) { |
698 | struct pci_dev *pdev = to_pci_dev(dev_data->dev); | ||
699 | |||
700 | if (!pci_ats_enabled(pdev)) | ||
701 | continue; | ||
702 | |||
703 | ret |= device_flush_iotlb(dev_data->dev, address, size); | ||
704 | } | ||
705 | |||
706 | WARN_ON(ret); | ||
606 | } | 707 | } |
607 | 708 | ||
608 | static void iommu_flush_pages(struct protection_domain *domain, | 709 | static void domain_flush_pages(struct protection_domain *domain, |
609 | u64 address, size_t size) | 710 | u64 address, size_t size) |
610 | { | 711 | { |
611 | __iommu_flush_pages(domain, address, size, 0); | 712 | __domain_flush_pages(domain, address, size, 0); |
612 | } | 713 | } |
613 | 714 | ||
614 | /* Flush the whole IO/TLB for a given protection domain */ | 715 | /* Flush the whole IO/TLB for a given protection domain */ |
615 | static void iommu_flush_tlb(struct protection_domain *domain) | 716 | static void domain_flush_tlb(struct protection_domain *domain) |
616 | { | 717 | { |
617 | __iommu_flush_pages(domain, 0, CMD_INV_IOMMU_ALL_PAGES_ADDRESS, 0); | 718 | __domain_flush_pages(domain, 0, CMD_INV_IOMMU_ALL_PAGES_ADDRESS, 0); |
618 | } | 719 | } |
619 | 720 | ||
620 | /* Flush the whole IO/TLB for a given protection domain - including PDE */ | 721 | /* Flush the whole IO/TLB for a given protection domain - including PDE */ |
621 | static void iommu_flush_tlb_pde(struct protection_domain *domain) | 722 | static void domain_flush_tlb_pde(struct protection_domain *domain) |
622 | { | 723 | { |
623 | __iommu_flush_pages(domain, 0, CMD_INV_IOMMU_ALL_PAGES_ADDRESS, 1); | 724 | __domain_flush_pages(domain, 0, CMD_INV_IOMMU_ALL_PAGES_ADDRESS, 1); |
624 | } | 725 | } |
625 | 726 | ||
626 | 727 | static void domain_flush_complete(struct protection_domain *domain) | |
627 | /* | ||
628 | * This function flushes the DTEs for all devices in domain | ||
629 | */ | ||
630 | static void iommu_flush_domain_devices(struct protection_domain *domain) | ||
631 | { | 728 | { |
632 | struct iommu_dev_data *dev_data; | 729 | int i; |
633 | unsigned long flags; | ||
634 | |||
635 | spin_lock_irqsave(&domain->lock, flags); | ||
636 | |||
637 | list_for_each_entry(dev_data, &domain->dev_list, list) | ||
638 | iommu_flush_device(dev_data->dev); | ||
639 | |||
640 | spin_unlock_irqrestore(&domain->lock, flags); | ||
641 | } | ||
642 | |||
643 | static void iommu_flush_all_domain_devices(void) | ||
644 | { | ||
645 | struct protection_domain *domain; | ||
646 | unsigned long flags; | ||
647 | 730 | ||
648 | spin_lock_irqsave(&amd_iommu_pd_lock, flags); | 731 | for (i = 0; i < amd_iommus_present; ++i) { |
732 | if (!domain->dev_iommu[i]) | ||
733 | continue; | ||
649 | 734 | ||
650 | list_for_each_entry(domain, &amd_iommu_pd_list, list) { | 735 | /* |
651 | iommu_flush_domain_devices(domain); | 736 | * Devices of this domain are behind this IOMMU |
652 | iommu_flush_complete(domain); | 737 | * We need to wait for completion of all commands. |
738 | */ | ||
739 | iommu_completion_wait(amd_iommus[i]); | ||
653 | } | 740 | } |
654 | |||
655 | spin_unlock_irqrestore(&amd_iommu_pd_lock, flags); | ||
656 | } | 741 | } |
657 | 742 | ||
658 | void amd_iommu_flush_all_devices(void) | ||
659 | { | ||
660 | iommu_flush_all_domain_devices(); | ||
661 | } | ||
662 | 743 | ||
663 | /* | 744 | /* |
664 | * This function uses heavy locking and may disable irqs for some time. But | 745 | * This function flushes the DTEs for all devices in domain |
665 | * this is no issue because it is only called during resume. | ||
666 | */ | 746 | */ |
667 | void amd_iommu_flush_all_domains(void) | 747 | static void domain_flush_devices(struct protection_domain *domain) |
668 | { | 748 | { |
669 | struct protection_domain *domain; | 749 | struct iommu_dev_data *dev_data; |
670 | unsigned long flags; | 750 | unsigned long flags; |
671 | 751 | ||
672 | spin_lock_irqsave(&amd_iommu_pd_lock, flags); | 752 | spin_lock_irqsave(&domain->lock, flags); |
673 | |||
674 | list_for_each_entry(domain, &amd_iommu_pd_list, list) { | ||
675 | spin_lock(&domain->lock); | ||
676 | iommu_flush_tlb_pde(domain); | ||
677 | iommu_flush_complete(domain); | ||
678 | spin_unlock(&domain->lock); | ||
679 | } | ||
680 | |||
681 | spin_unlock_irqrestore(&amd_iommu_pd_lock, flags); | ||
682 | } | ||
683 | |||
684 | static void reset_iommu_command_buffer(struct amd_iommu *iommu) | ||
685 | { | ||
686 | pr_err("AMD-Vi: Resetting IOMMU command buffer\n"); | ||
687 | |||
688 | if (iommu->reset_in_progress) | ||
689 | panic("AMD-Vi: ILLEGAL_COMMAND_ERROR while resetting command buffer\n"); | ||
690 | 753 | ||
691 | amd_iommu_reset_cmd_buffer(iommu); | 754 | list_for_each_entry(dev_data, &domain->dev_list, list) |
692 | amd_iommu_flush_all_devices(); | 755 | device_flush_dte(dev_data->dev); |
693 | amd_iommu_flush_all_domains(); | ||
694 | 756 | ||
695 | iommu->reset_in_progress = false; | 757 | spin_unlock_irqrestore(&domain->lock, flags); |
696 | } | 758 | } |
697 | 759 | ||
698 | /**************************************************************************** | 760 | /**************************************************************************** |
@@ -1410,17 +1472,22 @@ static bool dma_ops_domain(struct protection_domain *domain) | |||
1410 | return domain->flags & PD_DMA_OPS_MASK; | 1472 | return domain->flags & PD_DMA_OPS_MASK; |
1411 | } | 1473 | } |
1412 | 1474 | ||
1413 | static void set_dte_entry(u16 devid, struct protection_domain *domain) | 1475 | static void set_dte_entry(u16 devid, struct protection_domain *domain, bool ats) |
1414 | { | 1476 | { |
1415 | u64 pte_root = virt_to_phys(domain->pt_root); | 1477 | u64 pte_root = virt_to_phys(domain->pt_root); |
1478 | u32 flags = 0; | ||
1416 | 1479 | ||
1417 | pte_root |= (domain->mode & DEV_ENTRY_MODE_MASK) | 1480 | pte_root |= (domain->mode & DEV_ENTRY_MODE_MASK) |
1418 | << DEV_ENTRY_MODE_SHIFT; | 1481 | << DEV_ENTRY_MODE_SHIFT; |
1419 | pte_root |= IOMMU_PTE_IR | IOMMU_PTE_IW | IOMMU_PTE_P | IOMMU_PTE_TV; | 1482 | pte_root |= IOMMU_PTE_IR | IOMMU_PTE_IW | IOMMU_PTE_P | IOMMU_PTE_TV; |
1420 | 1483 | ||
1421 | amd_iommu_dev_table[devid].data[2] = domain->id; | 1484 | if (ats) |
1422 | amd_iommu_dev_table[devid].data[1] = upper_32_bits(pte_root); | 1485 | flags |= DTE_FLAG_IOTLB; |
1423 | amd_iommu_dev_table[devid].data[0] = lower_32_bits(pte_root); | 1486 | |
1487 | amd_iommu_dev_table[devid].data[3] |= flags; | ||
1488 | amd_iommu_dev_table[devid].data[2] = domain->id; | ||
1489 | amd_iommu_dev_table[devid].data[1] = upper_32_bits(pte_root); | ||
1490 | amd_iommu_dev_table[devid].data[0] = lower_32_bits(pte_root); | ||
1424 | } | 1491 | } |
1425 | 1492 | ||
1426 | static void clear_dte_entry(u16 devid) | 1493 | static void clear_dte_entry(u16 devid) |
@@ -1437,34 +1504,42 @@ static void do_attach(struct device *dev, struct protection_domain *domain) | |||
1437 | { | 1504 | { |
1438 | struct iommu_dev_data *dev_data; | 1505 | struct iommu_dev_data *dev_data; |
1439 | struct amd_iommu *iommu; | 1506 | struct amd_iommu *iommu; |
1507 | struct pci_dev *pdev; | ||
1508 | bool ats = false; | ||
1440 | u16 devid; | 1509 | u16 devid; |
1441 | 1510 | ||
1442 | devid = get_device_id(dev); | 1511 | devid = get_device_id(dev); |
1443 | iommu = amd_iommu_rlookup_table[devid]; | 1512 | iommu = amd_iommu_rlookup_table[devid]; |
1444 | dev_data = get_dev_data(dev); | 1513 | dev_data = get_dev_data(dev); |
1514 | pdev = to_pci_dev(dev); | ||
1515 | |||
1516 | if (amd_iommu_iotlb_sup) | ||
1517 | ats = pci_ats_enabled(pdev); | ||
1445 | 1518 | ||
1446 | /* Update data structures */ | 1519 | /* Update data structures */ |
1447 | dev_data->domain = domain; | 1520 | dev_data->domain = domain; |
1448 | list_add(&dev_data->list, &domain->dev_list); | 1521 | list_add(&dev_data->list, &domain->dev_list); |
1449 | set_dte_entry(devid, domain); | 1522 | set_dte_entry(devid, domain, ats); |
1450 | 1523 | ||
1451 | /* Do reference counting */ | 1524 | /* Do reference counting */ |
1452 | domain->dev_iommu[iommu->index] += 1; | 1525 | domain->dev_iommu[iommu->index] += 1; |
1453 | domain->dev_cnt += 1; | 1526 | domain->dev_cnt += 1; |
1454 | 1527 | ||
1455 | /* Flush the DTE entry */ | 1528 | /* Flush the DTE entry */ |
1456 | iommu_flush_device(dev); | 1529 | device_flush_dte(dev); |
1457 | } | 1530 | } |
1458 | 1531 | ||
1459 | static void do_detach(struct device *dev) | 1532 | static void do_detach(struct device *dev) |
1460 | { | 1533 | { |
1461 | struct iommu_dev_data *dev_data; | 1534 | struct iommu_dev_data *dev_data; |
1462 | struct amd_iommu *iommu; | 1535 | struct amd_iommu *iommu; |
1536 | struct pci_dev *pdev; | ||
1463 | u16 devid; | 1537 | u16 devid; |
1464 | 1538 | ||
1465 | devid = get_device_id(dev); | 1539 | devid = get_device_id(dev); |
1466 | iommu = amd_iommu_rlookup_table[devid]; | 1540 | iommu = amd_iommu_rlookup_table[devid]; |
1467 | dev_data = get_dev_data(dev); | 1541 | dev_data = get_dev_data(dev); |
1542 | pdev = to_pci_dev(dev); | ||
1468 | 1543 | ||
1469 | /* decrease reference counters */ | 1544 | /* decrease reference counters */ |
1470 | dev_data->domain->dev_iommu[iommu->index] -= 1; | 1545 | dev_data->domain->dev_iommu[iommu->index] -= 1; |
@@ -1476,7 +1551,7 @@ static void do_detach(struct device *dev) | |||
1476 | clear_dte_entry(devid); | 1551 | clear_dte_entry(devid); |
1477 | 1552 | ||
1478 | /* Flush the DTE entry */ | 1553 | /* Flush the DTE entry */ |
1479 | iommu_flush_device(dev); | 1554 | device_flush_dte(dev); |
1480 | } | 1555 | } |
1481 | 1556 | ||
1482 | /* | 1557 | /* |
@@ -1539,9 +1614,13 @@ out_unlock: | |||
1539 | static int attach_device(struct device *dev, | 1614 | static int attach_device(struct device *dev, |
1540 | struct protection_domain *domain) | 1615 | struct protection_domain *domain) |
1541 | { | 1616 | { |
1617 | struct pci_dev *pdev = to_pci_dev(dev); | ||
1542 | unsigned long flags; | 1618 | unsigned long flags; |
1543 | int ret; | 1619 | int ret; |
1544 | 1620 | ||
1621 | if (amd_iommu_iotlb_sup) | ||
1622 | pci_enable_ats(pdev, PAGE_SHIFT); | ||
1623 | |||
1545 | write_lock_irqsave(&amd_iommu_devtable_lock, flags); | 1624 | write_lock_irqsave(&amd_iommu_devtable_lock, flags); |
1546 | ret = __attach_device(dev, domain); | 1625 | ret = __attach_device(dev, domain); |
1547 | write_unlock_irqrestore(&amd_iommu_devtable_lock, flags); | 1626 | write_unlock_irqrestore(&amd_iommu_devtable_lock, flags); |
@@ -1551,7 +1630,7 @@ static int attach_device(struct device *dev, | |||
1551 | * left the caches in the IOMMU dirty. So we have to flush | 1630 | * left the caches in the IOMMU dirty. So we have to flush |
1552 | * here to evict all dirty stuff. | 1631 | * here to evict all dirty stuff. |
1553 | */ | 1632 | */ |
1554 | iommu_flush_tlb_pde(domain); | 1633 | domain_flush_tlb_pde(domain); |
1555 | 1634 | ||
1556 | return ret; | 1635 | return ret; |
1557 | } | 1636 | } |
@@ -1598,12 +1677,16 @@ static void __detach_device(struct device *dev) | |||
1598 | */ | 1677 | */ |
1599 | static void detach_device(struct device *dev) | 1678 | static void detach_device(struct device *dev) |
1600 | { | 1679 | { |
1680 | struct pci_dev *pdev = to_pci_dev(dev); | ||
1601 | unsigned long flags; | 1681 | unsigned long flags; |
1602 | 1682 | ||
1603 | /* lock device table */ | 1683 | /* lock device table */ |
1604 | write_lock_irqsave(&amd_iommu_devtable_lock, flags); | 1684 | write_lock_irqsave(&amd_iommu_devtable_lock, flags); |
1605 | __detach_device(dev); | 1685 | __detach_device(dev); |
1606 | write_unlock_irqrestore(&amd_iommu_devtable_lock, flags); | 1686 | write_unlock_irqrestore(&amd_iommu_devtable_lock, flags); |
1687 | |||
1688 | if (amd_iommu_iotlb_sup && pci_ats_enabled(pdev)) | ||
1689 | pci_disable_ats(pdev); | ||
1607 | } | 1690 | } |
1608 | 1691 | ||
1609 | /* | 1692 | /* |
@@ -1692,7 +1775,7 @@ static int device_change_notifier(struct notifier_block *nb, | |||
1692 | goto out; | 1775 | goto out; |
1693 | } | 1776 | } |
1694 | 1777 | ||
1695 | iommu_flush_device(dev); | 1778 | device_flush_dte(dev); |
1696 | iommu_completion_wait(iommu); | 1779 | iommu_completion_wait(iommu); |
1697 | 1780 | ||
1698 | out: | 1781 | out: |
@@ -1753,8 +1836,9 @@ static void update_device_table(struct protection_domain *domain) | |||
1753 | struct iommu_dev_data *dev_data; | 1836 | struct iommu_dev_data *dev_data; |
1754 | 1837 | ||
1755 | list_for_each_entry(dev_data, &domain->dev_list, list) { | 1838 | list_for_each_entry(dev_data, &domain->dev_list, list) { |
1839 | struct pci_dev *pdev = to_pci_dev(dev_data->dev); | ||
1756 | u16 devid = get_device_id(dev_data->dev); | 1840 | u16 devid = get_device_id(dev_data->dev); |
1757 | set_dte_entry(devid, domain); | 1841 | set_dte_entry(devid, domain, pci_ats_enabled(pdev)); |
1758 | } | 1842 | } |
1759 | } | 1843 | } |
1760 | 1844 | ||
@@ -1764,8 +1848,9 @@ static void update_domain(struct protection_domain *domain) | |||
1764 | return; | 1848 | return; |
1765 | 1849 | ||
1766 | update_device_table(domain); | 1850 | update_device_table(domain); |
1767 | iommu_flush_domain_devices(domain); | 1851 | |
1768 | iommu_flush_tlb_pde(domain); | 1852 | domain_flush_devices(domain); |
1853 | domain_flush_tlb_pde(domain); | ||
1769 | 1854 | ||
1770 | domain->updated = false; | 1855 | domain->updated = false; |
1771 | } | 1856 | } |
@@ -1924,10 +2009,10 @@ retry: | |||
1924 | ADD_STATS_COUNTER(alloced_io_mem, size); | 2009 | ADD_STATS_COUNTER(alloced_io_mem, size); |
1925 | 2010 | ||
1926 | if (unlikely(dma_dom->need_flush && !amd_iommu_unmap_flush)) { | 2011 | if (unlikely(dma_dom->need_flush && !amd_iommu_unmap_flush)) { |
1927 | iommu_flush_tlb(&dma_dom->domain); | 2012 | domain_flush_tlb(&dma_dom->domain); |
1928 | dma_dom->need_flush = false; | 2013 | dma_dom->need_flush = false; |
1929 | } else if (unlikely(amd_iommu_np_cache)) | 2014 | } else if (unlikely(amd_iommu_np_cache)) |
1930 | iommu_flush_pages(&dma_dom->domain, address, size); | 2015 | domain_flush_pages(&dma_dom->domain, address, size); |
1931 | 2016 | ||
1932 | out: | 2017 | out: |
1933 | return address; | 2018 | return address; |
@@ -1976,7 +2061,7 @@ static void __unmap_single(struct dma_ops_domain *dma_dom, | |||
1976 | dma_ops_free_addresses(dma_dom, dma_addr, pages); | 2061 | dma_ops_free_addresses(dma_dom, dma_addr, pages); |
1977 | 2062 | ||
1978 | if (amd_iommu_unmap_flush || dma_dom->need_flush) { | 2063 | if (amd_iommu_unmap_flush || dma_dom->need_flush) { |
1979 | iommu_flush_pages(&dma_dom->domain, flush_addr, size); | 2064 | domain_flush_pages(&dma_dom->domain, flush_addr, size); |
1980 | dma_dom->need_flush = false; | 2065 | dma_dom->need_flush = false; |
1981 | } | 2066 | } |
1982 | } | 2067 | } |
@@ -2012,7 +2097,7 @@ static dma_addr_t map_page(struct device *dev, struct page *page, | |||
2012 | if (addr == DMA_ERROR_CODE) | 2097 | if (addr == DMA_ERROR_CODE) |
2013 | goto out; | 2098 | goto out; |
2014 | 2099 | ||
2015 | iommu_flush_complete(domain); | 2100 | domain_flush_complete(domain); |
2016 | 2101 | ||
2017 | out: | 2102 | out: |
2018 | spin_unlock_irqrestore(&domain->lock, flags); | 2103 | spin_unlock_irqrestore(&domain->lock, flags); |
@@ -2039,7 +2124,7 @@ static void unmap_page(struct device *dev, dma_addr_t dma_addr, size_t size, | |||
2039 | 2124 | ||
2040 | __unmap_single(domain->priv, dma_addr, size, dir); | 2125 | __unmap_single(domain->priv, dma_addr, size, dir); |
2041 | 2126 | ||
2042 | iommu_flush_complete(domain); | 2127 | domain_flush_complete(domain); |
2043 | 2128 | ||
2044 | spin_unlock_irqrestore(&domain->lock, flags); | 2129 | spin_unlock_irqrestore(&domain->lock, flags); |
2045 | } | 2130 | } |
@@ -2104,7 +2189,7 @@ static int map_sg(struct device *dev, struct scatterlist *sglist, | |||
2104 | goto unmap; | 2189 | goto unmap; |
2105 | } | 2190 | } |
2106 | 2191 | ||
2107 | iommu_flush_complete(domain); | 2192 | domain_flush_complete(domain); |
2108 | 2193 | ||
2109 | out: | 2194 | out: |
2110 | spin_unlock_irqrestore(&domain->lock, flags); | 2195 | spin_unlock_irqrestore(&domain->lock, flags); |
@@ -2150,7 +2235,7 @@ static void unmap_sg(struct device *dev, struct scatterlist *sglist, | |||
2150 | s->dma_address = s->dma_length = 0; | 2235 | s->dma_address = s->dma_length = 0; |
2151 | } | 2236 | } |
2152 | 2237 | ||
2153 | iommu_flush_complete(domain); | 2238 | domain_flush_complete(domain); |
2154 | 2239 | ||
2155 | spin_unlock_irqrestore(&domain->lock, flags); | 2240 | spin_unlock_irqrestore(&domain->lock, flags); |
2156 | } | 2241 | } |
@@ -2200,7 +2285,7 @@ static void *alloc_coherent(struct device *dev, size_t size, | |||
2200 | goto out_free; | 2285 | goto out_free; |
2201 | } | 2286 | } |
2202 | 2287 | ||
2203 | iommu_flush_complete(domain); | 2288 | domain_flush_complete(domain); |
2204 | 2289 | ||
2205 | spin_unlock_irqrestore(&domain->lock, flags); | 2290 | spin_unlock_irqrestore(&domain->lock, flags); |
2206 | 2291 | ||
@@ -2232,7 +2317,7 @@ static void free_coherent(struct device *dev, size_t size, | |||
2232 | 2317 | ||
2233 | __unmap_single(domain->priv, dma_addr, size, DMA_BIDIRECTIONAL); | 2318 | __unmap_single(domain->priv, dma_addr, size, DMA_BIDIRECTIONAL); |
2234 | 2319 | ||
2235 | iommu_flush_complete(domain); | 2320 | domain_flush_complete(domain); |
2236 | 2321 | ||
2237 | spin_unlock_irqrestore(&domain->lock, flags); | 2322 | spin_unlock_irqrestore(&domain->lock, flags); |
2238 | 2323 | ||
@@ -2476,7 +2561,7 @@ static void amd_iommu_detach_device(struct iommu_domain *dom, | |||
2476 | if (!iommu) | 2561 | if (!iommu) |
2477 | return; | 2562 | return; |
2478 | 2563 | ||
2479 | iommu_flush_device(dev); | 2564 | device_flush_dte(dev); |
2480 | iommu_completion_wait(iommu); | 2565 | iommu_completion_wait(iommu); |
2481 | } | 2566 | } |
2482 | 2567 | ||
@@ -2542,7 +2627,7 @@ static int amd_iommu_unmap(struct iommu_domain *dom, unsigned long iova, | |||
2542 | unmap_size = iommu_unmap_page(domain, iova, page_size); | 2627 | unmap_size = iommu_unmap_page(domain, iova, page_size); |
2543 | mutex_unlock(&domain->api_lock); | 2628 | mutex_unlock(&domain->api_lock); |
2544 | 2629 | ||
2545 | iommu_flush_tlb_pde(domain); | 2630 | domain_flush_tlb_pde(domain); |
2546 | 2631 | ||
2547 | return get_order(unmap_size); | 2632 | return get_order(unmap_size); |
2548 | } | 2633 | } |
diff --git a/arch/x86/kernel/amd_iommu_init.c b/arch/x86/kernel/amd_iommu_init.c index 246d727b65b7..28b078133688 100644 --- a/arch/x86/kernel/amd_iommu_init.c +++ b/arch/x86/kernel/amd_iommu_init.c | |||
@@ -137,6 +137,7 @@ int amd_iommus_present; | |||
137 | 137 | ||
138 | /* IOMMUs have a non-present cache? */ | 138 | /* IOMMUs have a non-present cache? */ |
139 | bool amd_iommu_np_cache __read_mostly; | 139 | bool amd_iommu_np_cache __read_mostly; |
140 | bool amd_iommu_iotlb_sup __read_mostly = true; | ||
140 | 141 | ||
141 | /* | 142 | /* |
142 | * The ACPI table parsing functions set this variable on an error | 143 | * The ACPI table parsing functions set this variable on an error |
@@ -180,6 +181,12 @@ static u32 dev_table_size; /* size of the device table */ | |||
180 | static u32 alias_table_size; /* size of the alias table */ | 181 | static u32 alias_table_size; /* size of the alias table */ |
181 | static u32 rlookup_table_size; /* size if the rlookup table */ | 182 | static u32 rlookup_table_size; /* size if the rlookup table */ |
182 | 183 | ||
184 | /* | ||
185 | * This function flushes all internal caches of | ||
186 | * the IOMMU used by this driver. | ||
187 | */ | ||
188 | extern void iommu_flush_all_caches(struct amd_iommu *iommu); | ||
189 | |||
183 | static inline void update_last_devid(u16 devid) | 190 | static inline void update_last_devid(u16 devid) |
184 | { | 191 | { |
185 | if (devid > amd_iommu_last_bdf) | 192 | if (devid > amd_iommu_last_bdf) |
@@ -293,9 +300,23 @@ static void iommu_feature_disable(struct amd_iommu *iommu, u8 bit) | |||
293 | /* Function to enable the hardware */ | 300 | /* Function to enable the hardware */ |
294 | static void iommu_enable(struct amd_iommu *iommu) | 301 | static void iommu_enable(struct amd_iommu *iommu) |
295 | { | 302 | { |
296 | printk(KERN_INFO "AMD-Vi: Enabling IOMMU at %s cap 0x%hx\n", | 303 | static const char * const feat_str[] = { |
304 | "PreF", "PPR", "X2APIC", "NX", "GT", "[5]", | ||
305 | "IA", "GA", "HE", "PC", NULL | ||
306 | }; | ||
307 | int i; | ||
308 | |||
309 | printk(KERN_INFO "AMD-Vi: Enabling IOMMU at %s cap 0x%hx", | ||
297 | dev_name(&iommu->dev->dev), iommu->cap_ptr); | 310 | dev_name(&iommu->dev->dev), iommu->cap_ptr); |
298 | 311 | ||
312 | if (iommu->cap & (1 << IOMMU_CAP_EFR)) { | ||
313 | printk(KERN_CONT " extended features: "); | ||
314 | for (i = 0; feat_str[i]; ++i) | ||
315 | if (iommu_feature(iommu, (1ULL << i))) | ||
316 | printk(KERN_CONT " %s", feat_str[i]); | ||
317 | } | ||
318 | printk(KERN_CONT "\n"); | ||
319 | |||
299 | iommu_feature_enable(iommu, CONTROL_IOMMU_EN); | 320 | iommu_feature_enable(iommu, CONTROL_IOMMU_EN); |
300 | } | 321 | } |
301 | 322 | ||
@@ -651,7 +672,7 @@ static void __init set_device_exclusion_range(u16 devid, struct ivmd_header *m) | |||
651 | static void __init init_iommu_from_pci(struct amd_iommu *iommu) | 672 | static void __init init_iommu_from_pci(struct amd_iommu *iommu) |
652 | { | 673 | { |
653 | int cap_ptr = iommu->cap_ptr; | 674 | int cap_ptr = iommu->cap_ptr; |
654 | u32 range, misc; | 675 | u32 range, misc, low, high; |
655 | int i, j; | 676 | int i, j; |
656 | 677 | ||
657 | pci_read_config_dword(iommu->dev, cap_ptr + MMIO_CAP_HDR_OFFSET, | 678 | pci_read_config_dword(iommu->dev, cap_ptr + MMIO_CAP_HDR_OFFSET, |
@@ -667,6 +688,15 @@ static void __init init_iommu_from_pci(struct amd_iommu *iommu) | |||
667 | MMIO_GET_LD(range)); | 688 | MMIO_GET_LD(range)); |
668 | iommu->evt_msi_num = MMIO_MSI_NUM(misc); | 689 | iommu->evt_msi_num = MMIO_MSI_NUM(misc); |
669 | 690 | ||
691 | if (!(iommu->cap & (1 << IOMMU_CAP_IOTLB))) | ||
692 | amd_iommu_iotlb_sup = false; | ||
693 | |||
694 | /* read extended feature bits */ | ||
695 | low = readl(iommu->mmio_base + MMIO_EXT_FEATURES); | ||
696 | high = readl(iommu->mmio_base + MMIO_EXT_FEATURES + 4); | ||
697 | |||
698 | iommu->features = ((u64)high << 32) | low; | ||
699 | |||
670 | if (!is_rd890_iommu(iommu->dev)) | 700 | if (!is_rd890_iommu(iommu->dev)) |
671 | return; | 701 | return; |
672 | 702 | ||
@@ -1244,6 +1274,7 @@ static void enable_iommus(void) | |||
1244 | iommu_set_exclusion_range(iommu); | 1274 | iommu_set_exclusion_range(iommu); |
1245 | iommu_init_msi(iommu); | 1275 | iommu_init_msi(iommu); |
1246 | iommu_enable(iommu); | 1276 | iommu_enable(iommu); |
1277 | iommu_flush_all_caches(iommu); | ||
1247 | } | 1278 | } |
1248 | } | 1279 | } |
1249 | 1280 | ||
@@ -1274,8 +1305,8 @@ static void amd_iommu_resume(void) | |||
1274 | * we have to flush after the IOMMUs are enabled because a | 1305 | * we have to flush after the IOMMUs are enabled because a |
1275 | * disabled IOMMU will never execute the commands we send | 1306 | * disabled IOMMU will never execute the commands we send |
1276 | */ | 1307 | */ |
1277 | amd_iommu_flush_all_devices(); | 1308 | for_each_iommu(iommu) |
1278 | amd_iommu_flush_all_domains(); | 1309 | iommu_flush_all_caches(iommu); |
1279 | } | 1310 | } |
1280 | 1311 | ||
1281 | static int amd_iommu_suspend(void) | 1312 | static int amd_iommu_suspend(void) |
diff --git a/drivers/pci/intel-iommu.c b/drivers/pci/intel-iommu.c index d552d2c77844..6af6b628175b 100644 --- a/drivers/pci/intel-iommu.c +++ b/drivers/pci/intel-iommu.c | |||
@@ -39,6 +39,7 @@ | |||
39 | #include <linux/syscore_ops.h> | 39 | #include <linux/syscore_ops.h> |
40 | #include <linux/tboot.h> | 40 | #include <linux/tboot.h> |
41 | #include <linux/dmi.h> | 41 | #include <linux/dmi.h> |
42 | #include <linux/pci-ats.h> | ||
42 | #include <asm/cacheflush.h> | 43 | #include <asm/cacheflush.h> |
43 | #include <asm/iommu.h> | 44 | #include <asm/iommu.h> |
44 | #include "pci.h" | 45 | #include "pci.h" |
diff --git a/drivers/pci/iov.c b/drivers/pci/iov.c index 553d8ee55c1c..42fae4776515 100644 --- a/drivers/pci/iov.c +++ b/drivers/pci/iov.c | |||
@@ -13,6 +13,7 @@ | |||
13 | #include <linux/mutex.h> | 13 | #include <linux/mutex.h> |
14 | #include <linux/string.h> | 14 | #include <linux/string.h> |
15 | #include <linux/delay.h> | 15 | #include <linux/delay.h> |
16 | #include <linux/pci-ats.h> | ||
16 | #include "pci.h" | 17 | #include "pci.h" |
17 | 18 | ||
18 | #define VIRTFN_ID_LEN 16 | 19 | #define VIRTFN_ID_LEN 16 |
diff --git a/drivers/pci/pci.h b/drivers/pci/pci.h index a6ec200fe5ee..4020025f854e 100644 --- a/drivers/pci/pci.h +++ b/drivers/pci/pci.h | |||
@@ -250,15 +250,6 @@ struct pci_sriov { | |||
250 | u8 __iomem *mstate; /* VF Migration State Array */ | 250 | u8 __iomem *mstate; /* VF Migration State Array */ |
251 | }; | 251 | }; |
252 | 252 | ||
253 | /* Address Translation Service */ | ||
254 | struct pci_ats { | ||
255 | int pos; /* capability position */ | ||
256 | int stu; /* Smallest Translation Unit */ | ||
257 | int qdep; /* Invalidate Queue Depth */ | ||
258 | int ref_cnt; /* Physical Function reference count */ | ||
259 | unsigned int is_enabled:1; /* Enable bit is set */ | ||
260 | }; | ||
261 | |||
262 | #ifdef CONFIG_PCI_IOV | 253 | #ifdef CONFIG_PCI_IOV |
263 | extern int pci_iov_init(struct pci_dev *dev); | 254 | extern int pci_iov_init(struct pci_dev *dev); |
264 | extern void pci_iov_release(struct pci_dev *dev); | 255 | extern void pci_iov_release(struct pci_dev *dev); |
@@ -269,19 +260,6 @@ extern resource_size_t pci_sriov_resource_alignment(struct pci_dev *dev, | |||
269 | extern void pci_restore_iov_state(struct pci_dev *dev); | 260 | extern void pci_restore_iov_state(struct pci_dev *dev); |
270 | extern int pci_iov_bus_range(struct pci_bus *bus); | 261 | extern int pci_iov_bus_range(struct pci_bus *bus); |
271 | 262 | ||
272 | extern int pci_enable_ats(struct pci_dev *dev, int ps); | ||
273 | extern void pci_disable_ats(struct pci_dev *dev); | ||
274 | extern int pci_ats_queue_depth(struct pci_dev *dev); | ||
275 | /** | ||
276 | * pci_ats_enabled - query the ATS status | ||
277 | * @dev: the PCI device | ||
278 | * | ||
279 | * Returns 1 if ATS capability is enabled, or 0 if not. | ||
280 | */ | ||
281 | static inline int pci_ats_enabled(struct pci_dev *dev) | ||
282 | { | ||
283 | return dev->ats && dev->ats->is_enabled; | ||
284 | } | ||
285 | #else | 263 | #else |
286 | static inline int pci_iov_init(struct pci_dev *dev) | 264 | static inline int pci_iov_init(struct pci_dev *dev) |
287 | { | 265 | { |
@@ -304,21 +282,6 @@ static inline int pci_iov_bus_range(struct pci_bus *bus) | |||
304 | return 0; | 282 | return 0; |
305 | } | 283 | } |
306 | 284 | ||
307 | static inline int pci_enable_ats(struct pci_dev *dev, int ps) | ||
308 | { | ||
309 | return -ENODEV; | ||
310 | } | ||
311 | static inline void pci_disable_ats(struct pci_dev *dev) | ||
312 | { | ||
313 | } | ||
314 | static inline int pci_ats_queue_depth(struct pci_dev *dev) | ||
315 | { | ||
316 | return -ENODEV; | ||
317 | } | ||
318 | static inline int pci_ats_enabled(struct pci_dev *dev) | ||
319 | { | ||
320 | return 0; | ||
321 | } | ||
322 | #endif /* CONFIG_PCI_IOV */ | 285 | #endif /* CONFIG_PCI_IOV */ |
323 | 286 | ||
324 | static inline resource_size_t pci_resource_alignment(struct pci_dev *dev, | 287 | static inline resource_size_t pci_resource_alignment(struct pci_dev *dev, |
diff --git a/include/linux/pci-ats.h b/include/linux/pci-ats.h new file mode 100644 index 000000000000..655824fa4c76 --- /dev/null +++ b/include/linux/pci-ats.h | |||
@@ -0,0 +1,52 @@ | |||
1 | #ifndef LINUX_PCI_ATS_H | ||
2 | #define LINUX_PCI_ATS_H | ||
3 | |||
4 | /* Address Translation Service */ | ||
5 | struct pci_ats { | ||
6 | int pos; /* capability position */ | ||
7 | int stu; /* Smallest Translation Unit */ | ||
8 | int qdep; /* Invalidate Queue Depth */ | ||
9 | int ref_cnt; /* Physical Function reference count */ | ||
10 | unsigned int is_enabled:1; /* Enable bit is set */ | ||
11 | }; | ||
12 | |||
13 | #ifdef CONFIG_PCI_IOV | ||
14 | |||
15 | extern int pci_enable_ats(struct pci_dev *dev, int ps); | ||
16 | extern void pci_disable_ats(struct pci_dev *dev); | ||
17 | extern int pci_ats_queue_depth(struct pci_dev *dev); | ||
18 | /** | ||
19 | * pci_ats_enabled - query the ATS status | ||
20 | * @dev: the PCI device | ||
21 | * | ||
22 | * Returns 1 if ATS capability is enabled, or 0 if not. | ||
23 | */ | ||
24 | static inline int pci_ats_enabled(struct pci_dev *dev) | ||
25 | { | ||
26 | return dev->ats && dev->ats->is_enabled; | ||
27 | } | ||
28 | |||
29 | #else /* CONFIG_PCI_IOV */ | ||
30 | |||
31 | static inline int pci_enable_ats(struct pci_dev *dev, int ps) | ||
32 | { | ||
33 | return -ENODEV; | ||
34 | } | ||
35 | |||
36 | static inline void pci_disable_ats(struct pci_dev *dev) | ||
37 | { | ||
38 | } | ||
39 | |||
40 | static inline int pci_ats_queue_depth(struct pci_dev *dev) | ||
41 | { | ||
42 | return -ENODEV; | ||
43 | } | ||
44 | |||
45 | static inline int pci_ats_enabled(struct pci_dev *dev) | ||
46 | { | ||
47 | return 0; | ||
48 | } | ||
49 | |||
50 | #endif /* CONFIG_PCI_IOV */ | ||
51 | |||
52 | #endif /* LINUX_PCI_ATS_H*/ | ||
diff --git a/lib/dma-debug.c b/lib/dma-debug.c index 4bfb0471f106..db07bfd9298e 100644 --- a/lib/dma-debug.c +++ b/lib/dma-debug.c | |||
@@ -649,7 +649,7 @@ out_err: | |||
649 | return -ENOMEM; | 649 | return -ENOMEM; |
650 | } | 650 | } |
651 | 651 | ||
652 | static int device_dma_allocations(struct device *dev) | 652 | static int device_dma_allocations(struct device *dev, struct dma_debug_entry **out_entry) |
653 | { | 653 | { |
654 | struct dma_debug_entry *entry; | 654 | struct dma_debug_entry *entry; |
655 | unsigned long flags; | 655 | unsigned long flags; |
@@ -660,8 +660,10 @@ static int device_dma_allocations(struct device *dev) | |||
660 | for (i = 0; i < HASH_SIZE; ++i) { | 660 | for (i = 0; i < HASH_SIZE; ++i) { |
661 | spin_lock(&dma_entry_hash[i].lock); | 661 | spin_lock(&dma_entry_hash[i].lock); |
662 | list_for_each_entry(entry, &dma_entry_hash[i].list, list) { | 662 | list_for_each_entry(entry, &dma_entry_hash[i].list, list) { |
663 | if (entry->dev == dev) | 663 | if (entry->dev == dev) { |
664 | count += 1; | 664 | count += 1; |
665 | *out_entry = entry; | ||
666 | } | ||
665 | } | 667 | } |
666 | spin_unlock(&dma_entry_hash[i].lock); | 668 | spin_unlock(&dma_entry_hash[i].lock); |
667 | } | 669 | } |
@@ -674,6 +676,7 @@ static int device_dma_allocations(struct device *dev) | |||
674 | static int dma_debug_device_change(struct notifier_block *nb, unsigned long action, void *data) | 676 | static int dma_debug_device_change(struct notifier_block *nb, unsigned long action, void *data) |
675 | { | 677 | { |
676 | struct device *dev = data; | 678 | struct device *dev = data; |
679 | struct dma_debug_entry *uninitialized_var(entry); | ||
677 | int count; | 680 | int count; |
678 | 681 | ||
679 | if (global_disable) | 682 | if (global_disable) |
@@ -681,12 +684,17 @@ static int dma_debug_device_change(struct notifier_block *nb, unsigned long acti | |||
681 | 684 | ||
682 | switch (action) { | 685 | switch (action) { |
683 | case BUS_NOTIFY_UNBOUND_DRIVER: | 686 | case BUS_NOTIFY_UNBOUND_DRIVER: |
684 | count = device_dma_allocations(dev); | 687 | count = device_dma_allocations(dev, &entry); |
685 | if (count == 0) | 688 | if (count == 0) |
686 | break; | 689 | break; |
687 | err_printk(dev, NULL, "DMA-API: device driver has pending " | 690 | err_printk(dev, entry, "DMA-API: device driver has pending " |
688 | "DMA allocations while released from device " | 691 | "DMA allocations while released from device " |
689 | "[count=%d]\n", count); | 692 | "[count=%d]\n" |
693 | "One of leaked entries details: " | ||
694 | "[device address=0x%016llx] [size=%llu bytes] " | ||
695 | "[mapped with %s] [mapped as %s]\n", | ||
696 | count, entry->dev_addr, entry->size, | ||
697 | dir2name[entry->direction], type2name[entry->type]); | ||
690 | break; | 698 | break; |
691 | default: | 699 | default: |
692 | break; | 700 | break; |