diff options
46 files changed, 4214 insertions, 263 deletions
diff --git a/MAINTAINERS b/MAINTAINERS index b28dc111d4a7..f8d882e13200 100644 --- a/MAINTAINERS +++ b/MAINTAINERS | |||
@@ -10268,6 +10268,15 @@ S: Supported | |||
10268 | F: drivers/block/xen-blkback/* | 10268 | F: drivers/block/xen-blkback/* |
10269 | F: drivers/block/xen* | 10269 | F: drivers/block/xen* |
10270 | 10270 | ||
10271 | XEN PVSCSI DRIVERS | ||
10272 | M: Juergen Gross <jgross@suse.com> | ||
10273 | L: xen-devel@lists.xenproject.org (moderated for non-subscribers) | ||
10274 | L: linux-scsi@vger.kernel.org | ||
10275 | S: Supported | ||
10276 | F: drivers/scsi/xen-scsifront.c | ||
10277 | F: drivers/xen/xen-scsiback.c | ||
10278 | F: include/xen/interface/io/vscsiif.h | ||
10279 | |||
10271 | XEN SWIOTLB SUBSYSTEM | 10280 | XEN SWIOTLB SUBSYSTEM |
10272 | M: Konrad Rzeszutek Wilk <konrad.wilk@oracle.com> | 10281 | M: Konrad Rzeszutek Wilk <konrad.wilk@oracle.com> |
10273 | L: xen-devel@lists.xenproject.org (moderated for non-subscribers) | 10282 | L: xen-devel@lists.xenproject.org (moderated for non-subscribers) |
diff --git a/arch/arm/Kconfig b/arch/arm/Kconfig index 18f392f8b744..89c4b5ccc68d 100644 --- a/arch/arm/Kconfig +++ b/arch/arm/Kconfig | |||
@@ -1779,7 +1779,7 @@ config XEN_DOM0 | |||
1779 | depends on XEN | 1779 | depends on XEN |
1780 | 1780 | ||
1781 | config XEN | 1781 | config XEN |
1782 | bool "Xen guest support on ARM (EXPERIMENTAL)" | 1782 | bool "Xen guest support on ARM" |
1783 | depends on ARM && AEABI && OF | 1783 | depends on ARM && AEABI && OF |
1784 | depends on CPU_V7 && !CPU_V6 | 1784 | depends on CPU_V7 && !CPU_V6 |
1785 | depends on !GENERIC_ATOMIC64 | 1785 | depends on !GENERIC_ATOMIC64 |
diff --git a/arch/arm64/Kconfig b/arch/arm64/Kconfig index c49ca4c738bb..ac9afde76dea 100644 --- a/arch/arm64/Kconfig +++ b/arch/arm64/Kconfig | |||
@@ -349,7 +349,7 @@ config XEN_DOM0 | |||
349 | depends on XEN | 349 | depends on XEN |
350 | 350 | ||
351 | config XEN | 351 | config XEN |
352 | bool "Xen guest support on ARM64 (EXPERIMENTAL)" | 352 | bool "Xen guest support on ARM64" |
353 | depends on ARM64 && OF | 353 | depends on ARM64 && OF |
354 | select SWIOTLB_XEN | 354 | select SWIOTLB_XEN |
355 | help | 355 | help |
diff --git a/arch/x86/include/asm/pgtable_types.h b/arch/x86/include/asm/pgtable_types.h index 0f9724c9c510..07789647bf33 100644 --- a/arch/x86/include/asm/pgtable_types.h +++ b/arch/x86/include/asm/pgtable_types.h | |||
@@ -23,7 +23,6 @@ | |||
23 | #define _PAGE_BIT_SPECIAL _PAGE_BIT_SOFTW1 | 23 | #define _PAGE_BIT_SPECIAL _PAGE_BIT_SOFTW1 |
24 | #define _PAGE_BIT_CPA_TEST _PAGE_BIT_SOFTW1 | 24 | #define _PAGE_BIT_CPA_TEST _PAGE_BIT_SOFTW1 |
25 | #define _PAGE_BIT_SPLITTING _PAGE_BIT_SOFTW2 /* only valid on a PSE pmd */ | 25 | #define _PAGE_BIT_SPLITTING _PAGE_BIT_SOFTW2 /* only valid on a PSE pmd */ |
26 | #define _PAGE_BIT_IOMAP _PAGE_BIT_SOFTW2 /* flag used to indicate IO mapping */ | ||
27 | #define _PAGE_BIT_HIDDEN _PAGE_BIT_SOFTW3 /* hidden by kmemcheck */ | 26 | #define _PAGE_BIT_HIDDEN _PAGE_BIT_SOFTW3 /* hidden by kmemcheck */ |
28 | #define _PAGE_BIT_SOFT_DIRTY _PAGE_BIT_SOFTW3 /* software dirty tracking */ | 27 | #define _PAGE_BIT_SOFT_DIRTY _PAGE_BIT_SOFTW3 /* software dirty tracking */ |
29 | #define _PAGE_BIT_NX 63 /* No execute: only valid after cpuid check */ | 28 | #define _PAGE_BIT_NX 63 /* No execute: only valid after cpuid check */ |
@@ -52,7 +51,7 @@ | |||
52 | #define _PAGE_PSE (_AT(pteval_t, 1) << _PAGE_BIT_PSE) | 51 | #define _PAGE_PSE (_AT(pteval_t, 1) << _PAGE_BIT_PSE) |
53 | #define _PAGE_GLOBAL (_AT(pteval_t, 1) << _PAGE_BIT_GLOBAL) | 52 | #define _PAGE_GLOBAL (_AT(pteval_t, 1) << _PAGE_BIT_GLOBAL) |
54 | #define _PAGE_SOFTW1 (_AT(pteval_t, 1) << _PAGE_BIT_SOFTW1) | 53 | #define _PAGE_SOFTW1 (_AT(pteval_t, 1) << _PAGE_BIT_SOFTW1) |
55 | #define _PAGE_IOMAP (_AT(pteval_t, 1) << _PAGE_BIT_IOMAP) | 54 | #define _PAGE_SOFTW2 (_AT(pteval_t, 1) << _PAGE_BIT_SOFTW2) |
56 | #define _PAGE_PAT (_AT(pteval_t, 1) << _PAGE_BIT_PAT) | 55 | #define _PAGE_PAT (_AT(pteval_t, 1) << _PAGE_BIT_PAT) |
57 | #define _PAGE_PAT_LARGE (_AT(pteval_t, 1) << _PAGE_BIT_PAT_LARGE) | 56 | #define _PAGE_PAT_LARGE (_AT(pteval_t, 1) << _PAGE_BIT_PAT_LARGE) |
58 | #define _PAGE_SPECIAL (_AT(pteval_t, 1) << _PAGE_BIT_SPECIAL) | 57 | #define _PAGE_SPECIAL (_AT(pteval_t, 1) << _PAGE_BIT_SPECIAL) |
@@ -168,10 +167,10 @@ | |||
168 | #define __PAGE_KERNEL_LARGE_NOCACHE (__PAGE_KERNEL | _PAGE_CACHE_UC | _PAGE_PSE) | 167 | #define __PAGE_KERNEL_LARGE_NOCACHE (__PAGE_KERNEL | _PAGE_CACHE_UC | _PAGE_PSE) |
169 | #define __PAGE_KERNEL_LARGE_EXEC (__PAGE_KERNEL_EXEC | _PAGE_PSE) | 168 | #define __PAGE_KERNEL_LARGE_EXEC (__PAGE_KERNEL_EXEC | _PAGE_PSE) |
170 | 169 | ||
171 | #define __PAGE_KERNEL_IO (__PAGE_KERNEL | _PAGE_IOMAP) | 170 | #define __PAGE_KERNEL_IO (__PAGE_KERNEL) |
172 | #define __PAGE_KERNEL_IO_NOCACHE (__PAGE_KERNEL_NOCACHE | _PAGE_IOMAP) | 171 | #define __PAGE_KERNEL_IO_NOCACHE (__PAGE_KERNEL_NOCACHE) |
173 | #define __PAGE_KERNEL_IO_UC_MINUS (__PAGE_KERNEL_UC_MINUS | _PAGE_IOMAP) | 172 | #define __PAGE_KERNEL_IO_UC_MINUS (__PAGE_KERNEL_UC_MINUS) |
174 | #define __PAGE_KERNEL_IO_WC (__PAGE_KERNEL_WC | _PAGE_IOMAP) | 173 | #define __PAGE_KERNEL_IO_WC (__PAGE_KERNEL_WC) |
175 | 174 | ||
176 | #define PAGE_KERNEL __pgprot(__PAGE_KERNEL) | 175 | #define PAGE_KERNEL __pgprot(__PAGE_KERNEL) |
177 | #define PAGE_KERNEL_RO __pgprot(__PAGE_KERNEL_RO) | 176 | #define PAGE_KERNEL_RO __pgprot(__PAGE_KERNEL_RO) |
diff --git a/arch/x86/mm/fault.c b/arch/x86/mm/fault.c index a24194681513..83bb03bfa259 100644 --- a/arch/x86/mm/fault.c +++ b/arch/x86/mm/fault.c | |||
@@ -933,8 +933,17 @@ static int spurious_fault_check(unsigned long error_code, pte_t *pte) | |||
933 | * cross-processor TLB flush, even if no stale TLB entries exist | 933 | * cross-processor TLB flush, even if no stale TLB entries exist |
934 | * on other processors. | 934 | * on other processors. |
935 | * | 935 | * |
936 | * Spurious faults may only occur if the TLB contains an entry with | ||
937 | * fewer permission than the page table entry. Non-present (P = 0) | ||
938 | * and reserved bit (R = 1) faults are never spurious. | ||
939 | * | ||
936 | * There are no security implications to leaving a stale TLB when | 940 | * There are no security implications to leaving a stale TLB when |
937 | * increasing the permissions on a page. | 941 | * increasing the permissions on a page. |
942 | * | ||
943 | * Returns non-zero if a spurious fault was handled, zero otherwise. | ||
944 | * | ||
945 | * See Intel Developer's Manual Vol 3 Section 4.10.4.3, bullet 3 | ||
946 | * (Optional Invalidation). | ||
938 | */ | 947 | */ |
939 | static noinline int | 948 | static noinline int |
940 | spurious_fault(unsigned long error_code, unsigned long address) | 949 | spurious_fault(unsigned long error_code, unsigned long address) |
@@ -945,8 +954,17 @@ spurious_fault(unsigned long error_code, unsigned long address) | |||
945 | pte_t *pte; | 954 | pte_t *pte; |
946 | int ret; | 955 | int ret; |
947 | 956 | ||
948 | /* Reserved-bit violation or user access to kernel space? */ | 957 | /* |
949 | if (error_code & (PF_USER | PF_RSVD)) | 958 | * Only writes to RO or instruction fetches from NX may cause |
959 | * spurious faults. | ||
960 | * | ||
961 | * These could be from user or supervisor accesses but the TLB | ||
962 | * is only lazily flushed after a kernel mapping protection | ||
963 | * change, so user accesses are not expected to cause spurious | ||
964 | * faults. | ||
965 | */ | ||
966 | if (error_code != (PF_WRITE | PF_PROT) | ||
967 | && error_code != (PF_INSTR | PF_PROT)) | ||
950 | return 0; | 968 | return 0; |
951 | 969 | ||
952 | pgd = init_mm.pgd + pgd_index(address); | 970 | pgd = init_mm.pgd + pgd_index(address); |
diff --git a/arch/x86/mm/init_32.c b/arch/x86/mm/init_32.c index 7d05565ba781..c8140e12816a 100644 --- a/arch/x86/mm/init_32.c +++ b/arch/x86/mm/init_32.c | |||
@@ -537,7 +537,7 @@ static void __init pagetable_init(void) | |||
537 | permanent_kmaps_init(pgd_base); | 537 | permanent_kmaps_init(pgd_base); |
538 | } | 538 | } |
539 | 539 | ||
540 | pteval_t __supported_pte_mask __read_mostly = ~(_PAGE_NX | _PAGE_GLOBAL | _PAGE_IOMAP); | 540 | pteval_t __supported_pte_mask __read_mostly = ~(_PAGE_NX | _PAGE_GLOBAL); |
541 | EXPORT_SYMBOL_GPL(__supported_pte_mask); | 541 | EXPORT_SYMBOL_GPL(__supported_pte_mask); |
542 | 542 | ||
543 | /* user-defined highmem size */ | 543 | /* user-defined highmem size */ |
diff --git a/arch/x86/mm/init_64.c b/arch/x86/mm/init_64.c index 5621c47d7a1a..5d984769cbd8 100644 --- a/arch/x86/mm/init_64.c +++ b/arch/x86/mm/init_64.c | |||
@@ -151,7 +151,7 @@ early_param("gbpages", parse_direct_gbpages_on); | |||
151 | * around without checking the pgd every time. | 151 | * around without checking the pgd every time. |
152 | */ | 152 | */ |
153 | 153 | ||
154 | pteval_t __supported_pte_mask __read_mostly = ~_PAGE_IOMAP; | 154 | pteval_t __supported_pte_mask __read_mostly = ~0; |
155 | EXPORT_SYMBOL_GPL(__supported_pte_mask); | 155 | EXPORT_SYMBOL_GPL(__supported_pte_mask); |
156 | 156 | ||
157 | int force_personality32; | 157 | int force_personality32; |
diff --git a/arch/x86/pci/i386.c b/arch/x86/pci/i386.c index 2ae525e0d8ba..37c1435889ce 100644 --- a/arch/x86/pci/i386.c +++ b/arch/x86/pci/i386.c | |||
@@ -442,8 +442,6 @@ int pci_mmap_page_range(struct pci_dev *dev, struct vm_area_struct *vma, | |||
442 | */ | 442 | */ |
443 | prot |= _PAGE_CACHE_UC_MINUS; | 443 | prot |= _PAGE_CACHE_UC_MINUS; |
444 | 444 | ||
445 | prot |= _PAGE_IOMAP; /* creating a mapping for IO */ | ||
446 | |||
447 | vma->vm_page_prot = __pgprot(prot); | 445 | vma->vm_page_prot = __pgprot(prot); |
448 | 446 | ||
449 | if (io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, | 447 | if (io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, |
diff --git a/arch/x86/xen/efi.c b/arch/x86/xen/efi.c index a02e09e18f57..be14cc3e48d5 100644 --- a/arch/x86/xen/efi.c +++ b/arch/x86/xen/efi.c | |||
@@ -15,12 +15,14 @@ | |||
15 | * with this program. If not, see <http://www.gnu.org/licenses/>. | 15 | * with this program. If not, see <http://www.gnu.org/licenses/>. |
16 | */ | 16 | */ |
17 | 17 | ||
18 | #include <linux/bitops.h> | ||
18 | #include <linux/efi.h> | 19 | #include <linux/efi.h> |
19 | #include <linux/init.h> | 20 | #include <linux/init.h> |
20 | #include <linux/string.h> | 21 | #include <linux/string.h> |
21 | 22 | ||
22 | #include <xen/xen-ops.h> | 23 | #include <xen/xen-ops.h> |
23 | 24 | ||
25 | #include <asm/page.h> | ||
24 | #include <asm/setup.h> | 26 | #include <asm/setup.h> |
25 | 27 | ||
26 | void __init xen_efi_init(void) | 28 | void __init xen_efi_init(void) |
diff --git a/arch/x86/xen/enlighten.c b/arch/x86/xen/enlighten.c index c0cb11fb5008..acb0effd8077 100644 --- a/arch/x86/xen/enlighten.c +++ b/arch/x86/xen/enlighten.c | |||
@@ -1463,6 +1463,7 @@ static void __ref xen_setup_gdt(int cpu) | |||
1463 | pv_cpu_ops.load_gdt = xen_load_gdt; | 1463 | pv_cpu_ops.load_gdt = xen_load_gdt; |
1464 | } | 1464 | } |
1465 | 1465 | ||
1466 | #ifdef CONFIG_XEN_PVH | ||
1466 | /* | 1467 | /* |
1467 | * A PV guest starts with default flags that are not set for PVH, set them | 1468 | * A PV guest starts with default flags that are not set for PVH, set them |
1468 | * here asap. | 1469 | * here asap. |
@@ -1508,17 +1509,21 @@ static void __init xen_pvh_early_guest_init(void) | |||
1508 | return; | 1509 | return; |
1509 | 1510 | ||
1510 | xen_have_vector_callback = 1; | 1511 | xen_have_vector_callback = 1; |
1512 | |||
1513 | xen_pvh_early_cpu_init(0, false); | ||
1511 | xen_pvh_set_cr_flags(0); | 1514 | xen_pvh_set_cr_flags(0); |
1512 | 1515 | ||
1513 | #ifdef CONFIG_X86_32 | 1516 | #ifdef CONFIG_X86_32 |
1514 | BUG(); /* PVH: Implement proper support. */ | 1517 | BUG(); /* PVH: Implement proper support. */ |
1515 | #endif | 1518 | #endif |
1516 | } | 1519 | } |
1520 | #endif /* CONFIG_XEN_PVH */ | ||
1517 | 1521 | ||
1518 | /* First C function to be called on Xen boot */ | 1522 | /* First C function to be called on Xen boot */ |
1519 | asmlinkage __visible void __init xen_start_kernel(void) | 1523 | asmlinkage __visible void __init xen_start_kernel(void) |
1520 | { | 1524 | { |
1521 | struct physdev_set_iopl set_iopl; | 1525 | struct physdev_set_iopl set_iopl; |
1526 | unsigned long initrd_start = 0; | ||
1522 | int rc; | 1527 | int rc; |
1523 | 1528 | ||
1524 | if (!xen_start_info) | 1529 | if (!xen_start_info) |
@@ -1527,7 +1532,9 @@ asmlinkage __visible void __init xen_start_kernel(void) | |||
1527 | xen_domain_type = XEN_PV_DOMAIN; | 1532 | xen_domain_type = XEN_PV_DOMAIN; |
1528 | 1533 | ||
1529 | xen_setup_features(); | 1534 | xen_setup_features(); |
1535 | #ifdef CONFIG_XEN_PVH | ||
1530 | xen_pvh_early_guest_init(); | 1536 | xen_pvh_early_guest_init(); |
1537 | #endif | ||
1531 | xen_setup_machphys_mapping(); | 1538 | xen_setup_machphys_mapping(); |
1532 | 1539 | ||
1533 | /* Install Xen paravirt ops */ | 1540 | /* Install Xen paravirt ops */ |
@@ -1559,8 +1566,6 @@ asmlinkage __visible void __init xen_start_kernel(void) | |||
1559 | #endif | 1566 | #endif |
1560 | __supported_pte_mask &= ~(_PAGE_PWT | _PAGE_PCD); | 1567 | __supported_pte_mask &= ~(_PAGE_PWT | _PAGE_PCD); |
1561 | 1568 | ||
1562 | __supported_pte_mask |= _PAGE_IOMAP; | ||
1563 | |||
1564 | /* | 1569 | /* |
1565 | * Prevent page tables from being allocated in highmem, even | 1570 | * Prevent page tables from being allocated in highmem, even |
1566 | * if CONFIG_HIGHPTE is enabled. | 1571 | * if CONFIG_HIGHPTE is enabled. |
@@ -1667,10 +1672,16 @@ asmlinkage __visible void __init xen_start_kernel(void) | |||
1667 | new_cpu_data.x86_capability[0] = cpuid_edx(1); | 1672 | new_cpu_data.x86_capability[0] = cpuid_edx(1); |
1668 | #endif | 1673 | #endif |
1669 | 1674 | ||
1675 | if (xen_start_info->mod_start) { | ||
1676 | if (xen_start_info->flags & SIF_MOD_START_PFN) | ||
1677 | initrd_start = PFN_PHYS(xen_start_info->mod_start); | ||
1678 | else | ||
1679 | initrd_start = __pa(xen_start_info->mod_start); | ||
1680 | } | ||
1681 | |||
1670 | /* Poke various useful things into boot_params */ | 1682 | /* Poke various useful things into boot_params */ |
1671 | boot_params.hdr.type_of_loader = (9 << 4) | 0; | 1683 | boot_params.hdr.type_of_loader = (9 << 4) | 0; |
1672 | boot_params.hdr.ramdisk_image = xen_start_info->mod_start | 1684 | boot_params.hdr.ramdisk_image = initrd_start; |
1673 | ? __pa(xen_start_info->mod_start) : 0; | ||
1674 | boot_params.hdr.ramdisk_size = xen_start_info->mod_len; | 1685 | boot_params.hdr.ramdisk_size = xen_start_info->mod_len; |
1675 | boot_params.hdr.cmd_line_ptr = __pa(xen_start_info->cmd_line); | 1686 | boot_params.hdr.cmd_line_ptr = __pa(xen_start_info->cmd_line); |
1676 | 1687 | ||
diff --git a/arch/x86/xen/mmu.c b/arch/x86/xen/mmu.c index 16fb0099b7f2..f62af7647ec9 100644 --- a/arch/x86/xen/mmu.c +++ b/arch/x86/xen/mmu.c | |||
@@ -399,38 +399,14 @@ static pteval_t pte_pfn_to_mfn(pteval_t val) | |||
399 | if (unlikely(mfn == INVALID_P2M_ENTRY)) { | 399 | if (unlikely(mfn == INVALID_P2M_ENTRY)) { |
400 | mfn = 0; | 400 | mfn = 0; |
401 | flags = 0; | 401 | flags = 0; |
402 | } else { | 402 | } else |
403 | /* | 403 | mfn &= ~(FOREIGN_FRAME_BIT | IDENTITY_FRAME_BIT); |
404 | * Paramount to do this test _after_ the | ||
405 | * INVALID_P2M_ENTRY as INVALID_P2M_ENTRY & | ||
406 | * IDENTITY_FRAME_BIT resolves to true. | ||
407 | */ | ||
408 | mfn &= ~FOREIGN_FRAME_BIT; | ||
409 | if (mfn & IDENTITY_FRAME_BIT) { | ||
410 | mfn &= ~IDENTITY_FRAME_BIT; | ||
411 | flags |= _PAGE_IOMAP; | ||
412 | } | ||
413 | } | ||
414 | val = ((pteval_t)mfn << PAGE_SHIFT) | flags; | 404 | val = ((pteval_t)mfn << PAGE_SHIFT) | flags; |
415 | } | 405 | } |
416 | 406 | ||
417 | return val; | 407 | return val; |
418 | } | 408 | } |
419 | 409 | ||
420 | static pteval_t iomap_pte(pteval_t val) | ||
421 | { | ||
422 | if (val & _PAGE_PRESENT) { | ||
423 | unsigned long pfn = (val & PTE_PFN_MASK) >> PAGE_SHIFT; | ||
424 | pteval_t flags = val & PTE_FLAGS_MASK; | ||
425 | |||
426 | /* We assume the pte frame number is a MFN, so | ||
427 | just use it as-is. */ | ||
428 | val = ((pteval_t)pfn << PAGE_SHIFT) | flags; | ||
429 | } | ||
430 | |||
431 | return val; | ||
432 | } | ||
433 | |||
434 | __visible pteval_t xen_pte_val(pte_t pte) | 410 | __visible pteval_t xen_pte_val(pte_t pte) |
435 | { | 411 | { |
436 | pteval_t pteval = pte.pte; | 412 | pteval_t pteval = pte.pte; |
@@ -441,9 +417,6 @@ __visible pteval_t xen_pte_val(pte_t pte) | |||
441 | pteval = (pteval & ~_PAGE_PAT) | _PAGE_PWT; | 417 | pteval = (pteval & ~_PAGE_PAT) | _PAGE_PWT; |
442 | } | 418 | } |
443 | #endif | 419 | #endif |
444 | if (xen_initial_domain() && (pteval & _PAGE_IOMAP)) | ||
445 | return pteval; | ||
446 | |||
447 | return pte_mfn_to_pfn(pteval); | 420 | return pte_mfn_to_pfn(pteval); |
448 | } | 421 | } |
449 | PV_CALLEE_SAVE_REGS_THUNK(xen_pte_val); | 422 | PV_CALLEE_SAVE_REGS_THUNK(xen_pte_val); |
@@ -481,7 +454,6 @@ void xen_set_pat(u64 pat) | |||
481 | 454 | ||
482 | __visible pte_t xen_make_pte(pteval_t pte) | 455 | __visible pte_t xen_make_pte(pteval_t pte) |
483 | { | 456 | { |
484 | phys_addr_t addr = (pte & PTE_PFN_MASK); | ||
485 | #if 0 | 457 | #if 0 |
486 | /* If Linux is trying to set a WC pte, then map to the Xen WC. | 458 | /* If Linux is trying to set a WC pte, then map to the Xen WC. |
487 | * If _PAGE_PAT is set, then it probably means it is really | 459 | * If _PAGE_PAT is set, then it probably means it is really |
@@ -496,19 +468,7 @@ __visible pte_t xen_make_pte(pteval_t pte) | |||
496 | pte = (pte & ~(_PAGE_PCD | _PAGE_PWT)) | _PAGE_PAT; | 468 | pte = (pte & ~(_PAGE_PCD | _PAGE_PWT)) | _PAGE_PAT; |
497 | } | 469 | } |
498 | #endif | 470 | #endif |
499 | /* | 471 | pte = pte_pfn_to_mfn(pte); |
500 | * Unprivileged domains are allowed to do IOMAPpings for | ||
501 | * PCI passthrough, but not map ISA space. The ISA | ||
502 | * mappings are just dummy local mappings to keep other | ||
503 | * parts of the kernel happy. | ||
504 | */ | ||
505 | if (unlikely(pte & _PAGE_IOMAP) && | ||
506 | (xen_initial_domain() || addr >= ISA_END_ADDRESS)) { | ||
507 | pte = iomap_pte(pte); | ||
508 | } else { | ||
509 | pte &= ~_PAGE_IOMAP; | ||
510 | pte = pte_pfn_to_mfn(pte); | ||
511 | } | ||
512 | 472 | ||
513 | return native_make_pte(pte); | 473 | return native_make_pte(pte); |
514 | } | 474 | } |
@@ -2091,7 +2051,7 @@ static void xen_set_fixmap(unsigned idx, phys_addr_t phys, pgprot_t prot) | |||
2091 | 2051 | ||
2092 | default: | 2052 | default: |
2093 | /* By default, set_fixmap is used for hardware mappings */ | 2053 | /* By default, set_fixmap is used for hardware mappings */ |
2094 | pte = mfn_pte(phys, __pgprot(pgprot_val(prot) | _PAGE_IOMAP)); | 2054 | pte = mfn_pte(phys, prot); |
2095 | break; | 2055 | break; |
2096 | } | 2056 | } |
2097 | 2057 | ||
diff --git a/arch/x86/xen/p2m.c b/arch/x86/xen/p2m.c index 3172692381ae..9f5983b01ed9 100644 --- a/arch/x86/xen/p2m.c +++ b/arch/x86/xen/p2m.c | |||
@@ -173,6 +173,7 @@ | |||
173 | #include <xen/balloon.h> | 173 | #include <xen/balloon.h> |
174 | #include <xen/grant_table.h> | 174 | #include <xen/grant_table.h> |
175 | 175 | ||
176 | #include "p2m.h" | ||
176 | #include "multicalls.h" | 177 | #include "multicalls.h" |
177 | #include "xen-ops.h" | 178 | #include "xen-ops.h" |
178 | 179 | ||
@@ -180,12 +181,6 @@ static void __init m2p_override_init(void); | |||
180 | 181 | ||
181 | unsigned long xen_max_p2m_pfn __read_mostly; | 182 | unsigned long xen_max_p2m_pfn __read_mostly; |
182 | 183 | ||
183 | #define P2M_PER_PAGE (PAGE_SIZE / sizeof(unsigned long)) | ||
184 | #define P2M_MID_PER_PAGE (PAGE_SIZE / sizeof(unsigned long *)) | ||
185 | #define P2M_TOP_PER_PAGE (PAGE_SIZE / sizeof(unsigned long **)) | ||
186 | |||
187 | #define MAX_P2M_PFN (P2M_TOP_PER_PAGE * P2M_MID_PER_PAGE * P2M_PER_PAGE) | ||
188 | |||
189 | /* Placeholders for holes in the address space */ | 184 | /* Placeholders for holes in the address space */ |
190 | static RESERVE_BRK_ARRAY(unsigned long, p2m_missing, P2M_PER_PAGE); | 185 | static RESERVE_BRK_ARRAY(unsigned long, p2m_missing, P2M_PER_PAGE); |
191 | static RESERVE_BRK_ARRAY(unsigned long *, p2m_mid_missing, P2M_MID_PER_PAGE); | 186 | static RESERVE_BRK_ARRAY(unsigned long *, p2m_mid_missing, P2M_MID_PER_PAGE); |
@@ -202,16 +197,12 @@ static RESERVE_BRK_ARRAY(unsigned long, p2m_mid_identity_mfn, P2M_MID_PER_PAGE); | |||
202 | RESERVE_BRK(p2m_mid, PAGE_SIZE * (MAX_DOMAIN_PAGES / (P2M_PER_PAGE * P2M_MID_PER_PAGE))); | 197 | RESERVE_BRK(p2m_mid, PAGE_SIZE * (MAX_DOMAIN_PAGES / (P2M_PER_PAGE * P2M_MID_PER_PAGE))); |
203 | RESERVE_BRK(p2m_mid_mfn, PAGE_SIZE * (MAX_DOMAIN_PAGES / (P2M_PER_PAGE * P2M_MID_PER_PAGE))); | 198 | RESERVE_BRK(p2m_mid_mfn, PAGE_SIZE * (MAX_DOMAIN_PAGES / (P2M_PER_PAGE * P2M_MID_PER_PAGE))); |
204 | 199 | ||
205 | /* We might hit two boundary violations at the start and end, at max each | 200 | /* For each I/O range remapped we may lose up to two leaf pages for the boundary |
206 | * boundary violation will require three middle nodes. */ | 201 | * violations and three mid pages to cover up to 3GB. With |
207 | RESERVE_BRK(p2m_mid_extra, PAGE_SIZE * 2 * 3); | 202 | * early_can_reuse_p2m_middle() most of the leaf pages will be reused by the |
208 | 203 | * remapped region. | |
209 | /* When we populate back during bootup, the amount of pages can vary. The | 204 | */ |
210 | * max we have is seen is 395979, but that does not mean it can't be more. | 205 | RESERVE_BRK(p2m_identity_remap, PAGE_SIZE * 2 * 3 * MAX_REMAP_RANGES); |
211 | * Some machines can have 3GB I/O holes even. With early_can_reuse_p2m_middle | ||
212 | * it can re-use Xen provided mfn_list array, so we only need to allocate at | ||
213 | * most three P2M top nodes. */ | ||
214 | RESERVE_BRK(p2m_populated, PAGE_SIZE * 3); | ||
215 | 206 | ||
216 | static inline unsigned p2m_top_index(unsigned long pfn) | 207 | static inline unsigned p2m_top_index(unsigned long pfn) |
217 | { | 208 | { |
diff --git a/arch/x86/xen/p2m.h b/arch/x86/xen/p2m.h new file mode 100644 index 000000000000..ad8aee24ab72 --- /dev/null +++ b/arch/x86/xen/p2m.h | |||
@@ -0,0 +1,15 @@ | |||
1 | #ifndef _XEN_P2M_H | ||
2 | #define _XEN_P2M_H | ||
3 | |||
4 | #define P2M_PER_PAGE (PAGE_SIZE / sizeof(unsigned long)) | ||
5 | #define P2M_MID_PER_PAGE (PAGE_SIZE / sizeof(unsigned long *)) | ||
6 | #define P2M_TOP_PER_PAGE (PAGE_SIZE / sizeof(unsigned long **)) | ||
7 | |||
8 | #define MAX_P2M_PFN (P2M_TOP_PER_PAGE * P2M_MID_PER_PAGE * P2M_PER_PAGE) | ||
9 | |||
10 | #define MAX_REMAP_RANGES 10 | ||
11 | |||
12 | extern unsigned long __init set_phys_range_identity(unsigned long pfn_s, | ||
13 | unsigned long pfn_e); | ||
14 | |||
15 | #endif /* _XEN_P2M_H */ | ||
diff --git a/arch/x86/xen/setup.c b/arch/x86/xen/setup.c index 2e555163c2fe..af7216128d93 100644 --- a/arch/x86/xen/setup.c +++ b/arch/x86/xen/setup.c | |||
@@ -29,6 +29,7 @@ | |||
29 | #include <xen/features.h> | 29 | #include <xen/features.h> |
30 | #include "xen-ops.h" | 30 | #include "xen-ops.h" |
31 | #include "vdso.h" | 31 | #include "vdso.h" |
32 | #include "p2m.h" | ||
32 | 33 | ||
33 | /* These are code, but not functions. Defined in entry.S */ | 34 | /* These are code, but not functions. Defined in entry.S */ |
34 | extern const char xen_hypervisor_callback[]; | 35 | extern const char xen_hypervisor_callback[]; |
@@ -46,6 +47,9 @@ struct xen_memory_region xen_extra_mem[XEN_EXTRA_MEM_MAX_REGIONS] __initdata; | |||
46 | /* Number of pages released from the initial allocation. */ | 47 | /* Number of pages released from the initial allocation. */ |
47 | unsigned long xen_released_pages; | 48 | unsigned long xen_released_pages; |
48 | 49 | ||
50 | /* Buffer used to remap identity mapped pages */ | ||
51 | unsigned long xen_remap_buf[P2M_PER_PAGE] __initdata; | ||
52 | |||
49 | /* | 53 | /* |
50 | * The maximum amount of extra memory compared to the base size. The | 54 | * The maximum amount of extra memory compared to the base size. The |
51 | * main scaling factor is the size of struct page. At extreme ratios | 55 | * main scaling factor is the size of struct page. At extreme ratios |
@@ -151,107 +155,325 @@ static unsigned long __init xen_do_chunk(unsigned long start, | |||
151 | return len; | 155 | return len; |
152 | } | 156 | } |
153 | 157 | ||
154 | static unsigned long __init xen_release_chunk(unsigned long start, | 158 | /* |
155 | unsigned long end) | 159 | * Finds the next RAM pfn available in the E820 map after min_pfn. |
156 | { | 160 | * This function updates min_pfn with the pfn found and returns |
157 | return xen_do_chunk(start, end, true); | 161 | * the size of that range or zero if not found. |
158 | } | 162 | */ |
159 | 163 | static unsigned long __init xen_find_pfn_range( | |
160 | static unsigned long __init xen_populate_chunk( | ||
161 | const struct e820entry *list, size_t map_size, | 164 | const struct e820entry *list, size_t map_size, |
162 | unsigned long max_pfn, unsigned long *last_pfn, | 165 | unsigned long *min_pfn) |
163 | unsigned long credits_left) | ||
164 | { | 166 | { |
165 | const struct e820entry *entry; | 167 | const struct e820entry *entry; |
166 | unsigned int i; | 168 | unsigned int i; |
167 | unsigned long done = 0; | 169 | unsigned long done = 0; |
168 | unsigned long dest_pfn; | ||
169 | 170 | ||
170 | for (i = 0, entry = list; i < map_size; i++, entry++) { | 171 | for (i = 0, entry = list; i < map_size; i++, entry++) { |
171 | unsigned long s_pfn; | 172 | unsigned long s_pfn; |
172 | unsigned long e_pfn; | 173 | unsigned long e_pfn; |
173 | unsigned long pfns; | ||
174 | long capacity; | ||
175 | |||
176 | if (credits_left <= 0) | ||
177 | break; | ||
178 | 174 | ||
179 | if (entry->type != E820_RAM) | 175 | if (entry->type != E820_RAM) |
180 | continue; | 176 | continue; |
181 | 177 | ||
182 | e_pfn = PFN_DOWN(entry->addr + entry->size); | 178 | e_pfn = PFN_DOWN(entry->addr + entry->size); |
183 | 179 | ||
184 | /* We only care about E820 after the xen_start_info->nr_pages */ | 180 | /* We only care about E820 after this */ |
185 | if (e_pfn <= max_pfn) | 181 | if (e_pfn < *min_pfn) |
186 | continue; | 182 | continue; |
187 | 183 | ||
188 | s_pfn = PFN_UP(entry->addr); | 184 | s_pfn = PFN_UP(entry->addr); |
189 | /* If the E820 falls within the nr_pages, we want to start | 185 | |
190 | * at the nr_pages PFN. | 186 | /* If min_pfn falls within the E820 entry, we want to start |
191 | * If that would mean going past the E820 entry, skip it | 187 | * at the min_pfn PFN. |
192 | */ | 188 | */ |
193 | if (s_pfn <= max_pfn) { | 189 | if (s_pfn <= *min_pfn) { |
194 | capacity = e_pfn - max_pfn; | 190 | done = e_pfn - *min_pfn; |
195 | dest_pfn = max_pfn; | ||
196 | } else { | 191 | } else { |
197 | capacity = e_pfn - s_pfn; | 192 | done = e_pfn - s_pfn; |
198 | dest_pfn = s_pfn; | 193 | *min_pfn = s_pfn; |
199 | } | 194 | } |
195 | break; | ||
196 | } | ||
200 | 197 | ||
201 | if (credits_left < capacity) | 198 | return done; |
202 | capacity = credits_left; | 199 | } |
203 | 200 | ||
204 | pfns = xen_do_chunk(dest_pfn, dest_pfn + capacity, false); | 201 | /* |
205 | done += pfns; | 202 | * This releases a chunk of memory and then does the identity map. It's used as |
206 | *last_pfn = (dest_pfn + pfns); | 203 | * as a fallback if the remapping fails. |
207 | if (pfns < capacity) | 204 | */ |
208 | break; | 205 | static void __init xen_set_identity_and_release_chunk(unsigned long start_pfn, |
209 | credits_left -= pfns; | 206 | unsigned long end_pfn, unsigned long nr_pages, unsigned long *identity, |
207 | unsigned long *released) | ||
208 | { | ||
209 | WARN_ON(start_pfn > end_pfn); | ||
210 | |||
211 | /* Need to release pages first */ | ||
212 | *released += xen_do_chunk(start_pfn, min(end_pfn, nr_pages), true); | ||
213 | *identity += set_phys_range_identity(start_pfn, end_pfn); | ||
214 | } | ||
215 | |||
216 | /* | ||
217 | * Helper function to update both the p2m and m2p tables. | ||
218 | */ | ||
219 | static unsigned long __init xen_update_mem_tables(unsigned long pfn, | ||
220 | unsigned long mfn) | ||
221 | { | ||
222 | struct mmu_update update = { | ||
223 | .ptr = ((unsigned long long)mfn << PAGE_SHIFT) | MMU_MACHPHYS_UPDATE, | ||
224 | .val = pfn | ||
225 | }; | ||
226 | |||
227 | /* Update p2m */ | ||
228 | if (!early_set_phys_to_machine(pfn, mfn)) { | ||
229 | WARN(1, "Failed to set p2m mapping for pfn=%ld mfn=%ld\n", | ||
230 | pfn, mfn); | ||
231 | return false; | ||
210 | } | 232 | } |
211 | return done; | 233 | |
234 | /* Update m2p */ | ||
235 | if (HYPERVISOR_mmu_update(&update, 1, NULL, DOMID_SELF) < 0) { | ||
236 | WARN(1, "Failed to set m2p mapping for mfn=%ld pfn=%ld\n", | ||
237 | mfn, pfn); | ||
238 | return false; | ||
239 | } | ||
240 | |||
241 | return true; | ||
212 | } | 242 | } |
213 | 243 | ||
214 | static void __init xen_set_identity_and_release_chunk( | 244 | /* |
215 | unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, | 245 | * This function updates the p2m and m2p tables with an identity map from |
216 | unsigned long *released, unsigned long *identity) | 246 | * start_pfn to start_pfn+size and remaps the underlying RAM of the original |
247 | * allocation at remap_pfn. It must do so carefully in P2M_PER_PAGE sized blocks | ||
248 | * to not exhaust the reserved brk space. Doing it in properly aligned blocks | ||
249 | * ensures we only allocate the minimum required leaf pages in the p2m table. It | ||
250 | * copies the existing mfns from the p2m table under the 1:1 map, overwrites | ||
251 | * them with the identity map and then updates the p2m and m2p tables with the | ||
252 | * remapped memory. | ||
253 | */ | ||
254 | static unsigned long __init xen_do_set_identity_and_remap_chunk( | ||
255 | unsigned long start_pfn, unsigned long size, unsigned long remap_pfn) | ||
217 | { | 256 | { |
218 | unsigned long pfn; | 257 | unsigned long ident_pfn_iter, remap_pfn_iter; |
258 | unsigned long ident_start_pfn_align, remap_start_pfn_align; | ||
259 | unsigned long ident_end_pfn_align, remap_end_pfn_align; | ||
260 | unsigned long ident_boundary_pfn, remap_boundary_pfn; | ||
261 | unsigned long ident_cnt = 0; | ||
262 | unsigned long remap_cnt = 0; | ||
263 | unsigned long left = size; | ||
264 | unsigned long mod; | ||
265 | int i; | ||
266 | |||
267 | WARN_ON(size == 0); | ||
268 | |||
269 | BUG_ON(xen_feature(XENFEAT_auto_translated_physmap)); | ||
219 | 270 | ||
220 | /* | 271 | /* |
221 | * If the PFNs are currently mapped, clear the mappings | 272 | * Determine the proper alignment to remap memory in P2M_PER_PAGE sized |
222 | * (except for the ISA region which must be 1:1 mapped) to | 273 | * blocks. We need to keep track of both the existing pfn mapping and |
223 | * release the refcounts (in Xen) on the original frames. | 274 | * the new pfn remapping. |
224 | */ | 275 | */ |
225 | for (pfn = start_pfn; pfn <= max_pfn_mapped && pfn < end_pfn; pfn++) { | 276 | mod = start_pfn % P2M_PER_PAGE; |
226 | pte_t pte = __pte_ma(0); | 277 | ident_start_pfn_align = |
278 | mod ? (start_pfn - mod + P2M_PER_PAGE) : start_pfn; | ||
279 | mod = remap_pfn % P2M_PER_PAGE; | ||
280 | remap_start_pfn_align = | ||
281 | mod ? (remap_pfn - mod + P2M_PER_PAGE) : remap_pfn; | ||
282 | mod = (start_pfn + size) % P2M_PER_PAGE; | ||
283 | ident_end_pfn_align = start_pfn + size - mod; | ||
284 | mod = (remap_pfn + size) % P2M_PER_PAGE; | ||
285 | remap_end_pfn_align = remap_pfn + size - mod; | ||
286 | |||
287 | /* Iterate over each p2m leaf node in each range */ | ||
288 | for (ident_pfn_iter = ident_start_pfn_align, remap_pfn_iter = remap_start_pfn_align; | ||
289 | ident_pfn_iter < ident_end_pfn_align && remap_pfn_iter < remap_end_pfn_align; | ||
290 | ident_pfn_iter += P2M_PER_PAGE, remap_pfn_iter += P2M_PER_PAGE) { | ||
291 | /* Check we aren't past the end */ | ||
292 | BUG_ON(ident_pfn_iter + P2M_PER_PAGE > start_pfn + size); | ||
293 | BUG_ON(remap_pfn_iter + P2M_PER_PAGE > remap_pfn + size); | ||
294 | |||
295 | /* Save p2m mappings */ | ||
296 | for (i = 0; i < P2M_PER_PAGE; i++) | ||
297 | xen_remap_buf[i] = pfn_to_mfn(ident_pfn_iter + i); | ||
298 | |||
299 | /* Set identity map which will free a p2m leaf */ | ||
300 | ident_cnt += set_phys_range_identity(ident_pfn_iter, | ||
301 | ident_pfn_iter + P2M_PER_PAGE); | ||
302 | |||
303 | #ifdef DEBUG | ||
304 | /* Helps verify a p2m leaf has been freed */ | ||
305 | for (i = 0; i < P2M_PER_PAGE; i++) { | ||
306 | unsigned int pfn = ident_pfn_iter + i; | ||
307 | BUG_ON(pfn_to_mfn(pfn) != pfn); | ||
308 | } | ||
309 | #endif | ||
310 | /* Now remap memory */ | ||
311 | for (i = 0; i < P2M_PER_PAGE; i++) { | ||
312 | unsigned long mfn = xen_remap_buf[i]; | ||
313 | |||
314 | /* This will use the p2m leaf freed above */ | ||
315 | if (!xen_update_mem_tables(remap_pfn_iter + i, mfn)) { | ||
316 | WARN(1, "Failed to update mem mapping for pfn=%ld mfn=%ld\n", | ||
317 | remap_pfn_iter + i, mfn); | ||
318 | return 0; | ||
319 | } | ||
320 | |||
321 | remap_cnt++; | ||
322 | } | ||
227 | 323 | ||
228 | if (pfn < PFN_UP(ISA_END_ADDRESS)) | 324 | left -= P2M_PER_PAGE; |
229 | pte = mfn_pte(pfn, PAGE_KERNEL_IO); | 325 | } |
230 | 326 | ||
231 | (void)HYPERVISOR_update_va_mapping( | 327 | /* Max boundary space possible */ |
232 | (unsigned long)__va(pfn << PAGE_SHIFT), pte, 0); | 328 | BUG_ON(left > (P2M_PER_PAGE - 1) * 2); |
329 | |||
330 | /* Now handle the boundary conditions */ | ||
331 | ident_boundary_pfn = start_pfn; | ||
332 | remap_boundary_pfn = remap_pfn; | ||
333 | for (i = 0; i < left; i++) { | ||
334 | unsigned long mfn; | ||
335 | |||
336 | /* These two checks move from the start to end boundaries */ | ||
337 | if (ident_boundary_pfn == ident_start_pfn_align) | ||
338 | ident_boundary_pfn = ident_pfn_iter; | ||
339 | if (remap_boundary_pfn == remap_start_pfn_align) | ||
340 | remap_boundary_pfn = remap_pfn_iter; | ||
341 | |||
342 | /* Check we aren't past the end */ | ||
343 | BUG_ON(ident_boundary_pfn >= start_pfn + size); | ||
344 | BUG_ON(remap_boundary_pfn >= remap_pfn + size); | ||
345 | |||
346 | mfn = pfn_to_mfn(ident_boundary_pfn); | ||
347 | |||
348 | if (!xen_update_mem_tables(remap_boundary_pfn, mfn)) { | ||
349 | WARN(1, "Failed to update mem mapping for pfn=%ld mfn=%ld\n", | ||
350 | remap_pfn_iter + i, mfn); | ||
351 | return 0; | ||
352 | } | ||
353 | remap_cnt++; | ||
354 | |||
355 | ident_boundary_pfn++; | ||
356 | remap_boundary_pfn++; | ||
233 | } | 357 | } |
234 | 358 | ||
235 | if (start_pfn < nr_pages) | 359 | /* Finish up the identity map */ |
236 | *released += xen_release_chunk( | 360 | if (ident_start_pfn_align >= ident_end_pfn_align) { |
237 | start_pfn, min(end_pfn, nr_pages)); | 361 | /* |
362 | * In this case we have an identity range which does not span an | ||
363 | * aligned block so everything needs to be identity mapped here. | ||
364 | * If we didn't check this we might remap too many pages since | ||
365 | * the align boundaries are not meaningful in this case. | ||
366 | */ | ||
367 | ident_cnt += set_phys_range_identity(start_pfn, | ||
368 | start_pfn + size); | ||
369 | } else { | ||
370 | /* Remapped above so check each end of the chunk */ | ||
371 | if (start_pfn < ident_start_pfn_align) | ||
372 | ident_cnt += set_phys_range_identity(start_pfn, | ||
373 | ident_start_pfn_align); | ||
374 | if (start_pfn + size > ident_pfn_iter) | ||
375 | ident_cnt += set_phys_range_identity(ident_pfn_iter, | ||
376 | start_pfn + size); | ||
377 | } | ||
238 | 378 | ||
239 | *identity += set_phys_range_identity(start_pfn, end_pfn); | 379 | BUG_ON(ident_cnt != size); |
380 | BUG_ON(remap_cnt != size); | ||
381 | |||
382 | return size; | ||
240 | } | 383 | } |
241 | 384 | ||
242 | static unsigned long __init xen_set_identity_and_release( | 385 | /* |
243 | const struct e820entry *list, size_t map_size, unsigned long nr_pages) | 386 | * This function takes a contiguous pfn range that needs to be identity mapped |
387 | * and: | ||
388 | * | ||
389 | * 1) Finds a new range of pfns to use to remap based on E820 and remap_pfn. | ||
390 | * 2) Calls the do_ function to actually do the mapping/remapping work. | ||
391 | * | ||
392 | * The goal is to not allocate additional memory but to remap the existing | ||
393 | * pages. In the case of an error the underlying memory is simply released back | ||
394 | * to Xen and not remapped. | ||
395 | */ | ||
396 | static unsigned long __init xen_set_identity_and_remap_chunk( | ||
397 | const struct e820entry *list, size_t map_size, unsigned long start_pfn, | ||
398 | unsigned long end_pfn, unsigned long nr_pages, unsigned long remap_pfn, | ||
399 | unsigned long *identity, unsigned long *remapped, | ||
400 | unsigned long *released) | ||
401 | { | ||
402 | unsigned long pfn; | ||
403 | unsigned long i = 0; | ||
404 | unsigned long n = end_pfn - start_pfn; | ||
405 | |||
406 | while (i < n) { | ||
407 | unsigned long cur_pfn = start_pfn + i; | ||
408 | unsigned long left = n - i; | ||
409 | unsigned long size = left; | ||
410 | unsigned long remap_range_size; | ||
411 | |||
412 | /* Do not remap pages beyond the current allocation */ | ||
413 | if (cur_pfn >= nr_pages) { | ||
414 | /* Identity map remaining pages */ | ||
415 | *identity += set_phys_range_identity(cur_pfn, | ||
416 | cur_pfn + size); | ||
417 | break; | ||
418 | } | ||
419 | if (cur_pfn + size > nr_pages) | ||
420 | size = nr_pages - cur_pfn; | ||
421 | |||
422 | remap_range_size = xen_find_pfn_range(list, map_size, | ||
423 | &remap_pfn); | ||
424 | if (!remap_range_size) { | ||
425 | pr_warning("Unable to find available pfn range, not remapping identity pages\n"); | ||
426 | xen_set_identity_and_release_chunk(cur_pfn, | ||
427 | cur_pfn + left, nr_pages, identity, released); | ||
428 | break; | ||
429 | } | ||
430 | /* Adjust size to fit in current e820 RAM region */ | ||
431 | if (size > remap_range_size) | ||
432 | size = remap_range_size; | ||
433 | |||
434 | if (!xen_do_set_identity_and_remap_chunk(cur_pfn, size, remap_pfn)) { | ||
435 | WARN(1, "Failed to remap 1:1 memory cur_pfn=%ld size=%ld remap_pfn=%ld\n", | ||
436 | cur_pfn, size, remap_pfn); | ||
437 | xen_set_identity_and_release_chunk(cur_pfn, | ||
438 | cur_pfn + left, nr_pages, identity, released); | ||
439 | break; | ||
440 | } | ||
441 | |||
442 | /* Update variables to reflect new mappings. */ | ||
443 | i += size; | ||
444 | remap_pfn += size; | ||
445 | *identity += size; | ||
446 | *remapped += size; | ||
447 | } | ||
448 | |||
449 | /* | ||
450 | * If the PFNs are currently mapped, the VA mapping also needs | ||
451 | * to be updated to be 1:1. | ||
452 | */ | ||
453 | for (pfn = start_pfn; pfn <= max_pfn_mapped && pfn < end_pfn; pfn++) | ||
454 | (void)HYPERVISOR_update_va_mapping( | ||
455 | (unsigned long)__va(pfn << PAGE_SHIFT), | ||
456 | mfn_pte(pfn, PAGE_KERNEL_IO), 0); | ||
457 | |||
458 | return remap_pfn; | ||
459 | } | ||
460 | |||
461 | static unsigned long __init xen_set_identity_and_remap( | ||
462 | const struct e820entry *list, size_t map_size, unsigned long nr_pages, | ||
463 | unsigned long *released) | ||
244 | { | 464 | { |
245 | phys_addr_t start = 0; | 465 | phys_addr_t start = 0; |
246 | unsigned long released = 0; | ||
247 | unsigned long identity = 0; | 466 | unsigned long identity = 0; |
467 | unsigned long remapped = 0; | ||
468 | unsigned long last_pfn = nr_pages; | ||
248 | const struct e820entry *entry; | 469 | const struct e820entry *entry; |
470 | unsigned long num_released = 0; | ||
249 | int i; | 471 | int i; |
250 | 472 | ||
251 | /* | 473 | /* |
252 | * Combine non-RAM regions and gaps until a RAM region (or the | 474 | * Combine non-RAM regions and gaps until a RAM region (or the |
253 | * end of the map) is reached, then set the 1:1 map and | 475 | * end of the map) is reached, then set the 1:1 map and |
254 | * release the pages (if available) in those non-RAM regions. | 476 | * remap the memory in those non-RAM regions. |
255 | * | 477 | * |
256 | * The combined non-RAM regions are rounded to a whole number | 478 | * The combined non-RAM regions are rounded to a whole number |
257 | * of pages so any partial pages are accessible via the 1:1 | 479 | * of pages so any partial pages are accessible via the 1:1 |
@@ -269,22 +491,24 @@ static unsigned long __init xen_set_identity_and_release( | |||
269 | end_pfn = PFN_UP(entry->addr); | 491 | end_pfn = PFN_UP(entry->addr); |
270 | 492 | ||
271 | if (start_pfn < end_pfn) | 493 | if (start_pfn < end_pfn) |
272 | xen_set_identity_and_release_chunk( | 494 | last_pfn = xen_set_identity_and_remap_chunk( |
273 | start_pfn, end_pfn, nr_pages, | 495 | list, map_size, start_pfn, |
274 | &released, &identity); | 496 | end_pfn, nr_pages, last_pfn, |
275 | 497 | &identity, &remapped, | |
498 | &num_released); | ||
276 | start = end; | 499 | start = end; |
277 | } | 500 | } |
278 | } | 501 | } |
279 | 502 | ||
280 | if (released) | 503 | *released = num_released; |
281 | printk(KERN_INFO "Released %lu pages of unused memory\n", released); | ||
282 | if (identity) | ||
283 | printk(KERN_INFO "Set %ld page(s) to 1-1 mapping\n", identity); | ||
284 | 504 | ||
285 | return released; | 505 | pr_info("Set %ld page(s) to 1-1 mapping\n", identity); |
286 | } | 506 | pr_info("Remapped %ld page(s), last_pfn=%ld\n", remapped, |
507 | last_pfn); | ||
508 | pr_info("Released %ld page(s)\n", num_released); | ||
287 | 509 | ||
510 | return last_pfn; | ||
511 | } | ||
288 | static unsigned long __init xen_get_max_pages(void) | 512 | static unsigned long __init xen_get_max_pages(void) |
289 | { | 513 | { |
290 | unsigned long max_pages = MAX_DOMAIN_PAGES; | 514 | unsigned long max_pages = MAX_DOMAIN_PAGES; |
@@ -347,7 +571,6 @@ char * __init xen_memory_setup(void) | |||
347 | unsigned long max_pages; | 571 | unsigned long max_pages; |
348 | unsigned long last_pfn = 0; | 572 | unsigned long last_pfn = 0; |
349 | unsigned long extra_pages = 0; | 573 | unsigned long extra_pages = 0; |
350 | unsigned long populated; | ||
351 | int i; | 574 | int i; |
352 | int op; | 575 | int op; |
353 | 576 | ||
@@ -392,20 +615,11 @@ char * __init xen_memory_setup(void) | |||
392 | extra_pages += max_pages - max_pfn; | 615 | extra_pages += max_pages - max_pfn; |
393 | 616 | ||
394 | /* | 617 | /* |
395 | * Set P2M for all non-RAM pages and E820 gaps to be identity | 618 | * Set identity map on non-RAM pages and remap the underlying RAM. |
396 | * type PFNs. Any RAM pages that would be made inaccesible by | ||
397 | * this are first released. | ||
398 | */ | 619 | */ |
399 | xen_released_pages = xen_set_identity_and_release( | 620 | last_pfn = xen_set_identity_and_remap(map, memmap.nr_entries, max_pfn, |
400 | map, memmap.nr_entries, max_pfn); | 621 | &xen_released_pages); |
401 | |||
402 | /* | ||
403 | * Populate back the non-RAM pages and E820 gaps that had been | ||
404 | * released. */ | ||
405 | populated = xen_populate_chunk(map, memmap.nr_entries, | ||
406 | max_pfn, &last_pfn, xen_released_pages); | ||
407 | 622 | ||
408 | xen_released_pages -= populated; | ||
409 | extra_pages += xen_released_pages; | 623 | extra_pages += xen_released_pages; |
410 | 624 | ||
411 | if (last_pfn > max_pfn) { | 625 | if (last_pfn > max_pfn) { |
diff --git a/arch/x86/xen/smp.c b/arch/x86/xen/smp.c index 7005974c3ff3..c670d7518cf4 100644 --- a/arch/x86/xen/smp.c +++ b/arch/x86/xen/smp.c | |||
@@ -37,6 +37,7 @@ | |||
37 | #include <xen/hvc-console.h> | 37 | #include <xen/hvc-console.h> |
38 | #include "xen-ops.h" | 38 | #include "xen-ops.h" |
39 | #include "mmu.h" | 39 | #include "mmu.h" |
40 | #include "smp.h" | ||
40 | 41 | ||
41 | cpumask_var_t xen_cpu_initialized_map; | 42 | cpumask_var_t xen_cpu_initialized_map; |
42 | 43 | ||
@@ -99,10 +100,14 @@ static void cpu_bringup(void) | |||
99 | wmb(); /* make sure everything is out */ | 100 | wmb(); /* make sure everything is out */ |
100 | } | 101 | } |
101 | 102 | ||
102 | /* Note: cpu parameter is only relevant for PVH */ | 103 | /* |
103 | static void cpu_bringup_and_idle(int cpu) | 104 | * Note: cpu parameter is only relevant for PVH. The reason for passing it |
105 | * is we can't do smp_processor_id until the percpu segments are loaded, for | ||
106 | * which we need the cpu number! So we pass it in rdi as first parameter. | ||
107 | */ | ||
108 | asmlinkage __visible void cpu_bringup_and_idle(int cpu) | ||
104 | { | 109 | { |
105 | #ifdef CONFIG_X86_64 | 110 | #ifdef CONFIG_XEN_PVH |
106 | if (xen_feature(XENFEAT_auto_translated_physmap) && | 111 | if (xen_feature(XENFEAT_auto_translated_physmap) && |
107 | xen_feature(XENFEAT_supervisor_mode_kernel)) | 112 | xen_feature(XENFEAT_supervisor_mode_kernel)) |
108 | xen_pvh_secondary_vcpu_init(cpu); | 113 | xen_pvh_secondary_vcpu_init(cpu); |
@@ -374,11 +379,10 @@ cpu_initialize_context(unsigned int cpu, struct task_struct *idle) | |||
374 | ctxt->user_regs.fs = __KERNEL_PERCPU; | 379 | ctxt->user_regs.fs = __KERNEL_PERCPU; |
375 | ctxt->user_regs.gs = __KERNEL_STACK_CANARY; | 380 | ctxt->user_regs.gs = __KERNEL_STACK_CANARY; |
376 | #endif | 381 | #endif |
377 | ctxt->user_regs.eip = (unsigned long)cpu_bringup_and_idle; | ||
378 | |||
379 | memset(&ctxt->fpu_ctxt, 0, sizeof(ctxt->fpu_ctxt)); | 382 | memset(&ctxt->fpu_ctxt, 0, sizeof(ctxt->fpu_ctxt)); |
380 | 383 | ||
381 | if (!xen_feature(XENFEAT_auto_translated_physmap)) { | 384 | if (!xen_feature(XENFEAT_auto_translated_physmap)) { |
385 | ctxt->user_regs.eip = (unsigned long)cpu_bringup_and_idle; | ||
382 | ctxt->flags = VGCF_IN_KERNEL; | 386 | ctxt->flags = VGCF_IN_KERNEL; |
383 | ctxt->user_regs.eflags = 0x1000; /* IOPL_RING1 */ | 387 | ctxt->user_regs.eflags = 0x1000; /* IOPL_RING1 */ |
384 | ctxt->user_regs.ds = __USER_DS; | 388 | ctxt->user_regs.ds = __USER_DS; |
@@ -413,15 +417,18 @@ cpu_initialize_context(unsigned int cpu, struct task_struct *idle) | |||
413 | (unsigned long)xen_failsafe_callback; | 417 | (unsigned long)xen_failsafe_callback; |
414 | ctxt->user_regs.cs = __KERNEL_CS; | 418 | ctxt->user_regs.cs = __KERNEL_CS; |
415 | per_cpu(xen_cr3, cpu) = __pa(swapper_pg_dir); | 419 | per_cpu(xen_cr3, cpu) = __pa(swapper_pg_dir); |
416 | #ifdef CONFIG_X86_32 | ||
417 | } | 420 | } |
418 | #else | 421 | #ifdef CONFIG_XEN_PVH |
419 | } else | 422 | else { |
420 | /* N.B. The user_regs.eip (cpu_bringup_and_idle) is called with | 423 | /* |
421 | * %rdi having the cpu number - which means are passing in | 424 | * The vcpu comes on kernel page tables which have the NX pte |
422 | * as the first parameter the cpu. Subtle! | 425 | * bit set. This means before DS/SS is touched, NX in |
426 | * EFER must be set. Hence the following assembly glue code. | ||
423 | */ | 427 | */ |
428 | ctxt->user_regs.eip = (unsigned long)xen_pvh_early_cpu_init; | ||
424 | ctxt->user_regs.rdi = cpu; | 429 | ctxt->user_regs.rdi = cpu; |
430 | ctxt->user_regs.rsi = true; /* entry == true */ | ||
431 | } | ||
425 | #endif | 432 | #endif |
426 | ctxt->user_regs.esp = idle->thread.sp0 - sizeof(struct pt_regs); | 433 | ctxt->user_regs.esp = idle->thread.sp0 - sizeof(struct pt_regs); |
427 | ctxt->ctrlreg[3] = xen_pfn_to_cr3(virt_to_mfn(swapper_pg_dir)); | 434 | ctxt->ctrlreg[3] = xen_pfn_to_cr3(virt_to_mfn(swapper_pg_dir)); |
diff --git a/arch/x86/xen/smp.h b/arch/x86/xen/smp.h index c7c2d89efd76..963d62a35c82 100644 --- a/arch/x86/xen/smp.h +++ b/arch/x86/xen/smp.h | |||
@@ -8,4 +8,12 @@ extern void xen_send_IPI_allbutself(int vector); | |||
8 | extern void xen_send_IPI_all(int vector); | 8 | extern void xen_send_IPI_all(int vector); |
9 | extern void xen_send_IPI_self(int vector); | 9 | extern void xen_send_IPI_self(int vector); |
10 | 10 | ||
11 | #ifdef CONFIG_XEN_PVH | ||
12 | extern void xen_pvh_early_cpu_init(int cpu, bool entry); | ||
13 | #else | ||
14 | static inline void xen_pvh_early_cpu_init(int cpu, bool entry) | ||
15 | { | ||
16 | } | ||
17 | #endif | ||
18 | |||
11 | #endif | 19 | #endif |
diff --git a/arch/x86/xen/xen-head.S b/arch/x86/xen/xen-head.S index 485b69585540..674b222544b7 100644 --- a/arch/x86/xen/xen-head.S +++ b/arch/x86/xen/xen-head.S | |||
@@ -47,6 +47,41 @@ ENTRY(startup_xen) | |||
47 | 47 | ||
48 | __FINIT | 48 | __FINIT |
49 | 49 | ||
50 | #ifdef CONFIG_XEN_PVH | ||
51 | /* | ||
52 | * xen_pvh_early_cpu_init() - early PVH VCPU initialization | ||
53 | * @cpu: this cpu number (%rdi) | ||
54 | * @entry: true if this is a secondary vcpu coming up on this entry | ||
55 | * point, false if this is the boot CPU being initialized for | ||
56 | * the first time (%rsi) | ||
57 | * | ||
58 | * Note: This is called as a function on the boot CPU, and is the entry point | ||
59 | * on the secondary CPU. | ||
60 | */ | ||
61 | ENTRY(xen_pvh_early_cpu_init) | ||
62 | mov %rsi, %r11 | ||
63 | |||
64 | /* Gather features to see if NX implemented. */ | ||
65 | mov $0x80000001, %eax | ||
66 | cpuid | ||
67 | mov %edx, %esi | ||
68 | |||
69 | mov $MSR_EFER, %ecx | ||
70 | rdmsr | ||
71 | bts $_EFER_SCE, %eax | ||
72 | |||
73 | bt $20, %esi | ||
74 | jnc 1f /* No NX, skip setting it */ | ||
75 | bts $_EFER_NX, %eax | ||
76 | 1: wrmsr | ||
77 | #ifdef CONFIG_SMP | ||
78 | cmp $0, %r11b | ||
79 | jne cpu_bringup_and_idle | ||
80 | #endif | ||
81 | ret | ||
82 | |||
83 | #endif /* CONFIG_XEN_PVH */ | ||
84 | |||
50 | .pushsection .text | 85 | .pushsection .text |
51 | .balign PAGE_SIZE | 86 | .balign PAGE_SIZE |
52 | ENTRY(hypercall_page) | 87 | ENTRY(hypercall_page) |
@@ -124,6 +159,7 @@ NEXT_HYPERCALL(arch_6) | |||
124 | ELFNOTE(Xen, XEN_ELFNOTE_L1_MFN_VALID, | 159 | ELFNOTE(Xen, XEN_ELFNOTE_L1_MFN_VALID, |
125 | .quad _PAGE_PRESENT; .quad _PAGE_PRESENT) | 160 | .quad _PAGE_PRESENT; .quad _PAGE_PRESENT) |
126 | ELFNOTE(Xen, XEN_ELFNOTE_SUSPEND_CANCEL, .long 1) | 161 | ELFNOTE(Xen, XEN_ELFNOTE_SUSPEND_CANCEL, .long 1) |
162 | ELFNOTE(Xen, XEN_ELFNOTE_MOD_START_PFN, .long 1) | ||
127 | ELFNOTE(Xen, XEN_ELFNOTE_HV_START_LOW, _ASM_PTR __HYPERVISOR_VIRT_START) | 163 | ELFNOTE(Xen, XEN_ELFNOTE_HV_START_LOW, _ASM_PTR __HYPERVISOR_VIRT_START) |
128 | ELFNOTE(Xen, XEN_ELFNOTE_PADDR_OFFSET, _ASM_PTR 0) | 164 | ELFNOTE(Xen, XEN_ELFNOTE_PADDR_OFFSET, _ASM_PTR 0) |
129 | 165 | ||
diff --git a/drivers/block/xen-blkback/xenbus.c b/drivers/block/xen-blkback/xenbus.c index 3a8b810b4980..0b13b1c9a01e 100644 --- a/drivers/block/xen-blkback/xenbus.c +++ b/drivers/block/xen-blkback/xenbus.c | |||
@@ -907,22 +907,17 @@ static int connect_ring(struct backend_info *be) | |||
907 | return 0; | 907 | return 0; |
908 | } | 908 | } |
909 | 909 | ||
910 | |||
911 | /* ** Driver Registration ** */ | ||
912 | |||
913 | |||
914 | static const struct xenbus_device_id xen_blkbk_ids[] = { | 910 | static const struct xenbus_device_id xen_blkbk_ids[] = { |
915 | { "vbd" }, | 911 | { "vbd" }, |
916 | { "" } | 912 | { "" } |
917 | }; | 913 | }; |
918 | 914 | ||
919 | 915 | static struct xenbus_driver xen_blkbk_driver = { | |
920 | static DEFINE_XENBUS_DRIVER(xen_blkbk, , | 916 | .ids = xen_blkbk_ids, |
921 | .probe = xen_blkbk_probe, | 917 | .probe = xen_blkbk_probe, |
922 | .remove = xen_blkbk_remove, | 918 | .remove = xen_blkbk_remove, |
923 | .otherend_changed = frontend_changed | 919 | .otherend_changed = frontend_changed |
924 | ); | 920 | }; |
925 | |||
926 | 921 | ||
927 | int xen_blkif_xenbus_init(void) | 922 | int xen_blkif_xenbus_init(void) |
928 | { | 923 | { |
diff --git a/drivers/block/xen-blkfront.c b/drivers/block/xen-blkfront.c index 5deb235bd18f..37af03e9d859 100644 --- a/drivers/block/xen-blkfront.c +++ b/drivers/block/xen-blkfront.c | |||
@@ -2055,13 +2055,14 @@ static const struct xenbus_device_id blkfront_ids[] = { | |||
2055 | { "" } | 2055 | { "" } |
2056 | }; | 2056 | }; |
2057 | 2057 | ||
2058 | static DEFINE_XENBUS_DRIVER(blkfront, , | 2058 | static struct xenbus_driver blkfront_driver = { |
2059 | .ids = blkfront_ids, | ||
2059 | .probe = blkfront_probe, | 2060 | .probe = blkfront_probe, |
2060 | .remove = blkfront_remove, | 2061 | .remove = blkfront_remove, |
2061 | .resume = blkfront_resume, | 2062 | .resume = blkfront_resume, |
2062 | .otherend_changed = blkback_changed, | 2063 | .otherend_changed = blkback_changed, |
2063 | .is_ready = blkfront_is_ready, | 2064 | .is_ready = blkfront_is_ready, |
2064 | ); | 2065 | }; |
2065 | 2066 | ||
2066 | static int __init xlblk_init(void) | 2067 | static int __init xlblk_init(void) |
2067 | { | 2068 | { |
diff --git a/drivers/char/tpm/xen-tpmfront.c b/drivers/char/tpm/xen-tpmfront.c index 2064b4527040..441b44e54226 100644 --- a/drivers/char/tpm/xen-tpmfront.c +++ b/drivers/char/tpm/xen-tpmfront.c | |||
@@ -367,12 +367,13 @@ static const struct xenbus_device_id tpmfront_ids[] = { | |||
367 | }; | 367 | }; |
368 | MODULE_ALIAS("xen:vtpm"); | 368 | MODULE_ALIAS("xen:vtpm"); |
369 | 369 | ||
370 | static DEFINE_XENBUS_DRIVER(tpmfront, , | 370 | static struct xenbus_driver tpmfront_driver = { |
371 | .probe = tpmfront_probe, | 371 | .ids = tpmfront_ids, |
372 | .remove = tpmfront_remove, | 372 | .probe = tpmfront_probe, |
373 | .resume = tpmfront_resume, | 373 | .remove = tpmfront_remove, |
374 | .otherend_changed = backend_changed, | 374 | .resume = tpmfront_resume, |
375 | ); | 375 | .otherend_changed = backend_changed, |
376 | }; | ||
376 | 377 | ||
377 | static int __init xen_tpmfront_init(void) | 378 | static int __init xen_tpmfront_init(void) |
378 | { | 379 | { |
diff --git a/drivers/input/misc/xen-kbdfront.c b/drivers/input/misc/xen-kbdfront.c index fbfdc10573be..1af28b06c713 100644 --- a/drivers/input/misc/xen-kbdfront.c +++ b/drivers/input/misc/xen-kbdfront.c | |||
@@ -365,12 +365,13 @@ static const struct xenbus_device_id xenkbd_ids[] = { | |||
365 | { "" } | 365 | { "" } |
366 | }; | 366 | }; |
367 | 367 | ||
368 | static DEFINE_XENBUS_DRIVER(xenkbd, , | 368 | static struct xenbus_driver xenkbd_driver = { |
369 | .ids = xenkbd_ids, | ||
369 | .probe = xenkbd_probe, | 370 | .probe = xenkbd_probe, |
370 | .remove = xenkbd_remove, | 371 | .remove = xenkbd_remove, |
371 | .resume = xenkbd_resume, | 372 | .resume = xenkbd_resume, |
372 | .otherend_changed = xenkbd_backend_changed, | 373 | .otherend_changed = xenkbd_backend_changed, |
373 | ); | 374 | }; |
374 | 375 | ||
375 | static int __init xenkbd_init(void) | 376 | static int __init xenkbd_init(void) |
376 | { | 377 | { |
diff --git a/drivers/net/xen-netback/xenbus.c b/drivers/net/xen-netback/xenbus.c index 9c47b897b6d2..8079c31ac5e6 100644 --- a/drivers/net/xen-netback/xenbus.c +++ b/drivers/net/xen-netback/xenbus.c | |||
@@ -937,22 +937,18 @@ static int read_xenbus_vif_flags(struct backend_info *be) | |||
937 | return 0; | 937 | return 0; |
938 | } | 938 | } |
939 | 939 | ||
940 | |||
941 | /* ** Driver Registration ** */ | ||
942 | |||
943 | |||
944 | static const struct xenbus_device_id netback_ids[] = { | 940 | static const struct xenbus_device_id netback_ids[] = { |
945 | { "vif" }, | 941 | { "vif" }, |
946 | { "" } | 942 | { "" } |
947 | }; | 943 | }; |
948 | 944 | ||
949 | 945 | static struct xenbus_driver netback_driver = { | |
950 | static DEFINE_XENBUS_DRIVER(netback, , | 946 | .ids = netback_ids, |
951 | .probe = netback_probe, | 947 | .probe = netback_probe, |
952 | .remove = netback_remove, | 948 | .remove = netback_remove, |
953 | .uevent = netback_uevent, | 949 | .uevent = netback_uevent, |
954 | .otherend_changed = frontend_changed, | 950 | .otherend_changed = frontend_changed, |
955 | ); | 951 | }; |
956 | 952 | ||
957 | int xenvif_xenbus_init(void) | 953 | int xenvif_xenbus_init(void) |
958 | { | 954 | { |
diff --git a/drivers/net/xen-netfront.c b/drivers/net/xen-netfront.c index ca82f545ec2c..fa671442f420 100644 --- a/drivers/net/xen-netfront.c +++ b/drivers/net/xen-netfront.c | |||
@@ -2300,12 +2300,6 @@ static void xennet_sysfs_delif(struct net_device *netdev) | |||
2300 | 2300 | ||
2301 | #endif /* CONFIG_SYSFS */ | 2301 | #endif /* CONFIG_SYSFS */ |
2302 | 2302 | ||
2303 | static const struct xenbus_device_id netfront_ids[] = { | ||
2304 | { "vif" }, | ||
2305 | { "" } | ||
2306 | }; | ||
2307 | |||
2308 | |||
2309 | static int xennet_remove(struct xenbus_device *dev) | 2303 | static int xennet_remove(struct xenbus_device *dev) |
2310 | { | 2304 | { |
2311 | struct netfront_info *info = dev_get_drvdata(&dev->dev); | 2305 | struct netfront_info *info = dev_get_drvdata(&dev->dev); |
@@ -2338,12 +2332,18 @@ static int xennet_remove(struct xenbus_device *dev) | |||
2338 | return 0; | 2332 | return 0; |
2339 | } | 2333 | } |
2340 | 2334 | ||
2341 | static DEFINE_XENBUS_DRIVER(netfront, , | 2335 | static const struct xenbus_device_id netfront_ids[] = { |
2336 | { "vif" }, | ||
2337 | { "" } | ||
2338 | }; | ||
2339 | |||
2340 | static struct xenbus_driver netfront_driver = { | ||
2341 | .ids = netfront_ids, | ||
2342 | .probe = netfront_probe, | 2342 | .probe = netfront_probe, |
2343 | .remove = xennet_remove, | 2343 | .remove = xennet_remove, |
2344 | .resume = netfront_resume, | 2344 | .resume = netfront_resume, |
2345 | .otherend_changed = netback_changed, | 2345 | .otherend_changed = netback_changed, |
2346 | ); | 2346 | }; |
2347 | 2347 | ||
2348 | static int __init netif_init(void) | 2348 | static int __init netif_init(void) |
2349 | { | 2349 | { |
diff --git a/drivers/pci/xen-pcifront.c b/drivers/pci/xen-pcifront.c index 53df39a22c8a..116ca3746adb 100644 --- a/drivers/pci/xen-pcifront.c +++ b/drivers/pci/xen-pcifront.c | |||
@@ -1136,11 +1136,13 @@ static const struct xenbus_device_id xenpci_ids[] = { | |||
1136 | {""}, | 1136 | {""}, |
1137 | }; | 1137 | }; |
1138 | 1138 | ||
1139 | static DEFINE_XENBUS_DRIVER(xenpci, "pcifront", | 1139 | static struct xenbus_driver xenpci_driver = { |
1140 | .name = "pcifront", | ||
1141 | .ids = xenpci_ids, | ||
1140 | .probe = pcifront_xenbus_probe, | 1142 | .probe = pcifront_xenbus_probe, |
1141 | .remove = pcifront_xenbus_remove, | 1143 | .remove = pcifront_xenbus_remove, |
1142 | .otherend_changed = pcifront_backend_changed, | 1144 | .otherend_changed = pcifront_backend_changed, |
1143 | ); | 1145 | }; |
1144 | 1146 | ||
1145 | static int __init pcifront_init(void) | 1147 | static int __init pcifront_init(void) |
1146 | { | 1148 | { |
diff --git a/drivers/scsi/Kconfig b/drivers/scsi/Kconfig index e85e64a07d02..296619b7426c 100644 --- a/drivers/scsi/Kconfig +++ b/drivers/scsi/Kconfig | |||
@@ -587,6 +587,16 @@ config VMWARE_PVSCSI | |||
587 | To compile this driver as a module, choose M here: the | 587 | To compile this driver as a module, choose M here: the |
588 | module will be called vmw_pvscsi. | 588 | module will be called vmw_pvscsi. |
589 | 589 | ||
590 | config XEN_SCSI_FRONTEND | ||
591 | tristate "XEN SCSI frontend driver" | ||
592 | depends on SCSI && XEN | ||
593 | select XEN_XENBUS_FRONTEND | ||
594 | help | ||
595 | The XEN SCSI frontend driver allows the kernel to access SCSI Devices | ||
596 | within another guest OS (usually Dom0). | ||
597 | Only needed if the kernel is running in a XEN guest and generic | ||
598 | SCSI access to a device is needed. | ||
599 | |||
590 | config HYPERV_STORAGE | 600 | config HYPERV_STORAGE |
591 | tristate "Microsoft Hyper-V virtual storage driver" | 601 | tristate "Microsoft Hyper-V virtual storage driver" |
592 | depends on SCSI && HYPERV | 602 | depends on SCSI && HYPERV |
diff --git a/drivers/scsi/Makefile b/drivers/scsi/Makefile index 5f0d299b0093..59f1ce6df2d6 100644 --- a/drivers/scsi/Makefile +++ b/drivers/scsi/Makefile | |||
@@ -141,6 +141,7 @@ obj-$(CONFIG_SCSI_ESAS2R) += esas2r/ | |||
141 | obj-$(CONFIG_SCSI_PMCRAID) += pmcraid.o | 141 | obj-$(CONFIG_SCSI_PMCRAID) += pmcraid.o |
142 | obj-$(CONFIG_SCSI_VIRTIO) += virtio_scsi.o | 142 | obj-$(CONFIG_SCSI_VIRTIO) += virtio_scsi.o |
143 | obj-$(CONFIG_VMWARE_PVSCSI) += vmw_pvscsi.o | 143 | obj-$(CONFIG_VMWARE_PVSCSI) += vmw_pvscsi.o |
144 | obj-$(CONFIG_XEN_SCSI_FRONTEND) += xen-scsifront.o | ||
144 | obj-$(CONFIG_HYPERV_STORAGE) += hv_storvsc.o | 145 | obj-$(CONFIG_HYPERV_STORAGE) += hv_storvsc.o |
145 | 146 | ||
146 | obj-$(CONFIG_ARM) += arm/ | 147 | obj-$(CONFIG_ARM) += arm/ |
diff --git a/drivers/scsi/xen-scsifront.c b/drivers/scsi/xen-scsifront.c new file mode 100644 index 000000000000..34199d206ba6 --- /dev/null +++ b/drivers/scsi/xen-scsifront.c | |||
@@ -0,0 +1,1026 @@ | |||
1 | /* | ||
2 | * Xen SCSI frontend driver | ||
3 | * | ||
4 | * Copyright (c) 2008, FUJITSU Limited | ||
5 | * | ||
6 | * This program is free software; you can redistribute it and/or | ||
7 | * modify it under the terms of the GNU General Public License version 2 | ||
8 | * as published by the Free Software Foundation; or, when distributed | ||
9 | * separately from the Linux kernel or incorporated into other | ||
10 | * software packages, subject to the following license: | ||
11 | * | ||
12 | * Permission is hereby granted, free of charge, to any person obtaining a copy | ||
13 | * of this source file (the "Software"), to deal in the Software without | ||
14 | * restriction, including without limitation the rights to use, copy, modify, | ||
15 | * merge, publish, distribute, sublicense, and/or sell copies of the Software, | ||
16 | * and to permit persons to whom the Software is furnished to do so, subject to | ||
17 | * the following conditions: | ||
18 | * | ||
19 | * The above copyright notice and this permission notice shall be included in | ||
20 | * all copies or substantial portions of the Software. | ||
21 | * | ||
22 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
23 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
24 | * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE | ||
25 | * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||
26 | * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING | ||
27 | * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS | ||
28 | * IN THE SOFTWARE. | ||
29 | */ | ||
30 | |||
31 | #include <linux/module.h> | ||
32 | #include <linux/kernel.h> | ||
33 | #include <linux/device.h> | ||
34 | #include <linux/wait.h> | ||
35 | #include <linux/interrupt.h> | ||
36 | #include <linux/mutex.h> | ||
37 | #include <linux/spinlock.h> | ||
38 | #include <linux/sched.h> | ||
39 | #include <linux/blkdev.h> | ||
40 | #include <linux/pfn.h> | ||
41 | #include <linux/slab.h> | ||
42 | #include <linux/bitops.h> | ||
43 | |||
44 | #include <scsi/scsi_cmnd.h> | ||
45 | #include <scsi/scsi_device.h> | ||
46 | #include <scsi/scsi.h> | ||
47 | #include <scsi/scsi_host.h> | ||
48 | |||
49 | #include <xen/xen.h> | ||
50 | #include <xen/xenbus.h> | ||
51 | #include <xen/grant_table.h> | ||
52 | #include <xen/events.h> | ||
53 | #include <xen/page.h> | ||
54 | |||
55 | #include <xen/interface/grant_table.h> | ||
56 | #include <xen/interface/io/vscsiif.h> | ||
57 | #include <xen/interface/io/protocols.h> | ||
58 | |||
59 | #include <asm/xen/hypervisor.h> | ||
60 | |||
61 | |||
62 | #define GRANT_INVALID_REF 0 | ||
63 | |||
64 | #define VSCSIFRONT_OP_ADD_LUN 1 | ||
65 | #define VSCSIFRONT_OP_DEL_LUN 2 | ||
66 | |||
67 | /* Tuning point. */ | ||
68 | #define VSCSIIF_DEFAULT_CMD_PER_LUN 10 | ||
69 | #define VSCSIIF_MAX_TARGET 64 | ||
70 | #define VSCSIIF_MAX_LUN 255 | ||
71 | |||
72 | #define VSCSIIF_RING_SIZE __CONST_RING_SIZE(vscsiif, PAGE_SIZE) | ||
73 | #define VSCSIIF_MAX_REQS VSCSIIF_RING_SIZE | ||
74 | |||
75 | #define vscsiif_grants_sg(_sg) (PFN_UP((_sg) * \ | ||
76 | sizeof(struct scsiif_request_segment))) | ||
77 | |||
78 | struct vscsifrnt_shadow { | ||
79 | /* command between backend and frontend */ | ||
80 | unsigned char act; | ||
81 | uint16_t rqid; | ||
82 | |||
83 | unsigned int nr_grants; /* number of grants in gref[] */ | ||
84 | struct scsiif_request_segment *sg; /* scatter/gather elements */ | ||
85 | |||
86 | /* Do reset or abort function. */ | ||
87 | wait_queue_head_t wq_reset; /* reset work queue */ | ||
88 | int wait_reset; /* reset work queue condition */ | ||
89 | int32_t rslt_reset; /* reset response status: */ | ||
90 | /* SUCCESS or FAILED or: */ | ||
91 | #define RSLT_RESET_WAITING 0 | ||
92 | #define RSLT_RESET_ERR -1 | ||
93 | |||
94 | /* Requested struct scsi_cmnd is stored from kernel. */ | ||
95 | struct scsi_cmnd *sc; | ||
96 | int gref[vscsiif_grants_sg(SG_ALL) + SG_ALL]; | ||
97 | }; | ||
98 | |||
99 | struct vscsifrnt_info { | ||
100 | struct xenbus_device *dev; | ||
101 | |||
102 | struct Scsi_Host *host; | ||
103 | int host_active; | ||
104 | |||
105 | unsigned int evtchn; | ||
106 | unsigned int irq; | ||
107 | |||
108 | grant_ref_t ring_ref; | ||
109 | struct vscsiif_front_ring ring; | ||
110 | struct vscsiif_response ring_rsp; | ||
111 | |||
112 | spinlock_t shadow_lock; | ||
113 | DECLARE_BITMAP(shadow_free_bitmap, VSCSIIF_MAX_REQS); | ||
114 | struct vscsifrnt_shadow *shadow[VSCSIIF_MAX_REQS]; | ||
115 | |||
116 | wait_queue_head_t wq_sync; | ||
117 | unsigned int wait_ring_available:1; | ||
118 | |||
119 | char dev_state_path[64]; | ||
120 | struct task_struct *curr; | ||
121 | }; | ||
122 | |||
123 | static DEFINE_MUTEX(scsifront_mutex); | ||
124 | |||
125 | static void scsifront_wake_up(struct vscsifrnt_info *info) | ||
126 | { | ||
127 | info->wait_ring_available = 0; | ||
128 | wake_up(&info->wq_sync); | ||
129 | } | ||
130 | |||
131 | static int scsifront_get_rqid(struct vscsifrnt_info *info) | ||
132 | { | ||
133 | unsigned long flags; | ||
134 | int free; | ||
135 | |||
136 | spin_lock_irqsave(&info->shadow_lock, flags); | ||
137 | |||
138 | free = find_first_bit(info->shadow_free_bitmap, VSCSIIF_MAX_REQS); | ||
139 | __clear_bit(free, info->shadow_free_bitmap); | ||
140 | |||
141 | spin_unlock_irqrestore(&info->shadow_lock, flags); | ||
142 | |||
143 | return free; | ||
144 | } | ||
145 | |||
146 | static int _scsifront_put_rqid(struct vscsifrnt_info *info, uint32_t id) | ||
147 | { | ||
148 | int empty = bitmap_empty(info->shadow_free_bitmap, VSCSIIF_MAX_REQS); | ||
149 | |||
150 | __set_bit(id, info->shadow_free_bitmap); | ||
151 | info->shadow[id] = NULL; | ||
152 | |||
153 | return empty || info->wait_ring_available; | ||
154 | } | ||
155 | |||
156 | static void scsifront_put_rqid(struct vscsifrnt_info *info, uint32_t id) | ||
157 | { | ||
158 | unsigned long flags; | ||
159 | int kick; | ||
160 | |||
161 | spin_lock_irqsave(&info->shadow_lock, flags); | ||
162 | kick = _scsifront_put_rqid(info, id); | ||
163 | spin_unlock_irqrestore(&info->shadow_lock, flags); | ||
164 | |||
165 | if (kick) | ||
166 | scsifront_wake_up(info); | ||
167 | } | ||
168 | |||
169 | static struct vscsiif_request *scsifront_pre_req(struct vscsifrnt_info *info) | ||
170 | { | ||
171 | struct vscsiif_front_ring *ring = &(info->ring); | ||
172 | struct vscsiif_request *ring_req; | ||
173 | uint32_t id; | ||
174 | |||
175 | id = scsifront_get_rqid(info); /* use id in response */ | ||
176 | if (id >= VSCSIIF_MAX_REQS) | ||
177 | return NULL; | ||
178 | |||
179 | ring_req = RING_GET_REQUEST(&(info->ring), ring->req_prod_pvt); | ||
180 | |||
181 | ring->req_prod_pvt++; | ||
182 | |||
183 | ring_req->rqid = (uint16_t)id; | ||
184 | |||
185 | return ring_req; | ||
186 | } | ||
187 | |||
188 | static void scsifront_do_request(struct vscsifrnt_info *info) | ||
189 | { | ||
190 | struct vscsiif_front_ring *ring = &(info->ring); | ||
191 | int notify; | ||
192 | |||
193 | RING_PUSH_REQUESTS_AND_CHECK_NOTIFY(ring, notify); | ||
194 | if (notify) | ||
195 | notify_remote_via_irq(info->irq); | ||
196 | } | ||
197 | |||
198 | static void scsifront_gnttab_done(struct vscsifrnt_info *info, uint32_t id) | ||
199 | { | ||
200 | struct vscsifrnt_shadow *s = info->shadow[id]; | ||
201 | int i; | ||
202 | |||
203 | if (s->sc->sc_data_direction == DMA_NONE) | ||
204 | return; | ||
205 | |||
206 | for (i = 0; i < s->nr_grants; i++) { | ||
207 | if (unlikely(gnttab_query_foreign_access(s->gref[i]) != 0)) { | ||
208 | shost_printk(KERN_ALERT, info->host, KBUILD_MODNAME | ||
209 | "grant still in use by backend\n"); | ||
210 | BUG(); | ||
211 | } | ||
212 | gnttab_end_foreign_access(s->gref[i], 0, 0UL); | ||
213 | } | ||
214 | |||
215 | kfree(s->sg); | ||
216 | } | ||
217 | |||
218 | static void scsifront_cdb_cmd_done(struct vscsifrnt_info *info, | ||
219 | struct vscsiif_response *ring_rsp) | ||
220 | { | ||
221 | struct scsi_cmnd *sc; | ||
222 | uint32_t id; | ||
223 | uint8_t sense_len; | ||
224 | |||
225 | id = ring_rsp->rqid; | ||
226 | sc = info->shadow[id]->sc; | ||
227 | |||
228 | BUG_ON(sc == NULL); | ||
229 | |||
230 | scsifront_gnttab_done(info, id); | ||
231 | scsifront_put_rqid(info, id); | ||
232 | |||
233 | sc->result = ring_rsp->rslt; | ||
234 | scsi_set_resid(sc, ring_rsp->residual_len); | ||
235 | |||
236 | sense_len = min_t(uint8_t, VSCSIIF_SENSE_BUFFERSIZE, | ||
237 | ring_rsp->sense_len); | ||
238 | |||
239 | if (sense_len) | ||
240 | memcpy(sc->sense_buffer, ring_rsp->sense_buffer, sense_len); | ||
241 | |||
242 | sc->scsi_done(sc); | ||
243 | } | ||
244 | |||
245 | static void scsifront_sync_cmd_done(struct vscsifrnt_info *info, | ||
246 | struct vscsiif_response *ring_rsp) | ||
247 | { | ||
248 | uint16_t id = ring_rsp->rqid; | ||
249 | unsigned long flags; | ||
250 | struct vscsifrnt_shadow *shadow = info->shadow[id]; | ||
251 | int kick; | ||
252 | |||
253 | spin_lock_irqsave(&info->shadow_lock, flags); | ||
254 | shadow->wait_reset = 1; | ||
255 | switch (shadow->rslt_reset) { | ||
256 | case RSLT_RESET_WAITING: | ||
257 | shadow->rslt_reset = ring_rsp->rslt; | ||
258 | break; | ||
259 | case RSLT_RESET_ERR: | ||
260 | kick = _scsifront_put_rqid(info, id); | ||
261 | spin_unlock_irqrestore(&info->shadow_lock, flags); | ||
262 | kfree(shadow); | ||
263 | if (kick) | ||
264 | scsifront_wake_up(info); | ||
265 | return; | ||
266 | default: | ||
267 | shost_printk(KERN_ERR, info->host, KBUILD_MODNAME | ||
268 | "bad reset state %d, possibly leaking %u\n", | ||
269 | shadow->rslt_reset, id); | ||
270 | break; | ||
271 | } | ||
272 | spin_unlock_irqrestore(&info->shadow_lock, flags); | ||
273 | |||
274 | wake_up(&shadow->wq_reset); | ||
275 | } | ||
276 | |||
277 | static int scsifront_cmd_done(struct vscsifrnt_info *info) | ||
278 | { | ||
279 | struct vscsiif_response *ring_rsp; | ||
280 | RING_IDX i, rp; | ||
281 | int more_to_do = 0; | ||
282 | unsigned long flags; | ||
283 | |||
284 | spin_lock_irqsave(info->host->host_lock, flags); | ||
285 | |||
286 | rp = info->ring.sring->rsp_prod; | ||
287 | rmb(); /* ordering required respective to dom0 */ | ||
288 | for (i = info->ring.rsp_cons; i != rp; i++) { | ||
289 | |||
290 | ring_rsp = RING_GET_RESPONSE(&info->ring, i); | ||
291 | |||
292 | if (WARN(ring_rsp->rqid >= VSCSIIF_MAX_REQS || | ||
293 | test_bit(ring_rsp->rqid, info->shadow_free_bitmap), | ||
294 | "illegal rqid %u returned by backend!\n", | ||
295 | ring_rsp->rqid)) | ||
296 | continue; | ||
297 | |||
298 | if (info->shadow[ring_rsp->rqid]->act == VSCSIIF_ACT_SCSI_CDB) | ||
299 | scsifront_cdb_cmd_done(info, ring_rsp); | ||
300 | else | ||
301 | scsifront_sync_cmd_done(info, ring_rsp); | ||
302 | } | ||
303 | |||
304 | info->ring.rsp_cons = i; | ||
305 | |||
306 | if (i != info->ring.req_prod_pvt) | ||
307 | RING_FINAL_CHECK_FOR_RESPONSES(&info->ring, more_to_do); | ||
308 | else | ||
309 | info->ring.sring->rsp_event = i + 1; | ||
310 | |||
311 | info->wait_ring_available = 0; | ||
312 | |||
313 | spin_unlock_irqrestore(info->host->host_lock, flags); | ||
314 | |||
315 | wake_up(&info->wq_sync); | ||
316 | |||
317 | return more_to_do; | ||
318 | } | ||
319 | |||
320 | static irqreturn_t scsifront_irq_fn(int irq, void *dev_id) | ||
321 | { | ||
322 | struct vscsifrnt_info *info = dev_id; | ||
323 | |||
324 | while (scsifront_cmd_done(info)) | ||
325 | /* Yield point for this unbounded loop. */ | ||
326 | cond_resched(); | ||
327 | |||
328 | return IRQ_HANDLED; | ||
329 | } | ||
330 | |||
331 | static int map_data_for_request(struct vscsifrnt_info *info, | ||
332 | struct scsi_cmnd *sc, | ||
333 | struct vscsiif_request *ring_req, | ||
334 | struct vscsifrnt_shadow *shadow) | ||
335 | { | ||
336 | grant_ref_t gref_head; | ||
337 | struct page *page; | ||
338 | int err, ref, ref_cnt = 0; | ||
339 | int grant_ro = (sc->sc_data_direction == DMA_TO_DEVICE); | ||
340 | unsigned int i, off, len, bytes; | ||
341 | unsigned int data_len = scsi_bufflen(sc); | ||
342 | unsigned int data_grants = 0, seg_grants = 0; | ||
343 | struct scatterlist *sg; | ||
344 | unsigned long mfn; | ||
345 | struct scsiif_request_segment *seg; | ||
346 | |||
347 | ring_req->nr_segments = 0; | ||
348 | if (sc->sc_data_direction == DMA_NONE || !data_len) | ||
349 | return 0; | ||
350 | |||
351 | scsi_for_each_sg(sc, sg, scsi_sg_count(sc), i) | ||
352 | data_grants += PFN_UP(sg->offset + sg->length); | ||
353 | |||
354 | if (data_grants > VSCSIIF_SG_TABLESIZE) { | ||
355 | if (data_grants > info->host->sg_tablesize) { | ||
356 | shost_printk(KERN_ERR, info->host, KBUILD_MODNAME | ||
357 | "Unable to map request_buffer for command!\n"); | ||
358 | return -E2BIG; | ||
359 | } | ||
360 | seg_grants = vscsiif_grants_sg(data_grants); | ||
361 | shadow->sg = kcalloc(data_grants, | ||
362 | sizeof(struct scsiif_request_segment), GFP_ATOMIC); | ||
363 | if (!shadow->sg) | ||
364 | return -ENOMEM; | ||
365 | } | ||
366 | seg = shadow->sg ? : ring_req->seg; | ||
367 | |||
368 | err = gnttab_alloc_grant_references(seg_grants + data_grants, | ||
369 | &gref_head); | ||
370 | if (err) { | ||
371 | kfree(shadow->sg); | ||
372 | shost_printk(KERN_ERR, info->host, KBUILD_MODNAME | ||
373 | "gnttab_alloc_grant_references() error\n"); | ||
374 | return -ENOMEM; | ||
375 | } | ||
376 | |||
377 | if (seg_grants) { | ||
378 | page = virt_to_page(seg); | ||
379 | off = (unsigned long)seg & ~PAGE_MASK; | ||
380 | len = sizeof(struct scsiif_request_segment) * data_grants; | ||
381 | while (len > 0) { | ||
382 | bytes = min_t(unsigned int, len, PAGE_SIZE - off); | ||
383 | |||
384 | ref = gnttab_claim_grant_reference(&gref_head); | ||
385 | BUG_ON(ref == -ENOSPC); | ||
386 | |||
387 | mfn = pfn_to_mfn(page_to_pfn(page)); | ||
388 | gnttab_grant_foreign_access_ref(ref, | ||
389 | info->dev->otherend_id, mfn, 1); | ||
390 | shadow->gref[ref_cnt] = ref; | ||
391 | ring_req->seg[ref_cnt].gref = ref; | ||
392 | ring_req->seg[ref_cnt].offset = (uint16_t)off; | ||
393 | ring_req->seg[ref_cnt].length = (uint16_t)bytes; | ||
394 | |||
395 | page++; | ||
396 | len -= bytes; | ||
397 | off = 0; | ||
398 | ref_cnt++; | ||
399 | } | ||
400 | BUG_ON(seg_grants < ref_cnt); | ||
401 | seg_grants = ref_cnt; | ||
402 | } | ||
403 | |||
404 | scsi_for_each_sg(sc, sg, scsi_sg_count(sc), i) { | ||
405 | page = sg_page(sg); | ||
406 | off = sg->offset; | ||
407 | len = sg->length; | ||
408 | |||
409 | while (len > 0 && data_len > 0) { | ||
410 | /* | ||
411 | * sg sends a scatterlist that is larger than | ||
412 | * the data_len it wants transferred for certain | ||
413 | * IO sizes. | ||
414 | */ | ||
415 | bytes = min_t(unsigned int, len, PAGE_SIZE - off); | ||
416 | bytes = min(bytes, data_len); | ||
417 | |||
418 | ref = gnttab_claim_grant_reference(&gref_head); | ||
419 | BUG_ON(ref == -ENOSPC); | ||
420 | |||
421 | mfn = pfn_to_mfn(page_to_pfn(page)); | ||
422 | gnttab_grant_foreign_access_ref(ref, | ||
423 | info->dev->otherend_id, mfn, grant_ro); | ||
424 | |||
425 | shadow->gref[ref_cnt] = ref; | ||
426 | seg->gref = ref; | ||
427 | seg->offset = (uint16_t)off; | ||
428 | seg->length = (uint16_t)bytes; | ||
429 | |||
430 | page++; | ||
431 | seg++; | ||
432 | len -= bytes; | ||
433 | data_len -= bytes; | ||
434 | off = 0; | ||
435 | ref_cnt++; | ||
436 | } | ||
437 | } | ||
438 | |||
439 | if (seg_grants) | ||
440 | ring_req->nr_segments = VSCSIIF_SG_GRANT | seg_grants; | ||
441 | else | ||
442 | ring_req->nr_segments = (uint8_t)ref_cnt; | ||
443 | shadow->nr_grants = ref_cnt; | ||
444 | |||
445 | return 0; | ||
446 | } | ||
447 | |||
448 | static struct vscsiif_request *scsifront_command2ring( | ||
449 | struct vscsifrnt_info *info, struct scsi_cmnd *sc, | ||
450 | struct vscsifrnt_shadow *shadow) | ||
451 | { | ||
452 | struct vscsiif_request *ring_req; | ||
453 | |||
454 | memset(shadow, 0, sizeof(*shadow)); | ||
455 | |||
456 | ring_req = scsifront_pre_req(info); | ||
457 | if (!ring_req) | ||
458 | return NULL; | ||
459 | |||
460 | info->shadow[ring_req->rqid] = shadow; | ||
461 | shadow->rqid = ring_req->rqid; | ||
462 | |||
463 | ring_req->id = sc->device->id; | ||
464 | ring_req->lun = sc->device->lun; | ||
465 | ring_req->channel = sc->device->channel; | ||
466 | ring_req->cmd_len = sc->cmd_len; | ||
467 | |||
468 | BUG_ON(sc->cmd_len > VSCSIIF_MAX_COMMAND_SIZE); | ||
469 | |||
470 | memcpy(ring_req->cmnd, sc->cmnd, sc->cmd_len); | ||
471 | |||
472 | ring_req->sc_data_direction = (uint8_t)sc->sc_data_direction; | ||
473 | ring_req->timeout_per_command = sc->request->timeout / HZ; | ||
474 | |||
475 | return ring_req; | ||
476 | } | ||
477 | |||
478 | static int scsifront_queuecommand(struct Scsi_Host *shost, | ||
479 | struct scsi_cmnd *sc) | ||
480 | { | ||
481 | struct vscsifrnt_info *info = shost_priv(shost); | ||
482 | struct vscsiif_request *ring_req; | ||
483 | struct vscsifrnt_shadow *shadow = scsi_cmd_priv(sc); | ||
484 | unsigned long flags; | ||
485 | int err; | ||
486 | uint16_t rqid; | ||
487 | |||
488 | spin_lock_irqsave(shost->host_lock, flags); | ||
489 | if (RING_FULL(&info->ring)) | ||
490 | goto busy; | ||
491 | |||
492 | ring_req = scsifront_command2ring(info, sc, shadow); | ||
493 | if (!ring_req) | ||
494 | goto busy; | ||
495 | |||
496 | sc->result = 0; | ||
497 | |||
498 | rqid = ring_req->rqid; | ||
499 | ring_req->act = VSCSIIF_ACT_SCSI_CDB; | ||
500 | |||
501 | shadow->sc = sc; | ||
502 | shadow->act = VSCSIIF_ACT_SCSI_CDB; | ||
503 | |||
504 | err = map_data_for_request(info, sc, ring_req, shadow); | ||
505 | if (err < 0) { | ||
506 | pr_debug("%s: err %d\n", __func__, err); | ||
507 | scsifront_put_rqid(info, rqid); | ||
508 | spin_unlock_irqrestore(shost->host_lock, flags); | ||
509 | if (err == -ENOMEM) | ||
510 | return SCSI_MLQUEUE_HOST_BUSY; | ||
511 | sc->result = DID_ERROR << 16; | ||
512 | sc->scsi_done(sc); | ||
513 | return 0; | ||
514 | } | ||
515 | |||
516 | scsifront_do_request(info); | ||
517 | spin_unlock_irqrestore(shost->host_lock, flags); | ||
518 | |||
519 | return 0; | ||
520 | |||
521 | busy: | ||
522 | spin_unlock_irqrestore(shost->host_lock, flags); | ||
523 | pr_debug("%s: busy\n", __func__); | ||
524 | return SCSI_MLQUEUE_HOST_BUSY; | ||
525 | } | ||
526 | |||
527 | /* | ||
528 | * Any exception handling (reset or abort) must be forwarded to the backend. | ||
529 | * We have to wait until an answer is returned. This answer contains the | ||
530 | * result to be returned to the requestor. | ||
531 | */ | ||
532 | static int scsifront_action_handler(struct scsi_cmnd *sc, uint8_t act) | ||
533 | { | ||
534 | struct Scsi_Host *host = sc->device->host; | ||
535 | struct vscsifrnt_info *info = shost_priv(host); | ||
536 | struct vscsifrnt_shadow *shadow, *s = scsi_cmd_priv(sc); | ||
537 | struct vscsiif_request *ring_req; | ||
538 | int err = 0; | ||
539 | |||
540 | shadow = kmalloc(sizeof(*shadow), GFP_NOIO); | ||
541 | if (!shadow) | ||
542 | return FAILED; | ||
543 | |||
544 | spin_lock_irq(host->host_lock); | ||
545 | |||
546 | for (;;) { | ||
547 | if (!RING_FULL(&info->ring)) { | ||
548 | ring_req = scsifront_command2ring(info, sc, shadow); | ||
549 | if (ring_req) | ||
550 | break; | ||
551 | } | ||
552 | if (err) { | ||
553 | spin_unlock_irq(host->host_lock); | ||
554 | kfree(shadow); | ||
555 | return FAILED; | ||
556 | } | ||
557 | info->wait_ring_available = 1; | ||
558 | spin_unlock_irq(host->host_lock); | ||
559 | err = wait_event_interruptible(info->wq_sync, | ||
560 | !info->wait_ring_available); | ||
561 | spin_lock_irq(host->host_lock); | ||
562 | } | ||
563 | |||
564 | ring_req->act = act; | ||
565 | ring_req->ref_rqid = s->rqid; | ||
566 | |||
567 | shadow->act = act; | ||
568 | shadow->rslt_reset = RSLT_RESET_WAITING; | ||
569 | init_waitqueue_head(&shadow->wq_reset); | ||
570 | |||
571 | ring_req->nr_segments = 0; | ||
572 | |||
573 | scsifront_do_request(info); | ||
574 | |||
575 | spin_unlock_irq(host->host_lock); | ||
576 | err = wait_event_interruptible(shadow->wq_reset, shadow->wait_reset); | ||
577 | spin_lock_irq(host->host_lock); | ||
578 | |||
579 | if (!err) { | ||
580 | err = shadow->rslt_reset; | ||
581 | scsifront_put_rqid(info, shadow->rqid); | ||
582 | kfree(shadow); | ||
583 | } else { | ||
584 | spin_lock(&info->shadow_lock); | ||
585 | shadow->rslt_reset = RSLT_RESET_ERR; | ||
586 | spin_unlock(&info->shadow_lock); | ||
587 | err = FAILED; | ||
588 | } | ||
589 | |||
590 | spin_unlock_irq(host->host_lock); | ||
591 | return err; | ||
592 | } | ||
593 | |||
594 | static int scsifront_eh_abort_handler(struct scsi_cmnd *sc) | ||
595 | { | ||
596 | pr_debug("%s\n", __func__); | ||
597 | return scsifront_action_handler(sc, VSCSIIF_ACT_SCSI_ABORT); | ||
598 | } | ||
599 | |||
600 | static int scsifront_dev_reset_handler(struct scsi_cmnd *sc) | ||
601 | { | ||
602 | pr_debug("%s\n", __func__); | ||
603 | return scsifront_action_handler(sc, VSCSIIF_ACT_SCSI_RESET); | ||
604 | } | ||
605 | |||
606 | static int scsifront_sdev_configure(struct scsi_device *sdev) | ||
607 | { | ||
608 | struct vscsifrnt_info *info = shost_priv(sdev->host); | ||
609 | |||
610 | if (info && current == info->curr) | ||
611 | xenbus_printf(XBT_NIL, info->dev->nodename, | ||
612 | info->dev_state_path, "%d", XenbusStateConnected); | ||
613 | |||
614 | return 0; | ||
615 | } | ||
616 | |||
617 | static void scsifront_sdev_destroy(struct scsi_device *sdev) | ||
618 | { | ||
619 | struct vscsifrnt_info *info = shost_priv(sdev->host); | ||
620 | |||
621 | if (info && current == info->curr) | ||
622 | xenbus_printf(XBT_NIL, info->dev->nodename, | ||
623 | info->dev_state_path, "%d", XenbusStateClosed); | ||
624 | } | ||
625 | |||
626 | static struct scsi_host_template scsifront_sht = { | ||
627 | .module = THIS_MODULE, | ||
628 | .name = "Xen SCSI frontend driver", | ||
629 | .queuecommand = scsifront_queuecommand, | ||
630 | .eh_abort_handler = scsifront_eh_abort_handler, | ||
631 | .eh_device_reset_handler = scsifront_dev_reset_handler, | ||
632 | .slave_configure = scsifront_sdev_configure, | ||
633 | .slave_destroy = scsifront_sdev_destroy, | ||
634 | .cmd_per_lun = VSCSIIF_DEFAULT_CMD_PER_LUN, | ||
635 | .can_queue = VSCSIIF_MAX_REQS, | ||
636 | .this_id = -1, | ||
637 | .cmd_size = sizeof(struct vscsifrnt_shadow), | ||
638 | .sg_tablesize = VSCSIIF_SG_TABLESIZE, | ||
639 | .use_clustering = DISABLE_CLUSTERING, | ||
640 | .proc_name = "scsifront", | ||
641 | }; | ||
642 | |||
643 | static int scsifront_alloc_ring(struct vscsifrnt_info *info) | ||
644 | { | ||
645 | struct xenbus_device *dev = info->dev; | ||
646 | struct vscsiif_sring *sring; | ||
647 | int err = -ENOMEM; | ||
648 | |||
649 | /***** Frontend to Backend ring start *****/ | ||
650 | sring = (struct vscsiif_sring *)__get_free_page(GFP_KERNEL); | ||
651 | if (!sring) { | ||
652 | xenbus_dev_fatal(dev, err, | ||
653 | "fail to allocate shared ring (Front to Back)"); | ||
654 | return err; | ||
655 | } | ||
656 | SHARED_RING_INIT(sring); | ||
657 | FRONT_RING_INIT(&info->ring, sring, PAGE_SIZE); | ||
658 | |||
659 | err = xenbus_grant_ring(dev, virt_to_mfn(sring)); | ||
660 | if (err < 0) { | ||
661 | free_page((unsigned long)sring); | ||
662 | xenbus_dev_fatal(dev, err, | ||
663 | "fail to grant shared ring (Front to Back)"); | ||
664 | return err; | ||
665 | } | ||
666 | info->ring_ref = err; | ||
667 | |||
668 | err = xenbus_alloc_evtchn(dev, &info->evtchn); | ||
669 | if (err) { | ||
670 | xenbus_dev_fatal(dev, err, "xenbus_alloc_evtchn"); | ||
671 | goto free_gnttab; | ||
672 | } | ||
673 | |||
674 | err = bind_evtchn_to_irq(info->evtchn); | ||
675 | if (err <= 0) { | ||
676 | xenbus_dev_fatal(dev, err, "bind_evtchn_to_irq"); | ||
677 | goto free_gnttab; | ||
678 | } | ||
679 | |||
680 | info->irq = err; | ||
681 | |||
682 | err = request_threaded_irq(info->irq, NULL, scsifront_irq_fn, | ||
683 | IRQF_ONESHOT, "scsifront", info); | ||
684 | if (err) { | ||
685 | xenbus_dev_fatal(dev, err, "request_threaded_irq"); | ||
686 | goto free_irq; | ||
687 | } | ||
688 | |||
689 | return 0; | ||
690 | |||
691 | /* free resource */ | ||
692 | free_irq: | ||
693 | unbind_from_irqhandler(info->irq, info); | ||
694 | free_gnttab: | ||
695 | gnttab_end_foreign_access(info->ring_ref, 0, | ||
696 | (unsigned long)info->ring.sring); | ||
697 | |||
698 | return err; | ||
699 | } | ||
700 | |||
701 | static int scsifront_init_ring(struct vscsifrnt_info *info) | ||
702 | { | ||
703 | struct xenbus_device *dev = info->dev; | ||
704 | struct xenbus_transaction xbt; | ||
705 | int err; | ||
706 | |||
707 | pr_debug("%s\n", __func__); | ||
708 | |||
709 | err = scsifront_alloc_ring(info); | ||
710 | if (err) | ||
711 | return err; | ||
712 | pr_debug("%s: %u %u\n", __func__, info->ring_ref, info->evtchn); | ||
713 | |||
714 | again: | ||
715 | err = xenbus_transaction_start(&xbt); | ||
716 | if (err) | ||
717 | xenbus_dev_fatal(dev, err, "starting transaction"); | ||
718 | |||
719 | err = xenbus_printf(xbt, dev->nodename, "ring-ref", "%u", | ||
720 | info->ring_ref); | ||
721 | if (err) { | ||
722 | xenbus_dev_fatal(dev, err, "%s", "writing ring-ref"); | ||
723 | goto fail; | ||
724 | } | ||
725 | |||
726 | err = xenbus_printf(xbt, dev->nodename, "event-channel", "%u", | ||
727 | info->evtchn); | ||
728 | |||
729 | if (err) { | ||
730 | xenbus_dev_fatal(dev, err, "%s", "writing event-channel"); | ||
731 | goto fail; | ||
732 | } | ||
733 | |||
734 | err = xenbus_transaction_end(xbt, 0); | ||
735 | if (err) { | ||
736 | if (err == -EAGAIN) | ||
737 | goto again; | ||
738 | xenbus_dev_fatal(dev, err, "completing transaction"); | ||
739 | goto free_sring; | ||
740 | } | ||
741 | |||
742 | return 0; | ||
743 | |||
744 | fail: | ||
745 | xenbus_transaction_end(xbt, 1); | ||
746 | free_sring: | ||
747 | unbind_from_irqhandler(info->irq, info); | ||
748 | gnttab_end_foreign_access(info->ring_ref, 0, | ||
749 | (unsigned long)info->ring.sring); | ||
750 | |||
751 | return err; | ||
752 | } | ||
753 | |||
754 | |||
755 | static int scsifront_probe(struct xenbus_device *dev, | ||
756 | const struct xenbus_device_id *id) | ||
757 | { | ||
758 | struct vscsifrnt_info *info; | ||
759 | struct Scsi_Host *host; | ||
760 | int err = -ENOMEM; | ||
761 | char name[TASK_COMM_LEN]; | ||
762 | |||
763 | host = scsi_host_alloc(&scsifront_sht, sizeof(*info)); | ||
764 | if (!host) { | ||
765 | xenbus_dev_fatal(dev, err, "fail to allocate scsi host"); | ||
766 | return err; | ||
767 | } | ||
768 | info = (struct vscsifrnt_info *)host->hostdata; | ||
769 | |||
770 | dev_set_drvdata(&dev->dev, info); | ||
771 | info->dev = dev; | ||
772 | |||
773 | bitmap_fill(info->shadow_free_bitmap, VSCSIIF_MAX_REQS); | ||
774 | |||
775 | err = scsifront_init_ring(info); | ||
776 | if (err) { | ||
777 | scsi_host_put(host); | ||
778 | return err; | ||
779 | } | ||
780 | |||
781 | init_waitqueue_head(&info->wq_sync); | ||
782 | spin_lock_init(&info->shadow_lock); | ||
783 | |||
784 | snprintf(name, TASK_COMM_LEN, "vscsiif.%d", host->host_no); | ||
785 | |||
786 | host->max_id = VSCSIIF_MAX_TARGET; | ||
787 | host->max_channel = 0; | ||
788 | host->max_lun = VSCSIIF_MAX_LUN; | ||
789 | host->max_sectors = (host->sg_tablesize - 1) * PAGE_SIZE / 512; | ||
790 | host->max_cmd_len = VSCSIIF_MAX_COMMAND_SIZE; | ||
791 | |||
792 | err = scsi_add_host(host, &dev->dev); | ||
793 | if (err) { | ||
794 | dev_err(&dev->dev, "fail to add scsi host %d\n", err); | ||
795 | goto free_sring; | ||
796 | } | ||
797 | info->host = host; | ||
798 | info->host_active = 1; | ||
799 | |||
800 | xenbus_switch_state(dev, XenbusStateInitialised); | ||
801 | |||
802 | return 0; | ||
803 | |||
804 | free_sring: | ||
805 | unbind_from_irqhandler(info->irq, info); | ||
806 | gnttab_end_foreign_access(info->ring_ref, 0, | ||
807 | (unsigned long)info->ring.sring); | ||
808 | scsi_host_put(host); | ||
809 | return err; | ||
810 | } | ||
811 | |||
812 | static int scsifront_remove(struct xenbus_device *dev) | ||
813 | { | ||
814 | struct vscsifrnt_info *info = dev_get_drvdata(&dev->dev); | ||
815 | |||
816 | pr_debug("%s: %s removed\n", __func__, dev->nodename); | ||
817 | |||
818 | mutex_lock(&scsifront_mutex); | ||
819 | if (info->host_active) { | ||
820 | /* Scsi_host not yet removed */ | ||
821 | scsi_remove_host(info->host); | ||
822 | info->host_active = 0; | ||
823 | } | ||
824 | mutex_unlock(&scsifront_mutex); | ||
825 | |||
826 | gnttab_end_foreign_access(info->ring_ref, 0, | ||
827 | (unsigned long)info->ring.sring); | ||
828 | unbind_from_irqhandler(info->irq, info); | ||
829 | |||
830 | scsi_host_put(info->host); | ||
831 | |||
832 | return 0; | ||
833 | } | ||
834 | |||
835 | static void scsifront_disconnect(struct vscsifrnt_info *info) | ||
836 | { | ||
837 | struct xenbus_device *dev = info->dev; | ||
838 | struct Scsi_Host *host = info->host; | ||
839 | |||
840 | pr_debug("%s: %s disconnect\n", __func__, dev->nodename); | ||
841 | |||
842 | /* | ||
843 | * When this function is executed, all devices of | ||
844 | * Frontend have been deleted. | ||
845 | * Therefore, it need not block I/O before remove_host. | ||
846 | */ | ||
847 | |||
848 | mutex_lock(&scsifront_mutex); | ||
849 | if (info->host_active) { | ||
850 | scsi_remove_host(host); | ||
851 | info->host_active = 0; | ||
852 | } | ||
853 | mutex_unlock(&scsifront_mutex); | ||
854 | |||
855 | xenbus_frontend_closed(dev); | ||
856 | } | ||
857 | |||
858 | static void scsifront_do_lun_hotplug(struct vscsifrnt_info *info, int op) | ||
859 | { | ||
860 | struct xenbus_device *dev = info->dev; | ||
861 | int i, err = 0; | ||
862 | char str[64]; | ||
863 | char **dir; | ||
864 | unsigned int dir_n = 0; | ||
865 | unsigned int device_state; | ||
866 | unsigned int hst, chn, tgt, lun; | ||
867 | struct scsi_device *sdev; | ||
868 | |||
869 | dir = xenbus_directory(XBT_NIL, dev->otherend, "vscsi-devs", &dir_n); | ||
870 | if (IS_ERR(dir)) | ||
871 | return; | ||
872 | |||
873 | /* mark current task as the one allowed to modify device states */ | ||
874 | BUG_ON(info->curr); | ||
875 | info->curr = current; | ||
876 | |||
877 | for (i = 0; i < dir_n; i++) { | ||
878 | /* read status */ | ||
879 | snprintf(str, sizeof(str), "vscsi-devs/%s/state", dir[i]); | ||
880 | err = xenbus_scanf(XBT_NIL, dev->otherend, str, "%u", | ||
881 | &device_state); | ||
882 | if (XENBUS_EXIST_ERR(err)) | ||
883 | continue; | ||
884 | |||
885 | /* virtual SCSI device */ | ||
886 | snprintf(str, sizeof(str), "vscsi-devs/%s/v-dev", dir[i]); | ||
887 | err = xenbus_scanf(XBT_NIL, dev->otherend, str, | ||
888 | "%u:%u:%u:%u", &hst, &chn, &tgt, &lun); | ||
889 | if (XENBUS_EXIST_ERR(err)) | ||
890 | continue; | ||
891 | |||
892 | /* | ||
893 | * Front device state path, used in slave_configure called | ||
894 | * on successfull scsi_add_device, and in slave_destroy called | ||
895 | * on remove of a device. | ||
896 | */ | ||
897 | snprintf(info->dev_state_path, sizeof(info->dev_state_path), | ||
898 | "vscsi-devs/%s/state", dir[i]); | ||
899 | |||
900 | switch (op) { | ||
901 | case VSCSIFRONT_OP_ADD_LUN: | ||
902 | if (device_state != XenbusStateInitialised) | ||
903 | break; | ||
904 | |||
905 | if (scsi_add_device(info->host, chn, tgt, lun)) { | ||
906 | dev_err(&dev->dev, "scsi_add_device\n"); | ||
907 | xenbus_printf(XBT_NIL, dev->nodename, | ||
908 | info->dev_state_path, | ||
909 | "%d", XenbusStateClosed); | ||
910 | } | ||
911 | break; | ||
912 | case VSCSIFRONT_OP_DEL_LUN: | ||
913 | if (device_state != XenbusStateClosing) | ||
914 | break; | ||
915 | |||
916 | sdev = scsi_device_lookup(info->host, chn, tgt, lun); | ||
917 | if (sdev) { | ||
918 | scsi_remove_device(sdev); | ||
919 | scsi_device_put(sdev); | ||
920 | } | ||
921 | break; | ||
922 | default: | ||
923 | break; | ||
924 | } | ||
925 | } | ||
926 | |||
927 | info->curr = NULL; | ||
928 | |||
929 | kfree(dir); | ||
930 | } | ||
931 | |||
932 | static void scsifront_read_backend_params(struct xenbus_device *dev, | ||
933 | struct vscsifrnt_info *info) | ||
934 | { | ||
935 | unsigned int sg_grant; | ||
936 | int ret; | ||
937 | struct Scsi_Host *host = info->host; | ||
938 | |||
939 | ret = xenbus_scanf(XBT_NIL, dev->otherend, "feature-sg-grant", "%u", | ||
940 | &sg_grant); | ||
941 | if (ret == 1 && sg_grant) { | ||
942 | sg_grant = min_t(unsigned int, sg_grant, SG_ALL); | ||
943 | sg_grant = max_t(unsigned int, sg_grant, VSCSIIF_SG_TABLESIZE); | ||
944 | host->sg_tablesize = min_t(unsigned int, sg_grant, | ||
945 | VSCSIIF_SG_TABLESIZE * PAGE_SIZE / | ||
946 | sizeof(struct scsiif_request_segment)); | ||
947 | host->max_sectors = (host->sg_tablesize - 1) * PAGE_SIZE / 512; | ||
948 | } | ||
949 | dev_info(&dev->dev, "using up to %d SG entries\n", host->sg_tablesize); | ||
950 | } | ||
951 | |||
952 | static void scsifront_backend_changed(struct xenbus_device *dev, | ||
953 | enum xenbus_state backend_state) | ||
954 | { | ||
955 | struct vscsifrnt_info *info = dev_get_drvdata(&dev->dev); | ||
956 | |||
957 | pr_debug("%s: %p %u %u\n", __func__, dev, dev->state, backend_state); | ||
958 | |||
959 | switch (backend_state) { | ||
960 | case XenbusStateUnknown: | ||
961 | case XenbusStateInitialising: | ||
962 | case XenbusStateInitWait: | ||
963 | case XenbusStateInitialised: | ||
964 | break; | ||
965 | |||
966 | case XenbusStateConnected: | ||
967 | scsifront_read_backend_params(dev, info); | ||
968 | if (xenbus_read_driver_state(dev->nodename) == | ||
969 | XenbusStateInitialised) | ||
970 | scsifront_do_lun_hotplug(info, VSCSIFRONT_OP_ADD_LUN); | ||
971 | |||
972 | if (dev->state != XenbusStateConnected) | ||
973 | xenbus_switch_state(dev, XenbusStateConnected); | ||
974 | break; | ||
975 | |||
976 | case XenbusStateClosed: | ||
977 | if (dev->state == XenbusStateClosed) | ||
978 | break; | ||
979 | /* Missed the backend's Closing state -- fallthrough */ | ||
980 | case XenbusStateClosing: | ||
981 | scsifront_disconnect(info); | ||
982 | break; | ||
983 | |||
984 | case XenbusStateReconfiguring: | ||
985 | scsifront_do_lun_hotplug(info, VSCSIFRONT_OP_DEL_LUN); | ||
986 | xenbus_switch_state(dev, XenbusStateReconfiguring); | ||
987 | break; | ||
988 | |||
989 | case XenbusStateReconfigured: | ||
990 | scsifront_do_lun_hotplug(info, VSCSIFRONT_OP_ADD_LUN); | ||
991 | xenbus_switch_state(dev, XenbusStateConnected); | ||
992 | break; | ||
993 | } | ||
994 | } | ||
995 | |||
996 | static const struct xenbus_device_id scsifront_ids[] = { | ||
997 | { "vscsi" }, | ||
998 | { "" } | ||
999 | }; | ||
1000 | |||
1001 | static struct xenbus_driver scsifront_driver = { | ||
1002 | .ids = scsifront_ids, | ||
1003 | .probe = scsifront_probe, | ||
1004 | .remove = scsifront_remove, | ||
1005 | .otherend_changed = scsifront_backend_changed, | ||
1006 | }; | ||
1007 | |||
1008 | static int __init scsifront_init(void) | ||
1009 | { | ||
1010 | if (!xen_domain()) | ||
1011 | return -ENODEV; | ||
1012 | |||
1013 | return xenbus_register_frontend(&scsifront_driver); | ||
1014 | } | ||
1015 | module_init(scsifront_init); | ||
1016 | |||
1017 | static void __exit scsifront_exit(void) | ||
1018 | { | ||
1019 | xenbus_unregister_driver(&scsifront_driver); | ||
1020 | } | ||
1021 | module_exit(scsifront_exit); | ||
1022 | |||
1023 | MODULE_DESCRIPTION("Xen SCSI frontend driver"); | ||
1024 | MODULE_LICENSE("GPL"); | ||
1025 | MODULE_ALIAS("xen:vscsi"); | ||
1026 | MODULE_AUTHOR("Juergen Gross <jgross@suse.com>"); | ||
diff --git a/drivers/tty/hvc/hvc_xen.c b/drivers/tty/hvc/hvc_xen.c index 2967f0388d2c..f1e57425e39f 100644 --- a/drivers/tty/hvc/hvc_xen.c +++ b/drivers/tty/hvc/hvc_xen.c | |||
@@ -347,8 +347,6 @@ static int xen_console_remove(struct xencons_info *info) | |||
347 | } | 347 | } |
348 | 348 | ||
349 | #ifdef CONFIG_HVC_XEN_FRONTEND | 349 | #ifdef CONFIG_HVC_XEN_FRONTEND |
350 | static struct xenbus_driver xencons_driver; | ||
351 | |||
352 | static int xencons_remove(struct xenbus_device *dev) | 350 | static int xencons_remove(struct xenbus_device *dev) |
353 | { | 351 | { |
354 | return xen_console_remove(dev_get_drvdata(&dev->dev)); | 352 | return xen_console_remove(dev_get_drvdata(&dev->dev)); |
@@ -499,13 +497,14 @@ static const struct xenbus_device_id xencons_ids[] = { | |||
499 | { "" } | 497 | { "" } |
500 | }; | 498 | }; |
501 | 499 | ||
502 | 500 | static struct xenbus_driver xencons_driver = { | |
503 | static DEFINE_XENBUS_DRIVER(xencons, "xenconsole", | 501 | .name = "xenconsole", |
502 | .ids = xencons_ids, | ||
504 | .probe = xencons_probe, | 503 | .probe = xencons_probe, |
505 | .remove = xencons_remove, | 504 | .remove = xencons_remove, |
506 | .resume = xencons_resume, | 505 | .resume = xencons_resume, |
507 | .otherend_changed = xencons_backend_changed, | 506 | .otherend_changed = xencons_backend_changed, |
508 | ); | 507 | }; |
509 | #endif /* CONFIG_HVC_XEN_FRONTEND */ | 508 | #endif /* CONFIG_HVC_XEN_FRONTEND */ |
510 | 509 | ||
511 | static int __init xen_hvc_init(void) | 510 | static int __init xen_hvc_init(void) |
diff --git a/drivers/video/fbdev/xen-fbfront.c b/drivers/video/fbdev/xen-fbfront.c index 901014bbc821..09dc44736c1a 100644 --- a/drivers/video/fbdev/xen-fbfront.c +++ b/drivers/video/fbdev/xen-fbfront.c | |||
@@ -684,12 +684,13 @@ static const struct xenbus_device_id xenfb_ids[] = { | |||
684 | { "" } | 684 | { "" } |
685 | }; | 685 | }; |
686 | 686 | ||
687 | static DEFINE_XENBUS_DRIVER(xenfb, , | 687 | static struct xenbus_driver xenfb_driver = { |
688 | .ids = xenfb_ids, | ||
688 | .probe = xenfb_probe, | 689 | .probe = xenfb_probe, |
689 | .remove = xenfb_remove, | 690 | .remove = xenfb_remove, |
690 | .resume = xenfb_resume, | 691 | .resume = xenfb_resume, |
691 | .otherend_changed = xenfb_backend_changed, | 692 | .otherend_changed = xenfb_backend_changed, |
692 | ); | 693 | }; |
693 | 694 | ||
694 | static int __init xenfb_init(void) | 695 | static int __init xenfb_init(void) |
695 | { | 696 | { |
diff --git a/drivers/xen/Kconfig b/drivers/xen/Kconfig index 8bc01838daf9..b812462083fc 100644 --- a/drivers/xen/Kconfig +++ b/drivers/xen/Kconfig | |||
@@ -172,6 +172,15 @@ config XEN_PCIDEV_BACKEND | |||
172 | 172 | ||
173 | If in doubt, say m. | 173 | If in doubt, say m. |
174 | 174 | ||
175 | config XEN_SCSI_BACKEND | ||
176 | tristate "XEN SCSI backend driver" | ||
177 | depends on XEN && XEN_BACKEND && TARGET_CORE | ||
178 | help | ||
179 | The SCSI backend driver allows the kernel to export its SCSI Devices | ||
180 | to other guests via a high-performance shared-memory interface. | ||
181 | Only needed for systems running as XEN driver domains (e.g. Dom0) and | ||
182 | if guests need generic access to SCSI devices. | ||
183 | |||
175 | config XEN_PRIVCMD | 184 | config XEN_PRIVCMD |
176 | tristate | 185 | tristate |
177 | depends on XEN | 186 | depends on XEN |
diff --git a/drivers/xen/Makefile b/drivers/xen/Makefile index 84044b554e33..2140398a2a8c 100644 --- a/drivers/xen/Makefile +++ b/drivers/xen/Makefile | |||
@@ -36,6 +36,7 @@ obj-$(CONFIG_XEN_ACPI_HOTPLUG_MEMORY) += xen-acpi-memhotplug.o | |||
36 | obj-$(CONFIG_XEN_ACPI_HOTPLUG_CPU) += xen-acpi-cpuhotplug.o | 36 | obj-$(CONFIG_XEN_ACPI_HOTPLUG_CPU) += xen-acpi-cpuhotplug.o |
37 | obj-$(CONFIG_XEN_ACPI_PROCESSOR) += xen-acpi-processor.o | 37 | obj-$(CONFIG_XEN_ACPI_PROCESSOR) += xen-acpi-processor.o |
38 | obj-$(CONFIG_XEN_EFI) += efi.o | 38 | obj-$(CONFIG_XEN_EFI) += efi.o |
39 | obj-$(CONFIG_XEN_SCSI_BACKEND) += xen-scsiback.o | ||
39 | xen-evtchn-y := evtchn.o | 40 | xen-evtchn-y := evtchn.o |
40 | xen-gntdev-y := gntdev.o | 41 | xen-gntdev-y := gntdev.o |
41 | xen-gntalloc-y := gntalloc.o | 42 | xen-gntalloc-y := gntalloc.o |
diff --git a/drivers/xen/efi.c b/drivers/xen/efi.c index 31f618a49661..1f850c97482f 100644 --- a/drivers/xen/efi.c +++ b/drivers/xen/efi.c | |||
@@ -27,6 +27,8 @@ | |||
27 | #include <xen/interface/platform.h> | 27 | #include <xen/interface/platform.h> |
28 | #include <xen/xen.h> | 28 | #include <xen/xen.h> |
29 | 29 | ||
30 | #include <asm/page.h> | ||
31 | |||
30 | #include <asm/xen/hypercall.h> | 32 | #include <asm/xen/hypercall.h> |
31 | 33 | ||
32 | #define INIT_EFI_OP(name) \ | 34 | #define INIT_EFI_OP(name) \ |
diff --git a/drivers/xen/events/events_base.c b/drivers/xen/events/events_base.c index 5b5c5ff273fd..b4bca2d4a7e5 100644 --- a/drivers/xen/events/events_base.c +++ b/drivers/xen/events/events_base.c | |||
@@ -900,8 +900,8 @@ static int bind_ipi_to_irq(unsigned int ipi, unsigned int cpu) | |||
900 | return irq; | 900 | return irq; |
901 | } | 901 | } |
902 | 902 | ||
903 | static int bind_interdomain_evtchn_to_irq(unsigned int remote_domain, | 903 | int bind_interdomain_evtchn_to_irq(unsigned int remote_domain, |
904 | unsigned int remote_port) | 904 | unsigned int remote_port) |
905 | { | 905 | { |
906 | struct evtchn_bind_interdomain bind_interdomain; | 906 | struct evtchn_bind_interdomain bind_interdomain; |
907 | int err; | 907 | int err; |
@@ -914,6 +914,7 @@ static int bind_interdomain_evtchn_to_irq(unsigned int remote_domain, | |||
914 | 914 | ||
915 | return err ? : bind_evtchn_to_irq(bind_interdomain.local_port); | 915 | return err ? : bind_evtchn_to_irq(bind_interdomain.local_port); |
916 | } | 916 | } |
917 | EXPORT_SYMBOL_GPL(bind_interdomain_evtchn_to_irq); | ||
917 | 918 | ||
918 | static int find_virq(unsigned int virq, unsigned int cpu) | 919 | static int find_virq(unsigned int virq, unsigned int cpu) |
919 | { | 920 | { |
diff --git a/drivers/xen/grant-table.c b/drivers/xen/grant-table.c index c254ae036f18..7786291ba229 100644 --- a/drivers/xen/grant-table.c +++ b/drivers/xen/grant-table.c | |||
@@ -592,7 +592,7 @@ static int grow_gnttab_list(unsigned int more_frames) | |||
592 | return 0; | 592 | return 0; |
593 | 593 | ||
594 | grow_nomem: | 594 | grow_nomem: |
595 | for ( ; i >= nr_glist_frames; i--) | 595 | while (i-- > nr_glist_frames) |
596 | free_page((unsigned long) gnttab_list[i]); | 596 | free_page((unsigned long) gnttab_list[i]); |
597 | return -ENOMEM; | 597 | return -ENOMEM; |
598 | } | 598 | } |
diff --git a/drivers/xen/xen-pciback/xenbus.c b/drivers/xen/xen-pciback/xenbus.c index c214daab4829..ad8d30c088fe 100644 --- a/drivers/xen/xen-pciback/xenbus.c +++ b/drivers/xen/xen-pciback/xenbus.c | |||
@@ -719,11 +719,13 @@ static const struct xenbus_device_id xen_pcibk_ids[] = { | |||
719 | {""}, | 719 | {""}, |
720 | }; | 720 | }; |
721 | 721 | ||
722 | static DEFINE_XENBUS_DRIVER(xen_pcibk, DRV_NAME, | 722 | static struct xenbus_driver xen_pcibk_driver = { |
723 | .name = DRV_NAME, | ||
724 | .ids = xen_pcibk_ids, | ||
723 | .probe = xen_pcibk_xenbus_probe, | 725 | .probe = xen_pcibk_xenbus_probe, |
724 | .remove = xen_pcibk_xenbus_remove, | 726 | .remove = xen_pcibk_xenbus_remove, |
725 | .otherend_changed = xen_pcibk_frontend_changed, | 727 | .otherend_changed = xen_pcibk_frontend_changed, |
726 | ); | 728 | }; |
727 | 729 | ||
728 | const struct xen_pcibk_backend *__read_mostly xen_pcibk_backend; | 730 | const struct xen_pcibk_backend *__read_mostly xen_pcibk_backend; |
729 | 731 | ||
diff --git a/drivers/xen/xen-scsiback.c b/drivers/xen/xen-scsiback.c new file mode 100644 index 000000000000..3e32146472a5 --- /dev/null +++ b/drivers/xen/xen-scsiback.c | |||
@@ -0,0 +1,2126 @@ | |||
1 | /* | ||
2 | * Xen SCSI backend driver | ||
3 | * | ||
4 | * Copyright (c) 2008, FUJITSU Limited | ||
5 | * | ||
6 | * Based on the blkback driver code. | ||
7 | * Adaption to kernel taget core infrastructure taken from vhost/scsi.c | ||
8 | * | ||
9 | * This program is free software; you can redistribute it and/or | ||
10 | * modify it under the terms of the GNU General Public License version 2 | ||
11 | * as published by the Free Software Foundation; or, when distributed | ||
12 | * separately from the Linux kernel or incorporated into other | ||
13 | * software packages, subject to the following license: | ||
14 | * | ||
15 | * Permission is hereby granted, free of charge, to any person obtaining a copy | ||
16 | * of this source file (the "Software"), to deal in the Software without | ||
17 | * restriction, including without limitation the rights to use, copy, modify, | ||
18 | * merge, publish, distribute, sublicense, and/or sell copies of the Software, | ||
19 | * and to permit persons to whom the Software is furnished to do so, subject to | ||
20 | * the following conditions: | ||
21 | * | ||
22 | * The above copyright notice and this permission notice shall be included in | ||
23 | * all copies or substantial portions of the Software. | ||
24 | * | ||
25 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
26 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
27 | * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE | ||
28 | * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||
29 | * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING | ||
30 | * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS | ||
31 | * IN THE SOFTWARE. | ||
32 | */ | ||
33 | |||
34 | #include <stdarg.h> | ||
35 | |||
36 | #include <linux/module.h> | ||
37 | #include <linux/utsname.h> | ||
38 | #include <linux/interrupt.h> | ||
39 | #include <linux/slab.h> | ||
40 | #include <linux/wait.h> | ||
41 | #include <linux/sched.h> | ||
42 | #include <linux/list.h> | ||
43 | #include <linux/gfp.h> | ||
44 | #include <linux/delay.h> | ||
45 | #include <linux/spinlock.h> | ||
46 | #include <linux/configfs.h> | ||
47 | |||
48 | #include <generated/utsrelease.h> | ||
49 | |||
50 | #include <scsi/scsi_dbg.h> | ||
51 | #include <scsi/scsi_eh.h> | ||
52 | #include <scsi/scsi_tcq.h> | ||
53 | |||
54 | #include <target/target_core_base.h> | ||
55 | #include <target/target_core_fabric.h> | ||
56 | #include <target/target_core_configfs.h> | ||
57 | #include <target/target_core_fabric_configfs.h> | ||
58 | |||
59 | #include <asm/hypervisor.h> | ||
60 | |||
61 | #include <xen/xen.h> | ||
62 | #include <xen/balloon.h> | ||
63 | #include <xen/events.h> | ||
64 | #include <xen/xenbus.h> | ||
65 | #include <xen/grant_table.h> | ||
66 | #include <xen/page.h> | ||
67 | |||
68 | #include <xen/interface/grant_table.h> | ||
69 | #include <xen/interface/io/vscsiif.h> | ||
70 | |||
71 | #define DPRINTK(_f, _a...) \ | ||
72 | pr_debug("(file=%s, line=%d) " _f, __FILE__ , __LINE__ , ## _a) | ||
73 | |||
74 | #define VSCSI_VERSION "v0.1" | ||
75 | #define VSCSI_NAMELEN 32 | ||
76 | |||
77 | struct ids_tuple { | ||
78 | unsigned int hst; /* host */ | ||
79 | unsigned int chn; /* channel */ | ||
80 | unsigned int tgt; /* target */ | ||
81 | unsigned int lun; /* LUN */ | ||
82 | }; | ||
83 | |||
84 | struct v2p_entry { | ||
85 | struct ids_tuple v; /* translate from */ | ||
86 | struct scsiback_tpg *tpg; /* translate to */ | ||
87 | unsigned int lun; | ||
88 | struct kref kref; | ||
89 | struct list_head l; | ||
90 | }; | ||
91 | |||
92 | struct vscsibk_info { | ||
93 | struct xenbus_device *dev; | ||
94 | |||
95 | domid_t domid; | ||
96 | unsigned int irq; | ||
97 | |||
98 | struct vscsiif_back_ring ring; | ||
99 | int ring_error; | ||
100 | |||
101 | spinlock_t ring_lock; | ||
102 | atomic_t nr_unreplied_reqs; | ||
103 | |||
104 | spinlock_t v2p_lock; | ||
105 | struct list_head v2p_entry_lists; | ||
106 | |||
107 | wait_queue_head_t waiting_to_free; | ||
108 | }; | ||
109 | |||
110 | /* theoretical maximum of grants for one request */ | ||
111 | #define VSCSI_MAX_GRANTS (SG_ALL + VSCSIIF_SG_TABLESIZE) | ||
112 | |||
113 | /* | ||
114 | * VSCSI_GRANT_BATCH is the maximum number of grants to be processed in one | ||
115 | * call to map/unmap grants. Don't choose it too large, as there are arrays | ||
116 | * with VSCSI_GRANT_BATCH elements allocated on the stack. | ||
117 | */ | ||
118 | #define VSCSI_GRANT_BATCH 16 | ||
119 | |||
120 | struct vscsibk_pend { | ||
121 | uint16_t rqid; | ||
122 | |||
123 | uint8_t cmnd[VSCSIIF_MAX_COMMAND_SIZE]; | ||
124 | uint8_t cmd_len; | ||
125 | |||
126 | uint8_t sc_data_direction; | ||
127 | uint16_t n_sg; /* real length of SG list */ | ||
128 | uint16_t n_grants; /* SG pages and potentially SG list */ | ||
129 | uint32_t data_len; | ||
130 | uint32_t result; | ||
131 | |||
132 | struct vscsibk_info *info; | ||
133 | struct v2p_entry *v2p; | ||
134 | struct scatterlist *sgl; | ||
135 | |||
136 | uint8_t sense_buffer[VSCSIIF_SENSE_BUFFERSIZE]; | ||
137 | |||
138 | grant_handle_t grant_handles[VSCSI_MAX_GRANTS]; | ||
139 | struct page *pages[VSCSI_MAX_GRANTS]; | ||
140 | |||
141 | struct se_cmd se_cmd; | ||
142 | }; | ||
143 | |||
144 | struct scsiback_tmr { | ||
145 | atomic_t tmr_complete; | ||
146 | wait_queue_head_t tmr_wait; | ||
147 | }; | ||
148 | |||
149 | struct scsiback_nexus { | ||
150 | /* Pointer to TCM session for I_T Nexus */ | ||
151 | struct se_session *tvn_se_sess; | ||
152 | }; | ||
153 | |||
154 | struct scsiback_tport { | ||
155 | /* SCSI protocol the tport is providing */ | ||
156 | u8 tport_proto_id; | ||
157 | /* Binary World Wide unique Port Name for pvscsi Target port */ | ||
158 | u64 tport_wwpn; | ||
159 | /* ASCII formatted WWPN for pvscsi Target port */ | ||
160 | char tport_name[VSCSI_NAMELEN]; | ||
161 | /* Returned by scsiback_make_tport() */ | ||
162 | struct se_wwn tport_wwn; | ||
163 | }; | ||
164 | |||
165 | struct scsiback_tpg { | ||
166 | /* scsiback port target portal group tag for TCM */ | ||
167 | u16 tport_tpgt; | ||
168 | /* track number of TPG Port/Lun Links wrt explicit I_T Nexus shutdown */ | ||
169 | int tv_tpg_port_count; | ||
170 | /* xen-pvscsi references to tpg_nexus, protected by tv_tpg_mutex */ | ||
171 | int tv_tpg_fe_count; | ||
172 | /* list for scsiback_list */ | ||
173 | struct list_head tv_tpg_list; | ||
174 | /* Used to protect access for tpg_nexus */ | ||
175 | struct mutex tv_tpg_mutex; | ||
176 | /* Pointer to the TCM pvscsi I_T Nexus for this TPG endpoint */ | ||
177 | struct scsiback_nexus *tpg_nexus; | ||
178 | /* Pointer back to scsiback_tport */ | ||
179 | struct scsiback_tport *tport; | ||
180 | /* Returned by scsiback_make_tpg() */ | ||
181 | struct se_portal_group se_tpg; | ||
182 | /* alias used in xenstore */ | ||
183 | char param_alias[VSCSI_NAMELEN]; | ||
184 | /* list of info structures related to this target portal group */ | ||
185 | struct list_head info_list; | ||
186 | }; | ||
187 | |||
188 | #define SCSIBACK_INVALID_HANDLE (~0) | ||
189 | |||
190 | static bool log_print_stat; | ||
191 | module_param(log_print_stat, bool, 0644); | ||
192 | |||
193 | static int scsiback_max_buffer_pages = 1024; | ||
194 | module_param_named(max_buffer_pages, scsiback_max_buffer_pages, int, 0644); | ||
195 | MODULE_PARM_DESC(max_buffer_pages, | ||
196 | "Maximum number of free pages to keep in backend buffer"); | ||
197 | |||
198 | static struct kmem_cache *scsiback_cachep; | ||
199 | static DEFINE_SPINLOCK(free_pages_lock); | ||
200 | static int free_pages_num; | ||
201 | static LIST_HEAD(scsiback_free_pages); | ||
202 | |||
203 | /* Global spinlock to protect scsiback TPG list */ | ||
204 | static DEFINE_MUTEX(scsiback_mutex); | ||
205 | static LIST_HEAD(scsiback_list); | ||
206 | |||
207 | /* Local pointer to allocated TCM configfs fabric module */ | ||
208 | static struct target_fabric_configfs *scsiback_fabric_configfs; | ||
209 | |||
210 | static void scsiback_get(struct vscsibk_info *info) | ||
211 | { | ||
212 | atomic_inc(&info->nr_unreplied_reqs); | ||
213 | } | ||
214 | |||
215 | static void scsiback_put(struct vscsibk_info *info) | ||
216 | { | ||
217 | if (atomic_dec_and_test(&info->nr_unreplied_reqs)) | ||
218 | wake_up(&info->waiting_to_free); | ||
219 | } | ||
220 | |||
221 | static void put_free_pages(struct page **page, int num) | ||
222 | { | ||
223 | unsigned long flags; | ||
224 | int i = free_pages_num + num, n = num; | ||
225 | |||
226 | if (num == 0) | ||
227 | return; | ||
228 | if (i > scsiback_max_buffer_pages) { | ||
229 | n = min(num, i - scsiback_max_buffer_pages); | ||
230 | free_xenballooned_pages(n, page + num - n); | ||
231 | n = num - n; | ||
232 | } | ||
233 | spin_lock_irqsave(&free_pages_lock, flags); | ||
234 | for (i = 0; i < n; i++) | ||
235 | list_add(&page[i]->lru, &scsiback_free_pages); | ||
236 | free_pages_num += n; | ||
237 | spin_unlock_irqrestore(&free_pages_lock, flags); | ||
238 | } | ||
239 | |||
240 | static int get_free_page(struct page **page) | ||
241 | { | ||
242 | unsigned long flags; | ||
243 | |||
244 | spin_lock_irqsave(&free_pages_lock, flags); | ||
245 | if (list_empty(&scsiback_free_pages)) { | ||
246 | spin_unlock_irqrestore(&free_pages_lock, flags); | ||
247 | return alloc_xenballooned_pages(1, page, false); | ||
248 | } | ||
249 | page[0] = list_first_entry(&scsiback_free_pages, struct page, lru); | ||
250 | list_del(&page[0]->lru); | ||
251 | free_pages_num--; | ||
252 | spin_unlock_irqrestore(&free_pages_lock, flags); | ||
253 | return 0; | ||
254 | } | ||
255 | |||
256 | static unsigned long vaddr_page(struct page *page) | ||
257 | { | ||
258 | unsigned long pfn = page_to_pfn(page); | ||
259 | |||
260 | return (unsigned long)pfn_to_kaddr(pfn); | ||
261 | } | ||
262 | |||
263 | static unsigned long vaddr(struct vscsibk_pend *req, int seg) | ||
264 | { | ||
265 | return vaddr_page(req->pages[seg]); | ||
266 | } | ||
267 | |||
268 | static void scsiback_print_status(char *sense_buffer, int errors, | ||
269 | struct vscsibk_pend *pending_req) | ||
270 | { | ||
271 | struct scsiback_tpg *tpg = pending_req->v2p->tpg; | ||
272 | |||
273 | pr_err("xen-pvscsi[%s:%d] cmnd[0]=%02x -> st=%02x msg=%02x host=%02x drv=%02x\n", | ||
274 | tpg->tport->tport_name, pending_req->v2p->lun, | ||
275 | pending_req->cmnd[0], status_byte(errors), msg_byte(errors), | ||
276 | host_byte(errors), driver_byte(errors)); | ||
277 | |||
278 | if (CHECK_CONDITION & status_byte(errors)) | ||
279 | __scsi_print_sense("xen-pvscsi", sense_buffer, | ||
280 | SCSI_SENSE_BUFFERSIZE); | ||
281 | } | ||
282 | |||
283 | static void scsiback_fast_flush_area(struct vscsibk_pend *req) | ||
284 | { | ||
285 | struct gnttab_unmap_grant_ref unmap[VSCSI_GRANT_BATCH]; | ||
286 | struct page *pages[VSCSI_GRANT_BATCH]; | ||
287 | unsigned int i, invcount = 0; | ||
288 | grant_handle_t handle; | ||
289 | int err; | ||
290 | |||
291 | kfree(req->sgl); | ||
292 | req->sgl = NULL; | ||
293 | req->n_sg = 0; | ||
294 | |||
295 | if (!req->n_grants) | ||
296 | return; | ||
297 | |||
298 | for (i = 0; i < req->n_grants; i++) { | ||
299 | handle = req->grant_handles[i]; | ||
300 | if (handle == SCSIBACK_INVALID_HANDLE) | ||
301 | continue; | ||
302 | gnttab_set_unmap_op(&unmap[invcount], vaddr(req, i), | ||
303 | GNTMAP_host_map, handle); | ||
304 | req->grant_handles[i] = SCSIBACK_INVALID_HANDLE; | ||
305 | pages[invcount] = req->pages[i]; | ||
306 | put_page(pages[invcount]); | ||
307 | invcount++; | ||
308 | if (invcount < VSCSI_GRANT_BATCH) | ||
309 | continue; | ||
310 | err = gnttab_unmap_refs(unmap, NULL, pages, invcount); | ||
311 | BUG_ON(err); | ||
312 | invcount = 0; | ||
313 | } | ||
314 | |||
315 | if (invcount) { | ||
316 | err = gnttab_unmap_refs(unmap, NULL, pages, invcount); | ||
317 | BUG_ON(err); | ||
318 | } | ||
319 | |||
320 | put_free_pages(req->pages, req->n_grants); | ||
321 | req->n_grants = 0; | ||
322 | } | ||
323 | |||
324 | static void scsiback_free_translation_entry(struct kref *kref) | ||
325 | { | ||
326 | struct v2p_entry *entry = container_of(kref, struct v2p_entry, kref); | ||
327 | struct scsiback_tpg *tpg = entry->tpg; | ||
328 | |||
329 | mutex_lock(&tpg->tv_tpg_mutex); | ||
330 | tpg->tv_tpg_fe_count--; | ||
331 | mutex_unlock(&tpg->tv_tpg_mutex); | ||
332 | |||
333 | kfree(entry); | ||
334 | } | ||
335 | |||
336 | static void scsiback_do_resp_with_sense(char *sense_buffer, int32_t result, | ||
337 | uint32_t resid, struct vscsibk_pend *pending_req) | ||
338 | { | ||
339 | struct vscsiif_response *ring_res; | ||
340 | struct vscsibk_info *info = pending_req->info; | ||
341 | int notify; | ||
342 | struct scsi_sense_hdr sshdr; | ||
343 | unsigned long flags; | ||
344 | unsigned len; | ||
345 | |||
346 | spin_lock_irqsave(&info->ring_lock, flags); | ||
347 | |||
348 | ring_res = RING_GET_RESPONSE(&info->ring, info->ring.rsp_prod_pvt); | ||
349 | info->ring.rsp_prod_pvt++; | ||
350 | |||
351 | ring_res->rslt = result; | ||
352 | ring_res->rqid = pending_req->rqid; | ||
353 | |||
354 | if (sense_buffer != NULL && | ||
355 | scsi_normalize_sense(sense_buffer, VSCSIIF_SENSE_BUFFERSIZE, | ||
356 | &sshdr)) { | ||
357 | len = min_t(unsigned, 8 + sense_buffer[7], | ||
358 | VSCSIIF_SENSE_BUFFERSIZE); | ||
359 | memcpy(ring_res->sense_buffer, sense_buffer, len); | ||
360 | ring_res->sense_len = len; | ||
361 | } else { | ||
362 | ring_res->sense_len = 0; | ||
363 | } | ||
364 | |||
365 | ring_res->residual_len = resid; | ||
366 | |||
367 | RING_PUSH_RESPONSES_AND_CHECK_NOTIFY(&info->ring, notify); | ||
368 | spin_unlock_irqrestore(&info->ring_lock, flags); | ||
369 | |||
370 | if (notify) | ||
371 | notify_remote_via_irq(info->irq); | ||
372 | |||
373 | if (pending_req->v2p) | ||
374 | kref_put(&pending_req->v2p->kref, | ||
375 | scsiback_free_translation_entry); | ||
376 | } | ||
377 | |||
378 | static void scsiback_cmd_done(struct vscsibk_pend *pending_req) | ||
379 | { | ||
380 | struct vscsibk_info *info = pending_req->info; | ||
381 | unsigned char *sense_buffer; | ||
382 | unsigned int resid; | ||
383 | int errors; | ||
384 | |||
385 | sense_buffer = pending_req->sense_buffer; | ||
386 | resid = pending_req->se_cmd.residual_count; | ||
387 | errors = pending_req->result; | ||
388 | |||
389 | if (errors && log_print_stat) | ||
390 | scsiback_print_status(sense_buffer, errors, pending_req); | ||
391 | |||
392 | scsiback_fast_flush_area(pending_req); | ||
393 | scsiback_do_resp_with_sense(sense_buffer, errors, resid, pending_req); | ||
394 | scsiback_put(info); | ||
395 | } | ||
396 | |||
397 | static void scsiback_cmd_exec(struct vscsibk_pend *pending_req) | ||
398 | { | ||
399 | struct se_cmd *se_cmd = &pending_req->se_cmd; | ||
400 | struct se_session *sess = pending_req->v2p->tpg->tpg_nexus->tvn_se_sess; | ||
401 | int rc; | ||
402 | |||
403 | memset(pending_req->sense_buffer, 0, VSCSIIF_SENSE_BUFFERSIZE); | ||
404 | |||
405 | memset(se_cmd, 0, sizeof(*se_cmd)); | ||
406 | |||
407 | scsiback_get(pending_req->info); | ||
408 | rc = target_submit_cmd_map_sgls(se_cmd, sess, pending_req->cmnd, | ||
409 | pending_req->sense_buffer, pending_req->v2p->lun, | ||
410 | pending_req->data_len, 0, | ||
411 | pending_req->sc_data_direction, 0, | ||
412 | pending_req->sgl, pending_req->n_sg, | ||
413 | NULL, 0, NULL, 0); | ||
414 | if (rc < 0) { | ||
415 | transport_send_check_condition_and_sense(se_cmd, | ||
416 | TCM_LOGICAL_UNIT_COMMUNICATION_FAILURE, 0); | ||
417 | transport_generic_free_cmd(se_cmd, 0); | ||
418 | } | ||
419 | } | ||
420 | |||
421 | static int scsiback_gnttab_data_map_batch(struct gnttab_map_grant_ref *map, | ||
422 | struct page **pg, grant_handle_t *grant, int cnt) | ||
423 | { | ||
424 | int err, i; | ||
425 | |||
426 | if (!cnt) | ||
427 | return 0; | ||
428 | |||
429 | err = gnttab_map_refs(map, NULL, pg, cnt); | ||
430 | BUG_ON(err); | ||
431 | for (i = 0; i < cnt; i++) { | ||
432 | if (unlikely(map[i].status != GNTST_okay)) { | ||
433 | pr_err("xen-pvscsi: invalid buffer -- could not remap it\n"); | ||
434 | map[i].handle = SCSIBACK_INVALID_HANDLE; | ||
435 | err = -ENOMEM; | ||
436 | } else { | ||
437 | get_page(pg[i]); | ||
438 | } | ||
439 | grant[i] = map[i].handle; | ||
440 | } | ||
441 | return err; | ||
442 | } | ||
443 | |||
444 | static int scsiback_gnttab_data_map_list(struct vscsibk_pend *pending_req, | ||
445 | struct scsiif_request_segment *seg, struct page **pg, | ||
446 | grant_handle_t *grant, int cnt, u32 flags) | ||
447 | { | ||
448 | int mapcount = 0, i, err = 0; | ||
449 | struct gnttab_map_grant_ref map[VSCSI_GRANT_BATCH]; | ||
450 | struct vscsibk_info *info = pending_req->info; | ||
451 | |||
452 | for (i = 0; i < cnt; i++) { | ||
453 | if (get_free_page(pg + mapcount)) { | ||
454 | put_free_pages(pg, mapcount); | ||
455 | pr_err("xen-pvscsi: no grant page\n"); | ||
456 | return -ENOMEM; | ||
457 | } | ||
458 | gnttab_set_map_op(&map[mapcount], vaddr_page(pg[mapcount]), | ||
459 | flags, seg[i].gref, info->domid); | ||
460 | mapcount++; | ||
461 | if (mapcount < VSCSI_GRANT_BATCH) | ||
462 | continue; | ||
463 | err = scsiback_gnttab_data_map_batch(map, pg, grant, mapcount); | ||
464 | pg += mapcount; | ||
465 | grant += mapcount; | ||
466 | pending_req->n_grants += mapcount; | ||
467 | if (err) | ||
468 | return err; | ||
469 | mapcount = 0; | ||
470 | } | ||
471 | err = scsiback_gnttab_data_map_batch(map, pg, grant, mapcount); | ||
472 | pending_req->n_grants += mapcount; | ||
473 | return err; | ||
474 | } | ||
475 | |||
476 | static int scsiback_gnttab_data_map(struct vscsiif_request *ring_req, | ||
477 | struct vscsibk_pend *pending_req) | ||
478 | { | ||
479 | u32 flags; | ||
480 | int i, err, n_segs, i_seg = 0; | ||
481 | struct page **pg; | ||
482 | struct scsiif_request_segment *seg; | ||
483 | unsigned long end_seg = 0; | ||
484 | unsigned int nr_segments = (unsigned int)ring_req->nr_segments; | ||
485 | unsigned int nr_sgl = 0; | ||
486 | struct scatterlist *sg; | ||
487 | grant_handle_t *grant; | ||
488 | |||
489 | pending_req->n_sg = 0; | ||
490 | pending_req->n_grants = 0; | ||
491 | pending_req->data_len = 0; | ||
492 | |||
493 | nr_segments &= ~VSCSIIF_SG_GRANT; | ||
494 | if (!nr_segments) | ||
495 | return 0; | ||
496 | |||
497 | if (nr_segments > VSCSIIF_SG_TABLESIZE) { | ||
498 | DPRINTK("xen-pvscsi: invalid parameter nr_seg = %d\n", | ||
499 | ring_req->nr_segments); | ||
500 | return -EINVAL; | ||
501 | } | ||
502 | |||
503 | if (ring_req->nr_segments & VSCSIIF_SG_GRANT) { | ||
504 | err = scsiback_gnttab_data_map_list(pending_req, ring_req->seg, | ||
505 | pending_req->pages, pending_req->grant_handles, | ||
506 | nr_segments, GNTMAP_host_map | GNTMAP_readonly); | ||
507 | if (err) | ||
508 | return err; | ||
509 | nr_sgl = nr_segments; | ||
510 | nr_segments = 0; | ||
511 | for (i = 0; i < nr_sgl; i++) { | ||
512 | n_segs = ring_req->seg[i].length / | ||
513 | sizeof(struct scsiif_request_segment); | ||
514 | if ((unsigned)ring_req->seg[i].offset + | ||
515 | (unsigned)ring_req->seg[i].length > PAGE_SIZE || | ||
516 | n_segs * sizeof(struct scsiif_request_segment) != | ||
517 | ring_req->seg[i].length) | ||
518 | return -EINVAL; | ||
519 | nr_segments += n_segs; | ||
520 | } | ||
521 | if (nr_segments > SG_ALL) { | ||
522 | DPRINTK("xen-pvscsi: invalid nr_seg = %d\n", | ||
523 | nr_segments); | ||
524 | return -EINVAL; | ||
525 | } | ||
526 | } | ||
527 | |||
528 | /* free of (sgl) in fast_flush_area()*/ | ||
529 | pending_req->sgl = kmalloc_array(nr_segments, | ||
530 | sizeof(struct scatterlist), GFP_KERNEL); | ||
531 | if (!pending_req->sgl) | ||
532 | return -ENOMEM; | ||
533 | |||
534 | sg_init_table(pending_req->sgl, nr_segments); | ||
535 | pending_req->n_sg = nr_segments; | ||
536 | |||
537 | flags = GNTMAP_host_map; | ||
538 | if (pending_req->sc_data_direction == DMA_TO_DEVICE) | ||
539 | flags |= GNTMAP_readonly; | ||
540 | |||
541 | pg = pending_req->pages + nr_sgl; | ||
542 | grant = pending_req->grant_handles + nr_sgl; | ||
543 | if (!nr_sgl) { | ||
544 | seg = ring_req->seg; | ||
545 | err = scsiback_gnttab_data_map_list(pending_req, seg, | ||
546 | pg, grant, nr_segments, flags); | ||
547 | if (err) | ||
548 | return err; | ||
549 | } else { | ||
550 | for (i = 0; i < nr_sgl; i++) { | ||
551 | seg = (struct scsiif_request_segment *)( | ||
552 | vaddr(pending_req, i) + ring_req->seg[i].offset); | ||
553 | n_segs = ring_req->seg[i].length / | ||
554 | sizeof(struct scsiif_request_segment); | ||
555 | err = scsiback_gnttab_data_map_list(pending_req, seg, | ||
556 | pg, grant, n_segs, flags); | ||
557 | if (err) | ||
558 | return err; | ||
559 | pg += n_segs; | ||
560 | grant += n_segs; | ||
561 | } | ||
562 | end_seg = vaddr(pending_req, 0) + ring_req->seg[0].offset; | ||
563 | seg = (struct scsiif_request_segment *)end_seg; | ||
564 | end_seg += ring_req->seg[0].length; | ||
565 | pg = pending_req->pages + nr_sgl; | ||
566 | } | ||
567 | |||
568 | for_each_sg(pending_req->sgl, sg, nr_segments, i) { | ||
569 | sg_set_page(sg, pg[i], seg->length, seg->offset); | ||
570 | pending_req->data_len += seg->length; | ||
571 | seg++; | ||
572 | if (nr_sgl && (unsigned long)seg >= end_seg) { | ||
573 | i_seg++; | ||
574 | end_seg = vaddr(pending_req, i_seg) + | ||
575 | ring_req->seg[i_seg].offset; | ||
576 | seg = (struct scsiif_request_segment *)end_seg; | ||
577 | end_seg += ring_req->seg[i_seg].length; | ||
578 | } | ||
579 | if (sg->offset >= PAGE_SIZE || | ||
580 | sg->length > PAGE_SIZE || | ||
581 | sg->offset + sg->length > PAGE_SIZE) | ||
582 | return -EINVAL; | ||
583 | } | ||
584 | |||
585 | return 0; | ||
586 | } | ||
587 | |||
588 | static void scsiback_disconnect(struct vscsibk_info *info) | ||
589 | { | ||
590 | wait_event(info->waiting_to_free, | ||
591 | atomic_read(&info->nr_unreplied_reqs) == 0); | ||
592 | |||
593 | unbind_from_irqhandler(info->irq, info); | ||
594 | info->irq = 0; | ||
595 | xenbus_unmap_ring_vfree(info->dev, info->ring.sring); | ||
596 | } | ||
597 | |||
598 | static void scsiback_device_action(struct vscsibk_pend *pending_req, | ||
599 | enum tcm_tmreq_table act, int tag) | ||
600 | { | ||
601 | int rc, err = FAILED; | ||
602 | struct scsiback_tpg *tpg = pending_req->v2p->tpg; | ||
603 | struct se_cmd *se_cmd = &pending_req->se_cmd; | ||
604 | struct scsiback_tmr *tmr; | ||
605 | |||
606 | tmr = kzalloc(sizeof(struct scsiback_tmr), GFP_KERNEL); | ||
607 | if (!tmr) | ||
608 | goto out; | ||
609 | |||
610 | init_waitqueue_head(&tmr->tmr_wait); | ||
611 | |||
612 | transport_init_se_cmd(se_cmd, tpg->se_tpg.se_tpg_tfo, | ||
613 | tpg->tpg_nexus->tvn_se_sess, 0, DMA_NONE, MSG_SIMPLE_TAG, | ||
614 | &pending_req->sense_buffer[0]); | ||
615 | |||
616 | rc = core_tmr_alloc_req(se_cmd, tmr, act, GFP_KERNEL); | ||
617 | if (rc < 0) | ||
618 | goto out; | ||
619 | |||
620 | se_cmd->se_tmr_req->ref_task_tag = tag; | ||
621 | |||
622 | if (transport_lookup_tmr_lun(se_cmd, pending_req->v2p->lun) < 0) | ||
623 | goto out; | ||
624 | |||
625 | transport_generic_handle_tmr(se_cmd); | ||
626 | wait_event(tmr->tmr_wait, atomic_read(&tmr->tmr_complete)); | ||
627 | |||
628 | err = (se_cmd->se_tmr_req->response == TMR_FUNCTION_COMPLETE) ? | ||
629 | SUCCESS : FAILED; | ||
630 | |||
631 | out: | ||
632 | if (tmr) { | ||
633 | transport_generic_free_cmd(&pending_req->se_cmd, 1); | ||
634 | kfree(tmr); | ||
635 | } | ||
636 | |||
637 | scsiback_do_resp_with_sense(NULL, err, 0, pending_req); | ||
638 | |||
639 | kmem_cache_free(scsiback_cachep, pending_req); | ||
640 | } | ||
641 | |||
642 | /* | ||
643 | Perform virtual to physical translation | ||
644 | */ | ||
645 | static struct v2p_entry *scsiback_do_translation(struct vscsibk_info *info, | ||
646 | struct ids_tuple *v) | ||
647 | { | ||
648 | struct v2p_entry *entry; | ||
649 | struct list_head *head = &(info->v2p_entry_lists); | ||
650 | unsigned long flags; | ||
651 | |||
652 | spin_lock_irqsave(&info->v2p_lock, flags); | ||
653 | list_for_each_entry(entry, head, l) { | ||
654 | if ((entry->v.chn == v->chn) && | ||
655 | (entry->v.tgt == v->tgt) && | ||
656 | (entry->v.lun == v->lun)) { | ||
657 | kref_get(&entry->kref); | ||
658 | goto out; | ||
659 | } | ||
660 | } | ||
661 | entry = NULL; | ||
662 | |||
663 | out: | ||
664 | spin_unlock_irqrestore(&info->v2p_lock, flags); | ||
665 | return entry; | ||
666 | } | ||
667 | |||
668 | static int prepare_pending_reqs(struct vscsibk_info *info, | ||
669 | struct vscsiif_request *ring_req, | ||
670 | struct vscsibk_pend *pending_req) | ||
671 | { | ||
672 | struct v2p_entry *v2p; | ||
673 | struct ids_tuple vir; | ||
674 | |||
675 | pending_req->rqid = ring_req->rqid; | ||
676 | pending_req->info = info; | ||
677 | |||
678 | vir.chn = ring_req->channel; | ||
679 | vir.tgt = ring_req->id; | ||
680 | vir.lun = ring_req->lun; | ||
681 | |||
682 | v2p = scsiback_do_translation(info, &vir); | ||
683 | if (!v2p) { | ||
684 | pending_req->v2p = NULL; | ||
685 | DPRINTK("xen-pvscsi: doesn't exist.\n"); | ||
686 | return -ENODEV; | ||
687 | } | ||
688 | pending_req->v2p = v2p; | ||
689 | |||
690 | /* request range check from frontend */ | ||
691 | pending_req->sc_data_direction = ring_req->sc_data_direction; | ||
692 | if ((pending_req->sc_data_direction != DMA_BIDIRECTIONAL) && | ||
693 | (pending_req->sc_data_direction != DMA_TO_DEVICE) && | ||
694 | (pending_req->sc_data_direction != DMA_FROM_DEVICE) && | ||
695 | (pending_req->sc_data_direction != DMA_NONE)) { | ||
696 | DPRINTK("xen-pvscsi: invalid parameter data_dir = %d\n", | ||
697 | pending_req->sc_data_direction); | ||
698 | return -EINVAL; | ||
699 | } | ||
700 | |||
701 | pending_req->cmd_len = ring_req->cmd_len; | ||
702 | if (pending_req->cmd_len > VSCSIIF_MAX_COMMAND_SIZE) { | ||
703 | DPRINTK("xen-pvscsi: invalid parameter cmd_len = %d\n", | ||
704 | pending_req->cmd_len); | ||
705 | return -EINVAL; | ||
706 | } | ||
707 | memcpy(pending_req->cmnd, ring_req->cmnd, pending_req->cmd_len); | ||
708 | |||
709 | return 0; | ||
710 | } | ||
711 | |||
712 | static int scsiback_do_cmd_fn(struct vscsibk_info *info) | ||
713 | { | ||
714 | struct vscsiif_back_ring *ring = &info->ring; | ||
715 | struct vscsiif_request *ring_req; | ||
716 | struct vscsibk_pend *pending_req; | ||
717 | RING_IDX rc, rp; | ||
718 | int err, more_to_do; | ||
719 | uint32_t result; | ||
720 | uint8_t act; | ||
721 | |||
722 | rc = ring->req_cons; | ||
723 | rp = ring->sring->req_prod; | ||
724 | rmb(); /* guest system is accessing ring, too */ | ||
725 | |||
726 | if (RING_REQUEST_PROD_OVERFLOW(ring, rp)) { | ||
727 | rc = ring->rsp_prod_pvt; | ||
728 | pr_warn("xen-pvscsi: Dom%d provided bogus ring requests (%#x - %#x = %u). Halting ring processing\n", | ||
729 | info->domid, rp, rc, rp - rc); | ||
730 | info->ring_error = 1; | ||
731 | return 0; | ||
732 | } | ||
733 | |||
734 | while ((rc != rp)) { | ||
735 | if (RING_REQUEST_CONS_OVERFLOW(ring, rc)) | ||
736 | break; | ||
737 | pending_req = kmem_cache_alloc(scsiback_cachep, GFP_KERNEL); | ||
738 | if (!pending_req) | ||
739 | return 1; | ||
740 | |||
741 | ring_req = RING_GET_REQUEST(ring, rc); | ||
742 | ring->req_cons = ++rc; | ||
743 | |||
744 | act = ring_req->act; | ||
745 | err = prepare_pending_reqs(info, ring_req, pending_req); | ||
746 | if (err) { | ||
747 | switch (err) { | ||
748 | case -ENODEV: | ||
749 | result = DID_NO_CONNECT; | ||
750 | break; | ||
751 | default: | ||
752 | result = DRIVER_ERROR; | ||
753 | break; | ||
754 | } | ||
755 | scsiback_do_resp_with_sense(NULL, result << 24, 0, | ||
756 | pending_req); | ||
757 | kmem_cache_free(scsiback_cachep, pending_req); | ||
758 | return 1; | ||
759 | } | ||
760 | |||
761 | switch (act) { | ||
762 | case VSCSIIF_ACT_SCSI_CDB: | ||
763 | if (scsiback_gnttab_data_map(ring_req, pending_req)) { | ||
764 | scsiback_fast_flush_area(pending_req); | ||
765 | scsiback_do_resp_with_sense(NULL, | ||
766 | DRIVER_ERROR << 24, 0, pending_req); | ||
767 | kmem_cache_free(scsiback_cachep, pending_req); | ||
768 | } else { | ||
769 | scsiback_cmd_exec(pending_req); | ||
770 | } | ||
771 | break; | ||
772 | case VSCSIIF_ACT_SCSI_ABORT: | ||
773 | scsiback_device_action(pending_req, TMR_ABORT_TASK, | ||
774 | ring_req->ref_rqid); | ||
775 | break; | ||
776 | case VSCSIIF_ACT_SCSI_RESET: | ||
777 | scsiback_device_action(pending_req, TMR_LUN_RESET, 0); | ||
778 | break; | ||
779 | default: | ||
780 | pr_err_ratelimited("xen-pvscsi: invalid request\n"); | ||
781 | scsiback_do_resp_with_sense(NULL, DRIVER_ERROR << 24, | ||
782 | 0, pending_req); | ||
783 | kmem_cache_free(scsiback_cachep, pending_req); | ||
784 | break; | ||
785 | } | ||
786 | |||
787 | /* Yield point for this unbounded loop. */ | ||
788 | cond_resched(); | ||
789 | } | ||
790 | |||
791 | RING_FINAL_CHECK_FOR_REQUESTS(&info->ring, more_to_do); | ||
792 | return more_to_do; | ||
793 | } | ||
794 | |||
795 | static irqreturn_t scsiback_irq_fn(int irq, void *dev_id) | ||
796 | { | ||
797 | struct vscsibk_info *info = dev_id; | ||
798 | |||
799 | if (info->ring_error) | ||
800 | return IRQ_HANDLED; | ||
801 | |||
802 | while (scsiback_do_cmd_fn(info)) | ||
803 | cond_resched(); | ||
804 | |||
805 | return IRQ_HANDLED; | ||
806 | } | ||
807 | |||
808 | static int scsiback_init_sring(struct vscsibk_info *info, grant_ref_t ring_ref, | ||
809 | evtchn_port_t evtchn) | ||
810 | { | ||
811 | void *area; | ||
812 | struct vscsiif_sring *sring; | ||
813 | int err; | ||
814 | |||
815 | if (info->irq) | ||
816 | return -1; | ||
817 | |||
818 | err = xenbus_map_ring_valloc(info->dev, ring_ref, &area); | ||
819 | if (err) | ||
820 | return err; | ||
821 | |||
822 | sring = (struct vscsiif_sring *)area; | ||
823 | BACK_RING_INIT(&info->ring, sring, PAGE_SIZE); | ||
824 | |||
825 | err = bind_interdomain_evtchn_to_irq(info->domid, evtchn); | ||
826 | if (err < 0) | ||
827 | goto unmap_page; | ||
828 | |||
829 | info->irq = err; | ||
830 | |||
831 | err = request_threaded_irq(info->irq, NULL, scsiback_irq_fn, | ||
832 | IRQF_ONESHOT, "vscsiif-backend", info); | ||
833 | if (err) | ||
834 | goto free_irq; | ||
835 | |||
836 | return 0; | ||
837 | |||
838 | free_irq: | ||
839 | unbind_from_irqhandler(info->irq, info); | ||
840 | info->irq = 0; | ||
841 | unmap_page: | ||
842 | xenbus_unmap_ring_vfree(info->dev, area); | ||
843 | |||
844 | return err; | ||
845 | } | ||
846 | |||
847 | static int scsiback_map(struct vscsibk_info *info) | ||
848 | { | ||
849 | struct xenbus_device *dev = info->dev; | ||
850 | unsigned int ring_ref, evtchn; | ||
851 | int err; | ||
852 | |||
853 | err = xenbus_gather(XBT_NIL, dev->otherend, | ||
854 | "ring-ref", "%u", &ring_ref, | ||
855 | "event-channel", "%u", &evtchn, NULL); | ||
856 | if (err) { | ||
857 | xenbus_dev_fatal(dev, err, "reading %s ring", dev->otherend); | ||
858 | return err; | ||
859 | } | ||
860 | |||
861 | return scsiback_init_sring(info, ring_ref, evtchn); | ||
862 | } | ||
863 | |||
864 | /* | ||
865 | Add a new translation entry | ||
866 | */ | ||
867 | static int scsiback_add_translation_entry(struct vscsibk_info *info, | ||
868 | char *phy, struct ids_tuple *v) | ||
869 | { | ||
870 | int err = 0; | ||
871 | struct v2p_entry *entry; | ||
872 | struct v2p_entry *new; | ||
873 | struct list_head *head = &(info->v2p_entry_lists); | ||
874 | unsigned long flags; | ||
875 | char *lunp; | ||
876 | unsigned int lun; | ||
877 | struct scsiback_tpg *tpg_entry, *tpg = NULL; | ||
878 | char *error = "doesn't exist"; | ||
879 | |||
880 | lunp = strrchr(phy, ':'); | ||
881 | if (!lunp) { | ||
882 | pr_err("xen-pvscsi: illegal format of physical device %s\n", | ||
883 | phy); | ||
884 | return -EINVAL; | ||
885 | } | ||
886 | *lunp = 0; | ||
887 | lunp++; | ||
888 | if (kstrtouint(lunp, 10, &lun) || lun >= TRANSPORT_MAX_LUNS_PER_TPG) { | ||
889 | pr_err("xen-pvscsi: lun number not valid: %s\n", lunp); | ||
890 | return -EINVAL; | ||
891 | } | ||
892 | |||
893 | mutex_lock(&scsiback_mutex); | ||
894 | list_for_each_entry(tpg_entry, &scsiback_list, tv_tpg_list) { | ||
895 | if (!strcmp(phy, tpg_entry->tport->tport_name) || | ||
896 | !strcmp(phy, tpg_entry->param_alias)) { | ||
897 | spin_lock(&tpg_entry->se_tpg.tpg_lun_lock); | ||
898 | if (tpg_entry->se_tpg.tpg_lun_list[lun]->lun_status == | ||
899 | TRANSPORT_LUN_STATUS_ACTIVE) { | ||
900 | if (!tpg_entry->tpg_nexus) | ||
901 | error = "nexus undefined"; | ||
902 | else | ||
903 | tpg = tpg_entry; | ||
904 | } | ||
905 | spin_unlock(&tpg_entry->se_tpg.tpg_lun_lock); | ||
906 | break; | ||
907 | } | ||
908 | } | ||
909 | if (tpg) { | ||
910 | mutex_lock(&tpg->tv_tpg_mutex); | ||
911 | tpg->tv_tpg_fe_count++; | ||
912 | mutex_unlock(&tpg->tv_tpg_mutex); | ||
913 | } | ||
914 | mutex_unlock(&scsiback_mutex); | ||
915 | |||
916 | if (!tpg) { | ||
917 | pr_err("xen-pvscsi: %s:%d %s\n", phy, lun, error); | ||
918 | return -ENODEV; | ||
919 | } | ||
920 | |||
921 | new = kmalloc(sizeof(struct v2p_entry), GFP_KERNEL); | ||
922 | if (new == NULL) { | ||
923 | err = -ENOMEM; | ||
924 | goto out_free; | ||
925 | } | ||
926 | |||
927 | spin_lock_irqsave(&info->v2p_lock, flags); | ||
928 | |||
929 | /* Check double assignment to identical virtual ID */ | ||
930 | list_for_each_entry(entry, head, l) { | ||
931 | if ((entry->v.chn == v->chn) && | ||
932 | (entry->v.tgt == v->tgt) && | ||
933 | (entry->v.lun == v->lun)) { | ||
934 | pr_warn("xen-pvscsi: Virtual ID is already used. Assignment was not performed.\n"); | ||
935 | err = -EEXIST; | ||
936 | goto out; | ||
937 | } | ||
938 | |||
939 | } | ||
940 | |||
941 | /* Create a new translation entry and add to the list */ | ||
942 | kref_init(&new->kref); | ||
943 | new->v = *v; | ||
944 | new->tpg = tpg; | ||
945 | new->lun = lun; | ||
946 | list_add_tail(&new->l, head); | ||
947 | |||
948 | out: | ||
949 | spin_unlock_irqrestore(&info->v2p_lock, flags); | ||
950 | |||
951 | out_free: | ||
952 | mutex_lock(&tpg->tv_tpg_mutex); | ||
953 | tpg->tv_tpg_fe_count--; | ||
954 | mutex_unlock(&tpg->tv_tpg_mutex); | ||
955 | |||
956 | if (err) | ||
957 | kfree(new); | ||
958 | |||
959 | return err; | ||
960 | } | ||
961 | |||
962 | static void __scsiback_del_translation_entry(struct v2p_entry *entry) | ||
963 | { | ||
964 | list_del(&entry->l); | ||
965 | kref_put(&entry->kref, scsiback_free_translation_entry); | ||
966 | } | ||
967 | |||
968 | /* | ||
969 | Delete the translation entry specfied | ||
970 | */ | ||
971 | static int scsiback_del_translation_entry(struct vscsibk_info *info, | ||
972 | struct ids_tuple *v) | ||
973 | { | ||
974 | struct v2p_entry *entry; | ||
975 | struct list_head *head = &(info->v2p_entry_lists); | ||
976 | unsigned long flags; | ||
977 | |||
978 | spin_lock_irqsave(&info->v2p_lock, flags); | ||
979 | /* Find out the translation entry specified */ | ||
980 | list_for_each_entry(entry, head, l) { | ||
981 | if ((entry->v.chn == v->chn) && | ||
982 | (entry->v.tgt == v->tgt) && | ||
983 | (entry->v.lun == v->lun)) { | ||
984 | goto found; | ||
985 | } | ||
986 | } | ||
987 | |||
988 | spin_unlock_irqrestore(&info->v2p_lock, flags); | ||
989 | return 1; | ||
990 | |||
991 | found: | ||
992 | /* Delete the translation entry specfied */ | ||
993 | __scsiback_del_translation_entry(entry); | ||
994 | |||
995 | spin_unlock_irqrestore(&info->v2p_lock, flags); | ||
996 | return 0; | ||
997 | } | ||
998 | |||
999 | static void scsiback_do_add_lun(struct vscsibk_info *info, const char *state, | ||
1000 | char *phy, struct ids_tuple *vir) | ||
1001 | { | ||
1002 | if (!scsiback_add_translation_entry(info, phy, vir)) { | ||
1003 | if (xenbus_printf(XBT_NIL, info->dev->nodename, state, | ||
1004 | "%d", XenbusStateInitialised)) { | ||
1005 | pr_err("xen-pvscsi: xenbus_printf error %s\n", state); | ||
1006 | scsiback_del_translation_entry(info, vir); | ||
1007 | } | ||
1008 | } else { | ||
1009 | xenbus_printf(XBT_NIL, info->dev->nodename, state, | ||
1010 | "%d", XenbusStateClosed); | ||
1011 | } | ||
1012 | } | ||
1013 | |||
1014 | static void scsiback_do_del_lun(struct vscsibk_info *info, const char *state, | ||
1015 | struct ids_tuple *vir) | ||
1016 | { | ||
1017 | if (!scsiback_del_translation_entry(info, vir)) { | ||
1018 | if (xenbus_printf(XBT_NIL, info->dev->nodename, state, | ||
1019 | "%d", XenbusStateClosed)) | ||
1020 | pr_err("xen-pvscsi: xenbus_printf error %s\n", state); | ||
1021 | } | ||
1022 | } | ||
1023 | |||
1024 | #define VSCSIBACK_OP_ADD_OR_DEL_LUN 1 | ||
1025 | #define VSCSIBACK_OP_UPDATEDEV_STATE 2 | ||
1026 | |||
1027 | static void scsiback_do_1lun_hotplug(struct vscsibk_info *info, int op, | ||
1028 | char *ent) | ||
1029 | { | ||
1030 | int err; | ||
1031 | struct ids_tuple vir; | ||
1032 | char *val; | ||
1033 | int device_state; | ||
1034 | char phy[VSCSI_NAMELEN]; | ||
1035 | char str[64]; | ||
1036 | char state[64]; | ||
1037 | struct xenbus_device *dev = info->dev; | ||
1038 | |||
1039 | /* read status */ | ||
1040 | snprintf(state, sizeof(state), "vscsi-devs/%s/state", ent); | ||
1041 | err = xenbus_scanf(XBT_NIL, dev->nodename, state, "%u", &device_state); | ||
1042 | if (XENBUS_EXIST_ERR(err)) | ||
1043 | return; | ||
1044 | |||
1045 | /* physical SCSI device */ | ||
1046 | snprintf(str, sizeof(str), "vscsi-devs/%s/p-dev", ent); | ||
1047 | val = xenbus_read(XBT_NIL, dev->nodename, str, NULL); | ||
1048 | if (IS_ERR(val)) { | ||
1049 | xenbus_printf(XBT_NIL, dev->nodename, state, | ||
1050 | "%d", XenbusStateClosed); | ||
1051 | return; | ||
1052 | } | ||
1053 | strlcpy(phy, val, VSCSI_NAMELEN); | ||
1054 | kfree(val); | ||
1055 | |||
1056 | /* virtual SCSI device */ | ||
1057 | snprintf(str, sizeof(str), "vscsi-devs/%s/v-dev", ent); | ||
1058 | err = xenbus_scanf(XBT_NIL, dev->nodename, str, "%u:%u:%u:%u", | ||
1059 | &vir.hst, &vir.chn, &vir.tgt, &vir.lun); | ||
1060 | if (XENBUS_EXIST_ERR(err)) { | ||
1061 | xenbus_printf(XBT_NIL, dev->nodename, state, | ||
1062 | "%d", XenbusStateClosed); | ||
1063 | return; | ||
1064 | } | ||
1065 | |||
1066 | switch (op) { | ||
1067 | case VSCSIBACK_OP_ADD_OR_DEL_LUN: | ||
1068 | if (device_state == XenbusStateInitialising) | ||
1069 | scsiback_do_add_lun(info, state, phy, &vir); | ||
1070 | if (device_state == XenbusStateClosing) | ||
1071 | scsiback_do_del_lun(info, state, &vir); | ||
1072 | break; | ||
1073 | |||
1074 | case VSCSIBACK_OP_UPDATEDEV_STATE: | ||
1075 | if (device_state == XenbusStateInitialised) { | ||
1076 | /* modify vscsi-devs/dev-x/state */ | ||
1077 | if (xenbus_printf(XBT_NIL, dev->nodename, state, | ||
1078 | "%d", XenbusStateConnected)) { | ||
1079 | pr_err("xen-pvscsi: xenbus_printf error %s\n", | ||
1080 | str); | ||
1081 | scsiback_del_translation_entry(info, &vir); | ||
1082 | xenbus_printf(XBT_NIL, dev->nodename, state, | ||
1083 | "%d", XenbusStateClosed); | ||
1084 | } | ||
1085 | } | ||
1086 | break; | ||
1087 | /*When it is necessary, processing is added here.*/ | ||
1088 | default: | ||
1089 | break; | ||
1090 | } | ||
1091 | } | ||
1092 | |||
1093 | static void scsiback_do_lun_hotplug(struct vscsibk_info *info, int op) | ||
1094 | { | ||
1095 | int i; | ||
1096 | char **dir; | ||
1097 | unsigned int ndir = 0; | ||
1098 | |||
1099 | dir = xenbus_directory(XBT_NIL, info->dev->nodename, "vscsi-devs", | ||
1100 | &ndir); | ||
1101 | if (IS_ERR(dir)) | ||
1102 | return; | ||
1103 | |||
1104 | for (i = 0; i < ndir; i++) | ||
1105 | scsiback_do_1lun_hotplug(info, op, dir[i]); | ||
1106 | |||
1107 | kfree(dir); | ||
1108 | } | ||
1109 | |||
1110 | static void scsiback_frontend_changed(struct xenbus_device *dev, | ||
1111 | enum xenbus_state frontend_state) | ||
1112 | { | ||
1113 | struct vscsibk_info *info = dev_get_drvdata(&dev->dev); | ||
1114 | |||
1115 | switch (frontend_state) { | ||
1116 | case XenbusStateInitialising: | ||
1117 | break; | ||
1118 | |||
1119 | case XenbusStateInitialised: | ||
1120 | if (scsiback_map(info)) | ||
1121 | break; | ||
1122 | |||
1123 | scsiback_do_lun_hotplug(info, VSCSIBACK_OP_ADD_OR_DEL_LUN); | ||
1124 | xenbus_switch_state(dev, XenbusStateConnected); | ||
1125 | break; | ||
1126 | |||
1127 | case XenbusStateConnected: | ||
1128 | scsiback_do_lun_hotplug(info, VSCSIBACK_OP_UPDATEDEV_STATE); | ||
1129 | |||
1130 | if (dev->state == XenbusStateConnected) | ||
1131 | break; | ||
1132 | |||
1133 | xenbus_switch_state(dev, XenbusStateConnected); | ||
1134 | break; | ||
1135 | |||
1136 | case XenbusStateClosing: | ||
1137 | if (info->irq) | ||
1138 | scsiback_disconnect(info); | ||
1139 | |||
1140 | xenbus_switch_state(dev, XenbusStateClosing); | ||
1141 | break; | ||
1142 | |||
1143 | case XenbusStateClosed: | ||
1144 | xenbus_switch_state(dev, XenbusStateClosed); | ||
1145 | if (xenbus_dev_is_online(dev)) | ||
1146 | break; | ||
1147 | /* fall through if not online */ | ||
1148 | case XenbusStateUnknown: | ||
1149 | device_unregister(&dev->dev); | ||
1150 | break; | ||
1151 | |||
1152 | case XenbusStateReconfiguring: | ||
1153 | scsiback_do_lun_hotplug(info, VSCSIBACK_OP_ADD_OR_DEL_LUN); | ||
1154 | xenbus_switch_state(dev, XenbusStateReconfigured); | ||
1155 | |||
1156 | break; | ||
1157 | |||
1158 | default: | ||
1159 | xenbus_dev_fatal(dev, -EINVAL, "saw state %d at frontend", | ||
1160 | frontend_state); | ||
1161 | break; | ||
1162 | } | ||
1163 | } | ||
1164 | |||
1165 | /* | ||
1166 | Release the translation entry specfied | ||
1167 | */ | ||
1168 | static void scsiback_release_translation_entry(struct vscsibk_info *info) | ||
1169 | { | ||
1170 | struct v2p_entry *entry, *tmp; | ||
1171 | struct list_head *head = &(info->v2p_entry_lists); | ||
1172 | unsigned long flags; | ||
1173 | |||
1174 | spin_lock_irqsave(&info->v2p_lock, flags); | ||
1175 | |||
1176 | list_for_each_entry_safe(entry, tmp, head, l) | ||
1177 | __scsiback_del_translation_entry(entry); | ||
1178 | |||
1179 | spin_unlock_irqrestore(&info->v2p_lock, flags); | ||
1180 | } | ||
1181 | |||
1182 | static int scsiback_remove(struct xenbus_device *dev) | ||
1183 | { | ||
1184 | struct vscsibk_info *info = dev_get_drvdata(&dev->dev); | ||
1185 | |||
1186 | if (info->irq) | ||
1187 | scsiback_disconnect(info); | ||
1188 | |||
1189 | scsiback_release_translation_entry(info); | ||
1190 | |||
1191 | dev_set_drvdata(&dev->dev, NULL); | ||
1192 | |||
1193 | return 0; | ||
1194 | } | ||
1195 | |||
1196 | static int scsiback_probe(struct xenbus_device *dev, | ||
1197 | const struct xenbus_device_id *id) | ||
1198 | { | ||
1199 | int err; | ||
1200 | |||
1201 | struct vscsibk_info *info = kzalloc(sizeof(struct vscsibk_info), | ||
1202 | GFP_KERNEL); | ||
1203 | |||
1204 | DPRINTK("%p %d\n", dev, dev->otherend_id); | ||
1205 | |||
1206 | if (!info) { | ||
1207 | xenbus_dev_fatal(dev, -ENOMEM, "allocating backend structure"); | ||
1208 | return -ENOMEM; | ||
1209 | } | ||
1210 | info->dev = dev; | ||
1211 | dev_set_drvdata(&dev->dev, info); | ||
1212 | |||
1213 | info->domid = dev->otherend_id; | ||
1214 | spin_lock_init(&info->ring_lock); | ||
1215 | info->ring_error = 0; | ||
1216 | atomic_set(&info->nr_unreplied_reqs, 0); | ||
1217 | init_waitqueue_head(&info->waiting_to_free); | ||
1218 | info->dev = dev; | ||
1219 | info->irq = 0; | ||
1220 | INIT_LIST_HEAD(&info->v2p_entry_lists); | ||
1221 | spin_lock_init(&info->v2p_lock); | ||
1222 | |||
1223 | err = xenbus_printf(XBT_NIL, dev->nodename, "feature-sg-grant", "%u", | ||
1224 | SG_ALL); | ||
1225 | if (err) | ||
1226 | xenbus_dev_error(dev, err, "writing feature-sg-grant"); | ||
1227 | |||
1228 | err = xenbus_switch_state(dev, XenbusStateInitWait); | ||
1229 | if (err) | ||
1230 | goto fail; | ||
1231 | |||
1232 | return 0; | ||
1233 | |||
1234 | fail: | ||
1235 | pr_warn("xen-pvscsi: %s failed\n", __func__); | ||
1236 | scsiback_remove(dev); | ||
1237 | |||
1238 | return err; | ||
1239 | } | ||
1240 | |||
1241 | static char *scsiback_dump_proto_id(struct scsiback_tport *tport) | ||
1242 | { | ||
1243 | switch (tport->tport_proto_id) { | ||
1244 | case SCSI_PROTOCOL_SAS: | ||
1245 | return "SAS"; | ||
1246 | case SCSI_PROTOCOL_FCP: | ||
1247 | return "FCP"; | ||
1248 | case SCSI_PROTOCOL_ISCSI: | ||
1249 | return "iSCSI"; | ||
1250 | default: | ||
1251 | break; | ||
1252 | } | ||
1253 | |||
1254 | return "Unknown"; | ||
1255 | } | ||
1256 | |||
1257 | static u8 scsiback_get_fabric_proto_ident(struct se_portal_group *se_tpg) | ||
1258 | { | ||
1259 | struct scsiback_tpg *tpg = container_of(se_tpg, | ||
1260 | struct scsiback_tpg, se_tpg); | ||
1261 | struct scsiback_tport *tport = tpg->tport; | ||
1262 | |||
1263 | switch (tport->tport_proto_id) { | ||
1264 | case SCSI_PROTOCOL_SAS: | ||
1265 | return sas_get_fabric_proto_ident(se_tpg); | ||
1266 | case SCSI_PROTOCOL_FCP: | ||
1267 | return fc_get_fabric_proto_ident(se_tpg); | ||
1268 | case SCSI_PROTOCOL_ISCSI: | ||
1269 | return iscsi_get_fabric_proto_ident(se_tpg); | ||
1270 | default: | ||
1271 | pr_err("Unknown tport_proto_id: 0x%02x, using SAS emulation\n", | ||
1272 | tport->tport_proto_id); | ||
1273 | break; | ||
1274 | } | ||
1275 | |||
1276 | return sas_get_fabric_proto_ident(se_tpg); | ||
1277 | } | ||
1278 | |||
1279 | static char *scsiback_get_fabric_wwn(struct se_portal_group *se_tpg) | ||
1280 | { | ||
1281 | struct scsiback_tpg *tpg = container_of(se_tpg, | ||
1282 | struct scsiback_tpg, se_tpg); | ||
1283 | struct scsiback_tport *tport = tpg->tport; | ||
1284 | |||
1285 | return &tport->tport_name[0]; | ||
1286 | } | ||
1287 | |||
1288 | static u16 scsiback_get_tag(struct se_portal_group *se_tpg) | ||
1289 | { | ||
1290 | struct scsiback_tpg *tpg = container_of(se_tpg, | ||
1291 | struct scsiback_tpg, se_tpg); | ||
1292 | return tpg->tport_tpgt; | ||
1293 | } | ||
1294 | |||
1295 | static u32 scsiback_get_default_depth(struct se_portal_group *se_tpg) | ||
1296 | { | ||
1297 | return 1; | ||
1298 | } | ||
1299 | |||
1300 | static u32 | ||
1301 | scsiback_get_pr_transport_id(struct se_portal_group *se_tpg, | ||
1302 | struct se_node_acl *se_nacl, | ||
1303 | struct t10_pr_registration *pr_reg, | ||
1304 | int *format_code, | ||
1305 | unsigned char *buf) | ||
1306 | { | ||
1307 | struct scsiback_tpg *tpg = container_of(se_tpg, | ||
1308 | struct scsiback_tpg, se_tpg); | ||
1309 | struct scsiback_tport *tport = tpg->tport; | ||
1310 | |||
1311 | switch (tport->tport_proto_id) { | ||
1312 | case SCSI_PROTOCOL_SAS: | ||
1313 | return sas_get_pr_transport_id(se_tpg, se_nacl, pr_reg, | ||
1314 | format_code, buf); | ||
1315 | case SCSI_PROTOCOL_FCP: | ||
1316 | return fc_get_pr_transport_id(se_tpg, se_nacl, pr_reg, | ||
1317 | format_code, buf); | ||
1318 | case SCSI_PROTOCOL_ISCSI: | ||
1319 | return iscsi_get_pr_transport_id(se_tpg, se_nacl, pr_reg, | ||
1320 | format_code, buf); | ||
1321 | default: | ||
1322 | pr_err("Unknown tport_proto_id: 0x%02x, using SAS emulation\n", | ||
1323 | tport->tport_proto_id); | ||
1324 | break; | ||
1325 | } | ||
1326 | |||
1327 | return sas_get_pr_transport_id(se_tpg, se_nacl, pr_reg, | ||
1328 | format_code, buf); | ||
1329 | } | ||
1330 | |||
1331 | static u32 | ||
1332 | scsiback_get_pr_transport_id_len(struct se_portal_group *se_tpg, | ||
1333 | struct se_node_acl *se_nacl, | ||
1334 | struct t10_pr_registration *pr_reg, | ||
1335 | int *format_code) | ||
1336 | { | ||
1337 | struct scsiback_tpg *tpg = container_of(se_tpg, | ||
1338 | struct scsiback_tpg, se_tpg); | ||
1339 | struct scsiback_tport *tport = tpg->tport; | ||
1340 | |||
1341 | switch (tport->tport_proto_id) { | ||
1342 | case SCSI_PROTOCOL_SAS: | ||
1343 | return sas_get_pr_transport_id_len(se_tpg, se_nacl, pr_reg, | ||
1344 | format_code); | ||
1345 | case SCSI_PROTOCOL_FCP: | ||
1346 | return fc_get_pr_transport_id_len(se_tpg, se_nacl, pr_reg, | ||
1347 | format_code); | ||
1348 | case SCSI_PROTOCOL_ISCSI: | ||
1349 | return iscsi_get_pr_transport_id_len(se_tpg, se_nacl, pr_reg, | ||
1350 | format_code); | ||
1351 | default: | ||
1352 | pr_err("Unknown tport_proto_id: 0x%02x, using SAS emulation\n", | ||
1353 | tport->tport_proto_id); | ||
1354 | break; | ||
1355 | } | ||
1356 | |||
1357 | return sas_get_pr_transport_id_len(se_tpg, se_nacl, pr_reg, | ||
1358 | format_code); | ||
1359 | } | ||
1360 | |||
1361 | static char * | ||
1362 | scsiback_parse_pr_out_transport_id(struct se_portal_group *se_tpg, | ||
1363 | const char *buf, | ||
1364 | u32 *out_tid_len, | ||
1365 | char **port_nexus_ptr) | ||
1366 | { | ||
1367 | struct scsiback_tpg *tpg = container_of(se_tpg, | ||
1368 | struct scsiback_tpg, se_tpg); | ||
1369 | struct scsiback_tport *tport = tpg->tport; | ||
1370 | |||
1371 | switch (tport->tport_proto_id) { | ||
1372 | case SCSI_PROTOCOL_SAS: | ||
1373 | return sas_parse_pr_out_transport_id(se_tpg, buf, out_tid_len, | ||
1374 | port_nexus_ptr); | ||
1375 | case SCSI_PROTOCOL_FCP: | ||
1376 | return fc_parse_pr_out_transport_id(se_tpg, buf, out_tid_len, | ||
1377 | port_nexus_ptr); | ||
1378 | case SCSI_PROTOCOL_ISCSI: | ||
1379 | return iscsi_parse_pr_out_transport_id(se_tpg, buf, out_tid_len, | ||
1380 | port_nexus_ptr); | ||
1381 | default: | ||
1382 | pr_err("Unknown tport_proto_id: 0x%02x, using SAS emulation\n", | ||
1383 | tport->tport_proto_id); | ||
1384 | break; | ||
1385 | } | ||
1386 | |||
1387 | return sas_parse_pr_out_transport_id(se_tpg, buf, out_tid_len, | ||
1388 | port_nexus_ptr); | ||
1389 | } | ||
1390 | |||
1391 | static struct se_wwn * | ||
1392 | scsiback_make_tport(struct target_fabric_configfs *tf, | ||
1393 | struct config_group *group, | ||
1394 | const char *name) | ||
1395 | { | ||
1396 | struct scsiback_tport *tport; | ||
1397 | char *ptr; | ||
1398 | u64 wwpn = 0; | ||
1399 | int off = 0; | ||
1400 | |||
1401 | tport = kzalloc(sizeof(struct scsiback_tport), GFP_KERNEL); | ||
1402 | if (!tport) | ||
1403 | return ERR_PTR(-ENOMEM); | ||
1404 | |||
1405 | tport->tport_wwpn = wwpn; | ||
1406 | /* | ||
1407 | * Determine the emulated Protocol Identifier and Target Port Name | ||
1408 | * based on the incoming configfs directory name. | ||
1409 | */ | ||
1410 | ptr = strstr(name, "naa."); | ||
1411 | if (ptr) { | ||
1412 | tport->tport_proto_id = SCSI_PROTOCOL_SAS; | ||
1413 | goto check_len; | ||
1414 | } | ||
1415 | ptr = strstr(name, "fc."); | ||
1416 | if (ptr) { | ||
1417 | tport->tport_proto_id = SCSI_PROTOCOL_FCP; | ||
1418 | off = 3; /* Skip over "fc." */ | ||
1419 | goto check_len; | ||
1420 | } | ||
1421 | ptr = strstr(name, "iqn."); | ||
1422 | if (ptr) { | ||
1423 | tport->tport_proto_id = SCSI_PROTOCOL_ISCSI; | ||
1424 | goto check_len; | ||
1425 | } | ||
1426 | |||
1427 | pr_err("Unable to locate prefix for emulated Target Port: %s\n", name); | ||
1428 | kfree(tport); | ||
1429 | return ERR_PTR(-EINVAL); | ||
1430 | |||
1431 | check_len: | ||
1432 | if (strlen(name) >= VSCSI_NAMELEN) { | ||
1433 | pr_err("Emulated %s Address: %s, exceeds max: %d\n", name, | ||
1434 | scsiback_dump_proto_id(tport), VSCSI_NAMELEN); | ||
1435 | kfree(tport); | ||
1436 | return ERR_PTR(-EINVAL); | ||
1437 | } | ||
1438 | snprintf(&tport->tport_name[0], VSCSI_NAMELEN, "%s", &name[off]); | ||
1439 | |||
1440 | pr_debug("xen-pvscsi: Allocated emulated Target %s Address: %s\n", | ||
1441 | scsiback_dump_proto_id(tport), name); | ||
1442 | |||
1443 | return &tport->tport_wwn; | ||
1444 | } | ||
1445 | |||
1446 | static void scsiback_drop_tport(struct se_wwn *wwn) | ||
1447 | { | ||
1448 | struct scsiback_tport *tport = container_of(wwn, | ||
1449 | struct scsiback_tport, tport_wwn); | ||
1450 | |||
1451 | pr_debug("xen-pvscsi: Deallocating emulated Target %s Address: %s\n", | ||
1452 | scsiback_dump_proto_id(tport), tport->tport_name); | ||
1453 | |||
1454 | kfree(tport); | ||
1455 | } | ||
1456 | |||
1457 | static struct se_node_acl * | ||
1458 | scsiback_alloc_fabric_acl(struct se_portal_group *se_tpg) | ||
1459 | { | ||
1460 | return kzalloc(sizeof(struct se_node_acl), GFP_KERNEL); | ||
1461 | } | ||
1462 | |||
1463 | static void | ||
1464 | scsiback_release_fabric_acl(struct se_portal_group *se_tpg, | ||
1465 | struct se_node_acl *se_nacl) | ||
1466 | { | ||
1467 | kfree(se_nacl); | ||
1468 | } | ||
1469 | |||
1470 | static u32 scsiback_tpg_get_inst_index(struct se_portal_group *se_tpg) | ||
1471 | { | ||
1472 | return 1; | ||
1473 | } | ||
1474 | |||
1475 | static int scsiback_check_stop_free(struct se_cmd *se_cmd) | ||
1476 | { | ||
1477 | /* | ||
1478 | * Do not release struct se_cmd's containing a valid TMR | ||
1479 | * pointer. These will be released directly in scsiback_device_action() | ||
1480 | * with transport_generic_free_cmd(). | ||
1481 | */ | ||
1482 | if (se_cmd->se_cmd_flags & SCF_SCSI_TMR_CDB) | ||
1483 | return 0; | ||
1484 | |||
1485 | transport_generic_free_cmd(se_cmd, 0); | ||
1486 | return 1; | ||
1487 | } | ||
1488 | |||
1489 | static void scsiback_release_cmd(struct se_cmd *se_cmd) | ||
1490 | { | ||
1491 | struct vscsibk_pend *pending_req = container_of(se_cmd, | ||
1492 | struct vscsibk_pend, se_cmd); | ||
1493 | |||
1494 | kmem_cache_free(scsiback_cachep, pending_req); | ||
1495 | } | ||
1496 | |||
1497 | static int scsiback_shutdown_session(struct se_session *se_sess) | ||
1498 | { | ||
1499 | return 0; | ||
1500 | } | ||
1501 | |||
1502 | static void scsiback_close_session(struct se_session *se_sess) | ||
1503 | { | ||
1504 | } | ||
1505 | |||
1506 | static u32 scsiback_sess_get_index(struct se_session *se_sess) | ||
1507 | { | ||
1508 | return 0; | ||
1509 | } | ||
1510 | |||
1511 | static int scsiback_write_pending(struct se_cmd *se_cmd) | ||
1512 | { | ||
1513 | /* Go ahead and process the write immediately */ | ||
1514 | target_execute_cmd(se_cmd); | ||
1515 | |||
1516 | return 0; | ||
1517 | } | ||
1518 | |||
1519 | static int scsiback_write_pending_status(struct se_cmd *se_cmd) | ||
1520 | { | ||
1521 | return 0; | ||
1522 | } | ||
1523 | |||
1524 | static void scsiback_set_default_node_attrs(struct se_node_acl *nacl) | ||
1525 | { | ||
1526 | } | ||
1527 | |||
1528 | static u32 scsiback_get_task_tag(struct se_cmd *se_cmd) | ||
1529 | { | ||
1530 | struct vscsibk_pend *pending_req = container_of(se_cmd, | ||
1531 | struct vscsibk_pend, se_cmd); | ||
1532 | |||
1533 | return pending_req->rqid; | ||
1534 | } | ||
1535 | |||
1536 | static int scsiback_get_cmd_state(struct se_cmd *se_cmd) | ||
1537 | { | ||
1538 | return 0; | ||
1539 | } | ||
1540 | |||
1541 | static int scsiback_queue_data_in(struct se_cmd *se_cmd) | ||
1542 | { | ||
1543 | struct vscsibk_pend *pending_req = container_of(se_cmd, | ||
1544 | struct vscsibk_pend, se_cmd); | ||
1545 | |||
1546 | pending_req->result = SAM_STAT_GOOD; | ||
1547 | scsiback_cmd_done(pending_req); | ||
1548 | return 0; | ||
1549 | } | ||
1550 | |||
1551 | static int scsiback_queue_status(struct se_cmd *se_cmd) | ||
1552 | { | ||
1553 | struct vscsibk_pend *pending_req = container_of(se_cmd, | ||
1554 | struct vscsibk_pend, se_cmd); | ||
1555 | |||
1556 | if (se_cmd->sense_buffer && | ||
1557 | ((se_cmd->se_cmd_flags & SCF_TRANSPORT_TASK_SENSE) || | ||
1558 | (se_cmd->se_cmd_flags & SCF_EMULATED_TASK_SENSE))) | ||
1559 | pending_req->result = (DRIVER_SENSE << 24) | | ||
1560 | SAM_STAT_CHECK_CONDITION; | ||
1561 | else | ||
1562 | pending_req->result = se_cmd->scsi_status; | ||
1563 | |||
1564 | scsiback_cmd_done(pending_req); | ||
1565 | return 0; | ||
1566 | } | ||
1567 | |||
1568 | static void scsiback_queue_tm_rsp(struct se_cmd *se_cmd) | ||
1569 | { | ||
1570 | struct se_tmr_req *se_tmr = se_cmd->se_tmr_req; | ||
1571 | struct scsiback_tmr *tmr = se_tmr->fabric_tmr_ptr; | ||
1572 | |||
1573 | atomic_set(&tmr->tmr_complete, 1); | ||
1574 | wake_up(&tmr->tmr_wait); | ||
1575 | } | ||
1576 | |||
1577 | static void scsiback_aborted_task(struct se_cmd *se_cmd) | ||
1578 | { | ||
1579 | } | ||
1580 | |||
1581 | static ssize_t scsiback_tpg_param_show_alias(struct se_portal_group *se_tpg, | ||
1582 | char *page) | ||
1583 | { | ||
1584 | struct scsiback_tpg *tpg = container_of(se_tpg, struct scsiback_tpg, | ||
1585 | se_tpg); | ||
1586 | ssize_t rb; | ||
1587 | |||
1588 | mutex_lock(&tpg->tv_tpg_mutex); | ||
1589 | rb = snprintf(page, PAGE_SIZE, "%s\n", tpg->param_alias); | ||
1590 | mutex_unlock(&tpg->tv_tpg_mutex); | ||
1591 | |||
1592 | return rb; | ||
1593 | } | ||
1594 | |||
1595 | static ssize_t scsiback_tpg_param_store_alias(struct se_portal_group *se_tpg, | ||
1596 | const char *page, size_t count) | ||
1597 | { | ||
1598 | struct scsiback_tpg *tpg = container_of(se_tpg, struct scsiback_tpg, | ||
1599 | se_tpg); | ||
1600 | int len; | ||
1601 | |||
1602 | if (strlen(page) >= VSCSI_NAMELEN) { | ||
1603 | pr_err("param alias: %s, exceeds max: %d\n", page, | ||
1604 | VSCSI_NAMELEN); | ||
1605 | return -EINVAL; | ||
1606 | } | ||
1607 | |||
1608 | mutex_lock(&tpg->tv_tpg_mutex); | ||
1609 | len = snprintf(tpg->param_alias, VSCSI_NAMELEN, "%s", page); | ||
1610 | if (tpg->param_alias[len - 1] == '\n') | ||
1611 | tpg->param_alias[len - 1] = '\0'; | ||
1612 | mutex_unlock(&tpg->tv_tpg_mutex); | ||
1613 | |||
1614 | return count; | ||
1615 | } | ||
1616 | |||
1617 | TF_TPG_PARAM_ATTR(scsiback, alias, S_IRUGO | S_IWUSR); | ||
1618 | |||
1619 | static struct configfs_attribute *scsiback_param_attrs[] = { | ||
1620 | &scsiback_tpg_param_alias.attr, | ||
1621 | NULL, | ||
1622 | }; | ||
1623 | |||
1624 | static int scsiback_make_nexus(struct scsiback_tpg *tpg, | ||
1625 | const char *name) | ||
1626 | { | ||
1627 | struct se_portal_group *se_tpg; | ||
1628 | struct se_session *se_sess; | ||
1629 | struct scsiback_nexus *tv_nexus; | ||
1630 | |||
1631 | mutex_lock(&tpg->tv_tpg_mutex); | ||
1632 | if (tpg->tpg_nexus) { | ||
1633 | mutex_unlock(&tpg->tv_tpg_mutex); | ||
1634 | pr_debug("tpg->tpg_nexus already exists\n"); | ||
1635 | return -EEXIST; | ||
1636 | } | ||
1637 | se_tpg = &tpg->se_tpg; | ||
1638 | |||
1639 | tv_nexus = kzalloc(sizeof(struct scsiback_nexus), GFP_KERNEL); | ||
1640 | if (!tv_nexus) { | ||
1641 | mutex_unlock(&tpg->tv_tpg_mutex); | ||
1642 | return -ENOMEM; | ||
1643 | } | ||
1644 | /* | ||
1645 | * Initialize the struct se_session pointer | ||
1646 | */ | ||
1647 | tv_nexus->tvn_se_sess = transport_init_session(TARGET_PROT_NORMAL); | ||
1648 | if (IS_ERR(tv_nexus->tvn_se_sess)) { | ||
1649 | mutex_unlock(&tpg->tv_tpg_mutex); | ||
1650 | kfree(tv_nexus); | ||
1651 | return -ENOMEM; | ||
1652 | } | ||
1653 | se_sess = tv_nexus->tvn_se_sess; | ||
1654 | /* | ||
1655 | * Since we are running in 'demo mode' this call with generate a | ||
1656 | * struct se_node_acl for the scsiback struct se_portal_group with | ||
1657 | * the SCSI Initiator port name of the passed configfs group 'name'. | ||
1658 | */ | ||
1659 | tv_nexus->tvn_se_sess->se_node_acl = core_tpg_check_initiator_node_acl( | ||
1660 | se_tpg, (unsigned char *)name); | ||
1661 | if (!tv_nexus->tvn_se_sess->se_node_acl) { | ||
1662 | mutex_unlock(&tpg->tv_tpg_mutex); | ||
1663 | pr_debug("core_tpg_check_initiator_node_acl() failed for %s\n", | ||
1664 | name); | ||
1665 | goto out; | ||
1666 | } | ||
1667 | /* | ||
1668 | * Now register the TCM pvscsi virtual I_T Nexus as active with the | ||
1669 | * call to __transport_register_session() | ||
1670 | */ | ||
1671 | __transport_register_session(se_tpg, tv_nexus->tvn_se_sess->se_node_acl, | ||
1672 | tv_nexus->tvn_se_sess, tv_nexus); | ||
1673 | tpg->tpg_nexus = tv_nexus; | ||
1674 | |||
1675 | mutex_unlock(&tpg->tv_tpg_mutex); | ||
1676 | return 0; | ||
1677 | |||
1678 | out: | ||
1679 | transport_free_session(se_sess); | ||
1680 | kfree(tv_nexus); | ||
1681 | return -ENOMEM; | ||
1682 | } | ||
1683 | |||
1684 | static int scsiback_drop_nexus(struct scsiback_tpg *tpg) | ||
1685 | { | ||
1686 | struct se_session *se_sess; | ||
1687 | struct scsiback_nexus *tv_nexus; | ||
1688 | |||
1689 | mutex_lock(&tpg->tv_tpg_mutex); | ||
1690 | tv_nexus = tpg->tpg_nexus; | ||
1691 | if (!tv_nexus) { | ||
1692 | mutex_unlock(&tpg->tv_tpg_mutex); | ||
1693 | return -ENODEV; | ||
1694 | } | ||
1695 | |||
1696 | se_sess = tv_nexus->tvn_se_sess; | ||
1697 | if (!se_sess) { | ||
1698 | mutex_unlock(&tpg->tv_tpg_mutex); | ||
1699 | return -ENODEV; | ||
1700 | } | ||
1701 | |||
1702 | if (tpg->tv_tpg_port_count != 0) { | ||
1703 | mutex_unlock(&tpg->tv_tpg_mutex); | ||
1704 | pr_err("Unable to remove xen-pvscsi I_T Nexus with active TPG port count: %d\n", | ||
1705 | tpg->tv_tpg_port_count); | ||
1706 | return -EBUSY; | ||
1707 | } | ||
1708 | |||
1709 | if (tpg->tv_tpg_fe_count != 0) { | ||
1710 | mutex_unlock(&tpg->tv_tpg_mutex); | ||
1711 | pr_err("Unable to remove xen-pvscsi I_T Nexus with active TPG frontend count: %d\n", | ||
1712 | tpg->tv_tpg_fe_count); | ||
1713 | return -EBUSY; | ||
1714 | } | ||
1715 | |||
1716 | pr_debug("xen-pvscsi: Removing I_T Nexus to emulated %s Initiator Port: %s\n", | ||
1717 | scsiback_dump_proto_id(tpg->tport), | ||
1718 | tv_nexus->tvn_se_sess->se_node_acl->initiatorname); | ||
1719 | |||
1720 | /* | ||
1721 | * Release the SCSI I_T Nexus to the emulated xen-pvscsi Target Port | ||
1722 | */ | ||
1723 | transport_deregister_session(tv_nexus->tvn_se_sess); | ||
1724 | tpg->tpg_nexus = NULL; | ||
1725 | mutex_unlock(&tpg->tv_tpg_mutex); | ||
1726 | |||
1727 | kfree(tv_nexus); | ||
1728 | return 0; | ||
1729 | } | ||
1730 | |||
1731 | static ssize_t scsiback_tpg_show_nexus(struct se_portal_group *se_tpg, | ||
1732 | char *page) | ||
1733 | { | ||
1734 | struct scsiback_tpg *tpg = container_of(se_tpg, | ||
1735 | struct scsiback_tpg, se_tpg); | ||
1736 | struct scsiback_nexus *tv_nexus; | ||
1737 | ssize_t ret; | ||
1738 | |||
1739 | mutex_lock(&tpg->tv_tpg_mutex); | ||
1740 | tv_nexus = tpg->tpg_nexus; | ||
1741 | if (!tv_nexus) { | ||
1742 | mutex_unlock(&tpg->tv_tpg_mutex); | ||
1743 | return -ENODEV; | ||
1744 | } | ||
1745 | ret = snprintf(page, PAGE_SIZE, "%s\n", | ||
1746 | tv_nexus->tvn_se_sess->se_node_acl->initiatorname); | ||
1747 | mutex_unlock(&tpg->tv_tpg_mutex); | ||
1748 | |||
1749 | return ret; | ||
1750 | } | ||
1751 | |||
1752 | static ssize_t scsiback_tpg_store_nexus(struct se_portal_group *se_tpg, | ||
1753 | const char *page, | ||
1754 | size_t count) | ||
1755 | { | ||
1756 | struct scsiback_tpg *tpg = container_of(se_tpg, | ||
1757 | struct scsiback_tpg, se_tpg); | ||
1758 | struct scsiback_tport *tport_wwn = tpg->tport; | ||
1759 | unsigned char i_port[VSCSI_NAMELEN], *ptr, *port_ptr; | ||
1760 | int ret; | ||
1761 | /* | ||
1762 | * Shutdown the active I_T nexus if 'NULL' is passed.. | ||
1763 | */ | ||
1764 | if (!strncmp(page, "NULL", 4)) { | ||
1765 | ret = scsiback_drop_nexus(tpg); | ||
1766 | return (!ret) ? count : ret; | ||
1767 | } | ||
1768 | /* | ||
1769 | * Otherwise make sure the passed virtual Initiator port WWN matches | ||
1770 | * the fabric protocol_id set in scsiback_make_tport(), and call | ||
1771 | * scsiback_make_nexus(). | ||
1772 | */ | ||
1773 | if (strlen(page) >= VSCSI_NAMELEN) { | ||
1774 | pr_err("Emulated NAA Sas Address: %s, exceeds max: %d\n", | ||
1775 | page, VSCSI_NAMELEN); | ||
1776 | return -EINVAL; | ||
1777 | } | ||
1778 | snprintf(&i_port[0], VSCSI_NAMELEN, "%s", page); | ||
1779 | |||
1780 | ptr = strstr(i_port, "naa."); | ||
1781 | if (ptr) { | ||
1782 | if (tport_wwn->tport_proto_id != SCSI_PROTOCOL_SAS) { | ||
1783 | pr_err("Passed SAS Initiator Port %s does not match target port protoid: %s\n", | ||
1784 | i_port, scsiback_dump_proto_id(tport_wwn)); | ||
1785 | return -EINVAL; | ||
1786 | } | ||
1787 | port_ptr = &i_port[0]; | ||
1788 | goto check_newline; | ||
1789 | } | ||
1790 | ptr = strstr(i_port, "fc."); | ||
1791 | if (ptr) { | ||
1792 | if (tport_wwn->tport_proto_id != SCSI_PROTOCOL_FCP) { | ||
1793 | pr_err("Passed FCP Initiator Port %s does not match target port protoid: %s\n", | ||
1794 | i_port, scsiback_dump_proto_id(tport_wwn)); | ||
1795 | return -EINVAL; | ||
1796 | } | ||
1797 | port_ptr = &i_port[3]; /* Skip over "fc." */ | ||
1798 | goto check_newline; | ||
1799 | } | ||
1800 | ptr = strstr(i_port, "iqn."); | ||
1801 | if (ptr) { | ||
1802 | if (tport_wwn->tport_proto_id != SCSI_PROTOCOL_ISCSI) { | ||
1803 | pr_err("Passed iSCSI Initiator Port %s does not match target port protoid: %s\n", | ||
1804 | i_port, scsiback_dump_proto_id(tport_wwn)); | ||
1805 | return -EINVAL; | ||
1806 | } | ||
1807 | port_ptr = &i_port[0]; | ||
1808 | goto check_newline; | ||
1809 | } | ||
1810 | pr_err("Unable to locate prefix for emulated Initiator Port: %s\n", | ||
1811 | i_port); | ||
1812 | return -EINVAL; | ||
1813 | /* | ||
1814 | * Clear any trailing newline for the NAA WWN | ||
1815 | */ | ||
1816 | check_newline: | ||
1817 | if (i_port[strlen(i_port) - 1] == '\n') | ||
1818 | i_port[strlen(i_port) - 1] = '\0'; | ||
1819 | |||
1820 | ret = scsiback_make_nexus(tpg, port_ptr); | ||
1821 | if (ret < 0) | ||
1822 | return ret; | ||
1823 | |||
1824 | return count; | ||
1825 | } | ||
1826 | |||
1827 | TF_TPG_BASE_ATTR(scsiback, nexus, S_IRUGO | S_IWUSR); | ||
1828 | |||
1829 | static struct configfs_attribute *scsiback_tpg_attrs[] = { | ||
1830 | &scsiback_tpg_nexus.attr, | ||
1831 | NULL, | ||
1832 | }; | ||
1833 | |||
1834 | static ssize_t | ||
1835 | scsiback_wwn_show_attr_version(struct target_fabric_configfs *tf, | ||
1836 | char *page) | ||
1837 | { | ||
1838 | return sprintf(page, "xen-pvscsi fabric module %s on %s/%s on " | ||
1839 | UTS_RELEASE"\n", | ||
1840 | VSCSI_VERSION, utsname()->sysname, utsname()->machine); | ||
1841 | } | ||
1842 | |||
1843 | TF_WWN_ATTR_RO(scsiback, version); | ||
1844 | |||
1845 | static struct configfs_attribute *scsiback_wwn_attrs[] = { | ||
1846 | &scsiback_wwn_version.attr, | ||
1847 | NULL, | ||
1848 | }; | ||
1849 | |||
1850 | static char *scsiback_get_fabric_name(void) | ||
1851 | { | ||
1852 | return "xen-pvscsi"; | ||
1853 | } | ||
1854 | |||
1855 | static int scsiback_port_link(struct se_portal_group *se_tpg, | ||
1856 | struct se_lun *lun) | ||
1857 | { | ||
1858 | struct scsiback_tpg *tpg = container_of(se_tpg, | ||
1859 | struct scsiback_tpg, se_tpg); | ||
1860 | |||
1861 | mutex_lock(&tpg->tv_tpg_mutex); | ||
1862 | tpg->tv_tpg_port_count++; | ||
1863 | mutex_unlock(&tpg->tv_tpg_mutex); | ||
1864 | |||
1865 | return 0; | ||
1866 | } | ||
1867 | |||
1868 | static void scsiback_port_unlink(struct se_portal_group *se_tpg, | ||
1869 | struct se_lun *lun) | ||
1870 | { | ||
1871 | struct scsiback_tpg *tpg = container_of(se_tpg, | ||
1872 | struct scsiback_tpg, se_tpg); | ||
1873 | |||
1874 | mutex_lock(&tpg->tv_tpg_mutex); | ||
1875 | tpg->tv_tpg_port_count--; | ||
1876 | mutex_unlock(&tpg->tv_tpg_mutex); | ||
1877 | } | ||
1878 | |||
1879 | static struct se_portal_group * | ||
1880 | scsiback_make_tpg(struct se_wwn *wwn, | ||
1881 | struct config_group *group, | ||
1882 | const char *name) | ||
1883 | { | ||
1884 | struct scsiback_tport *tport = container_of(wwn, | ||
1885 | struct scsiback_tport, tport_wwn); | ||
1886 | |||
1887 | struct scsiback_tpg *tpg; | ||
1888 | u16 tpgt; | ||
1889 | int ret; | ||
1890 | |||
1891 | if (strstr(name, "tpgt_") != name) | ||
1892 | return ERR_PTR(-EINVAL); | ||
1893 | ret = kstrtou16(name + 5, 10, &tpgt); | ||
1894 | if (ret) | ||
1895 | return ERR_PTR(ret); | ||
1896 | |||
1897 | tpg = kzalloc(sizeof(struct scsiback_tpg), GFP_KERNEL); | ||
1898 | if (!tpg) | ||
1899 | return ERR_PTR(-ENOMEM); | ||
1900 | |||
1901 | mutex_init(&tpg->tv_tpg_mutex); | ||
1902 | INIT_LIST_HEAD(&tpg->tv_tpg_list); | ||
1903 | INIT_LIST_HEAD(&tpg->info_list); | ||
1904 | tpg->tport = tport; | ||
1905 | tpg->tport_tpgt = tpgt; | ||
1906 | |||
1907 | ret = core_tpg_register(&scsiback_fabric_configfs->tf_ops, wwn, | ||
1908 | &tpg->se_tpg, tpg, TRANSPORT_TPG_TYPE_NORMAL); | ||
1909 | if (ret < 0) { | ||
1910 | kfree(tpg); | ||
1911 | return NULL; | ||
1912 | } | ||
1913 | mutex_lock(&scsiback_mutex); | ||
1914 | list_add_tail(&tpg->tv_tpg_list, &scsiback_list); | ||
1915 | mutex_unlock(&scsiback_mutex); | ||
1916 | |||
1917 | return &tpg->se_tpg; | ||
1918 | } | ||
1919 | |||
1920 | static void scsiback_drop_tpg(struct se_portal_group *se_tpg) | ||
1921 | { | ||
1922 | struct scsiback_tpg *tpg = container_of(se_tpg, | ||
1923 | struct scsiback_tpg, se_tpg); | ||
1924 | |||
1925 | mutex_lock(&scsiback_mutex); | ||
1926 | list_del(&tpg->tv_tpg_list); | ||
1927 | mutex_unlock(&scsiback_mutex); | ||
1928 | /* | ||
1929 | * Release the virtual I_T Nexus for this xen-pvscsi TPG | ||
1930 | */ | ||
1931 | scsiback_drop_nexus(tpg); | ||
1932 | /* | ||
1933 | * Deregister the se_tpg from TCM.. | ||
1934 | */ | ||
1935 | core_tpg_deregister(se_tpg); | ||
1936 | kfree(tpg); | ||
1937 | } | ||
1938 | |||
1939 | static int scsiback_check_true(struct se_portal_group *se_tpg) | ||
1940 | { | ||
1941 | return 1; | ||
1942 | } | ||
1943 | |||
1944 | static int scsiback_check_false(struct se_portal_group *se_tpg) | ||
1945 | { | ||
1946 | return 0; | ||
1947 | } | ||
1948 | |||
1949 | static struct target_core_fabric_ops scsiback_ops = { | ||
1950 | .get_fabric_name = scsiback_get_fabric_name, | ||
1951 | .get_fabric_proto_ident = scsiback_get_fabric_proto_ident, | ||
1952 | .tpg_get_wwn = scsiback_get_fabric_wwn, | ||
1953 | .tpg_get_tag = scsiback_get_tag, | ||
1954 | .tpg_get_default_depth = scsiback_get_default_depth, | ||
1955 | .tpg_get_pr_transport_id = scsiback_get_pr_transport_id, | ||
1956 | .tpg_get_pr_transport_id_len = scsiback_get_pr_transport_id_len, | ||
1957 | .tpg_parse_pr_out_transport_id = scsiback_parse_pr_out_transport_id, | ||
1958 | .tpg_check_demo_mode = scsiback_check_true, | ||
1959 | .tpg_check_demo_mode_cache = scsiback_check_true, | ||
1960 | .tpg_check_demo_mode_write_protect = scsiback_check_false, | ||
1961 | .tpg_check_prod_mode_write_protect = scsiback_check_false, | ||
1962 | .tpg_alloc_fabric_acl = scsiback_alloc_fabric_acl, | ||
1963 | .tpg_release_fabric_acl = scsiback_release_fabric_acl, | ||
1964 | .tpg_get_inst_index = scsiback_tpg_get_inst_index, | ||
1965 | .check_stop_free = scsiback_check_stop_free, | ||
1966 | .release_cmd = scsiback_release_cmd, | ||
1967 | .put_session = NULL, | ||
1968 | .shutdown_session = scsiback_shutdown_session, | ||
1969 | .close_session = scsiback_close_session, | ||
1970 | .sess_get_index = scsiback_sess_get_index, | ||
1971 | .sess_get_initiator_sid = NULL, | ||
1972 | .write_pending = scsiback_write_pending, | ||
1973 | .write_pending_status = scsiback_write_pending_status, | ||
1974 | .set_default_node_attributes = scsiback_set_default_node_attrs, | ||
1975 | .get_task_tag = scsiback_get_task_tag, | ||
1976 | .get_cmd_state = scsiback_get_cmd_state, | ||
1977 | .queue_data_in = scsiback_queue_data_in, | ||
1978 | .queue_status = scsiback_queue_status, | ||
1979 | .queue_tm_rsp = scsiback_queue_tm_rsp, | ||
1980 | .aborted_task = scsiback_aborted_task, | ||
1981 | /* | ||
1982 | * Setup callers for generic logic in target_core_fabric_configfs.c | ||
1983 | */ | ||
1984 | .fabric_make_wwn = scsiback_make_tport, | ||
1985 | .fabric_drop_wwn = scsiback_drop_tport, | ||
1986 | .fabric_make_tpg = scsiback_make_tpg, | ||
1987 | .fabric_drop_tpg = scsiback_drop_tpg, | ||
1988 | .fabric_post_link = scsiback_port_link, | ||
1989 | .fabric_pre_unlink = scsiback_port_unlink, | ||
1990 | .fabric_make_np = NULL, | ||
1991 | .fabric_drop_np = NULL, | ||
1992 | #if 0 | ||
1993 | .fabric_make_nodeacl = scsiback_make_nodeacl, | ||
1994 | .fabric_drop_nodeacl = scsiback_drop_nodeacl, | ||
1995 | #endif | ||
1996 | }; | ||
1997 | |||
1998 | static int scsiback_register_configfs(void) | ||
1999 | { | ||
2000 | struct target_fabric_configfs *fabric; | ||
2001 | int ret; | ||
2002 | |||
2003 | pr_debug("xen-pvscsi: fabric module %s on %s/%s on "UTS_RELEASE"\n", | ||
2004 | VSCSI_VERSION, utsname()->sysname, utsname()->machine); | ||
2005 | /* | ||
2006 | * Register the top level struct config_item_type with TCM core | ||
2007 | */ | ||
2008 | fabric = target_fabric_configfs_init(THIS_MODULE, "xen-pvscsi"); | ||
2009 | if (IS_ERR(fabric)) | ||
2010 | return PTR_ERR(fabric); | ||
2011 | |||
2012 | /* | ||
2013 | * Setup fabric->tf_ops from our local scsiback_ops | ||
2014 | */ | ||
2015 | fabric->tf_ops = scsiback_ops; | ||
2016 | /* | ||
2017 | * Setup default attribute lists for various fabric->tf_cit_tmpl | ||
2018 | */ | ||
2019 | fabric->tf_cit_tmpl.tfc_wwn_cit.ct_attrs = scsiback_wwn_attrs; | ||
2020 | fabric->tf_cit_tmpl.tfc_tpg_base_cit.ct_attrs = scsiback_tpg_attrs; | ||
2021 | fabric->tf_cit_tmpl.tfc_tpg_attrib_cit.ct_attrs = NULL; | ||
2022 | fabric->tf_cit_tmpl.tfc_tpg_param_cit.ct_attrs = scsiback_param_attrs; | ||
2023 | fabric->tf_cit_tmpl.tfc_tpg_np_base_cit.ct_attrs = NULL; | ||
2024 | fabric->tf_cit_tmpl.tfc_tpg_nacl_base_cit.ct_attrs = NULL; | ||
2025 | fabric->tf_cit_tmpl.tfc_tpg_nacl_attrib_cit.ct_attrs = NULL; | ||
2026 | fabric->tf_cit_tmpl.tfc_tpg_nacl_auth_cit.ct_attrs = NULL; | ||
2027 | fabric->tf_cit_tmpl.tfc_tpg_nacl_param_cit.ct_attrs = NULL; | ||
2028 | /* | ||
2029 | * Register the fabric for use within TCM | ||
2030 | */ | ||
2031 | ret = target_fabric_configfs_register(fabric); | ||
2032 | if (ret < 0) { | ||
2033 | target_fabric_configfs_free(fabric); | ||
2034 | return ret; | ||
2035 | } | ||
2036 | /* | ||
2037 | * Setup our local pointer to *fabric | ||
2038 | */ | ||
2039 | scsiback_fabric_configfs = fabric; | ||
2040 | pr_debug("xen-pvscsi: Set fabric -> scsiback_fabric_configfs\n"); | ||
2041 | return 0; | ||
2042 | }; | ||
2043 | |||
2044 | static void scsiback_deregister_configfs(void) | ||
2045 | { | ||
2046 | if (!scsiback_fabric_configfs) | ||
2047 | return; | ||
2048 | |||
2049 | target_fabric_configfs_deregister(scsiback_fabric_configfs); | ||
2050 | scsiback_fabric_configfs = NULL; | ||
2051 | pr_debug("xen-pvscsi: Cleared scsiback_fabric_configfs\n"); | ||
2052 | }; | ||
2053 | |||
2054 | static const struct xenbus_device_id scsiback_ids[] = { | ||
2055 | { "vscsi" }, | ||
2056 | { "" } | ||
2057 | }; | ||
2058 | |||
2059 | static struct xenbus_driver scsiback_driver = { | ||
2060 | .ids = scsiback_ids, | ||
2061 | .probe = scsiback_probe, | ||
2062 | .remove = scsiback_remove, | ||
2063 | .otherend_changed = scsiback_frontend_changed | ||
2064 | }; | ||
2065 | |||
2066 | static void scsiback_init_pend(void *p) | ||
2067 | { | ||
2068 | struct vscsibk_pend *pend = p; | ||
2069 | int i; | ||
2070 | |||
2071 | memset(pend, 0, sizeof(*pend)); | ||
2072 | for (i = 0; i < VSCSI_MAX_GRANTS; i++) | ||
2073 | pend->grant_handles[i] = SCSIBACK_INVALID_HANDLE; | ||
2074 | } | ||
2075 | |||
2076 | static int __init scsiback_init(void) | ||
2077 | { | ||
2078 | int ret; | ||
2079 | |||
2080 | if (!xen_domain()) | ||
2081 | return -ENODEV; | ||
2082 | |||
2083 | scsiback_cachep = kmem_cache_create("vscsiif_cache", | ||
2084 | sizeof(struct vscsibk_pend), 0, 0, scsiback_init_pend); | ||
2085 | if (!scsiback_cachep) | ||
2086 | return -ENOMEM; | ||
2087 | |||
2088 | ret = xenbus_register_backend(&scsiback_driver); | ||
2089 | if (ret) | ||
2090 | goto out_cache_destroy; | ||
2091 | |||
2092 | ret = scsiback_register_configfs(); | ||
2093 | if (ret) | ||
2094 | goto out_unregister_xenbus; | ||
2095 | |||
2096 | return 0; | ||
2097 | |||
2098 | out_unregister_xenbus: | ||
2099 | xenbus_unregister_driver(&scsiback_driver); | ||
2100 | out_cache_destroy: | ||
2101 | kmem_cache_destroy(scsiback_cachep); | ||
2102 | pr_err("xen-pvscsi: %s: error %d\n", __func__, ret); | ||
2103 | return ret; | ||
2104 | } | ||
2105 | |||
2106 | static void __exit scsiback_exit(void) | ||
2107 | { | ||
2108 | struct page *page; | ||
2109 | |||
2110 | while (free_pages_num) { | ||
2111 | if (get_free_page(&page)) | ||
2112 | BUG(); | ||
2113 | free_xenballooned_pages(1, &page); | ||
2114 | } | ||
2115 | scsiback_deregister_configfs(); | ||
2116 | xenbus_unregister_driver(&scsiback_driver); | ||
2117 | kmem_cache_destroy(scsiback_cachep); | ||
2118 | } | ||
2119 | |||
2120 | module_init(scsiback_init); | ||
2121 | module_exit(scsiback_exit); | ||
2122 | |||
2123 | MODULE_DESCRIPTION("Xen SCSI backend driver"); | ||
2124 | MODULE_LICENSE("Dual BSD/GPL"); | ||
2125 | MODULE_ALIAS("xen-backend:vscsi"); | ||
2126 | MODULE_AUTHOR("Juergen Gross <jgross@suse.com>"); | ||
diff --git a/drivers/xen/xenbus/xenbus_client.c b/drivers/xen/xenbus/xenbus_client.c index 439c9dca9eee..ca744102b666 100644 --- a/drivers/xen/xenbus/xenbus_client.c +++ b/drivers/xen/xenbus/xenbus_client.c | |||
@@ -259,7 +259,6 @@ static char *error_path(struct xenbus_device *dev) | |||
259 | static void xenbus_va_dev_error(struct xenbus_device *dev, int err, | 259 | static void xenbus_va_dev_error(struct xenbus_device *dev, int err, |
260 | const char *fmt, va_list ap) | 260 | const char *fmt, va_list ap) |
261 | { | 261 | { |
262 | int ret; | ||
263 | unsigned int len; | 262 | unsigned int len; |
264 | char *printf_buffer = NULL; | 263 | char *printf_buffer = NULL; |
265 | char *path_buffer = NULL; | 264 | char *path_buffer = NULL; |
@@ -270,9 +269,7 @@ static void xenbus_va_dev_error(struct xenbus_device *dev, int err, | |||
270 | goto fail; | 269 | goto fail; |
271 | 270 | ||
272 | len = sprintf(printf_buffer, "%i ", -err); | 271 | len = sprintf(printf_buffer, "%i ", -err); |
273 | ret = vsnprintf(printf_buffer+len, PRINTF_BUFFER_SIZE-len, fmt, ap); | 272 | vsnprintf(printf_buffer+len, PRINTF_BUFFER_SIZE-len, fmt, ap); |
274 | |||
275 | BUG_ON(len + ret > PRINTF_BUFFER_SIZE-1); | ||
276 | 273 | ||
277 | dev_err(&dev->dev, "%s\n", printf_buffer); | 274 | dev_err(&dev->dev, "%s\n", printf_buffer); |
278 | 275 | ||
@@ -361,8 +358,8 @@ static void xenbus_switch_fatal(struct xenbus_device *dev, int depth, int err, | |||
361 | * @ring_mfn: mfn of ring to grant | 358 | * @ring_mfn: mfn of ring to grant |
362 | 359 | ||
363 | * Grant access to the given @ring_mfn to the peer of the given device. Return | 360 | * Grant access to the given @ring_mfn to the peer of the given device. Return |
364 | * 0 on success, or -errno on error. On error, the device will switch to | 361 | * a grant reference on success, or -errno on error. On error, the device will |
365 | * XenbusStateClosing, and the error will be saved in the store. | 362 | * switch to XenbusStateClosing, and the error will be saved in the store. |
366 | */ | 363 | */ |
367 | int xenbus_grant_ring(struct xenbus_device *dev, unsigned long ring_mfn) | 364 | int xenbus_grant_ring(struct xenbus_device *dev, unsigned long ring_mfn) |
368 | { | 365 | { |
diff --git a/drivers/xen/xenbus/xenbus_probe.c b/drivers/xen/xenbus/xenbus_probe.c index 3c0a74b3e9b1..564b31584860 100644 --- a/drivers/xen/xenbus/xenbus_probe.c +++ b/drivers/xen/xenbus/xenbus_probe.c | |||
@@ -297,9 +297,13 @@ void xenbus_dev_shutdown(struct device *_dev) | |||
297 | EXPORT_SYMBOL_GPL(xenbus_dev_shutdown); | 297 | EXPORT_SYMBOL_GPL(xenbus_dev_shutdown); |
298 | 298 | ||
299 | int xenbus_register_driver_common(struct xenbus_driver *drv, | 299 | int xenbus_register_driver_common(struct xenbus_driver *drv, |
300 | struct xen_bus_type *bus) | 300 | struct xen_bus_type *bus, |
301 | struct module *owner, const char *mod_name) | ||
301 | { | 302 | { |
303 | drv->driver.name = drv->name ? drv->name : drv->ids[0].devicetype; | ||
302 | drv->driver.bus = &bus->bus; | 304 | drv->driver.bus = &bus->bus; |
305 | drv->driver.owner = owner; | ||
306 | drv->driver.mod_name = mod_name; | ||
303 | 307 | ||
304 | return driver_register(&drv->driver); | 308 | return driver_register(&drv->driver); |
305 | } | 309 | } |
diff --git a/drivers/xen/xenbus/xenbus_probe.h b/drivers/xen/xenbus/xenbus_probe.h index 1085ec294a19..c9ec7ca1f7ab 100644 --- a/drivers/xen/xenbus/xenbus_probe.h +++ b/drivers/xen/xenbus/xenbus_probe.h | |||
@@ -60,7 +60,9 @@ extern int xenbus_match(struct device *_dev, struct device_driver *_drv); | |||
60 | extern int xenbus_dev_probe(struct device *_dev); | 60 | extern int xenbus_dev_probe(struct device *_dev); |
61 | extern int xenbus_dev_remove(struct device *_dev); | 61 | extern int xenbus_dev_remove(struct device *_dev); |
62 | extern int xenbus_register_driver_common(struct xenbus_driver *drv, | 62 | extern int xenbus_register_driver_common(struct xenbus_driver *drv, |
63 | struct xen_bus_type *bus); | 63 | struct xen_bus_type *bus, |
64 | struct module *owner, | ||
65 | const char *mod_name); | ||
64 | extern int xenbus_probe_node(struct xen_bus_type *bus, | 66 | extern int xenbus_probe_node(struct xen_bus_type *bus, |
65 | const char *type, | 67 | const char *type, |
66 | const char *nodename); | 68 | const char *nodename); |
diff --git a/drivers/xen/xenbus/xenbus_probe_backend.c b/drivers/xen/xenbus/xenbus_probe_backend.c index 5125dce11a60..04f7f85a5edf 100644 --- a/drivers/xen/xenbus/xenbus_probe_backend.c +++ b/drivers/xen/xenbus/xenbus_probe_backend.c | |||
@@ -234,13 +234,15 @@ int xenbus_dev_is_online(struct xenbus_device *dev) | |||
234 | } | 234 | } |
235 | EXPORT_SYMBOL_GPL(xenbus_dev_is_online); | 235 | EXPORT_SYMBOL_GPL(xenbus_dev_is_online); |
236 | 236 | ||
237 | int xenbus_register_backend(struct xenbus_driver *drv) | 237 | int __xenbus_register_backend(struct xenbus_driver *drv, struct module *owner, |
238 | const char *mod_name) | ||
238 | { | 239 | { |
239 | drv->read_otherend_details = read_frontend_details; | 240 | drv->read_otherend_details = read_frontend_details; |
240 | 241 | ||
241 | return xenbus_register_driver_common(drv, &xenbus_backend); | 242 | return xenbus_register_driver_common(drv, &xenbus_backend, |
243 | owner, mod_name); | ||
242 | } | 244 | } |
243 | EXPORT_SYMBOL_GPL(xenbus_register_backend); | 245 | EXPORT_SYMBOL_GPL(__xenbus_register_backend); |
244 | 246 | ||
245 | static int backend_probe_and_watch(struct notifier_block *notifier, | 247 | static int backend_probe_and_watch(struct notifier_block *notifier, |
246 | unsigned long event, | 248 | unsigned long event, |
diff --git a/drivers/xen/xenbus/xenbus_probe_frontend.c b/drivers/xen/xenbus/xenbus_probe_frontend.c index cb385c10d2b1..bcb53bdc469c 100644 --- a/drivers/xen/xenbus/xenbus_probe_frontend.c +++ b/drivers/xen/xenbus/xenbus_probe_frontend.c | |||
@@ -317,13 +317,15 @@ static void wait_for_devices(struct xenbus_driver *xendrv) | |||
317 | print_device_status); | 317 | print_device_status); |
318 | } | 318 | } |
319 | 319 | ||
320 | int xenbus_register_frontend(struct xenbus_driver *drv) | 320 | int __xenbus_register_frontend(struct xenbus_driver *drv, struct module *owner, |
321 | const char *mod_name) | ||
321 | { | 322 | { |
322 | int ret; | 323 | int ret; |
323 | 324 | ||
324 | drv->read_otherend_details = read_backend_details; | 325 | drv->read_otherend_details = read_backend_details; |
325 | 326 | ||
326 | ret = xenbus_register_driver_common(drv, &xenbus_frontend); | 327 | ret = xenbus_register_driver_common(drv, &xenbus_frontend, |
328 | owner, mod_name); | ||
327 | if (ret) | 329 | if (ret) |
328 | return ret; | 330 | return ret; |
329 | 331 | ||
@@ -332,7 +334,7 @@ int xenbus_register_frontend(struct xenbus_driver *drv) | |||
332 | 334 | ||
333 | return 0; | 335 | return 0; |
334 | } | 336 | } |
335 | EXPORT_SYMBOL_GPL(xenbus_register_frontend); | 337 | EXPORT_SYMBOL_GPL(__xenbus_register_frontend); |
336 | 338 | ||
337 | static DECLARE_WAIT_QUEUE_HEAD(backend_state_wq); | 339 | static DECLARE_WAIT_QUEUE_HEAD(backend_state_wq); |
338 | static int backend_state; | 340 | static int backend_state; |
diff --git a/include/xen/events.h b/include/xen/events.h index 8bee7a75e850..5321cd9636e6 100644 --- a/include/xen/events.h +++ b/include/xen/events.h | |||
@@ -28,6 +28,8 @@ int bind_ipi_to_irqhandler(enum ipi_vector ipi, | |||
28 | unsigned long irqflags, | 28 | unsigned long irqflags, |
29 | const char *devname, | 29 | const char *devname, |
30 | void *dev_id); | 30 | void *dev_id); |
31 | int bind_interdomain_evtchn_to_irq(unsigned int remote_domain, | ||
32 | unsigned int remote_port); | ||
31 | int bind_interdomain_evtchn_to_irqhandler(unsigned int remote_domain, | 33 | int bind_interdomain_evtchn_to_irqhandler(unsigned int remote_domain, |
32 | unsigned int remote_port, | 34 | unsigned int remote_port, |
33 | irq_handler_t handler, | 35 | irq_handler_t handler, |
diff --git a/include/xen/interface/elfnote.h b/include/xen/interface/elfnote.h index 6f4eae328ca7..f90b03454659 100644 --- a/include/xen/interface/elfnote.h +++ b/include/xen/interface/elfnote.h | |||
@@ -3,6 +3,24 @@ | |||
3 | * | 3 | * |
4 | * Definitions used for the Xen ELF notes. | 4 | * Definitions used for the Xen ELF notes. |
5 | * | 5 | * |
6 | * Permission is hereby granted, free of charge, to any person obtaining a copy | ||
7 | * of this software and associated documentation files (the "Software"), to | ||
8 | * deal in the Software without restriction, including without limitation the | ||
9 | * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or | ||
10 | * sell copies of the Software, and to permit persons to whom the Software is | ||
11 | * furnished to do so, subject to the following conditions: | ||
12 | * | ||
13 | * The above copyright notice and this permission notice shall be included in | ||
14 | * all copies or substantial portions of the Software. | ||
15 | * | ||
16 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
17 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
18 | * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE | ||
19 | * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||
20 | * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING | ||
21 | * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER | ||
22 | * DEALINGS IN THE SOFTWARE. | ||
23 | * | ||
6 | * Copyright (c) 2006, Ian Campbell, XenSource Ltd. | 24 | * Copyright (c) 2006, Ian Campbell, XenSource Ltd. |
7 | */ | 25 | */ |
8 | 26 | ||
@@ -18,12 +36,13 @@ | |||
18 | * | 36 | * |
19 | * LEGACY indicated the fields in the legacy __xen_guest string which | 37 | * LEGACY indicated the fields in the legacy __xen_guest string which |
20 | * this a note type replaces. | 38 | * this a note type replaces. |
39 | * | ||
40 | * String values (for non-legacy) are NULL terminated ASCII, also known | ||
41 | * as ASCIZ type. | ||
21 | */ | 42 | */ |
22 | 43 | ||
23 | /* | 44 | /* |
24 | * NAME=VALUE pair (string). | 45 | * NAME=VALUE pair (string). |
25 | * | ||
26 | * LEGACY: FEATURES and PAE | ||
27 | */ | 46 | */ |
28 | #define XEN_ELFNOTE_INFO 0 | 47 | #define XEN_ELFNOTE_INFO 0 |
29 | 48 | ||
@@ -137,10 +156,30 @@ | |||
137 | 156 | ||
138 | /* | 157 | /* |
139 | * Whether or not the guest supports cooperative suspend cancellation. | 158 | * Whether or not the guest supports cooperative suspend cancellation. |
159 | * This is a numeric value. | ||
160 | * | ||
161 | * Default is 0 | ||
140 | */ | 162 | */ |
141 | #define XEN_ELFNOTE_SUSPEND_CANCEL 14 | 163 | #define XEN_ELFNOTE_SUSPEND_CANCEL 14 |
142 | 164 | ||
143 | /* | 165 | /* |
166 | * The (non-default) location the initial phys-to-machine map should be | ||
167 | * placed at by the hypervisor (Dom0) or the tools (DomU). | ||
168 | * The kernel must be prepared for this mapping to be established using | ||
169 | * large pages, despite such otherwise not being available to guests. | ||
170 | * The kernel must also be able to handle the page table pages used for | ||
171 | * this mapping not being accessible through the initial mapping. | ||
172 | * (Only x86-64 supports this at present.) | ||
173 | */ | ||
174 | #define XEN_ELFNOTE_INIT_P2M 15 | ||
175 | |||
176 | /* | ||
177 | * Whether or not the guest can deal with being passed an initrd not | ||
178 | * mapped through its initial page tables. | ||
179 | */ | ||
180 | #define XEN_ELFNOTE_MOD_START_PFN 16 | ||
181 | |||
182 | /* | ||
144 | * The features supported by this kernel (numeric). | 183 | * The features supported by this kernel (numeric). |
145 | * | 184 | * |
146 | * Other than XEN_ELFNOTE_FEATURES on pre-4.2 Xen, this note allows a | 185 | * Other than XEN_ELFNOTE_FEATURES on pre-4.2 Xen, this note allows a |
@@ -153,6 +192,11 @@ | |||
153 | */ | 192 | */ |
154 | #define XEN_ELFNOTE_SUPPORTED_FEATURES 17 | 193 | #define XEN_ELFNOTE_SUPPORTED_FEATURES 17 |
155 | 194 | ||
195 | /* | ||
196 | * The number of the highest elfnote defined. | ||
197 | */ | ||
198 | #define XEN_ELFNOTE_MAX XEN_ELFNOTE_SUPPORTED_FEATURES | ||
199 | |||
156 | #endif /* __XEN_PUBLIC_ELFNOTE_H__ */ | 200 | #endif /* __XEN_PUBLIC_ELFNOTE_H__ */ |
157 | 201 | ||
158 | /* | 202 | /* |
diff --git a/include/xen/interface/io/vscsiif.h b/include/xen/interface/io/vscsiif.h new file mode 100644 index 000000000000..d07d7aca8d1c --- /dev/null +++ b/include/xen/interface/io/vscsiif.h | |||
@@ -0,0 +1,229 @@ | |||
1 | /****************************************************************************** | ||
2 | * vscsiif.h | ||
3 | * | ||
4 | * Based on the blkif.h code. | ||
5 | * | ||
6 | * Permission is hereby granted, free of charge, to any person obtaining a copy | ||
7 | * of this software and associated documentation files (the "Software"), to | ||
8 | * deal in the Software without restriction, including without limitation the | ||
9 | * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or | ||
10 | * sell copies of the Software, and to permit persons to whom the Software is | ||
11 | * furnished to do so, subject to the following conditions: | ||
12 | * | ||
13 | * The above copyright notice and this permission notice shall be included in | ||
14 | * all copies or substantial portions of the Software. | ||
15 | * | ||
16 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
17 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
18 | * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE | ||
19 | * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||
20 | * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING | ||
21 | * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER | ||
22 | * DEALINGS IN THE SOFTWARE. | ||
23 | * | ||
24 | * Copyright(c) FUJITSU Limited 2008. | ||
25 | */ | ||
26 | |||
27 | #ifndef __XEN__PUBLIC_IO_SCSI_H__ | ||
28 | #define __XEN__PUBLIC_IO_SCSI_H__ | ||
29 | |||
30 | #include "ring.h" | ||
31 | #include "../grant_table.h" | ||
32 | |||
33 | /* | ||
34 | * Feature and Parameter Negotiation | ||
35 | * ================================= | ||
36 | * The two halves of a Xen pvSCSI driver utilize nodes within the XenStore to | ||
37 | * communicate capabilities and to negotiate operating parameters. This | ||
38 | * section enumerates these nodes which reside in the respective front and | ||
39 | * backend portions of the XenStore, following the XenBus convention. | ||
40 | * | ||
41 | * Any specified default value is in effect if the corresponding XenBus node | ||
42 | * is not present in the XenStore. | ||
43 | * | ||
44 | * XenStore nodes in sections marked "PRIVATE" are solely for use by the | ||
45 | * driver side whose XenBus tree contains them. | ||
46 | * | ||
47 | ***************************************************************************** | ||
48 | * Backend XenBus Nodes | ||
49 | ***************************************************************************** | ||
50 | * | ||
51 | *------------------ Backend Device Identification (PRIVATE) ------------------ | ||
52 | * | ||
53 | * p-devname | ||
54 | * Values: string | ||
55 | * | ||
56 | * A free string used to identify the physical device (e.g. a disk name). | ||
57 | * | ||
58 | * p-dev | ||
59 | * Values: string | ||
60 | * | ||
61 | * A string specifying the backend device: either a 4-tuple "h:c:t:l" | ||
62 | * (host, controller, target, lun, all integers), or a WWN (e.g. | ||
63 | * "naa.60014054ac780582"). | ||
64 | * | ||
65 | * v-dev | ||
66 | * Values: string | ||
67 | * | ||
68 | * A string specifying the frontend device in form of a 4-tuple "h:c:t:l" | ||
69 | * (host, controller, target, lun, all integers). | ||
70 | * | ||
71 | *--------------------------------- Features --------------------------------- | ||
72 | * | ||
73 | * feature-sg-grant | ||
74 | * Values: unsigned [VSCSIIF_SG_TABLESIZE...65535] | ||
75 | * Default Value: 0 | ||
76 | * | ||
77 | * Specifies the maximum number of scatter/gather elements in grant pages | ||
78 | * supported. If not set, the backend supports up to VSCSIIF_SG_TABLESIZE | ||
79 | * SG elements specified directly in the request. | ||
80 | * | ||
81 | ***************************************************************************** | ||
82 | * Frontend XenBus Nodes | ||
83 | ***************************************************************************** | ||
84 | * | ||
85 | *----------------------- Request Transport Parameters ----------------------- | ||
86 | * | ||
87 | * event-channel | ||
88 | * Values: unsigned | ||
89 | * | ||
90 | * The identifier of the Xen event channel used to signal activity | ||
91 | * in the ring buffer. | ||
92 | * | ||
93 | * ring-ref | ||
94 | * Values: unsigned | ||
95 | * | ||
96 | * The Xen grant reference granting permission for the backend to map | ||
97 | * the sole page in a single page sized ring buffer. | ||
98 | * | ||
99 | * protocol | ||
100 | * Values: string (XEN_IO_PROTO_ABI_*) | ||
101 | * Default Value: XEN_IO_PROTO_ABI_NATIVE | ||
102 | * | ||
103 | * The machine ABI rules governing the format of all ring request and | ||
104 | * response structures. | ||
105 | */ | ||
106 | |||
107 | /* Requests from the frontend to the backend */ | ||
108 | |||
109 | /* | ||
110 | * Request a SCSI operation specified via a CDB in vscsiif_request.cmnd. | ||
111 | * The target is specified via channel, id and lun. | ||
112 | * | ||
113 | * The operation to be performed is specified via a CDB in cmnd[], the length | ||
114 | * of the CDB is in cmd_len. sc_data_direction specifies the direction of data | ||
115 | * (to the device, from the device, or none at all). | ||
116 | * | ||
117 | * If data is to be transferred to or from the device the buffer(s) in the | ||
118 | * guest memory is/are specified via one or multiple scsiif_request_segment | ||
119 | * descriptors each specifying a memory page via a grant_ref_t, a offset into | ||
120 | * the page and the length of the area in that page. All scsiif_request_segment | ||
121 | * areas concatenated form the resulting data buffer used by the operation. | ||
122 | * If the number of scsiif_request_segment areas is not too large (less than | ||
123 | * or equal VSCSIIF_SG_TABLESIZE) the areas can be specified directly in the | ||
124 | * seg[] array and the number of valid scsiif_request_segment elements is to be | ||
125 | * set in nr_segments. | ||
126 | * | ||
127 | * If "feature-sg-grant" in the Xenstore is set it is possible to specify more | ||
128 | * than VSCSIIF_SG_TABLESIZE scsiif_request_segment elements via indirection. | ||
129 | * The maximum number of allowed scsiif_request_segment elements is the value | ||
130 | * of the "feature-sg-grant" entry from Xenstore. When using indirection the | ||
131 | * seg[] array doesn't contain specifications of the data buffers, but | ||
132 | * references to scsiif_request_segment arrays, which in turn reference the | ||
133 | * data buffers. While nr_segments holds the number of populated seg[] entries | ||
134 | * (plus the set VSCSIIF_SG_GRANT bit), the number of scsiif_request_segment | ||
135 | * elements referencing the target data buffers is calculated from the lengths | ||
136 | * of the seg[] elements (the sum of all valid seg[].length divided by the | ||
137 | * size of one scsiif_request_segment structure). | ||
138 | */ | ||
139 | #define VSCSIIF_ACT_SCSI_CDB 1 | ||
140 | |||
141 | /* | ||
142 | * Request abort of a running operation for the specified target given by | ||
143 | * channel, id, lun and the operation's rqid in ref_rqid. | ||
144 | */ | ||
145 | #define VSCSIIF_ACT_SCSI_ABORT 2 | ||
146 | |||
147 | /* | ||
148 | * Request a device reset of the specified target (channel and id). | ||
149 | */ | ||
150 | #define VSCSIIF_ACT_SCSI_RESET 3 | ||
151 | |||
152 | /* | ||
153 | * Preset scatter/gather elements for a following request. Deprecated. | ||
154 | * Keeping the define only to avoid usage of the value "4" for other actions. | ||
155 | */ | ||
156 | #define VSCSIIF_ACT_SCSI_SG_PRESET 4 | ||
157 | |||
158 | /* | ||
159 | * Maximum scatter/gather segments per request. | ||
160 | * | ||
161 | * Considering balance between allocating at least 16 "vscsiif_request" | ||
162 | * structures on one page (4096 bytes) and the number of scatter/gather | ||
163 | * elements needed, we decided to use 26 as a magic number. | ||
164 | * | ||
165 | * If "feature-sg-grant" is set, more scatter/gather elements can be specified | ||
166 | * by placing them in one or more (up to VSCSIIF_SG_TABLESIZE) granted pages. | ||
167 | * In this case the vscsiif_request seg elements don't contain references to | ||
168 | * the user data, but to the SG elements referencing the user data. | ||
169 | */ | ||
170 | #define VSCSIIF_SG_TABLESIZE 26 | ||
171 | |||
172 | /* | ||
173 | * based on Linux kernel 2.6.18, still valid | ||
174 | * Changing these values requires support of multiple protocols via the rings | ||
175 | * as "old clients" will blindly use these values and the resulting structure | ||
176 | * sizes. | ||
177 | */ | ||
178 | #define VSCSIIF_MAX_COMMAND_SIZE 16 | ||
179 | #define VSCSIIF_SENSE_BUFFERSIZE 96 | ||
180 | |||
181 | struct scsiif_request_segment { | ||
182 | grant_ref_t gref; | ||
183 | uint16_t offset; | ||
184 | uint16_t length; | ||
185 | }; | ||
186 | |||
187 | #define VSCSIIF_SG_PER_PAGE (PAGE_SIZE / sizeof(struct scsiif_request_segment)) | ||
188 | |||
189 | /* Size of one request is 252 bytes */ | ||
190 | struct vscsiif_request { | ||
191 | uint16_t rqid; /* private guest value, echoed in resp */ | ||
192 | uint8_t act; /* command between backend and frontend */ | ||
193 | uint8_t cmd_len; /* valid CDB bytes */ | ||
194 | |||
195 | uint8_t cmnd[VSCSIIF_MAX_COMMAND_SIZE]; /* the CDB */ | ||
196 | uint16_t timeout_per_command; /* deprecated */ | ||
197 | uint16_t channel, id, lun; /* (virtual) device specification */ | ||
198 | uint16_t ref_rqid; /* command abort reference */ | ||
199 | uint8_t sc_data_direction; /* for DMA_TO_DEVICE(1) | ||
200 | DMA_FROM_DEVICE(2) | ||
201 | DMA_NONE(3) requests */ | ||
202 | uint8_t nr_segments; /* Number of pieces of scatter-gather */ | ||
203 | /* | ||
204 | * flag in nr_segments: SG elements via grant page | ||
205 | * | ||
206 | * If VSCSIIF_SG_GRANT is set, the low 7 bits of nr_segments specify the number | ||
207 | * of grant pages containing SG elements. Usable if "feature-sg-grant" set. | ||
208 | */ | ||
209 | #define VSCSIIF_SG_GRANT 0x80 | ||
210 | |||
211 | struct scsiif_request_segment seg[VSCSIIF_SG_TABLESIZE]; | ||
212 | uint32_t reserved[3]; | ||
213 | }; | ||
214 | |||
215 | /* Size of one response is 252 bytes */ | ||
216 | struct vscsiif_response { | ||
217 | uint16_t rqid; /* identifies request */ | ||
218 | uint8_t padding; | ||
219 | uint8_t sense_len; | ||
220 | uint8_t sense_buffer[VSCSIIF_SENSE_BUFFERSIZE]; | ||
221 | int32_t rslt; | ||
222 | uint32_t residual_len; /* request bufflen - | ||
223 | return the value from physical device */ | ||
224 | uint32_t reserved[36]; | ||
225 | }; | ||
226 | |||
227 | DEFINE_RING_TYPES(vscsiif, struct vscsiif_request, struct vscsiif_response); | ||
228 | |||
229 | #endif /*__XEN__PUBLIC_IO_SCSI_H__*/ | ||
diff --git a/include/xen/interface/xen.h b/include/xen/interface/xen.h index de082130ba4b..f68719f405af 100644 --- a/include/xen/interface/xen.h +++ b/include/xen/interface/xen.h | |||
@@ -3,6 +3,24 @@ | |||
3 | * | 3 | * |
4 | * Guest OS interface to Xen. | 4 | * Guest OS interface to Xen. |
5 | * | 5 | * |
6 | * Permission is hereby granted, free of charge, to any person obtaining a copy | ||
7 | * of this software and associated documentation files (the "Software"), to | ||
8 | * deal in the Software without restriction, including without limitation the | ||
9 | * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or | ||
10 | * sell copies of the Software, and to permit persons to whom the Software is | ||
11 | * furnished to do so, subject to the following conditions: | ||
12 | * | ||
13 | * The above copyright notice and this permission notice shall be included in | ||
14 | * all copies or substantial portions of the Software. | ||
15 | * | ||
16 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
17 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
18 | * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE | ||
19 | * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||
20 | * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING | ||
21 | * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER | ||
22 | * DEALINGS IN THE SOFTWARE. | ||
23 | * | ||
6 | * Copyright (c) 2004, K A Fraser | 24 | * Copyright (c) 2004, K A Fraser |
7 | */ | 25 | */ |
8 | 26 | ||
@@ -73,13 +91,23 @@ | |||
73 | * VIRTUAL INTERRUPTS | 91 | * VIRTUAL INTERRUPTS |
74 | * | 92 | * |
75 | * Virtual interrupts that a guest OS may receive from Xen. | 93 | * Virtual interrupts that a guest OS may receive from Xen. |
94 | * In the side comments, 'V.' denotes a per-VCPU VIRQ while 'G.' denotes a | ||
95 | * global VIRQ. The former can be bound once per VCPU and cannot be re-bound. | ||
96 | * The latter can be allocated only once per guest: they must initially be | ||
97 | * allocated to VCPU0 but can subsequently be re-bound. | ||
76 | */ | 98 | */ |
77 | #define VIRQ_TIMER 0 /* Timebase update, and/or requested timeout. */ | 99 | #define VIRQ_TIMER 0 /* V. Timebase update, and/or requested timeout. */ |
78 | #define VIRQ_DEBUG 1 /* Request guest to dump debug info. */ | 100 | #define VIRQ_DEBUG 1 /* V. Request guest to dump debug info. */ |
79 | #define VIRQ_CONSOLE 2 /* (DOM0) Bytes received on emergency console. */ | 101 | #define VIRQ_CONSOLE 2 /* G. (DOM0) Bytes received on emergency console. */ |
80 | #define VIRQ_DOM_EXC 3 /* (DOM0) Exceptional event for some domain. */ | 102 | #define VIRQ_DOM_EXC 3 /* G. (DOM0) Exceptional event for some domain. */ |
81 | #define VIRQ_DEBUGGER 6 /* (DOM0) A domain has paused for debugging. */ | 103 | #define VIRQ_TBUF 4 /* G. (DOM0) Trace buffer has records available. */ |
82 | #define VIRQ_PCPU_STATE 9 /* (DOM0) PCPU state changed */ | 104 | #define VIRQ_DEBUGGER 6 /* G. (DOM0) A domain has paused for debugging. */ |
105 | #define VIRQ_XENOPROF 7 /* V. XenOprofile interrupt: new sample available */ | ||
106 | #define VIRQ_CON_RING 8 /* G. (DOM0) Bytes received on console */ | ||
107 | #define VIRQ_PCPU_STATE 9 /* G. (DOM0) PCPU state changed */ | ||
108 | #define VIRQ_MEM_EVENT 10 /* G. (DOM0) A memory event has occured */ | ||
109 | #define VIRQ_XC_RESERVED 11 /* G. Reserved for XenClient */ | ||
110 | #define VIRQ_ENOMEM 12 /* G. (DOM0) Low on heap memory */ | ||
83 | 111 | ||
84 | /* Architecture-specific VIRQ definitions. */ | 112 | /* Architecture-specific VIRQ definitions. */ |
85 | #define VIRQ_ARCH_0 16 | 113 | #define VIRQ_ARCH_0 16 |
@@ -92,24 +120,68 @@ | |||
92 | #define VIRQ_ARCH_7 23 | 120 | #define VIRQ_ARCH_7 23 |
93 | 121 | ||
94 | #define NR_VIRQS 24 | 122 | #define NR_VIRQS 24 |
123 | |||
95 | /* | 124 | /* |
96 | * MMU-UPDATE REQUESTS | 125 | * enum neg_errnoval HYPERVISOR_mmu_update(const struct mmu_update reqs[], |
97 | * | 126 | * unsigned count, unsigned *done_out, |
98 | * HYPERVISOR_mmu_update() accepts a list of (ptr, val) pairs. | 127 | * unsigned foreigndom) |
99 | * A foreigndom (FD) can be specified (or DOMID_SELF for none). | 128 | * @reqs is an array of mmu_update_t structures ((ptr, val) pairs). |
100 | * Where the FD has some effect, it is described below. | 129 | * @count is the length of the above array. |
101 | * ptr[1:0] specifies the appropriate MMU_* command. | 130 | * @pdone is an output parameter indicating number of completed operations |
131 | * @foreigndom[15:0]: FD, the expected owner of data pages referenced in this | ||
132 | * hypercall invocation. Can be DOMID_SELF. | ||
133 | * @foreigndom[31:16]: PFD, the expected owner of pagetable pages referenced | ||
134 | * in this hypercall invocation. The value of this field | ||
135 | * (x) encodes the PFD as follows: | ||
136 | * x == 0 => PFD == DOMID_SELF | ||
137 | * x != 0 => PFD == x - 1 | ||
102 | * | 138 | * |
139 | * Sub-commands: ptr[1:0] specifies the appropriate MMU_* command. | ||
140 | * ------------- | ||
103 | * ptr[1:0] == MMU_NORMAL_PT_UPDATE: | 141 | * ptr[1:0] == MMU_NORMAL_PT_UPDATE: |
104 | * Updates an entry in a page table. If updating an L1 table, and the new | 142 | * Updates an entry in a page table belonging to PFD. If updating an L1 table, |
105 | * table entry is valid/present, the mapped frame must belong to the FD, if | 143 | * and the new table entry is valid/present, the mapped frame must belong to |
106 | * an FD has been specified. If attempting to map an I/O page then the | 144 | * FD. If attempting to map an I/O page then the caller assumes the privilege |
107 | * caller assumes the privilege of the FD. | 145 | * of the FD. |
108 | * FD == DOMID_IO: Permit /only/ I/O mappings, at the priv level of the caller. | 146 | * FD == DOMID_IO: Permit /only/ I/O mappings, at the priv level of the caller. |
109 | * FD == DOMID_XEN: Map restricted areas of Xen's heap space. | 147 | * FD == DOMID_XEN: Map restricted areas of Xen's heap space. |
110 | * ptr[:2] -- Machine address of the page-table entry to modify. | 148 | * ptr[:2] -- Machine address of the page-table entry to modify. |
111 | * val -- Value to write. | 149 | * val -- Value to write. |
112 | * | 150 | * |
151 | * There also certain implicit requirements when using this hypercall. The | ||
152 | * pages that make up a pagetable must be mapped read-only in the guest. | ||
153 | * This prevents uncontrolled guest updates to the pagetable. Xen strictly | ||
154 | * enforces this, and will disallow any pagetable update which will end up | ||
155 | * mapping pagetable page RW, and will disallow using any writable page as a | ||
156 | * pagetable. In practice it means that when constructing a page table for a | ||
157 | * process, thread, etc, we MUST be very dilligient in following these rules: | ||
158 | * 1). Start with top-level page (PGD or in Xen language: L4). Fill out | ||
159 | * the entries. | ||
160 | * 2). Keep on going, filling out the upper (PUD or L3), and middle (PMD | ||
161 | * or L2). | ||
162 | * 3). Start filling out the PTE table (L1) with the PTE entries. Once | ||
163 | * done, make sure to set each of those entries to RO (so writeable bit | ||
164 | * is unset). Once that has been completed, set the PMD (L2) for this | ||
165 | * PTE table as RO. | ||
166 | * 4). When completed with all of the PMD (L2) entries, and all of them have | ||
167 | * been set to RO, make sure to set RO the PUD (L3). Do the same | ||
168 | * operation on PGD (L4) pagetable entries that have a PUD (L3) entry. | ||
169 | * 5). Now before you can use those pages (so setting the cr3), you MUST also | ||
170 | * pin them so that the hypervisor can verify the entries. This is done | ||
171 | * via the HYPERVISOR_mmuext_op(MMUEXT_PIN_L4_TABLE, guest physical frame | ||
172 | * number of the PGD (L4)). And this point the HYPERVISOR_mmuext_op( | ||
173 | * MMUEXT_NEW_BASEPTR, guest physical frame number of the PGD (L4)) can be | ||
174 | * issued. | ||
175 | * For 32-bit guests, the L4 is not used (as there is less pagetables), so | ||
176 | * instead use L3. | ||
177 | * At this point the pagetables can be modified using the MMU_NORMAL_PT_UPDATE | ||
178 | * hypercall. Also if so desired the OS can also try to write to the PTE | ||
179 | * and be trapped by the hypervisor (as the PTE entry is RO). | ||
180 | * | ||
181 | * To deallocate the pages, the operations are the reverse of the steps | ||
182 | * mentioned above. The argument is MMUEXT_UNPIN_TABLE for all levels and the | ||
183 | * pagetable MUST not be in use (meaning that the cr3 is not set to it). | ||
184 | * | ||
113 | * ptr[1:0] == MMU_MACHPHYS_UPDATE: | 185 | * ptr[1:0] == MMU_MACHPHYS_UPDATE: |
114 | * Updates an entry in the machine->pseudo-physical mapping table. | 186 | * Updates an entry in the machine->pseudo-physical mapping table. |
115 | * ptr[:2] -- Machine address within the frame whose mapping to modify. | 187 | * ptr[:2] -- Machine address within the frame whose mapping to modify. |
@@ -119,6 +191,72 @@ | |||
119 | * ptr[1:0] == MMU_PT_UPDATE_PRESERVE_AD: | 191 | * ptr[1:0] == MMU_PT_UPDATE_PRESERVE_AD: |
120 | * As MMU_NORMAL_PT_UPDATE above, but A/D bits currently in the PTE are ORed | 192 | * As MMU_NORMAL_PT_UPDATE above, but A/D bits currently in the PTE are ORed |
121 | * with those in @val. | 193 | * with those in @val. |
194 | * | ||
195 | * @val is usually the machine frame number along with some attributes. | ||
196 | * The attributes by default follow the architecture defined bits. Meaning that | ||
197 | * if this is a X86_64 machine and four page table layout is used, the layout | ||
198 | * of val is: | ||
199 | * - 63 if set means No execute (NX) | ||
200 | * - 46-13 the machine frame number | ||
201 | * - 12 available for guest | ||
202 | * - 11 available for guest | ||
203 | * - 10 available for guest | ||
204 | * - 9 available for guest | ||
205 | * - 8 global | ||
206 | * - 7 PAT (PSE is disabled, must use hypercall to make 4MB or 2MB pages) | ||
207 | * - 6 dirty | ||
208 | * - 5 accessed | ||
209 | * - 4 page cached disabled | ||
210 | * - 3 page write through | ||
211 | * - 2 userspace accessible | ||
212 | * - 1 writeable | ||
213 | * - 0 present | ||
214 | * | ||
215 | * The one bits that does not fit with the default layout is the PAGE_PSE | ||
216 | * also called PAGE_PAT). The MMUEXT_[UN]MARK_SUPER arguments to the | ||
217 | * HYPERVISOR_mmuext_op serve as mechanism to set a pagetable to be 4MB | ||
218 | * (or 2MB) instead of using the PAGE_PSE bit. | ||
219 | * | ||
220 | * The reason that the PAGE_PSE (bit 7) is not being utilized is due to Xen | ||
221 | * using it as the Page Attribute Table (PAT) bit - for details on it please | ||
222 | * refer to Intel SDM 10.12. The PAT allows to set the caching attributes of | ||
223 | * pages instead of using MTRRs. | ||
224 | * | ||
225 | * The PAT MSR is as follows (it is a 64-bit value, each entry is 8 bits): | ||
226 | * PAT4 PAT0 | ||
227 | * +-----+-----+----+----+----+-----+----+----+ | ||
228 | * | UC | UC- | WC | WB | UC | UC- | WC | WB | <= Linux | ||
229 | * +-----+-----+----+----+----+-----+----+----+ | ||
230 | * | UC | UC- | WT | WB | UC | UC- | WT | WB | <= BIOS (default when machine boots) | ||
231 | * +-----+-----+----+----+----+-----+----+----+ | ||
232 | * | rsv | rsv | WP | WC | UC | UC- | WT | WB | <= Xen | ||
233 | * +-----+-----+----+----+----+-----+----+----+ | ||
234 | * | ||
235 | * The lookup of this index table translates to looking up | ||
236 | * Bit 7, Bit 4, and Bit 3 of val entry: | ||
237 | * | ||
238 | * PAT/PSE (bit 7) ... PCD (bit 4) .. PWT (bit 3). | ||
239 | * | ||
240 | * If all bits are off, then we are using PAT0. If bit 3 turned on, | ||
241 | * then we are using PAT1, if bit 3 and bit 4, then PAT2.. | ||
242 | * | ||
243 | * As you can see, the Linux PAT1 translates to PAT4 under Xen. Which means | ||
244 | * that if a guest that follows Linux's PAT setup and would like to set Write | ||
245 | * Combined on pages it MUST use PAT4 entry. Meaning that Bit 7 (PAGE_PAT) is | ||
246 | * set. For example, under Linux it only uses PAT0, PAT1, and PAT2 for the | ||
247 | * caching as: | ||
248 | * | ||
249 | * WB = none (so PAT0) | ||
250 | * WC = PWT (bit 3 on) | ||
251 | * UC = PWT | PCD (bit 3 and 4 are on). | ||
252 | * | ||
253 | * To make it work with Xen, it needs to translate the WC bit as so: | ||
254 | * | ||
255 | * PWT (so bit 3 on) --> PAT (so bit 7 is on) and clear bit 3 | ||
256 | * | ||
257 | * And to translate back it would: | ||
258 | * | ||
259 | * PAT (bit 7 on) --> PWT (bit 3 on) and clear bit 7. | ||
122 | */ | 260 | */ |
123 | #define MMU_NORMAL_PT_UPDATE 0 /* checked '*ptr = val'. ptr is MA. */ | 261 | #define MMU_NORMAL_PT_UPDATE 0 /* checked '*ptr = val'. ptr is MA. */ |
124 | #define MMU_MACHPHYS_UPDATE 1 /* ptr = MA of frame to modify entry for */ | 262 | #define MMU_MACHPHYS_UPDATE 1 /* ptr = MA of frame to modify entry for */ |
@@ -127,7 +265,12 @@ | |||
127 | /* | 265 | /* |
128 | * MMU EXTENDED OPERATIONS | 266 | * MMU EXTENDED OPERATIONS |
129 | * | 267 | * |
130 | * HYPERVISOR_mmuext_op() accepts a list of mmuext_op structures. | 268 | * enum neg_errnoval HYPERVISOR_mmuext_op(mmuext_op_t uops[], |
269 | * unsigned int count, | ||
270 | * unsigned int *pdone, | ||
271 | * unsigned int foreigndom) | ||
272 | */ | ||
273 | /* HYPERVISOR_mmuext_op() accepts a list of mmuext_op structures. | ||
131 | * A foreigndom (FD) can be specified (or DOMID_SELF for none). | 274 | * A foreigndom (FD) can be specified (or DOMID_SELF for none). |
132 | * Where the FD has some effect, it is described below. | 275 | * Where the FD has some effect, it is described below. |
133 | * | 276 | * |
@@ -164,9 +307,23 @@ | |||
164 | * cmd: MMUEXT_FLUSH_CACHE | 307 | * cmd: MMUEXT_FLUSH_CACHE |
165 | * No additional arguments. Writes back and flushes cache contents. | 308 | * No additional arguments. Writes back and flushes cache contents. |
166 | * | 309 | * |
310 | * cmd: MMUEXT_FLUSH_CACHE_GLOBAL | ||
311 | * No additional arguments. Writes back and flushes cache contents | ||
312 | * on all CPUs in the system. | ||
313 | * | ||
167 | * cmd: MMUEXT_SET_LDT | 314 | * cmd: MMUEXT_SET_LDT |
168 | * linear_addr: Linear address of LDT base (NB. must be page-aligned). | 315 | * linear_addr: Linear address of LDT base (NB. must be page-aligned). |
169 | * nr_ents: Number of entries in LDT. | 316 | * nr_ents: Number of entries in LDT. |
317 | * | ||
318 | * cmd: MMUEXT_CLEAR_PAGE | ||
319 | * mfn: Machine frame number to be cleared. | ||
320 | * | ||
321 | * cmd: MMUEXT_COPY_PAGE | ||
322 | * mfn: Machine frame number of the destination page. | ||
323 | * src_mfn: Machine frame number of the source page. | ||
324 | * | ||
325 | * cmd: MMUEXT_[UN]MARK_SUPER | ||
326 | * mfn: Machine frame number of head of superpage to be [un]marked. | ||
170 | */ | 327 | */ |
171 | #define MMUEXT_PIN_L1_TABLE 0 | 328 | #define MMUEXT_PIN_L1_TABLE 0 |
172 | #define MMUEXT_PIN_L2_TABLE 1 | 329 | #define MMUEXT_PIN_L2_TABLE 1 |
@@ -183,12 +340,18 @@ | |||
183 | #define MMUEXT_FLUSH_CACHE 12 | 340 | #define MMUEXT_FLUSH_CACHE 12 |
184 | #define MMUEXT_SET_LDT 13 | 341 | #define MMUEXT_SET_LDT 13 |
185 | #define MMUEXT_NEW_USER_BASEPTR 15 | 342 | #define MMUEXT_NEW_USER_BASEPTR 15 |
343 | #define MMUEXT_CLEAR_PAGE 16 | ||
344 | #define MMUEXT_COPY_PAGE 17 | ||
345 | #define MMUEXT_FLUSH_CACHE_GLOBAL 18 | ||
346 | #define MMUEXT_MARK_SUPER 19 | ||
347 | #define MMUEXT_UNMARK_SUPER 20 | ||
186 | 348 | ||
187 | #ifndef __ASSEMBLY__ | 349 | #ifndef __ASSEMBLY__ |
188 | struct mmuext_op { | 350 | struct mmuext_op { |
189 | unsigned int cmd; | 351 | unsigned int cmd; |
190 | union { | 352 | union { |
191 | /* [UN]PIN_TABLE, NEW_BASEPTR, NEW_USER_BASEPTR */ | 353 | /* [UN]PIN_TABLE, NEW_BASEPTR, NEW_USER_BASEPTR |
354 | * CLEAR_PAGE, COPY_PAGE, [UN]MARK_SUPER */ | ||
192 | xen_pfn_t mfn; | 355 | xen_pfn_t mfn; |
193 | /* INVLPG_LOCAL, INVLPG_ALL, SET_LDT */ | 356 | /* INVLPG_LOCAL, INVLPG_ALL, SET_LDT */ |
194 | unsigned long linear_addr; | 357 | unsigned long linear_addr; |
@@ -198,6 +361,8 @@ struct mmuext_op { | |||
198 | unsigned int nr_ents; | 361 | unsigned int nr_ents; |
199 | /* TLB_FLUSH_MULTI, INVLPG_MULTI */ | 362 | /* TLB_FLUSH_MULTI, INVLPG_MULTI */ |
200 | void *vcpumask; | 363 | void *vcpumask; |
364 | /* COPY_PAGE */ | ||
365 | xen_pfn_t src_mfn; | ||
201 | } arg2; | 366 | } arg2; |
202 | }; | 367 | }; |
203 | DEFINE_GUEST_HANDLE_STRUCT(mmuext_op); | 368 | DEFINE_GUEST_HANDLE_STRUCT(mmuext_op); |
@@ -225,10 +390,23 @@ DEFINE_GUEST_HANDLE_STRUCT(mmuext_op); | |||
225 | */ | 390 | */ |
226 | #define VMASST_CMD_enable 0 | 391 | #define VMASST_CMD_enable 0 |
227 | #define VMASST_CMD_disable 1 | 392 | #define VMASST_CMD_disable 1 |
393 | |||
394 | /* x86/32 guests: simulate full 4GB segment limits. */ | ||
228 | #define VMASST_TYPE_4gb_segments 0 | 395 | #define VMASST_TYPE_4gb_segments 0 |
396 | |||
397 | /* x86/32 guests: trap (vector 15) whenever above vmassist is used. */ | ||
229 | #define VMASST_TYPE_4gb_segments_notify 1 | 398 | #define VMASST_TYPE_4gb_segments_notify 1 |
399 | |||
400 | /* | ||
401 | * x86 guests: support writes to bottom-level PTEs. | ||
402 | * NB1. Page-directory entries cannot be written. | ||
403 | * NB2. Guest must continue to remove all writable mappings of PTEs. | ||
404 | */ | ||
230 | #define VMASST_TYPE_writable_pagetables 2 | 405 | #define VMASST_TYPE_writable_pagetables 2 |
406 | |||
407 | /* x86/PAE guests: support PDPTs above 4GB. */ | ||
231 | #define VMASST_TYPE_pae_extended_cr3 3 | 408 | #define VMASST_TYPE_pae_extended_cr3 3 |
409 | |||
232 | #define MAX_VMASST_TYPE 3 | 410 | #define MAX_VMASST_TYPE 3 |
233 | 411 | ||
234 | #ifndef __ASSEMBLY__ | 412 | #ifndef __ASSEMBLY__ |
@@ -260,6 +438,15 @@ typedef uint16_t domid_t; | |||
260 | */ | 438 | */ |
261 | #define DOMID_XEN (0x7FF2U) | 439 | #define DOMID_XEN (0x7FF2U) |
262 | 440 | ||
441 | /* DOMID_COW is used as the owner of sharable pages */ | ||
442 | #define DOMID_COW (0x7FF3U) | ||
443 | |||
444 | /* DOMID_INVALID is used to identify pages with unknown owner. */ | ||
445 | #define DOMID_INVALID (0x7FF4U) | ||
446 | |||
447 | /* Idle domain. */ | ||
448 | #define DOMID_IDLE (0x7FFFU) | ||
449 | |||
263 | /* | 450 | /* |
264 | * Send an array of these to HYPERVISOR_mmu_update(). | 451 | * Send an array of these to HYPERVISOR_mmu_update(). |
265 | * NB. The fields are natural pointer/address size for this architecture. | 452 | * NB. The fields are natural pointer/address size for this architecture. |
@@ -272,7 +459,9 @@ DEFINE_GUEST_HANDLE_STRUCT(mmu_update); | |||
272 | 459 | ||
273 | /* | 460 | /* |
274 | * Send an array of these to HYPERVISOR_multicall(). | 461 | * Send an array of these to HYPERVISOR_multicall(). |
275 | * NB. The fields are natural register size for this architecture. | 462 | * NB. The fields are logically the natural register size for this |
463 | * architecture. In cases where xen_ulong_t is larger than this then | ||
464 | * any unused bits in the upper portion must be zero. | ||
276 | */ | 465 | */ |
277 | struct multicall_entry { | 466 | struct multicall_entry { |
278 | xen_ulong_t op; | 467 | xen_ulong_t op; |
@@ -442,8 +631,48 @@ struct start_info { | |||
442 | unsigned long mod_start; /* VIRTUAL address of pre-loaded module. */ | 631 | unsigned long mod_start; /* VIRTUAL address of pre-loaded module. */ |
443 | unsigned long mod_len; /* Size (bytes) of pre-loaded module. */ | 632 | unsigned long mod_len; /* Size (bytes) of pre-loaded module. */ |
444 | int8_t cmd_line[MAX_GUEST_CMDLINE]; | 633 | int8_t cmd_line[MAX_GUEST_CMDLINE]; |
634 | /* The pfn range here covers both page table and p->m table frames. */ | ||
635 | unsigned long first_p2m_pfn;/* 1st pfn forming initial P->M table. */ | ||
636 | unsigned long nr_p2m_frames;/* # of pfns forming initial P->M table. */ | ||
445 | }; | 637 | }; |
446 | 638 | ||
639 | /* These flags are passed in the 'flags' field of start_info_t. */ | ||
640 | #define SIF_PRIVILEGED (1<<0) /* Is the domain privileged? */ | ||
641 | #define SIF_INITDOMAIN (1<<1) /* Is this the initial control domain? */ | ||
642 | #define SIF_MULTIBOOT_MOD (1<<2) /* Is mod_start a multiboot module? */ | ||
643 | #define SIF_MOD_START_PFN (1<<3) /* Is mod_start a PFN? */ | ||
644 | #define SIF_PM_MASK (0xFF<<8) /* reserve 1 byte for xen-pm options */ | ||
645 | |||
646 | /* | ||
647 | * A multiboot module is a package containing modules very similar to a | ||
648 | * multiboot module array. The only differences are: | ||
649 | * - the array of module descriptors is by convention simply at the beginning | ||
650 | * of the multiboot module, | ||
651 | * - addresses in the module descriptors are based on the beginning of the | ||
652 | * multiboot module, | ||
653 | * - the number of modules is determined by a termination descriptor that has | ||
654 | * mod_start == 0. | ||
655 | * | ||
656 | * This permits to both build it statically and reference it in a configuration | ||
657 | * file, and let the PV guest easily rebase the addresses to virtual addresses | ||
658 | * and at the same time count the number of modules. | ||
659 | */ | ||
660 | struct xen_multiboot_mod_list { | ||
661 | /* Address of first byte of the module */ | ||
662 | uint32_t mod_start; | ||
663 | /* Address of last byte of the module (inclusive) */ | ||
664 | uint32_t mod_end; | ||
665 | /* Address of zero-terminated command line */ | ||
666 | uint32_t cmdline; | ||
667 | /* Unused, must be zero */ | ||
668 | uint32_t pad; | ||
669 | }; | ||
670 | /* | ||
671 | * The console structure in start_info.console.dom0 | ||
672 | * | ||
673 | * This structure includes a variety of information required to | ||
674 | * have a working VGA/VESA console. | ||
675 | */ | ||
447 | struct dom0_vga_console_info { | 676 | struct dom0_vga_console_info { |
448 | uint8_t video_type; | 677 | uint8_t video_type; |
449 | #define XEN_VGATYPE_TEXT_MODE_3 0x03 | 678 | #define XEN_VGATYPE_TEXT_MODE_3 0x03 |
@@ -484,11 +713,6 @@ struct dom0_vga_console_info { | |||
484 | } u; | 713 | } u; |
485 | }; | 714 | }; |
486 | 715 | ||
487 | /* These flags are passed in the 'flags' field of start_info_t. */ | ||
488 | #define SIF_PRIVILEGED (1<<0) /* Is the domain privileged? */ | ||
489 | #define SIF_INITDOMAIN (1<<1) /* Is this the initial control domain? */ | ||
490 | #define SIF_PM_MASK (0xFF<<8) /* reserve 1 byte for xen-pm options */ | ||
491 | |||
492 | typedef uint64_t cpumap_t; | 716 | typedef uint64_t cpumap_t; |
493 | 717 | ||
494 | typedef uint8_t xen_domain_handle_t[16]; | 718 | typedef uint8_t xen_domain_handle_t[16]; |
diff --git a/include/xen/xenbus.h b/include/xen/xenbus.h index 0324c6d340c1..b78f21caf55a 100644 --- a/include/xen/xenbus.h +++ b/include/xen/xenbus.h | |||
@@ -86,6 +86,7 @@ struct xenbus_device_id | |||
86 | 86 | ||
87 | /* A xenbus driver. */ | 87 | /* A xenbus driver. */ |
88 | struct xenbus_driver { | 88 | struct xenbus_driver { |
89 | const char *name; /* defaults to ids[0].devicetype */ | ||
89 | const struct xenbus_device_id *ids; | 90 | const struct xenbus_device_id *ids; |
90 | int (*probe)(struct xenbus_device *dev, | 91 | int (*probe)(struct xenbus_device *dev, |
91 | const struct xenbus_device_id *id); | 92 | const struct xenbus_device_id *id); |
@@ -100,20 +101,22 @@ struct xenbus_driver { | |||
100 | int (*is_ready)(struct xenbus_device *dev); | 101 | int (*is_ready)(struct xenbus_device *dev); |
101 | }; | 102 | }; |
102 | 103 | ||
103 | #define DEFINE_XENBUS_DRIVER(var, drvname, methods...) \ | ||
104 | struct xenbus_driver var ## _driver = { \ | ||
105 | .driver.name = drvname + 0 ?: var ## _ids->devicetype, \ | ||
106 | .driver.owner = THIS_MODULE, \ | ||
107 | .ids = var ## _ids, ## methods \ | ||
108 | } | ||
109 | |||
110 | static inline struct xenbus_driver *to_xenbus_driver(struct device_driver *drv) | 104 | static inline struct xenbus_driver *to_xenbus_driver(struct device_driver *drv) |
111 | { | 105 | { |
112 | return container_of(drv, struct xenbus_driver, driver); | 106 | return container_of(drv, struct xenbus_driver, driver); |
113 | } | 107 | } |
114 | 108 | ||
115 | int __must_check xenbus_register_frontend(struct xenbus_driver *); | 109 | int __must_check __xenbus_register_frontend(struct xenbus_driver *drv, |
116 | int __must_check xenbus_register_backend(struct xenbus_driver *); | 110 | struct module *owner, |
111 | const char *mod_name); | ||
112 | int __must_check __xenbus_register_backend(struct xenbus_driver *drv, | ||
113 | struct module *owner, | ||
114 | const char *mod_name); | ||
115 | |||
116 | #define xenbus_register_frontend(drv) \ | ||
117 | __xenbus_register_frontend(drv, THIS_MODULE, KBUILD_MODNAME); | ||
118 | #define xenbus_register_backend(drv) \ | ||
119 | __xenbus_register_backend(drv, THIS_MODULE, KBUILD_MODNAME); | ||
117 | 120 | ||
118 | void xenbus_unregister_driver(struct xenbus_driver *drv); | 121 | void xenbus_unregister_driver(struct xenbus_driver *drv); |
119 | 122 | ||